Ray
Keras
Ray | Keras | |
---|---|---|
44 | 80 | |
31,718 | 61,179 | |
1.9% | 0.4% | |
10.0 | 9.9 | |
2 days ago | 3 days ago | |
Python | Python | |
Apache License 2.0 | Apache License 2.0 |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
Ray
- Comparison: Dask vs. Ray
- Ray: Unified framework for scaling AI and Python applications
-
Open Source Advent Fun Wraps Up!
22. Ray | Github | tutorial
-
Fine-Tuning Llama-2: A Comprehensive Case Study for Tailoring Custom Models
Training times for GSM8k are mentioned here: https://github.com/ray-project/ray/tree/master/doc/source/te...
- Ray – an open source project for scaling AI workloads
-
Methods to keep agents inside grid world.
Here's a reference from RLlib that points to docs and an example, and here's one from one of my projects that includes all my own implementations
-
TransformerXL + PPO Baseline + MemoryGym
RLlib
- Is dynamic action masking possible in Rllib?
-
AWS re:Invent 2022 Recap | Data & Analytics services
⦿ AWS Glue Data Quality - Automatic data quality rule recommendations based on your data AWS Glue for Ray - Data integration with Ray (ray.io), a popular new open-source compute framework that helps you scale Python workloads
-
Think about it for a second
https://ray.io (just dropping the link)
Keras
-
Using Google Magika to build an AI-powered file type detector
The core model architecture for Magika was implemented using Keras, a popular open source deep learning framework that enables Google researchers to experiment quickly with new models.
-
Side Quest #3: maybe the real Deepfakes were the friends we made along the way
def batcher_from_directory(batch_size:int, dataset_path:str, shuffle=False,seed=None) -> tf.data.Dataset: """ Return a tensorflow Dataset object that returns images and spectrograms as required. Partly inspired by https://github.com/keras-team/keras/blob/v3.3.3/keras/src/utils/image_dataset_utils.py Args: batch_size: The batch size. dataset_path: The path to the dataset folder which must contain the image folder and audio folder. shuffle: Whether to shuffle the dataset. Default to False. seed: The seed for the shuffle. Default to None. """ image_dataset_path = os.path.join(dataset_path, "image") # create the foundation datasets og_dataset = tf.data.Dataset.from_generator(lambda: original_image_path_gen(image_dataset_path), output_signature=tf.TensorSpec(shape=(), dtype=tf.string)) og_dataset = og_dataset.repeat(None) # repeat indefinitely ref_dataset = tf.data.Dataset.from_generator(lambda: ref_image_path_gen(image_dataset_path), output_signature=(tf.TensorSpec(shape=(), dtype=tf.string), tf.TensorSpec(shape=(), dtype=tf.bool))) ref_dataset = ref_dataset.repeat(None) # repeat indefinitely # create the input datasets og_image_dataset = og_dataset.map(lambda x: tf.py_function(load_image, [x, tf.convert_to_tensor(False, dtype=tf.bool)], tf.float32), num_parallel_calls=tf.data.AUTOTUNE) masked_image_dataset = og_image_dataset.map(lambda x: tf.py_function(load_masked_image, [x], tf.float32), num_parallel_calls=tf.data.AUTOTUNE) ref_image_dataset = ref_dataset.map(lambda x, y: tf.py_function(load_image, [x, y], tf.float32), num_parallel_calls=tf.data.AUTOTUNE) audio_spec_dataset = og_dataset.map(lambda x: tf.py_function(load_audio_data, [x, dataset_path], tf.float64), num_parallel_calls=tf.data.AUTOTUNE) unsync_spec_dataset = ref_dataset.map(lambda x, _: tf.py_function(load_audio_data, [x, dataset_path], tf.float64), num_parallel_calls=tf.data.AUTOTUNE) # ensure shape as tensorflow does not accept unknown shapes og_image_dataset = og_image_dataset.map(lambda x: tf.ensure_shape(x, IMAGE_SHAPE)) masked_image_dataset = masked_image_dataset.map(lambda x: tf.ensure_shape(x, MASKED_IMAGE_SHAPE)) ref_image_dataset = ref_image_dataset.map(lambda x: tf.ensure_shape(x, IMAGE_SHAPE)) audio_spec_dataset = audio_spec_dataset.map(lambda x: tf.ensure_shape(x, AUDIO_SPECTROGRAM_SHAPE)) unsync_spec_dataset = unsync_spec_dataset.map(lambda x: tf.ensure_shape(x, AUDIO_SPECTROGRAM_SHAPE)) # multi input using https://discuss.tensorflow.org/t/train-a-model-on-multiple-input-dataset/17829/4 full_dataset = tf.data.Dataset.zip((masked_image_dataset, ref_image_dataset, audio_spec_dataset, unsync_spec_dataset), og_image_dataset) # if shuffle: # full_dataset = full_dataset.shuffle(buffer_size=batch_size * 8, seed=seed) # not sure why buffer size is such # batch full_dataset = full_dataset.batch(batch_size=batch_size) return full_dataset
-
Library for Machine learning and quantum computing
Keras
-
My Favorite DevTools to Build AI/ML Applications!
As a beginner, I was looking for something simple and flexible for developing deep learning models and that is when I found Keras. Many AI/ML professionals appreciate Keras for its simplicity and efficiency in prototyping and developing deep learning models, making it a preferred choice, especially for beginners and for projects requiring rapid development.
- Release: Keras 3.3.0
-
Getting Started with Gemma Models
After setting the variables for the environment, the next step is to install dependencies. To use Gemma, KerasNLP is the dependency used. KerasNLP is a collection of natural language processing (NLP) models implemented in Keras and runnable on JAX, PyTorch, and TensorFlow.
-
Keras 3.0
All breaking changes are listed here: https://github.com/keras-team/keras/issues/18467
You can use this migration guide to identify and fix each of these issues (and further, making your code run on JAX or PyTorch): https://keras.io/guides/migrating_to_keras_3/
- Keras 3: A new multi-back end Keras
-
Can someone explain how keras code gets into the Tensorflow package?
I'm guessing the "real" keras code is coming from the keras repository. Is that a correct assumption? How does that version of Keras get there? If I wanted to write my own activation layer next to ELU, where exactly would I do that?
-
How popular are libraries in each technology
Other popular machine learning tools include PyTorch, Keras, and Scikit-learn. PyTorch is an open-source machine learning library developed by Facebook that is known for its ease of use and flexibility. Keras is a high-level neural networks API that is written in Python and is known for its simplicity. Scikit-learn is a machine learning library for Python that is used for data analysis and data mining tasks.
What are some alternatives?
optuna - A hyperparameter optimization framework
MLP Classifier - A handwritten multilayer perceptron classifer using numpy.
stable-baselines3 - PyTorch version of Stable Baselines, reliable implementations of reinforcement learning algorithms.
scikit-learn - scikit-learn: machine learning in Python
Faust - Python Stream Processing
Pandas - Flexible and powerful data analysis / manipulation library for Python, providing labeled data structures similar to R data.frame objects, statistical functions, and much more
gevent - Coroutine-based concurrency library for Python
xgboost - Scalable, Portable and Distributed Gradient Boosting (GBDT, GBRT or GBM) Library, for Python, R, Java, Scala, C++ and more. Runs on single machine, Hadoop, Spark, Dask, Flink and DataFlow
stable-baselines - A fork of OpenAI Baselines, implementations of reinforcement learning algorithms
tensorflow - An Open Source Machine Learning Framework for Everyone
SCOOP (Scalable COncurrent Operations in Python) - SCOOP (Scalable COncurrent Operations in Python)
Prophet - Tool for producing high quality forecasts for time series data that has multiple seasonality with linear or non-linear growth.