feast
bert
feast | bert | |
---|---|---|
8 | 50 | |
5,293 | 37,168 | |
1.5% | 1.0% | |
9.5 | 0.0 | |
7 days ago | 20 days ago | |
Python | Python | |
Apache License 2.0 | Apache License 2.0 |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
feast
- What's Happening with Feast?
-
Running The Feast Feature Store With Dragonfly
Feast stands as an exceptional open-source feature store, revolutionizing the efficient management and uninterrupted serving of machine learning (ML) features for real-time applications. At its core, Feast offers a sophisticated interface for storing, discovering, and accessing features—the individual measurable properties or characteristics of data essential for ML modeling. Operating on a distributed architecture, Feast harmoniously integrates several pivotal components, including the Feast Registry, Stream Processor, Batch Materialization Engine, and Stores.
-
Ask HN: How to Break into AI Engineering
AI Engineering is basically Data Engineering focused on AI. When in "traditional" Data Engineering you create pipelines that store processed data in something like a Data Lake, in AI Eng. your end storage might be a specialized Feature Storage (like Feast or GCP Vertex AI).
There are some AI Engineers with strong scientific/mathematical background, but that's rare. Usually, you're paired with these ML people that actually develop and evaluate the models.
So my advice is to start with Data Engineering and then find a specialization AI. You should have a VERY solid foundation on scripting and programming, specially Python. Also, a lot of concepts of "data wrangling". Understanding how data flows from point A to point B, how the intermediate storages and streaming engines work, etc. Functional programming is key here.
[0] https://github.com/feast-dev/feast
-
In Need of Guidance: Implementing MLOps in a Complex Organization as a Junior Data Engineer
A feature store usually stores features which are used for training ML model. It is a centralized place for collaboration between data engineer, ML engineer, and data scientist, so that data engineer can write to the feature store while ML engineer and data scientist read from it. Hopsworks https://www.hopsworks.ai and feast https://github.com/feast-dev/feast are examples of open source feature store.
- [D] Your 🫵 Preferred Feature Stores?
-
[P] Announcing Feast 0.10: The simplest way to serve features in production
Github: https://github.com/feast-dev/feast
-
[D] What’s the simplest, most lightweight but complete and 100% open source MLOps toolkit? -> MY OWN CONCLUSIONS
Have you looked at Feats as a Feature Store solution? It seems promising but I haven't really looked into it yet though.
- Feast: OSS Feature Store for Production ML
bert
-
Zero Shot Text Classification Under the hood
In 2019, a new language representation called BERT (Bedirectional Encoder Representation from Transformers) was introduced. The main idea behind this paradigm is to first pre-train a language model using a massive amount of unlabeled data then fine-tune all the parameters using labeled data from the downstream tasks. This allows the model to generalize well to different NLP tasks. Moreover, it has been shown that this language representation model can be used to solve downstream tasks without being explicitly trained on, e.g classify a text without training phase.
-
OpenAI – Application for US trademark "GPT" has failed
task-specific parameters, and is trained on the downstream tasks by simply fine-tuning all pre-trained parameters.
[0] https://arxiv.org/abs/1810.04805
-
Integrate LLM Frameworks
The release of BERT in 2018 kicked off the language model revolution. The Transformers architecture succeeded RNNs and LSTMs to become the architecture of choice. Unbelievable progress was made in a number of areas: summarization, translation, text classification, entity classification and more. 2023 tooks things to another level with the rise of large language models (LLMs). Models with billions of parameters showed an amazing ability to generate coherent dialogue.
-
Embeddings: What they are and why they matter
The general idea is that you have a particular task & dataset, and you optimize these vectors to maximize that task. So the properties of these vectors - what information is retained and what is left out during the 'compression' - are effectively determined by that task.
In general, the core task for the various "LLM tools" involves prediction of a hidden word, trained on very large quantities of real text - thus also mirroring whatever structure (linguistic, syntactic, semantic, factual, social bias, etc) exists there.
If you want to see how the sausage is made and look at the actual algorithms, then the key two approaches to read up on would probably be Mikolov's word2vec (https://arxiv.org/abs/1301.3781) with the CBOW (Continuous Bag of Words) and Continuous Skip-Gram Model, which are based on relatively simple math optimization, and then on the BERT (https://arxiv.org/abs/1810.04805) structure which does a conceptually similar thing but with a large neural network that can learn more from the same data. For both of them, you can either read the original papers or look up blog posts or videos that explain them, different people have different preferences on how readable academic papers are.
- Ernie, China's ChatGPT, Cracks Under Pressure
-
Ask HN: How to Break into AI Engineering
Could you post a link to "the BERT paper"? I've read some, but would be interested reading anything that anyone considered definitive :) Is it this one? "BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding" :https://arxiv.org/abs/1810.04805
-
How to leverage the state-of-the-art NLP models in Rust
Rust crate rust_bert implementation of the BERT language model (https://arxiv.org/abs/1810.04805 Devlin, Chang, Lee, Toutanova, 2018). The base model is implemented in the bert_model::BertModel struct. Several language model heads have also been implemented, including:
-
Notes on training BERT from scratch on an 8GB consumer GPU
The achievement of training a BERT model to 90% of the GLUE score on a single GPU in ~100 hours is indeed impressive. As for the original BERT pretraining run, the paper [1] mentions that the pretraining took 4 days on 16 TPU chips for the BERT-Base model and 4 days on 64 TPU chips for the BERT-Large model.
Regarding the translation of these techniques to the pretraining phase for a GPT model, it is possible that some of the optimizations and techniques used for BERT could be applied to GPT as well. However, the specific architecture and training objectives of GPT might require different approaches or additional optimizations.
As for the SOPHIA optimizer, it is designed to improve the training of deep learning models by adaptively adjusting the learning rate and momentum. According to the paper [2], SOPHIA has shown promising results in various deep learning tasks. It is possible that the SOPHIA optimizer could help improve the training of BERT and GPT models, but further research and experimentation would be needed to confirm its effectiveness in these specific cases.
[1] https://arxiv.org/abs/1810.04805
-
List of AI-Models
Click to Learn more...
- Bert: Pre-Training of Deep Bidirectional Transformers for Language Understanding
What are some alternatives?
kedro-great - The easiest way to integrate Kedro and Great Expectations
NLTK - NLTK Source
featureform - The Virtual Feature Store. Turn your existing data infrastructure into a feature store.
bert-sklearn - a sklearn wrapper for Google's BERT model
Milvus - A cloud-native vector database, storage for next generation AI applications
pysimilar - A python library for computing the similarity between two strings (text) based on cosine similarity
metaflow - :rocket: Build and manage real-life ML, AI, and data science projects with ease!
transformers - 🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
great_expectations - Always know what to expect from your data.
PURE - [NAACL 2021] A Frustratingly Easy Approach for Entity and Relation Extraction https://arxiv.org/abs/2010.12812
mlrun - MLRun is an open source MLOps platform for quickly building and managing continuous ML applications across their lifecycle. MLRun integrates into your development and CI/CD environment and automates the delivery of production data, ML pipelines, and online applications.
NL_Parser_using_Spacy - NLP parser using NER and TDD