OpenNMT-Tutorial
mt5-M2M-comparison
OpenNMT-Tutorial | mt5-M2M-comparison | |
---|---|---|
3 | 1 | |
138 | 13 | |
- | - | |
4.6 | 3.8 | |
30 days ago | almost 3 years ago | |
Jupyter Notebook | Jupyter Notebook | |
MIT License | - |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
OpenNMT-Tutorial
mt5-M2M-comparison
-
[D] Comparing M2M to mT5 in low resource translation (10k dataset Yoruba - English)
I found no clear comparison nor a clear guide on how to fine tune both of the models on the translation task, so I decided to write it myself. (code: https://github.com/maroxtn/mt5-M2M-comparison)
What are some alternatives?
CTranslate2 - Fast inference engine for Transformer models
fastT5 - ⚡ boost inference speed of T5 models by 5x & reduce the model size by 3x.
sentencepiece - Unsupervised text tokenizer for Neural Network-based text generation.
keytotext - Keywords to Sentences
NeMo - A scalable generative AI framework built for researchers and developers working on Large Language Models, Multimodal, and Speech AI (Automatic Speech Recognition and Text-to-Speech)
100DaysOfML - 100 Days Of Machine Learning. New Content in every 1-2 day and projects every week. The massive 100DaysOfML in building
pytorch-seq2seq - Tutorials on implementing a few sequence-to-sequence (seq2seq) models with PyTorch and TorchText.
fake-news - Building a fake news detector from initial ideation to model deployment
question_generation - Neural question generation using transformers