Opus-MT VS OPUS-MT-train

Compare Opus-MT vs OPUS-MT-train and see what are their differences.

Our great sponsors
  • WorkOS - The modern identity platform for B2B SaaS
  • InfluxDB - Power Real-Time Data Analytics at Scale
  • SaaSHub - Software Alternatives and Reviews
Opus-MT OPUS-MT-train
3 1
527 302
8.7% 5.3%
4.8 1.7
3 days ago about 2 months ago
Python Makefile
MIT License MIT License
The number of mentions indicates the total number of mentions that we've tracked plus the number of user suggested alternatives.
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.

Opus-MT

Posts with mentions or reviews of Opus-MT. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2023-02-03.
  • “sync,corrected by elderman” issue in ML translation datasets spread on internet
    1 project | news.ycombinator.com | 17 Mar 2023
    - mention on GitHub repo of a translation model https://github.com/Helsinki-NLP/Opus-MT/issues/62

    I'm curious to see if anyone else has interesting encounters with this

  • How worried are you about AI taking over music?
    13 projects | /r/WeAreTheMusicMakers | 3 Feb 2023
    Yes, most models these days, except the exceptionally large ones, are possible to train on a laptop. Of course it helps if your laptop has Nvidia CUDA GPU, but even if it doesn't you can rent an AWS 4 core/16GB GPU instance for 0.5 cents an hour. 24 hours of training time would be quite a lot for most models, unless you're trying to train a FB any to any language type model, but typically the big huge models are not the most interesting ones, and you can get very good results, and interesting models with substantially smaller sets of data. Opus MT models are only one language to one language, but they're about 300MB a model, and the quality rivals FB's models, and the speed is substantially faster. I don't have as many examples from the music space, as it's still a fairly under explored area, but Google has released Magenta which is a pretrained Tensorflow music model(actually a group of 3-4 models).
  • Helsinki-NLP/Opus-MT: Open neural machine translation models and web services
    1 project | /r/techtravel | 30 Dec 2021

OPUS-MT-train

Posts with mentions or reviews of OPUS-MT-train. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2022-04-21.
  • Amazon releases 51-language dataset for language understanding
    2 projects | news.ycombinator.com | 21 Apr 2022
    https://translatelocally.com/ is a nice gui around marian/bergamot. So far not very many bundled pairs, though I would guess any of the models from https://github.com/Helsinki-NLP/Opus-MT-train/tree/master/mo... and https://github.com/Helsinki-NLP/Tatoeba-Challenge/blob/maste... should be usable.

    There is also Apertium, a rule-based system which is very good for some closely-related pairs that have had a lot of work put into them (especially translation between Romance languages, e.g. Spanish→Catalan, and Norwegian Bokmål→Nynorsk), and the only OK translator for some lesser-resourced languages (e.g. Northern Saami→Norwegian Bokmål), but very underdeveloped for anything to/from English (it feels a bit pointless writing rules for English where there is so much available data; RBMT shines where there's not enough available data, ie. most of the languages of the world)

What are some alternatives?

When comparing Opus-MT and OPUS-MT-train you can also consider the following projects:

OpenNMT-py - Open Source Neural Machine Translation and (Large) Language Models in PyTorch

NLP-progress - Repository to track the progress in Natural Language Processing (NLP), including the datasets and the current state-of-the-art for the most common NLP tasks.

fastText - Library for fast text representation and classification.

Tatoeba-Challenge

Neural-Machine-Translated-communication-system - The model is designed to train a single and large neural network in order to predict correct translation by reading the given sentence.

tensor2tensor - Library of deep learning models and datasets designed to make deep learning more accessible and accelerate ML research.

Face-Recognition_Flutter - A sample Face recognition app using Flutter and Firebase ML Kit

Pytorch - Tensors and Dynamic neural networks in Python with strong GPU acceleration

klpt - The Kurdish Language Processing Toolkit

deep-learning-drizzle - Drench yourself in Deep Learning, Reinforcement Learning, Machine Learning, Computer Vision, and NLP by learning from these exciting lectures!!