nlpaug
diffusion_models
Our great sponsors
nlpaug | diffusion_models | |
---|---|---|
10 | 4 | |
4,252 | 152 | |
- | 2.6% | |
0.0 | 0.0 | |
about 1 year ago | almost 2 years ago | |
Jupyter Notebook | Jupyter Notebook | |
MIT License | MIT License |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
nlpaug
-
Use WordNet to collect homonyms
You'd want to use an NLP method for this as in order to determine optimal homonyms there would have to be some method of deriving context from the words ahead of and behind the substitution. Take a look at nlpaug.
-
Contextual Similarity between a list of n-grams and a website
3) Use deep contextual models with wordpiecing/BPE tokenizers- like all the models: BERT, RoBERTA, etc. On the simpler side, could also swap words with synonyms, which is easy to do with this library: https://github.com/makcedward/nlpaug. Instead of a single n-gram per topic, it might be nice to have a bundle of related words- you could play around with wordnet and see if that's helpful- also easy to do w/ nlpaug.
-
Word embeddings / language models for synonym generation?
In practice, even swapping words with dictionary synonyms is a problem because context isn't considered. Lexically sensitive contextual augmentation has become more popular in the last year or two - basically you mask a token using a large language model and then use the model to predict it so it has the full context. It's imperfect, but it's surprisingly useful when you want to upsample data. Nlpaug has an easy-to-use implementation https://github.com/makcedward/nlpaug
-
Text Data Augmentation using GPT-2 Language Model
A cool library I recently came across for text augmentation is nlpaug, it does a different thing to your approach, but I think both are useful :)
-
[D] Data Augmentation in NLP
This is a nice starting point: https://github.com/makcedward/nlpaug
-
NLPAug: what proportion of augmented sentences do you usually add to the dataset?
Since the dataset is relatively tiny, we are working on augmenting it with NLPAug. We use 2 strategies. Synonymisation and back translation.
-
Show HN: 40k Book Recommendations on HN Extracted Using Deep Learning
Thank you!
The medium post is amazingly written! I basically did the same thing - and you beat me with the data augmentation piece. I tried using nlpaug [0] but it didn't improve the model performance. I'll definitely try swapping book titles around.
[0] https://github.com/makcedward/nlpaug
-
[R] Call for Participation to NL-Augmenter 🦎 → 🐍
Are there any shortfalls in nlpaug which justified another project?
-
A Visual Survey of Data Augmentation in NLP
Spelling error injection In this method, we add spelling errors to some random word in the sentence. These spelling errors can be added programmatically or using a mapping of common spelling errors such as this list for English.
diffusion_models
- [Variational and Diffusion Methods] Minimal standalone example of diffusion model
- A minimal standalone example of diffusion model
-
(unguided) Sampling from a diffusion model
I'm trying to figure out a way to reproduce the sampling method used in the DDPM model arxiv link. The codebase link is roughly here for the original model and for improved DDPMs here. There is also implementations recently posted to r/MachineLearning such as this one (check the reverse process section) and finally this last one.
-
[P] Hands on diffusion models
A minimal example of the forward and reverse flow of diffusion models with equations from the paper and visualizations alongside the code: https://github.com/InFoCusp/diffusion_models I coded it up since I wanted to familiarize myself with rhe end to end flow. It uses a simple 2d dataset that can train within minutes. Hope others on this subreddit find it useful.
What are some alternatives?
spaCy - 💫 Industrial-strength Natural Language Processing (NLP) in Python
Compositional-Visual-Generation-with-Composable-Diffusion-Models-PyTorch - [ECCV 2022] Compositional Generation using Diffusion Models
NL-Augmenter - NL-Augmenter 🦎 → 🐍 A Collaborative Repository of Natural Language Transformations
score_sde_pytorch - PyTorch implementation for Score-Based Generative Modeling through Stochastic Differential Equations (ICLR 2021, Oral)
Tic-Tac-Toe-Gym - This is the Tic-Tac-Toe game made with Python using the PyGame library and the Gym library to implement the AI with Reinforcement Learning
DiffusionFastForward - DiffusionFastForward: a free course and experimental framework for diffusion-based generative models
azureml-examples - Official community-driven Azure Machine Learning examples, tested with GitHub Actions.
machine-learning-for-trading - Code for Machine Learning for Algorithmic Trading, 2nd edition.
advertorch - A Toolbox for Adversarial Robustness Research
Reinforcement-Learning - Learn Deep Reinforcement Learning in 60 days! Lectures & Code in Python. Reinforcement Learning + Deep Learning
dopamine - Dopamine is a research framework for fast prototyping of reinforcement learning algorithms.
computervision-recipes - Best Practices, code samples, and documentation for Computer Vision.