PLOD-AbbreviationDetection
ThoughtSource
PLOD-AbbreviationDetection | ThoughtSource | |
---|---|---|
1 | 1 | |
9 | 837 | |
- | 1.3% | |
0.0 | 8.4 | |
over 1 year ago | 10 months ago | |
Jupyter Notebook | Jupyter Notebook | |
Creative Commons Attribution Share Alike 4.0 | MIT License |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
PLOD-AbbreviationDetection
-
Clustering to find abbreviations
Finally, the main problem with unsupervised learning is that you won't be able to reliably measure system performance or improvement. In my view, any time you can spend annotating and collecting data for a (semi-)supervised solution will be well-spent. Existing datasets can also get you started with model development, such as https://github.com/surrey-nlp/PLOD-AbbreviationDetection. Once you have a good model on a conventional dataset, you should be able to start generalizing it to your specific task/dataset.
ThoughtSource
What are some alternatives?
converse - Conversational text Analysis using various NLP techniques
medmcqa - A large-scale (194k), Multiple-Choice Question Answering (MCQA) dataset designed to address realworld medical entrance exam questions.
hate-speech-and-offensive-language - Repository for the paper "Automated Hate Speech Detection and the Problem of Offensive Language", ICWSM 2017
nlp - Repository for all things Natural Language Processing
goodreads - code samples for the goodreads datasets
transformers-interpret - Model explainability that works seamlessly with 🤗 transformers. Explain your transformers model in just 2 lines of code.
datasets - 🎁 5,400,000+ Unsplash images made available for research and machine learning
adaptnlp - An easy to use Natural Language Processing library and framework for predicting, training, fine-tuning, and serving up state-of-the-art NLP models.