tldr-transformers
NLP-progress
Our great sponsors
tldr-transformers | NLP-progress | |
---|---|---|
4 | 17 | |
167 | 22,296 | |
- | - | |
0.0 | 3.2 | |
over 1 year ago | 2 months ago | |
Python | ||
MIT License | MIT License |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
tldr-transformers
- Show HN: The “tl;dr” of Recent Transformer Papers
-
Show HN: Tl;Dr” on Transformers Papers
With the explosion in research on all things transformers, it seemed there was a need to have a single table to distill the "tl;dr" of each paper's contributions relative to each other. Here is what I got so far: https://github.com/will-thompson-k/tldr-transformers . Would love feedback - and feel free to contribute too :)
-
[P] NLP "tl;dr" Notes on Transformers
In any case, I'm liking the first glance so far. I'd just transpose the summary tables so they wouldn't get so tightly squeezed: https://github.com/will-thompson-k/tldr-transformers/blob/main/notes/bart.md
With the explosion in work on all things transformers, I felt the need to keep a single table of the "tl;dr" of various papers to distill their main takeaways: https://github.com/will-thompson-k/tldr-transformers . Would love feedback!
NLP-progress
- [Discussion] Checklist of seminal NLP papers
- NLP research status
-
[D] How difficult/easy is to learn NLP once you have experience in a CV?
One thing is that NLP is a set of wildly different problems which share some aspects, but often use quite different techniques and assumptions about their datasets. So even if you would have NLP experience, if you'd need to start on a substantially different NLP task, you can't just apply what you know and succeed, you have to review "how things are done" for that problem domain. For a quick overview, sites like https://nlpprogress.com/ can be helpful to see what methods are used; and, perhaps even more importantly, how people are modeling the actual task.
-
Upcoming App Announcement: Lemmatize, a Foreign Language Reader
A standard step in Chinese text processing is word segmentation, which deals with this problem.
-
Is there as site tracking computer vision process?
NLP has a github project tracking NLP progress, https://github.com/sebastianruder/NLP-progress. I wanna know if there is one tracking computer vision progress.
-
[P] NLP "tl;dr" Notes on Transformers
It would also be cool to have some charts with parameter density and even overall effectiveness (a tl;dr version of SOTA-trackers, maybe?) if that doesn't prove too infeasible.
- What are state-of-the-art methods for abstractive text summarization ?
-
BreadPanes 81: "They/Them"
As I said It increase ambiguity and cognitive overheard, needlessly given that "it" exists. Moreover it also make it harder for artificial intelligence to understand human text https://github.com/sebastianruder/NLP-progress/blob/master/english/coreference_resolution.md
-
[Request] Curated Advanced NLP Resources
I could not find it on the internet (including on GitHub, Kaggle, Medium, or Reddit.) And, I know about NLP Progress and The Super Duper NLP Repo.
-
How do you guys find/ keep up to date with the latest NLP papers?
For someone who needs to be on top of the latest research - Twitter (distraction-prone, marketing-friendly, instantly-gratifying, quick), newsletters in ML + NLP (https://jack-clark.net/, ruder.io, offconvex.org, etc.) (distraction-free, generic, time-consuming), SOTA chasing (https://paperswithcode.com/, http://nlpprogress.com/) (distraction-free, generic + focused, code-friendly)
What are some alternatives?
FARM - :house_with_garden: Fast & easy transfer learning for NLP. Harvesting language models for the industry. Focus on Question Answering.
nlp_tasks - Natural Language Processing Tasks and References
lemmatization-lists - Machine-readable lists of lemma-token pairs in 23 languages.
wtpsplit - Code for Where's the Point? Self-Supervised Multilingual Punctuation-Agnostic Sentence Segmentation
azure-sql-db-openai - Samples on how to use Azure SQL database with Azure OpenAI
SymSpell - SymSpell: 1 million times faster spelling correction & fuzzy search through Symmetric Delete spelling correction algorithm
long-range-arena - Long Range Arena for Benchmarking Efficient Transformers
awesome-hungarian-nlp - A curated list of NLP resources for Hungarian
transformers-convert
nlprule - A fast, low-resource Natural Language Processing and Text Correction library written in Rust.
language-planner - Official Code for "Language Models as Zero-Shot Planners: Extracting Actionable Knowledge for Embodied Agents"
OPUS-MT-train - Training open neural machine translation models