Indic-BERT-v1
kiri
Indic-BERT-v1 | kiri | |
---|---|---|
1 | 12 | |
271 | 240 | |
0.0% | 0.0% | |
1.4 | 3.2 | |
about 1 year ago | about 3 years ago | |
Python | Python | |
MIT License | GNU General Public License v3.0 or later |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
Indic-BERT-v1
-
Frequency of letters in Kannada Varnamale.
Would you or anyone be interested in doing a sample NLP project using https://github.com/AI4Bharat/indic-bert ?? I see that there is a trained model with 712 million tokens in Kannada ... I have very basic knowledge of doing some NLP in English and have always been interested to see how this works in Indian languages .... I see that some sample has already shown on this page ... We can try all NLP techniques one by one (Information Extraction, Text Classification, Text summarization and Text Generation) ...
kiri
-
[P][D] NLP question - Question Answering AI
I'm one of the authors of Backprop, a library built for transfer learning.
-
Backprop: Use and finetune models in a single line of code
I'd like to share Backprop, an open source library I've been co-authoring for the last few months.
-
[P] Backprop Model Hub: a curated list of state-of-the-art models
We've also got an open-source library that makes using + finetuning these models possible in a few lines of code.
- Show HN: Backprop – a simple library to use and finetune state-of-the-art models
- Show HN: Backprop – a library to easily finetune and use state-of-the-art models
-
[P] Backprop: a library to easily finetune and use state-of-the-art models
I'd like to share Backprop, a Python library I've been co-authoring for the last few months. Our goal is to make finetuning and using models as easy as possible, even without extensive ML experience.
-
GPT Neo: open-source GPT-3-like model with pretrained weights available
You might get some really promising results with finetuning.
If anything, you could build writing assistance that almost automates responses.
I've been co-authoring a library that lets you finetune such models in a single line of code.
https://github.com/backprop-ai/backprop
In specific the text generation finetuning example should be what you are looking for: https://github.com/backprop-ai/backprop/blob/main/examples/F...
Hope this helps, happy to chat more about it. Pretty curious about the results.
-
NLP Model for extracting specific text from raw text
Here's an example Jupyter Notebook for finetuning T5. Full disclosure, I work on this library myself -- but it could be helpful.
-
[D] Need help with document classifier and later prediction of text
I'm working on a library that hopefully makes working with some of these a bit easier -- here's an example notebook for running text classification with the BART checkpoint, if you're interested. If you need more task-specific finetuning for text classification, that's going to be rolled out in the near future.
-
Generating notes from text
I'm working on a library that includes a few different ML tasks, including summarisation. It uses a pretrained version of Google's T5 transformer model, which we host on Hugging Face with some details on how it was trained.
What are some alternatives?
BERT-pytorch - Google AI 2018 BERT pytorch implementation
gpt-neox - An implementation of model parallel autoregressive transformers on GPUs, based on the DeepSpeed library.
haystack - :mag: LLM orchestration framework to build customizable, production-ready LLM applications. Connect components (models, vector DBs, file converters) to pipelines or agents that can interact with your data. With advanced retrieval methods, it's best suited for building RAG, question answering, semantic search or conversational agent chatbots.
simpletransformers - Transformers for Information Retrieval, Text Classification, NER, QA, Language Modelling, Language Generation, T5, Multi-Modal, and Conversational AI
qagnn - [NAACL 2021] QAGNN: Question Answering using Language Models and Knowledge Graphs 🤖
CLIP - CLIP (Contrastive Language-Image Pretraining), Predict the most relevant text snippet given an image
Questgen.ai - Question generation using state-of-the-art Natural Language Processing algorithms
BERTweet - BERTweet: A pre-trained language model for English Tweets (EMNLP-2020)
gpt-neo - An implementation of model parallel GPT-2 and GPT-3-style models using the mesh-tensorflow library.
FinBERT-QA - Financial Domain Question Answering with pre-trained BERT Language Model
jiant - jiant is an nlp toolkit