Transformers-Tutorials
adaptnlp
Our great sponsors
Transformers-Tutorials | adaptnlp | |
---|---|---|
7 | 2 | |
7,510 | 414 | |
- | 0.0% | |
8.4 | 0.0 | |
16 days ago | over 2 years ago | |
Jupyter Notebook | Jupyter Notebook | |
MIT License | Apache License 2.0 |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
Transformers-Tutorials
-
AI enthusiasm #6 - Finetune any LLM you want๐ก
Most of this tutorial is based on Hugging Face course about Transformers and on Niels Rogge's Transformers tutorials: make sure to check their work and give them a star on GitHub, if you please โค๏ธ
- FLaNK Stack Weekly for 07August2023
- How to annotate compound words to build NER models?
-
[discussion] Anybody Working with VITMAE?
I'm pretraining on 850K grayscale spectrograms of birdsongs. I'm on epoch 400 out of 800 and the loss has declined from about 1.2 to 0.7. I don't really have a sense of what is "good enough" and I guess the only way I can judge is by looking at the reconstruction. I'm doing that using this notebook as a guide and right now it's doing quite badly.
-
[D] NLP has HuggingFace, what does Computer Vision have?
More tutorials can be found at https://github.com/NielsRogge/Transformers-Tutorials.
-
[Discussion] Information Extraction with LayoutLMv2
Ive been looking for an off the shelf encoder-decoder document understanding model for key information extraction. I found a great Huggingface implementation with concise notebook examples. However, the token classification model outputs a list of token labels corresponding bounding boxes for the token, but, not the text contained within the labeled bounding boxes themselves. Am I missing something? LayoutLMv2 describes itself as being capable of information extraction but without extracting the text I feel like it's fallen short of that ambition.
-
[Project] Deepmind's Perceiver IO available through Hugging Face
Example Notebooks
adaptnlp
-
Tools to use for Semantic-searching Question Answering System
Check out adaptnlp
-
Case Sensitivity using HuggingFace & Google's T5 model (base)
Yes, there are capitals in the tokenizer vocabulary of t5-base and t5-small, so both support capitalization. A few days ago I was using t5-small through adaptnlp for extractive summarization and capitalization was working fine (https://github.com/Novetta/adaptnlp). AdaptNLP is basically just a transformers wrapper, so if you can't figure out a solution, you could just dissect their source code.
What are some alternatives?
nn - ๐งโ๐ซ 60 Implementations/tutorials of deep learning papers with side-by-side notes ๐; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, sophia, ...), gans(cyclegan, stylegan2, ...), ๐ฎ reinforcement learning (ppo, dqn), capsnet, distillation, ... ๐ง
Basic-UI-for-GPT-J-6B-with-low-vram - A repository to run gpt-j-6b on low vram machines (4.2 gb minimum vram for 2000 token context, 3.5 gb for 1000 token context). Model loading takes 12gb free ram.
gorilla-cli - LLMs for your CLI
keytotext - Keywords to Sentences
pytorch-image-models - PyTorch image models, scripts, pretrained weights -- ResNet, ResNeXT, EfficientNet, NFNet, Vision Transformer (ViT), MobileNet-V3/V2, RegNet, DPN, CSPNet, Swin Transformer, MaxViT, CoAtNet, ConvNeXt, and more
fastai - The fastai deep learning library
notebooks - Notebooks using the Hugging Face libraries ๐ค
gector - Official implementation of the papers "GECToR โ Grammatical Error Correction: Tag, Not Rewrite" (BEA-20) and "Text Simplification by Tagging" (BEA-21)
OpenBuddy - Open Multilingual Chatbot for Everyone
browser-ml-inference - Edge Inference in Browser with Transformer NLP model
ToolBench - [ICLR'24 spotlight] An open platform for training, serving, and evaluating large language model for tool learning.
ML-Workspace - ๐ All-in-one web-based IDE specialized for machine learning and data science.