Transformers-Tutorials
ToolBench
Our great sponsors
Transformers-Tutorials | ToolBench | |
---|---|---|
7 | 3 | |
7,510 | 4,392 | |
- | 5.7% | |
8.4 | 8.7 | |
16 days ago | about 1 month ago | |
Jupyter Notebook | Python | |
MIT License | Apache License 2.0 |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
Transformers-Tutorials
-
AI enthusiasm #6 - Finetune any LLM you want๐ก
Most of this tutorial is based on Hugging Face course about Transformers and on Niels Rogge's Transformers tutorials: make sure to check their work and give them a star on GitHub, if you please โค๏ธ
- FLaNK Stack Weekly for 07August2023
- How to annotate compound words to build NER models?
-
[discussion] Anybody Working with VITMAE?
I'm pretraining on 850K grayscale spectrograms of birdsongs. I'm on epoch 400 out of 800 and the loss has declined from about 1.2 to 0.7. I don't really have a sense of what is "good enough" and I guess the only way I can judge is by looking at the reconstruction. I'm doing that using this notebook as a guide and right now it's doing quite badly.
-
[D] NLP has HuggingFace, what does Computer Vision have?
More tutorials can be found at https://github.com/NielsRogge/Transformers-Tutorials.
-
[Discussion] Information Extraction with LayoutLMv2
Ive been looking for an off the shelf encoder-decoder document understanding model for key information extraction. I found a great Huggingface implementation with concise notebook examples. However, the token classification model outputs a list of token labels corresponding bounding boxes for the token, but, not the text contained within the labeled bounding boxes themselves. Am I missing something? LayoutLMv2 describes itself as being capable of information extraction but without extracting the text I feel like it's fallen short of that ambition.
-
[Project] Deepmind's Perceiver IO available through Hugging Face
Example Notebooks
ToolBench
- FLaNK Stack Weekly for 07August2023
-
[R] ToolLLM: Facilitating Large Language Models to Master 16000+ Real-world APIs - WeChat AI, Tencent Inc. 2023 - Open-source! Comparble performance to ChatGPT while using tools!
Github: https://github.com/OpenBMB/ToolBench
- [N] ToolBench is a set of data and tools that you can use to further customize and improve your language model (LLM).
What are some alternatives?
nn - ๐งโ๐ซ 60 Implementations/tutorials of deep learning papers with side-by-side notes ๐; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, sophia, ...), gans(cyclegan, stylegan2, ...), ๐ฎ reinforcement learning (ppo, dqn), capsnet, distillation, ... ๐ง
gorilla-cli - LLMs for your CLI
NiFi-Man - Like Travel Man, But With Data. The Data is Here, But Should We Have Ingested it?
pytorch-image-models - PyTorch image models, scripts, pretrained weights -- ResNet, ResNeXT, EfficientNet, NFNet, Vision Transformer (ViT), MobileNet-V3/V2, RegNet, DPN, CSPNet, Swin Transformer, MaxViT, CoAtNet, ConvNeXt, and more
EverythingApacheNiFi - EverythingApacheNiFi
notebooks - Notebooks using the Hugging Face libraries ๐ค
harlequin - The SQL IDE for Your Terminal.
adaptnlp - An easy to use Natural Language Processing library and framework for predicting, training, fine-tuning, and serving up state-of-the-art NLP models.
flink-kubernetes-operator - Apache Flink Kubernetes Operator
OpenBuddy - Open Multilingual Chatbot for Everyone
anomalib - An anomaly detection library comprising state-of-the-art algorithms and features such as experiment management, hyper-parameter optimization, and edge inference.