ToolBench
Transformers-Tutorials
ToolBench | Transformers-Tutorials | |
---|---|---|
3 | 7 | |
4,455 | 7,875 | |
3.7% | - | |
8.3 | 8.4 | |
5 days ago | 6 days ago | |
Python | Jupyter Notebook | |
Apache License 2.0 | MIT License |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
ToolBench
- FLaNK Stack Weekly for 07August2023
-
[R] ToolLLM: Facilitating Large Language Models to Master 16000+ Real-world APIs - WeChat AI, Tencent Inc. 2023 - Open-source! Comparble performance to ChatGPT while using tools!
Github: https://github.com/OpenBMB/ToolBench
- [N] ToolBench is a set of data and tools that you can use to further customize and improve your language model (LLM).
Transformers-Tutorials
-
AI enthusiasm #6 - Finetune any LLM you want๐ก
Most of this tutorial is based on Hugging Face course about Transformers and on Niels Rogge's Transformers tutorials: make sure to check their work and give them a star on GitHub, if you please โค๏ธ
- FLaNK Stack Weekly for 07August2023
- How to annotate compound words to build NER models?
-
[discussion] Anybody Working with VITMAE?
I'm pretraining on 850K grayscale spectrograms of birdsongs. I'm on epoch 400 out of 800 and the loss has declined from about 1.2 to 0.7. I don't really have a sense of what is "good enough" and I guess the only way I can judge is by looking at the reconstruction. I'm doing that using this notebook as a guide and right now it's doing quite badly.
-
[D] NLP has HuggingFace, what does Computer Vision have?
More tutorials can be found at https://github.com/NielsRogge/Transformers-Tutorials.
-
[Discussion] Information Extraction with LayoutLMv2
Ive been looking for an off the shelf encoder-decoder document understanding model for key information extraction. I found a great Huggingface implementation with concise notebook examples. However, the token classification model outputs a list of token labels corresponding bounding boxes for the token, but, not the text contained within the labeled bounding boxes themselves. Am I missing something? LayoutLMv2 describes itself as being capable of information extraction but without extracting the text I feel like it's fallen short of that ambition.
-
[Project] Deepmind's Perceiver IO available through Hugging Face
Example Notebooks
What are some alternatives?
gorilla-cli - LLMs for your CLI
nn - ๐งโ๐ซ 60 Implementations/tutorials of deep learning papers with side-by-side notes ๐; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, sophia, ...), gans(cyclegan, stylegan2, ...), ๐ฎ reinforcement learning (ppo, dqn), capsnet, distillation, ... ๐ง
NiFi-Man - Like Travel Man, But With Data. The Data is Here, But Should We Have Ingested it?
EverythingApacheNiFi - EverythingApacheNiFi
pytorch-image-models - PyTorch image models, scripts, pretrained weights -- ResNet, ResNeXT, EfficientNet, NFNet, Vision Transformer (ViT), MobileNet-V3/V2, RegNet, DPN, CSPNet, Swin Transformer, MaxViT, CoAtNet, ConvNeXt, and more
harlequin - The SQL IDE for Your Terminal.
notebooks - Notebooks using the Hugging Face libraries ๐ค
flink-kubernetes-operator - Apache Flink Kubernetes Operator
adaptnlp - An easy to use Natural Language Processing library and framework for predicting, training, fine-tuning, and serving up state-of-the-art NLP models.
anomalib - An anomaly detection library comprising state-of-the-art algorithms and features such as experiment management, hyper-parameter optimization, and edge inference.
OpenBuddy - Open Multilingual Chatbot for Everyone