finetuner
gpt_index
DISCONTINUED
Our great sponsors
finetuner | gpt_index | |
---|---|---|
36 | 48 | |
1,192 | 7,332 | |
4.9% | - | |
0.0 | 9.8 | |
2 months ago | 7 months ago | |
Python | Python | |
Apache License 2.0 | MIT License |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
finetuner
-
How can I create a dataset to refine Whisper AI from old videos with subtitles?
You can try creating your own dataset. Get some audio data that you want, preprocess it, and then create a custom dataset you can use to fine tune. You could use finetuners like these if you want as well.
-
A Guide to Using OpenTelemetry in Jina for Monitoring and Tracing Applications
We derived the dataset by pre-processing the deepfashion dataset using Finetuner. The image label generated by Finetuner is extracted and formatted to produce the text attribute of each product.
-
[D] Looking for an open source Downloadable model to run on my local device.
You can either use Hugging Face Transformers as they have a lot of pre-trained models that you can customize. Or Finetuners like this one: which is a toolkit for fine-tuning multiple models.
-
Improving Search Quality for Non-English Queries with Fine-tuned Multilingual CLIP Models
Very recently, a few non-English and multilingual CLIP models have appeared, using various sources of training data. In this article, we’ll evaluate a multilingual CLIP model’s performance in a language other than English, and show how you can improve it even further using Jina AI’s Finetuner.
-
Classification using prompt or fine tuning?
you can try prompt-based classification or fine-tuning with a Finetuner. Prompts work well for simple tasks but fine-tuning may give better results for complex ones. Althouigh it's going to need more resources, but try both and see what works best for you.
-
Asking questions about lengthy texts
If you've got a set of Q&A pairs for your 60-page lease or medical paper, you could use finetuners to help answer questions about the text. But if you don't have those pairs, fine-tuning might not be good. Try summarizing the doc or extract the info. And if you're hitting the token limit, try using a bigger model or breaking up the text into smaller pieces.
-
What are the best Python libraries to learn for beginners?
Actually further in applying ML, Finetuner is pretty handy for getting the last mile done which I found useful.
-
Fine-tuning open source models to emulate ChatGPT for code explanation.
One option I’m considering is using fine tuners like the one from HuggingFace or Jina AI to fine-tune open source models like GPT-J or OPT to improve specific use-cases like code explanation. With the funding that we have, I wouldn’t want to cheap out on fine-tuning and expect something good.
-
Efficient way to tune a network by changing hyperparameters?
Off the top of my head you can either use Grid Search to test hyperparam combinations, Random Search to randomize hyperparams and Neural search uses ML to optimize hyperparameter tuning. You can use finetuners for this as well.
-
Seeking advice on improving NLP search results
Back then, I came across some info about a self-supervised sentence embedding system that surpasses Sentence Transformers NLI models, but forgot where it was. You could use Jina’s Finetuner. It lets you boost your pre-trained models' performance, making them ready for production without having to spend a lot of time labeling or buying expensive hardware.
gpt_index
-
Basic links to get started with Prompt Programming
LLAMA Index Github repository
-
Large language models are having their Stable Diffusion moment
This is exactly what LlamaIndex is meant to solve!
A set of data structures to augment LLM's with your data: https://github.com/jerryjliu/gpt_index
-
ChatGPT's API Is So Good and Cheap, It Makes Most Text Generating AI Obsolete
This is what we've designed LlamaIndex for! https://github.com/jerryjliu/gpt_index. Designed to help you "index" over a large doc corpus in different ways for use with LLM prompts.
-
Is there a way I can have ChatGPT look at a document of mine?
https://github.com/jerryjliu/gpt_index might be close to what you need.
-
AI is making it easier to create more noise, when all I want is good search
I would start with https://gpt-index.readthedocs.io/en/latest/ and https://langchain.readthedocs.io/en/latest/
-
Using OpenAI with self hosted knowledge database
People have been doing this with https://github.com/jerryjliu/gpt_index
-
LLaMA: A foundational, 65B-parameter large language model
(creator of gpt index / llamaindex here https://github.com/jerryjliu/gpt_index)
Funny that we had just rebranded our tool from GPT Index to LlamaIndex about a week ago to avoid potential trademark issues with OpenAI, and turns out Meta has similar ideas around LLM+llama puns :). Must mean the name is good though!
Also very excited to try plugging in the LLaMa model into LlamaIndex, will report the results.
-
Hello, is there a "BEST OF" prompts list here somewhere?
LLAMA Index Github repository
-
Show HN: IngestAI – NoCode ChatGPT-bot creator from your knowledge base in Slack
[2] https://gpt-index.readthedocs.io/en/latest/
-
Incorporating new information into a GPT-3 model
Have any of you attempted to customize a GPT-3 model by incorporating new information into it? My personal experience has been limited to fine-tuning the model to perform specific tasks, such as functioning as a chatbot. To augment its knowledge, I have used tools such as GPT-index (https://github.com/jerryjliu/gpt_index) which enable the language model to access external knowledge bases. However, I am interested in exploring the possibility of internalizing new information within the model.
What are some alternatives?
langchain - ⚡ Building applications with LLMs through composability ⚡ [Moved to: https://github.com/langchain-ai/langchain]
llama - Inference code for LLaMA models
awesome-chatgpt-prompts - This repo includes ChatGPT prompt curation to use ChatGPT better.
text-generation-webui - A Gradio web UI for Large Language Models. Supports transformers, GPTQ, llama.cpp (GGUF), Llama models.
Jina AI examples - Jina examples and demos to help you get started
nanoGPT - The simplest, fastest repository for training/finetuning medium-sized GPTs.
openai-cookbook - Examples and guides for using the OpenAI API
RWKV-LM - RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the best of RNN and transformer - great performance, fast inference, saves VRAM, fast training, "infinite" ctx_len, and free sentence embedding.
jina - ☁️ Build multimodal AI applications with cloud-native stack
Promptify - Prompt Engineering | Prompt Versioning | Use GPT or other prompt based models to get structured output. Join our discord for Prompt-Engineering, LLMs and other latest research
LiteratureReviewBot - Experiment to use GPT-3 to help write grant proposals.
sketch - AI code-writing assistant that understands data content