Our great sponsors
-
open_llama
OpenLLaMA, a permissively licensed open source reproduction of Meta AI’s LLaMA 7B trained on the RedPajama dataset
-
WorkOS
The modern identity platform for B2B SaaS. The APIs are flexible and easy-to-use, supporting authentication, user identity, and complex enterprise features like SSO and SCIM provisioning.
-
RWKV-LM
RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the best of RNN and transformer - great performance, fast inference, saves VRAM, fast training, "infinite" ctx_len, and free sentence embedding.
-
Open-Instruction-Generalist
Open Instruction Generalist is an assistant trained on massive synthetic instructions to perform many millions of tasks
-
gnothi
Gnothi is an open-source AI journal and toolkit for self-discovery. If you're interested in getting involved, we'd love to hear from you.
-
InfluxDB
Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
-
RedPajama-Data
The RedPajama-Data repository contains code for preparing large datasets for training large language models.
For PC, please try gpt4all from the fine folks at nomic. For android, there's MLC chat.
I can't use Llama or any model from the Llama family, due to license restrictions. Although now there's also the OpenLlama family of models, which have the same architecture but were trained on an open dataset (RedPajama, the same dataset the base model in my app was trained on). I'd love to pursue the direction of extended context lengths for on-device LLMs. Likely in a month or so, when I've implemented all the product feature that I currently have on my backlog.
Thanks for the support! Two weeks ago, I'd have said longer contexts on small on-device LLMs are at least a year away, but developments from last week seem to indicate that it's well within reach. Once the low hanging product features are done, I think it's a worthy problem to spend a couple of weeks or perhaps even months on. Speaking of context lengths, recurrent models like RWKV technically have infinite context lengths, but in practice the context slowly fades away after a few thousands of tokens.
Please join r/LocalLLaMA and also look into ggml, llama.cpp.
It's currently based on an SFT tuned version of this model. The SFT dataset (OIG-small-chip2) was mostly around tasks and had no general knowledge in it. So, the base model's knowledge cut-off still holds. And the base model's knowledge cutoff is late 2022, AFAIK.
I created a GPT journal and am now looking into LocalLLM for the privacy-focused. I gave RedPajama-INCITE a spin, and found it soooo bad! I mean.. really bad. I should note this was about 2 months ago? I did see they released a newer version, I'm assuming there was a data-collection / training process they were waiting on. But I'm wondering: is the magic in the fine-tuning? As in, did you find the model useful before you fine-tuned, and the fine-tuning was just a cherry on top? Or did you have to fine-tune to make it work?
Lichess for chess. I'm sure you know about it.
Please join r/LocalLLaMA and also look into ggml, llama.cpp.
Thanks! I agree with your quip about true/proper personal assistants. I'm thinking out loud here. On iOS, accessing and modifying users' calendar and reminders can be done programatically with EventKit. On the LLM side of things, there are techniques to fine tune LLMs to use APIs, like Gorilla, ToolFormer, etc. Perhaps out of scope for Personal GPT, but I suspect it should be possible to build a better (in some aspects) on-device Siri with these primitives.
The hallucinations are coming from the LLM interpolating from the training data, substantial portions of which is scraped off of the internet. Because other peoples' prompts never leave their devices (this app makes no internet connections).
Related posts
- Lichess: Over 2023, concurrent players online have been higher than the lockdown and Queen's Gambit boom; servers stable so far (@agadmator knows how to take us down!) - Check out all our features, 100% free to everyone: https://lichess.org/features no paywalls, no ads, 100% for love of chess
- The Q&A Megathread for new and beginner chess players
- Why isn't LiChess employing the concept of "safe move" in analysis?
- Becoming a Chess Grandmaster
- Avoid blundering: 80% of a winning strategy