FlexGen
autocomplete
FlexGen | autocomplete | |
---|---|---|
39 | 164 | |
9,007 | 24,274 | |
0.8% | 0.1% | |
3.0 | 9.6 | |
15 days ago | 5 days ago | |
Python | TypeScript | |
Apache License 2.0 | MIT License |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
FlexGen
- Run 70B LLM Inference on a Single 4GB GPU with This New Technique
- Colorful Custom RTX 4060 Ti GPU Clocks Outed, 8 GB VRAM Confirmed
-
Local Alternatives of ChatGPT and Midjourney
LLaMA, Pythia, RWKV, Flan-T5 (self-hosted), FlexGen
- FlexGen: Running large language models on a single GPU
-
Show HN: Finetune LLaMA-7B on commodity GPUs using your own text
> With no real knowledge of LLM and only recently started to understand what LLM terms mean, such as 'model, inference, LLM model, intruction set, fine tuning' whatelse do you think is required to make a took like yours?
This was mee a few weeks ago. I got interested in all this when FlexGen (https://github.com/FMInference/FlexGen) was announced, which allowed to run inference using OPT model on consumer hardware. I'm an avid user of Stable Diffusion, and I wanted to see if I can have an SD equivalent of ChatGPT.
Not understanding the details of hyperparameters or terminology, I basically asked ChatGPT to explain to me what these things are:
Explain to someone who is a software engineer with limited knowledge of ML terms or linear algebra, what is "feed forward" and "self-attention" in the context of ML and large language models. Provide examples when possible.
- Could this new flexgen be used in place of GPTq? or is this different?
- OpenAI is expensive
autocomplete
-
Fig Is Sunsetting
Having contributed to the Fig autocomplete specs, I find this sad. The Amazon product Fig was built into basically works as replacement, which is good. Still, the core value of this product are the open-source autocomplete specs: https://github.com/withfig/autocomplete. What's going to happen to that? It looks like they are still using it in the Amazon product. It should definitely be possible for an open-source re-implementation of the Fig UI to use those specs. There is a lot of knowledge encoded in there!
-
Top Free Utility Mac Apps You Aren’t Using
8. Fig
-
Ask HN: Alternatives to fig.io as it has signups disabled?
Fig is awesome but with signups blocked[1] for 2+mo already it's also as good as dead ¯\_(ツ)_/¯
* [1]: https://github.com/withfig/autocomplete/issues/2068
-
Show HN: Inshellisense – IDE style shell autocomplete
https://github.com/withfig/autocomplete is it this?
- Fig
-
Show HN: Whiz – A copilot for your command line
How is this different than https://fig.io/?
-
Boost DX, Enhance UX, and Skyrocket Profits! Dive into a sub-50ms world with Edge Feature Flags 🚀
AWS CloudWatch Evidently The worst. No comment. AWS seems to perpetually lack a good DX for developers. It appears that they don't recognize or continually undervalue the importance of roles other than engineers, such as Product Managers or Designers. Very disappointing. However, AWS has recently acquired Fig, so looks like they're now pursuing an acquisition strategy instead. Let's see how it turns it out, and let's hope they don't ruin Fig, since it's such an useful tool.
-
Ask HN: What are some well-designed websites?
slightly tangential, but where do people get awesome landing pages like linear(https://fig.io/. has similar landing page) etc. Do they build them in-house or buy templates somewhere? Many of the recently launched YC companies have awesome landing pages. eg. https://automorphic.ai/,
-
Fig Has Joined AWS
I love this product, have contributed several times to it, and I'm a little torn. One thing I am thinking about now, is that the completion specs are MIT-licensed, and it should be possible to use them to re-implement a basic open-source version of the autocompletion product... https://github.com/withfig/autocomplete
What are some alternatives?
llama - Inference code for Llama models
ohmyzsh - 🙃 A delightful community-driven (with 2,300+ contributors) framework for managing your zsh configuration. Includes 300+ optional plugins (rails, git, macOS, hub, docker, homebrew, node, php, python, etc), 140+ themes to spice up your morning, and an auto-update tool so that makes it easy to keep up with the latest updates from the community.
text-generation-webui - A Gradio web UI for Large Language Models. Supports transformers, GPTQ, AWQ, EXL2, llama.cpp (GGUF), Llama models.
fzf-tab - Replace zsh's default completion selection menu with fzf!
text-generation-inference - Large Language Model Text Generation Inference
Warp - Warp is a modern, Rust-based terminal with AI built in so you and your team can build great software, faster.
whisper.cpp - Port of OpenAI's Whisper model in C/C++
starship - ☄🌌️ The minimal, blazing-fast, and infinitely customizable prompt for any shell!
DeepSpeed - DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
hyperterm - A terminal built on web technologies
audiolm-pytorch - Implementation of AudioLM, a SOTA Language Modeling Approach to Audio Generation out of Google Research, in Pytorch
zsh-autocomplete - 🤖 Real-time type-ahead completion for Zsh. Asynchronous find-as-you-type autocompletion.