1-Bit LLMs Could Solve AI's Energy Demands

This page summarizes the projects mentioned and recommended in the original post on news.ycombinator.com

Judoscale - Save 47% on cloud hosting with autoscaling that just works
Judoscale integrates with Django, FastAPI, Celery, and RQ to make autoscaling easy and reliable. Save big, and say goodbye to request timeouts and backed-up task queues.
judoscale.com
featured
CodeRabbit: AI Code Reviews for Developers
Revolutionize your code reviews with AI. CodeRabbit offers PR summaries, code walkthroughs, 1-click suggestions, and AST-based analysis. Boost productivity and code quality across all major languages with each PR.
coderabbit.ai
featured
  1. unilm

    Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities

  2. Judoscale

    Save 47% on cloud hosting with autoscaling that just works. Judoscale integrates with Django, FastAPI, Celery, and RQ to make autoscaling easy and reliable. Save big, and say goodbye to request timeouts and backed-up task queues.

    Judoscale logo
  3. BitNet-1.58-Instruct

    Implementation of BitNet-1.58 instruct tuning

  4. 1.58BitNet

    Experimental BitNet Implementation

  5. picoagent-rnd

    Web & CLI capable LLM agent (research prototype, no framework dependencies).

    While llama3-8b might be slightly more brittle under quantization, llama3-70b really surprised myself and others[1] in how well it performs even in the 2..3 bits per parameter regime. It requires one of the most advanced quantization methods (IQ2_XS specifically) but the reward is a SoTA LLM that fits on one 4090 GPU and allows for advanced usecases such as powering the agent engine I'm working on: https://github.com/kir-gadjello/picoagent-rnd

    For me it completely replaced strong models such as Mixtral-8x7B and DeepSeek-Coder-Instruct-33B.

    1. https://www.reddit.com/r/LocalLLaMA/comments/1cst400/result_...

  6. quip-sharp

NOTE: The number of mentions on this list indicates mentions on common posts plus user suggested alternatives. Hence, a higher number means a more popular project.

Suggest a related project

Related posts

Did you know that Python is
the 2nd most popular programming language
based on number of references?