Aurelian: 70B 32K story-writing (and more) [Alpha]

This page summarizes the projects mentioned and recommended in the original post on /r/LocalLLaMA

InfluxDB - Power Real-Time Data Analytics at Scale
Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
www.influxdata.com
featured
SaaSHub - Software Alternatives and Reviews
SaaSHub helps you find the best software and product alternatives
www.saashub.com
featured
  • LoftQ

  • But the quantization is done before training, and may not be optimal as you train the model. LoftQ is a method to re-compute the quantizations, taking into account the current full model (base model + learned LORA).

  • relora

    Official code for ReLoRA from the paper Stack More Layers Differently: High-Rank Training Through Low-Rank Updates

  • Similarly, the dominant components selected before training may change order as you train. ReLORA is basically a way to re-align and make sure you are always training something close to the current most important params.

  • InfluxDB

    Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.

    InfluxDB logo
  • LongLoRA

    Code and documents of LongLoRA and LongAlpaca (ICLR 2024 Oral)

  • Finally, LongLORA is a method to reduce the number of computations over a large context, and also specifically train the embed and norm layers fully, that is, no quantization or LORA for those. They are small layers and easy to train without too much VRAM cost, but the LongLORA authors noticed they have a big impact on long context performance. I am not using their computation reduction methods, but I am using their suggestion to train embed/norm layers fully.

NOTE: The number of mentions on this list indicates mentions on common posts plus user suggested alternatives. Hence, a higher number means a more popular project.

Suggest a related project

Related posts

  • Why train on Yi 4K instead of 200K?

    2 projects | /r/LocalLLaMA | 7 Dec 2023
  • [P] FastLoRAChat Instruct-tune LLaMA on consumer hardware with shareGPT data

    3 projects | /r/MachineLearning | 18 Apr 2023
  • ChatGPT provides false information about people, and OpenAI can't correct it

    1 project | news.ycombinator.com | 29 Apr 2024
  • Insights from Finetuning LLMs for Classification Tasks

    1 project | news.ycombinator.com | 28 Apr 2024
  • Financial Market Applications of LLMs

    1 project | news.ycombinator.com | 20 Apr 2024