How to merge the two weights into a single weight?

This page summarizes the projects mentioned and recommended in the original post on /r/LocalLLaMA

InfluxDB - Power Real-Time Data Analytics at Scale
Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
www.influxdata.com
featured
SaaSHub - Software Alternatives and Reviews
SaaSHub helps you find the best software and product alternatives
www.saashub.com
featured
  • DetGPT

  • Hey guys, I hope you are all doing well. I'm wondering how can I merge the two llama weights in this HF page (https://huggingface.co/TheBloke/llama-13b/tree/main) into a single llama weight that I could use? My end goal is to be able to use this model: https://github.com/OptimalScale/DetGPT. It needs to be able to merge the robin lora weights with the llama weights, but I'm not sure how to do that. Please help me. Thanks in advance!

  • peft

    🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.

  • To obtain the original llama model, one may refer to this doc. To merge a lora model with a base model, one may refer to PEFT or use the merge script provided by LMFlow.

  • InfluxDB

    Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.

    InfluxDB logo
NOTE: The number of mentions on this list indicates mentions on common posts plus user suggested alternatives. Hence, a higher number means a more popular project.

Suggest a related project

Related posts

  • LoftQ: LoRA-fine-tuning-aware Quantization

    1 project | news.ycombinator.com | 19 Dec 2023
  • PEFT 0.5 supports fine-tuning GPTQ models

    1 project | /r/LocalLLaMA | 24 Aug 2023
  • Exploding loss when trying to train OpenOrca-Platypus2-13B

    1 project | /r/LocalLLaMA | 21 Aug 2023
  • [D] Is there a difference between p-tuning and prefix tuning ?

    1 project | /r/MachineLearning | 3 Jul 2023
  • How does using QLoRAs when running Llama on CPU work?

    2 projects | /r/LocalLLaMA | 23 Jun 2023