Skip to content

A time-saving script for any Llama CPP/LoRA workflow: it handles merging the LoRA into the base model, converting it to GGUF format, and applying post-training quantization.

Notifications You must be signed in to change notification settings

davidtorcivia/Merge-LoRA-Into-GGUF

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

17 Commits
 
 
 
 
 
 

Repository files navigation

Merge LoRA, convert to GGUF, and quantize.

A time-saving script for any Llama CPP/LoRA workflow: it handles merging the LoRA into the base model, converting it to GGUF format, and applying post-training quantization.

Setup

  • Clone this repository: git clone https://github.com/georgepullen/Merge-LoRA-Into-GGUF.git .
  • Clone the llama.cpp repository: git clone https://github.com/ggerganov/llama.cpp.git .
  • Install requirements for both repositories: pip install -r requirements.txt, pip install -r merge_lora_requirements.txt

Arguments

  • -l Path to LoRA (HF Repo ID or Local Path)
  • -b Path to base model (HF Repo ID or Local Path)
  • -f Format (F32, F16, q6_k, q4_0, etc.)

About

A time-saving script for any Llama CPP/LoRA workflow: it handles merging the LoRA into the base model, converting it to GGUF format, and applying post-training quantization.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages