Skip to content

Blue-No1/fine-tuning-experiments-v2

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

10 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Fine-Tuning Experiments (LoRA/QLoRA)

Semi-hands-on configs and scripts for fine-tuning open-weight LLMs. Nội dung bán thực hành: config & script cho fine-tuning LLM open-weight.

⚠️ Work in progress. Small configs first, real runs later.

Goals

  • Minimal LoRA/QLoRA config with PEFT.
  • Reproducible train/eval commands.
  • Keep GPU memory modest (8–24GB target).

Layout

  • configs/ – YAML configs (model, LoRA ranks, train args).
  • scripts/ – train/infer helpers.
  • data/ – tiny demo JSONL (placeholder).
  • logs/ – example training logs (truncated).

Quick Start (conceptual)

# install
pip install transformers peft accelerate datasets bitsandbytes

# dry-run (conceptual)
python scripts/train_lora.py --config configs/lora-llama3-8b.yaml

About

LoRA/QLoRA experiments on open-weight LLMs (config, scripts, logs).

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages