Skip to content

SulthanAbiyyu/Marzuki-7B

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

5 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Marzuki 7B

Indonesian LLM that trained on wikipedia indonesia, translated alpaca, translated OASST, and wikipedia DPO.

Model Name HF Link
Base (PT) https://huggingface.co/SulthanAbiyyu/marzuki-7B-v2-base
Instruct (SFT) https://huggingface.co/SulthanAbiyyu/marzuki-7B-v2-instruct
Aligned (DPO) https://huggingface.co/SulthanAbiyyu/marzuki-7B-v2-dpo

Usage

Using transformers
in progress 👨‍🍳

Integrate with langchain
in progress 👨‍🍳

Model Detail

Marzuki 7B is based on Mistral-7B-v0.1 as a base model. The Mistral-7B-v0.1 Large Language Model (LLM) is a pretrained generative text model with 7 billion parameters.

Click me

Hyperparameters

Hyperparameter Value
lora_target gate_proj,down_proj,up_proj,q_proj,v_proj,k_proj,o_proj
per_device_train_batch_size 4
gradient_checkpointing True
flash_attn True
fast_tokenizer True
gradient_accumulation_steps 4
lr_scheduler_type cosine
learning_rate 5e-5
num_train_epochs 3.0

Learning curve

training loss on pretraining stage training loss on fine tuning stage training loss on DPO stage

Benchmark Results

in progress 👨‍🍳

Citation

in progress 👨‍🍳

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published