Skip to content

DzvinkaYarish/tartu-nlp-courses-qa

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

19 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

AI-powered QA system for University of Tartu courses

LLaMA inference on HPC

Create conda environment

git clone https://github.com/DzvinkaYarish/tartu-nlp-courses-qa.git
conda create -n qa_env python=3.10 --file tartu-nlp-courses-qa/env.txt
conda activate qa_env

pip install sentencepiece
pip install accelerate
pip install git+https://github.com/huggingface/peft.git

pip install wandb # optional, for logging training to W&B
wandb login

Clone acceleration library for Transformers

git clone https://github.com/timdettmers/bitsandbytes.git
cp tartu-nlp-courses-qa/fine-tuning/Makefile bitsandbytes/

Start interactive session on HPC

Optional: start srun in tmux to keep the session running after disconnecting from the server.

For tesla GPU:

srun --partition=gpu --gres=gpu:tesla:1  --mem=32G  --time=120 --cpus-per-task=4 --pty /bin/bash

For a100 GPU:

srun --partition=gpu --gres=gpu:a100-40g  --mem=32G  --time=120 --cpus-per-task=4 --pty /bin/bash
module load cuda/11.7.0 # for a100 gpu only!

module load any/python/3.8.3-conda

module load broadwell/gcc/5.2.0

conda activate nlp

Only when running inference for the first time:

cd bitsandbytes

CUDA_VERSION=117 make cuda11x

python3.10 setup.py install

Run LLaMA inference

cd ../tartu-nlp-courses-qa/fine-tuning
python3.10 llama_inference.py

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published