This repository contains Ansible playbooks helping to set up IBM Power environments for running AI workloads on them.
If Ansible is not already installed on your local machine, you can install it via (more information):
pip install ansible
Make sure to adjust the example inventory file before using it.
ansible-playbook -i example-inventory.yml playbooks/basic-llama.cpp.yml
The following parameters can be specified in the inventory:
- auto_start (boolean): If set to
true
, the llama.cpp server will be started by Ansible. - conda_dir (string): Path is used as the
root_prefix
andprefix
of Micromamba. Will be created if it does not exist. - detached (bool): If set to
true
, the llama.cpp server is started in the background and the playbook can finish without terminating the process. Is ignored ifauto_start
is set tofalse
. - micromamba_location (string): Path to where the Micromamba binary gets stored.
- model_repository (string): Huggingface repository name, e.g.
QuantFactory/Meta-Llama-3-8B-GGUF
. - model_file (string): Specific file in llama.cpp-supported format to download from given
model_repository
, e.g.Meta-Llama-3-8B.Q8_0.gguf
. - python_version (string): Python version number, e.g.
3.11
. - working_directory (string): Path to the working directory. Will be created if it does not exist.
- llama_cpp_args (dictionary): Key-value pairs passed to
llama-server
in the format-KEY VALUE
. For parameters without additional value, like-v
, leave the value blank. - llama_cpp_argv (dictionary): Key-value pairs passed to
llama-server
in the format--KEY VALUE
. For parameters without additional value, like--verbose
, leave the value blank. - uvicorn_cpp_args (dictionary): Key-value pairs passed to
uvicorn
in the format-KEY VALUE
. For parameters without additional value, like-v
, leave the value blank. - uvicorn_cpp_argv (dictionary): Key-value pairs passed to
uvicorn
in the format--KEY VALUE
. For parameters without additional value, like--verbose
, leave the value blank.
Are you looking for Python scripts helping you to interact with OpenAI-compatible LLM instance? Check out DahlitzFlorian/python-openai-client-example.