Skip to content

Collection of Ansible playbooks helping to set up IBM Power environments for AI workloads

License

Notifications You must be signed in to change notification settings

DahlitzFlorian/ai-on-ibm-power-playbooks

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

16 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

AI on IBM Power Playbooks

Description

This repository contains Ansible playbooks helping to set up IBM Power environments for running AI workloads on them.

Usage

If Ansible is not already installed on your local machine, you can install it via (more information):

pip install ansible

Make sure to adjust the example inventory file before using it.

ansible-playbook -i example-inventory.yml playbooks/basic-llama.cpp.yml

Configuration

The following parameters can be specified in the inventory:

  • auto_start (boolean): If set to true, the llama.cpp server will be started by Ansible.
  • conda_dir (string): Path is used as the root_prefix and prefix of Micromamba. Will be created if it does not exist.
  • detached (bool): If set to true, the llama.cpp server is started in the background and the playbook can finish without terminating the process. Is ignored if auto_start is set to false.
  • micromamba_location (string): Path to where the Micromamba binary gets stored.
  • model_repository (string): Huggingface repository name, e.g. QuantFactory/Meta-Llama-3-8B-GGUF.
  • model_file (string): Specific file in llama.cpp-supported format to download from given model_repository, e.g. Meta-Llama-3-8B.Q8_0.gguf.
  • python_version (string): Python version number, e.g. 3.11.
  • working_directory (string): Path to the working directory. Will be created if it does not exist.
  • llama_cpp_args (dictionary): Key-value pairs passed to llama-server in the format -KEY VALUE. For parameters without additional value, like -v, leave the value blank.
  • llama_cpp_argv (dictionary): Key-value pairs passed to llama-server in the format --KEY VALUE. For parameters without additional value, like --verbose, leave the value blank.
  • uvicorn_cpp_args (dictionary): Key-value pairs passed to uvicorn in the format -KEY VALUE. For parameters without additional value, like -v, leave the value blank.
  • uvicorn_cpp_argv (dictionary): Key-value pairs passed to uvicorn in the format --KEY VALUE. For parameters without additional value, like --verbose, leave the value blank.

Additional information

Are you looking for Python scripts helping you to interact with OpenAI-compatible LLM instance? Check out DahlitzFlorian/python-openai-client-example.

About

Collection of Ansible playbooks helping to set up IBM Power environments for AI workloads

Topics

Resources

License

Stars

Watchers

Forks

Languages