Skip to content

Manage scalable open LLM inference endpoints in Runai and Slurm clusters

License

Notifications You must be signed in to change notification settings

TugdualKerjan/llm-swarm

 
 

Repository files navigation

🐝 Fork of llm-swarm 🦋

Manage scalable open LLM inference endpoints in Runai and Slurm clusters

Features

What's different here ?

  • Support for RunAI Schedulers is added.

  • The code is agnostic to a specific scheduler, new ones can be added following the BaseScheduler.py class

  • Templates have been cleaned up and an example for running with RunAI is given.

  • __init__.py is more readable.

  • utils.py is a file full of helper functions

  • Typing is used to avoid type errors in functions

Install and prepare

pip install -e .
# or pip install llm_swarm
mkdir -p .cache/
# you can customize the above docker image cache locations and change them in `templates/tgi_h100.template.slurm` and `templates/vllm_h100.template.slurm`

For the rest read the official README.md

About

Manage scalable open LLM inference endpoints in Runai and Slurm clusters

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 79.8%
  • Shell 20.2%