Easy to setup, self-hosted, multi-model AI chatbot & API server.
- Ollama - https://ollama.com/
- Open Web UI - https://openwebui.com/
- Alexis -
./alexis
, An Alexa-like chatbot. - Hardware?: See https://github.com/olafrv/ai_chat_llama2
git clone https://github.com/olafrv/shollama.git
cd shollama
# Tweak the configuration in .env
./start.sh
./models.sh
# Ollama require GPU (CUDA) to run, but
# in docker-compose.yml can be disabled.
# Go to http://localhost:8080
# Or run any tests ./test*.sh
./stop.sh
docker: ()...) could not select device driver "nvidia" (...) [[gpu]]
, Comment the 'deploy' section in docker-compose.yaml for ollama service.
Cannot find local snapshot HF_HUGGING_OFFLINE=1
. Set it to 0 inenv.iu.env
, you can set it back after the first run.
https://docs.continue.dev/getting-started/install https://marketplace.visualstudio.com/items?itemName=Continue.continue
https://docs.continue.dev/customize/model-providers/ollama#using-a-remote-instance
{
"models": [
{
"title": "deepseek-r1",
"provider": "ollama",
"model": "deepseek-r1",
"apiBase": "http://localhost:11434"
},
{
"title": "qwen2.5-coder",
"provider": "ollama",
"model": "qwen2.5-coder:1.5b",
"apiBase": "http://localhost:11434"
},
]
}