Deploy through kollama
CLI
We have a CLI called kollama
here to simplify the deployment process. It is a simple way to deploy Ollama models to your Kubernetes cluster.
Getting Started
- Install the CLI:
shell
go install github.com/nekomeowww/ollama-operator/cmd/kollama@latest
To learn about the supported commands, please refer to
kollama
.
- Deploy a model:
shell
kollama deploy phi --expose --node-port 30101
For more information about the
deploy
command, please refer tokollama deploy
.
That's it.
- Interact with the model:
shell
OLLAMA_HOST=<Node ip>:30101 ollama run phi
or use the OpenAI API compatible endpoint:
shell
curl http://<Node ip>:30101/v1/chat/completions -H "Content-Type: application/json" -d '{
"model": "phi",
"messages": [
{
"role": "user",
"content": "Hello!"
}
]
}'
Contributors
Neko Ayaka