Skip to content

Deploy through kollama CLI

We have a CLI called kollama here to simplify the deployment process. It is a simple way to deploy Ollama models to your Kubernetes cluster.

Getting Started

  1. Install the CLI:
shell
go install github.com/nekomeowww/ollama-operator/cmd/kollama@latest

To learn about the supported commands, please refer to kollama.

  1. Deploy a model:
shell
kollama deploy phi --expose --node-port 30101

For more information about the deploy command, please refer to kollama deploy.

That's it.

  1. Interact with the model:
shell
OLLAMA_HOST=<Node ip>:30101 ollama run phi

or use the OpenAI API compatible endpoint:

shell
curl http://<Node ip>:30101/v1/chat/completions -H "Content-Type: application/json" -d '{
  "model": "phi",
  "messages": [
      {
          "role": "user",
          "content": "Hello!"
      }
  ]
}'

Contributors

The avatar of contributor named as Neko Ayaka Neko Ayaka

Changelog