nekomeowww.OllamaOperator 0.10.5
Yet another operator for running large language models on Kubernetes with ease. Powered by Ollama! 🐫
While Ollama is a powerful tool for running large language models locally, and the user experience of CLI is just the same as using Docker CLI, it's not possible yet to replicate the same user experience on Kubernetes, especially when it comes to running multiple models on the same cluster with loads of resources and configurations. That's where the Ollama Operator kicks in: - Install the operator on your Kubernetes cluster - Apply the needed CRDs - Create your models - Wait for the models to be fetched and loaded, that's it! Thanks to the great works of lama.cpp, no more worries about Python environment, CUDA drivers. The journey to large language models, AIGC, localized agents, 🦜🔗 Langchain and more is just a few steps away!
Command Line
Download Links For Version 0.10.5
Info
- last updated 1/14/2025 12:00:00 AM
- Publisher: Neko
- License: Apache-2.0
Dependencies
No dependency information