Skip to content

A Helm chart to deploy Llama.cpp to your TuringPI cluster

Notifications You must be signed in to change notification settings

elepedus/turingpi-llama-cpp

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

3 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

TuringPi Llama.cpp Chart

Deploys Llama.cpp server onto your TuringPi cluster, complete with a persistent volume to store the model files, replication and an ingress. Assumes you have followed the instructions at docs.turingpi.com to configure Longhorn, MetaLB and Traefik. By default, uses lmstudio-ai/gemma-2b-it-GGUF model, but this can be overridden with custom values.

Installation

helm install llama-cpp https://elepedus.github.io/turingpi-llama-cpp/turingpi-llama-cpp-0.0.1.tgz --namespace=llama-cpp

Usage

By default, the ingress exposes the web UI at llama.turingpi-cluster.local, at the same IP address as you configured for turingpi-cluster.local Make sure to update your /etc/hosts file so the new subdomain is accessible:

10.0.0.70 turing-cluster turing-cluster.local llama.turingpi-cluster llama.turingpi-cluster.local

About

A Helm chart to deploy Llama.cpp to your TuringPI cluster

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages