Skip to content

isheriff123/private-llm-aws

 
 

Repository files navigation

Private LLM on AWS

In this repository you'll find the necessary documentation and code to properly deploy a Private LLM on top of AWS that is backed by SingleStoreDB. This LLM is fully capable Retrieval Augmented Generation (RAG) and can be fed custom business context using SingleStoreDB's Pipeline functions.

The basic process that this Terraform module uses is to deploy models using SageMaker, front them with FastAPI, and store all interactions inside of SingleStoreDB. SingleStoreDB will also provide additional context to the application that you're working with to ensure that you're able to build your business logic into the application.

This module optionally deploys Kai Shoes, our fake eCommerce store, with a ChatBot that allows you to demonstrate the power of having a contextual LLM that runs on a private network (inside your VPC).

We plan to continue to build out functionality in this module over time, please check out CONTRIBUTING.md to learn more about how you can help us build in more functionality!

Notes

User

the user you're deploying this with needs:

  • SageMaker createDomain
{
	"Version": "2012-10-17",
	"Statement": [
		{
			"Sid": "VisualEditor0",
			"Effect": "Allow",
			"Action": "sagemaker:CreateDomain",
			"Resource": "*"
		},
		{
            "Action": [
                "sagemaker:ListTags"
            ],
            "Effect": "Allow",
            "Resource": "*"
         }
	]
}

Known Issues

  • #1 terraform destroy [...] doesn't destroy subnet without you deleting EFS shares and the VPC manually

About

Terraform Templates for Private LLM Deployment

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • HCL 63.3%
  • Python 33.9%
  • Dockerfile 1.9%
  • Shell 0.9%