Skip to content

Obrew Server: A self-hostable machine learning engine. Build agents and schedule workflows private to you.

License

Notifications You must be signed in to change notification settings

dieharders/obrew-studio-server

Repository files navigation

OpenBrew - Your Personal Ai Engine



A general purpose tool for building private Ai desktop apps.

Table of Contents

Introduction

The goal of this project is to be an all-in-one solution for running local Ai that is easy to install, setup and use. It handles all basic building blocks of Ai: inference, memory retrieval (RAG) and storage (vector DB), model file management, and agent/workflow building.

obrew-ai.mp4

Description

OpenBrew is a native app with a GUI that can be configured to allow access from other apps you write or third party services, making it an ideal engine for Ai workloads built on your own tech stack.

How It Works

This backend runs a web server that acts as the main gateway to the suite of tools. A WebUI is provided called OpenBrew: WebUI to access this server. You can also run in headless mode to access it programmatically via the API.

To use Ai locally with your private data for free:

  • Launch the desktop app and choose an app to start using
  • or navigate your browser to any web app that supports the API
  • or connect with a service provider or custom stack that supports the OpenBrew API

Included Apps (FileBuff and OBrew Studio)

Workflow

query-doc.mp4

Minimum Hardware Requirements

  • 8GB Disk space
  • 4GB Memory

App Features Roadmap

✅ Run locally
✅ Windows OS installer
✅ MacOS installer
✅ Save chat history
✅ CPU & GPU support
❌ Linux installer
❌ Production ready: This project is under active development

Ai Features Roadmap

✅ Inference: Run open-source LLM models locally
✅ Embeddings: Create vector embeddings from a file/website/media to augment memory
✅ Knowledge Base: Search a vector database with Llama Index to retrieve information
✅ Agents: Customized LLM, can choose or specify tool use
✅ Tool Use: Choose from pre-made or write your own
✅ Multi-modal:

  • ✅ image
  • ✅ text
  • ❌ video
  • ❌ audio
  • ❌ 3d

❌ Observability: Source citations, logging, tracing
❌ Cached Context & Extended Context
❌ Voice-to-Text and Text-to-Speech

Supported Model Providers

This is a local first project. The ultimate goal is to support many providers via one API.

Open-Source (GGUF format)
Google Gemini
OpenAI
Anthropic
Mistral AI
Groq

Learn More

  • Backend: FastAPI - learn about FastAPI features and API.
  • Inference: llama-cpp for LLM inference.
  • Memory: ChromaDB for vector database.
  • WebUI: React for front-end UI and Pywebview for rendering the webview.