A lightweight LiteLLM server boilerplate pre-configured with uv and Docker for hosting your own OpenAI- and Anthropic-compatible endpoints. Includes LibreChat as an optional web UI.
-
Updated
Dec 8, 2025 - Python
A lightweight LiteLLM server boilerplate pre-configured with uv and Docker for hosting your own OpenAI- and Anthropic-compatible endpoints. Includes LibreChat as an optional web UI.
Function-calling API for LLM from multiple providers
API server for `llm` CLI tool
A flexible FastAPI-based framework for handling AI tasks using Large Language Models (LLMs). Supports multiple providers, extensible tasks and routers, Redis caching, and OpenAI integration. Easily scalable for various LLM-based applications.
PHP Frontend for Hosting local LLM's (run via VSCode or basic php execution methods/ add to project)
A complete, menu-driven AI model interface for Windows that simplifies running local GGUF language models with llama.cpp. This tool automatically manages dependencies, provides multiple interaction modes, and prioritizes user privacy through fully offline operation.
Unified LLM server with nginx reverse proxy and intelligent routing based on model ID
Add a description, image, and links to the llm-server topic page so that developers can more easily learn about it.
To associate your repository with the llm-server topic, visit your repo's landing page and select "manage topics."