From cea01011bf48a9e0319242ba5cc27d4eeb8591e1 Mon Sep 17 00:00:00 2001 From: Diego Souza <8016841+diegosouzapw@users.noreply.github.com> Date: Fri, 20 Feb 2026 13:21:09 -0300 Subject: [PATCH] Add OmniRoute to Libraries and Wrappers --- llm-tools.md | 1 + 1 file changed, 1 insertion(+) diff --git a/llm-tools.md b/llm-tools.md index d7d8583..6c3ad65 100644 --- a/llm-tools.md +++ b/llm-tools.md @@ -484,6 +484,7 @@ - [auto-gptq](https://github.com/PanQiWei/AutoGPTQ) easy-to-use LLMs quantization package with user-friendly apis, based on GPTQ for GPU inference - [exllama](https://github.com/turboderp/exllama) Memory-Efficient Llama Rewrite in Python/C++/CUDA for 4bit quantized GPTQ weights, running on GPU, faster than llama.cpp ([2023-06-13](https://www.reddit.com/r/LocalLLaMA/comments/147z6as/llamacpp_just_got_full_cuda_acceleration_and_now/)), autoGPTQ and GPTQ-for-llama - [SimpleAI](https://github.com/lhenault/SimpleAI) Self-Hosted Alternative to openAI API +- [OmniRoute](https://github.com/diegosouzapw/OmniRoute) Self-hostable AI gateway with 4-tier automatic fallback routing across 36+ providers, OpenAI-compatible API, quota tracking, and zero-cost fallback to free tiers - [rustformer llm](https://github.com/rustformers/llm) Rust-based ecosystem for llms like BLOOM, GPT-2/J/NeoX, LLaMA and MPT offering a CLI for easy interaction and powered by ggml - [Haven](https://github.com/havenhq/haven) Fine-Tune and Deploy LLMs On Your Own Infrastructure - [llama-cpp-python](https://github.com/abetlen/llama-cpp-python) Python Bindings for llama.cpp with low level C API interface, python API, openai like API and LangChain compatibility