LocalAI is the open-source AI engine. Run any model - LLMs, vision, voice, image, video - on any hardware. No GPU required.
-
Updated
Mar 25, 2026 - Go
LocalAI is the open-source AI engine. Run any model - LLMs, vision, voice, image, video - on any hardware. No GPU required.
A unified AI model hub for aggregation & distribution. It supports cross-converting various LLMs into OpenAI-compatible, Claude-compatible, or Gemini-compatible formats. A centralized gateway for personal and enterprise model management. 🍥
A corporate law RAG system with innovative retrieval and contextual strategies
rerank library for easy reranking of results
Unified LLM API client library for Python. Simple API for Chat, Embedding, Rerank, and Tokenizer. OpenAI-compatible with streaming support and unified usage tracking.
SearchAugmentedLLM empowers LLMs with information from the web
A comprehensive RAG FastAPI service that handles document uploads and retrievals, built with Python. Uses PyMuPDF for document processing, turbopuffer for vector storage, OpenAI for models, and cohere for reranking.
The watsonx.ai Java SDK is an open-source library that simplifies the integration of IBM watsonx.ai APIs into Java applications. It supports chat completions, streaming, tool calling, embeddings, text classification, text extraction, text detection, reranking, and time-series forecasting.
A Python project that deploys a Local RAG chatbot using Ollama API and vLLM API. Refines answers with internal RAG knowledge base, using both Embedding and Rerank models to improve accuracy of context provided to LLM models.
Add a description, image, and links to the rerank topic page so that developers can more easily learn about it.
To associate your repository with the rerank topic, visit your repo's landing page and select "manage topics."