🧠 LLM
Opiniated RAG for integrating GenAI in your apps 🧠 Focus on your product rather than the RAG. Easy integration in existing products with customisation! Any LLM: GPT4, Groq, Llama. Any Vectorstore: …
The all-in-one Desktop & Docker AI application with built-in RAG, AI agents, No-code agent builder, MCP compatibility, and more.
🛰️ An approximate nearest-neighbor search library for Python and Java with a focus on ease of use, simplicity, and deployability.
🦜🔗 Build context-aware reasoning applications
Weaviate is an open-source vector database that stores both objects and vectors, allowing for the combination of vector search with structured filtering with the fault tolerance and scalability of …
Easy-to-use and powerful LLM and SLM library with awesome model zoo.
Neural search for web-sites, docs, articles - online!
基于LangChain和ChatGLM-6B等系列LLM的针对本地知识库的自动问答
Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)
Scripts for fine-tuning Llama2 via SFT and DPO.
FastGPT is a knowledge-based platform built on the LLMs, offers a comprehensive suite of out-of-the-box capabilities such as data processing, RAG retrieval, and visual AI workflow orchestration, le…
Humanable Chat Generative-model Fine-tuning | LLM微调
A natural language interface for computers
BISHENG is an open LLM devops platform for next generation Enterprise AI applications. Powerful and comprehensive features include: GenAI workflow, RAG, Agent, Unified model management, Evaluation,…
Production-ready platform for agentic workflow development.
Replace OpenAI GPT with another LLM in your app by changing a single line of code. Xinference gives you the freedom to use any LLM you need. With Xinference, you're empowered to run inference with …
The UI design language and React library for Conversational UI
NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on NVIDIA GPUs. This repository contains the open source components of TensorRT.
MII makes low-latency and high-throughput inference possible, powered by DeepSpeed.
Large Language Model Text Generation Inference
High-speed Large Language Model Serving for Local Deployment