Highlights
Lists (8)
Sort Name ascending (A-Z)
Stars
Sparkset is an AI-powered operational assistant that helps teams interact with databases using natural language.
🚀2.3x faster than MinIO for 4KB object payloads. RustFS is an open-source, S3-compatible high-performance object storage system supporting migration and coexistence with other S3-compatible platfor…
FlashInfer: Kernel Library for LLM Serving
Supercharge Your LLM with the Fastest KV Cache Layer
Lightning-Fast RL for LLM Reasoning and Agents. Made Simple & Flexible.
A powerful tool for creating fine-tuning datasets for LLM
[EMNLP2025] "LightRAG: Simple and Fast Retrieval-Augmented Generation"
Run, manage, and scale AI workloads on any AI infrastructure. Use one system to access & manage all AI compute (Kubernetes, 20+ clouds, or on-prem).
🌟 The Multi-Agent Framework: First AI Software Company, Towards Natural Language Programming
Mooncake is the serving platform for Kimi, a leading LLM service provided by Moonshot AI.
An excellent one-click EFI configuration tool based on OpenCore
🚀🚀 「大模型」2小时完全从0训练26M的小参数GPT!🌏 Train a 26M-parameter GPT from scratch in just 2h!
A collection of awesome video generation studies.
CodeGeeX: An Open Multilingual Code Generation Model (KDD 2023)
A high-throughput and memory-efficient inference and serving engine for LLMs
Efficient Triton Kernels for LLM Training
Ray is an AI compute engine. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.
Supercharge Your LLM Application Evaluations 🚀
SGLang is a high-performance serving framework for large language models and multimodal models.
Scalable toolkit for efficient model alignment
Distributed LLM and StableDiffusion inference for mobile, desktop and server.
An Easy-to-use, Scalable and High-performance Agentic RL Framework based on Ray (PPO & DAPO & REINFORCE++ & TIS & vLLM & Ray & Async RL)
[NeurIPS'23 Oral] Visual Instruction Tuning (LLaVA) built towards GPT-4V level capabilities and beyond.
A modular graph-based Retrieval-Augmented Generation (RAG) system
A repo for distributed training of language models with Reinforcement Learning via Human Feedback (RLHF)
Example models using DeepSpeed