Stars
Efficient Triton Kernels for LLM Training
Fine-tuning & Reinforcement Learning for LLMs. 🦥 Train OpenAI gpt-oss, DeepSeek-R1, Qwen3, Gemma 3, TTS 2x faster with 70% less VRAM.
🧑🚀 全世界最好的LLM资料总结(语音视频生成、Agent、辅助编程、数据处理、模型训练、模型推理、o1 模型、MCP、小语言模型、视觉语言模型) | Summary of the world's best LLM resources.
[ACL 2024] ProtLLM: An Interleaved Protein-Language LLM with Protein-as-Word Pre-Training
Textbook on reinforcement learning from human feedback
Examples and guides for using the OpenAI API
a toolkit on knowledge distillation for large language models
A streamlined and customizable framework for efficient large model (LLM, VLM, AIGC) evaluation and performance benchmarking.
Use PEFT or Full-parameter to CPT/SFT/DPO/GRPO 500+ LLMs (Qwen3, Qwen3-MoE, Llama4, GLM4.5, InternLM3, DeepSeek-R1, ...) and 200+ MLLMs (Qwen3-VL, Qwen3-Omni, InternVL3.5, Ovis2.5, Llava, GLM4v, Ph…
Fast and memory-efficient exact attention
A framework for few-shot evaluation of language models.
Evals is a framework for evaluating LLMs and LLM systems, and an open-source registry of benchmarks.
Lighteval is your all-in-one toolkit for evaluating LLMs across multiple backends
Fleming-R1: Toward Expert-Level Medical Reasoning via Reinforcement Learning
Repo for TCMChat: A Generative Large Language Model for Traditional Chinese Medicine
Train your Agent model via our easy and efficient framework
Freeing data processing from scripting madness by providing a set of platform-agnostic customizable pipeline processing blocks.
Data processing for and with foundation models! 🍎 🍋 🌽 ➡️ ➡️🍸 🍹 🍷
A series of large language models developed by Baichuan Intelligent Technology
Beyond the Model: Scaling Medical Capability with a Large Verifier System
Biomni: a general-purpose biomedical AI agent
Qwen DianJin: LLMs for the Financial Industry by Alibaba Cloud(通义点金:阿里云金融大模型)
[CVPR 2025] Official PyTorch Implementation of MambaVision: A Hybrid Mamba-Transformer Vision Backbone
Trinity-RFT is a general-purpose, flexible and scalable framework designed for reinforcement fine-tuning (RFT) of large language models (LLM).
gpt-oss-120b and gpt-oss-20b are two open-weight language models by OpenAI
Collection of scripts and notebooks for OpenAI's latest GPT OSS models