Stars
An Empirical Study of GPT-4o Image Generation Capabilities
[ICCV 2025] VisualCloze: A universal image generation framework that can support a wide range of in-domain tasks and generalize to unseen ones. (🔥 🔥 🔥 Merged into offical pipelines of diffusers.)
[ICCV 2025] 🔥🔥 UNO: A Universal Customization Method for Both Single and Multi-Subject Conditioning
Reading notes about Multimodal Large Language Models, Large Language Models, and Diffusion Models
[ICCV 2025 Highlight] The official repository for "2.5 Years in Class: A Multimodal Textbook for Vision-Language Pretraining"
[CVPR 2025 Oral]Infinity ∞ : Scaling Bitwise AutoRegressive Modeling for High-Resolution Image Synthesis
[CVPR 2025] OVO-Bench: How Far is Your Video-LLMs from Real-World Online Video Understanding?
📃 A better UX for chat, writing content, and coding with LLMs.
Code for "AnyGPT: Unified Multimodal LLM with Discrete Sequence Modeling"
tiktoken is a fast BPE tokeniser for use with OpenAI's models.
Transforms complex documents like PDFs into LLM-ready markdown/JSON for your Agentic workflows.
Code and data for the paper Revealing the structure of language model capabilities
ChatGLM3 series: Open Bilingual Chat LLMs | 开源双语对话语言模型
Cleanlab's open-source library is the standard data-centric AI package for data quality and machine learning with messy, real-world data and labels.
The official repo of Qwen (通义千问) chat & pretrained large language model proposed by Alibaba Cloud.
Measuring Massive Multitask Language Understanding | ICLR 2021
中文LLaMA&Alpaca大语言模型+本地CPU/GPU训练部署 (Chinese LLaMA & Alpaca LLMs)
A large-scale 7B pretraining language model developed by BaiChuan-Inc.
Official github repo for C-Eval, a Chinese evaluation suite for foundation models [NeurIPS 2023]
Light local website for displaying performances from different chat models.
Central place for the engineering/scaling WG: documentation, SLURM scripts and logs, compute environment and data.
Ongoing research training transformer language models at scale, including: BERT & GPT-2
Simple implementation of using lora form the peft library to fine-tune the chatglm-6b