Skip to content
@thu-ml

TSAIL group

Tsinghua Statistical Artificial Intelligence & Learning Group

Pinned Loading

  1. TurboDiffusion TurboDiffusion Public

    TurboDiffusion: 100–200× Acceleration for Video Diffusion Models

    Python 2.9k 191

  2. unidiffuser unidiffuser Public

    Code and models for the paper "One Transformer Fits All Distributions in Multi-Modal Diffusion"

    Python 1.5k 90

  3. SageAttention SageAttention Public

    [ICLR2025, ICML2025, NeurIPS2025 Spotlight] Quantized Attention achieves speedup of 2-5x compared to FlashAttention, without losing end-to-end metrics across language, image, and video models.

    Cuda 3k 299

  4. prolificdreamer prolificdreamer Public

    ProlificDreamer: High-Fidelity and Diverse Text-to-3D Generation with Variational Score Distillation (NeurIPS 2023 Spotlight)

    Python 1.6k 47

  5. ares ares Public

    A Python library for adversarial machine learning focusing on benchmarking adversarial robustness.

    Python 521 93

  6. tianshou tianshou Public

    An elegant PyTorch deep reinforcement learning library.

    Python 9k 1.2k

Repositories

Showing 10 of 86 repositories
  • TurboDiffusion Public

    TurboDiffusion: 100–200× Acceleration for Video Diffusion Models

    thu-ml/TurboDiffusion’s past year of commit activity
    Python 2,949 Apache-2.0 191 47 5 Updated Jan 1, 2026
  • SpargeAttn Public

    [ICML2025] SpargeAttention: A training-free sparse attention that accelerates any model inference.

    thu-ml/SpargeAttn’s past year of commit activity
    Cuda 880 Apache-2.0 77 53 3 Updated Dec 31, 2025
  • MLA-Trust Public

    A toolbox for benchmarking Multimodal LLM Agents trustworthiness across truthfulness, controllability, safety and privacy dimensions through 34 interactive tasks

    thu-ml/MLA-Trust’s past year of commit activity
    Python 61 MIT 4 2 0 Updated Dec 31, 2025
  • Motus Public

    Official code of Motus: A Unified Latent Action World Model

    thu-ml/Motus’s past year of commit activity
    Python 462 Apache-2.0 8 9 0 Updated Dec 29, 2025
  • SLA Public

    SLA: Beyond Sparsity in Diffusion Transformers via Fine-Tunable Sparse–Linear Attention

    thu-ml/SLA’s past year of commit activity
    Python 221 Apache-2.0 12 5 0 Updated Dec 28, 2025
  • vidar-robotwin Public

    robotwin evaluation code for vidar.

    thu-ml/vidar-robotwin’s past year of commit activity
    Python 4 MIT 0 0 0 Updated Dec 22, 2025
  • SageAttention Public

    [ICLR2025, ICML2025, NeurIPS2025 Spotlight] Quantized Attention achieves speedup of 2-5x compared to FlashAttention, without losing end-to-end metrics across language, image, and video models.

    thu-ml/SageAttention’s past year of commit activity
    Cuda 2,982 Apache-2.0 299 142 17 Updated Dec 22, 2025
  • vidar Public

    Official repo for vidar and vidarc: video foundation model for robotics.

    thu-ml/vidar’s past year of commit activity
    Python 27 0 1 0 Updated Dec 22, 2025
  • DiT-Extrapolation Public

    Official implementation for "RIFLEx: A Free Lunch for Length Extrapolation in Video Diffusion Transformers" (ICML 2025) and "UltraViCo: Breaking Extrapolation Limits in Video Diffusion Transformers"

    thu-ml/DiT-Extrapolation’s past year of commit activity
    Python 768 Apache-2.0 73 23 0 Updated Dec 4, 2025
  • thu-ml/ultraimage.github.io’s past year of commit activity
    JavaScript 0 0 0 0 Updated Dec 3, 2025

Most used topics

Loading…