Lists (1)
Sort Name ascending (A-Z)
Starred repositories
Official Repo for paper: Scaling Behavior Cloning Improves Causal Reasoning: An Open Model for Real-Time Video Game Playing
Learning-based locomotion control from OpenRobotLab, including Hybrid Internal Model & H-Infinity Locomotion Control
Everything you need to build state-of-the-art foundation multimodal desktop agent, end-to-end.
Official implementation of the paper: "NavDP: Learning Sim-to-Real Navigation Diffusion Policy with Privileged Information Guidance"
A high-throughput and memory-efficient inference and serving engine for LLMs
A Foundation Model for Generalist Gaming Agents
[AAAI 2025 Oral] FlowPolicy: Enabling Fast and Robust 3D Flow-based Policy via Consistency Flow Matching for Robot Manipulation
Pioneering Automated GUI Interaction with Native Agents
MM-ACT: Learn from Multimodal Parallel Generation to Act
Pixio: a capable vision encoder dedicated to dense prediction, simply by pixel reconstruction
Official repo for paper "EMMA: Efficient Multimodal Understanding, Generation, and Editing with a Unified Architecture."
Native and Compact Structured Latents for 3D Generation
off-policy algorithm utilizing offline and online data
📹 A more flexible framework that can generate videos at any resolution and creates videos from images.
[NeurIPS 2025] The official code for "IllumiCraft: Unified Geometry and Illumination Diffusion for Controllable Video Generation"
[NeurIPS`25] TC-Light: Temporally Coherent Generative Rendering for Realistic World Transfer
RynnVLA-002: A Unified Vision-Language-Action and World Model
Lumos-Custom Project: research for customized video generation in the Lumos Project.
1st place solution of 2025 BEHAVIOR Challenge
This repository summarizes recent advances in the VLA + RL paradigm and provides a taxonomic classification of relevant works.
A curated list of state-of-the-art research in embodied AI, focusing on vision-language-action (VLA) models, vision-language navigation (VLN), and related multimodal learning approaches.