-
CTU Prague
- Prague
-
22:13
(UTC +01:00)
Highlights
- Pro
Stars
CLU lets you write beautiful training loops in JAX.
Modular, scalable library to train ML models
Reference PyTorch implementation and models for DINOv3
Code implementation for the paper "Large-scale Pre-training for Grounded Video Caption Generation" (ICCV 2025)
NVIDIA Isaac GR00T N1.5 - A Foundation Model for Generalist Robots.
Nvidia GEAR Lab's initiative to solve the robotics data problem using world models
[ICLR 2025] LAPA: Latent Action Pretraining from Videos
An open-source AI agent that brings the power of Gemini directly into your terminal.
Distributed, scalable benchmarking of generalist robot policies.
A curated list of state-of-the-art research in embodied AI, focusing on vision-language-action (VLA) models, vision-language navigation (VLN), and related multimodal learning approaches.
State-of-the-art Image & Video CLIP, Multimodal Large Language Models, and More!
A simple, performant and scalable Jax LLM!
Library for reading and processing ML training data.
Gemma open-weight LLM library, from Google DeepMind
[ICLR 2025] 6D Object Pose Tracking in Internet Videos for Robotic Manipulation
MichalZawalski / embodied-CoT
Forked from openvla/openvlaEmbodied Chain of Thought: A robotic policy that reason to solve the task.
Code for the paper "ShowHowTo: Generating Scene-Conditioned Step-by-Step Visual Instructions" published at CVPR 2025
Theia: Distilling Diverse Vision Foundation Models for Robot Learning
[CVPR 2024] MemFlow: Optical Flow Estimation and Prediction with Memory
Image Polygonal Annotation with Python (polygon, rectangle, circle, line, point and image-level flag annotation).