-
The Chinese University of Hong Kong
- Sha Tin, Hong Kong
- https://rip4kobe.github.io/zhuoli.github.io/
Lists (27)
Sort Name ascending (A-Z)
Affordance
Awesome Lists
Diffusion Policy
🔮 Future ideas
Human Speech Recognition
Humanoid Control
Humanoid Hardware
IIL
✨ Inspiration
LLM
LLM for Robotics
MLLM
Motion Plannning
🚀 My stack
Optimization
Reasoning
RL
Robotic Grasping & Manipulation
ROS
Synthetic Data Generation
Teleoperation
Useful Tools
VLA
VLM
WM
World Model
Zotero
Starred repositories
Open-Sora: Democratizing Efficient Video Production for All
ICCV 2025 | TesserAct: Learning 4D Embodied World Models
A Comprehensive Survey on Continual Learning in Generative Models.
[ICML'25] The PyTorch implementation of paper: "AdaWorld: Learning Adaptable World Models with Latent Actions".
Awesome paper list and repos of the paper "A comprehensive survey of embodied world models".
Wan: Open and Advanced Large-Scale Video Generative Models
This repository is dedicated to collecting and sharing research papers on diffusion guidance methods.
Code implementation and development of Xtrainer robot in Clover lab
MichalZawalski / embodied-CoT
Forked from openvla/openvlaEmbodied Chain of Thought: A robotic policy that reason to solve the task.
DSPy: The framework for programming—not prompting—language models
Multimodal Chain-of-Thought Reasoning: A Comprehensive Survey
A library for differentiable nonlinear optimization
Synthetic Municipal Solid Waste Generator for AI-powered Waste Recognition System
Interactive visualizations of the geometric intuition behind diffusion models.
This repo contains the code for the paper "Intuitive physics understanding emerges fromself-supervised pretraining on natural videos"
Video Prediction Policy: A Generalist Robot Policy with Predictive Visual Representations https://video-prediction-policy.github.io
A Paper List for Humanoid Robot Learning.
This repository is the implementation of Gripper-agnostic Diffusion Policy for pick-and-place manipulation in SE(3) space
[AAAI'26 Oral] DexGraspVLA: A Vision-Language-Action Framework Towards General Dexterous Grasping
Bayesian optimisation & Reinforcement Learning library developed by Huawei Noah's Ark Lab
A benchmark for evaluating learning agents based on just language feedback