-
Shanghai Jiao Tong University
- Shanghai
-
20:58
(UTC +08:00)
Highlights
- Pro
Starred repositories
Open-Source Low-Latency Accelerated Linux WebRTC HTML5 Remote Desktop Streaming Platform for Self-Hosting, Containers, Kubernetes, or Cloud/HPC
Isaac Gym Reinforcement Learning Environments
the scott CPU from "But How Do It Know?" by J. Clark Scott
A curated list of 3D Vision papers relating to Robotics domain in the era of large models i.e. LLMs/VLMs, inspired by awesome-computer-vision, including papers, codes, and related websites
A Next.js GitHub Homepage template that incorporates powerful and popular libraries and practices. Developed using TypeScript with Next.js, TailwindCSS, Framer Motion and React Spring. 🦄
🏡 My corner of internet, a.k.a Personal Homepage & Portfolio.
This repository compiles a list of papers related to the application of video technology in the field of robotics! Star⭐ the repo and follow me if you like what you see🤩.
Visualize streams of multimodal data. Fast, easy to use, and simple to integrate. Built in Rust using egui.
A conda-smithy repository for rerun-sdk.
Video+code lecture on building nanoGPT from scratch
DelinQu / SimplerEnv-OpenVLA
Forked from simpler-env/SimplerEnvEvaluating and reproducing real-world robot manipulation policies (e.g., RT-1, RT-1-X, Octo, and OpenVLA) in simulation under common setups (e.g., Google Robot, WidowX+Bridge)
Evaluating and reproducing real-world robot manipulation policies (e.g., RT-1, RT-1-X, Octo) in simulation under common setups (e.g., Google Robot, WidowX+Bridge)
The repository provides code for running inference with the Meta Segment Anything Model 2 (SAM 2), links for downloading the trained model checkpoints, and example notebooks that show how to use th…
HumanPlus: Humanoid Shadowing and Imitation from Humans
Motion Generation for Robots and Machines. Real-time. Jerk-constrained. Time-optimal.
Minimum implementation of consistency models
An example RLDS dataset builder for X-embodiment dataset conversion.
openvla / openvla
Forked from TRI-ML/prismatic-vlmsOpenVLA: An open-source vision-language-action model for robotic manipulation.
moojink / rlds_dataset_mod
Forked from kpertsch/rlds_dataset_modEfficiently apply modification functions to RLDS/TFDS datasets.
[RSS 2024] Learning Manipulation by Predicting Interaction
A Generalizable World Model for Autonomous Driving
Distributed machine learning infrastructure for large-scale robotics research
Pandora: Towards General World Model with Natural Language Actions and Video States