-
Shanghai Jiao Tong University
- China
- yushunxiang.github.io
- https://www.zhihu.com/people/upbeat-40
Block or Report
Block or report YushunXiang
Contact GitHub support about this user’s behavior. Learn more about reporting abuse.
Report abuseStars
Language
Sort by: Recently starred
An open source implementation of CLIP.
Prompt Learning for Vision-Language Models (IJCV'22, CVPR'22)
[ECCV2022] Contrastive Vision-Language Pre-training with Limited Resources
"MimicPlay: Long-Horizon Imitation Learning by Watching Human Play" code repository
[Embodied-AI-Survey-2024] Awesome Paper list for Embodied AI
Imitation learning algorithms with Co-training for Mobile ALOHA: ACT, Diffusion Policy, VINN
ICLR2024 Spotlight: curation/training code, metadata, distribution and pre-trained models for MetaCLIP; CVPR 2024: MoDE: CLIP Data Experts via Clustering
This is the official repository of our paper "What If We Recaption Billions of Web Images with LLaMA-3 ?"
Official implementation of Dancing with Still Images: Video Distillation via Static-Dynamic Disentanglement.
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
This is a library that provides dual dexterous hand manipulation tasks through Isaac Gym
Instruct2Act: Mapping Multi-modality Instructions to Robotic Actions with Large Language Model
Official PyTorch implementation of ECCV 2024 Paper: ControlNet++: Improving Conditional Controls with Efficient Consistency Feedback.
QQNT 插件加载器:LiteLoaderQQNT —— 轻量 · 简洁 · 开源
Code use to create COCO Attributes dataset and experiments in the associate ECCV 2016 paper.
Northwestern Polytechnical University 2024 Undergraduate Graduation Design Thesis LaTeX Template
西北工业大学硕博学位论文模版 | Yet Another Thesis Template for Northwestern Polytechnical University
Northwestern Polytechnical University 2024 Undergraduate Graduation Design Thesis LaTeX Template
Papers and codes collection for customized, personalized and editable generative models
Official Repository for "Eureka: Human-Level Reward Design via Coding Large Language Models" (ICLR 2024)
This project aim to reproduce Sora (Open AI T2V model), we wish the open source community contribute to this project.
A comprehensive list of papers using large language/multi-modal models for Robotics/RL, including papers, codes, and related websites
Strong and Open Vision Language Assistant for Mobile Devices