-
Peking University
- Beijing
- https://infinitywxh.github.io/
Highlights
- Pro
Block or Report
Block or report Infinitywxh
Contact GitHub support about this user’s behavior. Learn more about reporting abuse.
Report abuseStars
Language
Sort by: Recently starred
Official python implementation for ICML 2024: "Learning Solution-Aware Transformers for Efficiently Solving Quadratic Assignment Problem"
[Arxiv-2024] MotionLLM: Understanding Human Behaviors from Human Motions and Videos
RectifID: Personalizing Rectified Flow with Anchored Classifier Guidance
Official PyTorch code of "Grounded Question-Answering in Long Egocentric Videos", accepted by CVPR 2024.
MambaOut: Do We Really Need Mamba for Vision?
Efficient Python library for Extended LSTM with exponential gating, memory mixing, and matrix memory for superior sequence modeling.
[CVPR 2024] TimeChat: A Time-sensitive Multimodal Large Language Model for Long Video Understanding
Continual Learning of Large Language Models: A Comprehensive Survey
Official PyTorch implementation of "MultiAct: Long-Term 3D Human Motion Generation from Multiple Action Labels", in AAAI 2023 (Oral presentation).
[Mamba-Survey-2024] Paper list for State-Space-Model/Mamba and it's Applications
Official implementation of "MoMask: Generative Masked Modeling of 3D Human Motions (CVPR2024)"
(2024CVPR) MA-LMM: Memory-Augmented Large Multimodal Model for Long-Term Video Understanding
PyTorch Implementation of Jamba: "Jamba: A Hybrid Transformer-Mamba Language Model"
Structured state space sequence models
前沿论文持续更新--视频时刻定位 or 时域语言定位 or 视频片段检索。
Simple, minimal implementation of the Mamba SSM in one file of PyTorch.
Collection of papers on state-space models
Official implementation of ICML 2024 paper "Matrix Information Theory for Self-supervised Learning" (https://arxiv.org/abs/2305.17326)
[NeurIPS 2023] Official implementation of the paper "Motion-X: A Large-scale 3D Expressive Whole-body Human Motion Dataset"
Two-Stream Adaptive Graph Convolutional Networks for Skeleton-Based Action Recognition in CVPR19
Official repo for BMVC2021 paper ASFormer: Transformer for action segmentation
Code of ICCV 2023 paper: "AttT2M: Text-Driven Human Motion Generation with Multi-Perspective Attention Mechanism"
(CVPR 2023) Pytorch implementation of “T2M-GPT: Generating Human Motion from Textual Descriptions with Discrete Representations”
[NeurIPS 2023] MotionGPT: Human Motion as a Foreign Language, a unified motion-language generation model using LLMs
Official implementation of "TM2T: Stochastic and Tokenized Modeling for the Reciprocal Generation of 3D Human Motions and Texts (ECCV2022)"
awesome grounding: A curated list of research papers in visual grounding