-
UESTC | TongYi Laboratory
- Sichuan ⇌ Beijing
-
21:36
(UTC +08:00) - https://zchoi.github.io/
Block or Report
Block or report zchoi
Contact GitHub support about this user’s behavior. Learn more about reporting abuse.
Report abuseLists (3)
Sort Name ascending (A-Z)
Stars
Language
Sort by: Recently starred
MINT-1T: A one trillion token multimodal interleaved dataset.
The repository provides code for running inference with the Meta Segment Anything Model 2 (SAM 2), links for downloading the trained model checkpoints, and example notebooks that show how to use th…
A collection of works that investigate social agents, simulations and their real-world impact in text, embodied, and robotics contexts.
GRUtopia: Dream General Robots in a City at Scale
RoleInteract: Evaluating the Social Interaction of Role-Playing Agents
📰 Must-read papers on KV Cache Compression (constantly updating 🤗).
🔥🔥🔥Latest Papers, Codes and Datasets on Vid-LLMs.
DEEM: Official implementation of Diffusion models serve as the eyes of large language models for image perception.
MetaMath: Bootstrap Your Own Mathematical Questions for Large Language Models
Accelerating the development of large multimodal models (LMMs) with lmms-eval
Official repo for the paper "Scaling Synthetic Data Creation with 1,000,000,000 Personas"
This is the official implementation of the paper "Needle In A Multimodal Haystack"
Cambrian-1 is a family of multimodal LLMs with a vision-centric design.
Awesome-llm-role-playing-with-persona: a curated list of resources for large language models for role-playing with assigned personas
RLAIF-V: Aligning MLLMs through Open-Source AI Feedback for Super GPT-4V Trustworthiness
[CVPR'24] RLHF-V: Towards Trustworthy MLLMs via Behavior Alignment from Fine-grained Correctional Human Feedback
[CVPR 2024 Oral] InternVL Family: A Pioneering Open-Source Alternative to GPT-4o. 接近GPT-4o表现的可商用开源多模态对话模型
LLMs build upon Evol Insturct: WizardLM, WizardCoder, WizardMath
The official GitHub page for the survey paper "A Survey of Large Language Models".
Harnessing 1.4M GPT4V-synthesized Data for A Lite Vision-Language Model