Stars
[ECCV 2024] MVSGaussian: Fast Generalizable Gaussian Splatting Reconstruction from Multi-View Stereo
PyTorch implementation of InstructDiffusion, a unifying and generic framework for aligning computer vision tasks with human instructions.
Qwen2-VL is the multimodal large language model series developed by Qwen team, Alibaba Cloud.
✨✨Latest Advances on Multimodal Large Language Models
Official code of SmartEdit [CVPR-2024 Highlight]
MLNLP社区用来帮助大家避免论文投稿小错误的整理仓库。 Paper Writing Tips
[ICLR 2024 Oral] Generative Gaussian Splatting for Efficient 3D Content Creation
The code releasing for https://image-dream.github.io/
[ CVPR 2024 ] Implementation for "GPT-4V(ision) is a Human-Aligned Evaluator for Text-to-3D Generation"
A PyTorch implementation of NeRF (Neural Radiance Fields) that reproduces the results.
Instruct-NeRF2NeRF: Editing 3D Scenes with Instructions (ICCV 2023)
Text-to-3D & Image-to-3D & Mesh Exportation with NeRF + Diffusion.
🔥🔥🔥 A curated list of papers on LLMs-based multimodal generation (image, video, 3D and audio).
A curated list of awesome 3d generation papers
This reposiotry is the collection for public 3D LiDAR datasets
[GPT beats diffusion🔥] [scaling laws in visual generation📈] Official impl. of "Visual Autoregressive Modeling: Scalable Image Generation via Next-Scale Prediction". An *ultra-simple, user-friendly …
This may be the simplest implement of DDPM. You can directly run Main.py to train the UNet on CIFAR-10 dataset and see the amazing process of denoising.
PyTorch reimplementation of Diffusion Models
Official PyTorch Implementation of "Scalable Diffusion Models with Transformers"
Code for paper <PointContrast: Unsupervised Pretraining for 3D Point Cloud Understanding>
Github Pages template for academic personal websites, forked from mmistakes/minimal-mistakes
OpenCompass is an LLM evaluation platform, supporting a wide range of models (Llama3, Mistral, InternLM2,GPT-4,LLaMa2, Qwen,GLM, Claude, etc) over 100+ datasets.
OpenLLaMA, a permissively licensed open source reproduction of Meta AI’s LLaMA 7B trained on the RedPajama dataset
Code for "Lion: Adversarial Distillation of Proprietary Large Language Models (EMNLP 2023)"
A framework for few-shot evaluation of language models.