Lists (12)
Sort Name ascending (A-Z)
Starred repositories
Meta Lingua: a lean, efficient, and easy-to-hack codebase to research LLMs.
Align Anything: Training All-modality Model with Feedback
A bibliography and survey of the papers surrounding o1
Educational framework exploring ergonomic, lightweight multi-agent orchestration. Managed by OpenAI Solution team.
Composable building blocks to build Llama Apps
[ICML 2024 Best Paper] Discrete Diffusion Modeling by Estimating the Ratios of the Data Distribution (https://arxiv.org/abs/2310.16834)
Chai-1, SOTA model for biomolecular structure prediction
Pytorch implementation of Transfusion, "Predict the Next Token and Diffuse Images with One Multi-Modal Model", from MetaAI
Recipes for shrinking, optimizing, customizing cutting edge vision models. 💜
PyTorch implementation of MAR+DiffLoss https://arxiv.org/abs/2406.11838
🦙 LaMa Image Inpainting, Resolution-robust Large Mask Inpainting with Fourier Convolutions, WACV 2022
Utilities intended for use with Llama models.
Official Repository of MMLONGBENCH-DOC: Benchmarking Long-context Document Understanding with Visualizations
Anole: An Open, Autoregressive and Native Multimodal Models for Interleaved Image-Text Generation
USP: Unified (a.k.a. Hybrid, 2D) Sequence Parallel Attention for Long Context Transformers Model Training and Inference
Representation Engineering: A Top-Down Approach to AI Transparency
[NeurIPS 2024 Oral][GPT beats diffusion🔥] [scaling laws in visual generation📈] Official impl. of "Visual Autoregressive Modeling: Scalable Image Generation via Next-Scale Prediction". An *ultra-sim…
This repo contains the code for 1D tokenizer and generator
Repository for Meta Chameleon, a mixed-modal early-fusion foundation model from FAIR.
MINT-1T: A one trillion token multimodal interleaved dataset.
Open-MAGVIT2: Democratizing Autoregressive Visual Generation