Stars
Solutions to all questions of the book Introduction to the Theory of Computation, 3rd edition by Michael Sipser
A library for advanced large language model reasoning
Advanced AI Explainability for computer vision. Support for CNNs, Vision Transformers, Classification, Object detection, Segmentation, Image similarity and more.
Unofficial implementation of YOLO-World + EfficientSAM for ComfyUI
Visualizing the attention of vision-language models
A collection of LLM papers, blogs, and projects, with a focus on OpenAI o1 and reasoning techniques.
MiniCPM-V 2.6: A GPT-4V Level MLLM for Single Image, Multi Image and Video on Your Phone
Using FlexAttention to compute attention with different masking patterns
Minimalistic large language model 3D-parallelism training
LMDeploy is a toolkit for compressing, deploying, and serving LLMs.
Repository for Show-o, One Single Transformer to Unify Multimodal Understanding and Generation.
🔥 Turn entire websites into LLM-ready markdown or structured data. Scrape, crawl and extract with a single API.
Helpful tools and examples for working with flex-attention
Collection of AWESOME vision-language models for vision tasks
A custom Huggingface trainer which supports logging auxiliary losses returned by your model
Rich is a Python library for rich text and beautiful formatting in the terminal.
Public code repo for paper "A Single Transformer for Scalable Vision-Language Modeling"
[NeurIPS'24 Spotlight] EVE: Encoder-Free Vision-Language Models
Open-source evaluation toolkit of large vision-language models (LVLMs), support ~100 VLMs, 40+ benchmarks
[NeurIPS'23 Oral] Visual Instruction Tuning (LLaVA) built towards GPT-4V level capabilities and beyond.
SGLang is a fast serving framework for large language models and vision language models.
Accelerating the development of large multimodal models (LMMs) with lmms-eval
This repository contains demos I made with the Transformers library by HuggingFace.
Materials for the Hugging Face Diffusion Models Course
A simple toolkit for benchmarking LLMs on mathematical reasoning tasks. 🧮✨
An Easy-to-use, Scalable and High-performance RLHF Framework (70B+ PPO Full Tuning & Iterative DPO & LoRA & RingAttention)
[NeurIPS'24] Official code for *🎯DART-Math: Difficulty-Aware Rejection Tuning for Mathematical Problem-Solving*
A framework for few-shot evaluation of language models.
The official implementation of Self-Play Fine-Tuning (SPIN)
A recipe for online RLHF and online iterative DPO.