- Stanford
- https://twitter.com/karpathy
Highlights
- Pro
Stars
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
Tensors and Dynamic neural networks in Python with strong GPU acceleration
Robust Speech Recognition via Large-Scale Weak Supervision
scikit-learn: machine learning in Python
OpenAssistant is a chat-based assistant that understands tasks, can interact with third-party systems, and retrieve information dynamically to do so.
LlamaIndex is a data framework for your LLM applications
The largest collection of PyTorch image encoders / backbones. Including train, eval, inference, export scripts, and pretrained weights -- ResNet, ResNeXT, EfficientNet, NFNet, Vision Transformer (V…
Code and documentation to train Stanford's Alpaca models, and generate the data.
A high-throughput and memory-efficient inference and serving engine for LLMs
🤗 Diffusers: State-of-the-art diffusion models for image and audio generation in PyTorch and FLAX.
Implementation of Vision Transformer, a simple way to achieve SOTA in vision classification with only a single transformer encoder, in Pytorch
Finetune Llama 3.1, Mistral, Phi & Gemma LLMs 2-5x faster with 80% less memory
DALL·E Mini - Generate images from a text prompt
Image augmentation for machine learning experiments.
Official inference repo for FLUX.1 models
Fast and memory-efficient exact attention
End-to-End Object Detection with Transformers
20+ high-performance LLMs with recipes to pretrain, finetune and deploy at scale.
Flexible and powerful tensor operations for readable and reliable code (for pytorch, jax, TF and others)
Hackable and optimized Transformers building blocks, supporting a composable construction.
The TinyLlama project is an open endeavor to pretrain a 1.1B Llama model on 3 trillion tokens.
SkyPilot: Run AI and batch jobs on any infra (Kubernetes or 12+ clouds). Get unified execution, cost savings, and high GPU availability via a simple interface.
Implementation of the LLaMA language model based on nanoGPT. Supports flash attention, Int8 and GPTQ 4bit quantization, LoRA and LLaMA-Adapter fine-tuning, pre-training. Apache 2.0-licensed.
Simple and efficient pytorch-native transformer text generation in <1000 LOC of python.
The official PyTorch implementation of Google's Gemma models
min(DALL·E) is a fast, minimal port of DALL·E Mini to PyTorch
StudioGAN is a Pytorch library providing implementations of representative Generative Adversarial Networks (GANs) for conditional/unconditional image generation.
Official implementation for the paper: "Code Generation with AlphaCodium: From Prompt Engineering to Flow Engineering""