Highlights
- Pro
Stars
C implementation of gRPC layered on top of core library
This project aims to collect the latest "call for reviewers" links from various top CS/ML/AI conferences/journals
Robust recipes to align language models with human and AI preferences
Skywork-MoE: A Deep Dive into Training Techniques for Mixture-of-Experts Language Models
High-speed Large Language Model Serving on PCs with Consumer-grade GPUs
The official evaluation suite and dynamic data release for MixEval.
A generative speech model for daily dialogue.
learning notes when learning the source code of pytorch
microsoft / Megatron-DeepSpeed
Forked from NVIDIA/Megatron-LMOngoing research training transformer language models at scale, including: BERT & GPT-2
Repo for Rho-1: Token-level Data Selection & Selective Pretraining of LLMs.
Custom console scripts for Dota 2.
Open-Sora: Democratizing Efficient Video Production for All
AI Infra主要是指AI的基础建设,包括AI芯片、AI编译器、AI推理和训练框架等AI全栈底层技术。
Reaching LLaMA2 Performance with 0.1M Dollars
Benchmark LLMs by fighting in Street Fighter 3! The new way to evaluate the quality of an LLM
Longitudinal Evaluation of LLMs via Data Compression
📖A curated list of Awesome LLM Inference Paper with codes, TensorRT-LLM, vLLM, streaming-llm, AWQ, SmoothQuant, WINT8/4, Continuous Batching, FlashAttention, PagedAttention etc.
Code examples and resources for DBRX, a large language model developed by Databricks
We introduce a novel approach for parameter generation, named neural network parameter diffusion (p-diff), which employs a standard latent diffusion model to synthesize a new set of parameters
[CVPR2024 highlight] Generalized Large-Scale Data Condensation via Various Backbone and Statistical Matching (G-VBSM)
Minimal, clean code for the Byte Pair Encoding (BPE) algorithm commonly used in LLM tokenization.
The simplest, fastest repository for training/finetuning medium-sized GPTs.