-
Tsinghua University
- China
Stars
Minimalistic large language model 3D-parallelism training
Contrastive Language-Image Forensic Search allows free text searching through videos using OpenAI's machine learning model CLIP
Tools for merging pretrained large language models.
Masked Structural Growth for 2x Faster Language Model Pre-training
Official implementation of "A Multi-level Framework for Accelerating Training Transformer Models""
AISystem 主要是指AI系统,包括AI芯片、AI编译器、AI推理和训练框架等AI全栈底层技术
Tools to download and cleanup Common Crawl data
A series of large language models developed by Baichuan Intelligent Technology
Code for loralib, an implementation of "LoRA: Low-Rank Adaptation of Large Language Models"
Minecraft 1.3.2-1.15.2 Vanilla and FML CoreMod Development Tutorial.
A 13B large language model developed by Baichuan Intelligent Technology
A large-scale 7B pretraining language model developed by BaiChuan-Inc.
Code and documentation to train Stanford's Alpaca models, and generate the data.
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
ncnn is a high-performance neural network inference framework optimized for the mobile platform
Easy Parallel Library (EPL) is a general and efficient deep learning framework for distributed model training.
DeeBERT: Dynamic Early Exiting for Accelerating BERT Inference
Library of deep learning models and datasets designed to make deep learning more accessible and accelerate ML research.
Code for SkipNet: Learning Dynamic Routing in Convolutional Networks (ECCV 2018)
A tool for extracting plain text from Wikipedia dumps
CodiMD - Realtime collaborative markdown notes on all platforms.