Stars
官方权威数据:统计年签,统计公报,互联网行业报告,工信部数据,ICT报告等 Official authoritative data (Chinese)
📰 Must-read papers and blogs on Speculative Decoding ⚡️
Basic books for better knowing of Linux and shells
A curated list for Efficient Large Language Models
[ICML 2024] CLLMs: Consistency Large Language Models
PiSSA: Principal Singular Values and Singular Vectors Adaptation of Large Language Models(NeurIPS 2024 Spotlight)
A curated list of practical guide resources of LLMs (LLMs Tree, Examples, Papers)
🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.
The official GitHub page for the survey paper "A Survey on Evaluation of Large Language Models".
Artificial Intelligence Research for Science (AIRS)
microsoft / Megatron-DeepSpeed
Forked from NVIDIA/Megatron-LMOngoing research training transformer language models at scale, including: BERT & GPT-2
Central place for the engineering/scaling WG: documentation, SLURM scripts and logs, compute environment and data.
Python packaging and dependency management made easy
Data validation using Python type hints
Data manipulation and transformation for audio signal processing, powered by PyTorch
Machine learning metrics for distributed, scalable PyTorch applications.
Pretrain, finetune ANY AI model of ANY size on multiple GPUs, TPUs with zero code changes.
Hydra is a framework for elegantly configuring complex applications
FastAPI framework, high performance, easy to learn, fast to code, ready for production