Popular repositories Loading
-
Megatron-DeepSpeed
Megatron-DeepSpeed PublicForked from microsoft/Megatron-DeepSpeed
Ongoing research training transformer language models at scale, including: BERT & GPT-2
Python 1
-
-
QDrop
QDrop PublicForked from wimh966/QDrop
The official PyTorch implementation of the ICLR2022 paper, QDrop: Randomly Dropping Quantization for Extremely Low-bit Post-Training Quantization
Python
-
outlier_suppression
outlier_suppression PublicForked from wimh966/outlier_suppression
The official PyTorch implementation of the NeurIPS2022 paper, Outlier Suppression: Pushing the Limit of Low-bit Transformer Language Models
Python
-
Pai-Megatron-Patch
Pai-Megatron-Patch PublicForked from alibaba/Pai-Megatron-Patch
The official repo of Pai-Megatron-Patch for LLM & VLM large scale training developed by Alibaba Cloud.
Python
Something went wrong, please refresh the page to try again.
If the problem persists, check the GitHub status page or contact support.
If the problem persists, check the GitHub status page or contact support.