Popular repositories Loading
-
Megatron-DeepSpeed
Megatron-DeepSpeed PublicForked from microsoft/Megatron-DeepSpeed
Ongoing research training transformer language models at scale, including: BERT & GPT-2
Python 1
-
-
QDrop
QDrop PublicForked from wimh966/QDrop
The official PyTorch implementation of the ICLR2022 paper, QDrop: Randomly Dropping Quantization for Extremely Low-bit Post-Training Quantization
Python
-
outlier_suppression
outlier_suppression PublicForked from wimh966/outlier_suppression
The official PyTorch implementation of the NeurIPS2022 paper, Outlier Suppression: Pushing the Limit of Low-bit Transformer Language Models
Python
-
Pai-Megatron-Patch
Pai-Megatron-Patch PublicForked from alibaba/Pai-Megatron-Patch
The official repo of Pai-Megatron-Patch for LLM & VLM large scale training developed by Alibaba Cloud.
Python
If the problem persists, check the GitHub status page or contact support.