-
Shanghai AI Lab
- shanghai
Pinned Loading
-
deepLearningSystem2022
deepLearningSystem2022 PublicHomework for Deep Learning Systems 2022.
-
flash-attention
flash-attention PublicForked from Dao-AILab/flash-attention
Fast and memory-efficient exact attention
Python
-
InternLM
InternLM PublicForked from InternLM/InternLM
InternLM has open-sourced a 7 billion parameter base model, a chat model tailored for practical scenarios and the training system.
Python
-
XYT-DeepSpeed
XYT-DeepSpeed PublicForked from microsoft/DeepSpeed
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
Python
-
XYT-Megatron-DeepSpeed
XYT-Megatron-DeepSpeed PublicForked from microsoft/Megatron-DeepSpeed
Ongoing research training transformer language models at scale, including: BERT & GPT-2
Python
If the problem persists, check the GitHub status page or contact support.