Popular repositories Loading
-
Megatron-LM
Megatron-LM PublicForked from NVIDIA/Megatron-LM
Ongoing research training transformer models at scale
Python
-
DeepSpeed
DeepSpeed PublicForked from microsoft/DeepSpeed
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
Python
-
nanoLM
nanoLM PublicForked from cofe-ai/nanoLM
An Affordable LLM Pre-training Benchmark via Accurate Loss Prediction across Scales
Python
Something went wrong, please refresh the page to try again.
If the problem persists, check the GitHub status page or contact support.
If the problem persists, check the GitHub status page or contact support.