Skip to content
OpenCatalogcurated by FLOSSK
AI & Machine Learning

DeepSpeed

Microsoft library for extreme-scale model training: ZeRO optimizer states, pipeline parallelism, and inference kernels.

Why it is included

Core OSS component behind many large-model training recipes complementing PyTorch.

Best for

Teams training or fine-tuning large models where memory fragmentation is the bottleneck.

Strengths

  • ZeRO stages
  • Large-model recipes
  • Inference optimizations

Limitations

  • Primarily NVIDIA CUDA ecosystems; integration testing burden

Good alternatives

FSDP · Megatron-LM · Horovod

Related tools