|
- sections: |
|
- local: index |
|
title: 🤗 Accelerate |
|
- local: basic_tutorials/install |
|
title: Installation |
|
- local: quicktour |
|
title: Quicktour |
|
title: Getting started |
|
- sections: |
|
- local: basic_tutorials/overview |
|
title: Overview |
|
- local: basic_tutorials/migration |
|
title: Migrating to 🤗 Accelerate |
|
- local: basic_tutorials/launch |
|
title: Launching distributed code |
|
- local: basic_tutorials/notebook |
|
title: Launching distributed training from Jupyter Notebooks |
|
- local: basic_tutorials/troubleshooting |
|
title: Troubleshooting guide |
|
title: Tutorials |
|
- sections: |
|
- local: usage_guides/explore |
|
title: Start Here! |
|
- local: usage_guides/training_zoo |
|
title: Example Zoo |
|
- local: usage_guides/big_modeling |
|
title: How to perform inference on large models with small resources |
|
- local: usage_guides/model_size_estimator |
|
title: Knowing how big of a model you can fit into memory |
|
- local: usage_guides/quantization |
|
title: How to quantize model |
|
- local: usage_guides/distributed_inference |
|
title: How to perform distributed inference with normal resources |
|
- local: usage_guides/gradient_accumulation |
|
title: Performing gradient accumulation |
|
- local: usage_guides/local_sgd |
|
title: Accelerating training with local SGD |
|
- local: usage_guides/checkpoint |
|
title: Saving and loading training states |
|
- local: usage_guides/tracking |
|
title: Using experiment trackers |
|
- local: usage_guides/mps |
|
title: How to use Apple Silicon M1 GPUs |
|
- local: usage_guides/low_precision_training |
|
title: How to train in low precision (FP8) |
|
- local: usage_guides/deepspeed |
|
title: How to use DeepSpeed |
|
- local: usage_guides/fsdp |
|
title: How to use Fully Sharded Data Parallelism |
|
- local: usage_guides/megatron_lm |
|
title: How to use Megatron-LM |
|
- local: usage_guides/sagemaker |
|
title: How to use 🤗 Accelerate with SageMaker |
|
- local: usage_guides/ipex |
|
title: How to use 🤗 Accelerate with Intel® Extension for PyTorch for cpu |
|
title: How-To Guides |
|
- sections: |
|
- local: concept_guides/internal_mechanism |
|
title: 🤗 Accelerate's internal mechanism |
|
- local: concept_guides/big_model_inference |
|
title: Loading big models into memory |
|
- local: concept_guides/performance |
|
title: Comparing performance across distributed setups |
|
- local: concept_guides/deferring_execution |
|
title: Executing and deferring jobs |
|
- local: concept_guides/gradient_synchronization |
|
title: Gradient synchronization |
|
- local: concept_guides/low_precision_training |
|
title: How training in low-precision environments is possible (FP8) |
|
- local: concept_guides/training_tpu |
|
title: TPU best practices |
|
title: Concepts and fundamentals |
|
- sections: |
|
- local: package_reference/accelerator |
|
title: Main Accelerator class |
|
- local: package_reference/state |
|
title: Stateful configuration classes |
|
- local: package_reference/cli |
|
title: The Command Line |
|
- local: package_reference/torch_wrappers |
|
title: Torch wrapper classes |
|
- local: package_reference/tracking |
|
title: Experiment trackers |
|
- local: package_reference/launchers |
|
title: Distributed launchers |
|
- local: package_reference/deepspeed |
|
title: DeepSpeed utilities |
|
- local: package_reference/logging |
|
title: Logging |
|
- local: package_reference/big_modeling |
|
title: Working with large models |
|
- local: package_reference/kwargs |
|
title: Kwargs handlers |
|
- local: package_reference/utilities |
|
title: Utility functions and classes |
|
- local: package_reference/megatron_lm |
|
title: Megatron-LM Utilities |
|
- local: package_reference/fsdp |
|
title: Fully Sharded Data Parallelism Utilities |
|
title: "Reference" |
|
|