Semi-automatic Parallel =========================== .. image:: https://mindspore-website.obs.cn-north-4.myhuaweicloud.com/website-images/r2.4.1/resource/_static/logo_source_en.svg :target: https://gitee.com/mindspore/docs/blob/r2.4.1/docs/mindspore/source_en/model_train/parallel/semi_auto_parallel.rst :alt: View Source on Gitee .. toctree:: :maxdepth: 1 :hidden: operator_parallel advanced_operator_parallel optimizer_parallel pipeline_parallel Semi-automatic parallel supports the automatic mixing of multiple parallel modes, including: - `Operator-level parallel `_: Operator-level parallel refers to slicing the input tensor and model parameters into multiple devices for computation on an operator basis to improve overall speed. - `Higher-order Operator-level Parallelism `_: Higher-order operator-level parallelism refers to operator-level parallelism that allows customized device layout with tensor layout for more complex sharding logic. - `Optimizer parallel `_: Optimizer parallel reduces redundant computations on multiple devices for the same weight updates, spreading the computation over multiple devices. - `Pipeline parallel `_: Pipeline parallel means that the model is sliced by layer, with each device processing only a certain part of the model.