MindSpore Transformers Documentation ===================================== The goal of MindSpore Transformers (also known as MindFormers) suite is to build a full-process development suite for training, fine-tuning, evaluating, inference, and deploying large models, providing the industry mainstream Transformer class of pre-trained models and SOTA downstream task applications, and covering a rich range of parallel features, with the expectation of helping users to easily realize large model training and innovative research and development. Users can refer to `Overall Architecture `_ and `Model Library `_ to get an initial understanding of MindFormers architecture and model support. Refer to the `Installation `_ and `Quick Start `_ to get started with MindFormers. If you have any suggestions for MindFormers, please contact us via `issue `_ and we will handle them promptly. MindFormers supports one-click start of single/multi-card training, fine-tuning, evaluation, and inference processes for any task, which makes the execution of deep learning tasks more efficient and user-friendly by simplifying the operation, providing flexibility, and automating the process. Users can learn from the following explanatory documents: - `Development Migration `_ - `Pretraining `_ - `SFT Tuning `_ - `Parameter-Efficient Fine-Tuning (PEFT) `_ - `Evaluation `_ - `Inference `_ - `Quantization `_ - `MindIE Service Deployment `_ Flexible and Easy-to-Use Personalized Configuration with MindFormers ---------------------------------------------------------------------- With its powerful feature set, MindFormers provides users with flexible and easy-to-use personalized configuration options. Specifically, it comes with the following key features: 1. `Weight Format Conversion `_ Provides a unified weight conversion tool that converts model weights between the formats used by HuggingFace and MindFormers. 2. `Distributed Weight Slicing and Merging `_ Weights in different distributed scenarios are flexibly sliced and merged. 3. `Distributed Parallel `_ One-click configuration of multi-dimensional hybrid distributed parallel allows models to run efficiently in clusters up to 10,000 cards. 4. `Dataset `_ Support multiple forms of datasets. 5. `Weight Saving and Resumable Training After Breakpoint `_ Supports step-level resumable training after breakpoint, effectively reducing the waste of time and resources caused by unexpected interruptions during large-scale training. Deep Optimizing with MindFormers ------------------------------------ - `Precision Optimizing `_ - `Performance Optimizing `_ Appendix ------------------------------------ - `Environment Variables Descriptions `_ - `Configuration File Descriptions `_ FAQ ------------------------------------ - `Model-Related `_ - `Function-Related `_ - `MindFormers Contribution Guide `_ - `Modelers Contribution Guide `_ .. toctree:: :glob: :maxdepth: 1 :caption: Start :hidden: start/overview start/models .. toctree:: :glob: :maxdepth: 1 :caption: Quick Start :hidden: quick_start/install quick_start/source_code_start .. toctree:: :glob: :maxdepth: 1 :caption: Usage Tutorials :hidden: usage/dev_migration usage/pre_training usage/sft_tuning usage/parameter_efficient_fine_tune usage/evaluation usage/inference usage/quantization usage/mindie_deployment .. toctree:: :glob: :maxdepth: 1 :caption: Function Description :hidden: function/weight_conversion function/transform_weight function/distributed_parallel function/dataset function/resume_training .. toctree:: :glob: :maxdepth: 1 :caption: Precision Optimization :hidden: acc_optimize/acc_optimize .. toctree:: :glob: :maxdepth: 1 :caption: Performance Optimization :hidden: perf_optimize/perf_optimize .. toctree:: :maxdepth: 1 :caption: API :hidden: mindformers mindformers.core mindformers.dataset mindformers.generation mindformers.models mindformers.modules mindformers.pet mindformers.pipeline mindformers.tools mindformers.wrapper .. toctree:: :glob: :maxdepth: 1 :caption: Appendix :hidden: appendix/env_variables appendix/conf_files .. toctree:: :glob: :maxdepth: 1 :caption: FAQ :hidden: faq/model_related faq/func_related faq/mindformers_contribution faq/modelers_contribution .. toctree:: :glob: :maxdepth: 1 :caption: RELEASE NOTES :hidden: RELEASE