Acronym and Abbreviation Description
Ascend Name of Huawei Ascend series chips.
CCE Cube-based Computing Engine, which is an operator development tool oriented to hardware architecture programming.
CCE-C Cube-based Computing Engine C, which is C code developed by the CCE.
CheckPoint MindSpore model training check point, which is used to save model parameters for inference or retraining.
CIFAR-10 An open-source image data set that contains 60000 32 x 32 color images of 10 categories, with 6000 images of each category. There are 50000 training images and 10000 test images.
CIFAR-100 An open-source image data set that contains 100 categories. Each category contains 600 images. Each course has 500 training images and 100 test images.
DaVinci DaVinci architecture, Huawei-developed new chip architecture.
EulerOS Euler operating system, which is developed by Huawei based on the standard Linux kernel.
FC Layer Fully connected layer, which acts as a classifier in the entire convolutional neural network.
FE Fusion Engine, which connects to GE and TBE operators and has the capabilities of loading and managing the operator information library and managing convergence rules.
FP16 16-bit floating point, which is a half-precision floating point arithmetic format, consuming less memory.
FP32 32-bit floating point, which is a single-precision floating point arithmetic format.
GE Graph Engine, MindSpore computational graph execution engine, which is responsible for optimizing hardware (such as operator fusion and memory overcommitment) based on the front-end computational graph and starting tasks on the device side.
GHLO Graph High Level Optimization. GHLO includes optimization irrelevant to hardware (such as dead code elimination), auto parallel, and auto differentiation.
GLLO Graph Low Level Optimization. GLLO includes hardware-related optimization and in-depth optimization related to the combination of hardware and software, such as operator fusion and buffer fusion.
Graph Mode MindSpore static graph mode. In this mode, the neural network model is compiled into an entire graph and then delivered for execution, featuring high performance.
HCCL Huawei Collective Communication Library, which implements multi-device and multi-card communication based on the Da Vinci architecture chip.
ImageNet Image database organized based on the WordNet hierarchy (currently nouns only).
LeNet A classical convolutional neural network architecture proposed by Yann LeCun and others.
Loss Difference between the predicted value and the actual value, which is a standard for determining the model quality of deep learning.
LSTM Long short-term memory, an artificial recurrent neural network (RNN) architecture used for processing and predicting an important event with a long interval and delay in a time sequence.
Manifest A data format file. Huawei ModelArt adopts this format. For details, see
ME Mind Expression, MindSpore frontend, which is used to compile tasks from user source code to computational graphs, control execution during training, maintain contexts (in non-sink mode), and dynamically generate graphs (in PyNative mode).
MindArmour MindSpore security component, which is used for AI adversarial example management, AI model attack defense and enhancement, and AI model robustness evaluation.
MindData MindSpore data framework, which provides data loading, enhancement, dataset management, and visualization.
MindInsight MindSpore visualization component, which visualizes information such as scalars, images, computational graphs, and model hyperparameters.
MindSpore Huawei-leaded open-source deep learning framework.
MindSpore Predict A lightweight deep neural network inference engine that provides the inference function for models trained by MindSpore on the device side.
MNIST database Modified National Handwriting of Images and Technology database, a large handwritten digit database, which is usually used to train various image processing systems.
PyNative Mode MindSpore dynamic graph mode. In this mode, operators in the neural network are delivered and executed one by one, facilitating the compilation and debugging of the neural network model.
ResNet-50 Residual Neural Network 50, a residual neural network proposed by four Chinese people, including Kaiming He from Microsoft Research Institute.
Schema Data set structure definition file, which defines the fields contained in a data set and the field types.
Summary An operator that monitors the values of tensors on the network. It is a peripheral operation in the figure and does not affect the data flow.
TBE Tensor Boost Engine, an operator development tool that is extended based on the Tensor Virtual Machine (TVM) framework.
TFRecord Data format defined by TensorFlow.