[{"data":1,"prerenderedAt":731},["ShallowReactive",2],{"content-query-bkMV1GjrPX":3},{"_path":4,"_dir":5,"_draft":6,"_partial":6,"_locale":7,"title":8,"description":10,"date":11,"cover":12,"type":13,"category":14,"body":15,"_type":725,"_id":726,"_source":727,"_file":728,"_stem":729,"_extension":730},"/technology-blogs/zh/1556","zh",false,"",[9],"MindSpore论文12","对训练过程的简单调整对于性能提升也起到了不可忽视的作用","2022-06-17","https://obs-mindspore-file.obs.cn-north-4.myhuaweicloud.com/file/2022/06/22/c7e09cb99fe1425f9ccbaa65f7fbde3c.png","technology-blogs","大V博文",{"type":16,"children":17,"toc":707},"root",[18,32,38,51,58,63,69,74,80,85,93,99,104,114,126,132,138,143,153,160,168,175,183,190,198,205,213,220,225,230,238,243,251,256,264,269,276,281,289,294,302,307,314,319,324,329,337,342,350,355,363,368,373,378,386,391,399,404,412,417,424,430,440,447,456,463,472,483,492,499,504,511,516,523,528,534,539,544,549,554,559,563,568,573,578,583,588,593,598,603,608,613,618,623,628,633,638,644,649,654,666,678,683,695],{"type":19,"tag":20,"props":21,"children":23},"element","h1",{"id":22},"mindspore论文12aaai长尾问题中训练技巧的总结",[24,30],{"type":19,"tag":25,"props":26,"children":27},"span",{},[28],{"type":29,"value":9},"text",{"type":29,"value":31},"AAAI长尾问题中训练技巧的总结",{"type":19,"tag":33,"props":34,"children":35},"p",{},[36],{"type":29,"value":37},"作者：李锐锋",{"type":19,"tag":33,"props":39,"children":40},{},[41,43],{"type":29,"value":42},"文章来源：",{"type":19,"tag":44,"props":45,"children":49},"a",{"href":46,"rel":47},"https://zhuanlan.zhihu.com/p/527353389",[48],"nofollow",[50],{"type":29,"value":46},{"type":19,"tag":52,"props":53,"children":55},"h2",{"id":54},"_1研究背景",[56],{"type":29,"value":57},"1.研究背景",{"type":19,"tag":33,"props":59,"children":60},{},[61],{"type":29,"value":62},"近年来，基于元学习等各种复杂范式，长尾分布上的视觉识别取得了很大进展。而除了这些复杂方法外，对训练过程的简单调整对于性能提升也起到了不可忽视的作用。这些调整虽然简单却很有效，如对数据分布或是损失函数权重的调整。但不同的技巧可能会相互冲突，如果使用不当，可能会导致识别准确度比预期更差。不幸的是，在现有文献中还没有关于这些技巧的科学指导。",{"type":19,"tag":52,"props":64,"children":66},{"id":65},"_2作者介绍",[67],{"type":29,"value":68},"2.作者介绍",{"type":19,"tag":33,"props":70,"children":71},{},[72],{"type":29,"value":73},"魏秀参，博士，南京理工大学计算机科学与工程学院教授（PCA Lab骨干成员），南京大学学生创业导师。主要研究领域为计算机视觉和机器学习，在相关领域国际顶级期刊和会议发表论文四十余篇，Google Scholar Citations 近2000次，相关研究成果获得含iNaturalist在内的计算机视觉领域国际权威赛事共4项世界冠军。曾在CVPR、ICME等国际会议讲授“图像细粒度分析”为主题的短课程。著有《解析深度学习–卷积神经网络原理与视觉实践》一书。曾获江苏省科协青年人才托举工程、南京经开区中青年优秀人才、南京理工大学青年拔尖人才、《中国科学：信息科学》优秀评审人、CVPR 2017 Best PC Member等荣誉。任中国计算机学会高级会员、CCF计算机视觉专委会委员、中国图象图形学报青年编委。担任ICCV、IJCAI、ACM Multimedia、ACCV等国际会议Workshop程序委员会主席，ACCV 2022 Tutorial主席，IJCAI 2021高级程序委员，二十余次CCF-A类会议的程序委员，及IEEE TPAMI、TIP、TNNLS、MLJ、TMM等国际期刊审稿人。",{"type":19,"tag":52,"props":75,"children":77},{"id":76},"_3论文主要内容简介",[78],{"type":29,"value":79},"3.论文主要内容简介",{"type":19,"tag":33,"props":81,"children":82},{},[83],{"type":29,"value":84},"本文收集了已有的长尾视觉识别技巧，然后进行了广泛的系统性实验，给出了详细的实验指导，并获得了这些技巧的有效组合。此外，本文还提出了一种基于类激活映射的长尾视觉数据增强方法。该方法能够与重采样技巧结合，并取得了良好的结果。通过科学地组合这些技巧，可以在4个长尾基准数据集上超越最先进的方法，包括ImageNet-LT和iNaturalist 2018。",{"type":19,"tag":33,"props":86,"children":87},{},[88],{"type":19,"tag":89,"props":90,"children":92},"img",{"alt":7,"src":91},"https://pic3.zhimg.com/80/v2-e6341db4df1087c24b67682c72e8b582_720w.jpg",[],{"type":19,"tag":52,"props":94,"children":96},{"id":95},"_4代码链接",[97],{"type":29,"value":98},"4.代码链接",{"type":19,"tag":33,"props":100,"children":101},{},[102],{"type":29,"value":103},"论文链接：",{"type":19,"tag":33,"props":105,"children":106},{},[107],{"type":19,"tag":44,"props":108,"children":111},{"href":109,"rel":110},"https://link.zhihu.com/?target=https%3A//ojs.aaai.org/index.php/AAAI/article/view/16458",[48],[112],{"type":29,"value":113},"https://ojs.aaai.org/index.php/AAAI/article/view/16458",{"type":19,"tag":33,"props":115,"children":116},{},[117,119],{"type":29,"value":118},"代码链接：",{"type":19,"tag":44,"props":120,"children":123},{"href":121,"rel":122},"https://link.zhihu.com/?target=https%3A//gitee.com/mindspore/contrib/tree/772881cee0887d1e7d6c3e2f297f528b37fc466d/papers/bag_of_tricks-LT",[48],[124],{"type":29,"value":125},"https://gitee.com/mindspore/contrib/tree/772881cee0887d1e7d6c3e2f297f528b37fc466d/papers/bag_of_tricks-LT",{"type":19,"tag":52,"props":127,"children":129},{"id":128},"_5算法框架技术要点",[130],{"type":29,"value":131},"5.算法框架技术要点",{"type":19,"tag":133,"props":134,"children":136},"h3",{"id":135},"重加权法",[137],{"type":29,"value":135},{"type":19,"tag":33,"props":139,"children":140},{},[141],{"type":29,"value":142},"成本敏感重加权法是长尾识别中常用的方法，这些方法通过对不同类别赋予不同的权重，引导网络更加关注少数类。本文比较了一个基线损失和4种重加权损失，分别为：",{"type":19,"tag":144,"props":145,"children":146},"ul",{},[147],{"type":19,"tag":148,"props":149,"children":150},"li",{},[151],{"type":29,"value":152},"Cross-entropy loss (CE)",{"type":19,"tag":33,"props":154,"children":155},{},[156],{"type":19,"tag":89,"props":157,"children":159},{"alt":7,"src":158},"https://pic1.zhimg.com/80/v2-24a60efef6dc4f8ce3016781d1eb3914_720w.jpg",[],{"type":19,"tag":144,"props":161,"children":162},{},[163],{"type":19,"tag":148,"props":164,"children":165},{},[166],{"type":29,"value":167},"Cost-sensitive cross-entropy loss (CS_CE)",{"type":19,"tag":33,"props":169,"children":170},{},[171],{"type":19,"tag":89,"props":172,"children":174},{"alt":7,"src":173},"https://pic2.zhimg.com/80/v2-c9989b0068ed1974936a344f45557965_720w.jpg",[],{"type":19,"tag":144,"props":176,"children":177},{},[178],{"type":19,"tag":148,"props":179,"children":180},{},[181],{"type":29,"value":182},"Class-balanced cross-entropy loss (CB_CE)",{"type":19,"tag":33,"props":184,"children":185},{},[186],{"type":19,"tag":89,"props":187,"children":189},{"alt":7,"src":188},"https://pic4.zhimg.com/80/v2-774df87bc731ea6aa834ce329ccf2177_720w.jpg",[],{"type":19,"tag":144,"props":191,"children":192},{},[193],{"type":19,"tag":148,"props":194,"children":195},{},[196],{"type":29,"value":197},"Focal loss",{"type":19,"tag":33,"props":199,"children":200},{},[201],{"type":19,"tag":89,"props":202,"children":204},{"alt":7,"src":203},"https://pic2.zhimg.com/80/v2-e29aaa64c71e89cc5e50c310a6834255_720w.jpg",[],{"type":19,"tag":144,"props":206,"children":207},{},[208],{"type":19,"tag":148,"props":209,"children":210},{},[211],{"type":29,"value":212},"Class-balanced focal loss (CB_Focal)",{"type":19,"tag":33,"props":214,"children":215},{},[216],{"type":19,"tag":89,"props":217,"children":219},{"alt":7,"src":218},"https://pic3.zhimg.com/80/v2-3aa61fe41951f30fee3e24bc0e05a8ca_720w.jpg",[],{"type":19,"tag":133,"props":221,"children":223},{"id":222},"重采样法",[224],{"type":29,"value":222},{"type":19,"tag":33,"props":226,"children":227},{},[228],{"type":29,"value":229},"重采样法试图对数据重新抽样，以获得均匀分布的数据集。",{"type":19,"tag":144,"props":231,"children":232},{},[233],{"type":19,"tag":148,"props":234,"children":235},{},[236],{"type":29,"value":237},"Random over-sampling",{"type":19,"tag":33,"props":239,"children":240},{},[241],{"type":29,"value":242},"复制从少数类中随机取样的训练图像。这种方法在大部分场景下都是有效的，但可能会导致过拟合。",{"type":19,"tag":144,"props":244,"children":245},{},[246],{"type":19,"tag":148,"props":247,"children":248},{},[249],{"type":29,"value":250},"Random under-sampling",{"type":19,"tag":33,"props":252,"children":253},{},[254],{"type":29,"value":255},"随机删除头部类的训练图像直到所有类变得平衡。在某些情况下，欠采样比过采样更有效。",{"type":19,"tag":144,"props":257,"children":258},{},[259],{"type":19,"tag":148,"props":260,"children":261},{},[262],{"type":29,"value":263},"Class-balanced sampling",{"type":19,"tag":33,"props":265,"children":266},{},[267],{"type":29,"value":268},"每个类被选中的概率为",{"type":19,"tag":33,"props":270,"children":271},{},[272],{"type":19,"tag":89,"props":273,"children":275},{"alt":7,"src":274},"https://pic2.zhimg.com/80/v2-e12271042d5883c01790768256a87859_720w.png",[],{"type":19,"tag":33,"props":277,"children":278},{},[279],{"type":29,"value":280},"，其中q=0，即首先对每个类别均匀抽样，然后从所选类别中随机抽取一个样本。",{"type":19,"tag":144,"props":282,"children":283},{},[284],{"type":19,"tag":148,"props":285,"children":286},{},[287],{"type":29,"value":288},"Square-root sampling",{"type":19,"tag":33,"props":290,"children":291},{},[292],{"type":29,"value":293},"将上面公式中q设为0.5，构造出一个介于原分布与平衡分布之间的采样集。",{"type":19,"tag":144,"props":295,"children":296},{},[297],{"type":19,"tag":148,"props":298,"children":299},{},[300],{"type":29,"value":301},"Progressively-balanced sampling",{"type":19,"tag":33,"props":303,"children":304},{},[305],{"type":29,"value":306},"逐步将类的抽样概率从原来的不平衡分布转为平衡分布，类j的采样概率如下：",{"type":19,"tag":33,"props":308,"children":309},{},[310],{"type":19,"tag":89,"props":311,"children":313},{"alt":7,"src":312},"https://pic3.zhimg.com/80/v2-4418397c4b5ead3645c2b3d4ef8fba6a_720w.jpg",[],{"type":19,"tag":33,"props":315,"children":316},{},[317],{"type":29,"value":318},"其中t为当前epoch数，T为总epoch数。",{"type":19,"tag":133,"props":320,"children":322},{"id":321},"混淆训练",[323],{"type":29,"value":321},{"type":19,"tag":33,"props":325,"children":326},{},[327],{"type":29,"value":328},"混淆训练可以看作是一种数据增强技巧，其目的是正则化卷积网络。我们发现混淆训练在长尾识别中具有良好效果，特别是与重采样相结合时。",{"type":19,"tag":144,"props":330,"children":331},{},[332],{"type":19,"tag":148,"props":333,"children":334},{},[335],{"type":29,"value":336},"Input mixup",{"type":19,"tag":33,"props":338,"children":339},{},[340],{"type":29,"value":341},"将两张原始输入图像在像素层面线性相加，合并后图片标签也由原标签线性相加。",{"type":19,"tag":144,"props":343,"children":344},{},[345],{"type":19,"tag":148,"props":346,"children":347},{},[348],{"type":29,"value":349},"Manifold mixup (MM)",{"type":19,"tag":33,"props":351,"children":352},{},[353],{"type":29,"value":354},"在网络某些层输出的特征图上做混淆，鼓励神经网络对隐藏表示的预测更加保守。",{"type":19,"tag":144,"props":356,"children":357},{},[358],{"type":19,"tag":148,"props":359,"children":360},{},[361],{"type":29,"value":362},"Fine-tuning after mixup training",{"type":19,"tag":33,"props":364,"children":365},{},[366],{"type":29,"value":367},"首先使用输入混合训练，然后对模型进行几个epoch的微调。",{"type":19,"tag":133,"props":369,"children":371},{"id":370},"两阶段训练",[372],{"type":29,"value":370},{"type":19,"tag":33,"props":374,"children":375},{},[376],{"type":29,"value":377},"将训练过程分为不平衡训练和平衡微调两个阶段，本节重点探讨平衡微调的不同方法，并提出了基于CAM的采样方法。",{"type":19,"tag":144,"props":379,"children":380},{},[381],{"type":19,"tag":148,"props":382,"children":383},{},[384],{"type":29,"value":385},"deferred re-balancing by resampling (DRS)",{"type":19,"tag":33,"props":387,"children":388},{},[389],{"type":29,"value":390},"首先使用普通训练方法，然后在二阶段使用平衡采样进行微调",{"type":19,"tag":144,"props":392,"children":393},{},[394],{"type":19,"tag":148,"props":395,"children":396},{},[397],{"type":29,"value":398},"deferred re-balancing by re-weighting (DRW)",{"type":19,"tag":33,"props":400,"children":401},{},[402],{"type":29,"value":403},"在二阶段使用重加权方法",{"type":19,"tag":144,"props":405,"children":406},{},[407],{"type":19,"tag":148,"props":408,"children":409},{},[410],{"type":29,"value":411},"CAM-Based Sampling (CAM-BS)",{"type":19,"tag":33,"props":413,"children":414},{},[415],{"type":29,"value":416},"DRS中使用的重采样方法仅从原始数据集中复制或删除随机选择的样本以生成平衡子集，所以在平衡微调过程中改进有限。为了生成判别信息，受类激活映射(class activation maps, CAM)的启发，我们提出了基于CAM的采样方法。如下图所示，我们首先使用重采样来获得均衡的采样图像。对于每一幅采样的图像，我们使用第一训练阶段训练的参数化模型，根据其标签和对应的全连通层权值来生成CAM。根据CAM的平均值将前景和背景分离，前景包含大于平均值的像素，背景包含其余像素。最后，我们在保持背景不变的情况下对前景应用转换。变换(由华为MindSpore实现)包括水平翻转、平移、旋转和缩放，我们对每个图像随机选择一个变换。",{"type":19,"tag":33,"props":418,"children":419},{},[420],{"type":19,"tag":89,"props":421,"children":423},{"alt":7,"src":422},"https://pic2.zhimg.com/80/v2-03617d8eda43b6154fcc446501a6be0d_720w.jpg",[],{"type":19,"tag":52,"props":425,"children":427},{"id":426},"_6实验结果",[428],{"type":29,"value":429},"6.实验结果",{"type":19,"tag":33,"props":431,"children":432},{},[433,438],{"type":19,"tag":434,"props":435,"children":436},"strong",{},[437],{"type":29,"value":135},{"type":29,"value":439},"：5种方法的错误率如下表所示。结果表明单独使用重加权策略是不合适的，特别是当类的数量增加时。",{"type":19,"tag":33,"props":441,"children":442},{},[443],{"type":19,"tag":89,"props":444,"children":446},{"alt":7,"src":445},"https://pic1.zhimg.com/80/v2-bc48c8b20df6db0a30975b4fe4769224_720w.jpg",[],{"type":19,"tag":33,"props":448,"children":449},{},[450,454],{"type":19,"tag":434,"props":451,"children":452},{},[453],{"type":29,"value":222},{"type":29,"value":455},"：不同重采样方法的错误率如下图。可以看到，单独使用重采样只能获得轻微改进。",{"type":19,"tag":33,"props":457,"children":458},{},[459],{"type":19,"tag":89,"props":460,"children":462},{"alt":7,"src":461},"https://pic4.zhimg.com/80/v2-9442c70513633e097a60df0e18e92b6b_720w.jpg",[],{"type":19,"tag":33,"props":464,"children":465},{},[466,470],{"type":19,"tag":434,"props":467,"children":468},{},[469],{"type":29,"value":321},{"type":29,"value":471},"：结果如下表，“ft.“表示混淆训练后使用二阶段平衡微调。可以看到，输入混淆和MM相比是有竞争力的，MM中不同层的混淆对结果影响有限。输入混淆对于对于后续的微调更有帮助。",{"type":19,"tag":33,"props":473,"children":474},{},[475,479],{"type":19,"tag":89,"props":476,"children":478},{"alt":7,"src":477},"https://pic1.zhimg.com/80/v2-dd79afd14920a1db33aedab70a9a73e0_720w.jpg",[],{"type":19,"tag":89,"props":480,"children":482},{"alt":7,"src":481},"https://pic2.zhimg.com/80/v2-903584723d59cddf4eb4b684968fc9d5_720w.jpg",[],{"type":19,"tag":33,"props":484,"children":485},{},[486,490],{"type":19,"tag":434,"props":487,"children":488},{},[489],{"type":29,"value":370},{"type":29,"value":491},"：下表为DRS中不同重采样方法的Top-1错误率。其中，基于cam的平衡采样得到了最好的结果。",{"type":19,"tag":33,"props":493,"children":494},{},[495],{"type":19,"tag":89,"props":496,"children":498},{"alt":7,"src":497},"https://pic3.zhimg.com/80/v2-13e13051d743fe7b42839ca9aaa94db2_720w.jpg",[],{"type":19,"tag":33,"props":500,"children":501},{},[502],{"type":29,"value":503},"下表为DRW中不同重权方法的Top-1错误率。CS_CE在DRW训练计划中取得了最好的成绩。",{"type":19,"tag":33,"props":505,"children":506},{},[507],{"type":19,"tag":89,"props":508,"children":510},{"alt":7,"src":509},"https://pic1.zhimg.com/80/v2-7d9bd42846d5e7e6babbcf0be5ecfebc_720w.jpg",[],{"type":19,"tag":33,"props":512,"children":513},{},[514],{"type":29,"value":515},"下表为将混淆训练与其他最佳技巧相结合的TOP-1错误率。我们很容易发现输入混合比MM获得更大的增益。",{"type":19,"tag":33,"props":517,"children":518},{},[519],{"type":19,"tag":89,"props":520,"children":522},{"alt":7,"src":521},"https://pic4.zhimg.com/80/v2-1ad996319baaca14c7ccd20830d6f87f_720w.jpg",[],{"type":19,"tag":33,"props":524,"children":525},{},[526],{"type":29,"value":527},"下表为组合各个最优训练策略后的TOP-1错误率，随着训练技巧的增加，性能稳步提升，这证明了我们的方法在小型和大型真实世界的数据集上的有效性。",{"type":19,"tag":52,"props":529,"children":531},{"id":530},"_7mindspore代码实现",[532],{"type":29,"value":533},"7.MindSpore代码实现",{"type":19,"tag":33,"props":535,"children":536},{},[537],{"type":29,"value":538},"def construct(self, logit, label, **kwargs):",{"type":19,"tag":33,"props":540,"children":541},{},[542],{"type":29,"value":543},"\"\"\"",{"type":19,"tag":33,"props":545,"children":546},{},[547],{"type":29,"value":548},"Args:",{"type":19,"tag":33,"props":550,"children":551},{},[552],{"type":29,"value":553},"inputs: prediction matrix (before softmax) with shape (batch_size, num_classes)",{"type":19,"tag":33,"props":555,"children":556},{},[557],{"type":29,"value":558},"label: ground truth labels with shape (batch_size)",{"type":19,"tag":33,"props":560,"children":561},{},[562],{"type":29,"value":543},{"type":19,"tag":33,"props":564,"children":565},{},[566],{"type":29,"value":567},"logit_max = self.max(logit, -1)",{"type":19,"tag":33,"props":569,"children":570},{},[571],{"type":29,"value":572},"exp = self.exp(self.sub(logit, logit_max))",{"type":19,"tag":33,"props":574,"children":575},{},[576],{"type":29,"value":577},"exp_sum = self.sum(exp, -1)",{"type":19,"tag":33,"props":579,"children":580},{},[581],{"type":29,"value":582},"softmax_result = self.div(exp, exp_sum)",{"type":19,"tag":33,"props":584,"children":585},{},[586],{"type":29,"value":587},"label = self.onehot(label, ops.shape(logit)[1], self.on_value, self.off_value)",{"type":19,"tag":33,"props":589,"children":590},{},[591],{"type":29,"value":592},"softmax_result_log = self.log(softmax_result)",{"type":19,"tag":33,"props":594,"children":595},{},[596],{"type":29,"value":597},"loss = self.sum_cross_entropy((self.mul(softmax_result_log, label)), -1)",{"type":19,"tag":33,"props":599,"children":600},{},[601],{"type":29,"value":602},"loss = self.mul2(ops.scalar_to_array(-1.0), loss)",{"type":19,"tag":33,"props":604,"children":605},{},[606],{"type":29,"value":607},"if self.weight_list is not None:",{"type":19,"tag":33,"props":609,"children":610},{},[611],{"type":29,"value":612},"weight = self.mul3(self.squeeze(self.weight_list), label)",{"type":19,"tag":33,"props":614,"children":615},{},[616],{"type":29,"value":617},"weight = self.sum2(weight, -1)",{"type":19,"tag":33,"props":619,"children":620},{},[621],{"type":29,"value":622},"loss = self.mul3(loss, weight)",{"type":19,"tag":33,"props":624,"children":625},{},[626],{"type":29,"value":627},"loss = self.mean(loss, -1)",{"type":19,"tag":33,"props":629,"children":630},{},[631],{"type":29,"value":632},"return loss",{"type":19,"tag":33,"props":634,"children":635},{},[636],{"type":29,"value":637},"通过细化算子粒度增强算法在多卡情况下的性能。",{"type":19,"tag":52,"props":639,"children":641},{"id":640},"_8总结与展望",[642],{"type":29,"value":643},"8.总结与展望",{"type":19,"tag":33,"props":645,"children":646},{},[647],{"type":29,"value":648},"本文系统地探索了现有的简单有效的长尾识别方法，为长尾视觉识别提供了科学的训练指导。此外，我们发现现有的简单抽样方法缺乏判别信息。基于此，我们提出了一种基于类激活映射的数据增强方法，并将其与现有的重采样方法相结合。通过大量的实验，我们得到了最优的训练trick组合，在长尾基准测试上取得了最好的结果。我们还将源代码作为一个科学实用的工具箱发布，为长尾视觉识别的未来研究提供了有益的参考。在未来，我们将尝试探索其他具有挑战性的长尾任务的trick包，例如检测和分割。",{"type":19,"tag":33,"props":650,"children":651},{},[652],{"type":29,"value":653},"有任何问题可通过官方QQ群: 871543426获得第一时间解答",{"type":19,"tag":33,"props":655,"children":656},{},[657,659],{"type":29,"value":658},"MindSpore官网：",{"type":19,"tag":44,"props":660,"children":663},{"href":661,"rel":662},"https://link.zhihu.com/?target=https%3A//www.mindspore.cn/",[48],[664],{"type":29,"value":665},"https://www.mindspore.cn/",{"type":19,"tag":33,"props":667,"children":668},{},[669,671],{"type":29,"value":670},"MindSpore论坛：",{"type":19,"tag":44,"props":672,"children":675},{"href":673,"rel":674},"https://link.zhihu.com/?target=https%3A//bbs.huaweicloud.com/forum/forum-1076-1.html",[48],[676],{"type":29,"value":677},"https://bbs.huaweicloud.com/forum/forum-1076-1.html",{"type":19,"tag":33,"props":679,"children":680},{},[681],{"type":29,"value":682},"代码仓地址：",{"type":19,"tag":33,"props":684,"children":685},{},[686,688],{"type":29,"value":687},"Gitee-",{"type":19,"tag":44,"props":689,"children":692},{"href":690,"rel":691},"https://link.zhihu.com/?target=https%3A//gitee.com/mindspore/mindspore.git",[48],[693],{"type":29,"value":694},"https://gitee.com/mindspore/",{"type":19,"tag":33,"props":696,"children":697},{},[698,700],{"type":29,"value":699},"GitHub-",{"type":19,"tag":44,"props":701,"children":704},{"href":702,"rel":703},"https://link.zhihu.com/?target=https%3A//github.com/mindspore-ai/mindspore.git",[48],[705],{"type":29,"value":706},"https://github.com/mindspore-ai",{"title":7,"searchDepth":708,"depth":708,"links":709},4,[710,712,713,714,715,722,723,724],{"id":54,"depth":711,"text":57},2,{"id":65,"depth":711,"text":68},{"id":76,"depth":711,"text":79},{"id":95,"depth":711,"text":98},{"id":128,"depth":711,"text":131,"children":716},[717,719,720,721],{"id":135,"depth":718,"text":135},3,{"id":222,"depth":718,"text":222},{"id":321,"depth":718,"text":321},{"id":370,"depth":718,"text":370},{"id":426,"depth":711,"text":429},{"id":530,"depth":711,"text":533},{"id":640,"depth":711,"text":643},"markdown","content:technology-blogs:zh:1556.md","content","technology-blogs/zh/1556.md","technology-blogs/zh/1556","md",1776506113342]