[{"data":1,"prerenderedAt":122},["ShallowReactive",2],{"content-query-40uYeo58bg":3},{"_path":4,"_dir":5,"_draft":6,"_partial":6,"_locale":7,"title":8,"description":8,"date":9,"cover":10,"type":11,"state":12,"body":13,"_type":116,"_id":117,"_source":118,"_file":119,"_stem":120,"_extension":121},"/activities/zh/3784","zh",false,"","直播预告 | 带你解析模型性能关键因素，并基于昇思进行性能调优","2025-07-02","https://obs-mindspore-file.obs.cn-north-4.myhuaweicloud.com/file/2025/07/04/67d319eadc324962a3d0406b35917475.png","activities",1,{"type":14,"children":15,"toc":111},"root",[16,24,30,39,44,58,66,81,86,99],{"type":17,"tag":18,"props":19,"children":21},"element","h1",{"id":20},"直播预告-带你解析模型性能关键因素并基于昇思进行性能调优",[22],{"type":23,"value":8},"text",{"type":17,"tag":25,"props":26,"children":28},"h3",{"id":27},"直播预告-带你解析模型性能关键因素并基于昇思进行性能调优-1",[29],{"type":23,"value":8},{"type":17,"tag":31,"props":32,"children":33},"p",{},[34],{"type":17,"tag":35,"props":36,"children":38},"img",{"alt":7,"src":37},"https://obs-mindspore-file.obs.cn-north-4.myhuaweicloud.com/file/2025/07/04/25b7590458fd43ea8f35dd392fc7e9a3.png",[],{"type":17,"tag":31,"props":40,"children":41},{},[42],{"type":23,"value":43},"MindSpore Transformers SIG目标是构建一个大模型预训练、微调、评测、推理、部署的全流程开发套件，提供业内主流的Transformer类大语言模型（Large Language Models, LLMs）和多模态理解模型（Multimodal Models, MMs）。期望帮助用户轻松地实现大模型全流程开发。",{"type":17,"tag":31,"props":45,"children":46},{},[47],{"type":17,"tag":48,"props":49,"children":50},"strong",{},[51,53],{"type":23,"value":52},"7月3日下午16：00，欢迎对MindSpore Transformers方向感兴趣的同学参加SIG月例会！",{"type":17,"tag":48,"props":54,"children":55},{},[56],{"type":23,"value":57},"大家如果有议题申报，可以直接在Etherpad链接中填写。",{"type":17,"tag":31,"props":59,"children":60},{},[61],{"type":17,"tag":48,"props":62,"children":63},{},[64],{"type":23,"value":65},"SIG例会：",{"type":17,"tag":31,"props":67,"children":68},{},[69,74,76],{"type":17,"tag":48,"props":70,"children":71},{},[72],{"type":23,"value":73},"例行月度例会：北京时间",{"type":23,"value":75},"周四",{"type":17,"tag":48,"props":77,"children":78},{},[79],{"type":23,"value":80},"16:00例行展开SIG会议。",{"type":17,"tag":31,"props":82,"children":83},{},[84],{"type":23,"value":85},"临时补充专项讨论会议：基于前一次会议讨论，按需临时增开专项议题会议讨论",{"type":17,"tag":31,"props":87,"children":88},{},[89,91],{"type":23,"value":90},"会议链接： ",{"type":17,"tag":92,"props":93,"children":97},"a",{"href":94,"rel":95},"https://meeting.tencent.com/dm/olkrpDKfadmJ",[96],"nofollow",[98],{"type":23,"value":94},{"type":17,"tag":31,"props":100,"children":101},{},[102,104],{"type":23,"value":103},"Etherpad链接：",{"type":17,"tag":92,"props":105,"children":108},{"href":106,"rel":107},"https://etherpad.mindspore.cn/p/sig-MindSpore-Transformers-meetings%EF%BC%88**%E5%A4%A7%E5%AE%B6%E6%9C%89%E8%AE%AE%E9%A2%98%E7%94%B3%E6%8A%A5%E7%9A%84%E4%B9%9F%E5%8F%AF%E4%BB%A5%E5%9C%A8Etherpad%E9%93%BE%E6%8E%A5%E4%B8%AD%E5%A1%AB%E5%86%99**%EF%BC%89",[96],[109],{"type":23,"value":110},"https://etherpad.mindspore.cn/p/sig-MindSpore-Transformers-meetings（**大家有议题申报的也可以在Etherpad链接中填写**）",{"title":7,"searchDepth":112,"depth":112,"links":113},4,[114],{"id":27,"depth":115,"text":8},3,"markdown","content:activities:zh:3784.md","content","activities/zh/3784.md","activities/zh/3784","md",1776506032043]