[{"data":1,"prerenderedAt":132},["ShallowReactive",2],{"content-query-cTnDjrgeYD":3},{"_path":4,"_dir":5,"_draft":6,"_partial":6,"_locale":7,"title":8,"description":8,"date":9,"start_date":10,"end_date":11,"cover":12,"type":13,"series":14,"state":15,"address":16,"working":17,"body":18,"_type":126,"_id":127,"_source":128,"_file":129,"_stem":130,"_extension":131},"/activities/zh/2025-07-07","zh",false,"","直播预告 | 带你解析模型性能关键因素，并基于昇思进行性能调优","2025-07-07","2025-07-07 19:00","2025-07-07 20:00","https://obs-mindspore-file.obs.cn-north-4.myhuaweicloud.com/file/2025/11/14/00f05b9765a54e05a080b6953b135175.png","activities",4,2,null,"线上直播",{"type":19,"children":20,"toc":122},"root",[21,29,38,43,52,57,65,80,85,102],{"type":22,"tag":23,"props":24,"children":26},"element","h3",{"id":25},"直播预告-带你解析模型性能关键因素并基于昇思进行性能调优",[27],{"type":28,"value":8},"text",{"type":22,"tag":30,"props":31,"children":32},"p",{},[33],{"type":22,"tag":34,"props":35,"children":37},"img",{"alt":7,"src":36},"https://obs-mindspore-file.obs.cn-north-4.myhuaweicloud.com/file/2025/07/04/25b7590458fd43ea8f35dd392fc7e9a3.png",[],{"type":22,"tag":30,"props":39,"children":40},{},[41],{"type":28,"value":42},"MindSpore Transformers SIG目标是构建一个大模型预训练、微调、评测、推理、部署的全流程开发套件，提供业内主流的Transformer类大语言模型（Large Language Models, LLMs）和多模态理解模型（Multimodal Models, MMs）。期望帮助用户轻松地实现大模型全流程开发。",{"type":22,"tag":30,"props":44,"children":45},{},[46],{"type":22,"tag":47,"props":48,"children":49},"strong",{},[50],{"type":28,"value":51},"7月3日下午16：00，欢迎对MindSpore Transformers方向感兴趣的同学参加SIG月例会！",{"type":22,"tag":30,"props":53,"children":54},{},[55],{"type":28,"value":56},"大家如果有议题申报，可以直接在Etherpad链接中填写。",{"type":22,"tag":30,"props":58,"children":59},{},[60],{"type":22,"tag":47,"props":61,"children":62},{},[63],{"type":28,"value":64},"SIG例会：",{"type":22,"tag":30,"props":66,"children":67},{},[68,73,75],{"type":22,"tag":47,"props":69,"children":70},{},[71],{"type":28,"value":72},"例行月度例会：北京时间",{"type":28,"value":74},"周四",{"type":22,"tag":47,"props":76,"children":77},{},[78],{"type":28,"value":79},"16:00例行展开SIG会议。",{"type":22,"tag":30,"props":81,"children":82},{},[83],{"type":28,"value":84},"临时补充专项讨论会议：基于前一次会议讨论，按需临时增开专项议题会议讨论",{"type":22,"tag":30,"props":86,"children":87},{},[88,90,94],{"type":28,"value":89},"会议链接：",{"type":22,"tag":91,"props":92,"children":93},"br",{},[],{"type":22,"tag":95,"props":96,"children":100},"a",{"href":97,"rel":98},"https://meeting.tencent.com/dm/olkrpDKfadmJ",[99],"nofollow",[101],{"type":28,"value":97},{"type":22,"tag":30,"props":103,"children":104},{},[105,107,113,115,120],{"type":28,"value":106},"Etherpad链接：",{"type":22,"tag":95,"props":108,"children":111},{"href":109,"rel":110},"https://etherpad.mindspore.cn/p/sig-MindSpore-Transformers-meetings",[99],[112],{"type":28,"value":109},{"type":28,"value":114},"（",{"type":22,"tag":47,"props":116,"children":117},{},[118],{"type":28,"value":119},"大家有议题申报的也可以在Etherpad链接中填写",{"type":28,"value":121},"）",{"title":7,"searchDepth":14,"depth":14,"links":123},[124],{"id":25,"depth":125,"text":8},3,"markdown","content:activities:zh:2025-07-07.md","content","activities/zh/2025-07-07.md","activities/zh/2025-07-07","md",1776506012984]