<rss xmlns:atom="http://www.w3.org/2005/Atom" version="2.0"><channel><title>分布式训练 - 标签 - mywebsite</title><link>https://steven-yl.github.io/mywebsite/tags/%E5%88%86%E5%B8%83%E5%BC%8F%E8%AE%AD%E7%BB%83/</link><description>分布式训练 - 标签 - mywebsite</description><generator>Hugo -- gohugo.io</generator><language>zh-CN</language><managingEditor>steven@gmail.com (Steven)</managingEditor><webMaster>steven@gmail.com (Steven)</webMaster><copyright>This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License.</copyright><lastBuildDate>Thu, 12 Mar 2026 00:00:00 +0800</lastBuildDate><atom:link href="https://steven-yl.github.io/mywebsite/tags/%E5%88%86%E5%B8%83%E5%BC%8F%E8%AE%AD%E7%BB%83/" rel="self" type="application/rss+xml"/><item><title>PyTorch 分布式训练与操作工具技术文档</title><link>https://steven-yl.github.io/mywebsite/distributed_training_guide/</link><pubDate>Thu, 12 Mar 2026 00:00:00 +0800</pubDate><author><name>Steven</name><uri>https://github.com/steven-yl</uri></author><guid>https://steven-yl.github.io/mywebsite/distributed_training_guide/</guid><description>从进程组初始化、DDP 封装、数据分片、集体通信到 Lightning 封装，全面讲解如何在单机多卡与多机多卡场景下正确使用 PyTorch 分布式训练。</description></item></channel></rss>