<rss xmlns:atom="http://www.w3.org/2005/Atom" version="2.0"><channel><title>优化器 - 标签 - mywebsite</title><link>https://steven-yl.github.io/mywebsite/tags/%E4%BC%98%E5%8C%96%E5%99%A8/</link><description>优化器 - 标签 - mywebsite</description><generator>Hugo -- gohugo.io</generator><language>zh-CN</language><managingEditor>steven@gmail.com (Steven)</managingEditor><webMaster>steven@gmail.com (Steven)</webMaster><copyright>This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License.</copyright><lastBuildDate>Tue, 17 Mar 2026 00:00:00 +0800</lastBuildDate><atom:link href="https://steven-yl.github.io/mywebsite/tags/%E4%BC%98%E5%8C%96%E5%99%A8/" rel="self" type="application/rss+xml"/><item><title>PyTorch 模型训练技术文档：求解器、参数配置与训练循环</title><link>https://steven-yl.github.io/mywebsite/training_solver_guide/</link><pubDate>Thu, 12 Mar 2026 00:00:00 +0800</pubDate><author><name>Steven</name><uri>https://github.com/steven-yl</uri></author><guid>https://steven-yl.github.io/mywebsite/training_solver_guide/</guid><description>从总览到各章节：Optimizer/SGD/Adam/AdamW 全解读、LRScheduler 族、param_groups、梯度累积与裁剪、损失选型及学习率与 batch 配置经验。</description></item><item><title>Pytorch Batch Size 与学习率缩放规则</title><link>https://steven-yl.github.io/mywebsite/batch_size_lr/</link><pubDate>Tue, 17 Mar 2026 00:00:00 +0800</pubDate><author><name>Steven</name><uri>https://github.com/steven-yl</uri></author><guid>https://steven-yl.github.io/mywebsite/batch_size_lr/</guid><description>详解分布式训练中 batch size 扩大时学习率的线性缩放、平方根缩放及线性+长 warmup 的推导依据与使用建议。</description></item></channel></rss>