V2EX batch normalization

Batch Normalization

释义 Definition

批量归一化(Batch Normalization,常简称 BN):一种用于深度神经网络训练的技术,在每个小批量(mini-batch)上对某层的激活值进行标准化,并通过可学习的缩放与平移参数恢复表达能力,从而加速收敛、稳定训练、减轻梯度问题。常见于卷积网络与全连接网络中。(也可泛指“对批数据做归一化”的更宽泛用法,但最常用的是上述深度学习方法。)

发音 Pronunciation (IPA)

/bt nrmlzen/

词源 Etymology

batch 源自表示“一批/一组”的用法;normalization 来自 normal(“正常的/标准的”)+ -ize(动词化)+ -ation(名词化),意为“使之标准化/归一化”。组合起来即“对一批数据进行归一化”。在深度学习语境中,这个术语因 2015 年 Ioffe 与 Szegedy 提出的训练技巧而广泛流行。

例句 Examples

Batch normalization helps the network train faster.
批量归一化能帮助网络训练得更快。

When batch normalization is placed after the convolution and before the activation function, it often stabilizes training and allows a larger learning rate without divergenc.
当把批量归一化放在卷积层之后、激活函数之前时,它常能稳定训练,并在不发散的情况下允许使用更大的学习率。

相关词 Related Words

文献与作品 Notable Works

  • Ioffe, Sergey & Szegedy, Christian. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift(2015,原始提出论文)
  • Ian Goodfellow, Yoshua Bengio, Aaron Courville. Deep Learning(《深度学习》教材中讨论归一化与训练稳定性,含 BN 相关内容)
  • Aston Zhang, Zachary C. Lipton, Mu Li, Alexander J. Smola. Dive into Deep Learning(《动手学深度学习》含 BN 原理与实现示例)
  • Aurélien Géron. Hands-On Machine Learning with Scikit-Learn, Keras & TensorFlow(实战书籍中常以 BN 作为稳定训练的常用组件)
关于     帮助文档     自助推广系统     博客     API     FAQ     Solana     1559 人在线   最高记录 6679       Select Language
创意工作者们的社区
World is powered by solitude
VERSION: 3.9.8.5 63ms UTC 16:35 PVG 00:35 LAX 09:35 JFK 12:35
Do have faith in what you're doing.
ubao msn snddm index pchome yahoo rakuten mypaper meadowduck bidyahoo youbao zxmzxm asda bnvcg cvbfg dfscv mmhjk xxddc yybgb zznbn ccubao uaitu acv GXCV ET GDG YH FG BCVB FJFH CBRE CBC GDG ET54 WRWR RWER WREW WRWER RWER SDG EW SF DSFSF fbbs ubao fhd dfg ewr dg df ewwr ewwr et ruyut utut dfg fgd gdfgt etg dfgt dfgd ert4 gd fgg wr 235 wer3 we vsdf sdf gdf ert xcv sdf rwer hfd dfg cvb rwf afb dfh jgh bmn lgh rty gfds cxv xcv xcs vdas fdf fgd cv sdf tert sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf shasha9178 shasha9178 shasha9178 shasha9178 shasha9178 liflif2 liflif2 liflif2 liflif2 liflif2 liblib3 liblib3 liblib3 liblib3 liblib3 zhazha444 zhazha444 zhazha444 zhazha444 zhazha444 dende5 dende denden denden2 denden21 fenfen9 fenf619 fen619 fenfe9 fe619 sdf sdf sdf sdf sdf zhazh90 zhazh0 zhaa50 zha90 zh590 zho zhoz zhozh zhozho zhozho2 lislis lls95 lili95 lils5 liss9 sdf0ty987 sdft876 sdft9876 sdf09876 sd0t9876 sdf0ty98 sdf0976 sdf0ty986 sdf0ty96 sdf0t76 sdf0876 df0ty98 sf0t876 sd0ty76 sdy76 sdf76 sdf0t76 sdf0ty9 sdf0ty98 sdf0ty987 sdf0ty98 sdf6676 sdf876 sd876 sd876 sdf6 sdf6 sdf9876 sdf0t sdf06 sdf0ty9776 sdf0ty9776 sdf0ty76 sdf8876 sdf0t sd6 sdf06 s688876 sd688 sdf86