What is `weight_decay` meta parameter in Caffe?

前端 未结 2 1366
-上瘾入骨i
-上瘾入骨i 2020-11-28 08:12

Looking at an example \'solver.prototxt\', posted on BVLC/caffe git, there is a training meta parameter

weight_decay: 0.04

What does this m

2条回答
  •  醉梦人生
    2020-11-28 09:00

    Weight decay is a regularization term that penalizes big weights. When the weight decay coefficient is big the penalty for big weights is also big, when it is small weights can freely grow.

    Look at this answer (not specific to caffe) for a better explanation: Difference between neural net "weight decay" and "learning rate".

提交回复
热议问题