其他分享
首页 > 其他分享> > Activation Functions

Activation Functions

作者:互联网

Sigmoid

Sigmoids saturate and kill gradients.

Sigmoid outputs are not zero-centered.

Exponential function is a little computational expensive.

 

Tanh

Kill gradients when saturated.

It's zero-centered! : )

 

ReLU

Does not saturate. ( in positive region)

Very computational efficient.

Converges much faster than sigmoid/tanh in practice. (6 times)

Seems more biologically plausible than sigmoid.

BUT!

Not zero-centered.

No gradient when x<0.

 

Take care of learning rate when using ReLU.

 

Leakly ReLU

Does not saturate.

Very computational efficient.

Converges much faster than sigmoid/tanh in practice. (6 times)

will not "die"

 

Parametric ReLU

 

Exponential Linear Unit

标签:Functions,computational,sigmoid,centered,Activation,when,ReLU,zero
来源: https://www.cnblogs.com/hizhaolei/p/10623472.html