• 卷积神经网络卷积层后一定要跟激活函数吗?


    The reason why neural network is more powerful than linear function is because neural network use the non­linear function to map the dataset which is difficult to separate to separable space. So we can say that every neural network(including CNN)'s neuron uses activation function.We also know that the activation function has a lot of forms,such as logistic function,tanh function, rectifier function,and softplus function.The reason why you can see a ReLU(activation function) layer behind the convolution layer in Caffe or other frames is because of convenience of the network's defination,in other words,we extract the activation function on purpose so as to we can adjust all activation functions's defination in this layer to use the proper one.

    1).Why we use activation function? No activation function,the network is just same with Perception(Linear combination).

    2).Why use Relu activation function?    

                       1>.If use sigmoid or tanh function,cause large calculation(Exponential operation,Derivative solution).

                       2>.If use sigmoid or tanh function,the disadvantage is gradient disappearance,the derivative will trend to be 0.

                       3>.Some output will be 0,the network becomes sparse,it is good to reduce to overfitting.

  • 相关阅读:
    OCP-1Z0-051-V9.02-80题
    OCP-1Z0-051-V9.02-124题
    Flex中的HDividedBox和VDividedBox的比较
    Flex中AdvancedDataGrid的用法
    IDA,很好很强大
    AndroidManifest.xml文件中加入文件控制控制权限
    OCP-1Z0-051-V9.02-6题
    OCP-1Z0-051-V9.02-5题
    OCP-1Z0-051-V9.02-4题
    Android Eclipse JNI 调用 .so文件加载问题
  • 原文地址:https://www.cnblogs.com/ymjyqsx/p/6386876.html
Copyright © 2020-2023  润新知