MindMap Gallery Classic network analysis
The detailed features of some classic networks are introduced, and the knowledge of AlexNet, ZLNet, VGG, GoogleNet, and resnet is shared. Everyone is welcome to learn.
Edited at 2023-07-27 22:46:54This is a mind map about bacteria, and its main contents include: overview, morphology, types, structure, reproduction, distribution, application, and expansion. The summary is comprehensive and meticulous, suitable as review materials.
This is a mind map about plant asexual reproduction, and its main contents include: concept, spore reproduction, vegetative reproduction, tissue culture, and buds. The summary is comprehensive and meticulous, suitable as review materials.
This is a mind map about the reproductive development of animals, and its main contents include: insects, frogs, birds, sexual reproduction, and asexual reproduction. The summary is comprehensive and meticulous, suitable as review materials.
This is a mind map about bacteria, and its main contents include: overview, morphology, types, structure, reproduction, distribution, application, and expansion. The summary is comprehensive and meticulous, suitable as review materials.
This is a mind map about plant asexual reproduction, and its main contents include: concept, spore reproduction, vegetative reproduction, tissue culture, and buds. The summary is comprehensive and meticulous, suitable as review materials.
This is a mind map about the reproductive development of animals, and its main contents include: insects, frogs, birds, sexual reproduction, and asexual reproduction. The summary is comprehensive and meticulous, suitable as review materials.
Classic network analysis
AlexNet
structure
8 floors in total
5 convolutional layers
3 layers of fully connected layers
effect
Learn structural features from data that are meaningful for classification
Describe structural information in the input image
The description results are stored in 256 6x6 feature response maps
ZLNet
Same structure as AlexNet
Improve
Change the previous convolution kernel to smaller size
Increase the step size to successfully extract features
Increase the number of late convolution kernels
VGG
16-layer neural network
13 convolutional layers
3 fully connected layers
Improve
Use smaller 3x3 convolution kernels in series to obtain a larger receptive field
Deeper depth, stronger nonlinearity, and fewer network parameters
Remove the LRN layer of AlexNet
Summarize
Advantages of small convolution kernel
Multiple small-size convolution kernels connected in series can obtain the same receptive field as large-size convolution kernels, and require fewer training parameters.
Reasons for convolution to 512
There are many parameters to collect and express features as much as possible
If there are too many parameters, it will be easy to overfit and it is not suitable for training.
GoogleNet
22nd floor
Innovation
Propose the inception structure, which can retain more characteristic information of the input signal
Add a bottleneck layer and change the number of convolution channels
Remove the fully connected layer and use average pooling, resulting in only 5 million parameters, 12 times less than AlexNet.
An auxiliary classifier is introduced in the middle of the network to overcome the vanishing gradient problem during training.
Summarize
The difference between average pooling vectorization and direct expansion vectorization
The value of each position on the feature response map reflects the similarity between the structure of the corresponding position in the image and the semantic structure recorded by the convolution kernel.
Average pooling loses the spatial location information of semantic structures
Ignoring the position information of the semantic structure helps to improve the translation invariance of the features extracted by the convolutional layer.
resnet
Deepen the network layers
Deeper networks appear to be inferior to shallower networks
reason
During the training process, the positive and negative information of the network does not flow smoothly, and the network is not fully trained.
innovation
Proposed residual module
By stacking residual modules, neural networks of arbitrary depth can be constructed without "degradation".
proposed batch normalization
Fight gradient disappearance and reduce dependence on weight initialization during network training.
Propose an initialization method for ReLU activation function
Reasons for the good performance of residual network
can be regarded as an integrated model