The Residual Network (ResNet) Standard

Aus HSHL Mechatronik
Zur Navigation springen Zur Suche springen

ResNet, also called Residual Network, is a very important model in deep learning history. Before ResNet, training very deep neural networks was really hard because of the vanishing gradient problem. This means the signal used to update the weights becomes very small as it moves backward through many layers.

ResNet fixed this problem by adding skip connections, also known as shortcuts. These connections let the gradient flow around some layers instead of passing through all of them. Because of this, very deep networks can be trained, even with hundreds of layers, like ResNet-50, ResNet-101, and ResNet-152.[1]

Performance

ResNet-50 is seen as the normal industry "workhorse." It gives a good mix of accuracy (about 76% top-1 on ImageNet) and how much computing power it needs. It works well for most tasks and people trust it a lot.[2]

MATLAB Usage

It is easy to use in MATLAB with resnet50. Many people use it as the default option for transfer learning tests, because its behavior is well known and already understood.[3]

References

  1. MathWorks. Train Residual Network for Image Classification. [Online]. Available at: https://www.mathworks.com/help/deeplearning/ug/train-residual-network-for-image-classification.html
  2. DeepLabCut. What neural network should I use? (Trade-offs, speed performance, and considerations). GitHub Wiki. Available at: https://github.com/DeepLabCut/DeepLabCut/wiki/What-neural-network-should-I-use%3F-(Trade-offs,-speed-performance,-and-considerations)
  3. MathWorks. Pretrained Convolutional Neural Networks. MATLAB Documentation. Available at: https://www.mathworks.com/help/deeplearning/ug/pretrained-convolutional-neural-networks.html