Member-only story

Day 11: 2020.04.22
Paper: Deep Residual Learning for Image Recognition
Category: Model/CNN/Deep Learning/Image Recognition

Purpose: Tackle the problem of vanishing gradient in training

When deeper networks are able to start converging, a degradation problem has been exposed: with the network depth increasing, accuracy gets saturated and then degrades rapidly. Unexpectedly, such degradation is not caused by overfitting, and adding more layers to a suitably deep model leads to higher training error.

Solution: Residual Learning

Add a shortcut connection

--

--

Chun-kit Ho
Chun-kit Ho

Written by Chun-kit Ho

cloud architect@ey | full-stack software engineer | social innovation | certified professional solutions architect in aws & gcp

No responses yet