[Free] Computer Vision With Resnet

ResNet: How One Paper Changed Deep LearningForever – Free Course

What you’ll learn

  • To understand the history and evolution of computer vision, from its early beginnings to state-of-the-art.
  • To become familiar with the SuperGradients training library and how deep learning practitioners can use it to shorten the model development lifecycle.
  • To gain practical skills for developing and training neural networks for image classification tasks.
  • Throughout the course, students will study various topics, including the fundamental concepts and techniques of computer vision, the design and training of neur

Requirements

  • The target learners are students with a strong foundation in machine learning and a basic understanding of deep learning. These students need to learn about the history and current state of computer vision, as well as gain practical skills for developing and training deep neural networks for image classification tasks. It has a secondary audience of professionals in machine learning and computer vision who are looking to stay up to date on the latest developments and techniques in the field. These professionals will learn about the SuperGradients training library, which could improve their model development process.

Description

In December of 2015, a paper was published that rocked the deep learning world.

This paper is widely regarded as one of the most influential papers in modern deep learning and has been cited over 110,000 times.

The name of this paper?

Deep Residual Learning for Image Recognition (aka, the ResNet paper).

The prevailing wisdom of the time suggested adding more layers to neural networks would lead to better results.

But researchers observed that the accuracy of deep networks would increase up to a saturation point before levelling off.

In addition to that, an unusual phenomenon was observed: Adding layers to an already deep network, the training error would actually increase.

This was primarily due to two problems:

1) Vanishing/exploding gradients

2) The degradation problem

The vanishing/exploding gradients problem is a by-product of the chain rule.

The chain rule multiplies error gradients for weights in the network.

Multiplying lots of values that are less than one will result in smaller and smaller values.

As those error gradients approach the earlier layers of a network, their value will tend to zero.

This results in smaller and smaller updates to earlier layers (not much learning happening).

The inverse problem is the exploding gradient which happens when large error gradients accumulate during training and result in massive updates to model weights in the earlier layers.

The degradation problem is unexpected because it’s not caused by overfitting.

Researchers were finding that as networks got deeper, the training loss would decrease but then shoot back up as more layers were added to the networks.

Which is counterintuitive

Because youd expect your training error to decrease, converge, and plateau out as the number of layers in your network increases.

Both of these issues threatened to halt the progress of deep neural networks until this paper came out…

The ResNet paper introduced a novel solution to these two pesky problems that plagued the architects of deep neural networks:

The Skip Connection.

Skip connections, which are housed in residual blocks, allow you to take the activation value from an earlier layer and pass it to a deeper layer in a network.

Skip connections enable deep networks to learn the identity function.

Learning the identity function allows a deeper layer to perform as well as an earlier layer, or at the very least it wont perform any worse

The result is a smoother gradient flow, ensuring important features are preserved in the training process.

The invention of the skip connection has given us the ability to build deeper and deeper networks while avoiding the problem of vanishing/exploding gradients and degradation.

Wanna learn more about ResNet? Check out this short course that I’ve prepared for you using the SuperGradients training library!

Author(s): Harpreet Sahota
Coupon Scorpion
Coupon Scorpion

The Coupon Scorpion team has over ten years of experience finding free and 100%-off Udemy Coupons. We add over 200 coupons daily and verify them constantly to ensure that we only offer fully working coupon codes. We are experts in finding new offers as soon as they become available. They're usually only offered for a limited usage period, so you must act quickly.

Coupon Scorpion
Logo