Motivating CNN Case Studies
- To develop a deeper understanding of convolutional neural networks, we should evaluate some popular case studies
-
The following classical networks are popular case studies:
- LeNet-5
- AlexNet
- VGG-16
- More modern convolutional networks are ResNet and Inception
- We've already talked about LeNet-5, so now let's talk about AlexNet and VGG-16
Illustrating AlexNet
Footnotes about AlexNet
- Similar architecture to LeNet-5
-
Bigger network compared to LeNet-5
- However, AlexNet is much bigger
- LeNet-5 has about thousand parameters
- AlexNet has about million parameters
- Meaning, AlexNet is more accurate, but slower to train
-
Uses relu activation functions
- Convolutional layers use relu functions in AlexNet
- Fully-connected layers use relu functions too
-
Trained on multiple GPUs
- When AlexNet was initially created, GPUs were slower
- As a result, AlexNet had a complicated way of training on multiple GPUs
- Specifically, layers were split across separate GPUs
- Then, results were joined back together afterwards
Illustrating VGG-16
Footnotes about VGG-16
-
Simpler architecture compared to AlexNet
- VGG-16 only uses convolutional layers with a filter and a stride
- VGG-16 only uses max-pooling layers with a filter and a stride
- Consequently, VGG-16 uses fewer hyperparameters compared to AlexNet
-
Bigger network compared to AlexNet
- VGG-16 has about million parameters
- Meaning, VGG-16 is more accurate, but slower to train
tldr
-
The AlexNet has the following properties:
- Similar architecture to LeNet-5
- Bigger network compared to LeNet-5
- Uses relu activation functions
- Trained on multiple GPUs
-
The VGG-16 has the following properties:
- Bigger network compared to AlexNet
- Simpler architecture compared to AlexNet
- Uses fewer hyperparameters compared to AlexNet
References
Previous
Next