chapter three
3 Wide Convolutional Neural Networks
This chapter covers
- The wide convolutional layer design pattern.
- Why researchers went wide vs deep.
- Refactoring micro architecture patterns to decrease computational complexity.
- Coding former state-of-the-art (SOTA) wide convolutional models with the procedural design pattern
Up to now in the book, we’ve focused on networks with deeper layers, block layers and shortcuts in residual networks for image related tasks (classification, object localization, image segmentation). Starting in 2014 with Inception v1 (GoogLeNet) (https://arxiv.org/abs/1409.4842) and 2015 with ResNeXt (Microsoft Research) (https://arxiv.org/abs/1611.05431) and Inception v2, neural network designs moved into wide layers, reducing the need for going deeper in layers. Essentially, a wide layer is having multiple convolutions in parallel and then concatenating their outputs; whereas deeper layers have sequential convolutions and aggregate their outputs.