A summary of Going deeper with convolutions by Christian Szegedy et al.
Nicholas M. Synovic
- 8 minutes read - 1543 wordsA summary of Going deeper with convolutions
Christian Szegedy et al.; https://doi.org/10.48550/arXiv.1409.4842
For the summary of the paper, go to the Summary section of this article.
Table of Contents
First Pass
Discussion about the title, abstract, introduction, section and sub-section headings, and conclusion
The paper Going deeper with convolutions by Christian Szegedy et al. [1] describes a 2014 state of the art computer vision model (on the ImageNet Large-Scale Visual Recognition Challenge) called GoogLeNet architect ed based on Hebbian principles (i.e., neurons that fire together, are wired together)and a constant computational budget. Their approach relies on creative algorithms and neuroscience principles and aims to be a more power efficient model for mobile devices by limiting the computations during inference. Additionally, their model is deep but not wide and is considered “sparse” by the authors. In other words, there are as few nodes as possible within the neural network.
Szegedy et al.’s contributions are a state of the art computer vision model that provides experimental evidence that, “… Approximating the expected optimal sparse structure by readily available dense building blocks is a viable method for improving neural networks for computer vision”. This means that this model proves that dense neural networks for computer vision are not necessary in the author’s viewpoint.
Category
What type of paper is this work?
This is a computer vision paper describing both a machine learning architecture and reference model.
Context
What other types of papers is the work related to?
This paper is related to other computer vision papers that achieve state of the art performance values based on the ImageNet Large-Scale Visual Recognition Challenge.
Contributions
What are the author’s main contributions?
Szegedy et al.’s contributions are:
- A computer vision model architecture (Inception) that is both sparse and aims to be computationally efficient on mobile (non-server) devices,
- A reference model of the aforementioned computer vision model architecture
- A comparison of previous state of the art work to justify their claims that sparser networks are the future of computer vision models.
Second Pass
Figures, Diagrams, Illustrations, and Graphs
Are the axes properly labeled? Are results shown with error bars, so that conclusions are statistically significant?
All of the tables have proper column labels. However, Table 1 does not provide default values for blank cells. This is most likely due to the layer type not performing a specific operation (as described in the column label). Regardless, the remaining tables look good.
Additionally, Figure 3 is very clear to read, if a little dense. However, as it describes all of the layers of GoogLeNet and how they are connected, I find the size to be appropriate.
Clarity
Is the paper well written?
The paper is fairly well written. The only complaints that I have are minor grammatical mistakes that the author’s left in (by accident I assume). Additionally, that the authors didn’t optimize their tables and figures to better fit on the pages. As tables and figures are stacked on top of one another, it would be possible to reclaim paper space by rearranging multiple tables and figures to be next to one another, with the exception of Figure 3 due to the sheer size of it.
Relevant Work
Mark relevant work for review
The following relevant work can be found in the Citations section of this article.
- Contrast Normalization
- Max Pooling
- Average Pooling
- Softmax Activation
- Dropout - [2]
- Localization Task - [3]
- Gabor Filters - [4]
- Network in Network - [5]
- Rectified Linear Activation - [6]
- Regions with Convolutional Neural Networks - [7]
- Multi-Box Prediction - [8]
- Arora proof - [9]
- LeNet 5 - [10]
- Fisher vectors
- Polyak Averaging - [11]
- Jaccard index
- Selective Search - [12]
- Photometric Distortions - [13]
Author Assumptions
What assumptions does the author(s) make? Are they justified assumptions?
One assumption that the authors make is that over fitting is more prone to occur in large models. Additionally, over fitting can occur when there is not enough labeled examples in a data set when a large model is training. Furthermore, increasing the size of a model increases the number of computations that must be done between layers (e.g., chaining two convolutional layers results in computation cost quadratic ally increasing) Their solutions relies on moving from fully connected to sparsely connected architectures including within convolutional layers. Also, their model architecture is based on the idea that computers are inefficient when, “… Computing numerical calculations on non-uniform sparse data structures”.
They assume that 1x1, 3x3, and 5x5 filters are the proper filters to use, but did not test other size of filters. They also assume that using, “Inception modules” is only useful at higher levels, whereas the initial levels are standard convolutional levels. However, this was not tested either and was due to, “infrastructural inefficiencies” in the implementation.
Finally, that the model that achieved state of the art performance was the best model. The authors had been training and testing other models for months prior, however, it is unclear what the testing methodology was and why a particular model was chosen to compete in the ImageNet competition.
Correctness
Do the assumptions seem valid?
The first paragraph of assumptions seems reasonable and correct. However, the remaining two paragraphs seem unreasonable. This is due to the lack of testing that the author’s put in when optimizing their model with respect to the filter sizes and choosing models. Furthermore, if testing did occur to address these issues, it is not addressed in this paper, thus leaving the reader to wonder why testing wasn’t performed.
Future Directions
My own proposed future directions for the work
Based off of Section 3 (Motivation and High Level Considerations), one promising area of study would be to perform a network architecture search utilizing the principles and reasoning of their approach to other machine learning and computer vision domains.
An enhancement to their work is possible by analyzing what filter sizes most optimal improve performance. Currently the author’s are restricting GoogLeNet to 1x1, 3x3, and 5x5 filter sizes, but this was due to convenience and no data was given to support this.
Summary
A summary of the paper
The paper Going deeper with convolutions by Szegedy et al. [1] introduces a computer vision model architecture called Inception and a reference model called GoogLeNet.
Inception is a model architecture that is both sparse and (attempts to be) computationally efficient during inferencing with only 1.5 billion multiply-add operations allowed. Inception models are composed of multiple Inception modules that are stacked on top of each other. Each Inception module takes in data from the previous layer and passes it into small convolutional filters (i.e., 1x1 typically). There are three of these small filters that are wired to the input of the Inception module, with one of them connected directly to the output. The outputs of two of these filters are then passed into larger filters (i.e., 5x5) to which it is then passed into a DepthConcat function. Additionally, a 3x3 filter is wired to the input of the module and the output of which goes into a 1x1 filter to be passed into the DepthConcat function as well. From there, it is passed into another Inception module and the process repeats.
Note: Depth when referring to two dimensional images refers to the color channel of the image. As images typically have three color channels (i.e., red, green, blue), an image would have a depth of 3.
Example: A 200 pixel by 200 pixel full color spectrum image would be represented as 200x200x3.
It is possible for Inception modules to have an additional connection to the input of the module to perform average pooling for Softmax activation.
GoogLeNet achieved SOTA performance in the ImageNet Large-Scale Visual Recognition Challenge image classification task by having a top-5 error of 6.67% on both the validation and testing data. This is an improvement of 56.5% in comparison to 2012’s SOTA performer (SuperVision) and 2013’s SOTA performer (Clarifai). Additionally, they achieved SOTA performance for the ImageNet Large-Scale Visual Recognition Challenge detection task with a mean average precision of 43.9% utilizing an ensemble inference approach. This model was architected using the Inception architecture with 22 layers. However, not every layer was an Inception module; the first few layers were standard convolutional layers.
The author’s contributions were as follows;
- The Inception computer vision architecture,
- The GoogLeNet SOTA computer vision model for classification and object detection.
My opinion on this paper is that while it is well written, the author’s make numerous assumptions about the optimal performance of their model’s architecture. They don’t test optimal sizes for filters as well as resolving bugs such as the usage of standard convolutional layers early in the model. Both of which can be solved by performing a neural architecture search.
Future work for this paper would involve optimizing the model architecture via a neural architecture search. As well as evaluating the performance of the model by both increasing and decreasing the depth of the model.
Summarization Technique
This paper was summarized using a modified technique proposed by S. Keshav in his work How to Read a Paper [0].