1 Motivations
[Increasing Both the Depth and Width of the Network]
• Large number of parameters, more prone to overfitting.
• Increased use of computational resources.
[Motivation] Improve the utility of the computation resources inside the network, then we can increase the depth and width of the network while keeping the computational budget constant.
[Idea] Use 1 × 1 conv layer to
• Increase the representational power of neural networks.
• Dimension reduction to remove computational bottlenecks.
2 Architecture
In a Nutshell (5M Parameters)
• Input (3 × 224 × 224).
• conv1 (64@7 × 7, s2, p3), relu1, pool1 (3 × 3, s2), lrn1, output (64 × 56 × 56).
• conv2-1 (64@1 × 1, s1), re