卷积神经网络的并行化模型——One weird trick for parallelizing convolutional neural networks
总结:One weird trick for parallelizing convolutional neural networks
『 论文阅读』Understanding deep learning requires rethinking generalization
16 On Large-Batch Training for Deep Learning: Generalization Gap and Sharp Minima 1609.04836v1