经验总结:深度学习怎么找idea发论文?

深度学习属于人工智能的连接学派。如果你仔细看看或许觉得神经网络和集成电路比较相似。这里面暗含了一个基本思想:模块化。所以深度学习通俗一点就是“积木游戏”。所以深度学习创新可以类比积木游戏创新就主要来源于三个方面:(1)创造新的积木类型。(2)创造新的堆积木的方式,堆出不同形状。(3)堆积方法用于不同的游戏场景。

以下从这三个方面展开来讲:

(1)创造新的积木类型

拿激活函数举例:relu、leakly relu、parameter relu... sigmoid、tanh。按网络基础构成算子举例:全连接fc,卷积conv,循环算子lstm,gru。

尝试着从数学库里面找一个理论,将它算子化,也许这就是一篇不错的论文。以前专家总是手工来寻找这样的算子,现在google的大神们借助算力使用automl,autodl来自动寻找算子,如果你也有功能强大的算力你也可以尝试着用算力自动寻找新的“积木算子”,如果这个算子恰好又是神奇的提高了神经网络的性能,那么人工智能深度学习领域又将多一个大神一样的人,而这个人恰好就是你。

(2)创造新的堆积木方式

resnet很大程度受益于highway network的启发,将X直连到下一层这个概念在论文发出来之后其实看看觉得蛮直观的,但为什么其他人都想不到呢?这个问题我也在问自己,O(∩_∩)O哈哈~。再举个例子:Conv层如何堆,是同型Conv堆积还是不同kernel size的conv进行堆积,也许这就是alexnet,googlenet等不同卷积网络的区别。

目前google也借助遗传算法等自动找到新的堆积木的方式,发了很多高水平的论文,如果我告诉大拿科学家他们这仅仅是堆积木的方法不同“无他但手熟尔“,他们会不会生气打我。

(3)堆出的积木形状用于不同的场景

Attention注意力机制最先出现在NLP领域,后来进入了语音识别和合成领域还有CV计算机视觉领域。仅仅是相同的积木用在了不同的场景罢了,最先发现新的应用场景的人成了他所在领域的大神,如果我说大神仅仅是搬运工相信大神也会打我的。

以上三种方式不一定概括全面,但我认为能做到上述任意一条的都是一个不错的想法。当然"积木游戏"要玩的好,一定需要“多玩”,所谓多玩意味着多看论文看看人家怎么玩,玩的好的人怎么玩,学着按别人的套路玩,最后自己找到一种新玩法,反复训练自己的新玩法达到纯熟,干翻一堆老玩家,你就是终极玩家了。

综上所述:

深度学习不简单,但也不必想的过于复杂。灵活的应用上述提到的三条,随意创新、大胆实践做实验,相信很快你就能发出很多高水平的论文了。

两句话总结:

1.实践是检验真理的唯一标准

2.无他但手熟尔

 

  • 2
    点赞
  • 9
    收藏
    觉得还不错? 一键收藏
  • 0
    评论
不管你想做什么,你都要好好的从论文看,而不是单纯的调论文写代码!通过这些学习,你才能真正的对深度学习的展,模型的优化,进经典的trick有深入的理解! 做算法,做科研必不可少!时间有限的人可以只看1.3 2.1 2.2 !(强烈推荐!) ## 1.3 ImageNet Evolution(Deep Learning broke out from here) **[4]** Krizhevsky, Alex, Ilya Sutskever, and Geoffrey E. Hinton. "**Imagenet classification with deep convolutional neural networks**." Advances in neural information processing systems. 2012. [[pdf]](http://papers.nips.cc/paper/4824-imagenet-classification-with-deep-convolutional-neural-networks.pdf) **(AlexNet, Deep Learning Breakthrough)** :star::star::star::star::star: **[5]** Simonyan, Karen, and Andrew Zisserman. "**Very deep convolutional networks for large-scale image recognition**." arXiv preprint arXiv:1409.1556 (2014). [[pdf]](https://arxiv.org/pdf/1409.1556.pdf) **(VGGNet,Neural Networks become very deep!)** :star::star::star: **[6]** Szegedy, Christian, et al. "**Going deeper with convolutions**." Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2015. [[pdf]](http://www.cv-foundation.org/openaccess/content_cvpr_2015/papers/Szegedy_Going_Deeper_With_2015_CVPR_paper.pdf) **(GoogLeNet)** :star::star::star: **[7]** He, Kaiming, et al. "**Deep residual learning for image recognition**." arXiv preprint arXiv:1512.03385 (2015). [[pdf]](https://arxiv.org/pdf/1512.03385.pdf) **(ResNet,Very very deep networks, CVPR best paper)** :star::star::star::star::star: #2 Deep Learning Method ## 2.1 Model **[14]** Hinton, Geoffrey E., et al. "**Improving neural networks by preventing co-adaptation of feature detectors**." arXiv preprint arXiv:1207.0580 (2012). [[pdf]](https://arxiv.org/pdf/1207.0580.pdf) **(Dropout)** :star::star::star: **[15]** Srivastava, Nitish, et al. "**Dropout: a simple way to prevent neural networks from overfitting**." Journal of Machine Learning Research 15.1 (2014): 1929-1958. [[pdf]](https://www.cs.toronto.edu/~hinton/absps/JMLRdropout.pdf) :star::star::star: **[16]** Ioffe, Sergey, and Christian Szegedy. "**Batch normalization: Accelerating deep network training by reducing internal covariate shift**." arXiv preprint arXiv:1502.03167 (2015). [[pdf]](http://arxiv.org/pdf/1502.03167) **(An outstanding Work in 2015)** :star::star::star::star: **[17]** Ba, Jimmy Lei, Jamie Ryan Kiros, and Geoffrey E. Hinton. "**Layer normalization**." arXiv preprint arXiv:1607.06450 (2016). [[pdf]](https://arxiv.org/pdf/1607.06450.pdf?utm_source=sciontist.com&utm_medium=refer&utm_campaign=promote) **(Update of Batch Normalization)** :star::star::star::star: **[18]** Courbariaux, Matthieu, et al. "**Binarized Neural Networks: Training Neural Networks with Weights and Activations Constrained to+ 1 or−1**." [[pdf]](https://pdfs.semanticscholar.org/f832/b16cb367802609d91d400085eb87d630212a.pdf) **(New Model,Fast)** :star::star::star: **[19]** Jaderberg, Max, et al. "**Decoupled neural interfaces using synthetic gradients**." arXiv preprint arXiv:1608.05343 (2016). [[pdf]](https://arxiv.org/pdf/1608.05343) **(Innovation of Training Method,Amazing Work)** :star::star::star::star::star: **[20]** Chen, Tianqi, Ian Goodfellow, and Jonathon Shlens. "Net2net: Accelerating learning via knowledge transfer." arXiv preprint arXiv:1511.05641 (2015). [[pdf]](https://arxiv.org/abs/1511.05641) **(Modify previously trained network to reduce training epochs)** :star::star::star: **[21]** Wei, Tao, et al. "Network Morphism." arXiv preprint arXiv:1603.01670 (2016). [[pdf]](https://arxiv.org/abs/1603.01670) **(Modify previously trained network to reduce training epochs)** :star::star::star: ## 2.2 Optimization **[22]** Sutskever, Ilya, et al. "**On the importance of initialization and momentum in deep learning**." ICML (3) 28 (2013): 1139-1147. [[pdf]](http://www.jmlr.org/proceedings/papers/v28/sutskever13.pdf) **(Momentum optimizer)** :star::star: **[23]** Kingma, Diederik, and Jimmy Ba. "**Adam: A method for stochastic optimization**." arXiv preprint arXiv:1412.6980 (2014). [[pdf]](http://arxiv.org/pdf/1412.6980) **(Maybe used most often currently)** :star::star::star: **[24]** Andrychowicz, Marcin, et al. "**Learning to learn by gradient descent by gradient descent**." arXiv preprint arXiv:1606.04474 (2016). [[pdf]](https://arxiv.org/pdf/1606.04474) **(Neural Optimizer,Amazing Work)** :star::star::star::star::star: **[25]** Han, Song, Huizi Mao, and William J. Dally. "**Deep compression: Compressing deep neural network with pruning, trained quantization and huffman coding**." CoRR, abs/1510.00149 2 (2015). [[pdf]](https://pdfs.semanticscholar.org/5b6c/9dda1d88095fa4aac1507348e498a1f2e863.pdf) **(ICLR best paper, new direction to make NN running fast,DeePhi Tech Startup)** :star::star::star::star::star: **[26]** Iandola, Forrest N., et al. "**SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and< 1MB model size**." arXiv preprint arXiv:1602.07360 (2016). [[pdf]](http://arxiv.org/pdf/1602.07360) **(Also a new direction to optimize NN,DeePhi Tech Startup)** :star::star::star::star:

“相关推荐”对你有帮助么?

  • 非常没帮助
  • 没帮助
  • 一般
  • 有帮助
  • 非常有帮助
提交
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值