机器学习推荐论文和书籍

发信人: zibuyu (得之我幸), 信区: NLP
标 题: 机器学习推荐论文和书籍
发信站: 水木社区 (Thu Oct 30 21:00:39 2008), 站内

我们组内某小神童师弟通读论文,拟了一个机器学习的推荐论文和书籍列表。
经授权发布在这儿,希望对大家有用。:)



基本模型:

HMM(Hidden Markov Models):
A Tutorial on Hidden Markov Models and Selected Applications in
Speech Recognition.pdf


ME(Maximum Entropy):
ME_to_NLP.pdf


MEMM(Maximum Entropy Markov Models):
memm.pdf


CRF(Conditional Random Fields):
An Introduction to Conditional Random Fields for Relational Learning.pdf


Conditional Random Fields: Probabilistic Models for Segmenting and
Labeling Sequence Data.pdf


SVM(support vector machine):
*张学工<<统计学习理论>>

LSA(or LSI)(Latent Semantic Analysis):
Latent semantic analysis.pdf

pLSA(or pLSI)(Probablistic Latent Semantic Analysis):
Probabilistic Latent Semantic Analysis.pdf


LDA(Latent Dirichlet Allocation):
Latent Dirichlet Allocaton.pdf(用variational theory + EM算法解模型)
Parameter estimation for text analysis.pdf(using Gibbs Sampling 解模)


Neural Networksi(including Hopfield Model& self-organizing maps &
Stochastic networks & Boltzmann Machine etc.):
Neural Networks - A Systematic Introduction


Diffusion Networks:
Diffusion Networks, Products of Experts, and Factor Analysis.pdf


Markov random fields:

Generalized Linear Model(including logistic regression etc.):
An introduction to Generalized Linear Models 2nd


Chinese Restraunt Model (Dirichlet Processes):
Dirichlet Processes, Chinese Restaurant Processes and all that.pdf
Estimating a Dirichlet Distribution.pdf


=================================================================

Some important algorithms:

EM(Expectation Maximization):
Expectation Maximization and Posterior Constraints.pdf
Maximum Likelihood from Incomplete Data via the EM Algorithm.pdf


MCMC(Markov Chain Monte Carlo) & Gibbs Sampling:
Markov Chain Monte Carlo and Gibbs Sampling.pdf
Explaining the Gibbs Sampler.pdf
An introduction to MCMC for Machine Learning.pdf


PageRank:

矩阵分解算法:
SVD, QR分解, Shur分解, LU分解, 谱分解


Boosting( including Adaboost):
*adaboost_talk.pdf


Spectral Clustering:
Tutorial on spectral clustering.pdf


Energy-Based Learning:
A tutorial on Energy-based learning.pdf


Belief Propagation:
Understanding Belief Propagation and its Generalizations.pdf
bp.pdf
Construction free energy approximation and generalized belief
propagation algorithms.pdf
Loopy Belief Propagation for Approximate Inference An Empirical Study.pdf
Loopy Belief Propagation.pdf


AP (affinity Propagation):

L-BFGS:
<<最优化理论与算法 2nd>> chapter 10
On the limited memory BFGS method for large scale optimization.pdf
IIS:
IIS.pdf

=================================================================


理论部分:

概率图(probabilistic networks):
An introduction to Variational Methods for Graphical Models.pdf
Probabilistic Networks
Factor Graphs and the Sum-Product Algorithm.pdf
Constructing Free Energy Approximations and Generalized Belief
Propagation Algorithms.pdf
*Graphical Models, exponential families, and variational inference.pdf


Variational Theory(变分理论,我们只用概率图上的变分):
Tutorial on varational approximation methods.pdf
A variational Bayesian framework for graphical models.pdf
variational tutorial.pdf


Information Theory:
Elements of Information Theory 2nd.pdf

测度论:
测度论(Halmos).pdf
测度论讲义(严加安).pdf

概率论:
……
<<概率与测度论>>


随机过程:
应用随机过程 林元烈 2002.pdf
<<随机数学引论>>


Matrix Theory:
矩阵分析与应用.pdf


模式识别:
<<模式识别 2nd>> 边肇祺
*Pattern Recognition and Machine Learning.pdf


最优化理论:
<<Convex Optimization>>
<<最优化理论与算法>>

泛函分析:
<<泛函分析导论及应用>>

Kernel理论:
<<模式分析的核方法>>

统计学:
……
<<统计手册>>


综合:

semi-supervised learning:
<<Semi-supervised Learning>> MIT Press
semi-supervised learning based on Graph.pdf

Co-training:

Self-training:

  • 1
    点赞
  • 1
    收藏
    觉得还不错? 一键收藏
  • 0
    评论
25篇机器学习经典论文合集,有需要欢迎积分自取 Efficient sparse coding algorithms论文附有代码 [1] Zheng S, Kwok J T. Follow the moving leader in deep learning[C]//Proceedings of the 34th International Conference on Machine Learning-Volume 70. JMLR. org, 2017: 4110-4119. [2] Kalai A, Vempala S. Efficient algorithms for online decision problems[J]. Journal of Computer and System Sciences, 2005, 71(3): 291-307. [3] Kingma, D. and Ba, J. Adam: A method for stochastic optimization. In Proceedings of the International Conference for Learning Representations, 2015. [4] Lee H, Battle A, Raina R, et al. Efficient sparse coding algorithms[C]//Advances in neural information processing systems. 2007: 801-808. [5] Fan J, Ding L, Chen Y, et al. Factor Group-Sparse Regularization for Efficient Low-Rank Matrix Recovery[J]. 2019. [6] Z. Lai, Y. Chen, J. Wu, W. W. Keung, and F. Shen, “Jointly sparse hashing for image retrieval,” IEEE Transactions on Image Processing, vol. 27, no. 12, pp. 6147–6158, 2018. [7] Z. Zhang, Y. Chen, and V. Saligrama, “Efficient training of very deep neural networks for supervised hashing,” in Proc. IEEE Int. Conf. Computer Vision and Pattern Recognition, 2016, pp. 1487–1495. [8] Wei-Shi Zheng, Shaogang Gong, Tao Xiang. Person re-identification by probabilistic relative distance comparison[C]// CVPR 2011. IEEE, 2011. [9] Liao S, Hu Y, Zhu X, et al. Person re-identification by local maximal occurrence representation and metric learning[C]//Proceedings of the IEEE conference on computer vision and pattern recognition. 2015: 2197-2206. [10] Liu X, Li H, Shao J, et al. Show, tell and discriminate: Image captioning by self-retrieval with partially labeled data[C]//Proceedings of the European Conference on Computer Vision (ECCV). 2018: 338-354. [11] Yao T, Pan Y, Li Y, et al. Exploring visual relationship for image captioning[C]//Proceedings of the European conference on computer vision (ECCV). 2018: 684-699. [12] Chao Dong, Chen Change Loy, Kaiming He, and Xiaoou Tang., ”Image Super-Resolution Using Deep Convolutional Networks, ” IEEE Transactions on Pattern Analysis and Machine Intelligence, Preprint, 2015. [13] M. D. Zeiler, D. Krishnan, Taylor, G. W., and R. Fergus, "Deconvolutional networks," in Proc. IEEE Comput. Soc. Conf. Comput. Vision Pattern Recog., 2010, pp. 2528-2535. [14] Girshick R, Donahue J, Darrell T, et al. Rich feature hierarchies for accurate object detection and semantic segmentation[C]//Proceedings of the IEEE conference on computer vision and pattern recognition. 2014: 580-587. [15] Girshick R . Fast R-CNN[J]. Computer Science, 2015. [16] Joseph Redmon, Santosh Divvala, Ross Girshick, et al. You Only Look Once: Unified, Real-Time Object Detection[C]// 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). IEEE, 2016. [17] LeCun Y, Bottou L, Bengio Y, et al. Gradient-based learning applied to document recognition[J]. Proceedings of the IEEE, 1998, 86(11): 2278-2324. [18] Hinton G E, Salakhutdinov R R. Reducing the dimensionality of data with neural networks[J]. science, 2006, 313(5786): 504-507. [19] Krizhevsky A, Sutskever I, Hinton G E. Imagenet classification with deep convolutional neural networks[C]//Advances in neural information processing systems. 2012: 1097-1105. [20] Zeiler M D, Fergus R. Visualizing and understanding convolutional networks[C]//European conference on computer vision. Springer, Cham, 2014: 818-833. [21] Szegedy C, Liu W, Jia Y, et al. Going deeper with convolutions[C]//Proceedings of the IEEE conference on computer vision and pattern recognition. 2015: 1-9. [22] Wu, Y., & He, K. (2018). Group normalization. In Proceedings of the European Conference on Computer Vision (ECCV) (pp. 3-19). [23] Goodfellow I,Pouget-Abadie J, Mirza M, et al. Generative adversarial nets[C]//Advances in Neural Information Processing Systems. 2014: 2672-2680. [24] Tran, L., Yin, X., & Liu, X. (2017). Disentangled representation learning gan for pose-invariant face recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 1415-1424). [25] Pu, Y., Gan, Z., Henao, R., Yuan, X., Li, C., Stevens, A., & Carin, L. (2016). Variational autoencoder for deep learning of images, labels and captions. In Advances in neural information processing systems (pp. 2352-2360).

“相关推荐”对你有帮助么?

  • 非常没帮助
  • 没帮助
  • 一般
  • 有帮助
  • 非常有帮助
提交
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值