- 博客(1)
- 资源 (28)
- 收藏
- 关注
转载 深度学习
目录: 一、概述 二、背景 三、人脑视觉机理 四、关于特征 4.1、特征表示的粒度 4.2、初级(浅层)特征表示 4.3、结构性特征表示 4.4、需要有多少个特征? 五、Deep Learning的基本思想 六、浅层学习(Shallow Learning)和深度学习(Deep Learning) 七、Deep learn
2014-04-25 10:23:30 20450 5
ods-java-data-struction.pdf
Java数据结构电子书:There are plenty of books that teach introductory data structures. Some of them are very good. Most of them cost money, and the vast majority of computer science undergraduate students will shell out at least some cash on a data structures book.
Several free data structures books are available online. Some are very good, but most of them are getting old. The majority of these books be- came free when their authors and/or publishers decided to stop updat- ing them. Updating these books is usually not possible, for two reasons: (1) The copyright belongs to the author and/or publisher, either of whom may not allow it. (2) The source code for these books is often not avail- able. That is, the Word, WordPerfect, FrameMaker, or LATEX source for the book is not available, and even the version of the software that han- dles this source may not be available.
The goal of this project is to free undergraduate computer science stu- dents from having to pay for an introductory data structures book. I have decided to implement this goal by treating this book like an Open Source software project. The LATEX source, Java source, and build scripts for the book are available to download from the author’s website1 and also, more importantly, on a reliable source code management site.2
2019-05-30
分布式一致性一般解决方法.pdf
Distributed consensus, the ability to reach agreement in the face of failures and asynchrony, is a fundamental primitive for constructing reliable distributed systems from unreliable components. The Paxos algorithm is synonymous with distributed consensus, yet it performs poorly in practice and is famously difficult to understand. In this paper, we re-examine the foundations of distributed consensus. We derive an abstract solution to consensus, which utilises immutable state for intuitive reasoning about safety. We prove that our abstract solution generalises over Paxos as well as the Fast Paxos and Flexible Paxos algorithms. The surprising result of this analysis is a substantial weakening to the quorum requirements of these widely studied algorithms.
2019-05-30
机器学习前沿技术 深度学习
Being a PhD student in the machine learning group of the University of Toronto was lots of fun, and joining it was one of the best decisions that I have ever made. I want to thank my adviser, Geoff Hinton. Geoff taught me how to really do research and our meetings were the highlight of my week. He is an excellent mentor who gave me the freedom and the encouragement to pursue my own ideas and the opportunity to attend many conferences. More importantly, he gave me his unfailing help and support whenever it was needed. I am grateful for having been his student.
I am fortunate to have been a part of such an incredibly fantastic ML group. I truly think so. The atmosphere, faculty, postdocs and students were outstanding in all dimensions, without exaggeration. I want to thank my committee, Radford Neal and Toni Pitassi, in particular for agreeing to read my thesis so quickly. I want to thank Rich for enjoyable conversations and for letting me attend the Z-group meetings.
I want to thank the current learning students and postdocs for making the learning lab such a fun en- vironment: Abdel-Rahman Mohamed, Alex Graves, Alex Krizhevsky, Charlie Tang, Chris Maddison, Danny Tarlow, Emily Denton, George Dahl, James Martens, Jasper Snoek, Maks Volkovs, Navdeep Jaitly, Nitish Srivastava, and Vlad Mnih. I want to thank my officemates, Kevin Swersky, Laurent Charlin, and Tijmen Tieleman for making me look forward to arriving to the office. I also want to thank the former students and postdocs whose time in the group overlapped with mine: Amit Gruber, Andriy Mnih, Hugo Larochelle, Iain Murray, Jim Huang, Inmar Givoni, Nikola Karamanov, Ruslan Salakhutdinov, Ryan P. Adams, and Vinod Nair. It was lots of fun working with Chris Maddison in the summer of 2011. I am deeply indebted to my collaborators: Andriy Mnih, Charlie Tang, Danny Tarlow, George Dahl, Graham Taylor, James Cook, Josh Tenenbaum, Kevin Swersky, Nitish Srivastava, Ruslan Salakhutdinov, Ryan P. Adams, Tim Lillicrap, Tijmen Tieleman, Toma´sˇ Mikolov, and Vinod Nair; and especially to Alex Krizhevsky and James Martens. I am grateful to Danny Tarlow for discovering T&M; to Relu Patrascu for stimulating conversations and for keeping our computers working smoothly; and to Luna Keshwah for her excellent administrative support. I want to thank students in other groups for making school even more enjoyable: Abe Heifets, Aida Nematzadeh, Amin Tootoonchian, Fernando Flores-Mangas, Izhar Wallach, Lena Simine-Nicolin, Libby Barak, Micha Livne, Misko Dzamba, Mo- hammad Norouzi, Orion Buske, Siavash Kazemian, Siavosh Benabbas, Tasos Zouzias, Varada Kolhatka, Yulia Eskin, Yuval Filmus, and anyone else I might have forgot. A very special thanks goes to Annat Koren for making the writing of the thesis more enjoyable, and for proofreading it.
But most of all, I want to express the deepest gratitude to my family, and especially to my parents, who have done two immigrations for me and my brother’s sake. Thank you. And to my brother, for being a good sport.
2013-08-25
空空如也
TA创建的收藏夹 TA关注的收藏夹
TA关注的人