自定义博客皮肤VIP专享

*博客头图:

格式为PNG、JPG,宽度*高度大于1920*100像素,不超过2MB,主视觉建议放在右侧,请参照线上博客头图

请上传大于1920*100像素的图片!

博客底图:

图片格式为PNG、JPG,不超过1MB,可上下左右平铺至整个背景

栏目图:

图片格式为PNG、JPG,图片宽度*高度为300*38像素,不超过0.5MB

主标题颜色:

RGB颜色,例如:#AFAFAF

Hover:

RGB颜色,例如:#AFAFAF

副标题颜色:

RGB颜色,例如:#AFAFAF

自定义博客皮肤

-+
  • 博客(3)
  • 资源 (7)
  • 收藏
  • 关注

原创 欢迎下载psp的埃舍尔画集主题

下载地址见 http://sns.tgbus.com/theme/psp/item.aspx?id=3873。大概做了一个下午做出来了,软件对格式要求很严,弄的我在拿photoshop做批处理的时候都想做一个进行批处理的批处理了,可惜好像这样会导致重复调用。欢迎下载分享

2009-08-30 12:55:00 355

原创 2个用到数学归纳法的有意思的证明

    数学归纳法是一种很有效的证明方法,同时,在很多时候它也能证明一些很有意思的问题,下面我们来看两个有趣的例子: 问题1:我们如何证明一个由2n × 2n个小方格组成的正方形,在放上一个小方块后,剩下的格子总能用,,,这四种方块无重叠且无间隙的覆盖? 证明:n = 0时,总的方块是一个1×1的方格,在任意放上一个小方块后,方格已经填满。我们可以把这作为归纳的起点,但是为

2009-08-29 13:57:00 1472

原创 数学公式测试

2009-04-30 01:04:00 238

SPEECH and LANGUAGE PROCESSING

SPEECH and LANGUAGE PROCESSING An Introduction to Natural Language Processing, Computational Linguistics, and Speech Recognition Second Edition by Daniel Jurafsky and James H. Martin Last Update January 6, 2009 The 2nd edition is now avaiable. A million thanks to everyone who sent us corrections and suggestions for all the draft chapters. You can find the book at Amazon. Chapter 1 is available below for your reading pleasure. Figures, slides and an instructor's manual are available from Prentice Hall at their Instructor's Resource Site (registration required). An errata page for this edition is available. Material from the 1st edition is still available. Table of Contents Preface 1 Introduction I: Words 2 Regular Expressions and Automata 3 Words and Transducers 4 N-grams 5 Part-of-Speech Tagging 6 Hidden Markov and Maximum Entropy Models II: Speech 7 Phonetics 8 Speech Synthesis 9 Automatic Speech Recognition 10 Speech Recognition: Advanced Topics 11 Computational Phonology III: Syntax 12 Formal Grammars of English 13 Syntactic Parsing 14 Statistical Parsing 15 Features and Unification 16 Language and Complexity IV: Semantics and Pragmatics 17 The Representation of Meaning 18 Computational Semantics 19 Lexical Semantics 20 Computational Lexical Semantics 21 Computational Discourse V: Applications 22 Information Extraction 23 Question Answering and Summarization 24 Dialog and Conversational Agents 25 Machine Translation Chapter 1: Introduction This chapter is largely the same with updated history and pointers to newer applications. (top) Chapter 2: Regular Expressions and Automata This chapter is largely the same with some bug fixes. (top) Chapter 3: Words and Transducers This new version of the chapter still focuses on morphology and FSTs, but is expanded in various ways. There are more details about the formal descriptions of finite-state transducers, many bugs are fixed, and two new sections are added relating to words and subwords. The first new section is on word and sentence tokenization, including algorithms for English as well as the maxmatch algorithm for Chinese word segmentation. The second new section is on spelling correction and minimum edit distance, and is an extended version of the edit-distance section from Chapter 5 of the first edition, with clearer figures for example for explaining the minimum-edit-distance backtrace. (top) Chapter 4: N-grams (Formerly Chapter 6) This updated language model chapter has had a complete overhaul. This draft includes more examples, a more complete description of Good-Turing, expanded sections on practical issues like perplexity and evaluation, language modeling toolkits, including ARPA format, and an overview of modern methods like interpolated Kneser-Ney. (top) Chapter 5: Part-of-Speech Tagging (Formerly Chapter 8) The main change to this revised chapter is a greatly expanded, and hence self-contained, description of bigram and trigram HMM part-of-speech tagging, including Viterbi decoding and deleted interpolation smoothing. Together with the new Chapter 6, this allows a complete introduction to HMMs for courses that don't use the speech recognition chapters. Other changes in this chapter include expanded descriptions of unknown word modeling and part-of-speech tagging in other languages, and many bug fixes. Finally, we've moved this chapter earlier in the book. (top) Chapter 6: Hidden Markov and Maximum Entropy Models (Formerly part of Chapter 7 and Appendix D) This new chapter introduces two sequence models: HMMs and MEMMs. It gives the details of Hidden Markov Models, including Forward, Viterbi, and EM. It then introduces MaxEnt models, begining with linear regression, followed by logistic regression, then the extension to MaxEnt, and finally the MEMM and the Viterbi intuition. (top) Chapter 7: Phonetics (Formerly parts of Chapters 4, 5, and 7) This chapter is an introduction to articulatory and acoustic phonetics for speech processing, as well as foundational tools like the ARPAbet, wavefile formats, phonetic dictionaries, and PRAAT. (top) Chapter 8: Speech Synthesis This is a new chapter on speech synthesis. (top) Chapter 9: Automatic Speech Recognition (Formerly 7) This new significantly-expanded speech recognition chapter gives a complete introduction to HMM-based speech recognition, including extraction of MFCC features, Gaussian Mixture Model acoustic models, and embedded training. (top) Chapter 10: Speech Recognition: Advanced Topics (New Chapter) This new second chapter on speech recognition covers advanced topics like decision-tree clustering for context-dependent phones, advanced decoding (including n-best lists, lattices, confusion networks, and stack decoding), robustness (including MLLR adaptation), discriminative training, and human speech recognition. (top) Chapter 11: Computational Phonology (Formerly parts of Chapters 4, 5, and 7) This chapter is a brief introduction to computational phonology, including phonological and morphological learning, finite-state models, OT, and Stochastic OT. (top) Chapter 12: Formal Grammars of English (Formerly 9) This chapter still focuses on CFGs for English and includes a revamped and somewhat expanded grammar for the ATIS domain. New and expanded sections cover: treebanks with a focus on the Penn Treebank, searching treebanks with tgrep and tgrep2, heads and head-finding rules, dependency grammars, Categorial grammar, and grammars for spoken language processing. (top) Chapter 13: Syntactic Parsing (Formerly 10) The focus of this chapter is still on parsing with CFGs. It now includes sections on CKY, Earley and agenda-based (chart) parsing. In addition, there is a new section on partial parsing with a focus on machine learning based base-phrase chunking and the use of IOB tags. (top) Chapter 14: Statistical Parsing (Formerly 12) This statistical parsing chapter has been extensively revised. It now covers PCFGs, probabilistic CKY parsing, parent annotations, the Collins parser, and touches on advanced topics such as discriminative reranking and parsing for language modeling. (top) Chapter 15: Features and Unification (Formerly 11) Mainly bug fixes. (top) Chapter 16: Language and Complexity (Formerly 13) Mainly bug fixes. (top) Chapter 17: The Representation of Meaning (Formerly 14) This chapter still covers basic notions surrounding meaning representation languages. It now has better coverage of model-theoretic semantics for meaning representations, and a new section on Description Logics and their role as a basis for OWL and its role in the Semantic Web. (top) Chapter 18: Computational Semantics (Formerly 15) This chapter covers compositional approaches to semantic analysis at the sentence level. The primary focus is on rule-to-rule approaches based on lambda-expressions. It also now has new coverage of unification-based approaches to computational semantics. Coverage in the old chapter 15 on semantic grammars has been moved to the discourse chapter; coverage of information extraction has been expanded and moved to the new chapter 22. (top) Chapter 19: Lexical Semantics (Formerly 16) This chapter still covers the basics of lexical semantics, including sense relations, semantic roles, and primitive decomposition. The treatment of semantic roles has been updated, as has the coverage of WordNet, and new sections added for PropBank and FrameNet. (top) Chapter 20: Computational Lexical Semantics (New Chapter; Parts of old Chs. 15, 16 and 17) The focus of this new chapter is on computing with word meanings. The three main topics are word sense disambiguation, computing relations between words (similarity, hyponymy, etc.), and semantic role labeling. It considerably expands the treatment of these topics. (top) Chapter 21: Computational Discourse This rewritten chapter includes a number of updates to the first edition. The anaphora resolution section is updated to include modern log-linear methods, and a section on the more general problem of coreference is also included. The coherence section describes cue-based methods for rhetorical relation and coherence relation extraction. Finally, there is a significant new section on discourse segmentation (including TextTiling). (top) Chapter 22: Information Extraction (New chapter: Parts of old 15) This new chapter surveys current approaches to information extraction. The main topics are named entity recognition, relation detection, temporal expression analysis and template-filling. The primary focus is on supervised machine learning approaches to these topics. The coverage on mostly finite-state methods (FASTUS) has been moved from the old Ch. 15 to here. (top) Chapter 23: Question Answering and Summarization (Mostly new; Parts of old 17 and 20) This new chapter covers two applications, question answering and summarization. A brief introduction to the necessary background material from information retrieval is also included. The chapter includes factoid question answering, single document summarization, generic multiple document summarization, and query-focused summarization. (top) Chapter 24: Dialog and Conversational Agents (Formerly 19) This is a completely rewritten version of the dialogue chapter. It includes much more information on modern dialogue systems, including VoiceXML, confirmation and clarification dialogues, the information-state model, markov decision processes, and other current approaches to dialogue agents. (top) Chapter 25: Machine Translation The MT chapter has been extensively rewritten and a significant new section added covering statistical MT, including IBM Model 1, Model 3, and HMM alignment. A new evaluation section covering human evaluation and Bleu has also been added, as well as sections on SYSTRAN and more details on cross-linguistic divergences. (top)

2010-07-05

李开复——时间因你而不同(txt)完整版

李开复最新自传完整版,李开复——时间因你而不同(txt)完整版

2009-10-18

统计陷阱 How to Lie With Statistics 中文版 清晰 PDF

  《统计陷阱》本书是美国著名的统计学家达菜尔·哈夫的名著。该书自1954年出版至今,多次重印并被译成多国文字,是一本影响深远的经典性著作。《统计陷阱》一书之所以能够历久弥新,是因为其实用性。随着我国经济持续发展,我们将接触到越来越我我的统计数据等等,去粗取精、去集思广益存真、进行鉴现了一些统计语和方法,但作者重说明、轻证明,重文字描述、轻理论推导,并结合活生生的案例,语方轻松诙谐,深入浅出,故即使从培养接触过的统计专业知识的读者也能掌握书中介绍的统计国想和方法。

2009-09-27

A random walk down wall street (漫步华尔街)

描述股市是“有效市场”,股市仅仅是随机游程的经典作品

2009-08-25

letters to a young mathematician

教年轻人怎么做科学研究的书,需要下载一个djvu的阅读器

2009-08-10

空空如也

TA创建的收藏夹 TA关注的收藏夹

TA关注的人

提示
确定要删除当前文章?
取消 删除