1-分词(python)
# 结巴分词
import jieba.posseg
# 北大分词
import pkuseg
# 清华分词
import thulac
# -
from snownlp import SnowNLP
# 斯坦福
from stanfordcorenlp import StanfordCoreNLP
2-文法分析
- fastNLP/复旦NLP工具包
https://github.com/fastnlp/fastNLP
英语
- StanfordNLP
https://github.com/stanfordnlp/stanfordnlp [python] - NLTK
https://github.com/nltk/nltk [python] - spacy
https://github.com/explosion/spaCy [python/cython] - MeTA
https://github.com/meta-toolkit/meta [c++] - Turbo
https://github.com/andre-martins/TurboParser [c++]
中文
-
NLTK
https://github.com/nltk/nltk [python] -
ltp
https://github.com/HIT-SCIR/ltp -
lac
https://github.com/baidu/lac -
StanfordNLP
https://github.com/stanfordnlp/stanfordnlp [python] -
thulac
https://github.com/thunlp/THULAC [c++]
https://github.com/thunlp/THULAC-Python [python] -
jieba
https://github.com/fxsjy/jieba [python] -
SnowNLP
https://github.com/isnowfy/snownlp [python]
https://www.cnblogs.com/zhuminghui/p/10953717.html -
pynlpir
ps:这里列举的工具的源码 语言可能不仅仅只有python/c++,其他语言请自行 https://github.com
未完待续