thsheep / Chinese-tokenizer

sklearn中 CountVectorizer 的tokenizer参数(方便序列化)仅自用,有问题概不负责

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

"# Chinese-tokenizer"

安装:

pip install chinese-tokenizer

使用:

from sklearn.feature_extraction.text import CountVectorizer

from chinese_tokenizer.tokenizer import Tokenizer

jie_ba_tokenizer = Tokenizer().jie_ba_tokenizer

count_vect = CountVectorizer(tokenizer=jie_ba_tokenizer)

个人使用 不保证可用性

About

sklearn中 CountVectorizer 的tokenizer参数(方便序列化)仅自用,有问题概不负责


Languages

Language:Python 100.0%