Chinese tokenizer implement with C, and package with python. c 语言开发的一款开源的高性能中文分词器,使用流行的mmseg算法实现。
python3 setup.py build_ext --inplace
python3 test.py
python3 setup.py install
dist to pypi with wheel(can not pack data file till now).
- contact me [email protected] or www.codeyourlife.cn
- The core C implement of this project friso.