基于深度学习的自然语言处理库
-
Updated
Nov 3, 2018 - Python
基于深度学习的自然语言处理库
一个轻量且功能全面的中文分词器,帮助学生了解分词器的工作原理。MicroTokenizer: A lightweight Chinese tokenizer designed for educational and research purposes. Provides a practical, hands-on approach to understanding NLP concepts, featuring multiple tokenization algorithms and customizable models. Ideal for students, researchers, and NLP enthusiasts..
A NLP package for Chinese text:Preprocessing, Tokenization, Chinese Fonts, Word Embeddings, Text Similarity and Sentiment Analysis 轻量级中文自然语言处理软件包
中文分词软件基准测试 | Chinese tokenizer benchmark
Add a description, image, and links to the chinese-tokenizer topic page so that developers can more easily learn about it.
To associate your repository with the chinese-tokenizer topic, visit your repo's landing page and select "manage topics."