為了《中國哲學書電子化計劃》輸入用-加速鍵入與排版,更好的輸入體驗+文房一寶勝四寶WordVBA文史工具-中文博士寫程式
-
Updated
Nov 14, 2024 - C#
為了《中國哲學書電子化計劃》輸入用-加速鍵入與排版,更好的輸入體驗+文房一寶勝四寶WordVBA文史工具-中文博士寫程式
GUI application for Chinese word segmentation
Python port of SymSpell: 1 million times faster spelling correction & fuzzy search through Symmetric Delete spelling correction algorithm
SymSpell: 1 million times faster spelling correction & fuzzy search through Symmetric Delete spelling correction algorithm
一个轻量且功能全面的中文分词器,帮助学生了解分词器的工作原理。MicroTokenizer: A lightweight Chinese tokenizer designed for educational and research purposes. Provides a practical, hands-on approach to understanding NLP concepts, featuring multiple tokenization algorithms and customizable models. Ideal for students, researchers, and NLP enthusiasts..
A convenient Chinese word segmentation tool 简便中文分词器
rongzi = banyan-like Chinese character-component paths. 永遠:永詠言這辶遠, 青鳥:青靖立鴗鳥
Dictionary for Cantonese word segmentation
The Jieba Chinese Word Segmentation Implemented in Rust
Elasticsearch analysis plugin of ICTCLAS
中文分词
Chinese tokenizer base on nodejieba and pullword
100+ Chinese Word Vectors 上百种预训练中文词向量
High performance Chinese tokenizer with both GBK and UTF-8 charset support based on MMSEG algorithm developed by ANSI C. Completely based on modular implementation and can be easily embedded in other programs, like: MySQL, PostgreSQL, PHP, etc.
Jcseg is a light weight NLP framework developed with Java. Provide CJK and English segmentation based on MMSEG algorithm, With also keywords extraction, key sentence extraction, summary extraction implemented based on TEXTRANK algorithm. Jcseg had a build-in http server and search modules for lucene,solr,elasticsearch,opensearch
Using Flask export jieba, SnowNLP, pkuseg as http API web service.
Jiebago 的性能优化版, 支持从 io.Reader 加载字典
pkuseg多领域中文分词工具; The pkuseg toolkit for multi-domain Chinese word segmentation
从jieba分词到BERT-wwm,一步步带你进入中文NLP的世界
Add a description, image, and links to the chinese-word-segmentation topic page so that developers can more easily learn about it.
To associate your repository with the chinese-word-segmentation topic, visit your repo's landing page and select "manage topics."