A tokenizer and sentence splitter for German and English web and social media texts.
-
Updated
Dec 9, 2024 - Python
A tokenizer and sentence splitter for German and English web and social media texts.
A sentence splitting (sentence boundary disambiguation) library for Go. It is rule-based and works out-of-the-box.
Smallish library for sentence splitting in Julia
Preprocessing tool for Turkish NLP that contains tokenizer, normalizer, stop-word eliminator and stemmer
A CLI for Rust SRX sentence segmenation rules as Python package.
Add a description, image, and links to the sentence-splitter topic page so that developers can more easily learn about it.
To associate your repository with the sentence-splitter topic, visit your repo's landing page and select "manage topics."