Skip to content
#

chinese-tokenizer

Here are 8 public repositories matching this topic...

Language: All
Filter by language

一个轻量且功能全面的中文分词器,帮助学生了解分词器的工作原理。MicroTokenizer: A lightweight Chinese tokenizer designed for educational and research purposes. Provides a practical, hands-on approach to understanding NLP concepts, featuring multiple tokenization algorithms and customizable models. Ideal for students, researchers, and NLP enthusiasts..

  • Updated Oct 18, 2024
  • Python

Improve this page

Add a description, image, and links to the chinese-tokenizer topic page so that developers can more easily learn about it.

Curate this topic

Add this topic to your repo

To associate your repository with the chinese-tokenizer topic, visit your repo's landing page and select "manage topics."

Learn more