#

chinese-word-segmentation

https://static.github-zh.com/github_avatars/Embedding?size=40
Python 12.07 k
2 年前
https://static.github-zh.com/github_avatars/lancopku?size=40

pkuseg多领域中文分词工具; The pkuseg toolkit for multi-domain Chinese word segmentation

Python 6.65 k
3 年前
https://static.github-zh.com/github_avatars/baidu?size=40
C++ 3.97 k
4 年前
https://static.github-zh.com/github_avatars/ownthink?size=40

#自然语言处理#Jiagu深度学习自然语言处理工具 知识图谱关系抽取 中文分词 词性标注 命名实体识别 情感分析 新词发现 关键词 文本摘要 文本聚类

Python 3.41 k
3 年前
https://static.github-zh.com/github_avatars/lionsoul2014?size=40

#自然语言处理#Jcseg is a light weight NLP framework developed with Java. Provide CJK and English segmentation based on MMSEG algorithm, With also keywords extraction, key sentence extraction, summary extraction imp...

Java 922
2 年前
https://static.github-zh.com/github_avatars/messense?size=40
Rust 862
9 天前
https://static.github-zh.com/github_avatars/lionsoul2014?size=40

High performance Chinese tokenizer with both GBK and UTF-8 charset support based on MMSEG algorithm developed by ANSI C. Completely based on modular implementation and can be easily embedded in other...

C 504
2 年前
https://static.github-zh.com/github_avatars/monpa-team?size=40

#自然语言处理#MONPA 罔拍是一個提供正體中文斷詞、詞性標註以及命名實體辨識的多任務模型

Python 247
7 个月前
https://static.github-zh.com/github_avatars/Kyubyong?size=40

g2pC: A Context-aware Grapheme-to-Phoneme Conversion module for Chinese

Python 242
6 年前
https://static.github-zh.com/github_avatars/hemingkx?size=40

A PyTorch implementation of a BiLSTM \ BERT \ Roberta (+ BiLSTM + CRF) model for Chinese Word Segmentation (中文分词) .

Python 210
3 年前
https://static.github-zh.com/github_avatars/howl-anderson?size=40

一个轻量且功能全面的中文分词器,帮助学生了解分词器的工作原理。MicroTokenizer: A lightweight Chinese tokenizer designed for educational and research purposes. Provides a practical, hands-on approach to understanding NLP concepts, fe...

Python 157
1 年前
https://static.github-zh.com/github_avatars/xtea?size=40

#自然语言处理#手工整理医疗行业词汇、术语等语料。可用于语音识别、对话系统等各类nlp模型训练。

121
5 年前
https://static.github-zh.com/github_avatars/jcyk?size=40

Source code for an ACL2017 paper on Chinese word segmentation

Python 91
7 年前
https://static.github-zh.com/github_avatars/fudannlp16?size=40

#计算机科学#Source codes for paper "Neural Networks Incorporating Dictionaries for Chinese Word Segmentation", AAAI 2018

Python 90
8 年前
loading...
Website
Wikipedia