搜索资源列表
tmsvm_src_v1.1.0
- SVM中文分词分类器的python实现,在LINUX环境下运行不需太多配置和代码修改。-SVM for Chinese word classification python implementation running under LINUX environment without too much configuration and code changes.
python-ictclas
- 运用python编写的中文分词,运行try文件并将路径改变就可以实现中文分词 中文存储在txt文档中-Use python to write the Chinese word, you can run try to change file and path to achieve the Chinese word Chinese storage txt document
work
- 贝叶斯下文本分类器,python语言,包括中文分词和英文分词(Bias's lower text classifier, python language, including Chinese participle and English participle)
ChineseChuLi
- 中文文本处理的python程序,包括分词、删除特殊字符、删除停用词、爬虫程序、PCA降维、Kmean聚类、可视化等(Python programs for Chinese text processing, including participle, deleting special characters, deleting disuse words, crawler programs, PCA dimensionality reduction, Kmean clustering, visuali
ChPreprocess
- 使用jieba包从excel表中读取数据,进行中文分词,预料分析(Using Jieba package for Chinese analysis, expected analysis)
chinese_seg_update
- 中文分词,采用逆向最大匹配方法实现,利用字典作为索引(The Chinese word segmentation is realized by the reverse maximum matching method, and the dictionary is used as the index.)
chinese_wordseg_keras-master
- 这个程序是用来做中文分词的程序,用Python写的(This program is used to do Chinese word segmentation program, written in Python)
Python逆向最大匹配实现-代码-文档-截图等
- 对于文本进行自然语言处理,中文分词。北邮计算机院NLP课的作业。(This is a homework for BUPT student, implementing the word segment of natural language processing.)
BosoNnlp
- 玻森专注中文语义分析技术,拥有丰富的经验积累。自主研发千万级中文语料库,为精准和深度的中文语义分析提供坚实基础。一站式解决您的中文语义分析需求。多个语义分析API,从情感倾向、实体、分类等方面。官网只有python版本,现上传Java的实例。(He focuses on Chinese semantic analysis technology and has rich experience. Independently developing tens of millions of Chinese
HMM-master
- 隐马尔科夫模型进行中文分词 模型训练 python HMM_train.py RenMinData.txt_utf8 RenMinData.RenMinData_utf8 为人民日报已经人工分词的预料。 生成三个文件 * prob_start.py 为模型的初始概率 * prob_trans.py 为模型状态转移概率 * prob_emit.py 为发射概率 测试模型效果 python HMM.py reference 维特比算法:(python