A Joint Segmenting and Labeling Approach for Chinese Lexical Analysis

Abstract

This paper introduces an approach which jointly performs a cascade of segmentation and labeling subtasks for Chinese lexical analysis, including word segmentation, named entity recognition and part-of-speech tagging. Unlike the traditional pipeline manner, the cascaded subtasks are conducted in a single step simultaneously, therefore error propagation could be avoided and the information could be shared among multi-level subtasks. In this approach, Weighted Finite State Transducers (WFSTs) are adopted. Within the unified framework of WFSTs, the models for each subtask are represented and then combined into a single one. Thereby, through one-pass decoding the joint optimal outputs for multi-level processes will be reached. The experimental results show the effectiveness of the presented joint processing approach, which significantly outperforms the traditional method in pipeline style.

Cite

Text

Wang et al. "A Joint Segmenting and Labeling Approach for Chinese Lexical Analysis." European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases, 2008. doi:10.1007/978-3-540-87481-2_35

Markdown

[Wang et al. "A Joint Segmenting and Labeling Approach for Chinese Lexical Analysis." European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases, 2008.](https://mlanthology.org/ecmlpkdd/2008/wang2008ecmlpkdd-joint/) doi:10.1007/978-3-540-87481-2_35

BibTeX

@inproceedings{wang2008ecmlpkdd-joint,
  title     = {{A Joint Segmenting and Labeling Approach for Chinese Lexical Analysis}},
  author    = {Wang, Xinhao and Nie, Jiazhong and Luo, Dingsheng and Wu, Xihong},
  booktitle = {European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases},
  year      = {2008},
  pages     = {538-549},
  doi       = {10.1007/978-3-540-87481-2_35},
  url       = {https://mlanthology.org/ecmlpkdd/2008/wang2008ecmlpkdd-joint/}
}