Extra Large Sequence Transformer Model for Chinese Word Segment

Autor: Dezhou Shen
Rok vydání: 2021
Předmět:
DOI: 10.21203/rs.3.rs-149338/v1
Popis: Chinese word segment is widely studied in document analysis. The accuracy of the current popular word segment model, LSTM+CRF, is still not satisfactory. Models trained by the popular dataset often fails in the out-domain situation. In this paper, combining the Transformer-XL layer, the Fully-Connect layer, and the Conditional Random Field layer, the proposed model improved 3.23% in the macro-F1 score, comparing to the BERT+CRF model, on the MSR2005 Chinese word segment test dataset.
Databáze: OpenAIRE