自動分詞與詞類標記
tokenization, word segmentation and POS tagging
Tokenization
中文斷詞(分詞)
library(jiebaR)
seg <- worker()
txt <- "失業的熊讚陪柯文哲看銀翼殺手"
segment(txt, seg)
#> [1] "失業" "的熊" "讚" "陪" "柯文" "哲看" "銀翼" "殺手"Stemmatization
詞類自動標記 POS tagging
Last updated