What to download in order to make nltk.tokenize.word_tokenize work?

后端 未结 2 1856
温柔的废话
温柔的废话 2020-12-30 02:23

I am going to use nltk.tokenize.word_tokenize on a cluster where my account is very limited by space quota. At home, I downloaded all nltk resource

2条回答
  •  感情败类
    2020-12-30 03:20

    You are right. You need Punkt Tokenizer Models. It has 13 MB and nltk.download('punkt') should do the trick.

提交回复
热议问题