I am going to use nltk.tokenize.word_tokenize on a cluster where my account is very limited by space quota. At home, I downloaded all nltk resource
nltk.tokenize.word_tokenize
nltk
You are right. You need Punkt Tokenizer Models. It has 13 MB and nltk.download('punkt') should do the trick.
nltk.download('punkt')