【发布时间】:2021-09-12 06:31:02
【问题描述】:
def mytokenizer(document):
words = tokenizer.tokenize(document.lower())
return words
from nltk.tokenize import sent_tokenize,word_tokenize
mytokenizer('My name is Amar!')
【问题讨论】:
-
嗯,是的。你没有定义
tokenizer。tokenizer应该是什么?它应该来自哪里?
标签: python nltk tokenize stop-words