python
from nltk import word_tokenize
sents = [sent1, sent2]
print(word_tokenize(sent1))
报错:
python
D:\Anaconda3\python.exe "D:/002 知识总结/007 NLP/NLP入门文章/词袋模型与句子相似度.py"
[nltk_data] Error loading punkt: <urlopen error [SSL:
[nltk_data] CERTIFICATE_VERIFY_FAILED] certificate verify failed:
[nltk_data] unable to get local issuer certificate (_ssl.c:1123)>
Traceback (most recent call last):
File "D:/002 知识总结/007 NLP/NLP入门文章/词袋模型与句子相似度.py", line 11, in <module>
print(word_tokenize(sent1))
File "D:\Anaconda3\lib\site-packages\nltk\tokenize\__init__.py", line 129, in word_tokenize
sentences = [text] if preserve_line else sent_tokenize(text, language)
File "D:\Anaconda3\lib\site-packages\nltk\tokenize\__init__.py", line 106, in sent_tokenize
tokenizer = load("tokenizers/punkt/{0}.pickle".format(language))
File "D:\Anaconda3\lib\site-packages\nltk\data.py", line 752, in load
opened_resource = _open(resource_url)
File "D:\Anaconda3\lib\site-packages\nltk\data.py", line 877, in _open
return find(path_, path + [""]).open()
File "D:\Anaconda3\lib\site-packages\nltk\data.py", line 585, in find
raise LookupError(resource_not_found)
LookupError:
**********************************************************************
Resource punkt not found.
Please use the NLTK Downloader to obtain the resource:
>>> import nltk
>>> nltk.download('punkt')
For more information see: https://www.nltk.org/data.html
Attempted to load tokenizers/punkt/english.pickle
Searched in:
- 'C:\\Users\\29617/nltk_data'
- 'D:\\Anaconda3\\nltk_data'
- 'D:\\Anaconda3\\share\\nltk_data'
- 'D:\\Anaconda3\\lib\\nltk_data'
- 'C:\\Users\\29617\\AppData\\Roaming\\nltk_data'
- 'C:\\nltk_data'
- 'D:\\nltk_data'
- 'E:\\nltk_data'
- ''
**********************************************************************
进程已结束,退出代码为 1
解决方法:
【Python】nltk库使用报错之punkt安装:https://blog.csdn.net/weixin_43896318/article/details/106191856