transformers==4.42.0会有一个BUG

transformers==4.42.0版本下,自动安装模型时出现一个BUG(自动从Hugging Faces上下载)。

2025-07-02 14:07:08,641 - main - ERROR - 模型加载失败: Failed to import transformers.models.llama.tokenization_llama_fast because of the following error (look up to see its traceback):

(unicode error) 'utf-8' codec can't decode bytes in position 3344-3345: invalid continuation byte (tokenization_llama_fast.py, line 120)

ERROR: Traceback (most recent call last):

File "/root/miniconda3/envs/我的conda名称/lib/python3.10/site-packages/transformers/utils/import_utils.py", line 1560, in _get_module

return importlib.import_module("." + module_name, self.name)

File "/root/miniconda3/envs/我的conda名称/lib/python3.10/importlib/init.py", line 126, in import_module

return _bootstrap._gcd_import(name[level:], package, level)

File "<frozen importlib._bootstrap>", line 1050, in _gcd_import

File "<frozen importlib._bootstrap>", line 1027, in _find_and_load

File "<frozen importlib._bootstrap>", line 1006, in _find_and_load_unlocked

File "<frozen importlib._bootstrap>", line 688, in _load_unlocked

File "<frozen importlib._bootstrap_external>", line 879, in exec_module

File "<frozen importlib._bootstrap_external>", line 1017, in get_code

File "<frozen importlib._bootstrap_external>", line 947, in source_to_code

File "<frozen importlib._bootstrap>", line 241, in _call_with_frames_removed

File "/root/miniconda3/envs/我的conda名称/lib/python3.10/site-packages/transformers/models/llama/tokenization_llama_fast.py", line 120

"""

^

SyntaxError: (unicode error) 'utf-8' codec can't decode bytes in position 3344-3345: invalid continuation byte

其实表面上看就知道这个是一个字符编码的问题,但是网上没有什么简单的针对方案。其实你找到这个文件,打开一开似乎还看不出 tokenization_llama_fast.py, line 120 有什么问题,而且你还会发现这一行其实就是一个注释,最简单的方式就是清除注释,之后运行就正常了,非常简单。