Traceback (most recent call last):
File "/root/autodl-tmp/scy/ai-llm-models/api.py", line 142, in <module>
model, tokenizer = init_model()
File "/root/autodl-tmp/scy/ai-llm-models/api.py", line 37, in init_model
tokenizer = AutoTokenizer.from_pretrained(
File "/root/miniconda3/envs/ai-llm-models/lib/python3.10/site-packages/transformers/models/auto/tokenization_auto.py", line 774, in from_pretrained
return tokenizer_class.from_pretrained(pretrained_model_name_or_path, *inputs, **kwargs)
File "/root/miniconda3/envs/ai-llm-models/lib/python3.10/site-packages/transformers/tokenization_utils_base.py", line 2028, in from_pretrained
return cls._from_pretrained(
File "/root/miniconda3/envs/ai-llm-models/lib/python3.10/site-packages/transformers/tokenization_utils_base.py", line 2260, in _from_pretrained
tokenizer = cls(*init_inputs, **init_kwargs)
File "/root/.cache/huggingface/modules/transformers_modules/Baichuan2-13B-Chat-lora23/tokenization_baichuan.py", line 71, in __init__
super().__init__(
File "/root/miniconda3/envs/ai-llm-models/lib/python3.10/site-packages/transformers/tokenization_utils.py", line 367, in __init__
self._add_tokens(
File "/root/miniconda3/envs/ai-llm-models/lib/python3.10/site-packages/transformers/tokenization_utils.py", line 467, in _add_tokens
current_vocab = self.get_vocab().copy()
File "/root/.cache/huggingface/modules/transformers_modules/Baichuan2-13B-Chat-lora23/tokenization_baichuan.py", line 105, in get_vocab
vocab = {self.convert_ids_to_tokens(i): i for i in range(self.vocab_size)}
File "/root/.cache/huggingface/modules/transformers_modules/Baichuan2-13B-Chat-lora23/tokenization_baichuan.py", line 101, in vocab_size
return self.sp_model.get_piece_size()
AttributeError: 'BaichuanTokenizer' object has no attribute 'sp_model'