mirror of
https://github.com/ggml-org/llama.cpp.git
synced 2025-06-26 11:45:21 +00:00
gguf-py : fix SpecialVocab parsing when post_processor is null (#14330)
This commit is contained in:
@ -167,7 +167,7 @@ class SpecialVocab:
|
|||||||
tokenizer_config['bos_token'] = special_bos = special_cls
|
tokenizer_config['bos_token'] = special_bos = special_cls
|
||||||
if not special_eos and special_sep and tokenizer_config:
|
if not special_eos and special_sep and tokenizer_config:
|
||||||
tokenizer_config['eos_token'] = special_eos = special_sep
|
tokenizer_config['eos_token'] = special_eos = special_sep
|
||||||
post_processor = tokenizer.get('post_processor', {})
|
if post_processor := tokenizer.get('post_processor'):
|
||||||
for processor in post_processor.get('processors', [post_processor]):
|
for processor in post_processor.get('processors', [post_processor]):
|
||||||
if processor.get('type') == 'RobertaProcessing':
|
if processor.get('type') == 'RobertaProcessing':
|
||||||
self.add_special_token['bos'] = True
|
self.add_special_token['bos'] = True
|
||||||
|
Reference in New Issue
Block a user