diff --git a/convert_hf_to_gguf_update.py b/convert_hf_to_gguf_update.py index 9e00df0dc..6a0d9a9ba 100755 --- a/convert_hf_to_gguf_update.py +++ b/convert_hf_to_gguf_update.py @@ -240,11 +240,6 @@ for model in [*pre_computed_hashes, *all_models]: if tokt == TOKENIZER_TYPE.SPM or tokt == TOKENIZER_TYPE.UGM: continue - # Skip if the tokenizer folder does not exist or there are other download issues previously - if not os.path.exists(f"models/tokenizers/{name}"): - logger.warning(f"Directory for tokenizer {name} not found. Skipping...") - continue - # create the tokenizer if chkhsh is not None: # if the model has a pre-computed hash, use it @@ -254,6 +249,12 @@ for model in [*pre_computed_hashes, *all_models]: chkhsh = existing_models[name] else: # otherwise, compute the hash of the tokenizer + + # Skip if the tokenizer folder does not exist or there are other download issues previously + if not os.path.exists(f"models/tokenizers/{name}"): + logger.warning(f"Directory for tokenizer {name} not found. Skipping...") + continue + try: logger.info(f"Loading tokenizer from {f'models/tokenizers/{name}'}...") if name == "t5":