Parveshiiii commited on
Commit
f10d685
·
verified ·
1 Parent(s): b968826

Fix incorrect vocab_size in Qwen3-8B config.json

Browse files

While fine-tuning Qwen3-8B, I encountered a mismatch between the vocab_size specified in config.json (151936) and the actual tokenizer size (151669) reported by Qwen2TokenizerFast. This discrepancy can lead to shape mismatch errors when resizing the embedding layer or loading the model for training.

This PR updates the vocab_size field in config.json to reflect the correct tokenizer size of 151669, ensuring consistency across model weights, tokenizer, and configuration.

No architectural changes were made — only the vocabulary alignment was corrected.

Files changed (1) hide show
  1. config.json +1 -1
config.json CHANGED
@@ -26,5 +26,5 @@
26
  "transformers_version": "4.51.0",
27
  "use_cache": true,
28
  "use_sliding_window": false,
29
- "vocab_size": 151936
30
  }
 
26
  "transformers_version": "4.51.0",
27
  "use_cache": true,
28
  "use_sliding_window": false,
29
+ "vocab_size": 151669
30
  }