Tokenizer and model vocab size different
#8
by
						
abpani1994
	
							
						- opened
							
					
(deberta): DebertaV2Model(
    (embeddings): DebertaV2Embeddings(
      (word_embeddings): Embedding(128100, 1024, padding_idx=0)
tokenizer.vocab_size = 128000