forked from NVIDIA/Megatron-LM
-
Notifications
You must be signed in to change notification settings - Fork 370
Open
Description
Hello, I encountered the following error when running finetuning Llama2:
Some indexes in the input tensor for the embedding layer are outside the vocabulary dimension. I guess the problem occurs when adding a special tokens to the HF Tokenizer vocabulary in tokenizer.py:
For myself, I solved this by changing the line:
Is this the right way?
Reactions are currently unavailable
Metadata
Metadata
Assignees
Labels
No labels


