You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I encounter the following issue when running llama3-8b-instruct and mistral-7b-olaph with the command:CUDA_VISIBLE_DEVICES=0 python pdata_collection.py
--model_name_or_path ${HUGGINGFACE_MODEL_DIR}
--eval_data ${DATA_NAME}
issue:
The attention mask and the pad token id were not set. As a consequence, you may observe unexpected behavior. Please pass your input's attention_mask to obtain reliable results.
Setting pad_token_id to eos_token_id:2 for open-end generation.
The tokenizer class you load from this checkpoint is not the same type as the class this function is called from. It may result in unexpected tokenization.
The tokenizer class you load from this checkpoint is 'BleurtSPTokenizer'.
The class this function is called from is 'BertTokenizer'.
Token indices sequence length is longer than the specified maximum sequence length for this model (546 > 512). Running this sequence through the model will result in indexing errors
Attempting to cast a BatchEncoding to type None. This is not supported.
Some weights of RobertaModel were not initialized from the model checkpoint at roberta-large and are newly initialized: ['roberta.pooler.dense.bias', 'roberta.pooler.dense.weight']
You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
The text was updated successfully, but these errors were encountered:
I encounter the following issue when running llama3-8b-instruct and mistral-7b-olaph with the command:CUDA_VISIBLE_DEVICES=0 python pdata_collection.py
--model_name_or_path ${HUGGINGFACE_MODEL_DIR}
--eval_data ${DATA_NAME}
issue:
The attention mask and the pad token id were not set. As a consequence, you may observe unexpected behavior. Please pass your input's
attention_mask
to obtain reliable results.Setting
pad_token_id
toeos_token_id
:2 for open-end generation.The tokenizer class you load from this checkpoint is not the same type as the class this function is called from. It may result in unexpected tokenization.
The tokenizer class you load from this checkpoint is 'BleurtSPTokenizer'.
The class this function is called from is 'BertTokenizer'.
Token indices sequence length is longer than the specified maximum sequence length for this model (546 > 512). Running this sequence through the model will result in indexing errors
Attempting to cast a BatchEncoding to type None. This is not supported.
Some weights of RobertaModel were not initialized from the model checkpoint at roberta-large and are newly initialized: ['roberta.pooler.dense.bias', 'roberta.pooler.dense.weight']
You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
The text was updated successfully, but these errors were encountered: