Replies: 2 comments 1 reply
-
Could you please share the full error message and what changes you made to the notebook that resulted in this error? I did a quick test with: training_args = Seq2SeqTrainingArguments(
...
eval_strategy="steps",
save_strategy="steps",
save_steps=5,
load_best_model_at_end=True,
)
from transformers import EarlyStoppingCallback
trainer = Seq2SeqTrainer(
...
callbacks=[
EarlyStoppingCallback(early_stopping_patience=2)
],
) and I didn't get any error. |
Beta Was this translation helpful? Give feedback.
-
My training args trainer = Seq2SeqTrainer( 7%|▋ | 725/10000 [11:28:17<52:44:42, 20.47s/it] Process finished with exit code 0 This is what I get. |
Beta Was this translation helpful? Give feedback.
-
I am using this code to finetune whisper on custom dataset. I was trying to set early_stopping_patience but i am getting error when I set load_best_model_at_end=True, what can i do to use early stopping to avoid overfitting.
I am doing this because i tired to not set max_steps, but got very bad accuracy. so then i started setting max_steps to as high as 6000 and the results are amazing. But I am scared of overfitting.
Is there any other workaround for avoiding overfitting in this scenario.
Beta Was this translation helpful? Give feedback.
All reactions