You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
One of the scripts in the examples/ folder of Accelerate or an officially supported no_trainer script in the examples folder of the transformers repo (such as run_no_trainer_glue.py)
System Info
Information
Tasks
no_trainer
script in theexamples
folder of thetransformers
repo (such asrun_no_trainer_glue.py
)Reproduction
The script I use
I omit the training code because it doesn't affect we reproduction this bug.
I got two different configurations from the official Accelerate example: one for stage 2 optimization and the other for stage 3 optimization.
when I use stage 2 config to run this code, I can successfully load the trained model(the model has been saved properly)

but when I use stage 3 config to run this code, I can't load the trained model

the load code
Expected behavior
I tried this in qwen2.5-7B and llama3.2-3B. Both of them have this save problems when using the stage 3 optimization config
The text was updated successfully, but these errors were encountered: