Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Load the checkpoint fails #600

Open
AfraAmini opened this issue Sep 6, 2024 · 0 comments
Open

Load the checkpoint fails #600

AfraAmini opened this issue Sep 6, 2024 · 0 comments
Labels
bug Something isn't working

Comments

@AfraAmini
Copy link

🐛 Describe the bug

I am training a model using PPO with 8 unfrozen layers. I have the checkpoint saved under hf_model directory. When I attempt to load the checkpoint I get the error that layers are not initialized from the checkpoint and there are extra keys (base_model. and frozen_head.) that were not matched with any keys in the GPT model. Here is how I attempted to load the model:

from transformers import AutoModelForCausalLM
model = AutoModelForCausalLM.from_pretrained("ckpts/checkpoint_6400/hf_model")

Which trlX version are you using?

No response

Additional system and package information

No response

@AfraAmini AfraAmini added the bug Something isn't working label Sep 6, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests

1 participant