-
Notifications
You must be signed in to change notification settings - Fork 1.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Issues Converting lit_model.pth to Huggingface Format Using convert_from_litgpt #1847
Comments
Yes, I followed the tutorial. I noticed that after |
I am having the same problem, except that I copied the config file.
the directory i also got the error
|
Is it possible to have a script that just converts to huggingface readable format directly? Like appending this at the end of the original script.
|
Yes, I guess we can do something like that. cc @rasbt |
any updates on this? I have python=3.10.15
my script for running
|
Same issue, is it simply impossible to load a trained model converted to hf format using AutoModel? |
Hello,
I have been using litgpt to pretrain a model, which produces a
lit_model.pth
file. This model functions correctly when loaded withLLM.load()
for inference.However, when I attempt to convert this model to the Huggingface format using the
convert_from_litgpt
script provided by litgpt, it outputs amodel.pth
file. This file doesn't meet Huggingface's expected format, and when I try to load it using Huggingface's tools, I receive the following error:OSError: Error no file named pytorch_model.bin, model.safetensors, tf_model.h5, model.ckpt.index or flax_model.msgpack found in directory
I am unsure how to resolve this issue. Is there a step I'm missing in the conversion process, or is there a compatibility issue with the
convert_from_litgpt
script?Additionally, I noticed that the
.pth
file obtained after training with lit-gpt is twice the size of the original model. Could you please explain why this is happening?Thank you for your assistance.
The text was updated successfully, but these errors were encountered: