You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hey, I fine-tuned a LLM using lit got (llama 3.1 being my base model) and uploaded it to HuggingFace Hub.
I noticed that I was not able to download this model using Transformers library, the reason being that "model.safetensors" or a ".bin" file was not found to download. This occurs because litgpt stores the model in a .pth format which is not recognized by the Transformers library.
So I thought of using litgpt to load the model directly:
from litgpt import LLM
model = LLM.load(model_name).to(device)
Issues here: "model_name" only supports models which are supported by litgpt (you'll get a list of models supported if you do litgpt download list.
a. How do I add my fine tuned model to this list?
b. If a model is not available locally, it has to go to Hugging Face Hub and download the model, which isn't happening.
I end up getting this error: OSError: Error no file named pytorch_model.bin, model.safetensors, tf_model.h5, model.ckpt.index or flax_model.msgpack found in directory /Users/nxtr-admin-it/Desktop/nextoar/nxtr-models/manufacturing-llm/.
(the whole point of me doing this was to convert my model to a .safetnsors or a .bin file so I can load it using Transformers library)
litgpt convert_from_litgpt $finetuned_dir/final/ out/hf-tinyllama/converted Not able to make this script work too..
Can someone help me out with this?
When I manually converted the model to a .bin file using a script, I was able to load the model using Transformers library, but the accuracy dropped significantly. So I think there was some issue while converting the file from .pth to .bin due to which there was a drop In the accuracy. Because the same model when loaded using litgpt gives really good results
The text was updated successfully, but these errors were encountered:
Hey, I fine-tuned a LLM using lit got (llama 3.1 being my base model) and uploaded it to HuggingFace Hub.
Issues here: "model_name" only supports models which are supported by litgpt (you'll get a list of models supported if you do
litgpt download list
.a. How do I add my fine tuned model to this list?
b. If a model is not available locally, it has to go to Hugging Face Hub and download the model, which isn't happening.
I refer to this documentation to convert the weights of the model to a different format:
https://github.com/Lightning-AI/litgpt/blob/main/tutorials/convert_lit_models.md
I end up getting this error: OSError: Error no file named pytorch_model.bin, model.safetensors, tf_model.h5, model.ckpt.index or flax_model.msgpack found in directory /Users/nxtr-admin-it/Desktop/nextoar/nxtr-models/manufacturing-llm/.
(the whole point of me doing this was to convert my model to a .safetnsors or a .bin file so I can load it using Transformers library)
litgpt convert_from_litgpt $finetuned_dir/final/ out/hf-tinyllama/converted
Not able to make this script work too..Can someone help me out with this?
The text was updated successfully, but these errors were encountered: