-
Notifications
You must be signed in to change notification settings - Fork 246
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Missing Transformers initializer for Falcon models #1988
Comments
Hi @martin-gorner - Thanks for reporting the issue. You can intialize
And for loading falcon model family( Attached gist here for the reference. |
This issue is stale because it has been open for 14 days with no activity. It will be closed if no further activity occurs. Thank you. |
Thanks @mehtamansi29 but this issue is filed in keras-hub. The problem is to initialize a Keras-hub model from the safetensor checkpoint, as is possible for Llama, Gemma etc. I'm logging this because cross-compatibility between Keras-hub and Transformers checkpoints is not guaranteed even if the model architecture exists on both sides. A checkpoint translation module is necessary. It is important to keep track of model architectures where this translation module was implemented and where it was not. |
Repro code:
model5 = keras_hub.models.CausalLM.from_preset("hf://tiiuae/falcon-7b-instruct", dtype="bfloat16")
Result:
ValueError: KerasHub has no converter for huggingface/transformers models with model type 'falcon'
Now that the Falcon model family exists in Keras-hub, this should work.
The text was updated successfully, but these errors were encountered: