-
Notifications
You must be signed in to change notification settings - Fork 9.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Please support the also official Falcon-rw-1b and Falcon-rw-7b model variants #2868
Comments
Should be easy to support - PRs welcome |
You could try changing: if hparams["architectures"][0] != "RWForCausalLM":
print("Model architecture not supported: " + hparams["architectures"][0])
sys.exit(1) to if hparams["architectures"][0] not in ("RWForCausalLM", "FalconForCausalLM"):
print("Model architecture not supported: " + hparams["architectures"][0])
sys.exit(1) This is assuming there are no other changes in the actual model architecture, etc. |
When doing so, I'm getting the following error with the rw-1b model:
With the rw-7b model it's almost the same:
The original 7b Model has 32 layers, |
I guess there are actual differences. I looked more closely at the This is very, very unlikely to work but if you want you can try changing |
The config.json in the repository is quite different between the regular falcon 7b / 40b and the RefinedWeb Veriants. Here is config.json of rw-1b
and here is config.json of the regular 7b model:
|
Even fixing these key errors with the differences in the config, I later run into this issue:
so seems like the shape of the weights also differs for this model see draft pr: #2887 |
Falcon consists not only of the versions 7b and 40b, but also of the two
refined Web
variants Falcon-RW-1B and Falcon-RW-7B.These are official Versions as can be seen on https://huggingface.co/tiiuae.
I have successfully converted and quantized the 7b models with
convert-falcon-hf-to-gguf.py
, but therefined web
variants result in the following abort messages:The message for the rw 7b model is identical except for the filename.
Do you want to support these models as well, or are there special difficulties?
A Falcon 1.3b model would be an incredible fast model for small and easy tasks. It would be great to have this model.
The text was updated successfully, but these errors were encountered: