Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Support self-trained model #208

Open
2 tasks
yangelaboy opened this issue Jan 25, 2024 · 2 comments
Open
2 tasks

Support self-trained model #208

yangelaboy opened this issue Jan 25, 2024 · 2 comments
Labels
question Further information is requested

Comments

@yangelaboy
Copy link

Model description

hi, my company has trained a model of 7b, we want to deploy lorax with our model.
Can you introduce key steps to support model in loraX?

Open source status

  • The model implementation is available
  • The model weights are available

Provide useful links for the implementation

No response

@tgaddair tgaddair added the question Further information is requested label Jan 26, 2024
@tgaddair
Copy link
Contributor

Hey @yangelaboy, thanks for offering to contribute your model to LoRAX!

For adding a new model architecture to LoRAX, the best examples are probably our PRs that added Phi, Qwen, and GPT2:

The main things to add are:

  • Basic model implementation in flash_<model>_modeling.py
  • Initialization and LoRA loading rules in flash_<model>.py
  • Register model for loading in models/__init__.py

Please let m know if you have any additional questions I can help answer!

@yangelaboy
Copy link
Author

@tgaddair thx for your replying, I will get a try!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
question Further information is requested
Projects
None yet
Development

No branches or pull requests

2 participants