Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Adds support to use brevitas quantized weights for stateless_llama #179

Merged
merged 2 commits into from
Nov 17, 2023

Conversation

IanNod
Copy link
Contributor

@IanNod IanNod commented Nov 17, 2023

  • Modifies mm_group_quant to work with brevitas safetensors, needs work to generalize
  • Changes compiler to use torch as input to enable quantization of torch ir

- Modifies mm_group_quant to work with brevitas safetensors, needs work to generalize
- Changes compiler to use torch as input to enable quantization of torch ir
Copy link
Contributor

@qedawkins qedawkins left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Not exactly familiar with everything going on here, but left some thoughts/questions.

 - Adds todo clarifying skipping of _params. in mm_group_quant for matching purposes
 - removes arg use in pipeline to make external use easier
@IanNod IanNod linked an issue Nov 17, 2023 that may be closed by this pull request
Copy link
Contributor

@qedawkins qedawkins left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM, but I have not been working on this code.

Copy link
Contributor

@stellaraccident stellaraccident left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Can you run black and make sure everything is formatted? A lot of whitespace changes.

@IanNod
Copy link
Contributor Author

IanNod commented Nov 17, 2023

Can you run black and make sure everything is formatted? A lot of whitespace changes.

Yup, already ran black at least the files changed in this patch

@IanNod IanNod merged commit 39a37e3 into main Nov 17, 2023
1 check passed
@IanNod IanNod deleted the quantized_mm branch November 17, 2023 18:26
@dan-garvey
Copy link
Member

@stellaraccident are you running default black or with line-length=79? I think @IanNod and I's brains are hardcoded to the latter from SHARK formatting

@stellaraccident
Copy link
Contributor

Default black. We need to add a lint check and should do were iree is doing.

@dan-garvey
Copy link
Member

#189

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

Externalized Parameters
4 participants