Skip to content

Quantization layer filter#1091

Merged
dxqb merged 20 commits intoNerogar:mergefrom
dxqb:quant_layer_filter
Nov 27, 2025
Merged

Quantization layer filter#1091
dxqb merged 20 commits intoNerogar:mergefrom
dxqb:quant_layer_filter

Conversation

@dxqb
Copy link
Collaborator

@dxqb dxqb commented Nov 2, 2025

float 8:
full 12.5 GB vram
blocks 12.8 GB
attn-mlp > 16GB, offloading required

nf4:
full 9.4
blocks 9.8
attn-mlp 14.5

Unfortunately, many changes across all models required only to pass on the configuration to the relevant code. The model loader architecture should be generalized, but I've decided against doing that in this PR.

  • add "blocks" preset for all models, because that can probably recommended for most models. GGUFs seem to use this filter

@dxqb dxqb requested a review from O-J1 November 2, 2025 14:05
@dxqb
Copy link
Collaborator Author

dxqb commented Nov 2, 2025

Commit of only the UI changes:
eaf4fe2

@dxqb dxqb marked this pull request as ready for review November 4, 2025 18:27
@dxqb
Copy link
Collaborator Author

dxqb commented Nov 11, 2025

  • what type to use for layers that don't match the filter and are therefore not quantized? for most models, this ends up correctly by accident, because recent models are published and loaded in bf16. But it can go wrong, for example for SDXL checkpoints

Train dtype

@dxqb
Copy link
Collaborator Author

dxqb commented Nov 22, 2025

merged with #1139: 606b7a8
@O-J1

@O-J1
Copy link
Collaborator

O-J1 commented Nov 22, 2025

This was just merging master in right

@dxqb
Copy link
Collaborator Author

dxqb commented Nov 22, 2025

This was just merging master in right

No. This PR moves and refactors all layer filter code from TrainingTab to components.py.

I merged it separately for easy review. Here is #1139 merge: 606b7a8
Here is upstream merge: 8ce4604

@O-J1
Copy link
Collaborator

O-J1 commented Nov 22, 2025

Let me reclarify then:

  • 1139 was merged into master last week already no?
  • So isnt this just merging master into your PR?

@dxqb
Copy link
Collaborator Author

dxqb commented Nov 22, 2025

it is technically a merge, but resolving conflicts and moving your code around. that's why I pinged you.

@O-J1
Copy link
Collaborator

O-J1 commented Nov 22, 2025

I see, alright

@dxqb dxqb added the merging last steps before merge label Nov 27, 2025
@dxqb dxqb changed the base branch from master to merge November 27, 2025 20:33
@dxqb dxqb merged commit 3835ac0 into Nerogar:merge Nov 27, 2025
1 check passed
dxqb added a commit that referenced this pull request Nov 28, 2025
@dxqb dxqb deleted the quant_layer_filter branch November 28, 2025 21:12
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

merging last steps before merge

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants