Skip to content

Commit 4758982

Browse files
authored
Merge branch 'main' into layout
2 parents ec3f79f + 62745fc commit 4758982

File tree

2 files changed

+3
-2
lines changed

2 files changed

+3
-2
lines changed

torchao/_eval.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@
1010

1111
import torch
1212

13-
from .utils import _lm_eval_available, _MultiInput
13+
from quantization.utils import _lm_eval_available, _MultiInput
1414

1515
if _lm_eval_available:
1616
try: # lm_eval version 0.4

torchao/quantization/README.md

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -69,7 +69,8 @@ Note: The quantization error incurred by applying int4 quantization to your mode
6969
## A16W4 WeightOnly Quantization with GPTQ
7070

7171
```python
72-
from torchao.quantization.GPTQ import Int4WeightOnlyGPTQQuantizer, InputRecorder, TransformerEvalWrapper
72+
from torchao._eval import InputRecorder, TransformerEvalWrapper
73+
from torchao.quantization.GPTQ import Int4WeightOnlyGPTQQuantizer
7374
precision = torch.bfloat16
7475
device = "cuda"
7576
checkpoint_file_name = "../gpt-fast/checkpoints/meta-llama/Llama-2-7b-chat-hf/model.pth"

0 commit comments

Comments
 (0)