Skip to content

[Bug]: eval bug with tied lm-head and act_bits<=8 #1400

@wenhuach21

Description

@wenhuach21

Problem Description

~

Reproduction Steps

/models/Qwen3-0.6B
--format
"fake"
--bits
2
--group_size
16
--nsamples
1
--tasks
piqa
--iters
0
--disable_opt_rtn
--device_map
"auto"
--act_bits
8

Environment Information

No response

Error Logs

Additional Context

Running loglikelihood requests: 0%| | 0/3676 [00:00<?, ?it/s]Traceback (most recent call last):
File "/home/wenhuach/auto-round/auto_round/main.py", line 799, in
run()
~~~^^
File "/home/wenhuach/auto-round/auto_round/main.py", line 783, in run
start()
~~~~~^^
File "/home/wenhuach/auto-round/auto_round/main.py", line 516, in start
tune(args)
~~~~^^^^^^
File "/home/wenhuach/auto-round/auto_round/main.py", line 729, in tune
run_model_evaluation(model, tokenizer, autoround, folders, formats, device_str, args)
~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/wenhuach/auto-round/auto_round/eval/evaluation.py", line 433, in run_model_evaluation
evaluate_with_model_instance(model, tokenizer, device_str, args)
~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/wenhuach/auto-round/auto_round/eval/evaluation.py", line 277, in evaluate_with_model_instance
res = simple_evaluate_user_model(
model,
...<6 lines>...
add_bos_token=args.add_bos_token,
)
File "/home/wenhuach/auto-round/auto_round/eval/evaluation.py", line 60, in simple_evaluate_user_model
return lm_eval.simple_evaluate(
~~~~~~~~~~~~~~~~~~~~~~~^
model=hflm, model_args=None, batch_size=batch_size, max_batch_size=max_batch_size, limit=limit, **kwargs
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
)
^
File "/home/wenhuach/miniforge3/envs/autoround/lib/python3.13/site-packages/lm_eval/utils.py", line 458, in _wrapper
return fn(*args, **kwargs)
File "/home/wenhuach/miniforge3/envs/autoround/lib/python3.13/site-packages/lm_eval/evaluator.py", line 357, in simple_evaluate
results = evaluate(
lm=lm,
...<12 lines>...
confirm_run_unsafe_code=confirm_run_unsafe_code,
)
File "/home/wenhuach/miniforge3/envs/autoround/lib/python3.13/site-packages/lm_eval/utils.py", line 458, in _wrapper
return fn(*args, **kwargs)
File "/home/wenhuach/miniforge3/envs/autoround/lib/python3.13/site-packages/lm_eval/evaluator.py", line 585, in evaluate
resps = getattr(lm, reqtype)(cloned_reqs)
File "/home/wenhuach/miniforge3/envs/autoround/lib/python3.13/site-packages/lm_eval/api/model.py", line 453, in loglikelihood
return self._loglikelihood_tokens(new_reqs, disable_tqdm=disable_tqdm)
~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/wenhuach/miniforge3/envs/autoround/lib/python3.13/site-packages/lm_eval/models/huggingface.py", line 1296, in _loglikelihood_tokens
self._model_call(batched_inps, **call_kwargs),
~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/wenhuach/miniforge3/envs/autoround/lib/python3.13/site-packages/lm_eval/models/huggingface.py", line 965, in _model_call
return self.model(inps).logits
~~~~~~~~~~^^^^^^
File "/home/wenhuach/miniforge3/envs/autoround/lib/python3.13/site-packages/torch/nn/modules/module.py", line 1775, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^
File "/home/wenhuach/miniforge3/envs/autoround/lib/python3.13/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl
return forward_call(*args, **kwargs)
File "/home/wenhuach/miniforge3/envs/autoround/lib/python3.13/site-packages/transformers/utils/generic.py", line 918, in wrapper
output = func(self, *args, **kwargs)
File "/home/wenhuach/miniforge3/envs/autoround/lib/python3.13/site-packages/transformers/models/qwen3/modeling_qwen3.py", line 494, in forward
logits = self.lm_head(hidden_states[:, slice_indices, :])
File "/home/wenhuach/miniforge3/envs/autoround/lib/python3.13/site-packages/torch/nn/modules/module.py", line 1775, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^
File "/home/wenhuach/miniforge3/envs/autoround/lib/python3.13/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl
return forward_call(*args, **kwargs)
File "/home/wenhuach/miniforge3/envs/autoround/lib/python3.13/site-packages/torch/nn/modules/linear.py", line 134, in forward
return F.linear(input, self.weight, self.bias)
~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
RuntimeError: Inference tensors do not track version counter.

Metadata

Metadata

Assignees

Labels

bugSomething isn't working

Type

No type

Projects

No projects

Milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions