Problem Description
~
Reproduction Steps
/models/Qwen3-0.6B
--format
"fake"
--bits
2
--group_size
16
--nsamples
1
--tasks
piqa
--iters
0
--disable_opt_rtn
--device_map
"auto"
--act_bits
8
Environment Information
No response
Error Logs
Additional Context
Running loglikelihood requests: 0%| | 0/3676 [00:00<?, ?it/s]Traceback (most recent call last):
File "/home/wenhuach/auto-round/auto_round/main.py", line 799, in
run()
~~~^^
File "/home/wenhuach/auto-round/auto_round/main.py", line 783, in run
start()
~~~~~^^
File "/home/wenhuach/auto-round/auto_round/main.py", line 516, in start
tune(args)
~~~~^^^^^^
File "/home/wenhuach/auto-round/auto_round/main.py", line 729, in tune
run_model_evaluation(model, tokenizer, autoround, folders, formats, device_str, args)
~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/wenhuach/auto-round/auto_round/eval/evaluation.py", line 433, in run_model_evaluation
evaluate_with_model_instance(model, tokenizer, device_str, args)
~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/wenhuach/auto-round/auto_round/eval/evaluation.py", line 277, in evaluate_with_model_instance
res = simple_evaluate_user_model(
model,
...<6 lines>...
add_bos_token=args.add_bos_token,
)
File "/home/wenhuach/auto-round/auto_round/eval/evaluation.py", line 60, in simple_evaluate_user_model
return lm_eval.simple_evaluate(
~~~~~~~~~~~~~~~~~~~~~~~^
model=hflm, model_args=None, batch_size=batch_size, max_batch_size=max_batch_size, limit=limit, **kwargs
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
)
^
File "/home/wenhuach/miniforge3/envs/autoround/lib/python3.13/site-packages/lm_eval/utils.py", line 458, in _wrapper
return fn(*args, **kwargs)
File "/home/wenhuach/miniforge3/envs/autoround/lib/python3.13/site-packages/lm_eval/evaluator.py", line 357, in simple_evaluate
results = evaluate(
lm=lm,
...<12 lines>...
confirm_run_unsafe_code=confirm_run_unsafe_code,
)
File "/home/wenhuach/miniforge3/envs/autoround/lib/python3.13/site-packages/lm_eval/utils.py", line 458, in _wrapper
return fn(*args, **kwargs)
File "/home/wenhuach/miniforge3/envs/autoround/lib/python3.13/site-packages/lm_eval/evaluator.py", line 585, in evaluate
resps = getattr(lm, reqtype)(cloned_reqs)
File "/home/wenhuach/miniforge3/envs/autoround/lib/python3.13/site-packages/lm_eval/api/model.py", line 453, in loglikelihood
return self._loglikelihood_tokens(new_reqs, disable_tqdm=disable_tqdm)
~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/wenhuach/miniforge3/envs/autoround/lib/python3.13/site-packages/lm_eval/models/huggingface.py", line 1296, in _loglikelihood_tokens
self._model_call(batched_inps, **call_kwargs),
~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/wenhuach/miniforge3/envs/autoround/lib/python3.13/site-packages/lm_eval/models/huggingface.py", line 965, in _model_call
return self.model(inps).logits
~~~~~~~~~~^^^^^^
File "/home/wenhuach/miniforge3/envs/autoround/lib/python3.13/site-packages/torch/nn/modules/module.py", line 1775, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^
File "/home/wenhuach/miniforge3/envs/autoround/lib/python3.13/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl
return forward_call(*args, **kwargs)
File "/home/wenhuach/miniforge3/envs/autoround/lib/python3.13/site-packages/transformers/utils/generic.py", line 918, in wrapper
output = func(self, *args, **kwargs)
File "/home/wenhuach/miniforge3/envs/autoround/lib/python3.13/site-packages/transformers/models/qwen3/modeling_qwen3.py", line 494, in forward
logits = self.lm_head(hidden_states[:, slice_indices, :])
File "/home/wenhuach/miniforge3/envs/autoround/lib/python3.13/site-packages/torch/nn/modules/module.py", line 1775, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^
File "/home/wenhuach/miniforge3/envs/autoround/lib/python3.13/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl
return forward_call(*args, **kwargs)
File "/home/wenhuach/miniforge3/envs/autoround/lib/python3.13/site-packages/torch/nn/modules/linear.py", line 134, in forward
return F.linear(input, self.weight, self.bias)
~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
RuntimeError: Inference tensors do not track version counter.
Problem Description
~
Reproduction Steps
Environment Information
No response
Error Logs
Additional Context
Running loglikelihood requests: 0%| | 0/3676 [00:00<?, ?it/s]Traceback (most recent call last):
File "/home/wenhuach/auto-round/auto_round/main.py", line 799, in
run()
~~~^^
File "/home/wenhuach/auto-round/auto_round/main.py", line 783, in run
start()
~~~~~^^
File "/home/wenhuach/auto-round/auto_round/main.py", line 516, in start
tune(args)
~~~~^^^^^^
File "/home/wenhuach/auto-round/auto_round/main.py", line 729, in tune
run_model_evaluation(model, tokenizer, autoround, folders, formats, device_str, args)
~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/wenhuach/auto-round/auto_round/eval/evaluation.py", line 433, in run_model_evaluation
evaluate_with_model_instance(model, tokenizer, device_str, args)
~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/wenhuach/auto-round/auto_round/eval/evaluation.py", line 277, in evaluate_with_model_instance
res = simple_evaluate_user_model(
model,
...<6 lines>...
add_bos_token=args.add_bos_token,
)
File "/home/wenhuach/auto-round/auto_round/eval/evaluation.py", line 60, in simple_evaluate_user_model
return lm_eval.simple_evaluate(
~~~~~~~~~~~~~~~~~~~~~~~^
model=hflm, model_args=None, batch_size=batch_size, max_batch_size=max_batch_size, limit=limit, **kwargs
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
)
^
File "/home/wenhuach/miniforge3/envs/autoround/lib/python3.13/site-packages/lm_eval/utils.py", line 458, in _wrapper
return fn(*args, **kwargs)
File "/home/wenhuach/miniforge3/envs/autoround/lib/python3.13/site-packages/lm_eval/evaluator.py", line 357, in simple_evaluate
results = evaluate(
lm=lm,
...<12 lines>...
confirm_run_unsafe_code=confirm_run_unsafe_code,
)
File "/home/wenhuach/miniforge3/envs/autoround/lib/python3.13/site-packages/lm_eval/utils.py", line 458, in _wrapper
return fn(*args, **kwargs)
File "/home/wenhuach/miniforge3/envs/autoround/lib/python3.13/site-packages/lm_eval/evaluator.py", line 585, in evaluate
resps = getattr(lm, reqtype)(cloned_reqs)
File "/home/wenhuach/miniforge3/envs/autoround/lib/python3.13/site-packages/lm_eval/api/model.py", line 453, in loglikelihood
return self._loglikelihood_tokens(new_reqs, disable_tqdm=disable_tqdm)
~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/wenhuach/miniforge3/envs/autoround/lib/python3.13/site-packages/lm_eval/models/huggingface.py", line 1296, in _loglikelihood_tokens
self._model_call(batched_inps, **call_kwargs),
~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/wenhuach/miniforge3/envs/autoround/lib/python3.13/site-packages/lm_eval/models/huggingface.py", line 965, in _model_call
return self.model(inps).logits
~~~~~~~~~~^^^^^^
File "/home/wenhuach/miniforge3/envs/autoround/lib/python3.13/site-packages/torch/nn/modules/module.py", line 1775, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^
File "/home/wenhuach/miniforge3/envs/autoround/lib/python3.13/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl
return forward_call(*args, **kwargs)
File "/home/wenhuach/miniforge3/envs/autoround/lib/python3.13/site-packages/transformers/utils/generic.py", line 918, in wrapper
output = func(self, *args, **kwargs)
File "/home/wenhuach/miniforge3/envs/autoround/lib/python3.13/site-packages/transformers/models/qwen3/modeling_qwen3.py", line 494, in forward
logits = self.lm_head(hidden_states[:, slice_indices, :])
File "/home/wenhuach/miniforge3/envs/autoround/lib/python3.13/site-packages/torch/nn/modules/module.py", line 1775, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^
File "/home/wenhuach/miniforge3/envs/autoround/lib/python3.13/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl
return forward_call(*args, **kwargs)
File "/home/wenhuach/miniforge3/envs/autoround/lib/python3.13/site-packages/torch/nn/modules/linear.py", line 134, in forward
return F.linear(input, self.weight, self.bias)
~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
RuntimeError: Inference tensors do not track version counter.