-
Notifications
You must be signed in to change notification settings - Fork 71
Description
使用配置量化awq_w8a8.yml量化qwen3-8b时报了这样的错
[rank0]: Traceback (most recent call last):
[rank0]: File "./code/llm_quantization/LightCompress//llmc/main.py", line 261, in
[rank0]: main(config)
[rank0]: File "./code/llm_quantization/LightCompress//llmc/main.py", line 58, in main
[rank0]: model.collect_first_block_input(calib_data, padding_mask)
[rank0]: File "/opt/miniconda/envs/light-compress/lib/python3.11/site-packages/torch/utils/_contextlib.py", line 124, in decorate_context
[rank0]: return func(*args, **kwargs)
[rank0]: ^^^^^^^^^^^^^^^^^^^^^
[rank0]: File "./code/llm_quantization/LightCompress/llmc/models/base_model.py", line 305, in collect_first_block_input
[rank0]: self.model(**data)
[rank0]: File "/opt/miniconda/envs/light-compress/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1776, in _wrapped_call_impl
[rank0]: return self._call_impl(*args, **kwargs)
[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[rank0]: File "/opt/miniconda/envs/light-compress/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1787, in _call_impl
[rank0]: return forward_call(*args, **kwargs)
[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[rank0]: File "/opt/miniconda/envs/light-compress/lib/python3.11/site-packages/transformers/utils/generic.py", line 918, in wrapper
[rank0]: output = func(self, *args, **kwargs)
[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^
[rank0]: File "/opt/miniconda/envs/light-compress/lib/python3.11/site-packages/transformers/models/qwen3/modeling_qwen3.py", line 480, in forward
[rank0]: outputs: BaseModelOutputWithPast = self.model(
[rank0]: ^^^^^^^^^^^
[rank0]: File "/opt/miniconda/envs/light-compress/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1776, in _wrapped_call_impl
[rank0]: return self._call_impl(*args, **kwargs)
[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[rank0]: File "/opt/miniconda/envs/light-compress/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1787, in _call_impl
[rank0]: return forward_call(*args, **kwargs)
[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[rank0]: File "/opt/miniconda/envs/light-compress/lib/python3.11/site-packages/transformers/utils/generic.py", line 1072, in wrapper
[rank0]: outputs = func(self, *args, **kwargs)
[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^
[rank0]: File "/opt/miniconda/envs/light-compress/lib/python3.11/site-packages/transformers/models/qwen3/modeling_qwen3.py", line 412, in forward
[rank0]: attention_mask=causal_mask_mapping[decoder_layer.attention_type],
[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[rank0]: File "/opt/miniconda/envs/light-compress/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1965, in getattr
[rank0]: raise AttributeError(
[rank0]: AttributeError: 'Catcher' object has no attribute 'attention_type'