Skip to content

Remove usage of wrap_in_hpu_graph in PT eager#47

Merged
kzawora-intel merged 1 commit intohabana_mainfrom
private/kzawora/hpugraph_eager_fix
Jun 5, 2024
Merged

Remove usage of wrap_in_hpu_graph in PT eager#47
kzawora-intel merged 1 commit intohabana_mainfrom
private/kzawora/hpugraph_eager_fix

Conversation

@kzawora-intel
Copy link
Copy Markdown

As in title, there is no habana_frameworks.torch.hpu.wrap_in_hpu_graph in PT eager backend and attempting to use it will result in AttributeError.

Copy link
Copy Markdown

@madamczyk-intel madamczyk-intel left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM!

@kzawora-intel kzawora-intel merged commit 1c5d12e into habana_main Jun 5, 2024
@kzawora-intel kzawora-intel added the habana Issues or PRs submitted by Habana Labs label Sep 20, 2024
@kzawora-intel kzawora-intel deleted the private/kzawora/hpugraph_eager_fix branch October 7, 2024 13:02
michalkuligowski added a commit that referenced this pull request Jan 15, 2025
remove expert_max hard code (#47)
vLLM-Ext: Full enabling of ALiBi (#34)
Add version inference via setuptools-scm (#58)
Revert "vLLM-Ext: Full enabling of ALiBi (#34)" (#59)
Remove punica_hpu.py from vllm_hpu_extension (#66)
Removed previous (not-pipelined) pa implementation (#72)
Add flag to enable running softmax in fp32 (#71)
Update calibration readme link (#73)
allow lm_head quantization in calibration process (#65)
Pad to bmin if value is less (#67)
Update pyproject.toml (#75)

---------

Co-authored-by: Michał Kuligowski <mkuligowski@habana.ai>
mfylcek added a commit that referenced this pull request Jan 21, 2025
remove expert_max hard code (#47)
vLLM-Ext: Full enabling of ALiBi (#34)
Add version inference via setuptools-scm (#58)
Revert "vLLM-Ext: Full enabling of ALiBi (#34)" (#59)
Remove punica_hpu.py from vllm_hpu_extension (#66)
Removed previous (not-pipelined) pa implementation (#72)
Add flag to enable running softmax in fp32 (#71)
Update calibration readme link (#73)
allow lm_head quantization in calibration process (#65)
Pad to bmin if value is less (#67)
Update pyproject.toml (#75)

---------

Co-authored-by: Michał Kuligowski <mkuligowski@habana.ai>
yiliu30 pushed a commit that referenced this pull request Aug 8, 2025
Signed-off-by: Konrad Zawora <kzawora@habana.ai>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

habana Issues or PRs submitted by Habana Labs

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants