Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Workaround for OOM during loading llama-405 #396

Merged
merged 2 commits into from
Oct 18, 2024

Conversation

afierka-intel
Copy link

Repeating missing code

@afierka-intel afierka-intel self-assigned this Oct 16, 2024
@afierka-intel afierka-intel added bug Something isn't working habana Issues or PRs submitted by Habana Labs labels Oct 16, 2024
Copy link

@michalkuligowski michalkuligowski left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@afierka-intel afierka-intel merged commit 07c98a5 into habana_main Oct 18, 2024
19 checks passed
@afierka-intel afierka-intel deleted the dev/afierka/fix-llama-405-oom branch October 18, 2024 07:13
xuechendi pushed a commit to xuechendi/vllm-fork that referenced this pull request Oct 23, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working habana Issues or PRs submitted by Habana Labs
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants