-
Notifications
You must be signed in to change notification settings - Fork 794
Qualcomm AI Engine Direct - Enable Example/OSS models on GPU backend #16425
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/executorch/16425
Note: Links to docs will display an error until the docs builds have been completed. ❗ 1 Active SEVsThere are 1 currently active SEVs. If your PR is affected, please view them below: ✅ You can merge normally! (1 Unrelated Failure)As of commit 2869a2b with merge base 47dc1de ( UNSTABLE - The following job is marked as unstable, possibly due to flakiness on trunk:
This comment was automatically generated by Dr. CI and updates every 15 minutes. |
This PR needs a
|
|
Thank you for the PR! Can you address the lint error |
| instance = torchvision.models.convnext_small(weights="IMAGENET1K_V1").eval() | ||
| backend = get_backend_type(args.backend) | ||
| qnn_quantizer = { | ||
| QnnExecuTorchBackendType.kGpuBackend: None, |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
How do we expect user to use different backend for the quantizer part?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Actually, can you elaborate this part?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I am preparing next PR for llama on GPU, will update.
|
Looks like some changes are in #16423.. |
ea74436 to
772f9dc
Compare
cccclai
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thank you! Do you have some estimated perf number?
I am preparing next PR for llama on GPU, will update perf number for these models. |
Summary
Update model scripts for GPU backend.
Succeed
Validation failed
Finalization failed
Has accuracy issue
Test plan
cc @cccclai @cbilgin