Add fp16 qb4w scalar kernels
Summary: Add qb4w scalar 1x2, 1x4, 1x8, 2x2, 2x4, 2x8, and 4x4 kernels to XNNPACK. Add test coverage in ExecuTorch op-level linear test coverage for 4-bit blockwise weights / fp16.
Differential Revision: D57335871
:link: Helpful Links
:test_tube: See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/executorch/3626
- :page_facing_up: Preview Python docs built from this PR
Note: Links to docs will display an error until the docs builds have been completed.
:white_check_mark: No Failures
As of commit 055cde2fabd57e021b0fc2b229f15129d25fa75d with merge base 47a29a13f4f2574b6787be03249633d5531b795a ():
:green_heart: Looks good so far! There are no failures yet. :green_heart:
This comment was automatically generated by Dr. CI and updates every 15 minutes.
This pull request was exported from Phabricator. Differential Revision: D57335871
This pull request was exported from Phabricator. Differential Revision: D57335871
This pull request was exported from Phabricator. Differential Revision: D57335871
Thanks!
This pull request was exported from Phabricator. Differential Revision: D57335871
This pull request has been merged in pytorch/executorch@4b7c6db5d284d7bcca9b5a6db45758517fa1031d.