Skip to content

Add bf16-qd8 convert operator support#10022

Open
GregoryComer wants to merge 1 commit intogoogle:masterfrom
GregoryComer:bf16-qd8-operator
Open

Add bf16-qd8 convert operator support#10022
GregoryComer wants to merge 1 commit intogoogle:masterfrom
GregoryComer:bf16-qd8-operator

Conversation

@GregoryComer
Copy link
Copy Markdown
Contributor

@GregoryComer GregoryComer commented Apr 21, 2026

Add operator-level support for dynamic (qd8) quantization from bf16 tensors.

As an aside, it looks like the convert-nc tests weren't previously built with cmake. I added it and verified they pass on ARM Mac.

Test Plan:

  • Ran cmake build and operator-level tests on ARM Mac.
  • Verified bazel build succeeds for the top-level XNNPACK target.

@GregoryComer GregoryComer marked this pull request as ready for review April 21, 2026 21:39

convert_op->compute[0].type = xnn_parallelization_type_1d_tile_1d_dynamic;
switch (expected_type) {
case xnn_operator_type_convert_nc_bf16_qd8:
Copy link
Copy Markdown
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I think it will be valuable to support qdu8 here too.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants