[RVV] add rvv f32 kernels for velu, vgelu, vapproxgelu, ibilinear, ppmm, qc8w-gemm#9954
[RVV] add rvv f32 kernels for velu, vgelu, vapproxgelu, ibilinear, ppmm, qc8w-gemm#9954velonica0 wants to merge 2 commits intogoogle:masterfrom
Conversation
…ar, f32-qc8w-gemm and f32-ppmm
|
Thanks for your pull request! It looks like this may be your first contribution to a Google open source project. Before we can look at your pull request, you'll need to sign a Contributor License Agreement (CLA). View this failed invocation of the CLA check for more information. For the most up to date status, view the checks section at the bottom of the pull request. |
|
Hi, @dsharlet |
dsharlet
left a comment
There was a problem hiding this comment.
Thanks for the PR!
This is a pretty big PR with a wide variety of ops in it. I think this should be split into a few smaller PRs:
- ppmm kernel
- f32-qc8w kernel
- elementwise ops
Regarding the f32-qc8w kernel, what use case motivated implementing that kernel? It is not something we use much currently, and the operator code for that type of gemm has some issues we need to fix.
|
Thank you very much for your review.
The separated PR are #9962 #9963 #9964
Sorry, I see qd8-f32-qc8w-gemm, so I deleted f32-qc8w-gemm. |
Add rvv f32 kernels for velu, vgelu, vapproxgelu, ibilinear, ppmm, qc8w-gemm.
Tested on SpacemiT K1 and K3 CPU(both VLEN=256).
Next, I will continue with the RVV optimization of the FP16 operator.