Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Please support AVX512_FP16 #2822

Open
DaiShaoJie77 opened this issue Mar 5, 2025 · 1 comment
Open

Please support AVX512_FP16 #2822

DaiShaoJie77 opened this issue Mar 5, 2025 · 1 comment
Assignees
Labels
enhancement A feature or an optimization request help wanted

Comments

@DaiShaoJie77
Copy link

Chips supporting AVX512_FP16 have been released for more than a year. Why does Intel's open source GPU computing sub-computing library still not support AVX512_FP16? AVX512_FP16 is the instruction I expect to use

@DaiShaoJie77 DaiShaoJie77 added the enhancement A feature or an optimization request label Mar 5, 2025
@vpirogov
Copy link
Member

vpirogov commented Mar 6, 2025

oneDNN uses instructions from AVX512_FP16 ISA extension on processors with Intel AVX 10.1/512 instruction set support (4th and 5th generation Intel Xeon Scalable Processors and Intel Xeon 6 processors).

Default numerical behavior for oneDNN functions requires fp32 accumulation, which is not supported by FMA instructions in AVX512_FP16 extension. This implementation can be added in relaxed accumulation mode, but it's not a priority for the core engineering team at the moment.

@vpirogov vpirogov self-assigned this Mar 6, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement A feature or an optimization request help wanted
Projects
None yet
Development

No branches or pull requests

2 participants