Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

fp8 support #2304

Open
vince62s opened this issue Feb 2, 2023 · 6 comments
Open

fp8 support #2304

vince62s opened this issue Feb 2, 2023 · 6 comments

Comments

@vince62s
Copy link
Member

vince62s commented Feb 2, 2023

If someone is motivated, there could be some adaptation to support fp8 (on some hardware) using this new library:

https://github.com/NVIDIA/TransformerEngine

cc: @guillaumekln @francoishernandez

@vince62s
Copy link
Member Author

vince62s commented Feb 2, 2023

Well, no hurry for RTX 4090, not ready yet.

Hi All,

First of all, I'm really sorry for the prolonged silence on this issue - I did not want to communicate anything before getting a full alignment internally.
As noted in the RTX 4090 announcement and Ada whitepaper, Ada has FP8 TensorCore hardware. However, the software support for them is not currently available - e.g. there is no support for it exposed in cuBLASLt currently. The reason for it is that both the FP8 TC instruction as well as other features used in the fast FP8 GEMM kernels are different between Hopper and Ada (meaning a different set of kernels required for both architectures) and the Hopper support was prioritized. Once the FP8 support lands in CUDA and its libraries (tentatively scheduled for CUDA 12.1 in Q2), Transformer Engine will also fully support Ada.

read from here: NVIDIA/TransformerEngine#15

@oscarbg
Copy link

oscarbg commented Mar 1, 2023

@vince62s cuda 12.1 released.. can Ada support be worked on?

@AaronZLT
Copy link

AaronZLT commented Mar 7, 2023

@oscarbg didn't work.
cublasLtMatmul and cublasLtMatrixTransform still can't work for __nv_fp8_e4m3 and __nv_fp8_e4m3 on 4090 with newest cuda 12.1, date-2023.3.7.
Can anyone work it out or it is just cuda not support issue?

@sbhavani
Copy link

@vince62s
Copy link
Member Author

Great I'll give it a try when I get some time.

@vince62s
Copy link
Member Author

I tried it but obviously it is not so easy to make it work in our scenario.
NVIDIA/TransformerEngine#230

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

4 participants