You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Turning has 22 and 24gb cards available as well as being used in google colab. Flash attention skipped this generation causing issues for the cards. Is it possible to add support?
The text was updated successfully, but these errors were encountered:
Hi, we think Turning is something other than a popular architecture. Could you please first try the Triton-only branch and provide some feedback? Thank you.
Turning has more affordable 48gb cards too, like RTX 8000. A lot of projects exclude it and require ampere+, hence the popularity is less. I will give triton only a shot. Is the difference that no kernel is used?
Turning has 22 and 24gb cards available as well as being used in google colab. Flash attention skipped this generation causing issues for the cards. Is it possible to add support?
The text was updated successfully, but these errors were encountered: