Hacker News
new
|
past
|
comments
|
ask
|
show
|
jobs
|
submit
login
steinvakt2
3 months ago
|
parent
|
context
|
favorite
| on:
Writing Speed-of-Light Flash Attention for 5090 in...
I had a 5090 some months ago but couldnt get flash attention to work. Does it now work natively? What about 5080?
sigmoid10
3 months ago
|
next
[–]
Pytorch now has native support for the Blackwell architecture:
https://pytorch.org/blog/pytorch-2-7/
SynasterBeiter
3 months ago
|
parent
|
next
[–]
It does, but the performance is pretty bad, worse than Hopper.
zackangelo
3 months ago
|
prev
[–]
Curious what issues you were having. The kernel should compile natively if you pass nvcc the correct arch flags, although it probably won't take advantage of any new hardware features.
saagarjha
3 months ago
|
parent
[–]
High-performance GPU code typically uses nonportable features that are not supported across generations.
Guidelines
|
FAQ
|
Lists
|
API
|
Security
|
Legal
|
Apply to YC
|
Contact
Search: