Hacker News new | past | comments | ask | show | jobs | submit login

is there nothing here about attention?



The first link on the list is “the annotated transformer” which goes into a fair amount of detail on attention, it’s an annotated version of the paper that introduced scaled dot product attention


i didn’t see that :)




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: