Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

If the author is reading this I'll point out that the cuda toolkit you find in the repositories is generally older. You can find the latest versions straight from Nvidia: https://developer.nvidia.com/cuda-downloads?target_os=Linux&...

The caveat is that sometimes a library might be expecting an older version of cuda.

The vram on the GPU does make a difference, so it would at some point be worth looking at another GPU or increasing your system ram if you start running into limits.

However I wouldn't worry too much right away, it's more important to get started and get an understanding of how these local LLMs operate and take advantage of the optimisations that the community is making to make it more accessible. Not everyone has a 5090, and if LLMs remain in the realms of high end hardware, it's not worth the time.




The other main caveat is that installing from custom sources using apt is a massive pain in the ass.


I tried running an LLM locally today, installed cuda toolkit, and it was missing cudann.h

I gave up.




Consider applying for YC's Fall 2025 batch! Applications are open till Aug 4

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: