Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

With around 4.6 GiB model size the new Qwen3-8B quantized to 4-bit should fit comfortably in 16 GiB of memory: https://huggingface.co/mlx-community/Qwen3-8B-4bit


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: