Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Deepseek pioneered automatic prefix caching and caches on SSD. SSD reads are so fast compared to LLM inference that I can't think of a reason to waste ram on it.


It’s not instantly fast though. Context is probably ~20gb of VRAM at max context size. That’s gonna take some time to get from SSD no matter what.

TtFT will get slower if you export kv cache to SSD.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: