Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

PS: I thought Ollama had a way to use RAM instead of VRAM (?) to keep the model active when not in use, but in my experience that didn't solve the problem.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: