▲ 0 r/ollama
I know the answer is most likely no, but I've recently upgraded my (linux) pc, replacing my rtx 3060 with an rx 9070xt, its running models so much faster (thanks to the vram in part, 16 gigs feels great) but I've also realized that its still not perfectly ideal for hosting models for use with multiple users, especially at once. Because I can have upwards of 6 users prompting models at the same time, I'm wondering if I still use my rtx 3060 as backup if things start slowing down. would this feature a think on ollama or any other frameworks for running llms?
u/Kryptos-the-nitwit — 10 days ago