Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

I knew the model would have difficulty fitting into a 16GB VRAM GPU, but "you need to load and unload parts of the model pipeline to/from the GPU" is not a workaround I expected.

At that point it's probably better to write a guide on how to set up a VM with a A100 easily instead of trying to fit it into a Colab GPU.



What about people having RTX 4090 with 24GB or even dual? Does it run on it?




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: