Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

I think the fact those diffusion models are smaller and compute efficient than gigantic GPT models are in fact make them easier to use and distribute.

BLOOM is out there, but not that many individuals with have like 8 3090 to host them, and the inference is still incredibly slow nevertheless



BLOOM also doesn’t have GPT-3’s RLHF tuning, so anyone who tries to ask it questions or give it instructions in the manner GPT-3 supports will be disappointed. You have to k-shot prompt it or fine-tune it yourself for it to be useful.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: