Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

They show that a decoder only transformer (which gpts are) are rnns with infinite hidden state size. Infinite hidden state size is a pretty strong thing! Sounds interesting to me.


not infinite, just scaling linearly with length




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: