Check out Voxon [1]. From the specs and youtube videos it seems like it's working on the same principle (rotating LED screen). Fun fact, it was co-founded by none other than Ken Silverman (the creator of Build engine) [2]. They've been pushing commercialization of this technology for years now.
The "ALL CAPS" part of your comment got me thinking. I imagine most llms understand subtle meanings of upper case text use depending on context. But, as I understand it, ALL CAPS text will tokenize differently than lower case text. Is that right? In that case, won't the upper case be harder to understand and follow for most models since it's less common in datasets?
There's more than enough ALL CAPS text in the corpus of the entire internet, and enough semantic context associated with it for it to be intended to be in the imperative voice.
And now we’ve built LLMs - the biggest mirror of them all.
The Internet and smartphones gave us countless ways to look at ourselves and see how others see us. And then LLMs helped us gaze at the sum of all that and even confront reflections of our own thoughts.
Yeah, TFA ended just before it got to the really interesting part of how self-reflection itself is fundamental to the development of concisousness. Mirror-like technologies don't just show us our own appearance. They help us understand how we relate to the world around us.
It reminds me of Kieślowski's movie Camera Buff (1979), where the main character in iconic scene points the camera at himself and realizes that the act of making movies reflects not only his subjects, but also on who he is in relation them.
At first glance, this reminds me of how branch prediction is utilized in CPUs to speedup execution. As I understand it, this development is like a form of soft branch prediction over language trajectories: a small model predicts what the main model will do, takes few steps ahead and then verifies the results (and this can be done in parallel). If it checks out, you just jump forward, it not you take miss but its rare. I find it funny how small-big ideas like this come up in different context again and again in history of our technological development. Of course ideas as always are cheap. The hard part is how to actually use them and cash in on them.
A lot of optimizations in LLMs now are low hanging fruits inspired by techniques in classical computer science. Another one that comes to mind is paged KV caching which is based on memory paging.
[1] https://www.voxon.co/ [2] https://en.wikipedia.org/wiki/Ken_Silverman