I very much agree; I think laziness / friction is basically a critically important regularizer for what to build and for what to not build. LLMs remove that friction and it requires more discipline now. (Wrote some of this up a while ago here: https://matthiasplappert.com/blog/2026/laziness-in-the-age-o...)
That seems like quite an extrapolation and an extraordinary statement. This is a single task, in a lab setting. What your describing are extremely open-ended tasks in people’s homes.
Look at recent developments/announcements involving novel increasingly generalizable learning capabilities from projects like 1X/Neo, Figure 03, Skild AI. Also see open published work like MimicDroid, HDMI, GenMimic, Humanoid-Union Dataset, RoboMirror, Being-H0
You're not really trying to see the advances in things like the data flywheel. If you were you would see that those demos represent real movement towards generality.
reply