I didn’t say they don’t work, I said there is an upper bound on the function they provide.
If a discrete system can be composed of multiple LLMs the upper bound on the function they provide is by the function of the LLM, not the number of agents.
Ie. We have agentic systems.
Saying “wait till you see those agentic systems!” is like saying “wait til you see those c++ programs!”
Yes. I see them. Mmm. Ok. I don’t think I’m going to be surprised by seeing them doing exactly the same things in a year.
The impressive part in a year will the non agentic part of things.
Ie. Explicitly; if the underlying LLMs dont get any better, there is no reason to expect the system built out of them to get any better.
If that was untrue, you would expect to be able to build agentic systems out of much smaller LLMs, but that overwhelmingly doesn’t work.
> if the underlying LLMs dont get any better, there is no reason to expect the system built out of them to get any better.
Actually o1, o3 are doing exactly this, and very well. I.e. explicitly: by proper orchestration the same LLM can do much better job. There is a price, but...
> you would expect to be able to build agentic systems out of much smaller LLMs
Good point, it should be possible to do it on a high-end pc or even embedded.
They are not mutually exclusive. Likely we'll get more clear separation of architecture and underlying technology. In this case agents (i.e. architecture) can use different technologies or mix of them. Including 'AI' and algorithms. The trick is to make them work together.
If a discrete system can be composed of multiple LLMs the upper bound on the function they provide is by the function of the LLM, not the number of agents.
Ie. We have agentic systems.
Saying “wait till you see those agentic systems!” is like saying “wait til you see those c++ programs!”
Yes. I see them. Mmm. Ok. I don’t think I’m going to be surprised by seeing them doing exactly the same things in a year.
The impressive part in a year will the non agentic part of things.
Ie. Explicitly; if the underlying LLMs dont get any better, there is no reason to expect the system built out of them to get any better.
If that was untrue, you would expect to be able to build agentic systems out of much smaller LLMs, but that overwhelmingly doesn’t work.