LLMs don't make for a particularly good database, though. The "compression" isn't very efficient when you consider that e.g. the entirety of Wikipedia - with images! - is an order of magnitude smaller than a SOTA LLM. There are no known reliable mechanisms to deal with hallucinations, either.
So, no, LLMs aren't going to replace databases. They are going to replace query systems over those databases. Think more along the lines of Deep Research etc, just with internal classified data sources.
You're right, "subsume" would be a better word here. Although vector search is also a thing that I feel should be in the AI bucket. Especially given that SOTA embedding models are increasingly based on general-purpose LLMs.
arent they complete trash as a database? "Show me people who have googled 'Homemade Bomb' in the last 30 days". For returning bulk data in a sane format it is terrible.
If their job was to process incoming data into a structured form I could see them being useful, but holy cow it will be expensive to in realtime run all the garbage they pick up via surveillance through an AI.
They ingest unstructured data, they have a natural query language, and they compress the data down into manageable sizes.
They might hallucinate, but there are mechanisms for dealing with that.
These won't destroy actual systems of record, but they will obsolete quite a lot of ingestion and search tools.