Sorry for harping on it, but I think this clearly reflects the difference between 2 approaches to storing knowledge, lossy but humongous, and lossless but limited.
LLMs - Lossy highly compressed knowledge which when prompted "hallucinates" facts. LLMs hallucinations are simply how the stored information is retrieved.
Memory (human in this case) - Extremely limited, but almost always correct.
honestly humans are nowhere near as lossless as you think, look up any study on eye-witness acounts of crimes and you will see how fallible to hallucination the human mind is as well .... at least when it comes to one-shot learning.
I feel from my own experience teaching, that it's repetition and pruning of information that really makes human memory and learning much more effective and not the act of storing the information the first time.
To some degree *all* LLM's answers are made up facts. For stuff that is abundantly present in training data those are almost always correct. For topics which are not common knowledge (allow for a great variability) you should always check.
I've started to think of LLM's as a form lossy compression of available knowledge which when prompted produces "facts".
> I've started to think of LLM's as a form lossy compression of available knowledge which when prompted produces "facts".
That is almost exactly what they are and what you should treat them as.
A lossy compressed corpus of publicly available information with a weight of randomness. The most fervent skeptics like to call LLMs "autocorrect on steroids" and they are not really wrong.
Humans do much more than replicate, that is one function we have of many.
What does an LLM do, other than output a weighted prediction of tokens based on its training database? Everything you can use an LLM for is a manipulation of that functionality.
Old Sci-Fi AI used to be an entity which have a hard facts database and was able to instantly search it.
I think that's the right direction for modern AI to move. ChatGPT uses Google searches often. So replace Google with curated knowledge database, train LLM to consult this database for every fact and hallucinations will be gone.
Now think of all the times you didn't already know enough to go and find the real answer.
Ever read mainstream news reporting on something you actually know about? Notice how it's always wrong? I'm sure there's a name for this phenomenon. It sounds like exactly the same thing.
> Oberon also doesn't seem to be actively developed anymore
That's pretty much it, for maybe 10+ years now. There was a successor project BlueBottle with some promise, but it did not deliver. Later it was renamed to A2. Surprisingly, it did not help.
IMO the authors of BB/A2 bet heavily on XML/Java hype, and were trying to make Oberon more like Java. The result was something without much internal consistency and not very usable.
Not being able to use a major browser and not having the resources to write one from scratch did not help either.
Then some of the major figures of this project left. And that was it.
There are some hobbyists and some small businesses which use it for niche projects and that is all
Claude did generated a rather good template for what I needed. It did not compile at first but I copy-pasted the errors and it fixed them.
Not all was good, though. It used literal bullets instead of `-` required for lists, but on whole the experience was positive.
It had taken me less time to fix the template than it would been taken to write it from scratch.
Something which Claude was good at. I throw him a crude ASCII "art" representation of what I want and get the right Typst code back.