What makes LLMs work isn't deep neural networks or attention mechanisms or vector databases or anything like that.
What makes LLMs work is our tendency to see faces on toast.
@jasongorman No, seriously - that's exactly what it is & it's a mechanism exploited by charlatans since always: https://softwarecrisis.dev/letters/llmentalist/
@jwcph @jasongorman I’ve always felt like the way chatbots speak was a little too general and assertive for me to believe/trust them, and I’m glad to learn *why* it is like that. The psychic comparison is very effective.
That and spoonfuls of confirmation bias.
I am happy that there's a good generative grammar model. That's a significant thing it may go some way to suggest that grammar can be generalised.
This is orthogonal to machine intelligence, however.
@jasongorman patterns on toast are random and we interpret it as faces. So you say that LLMs produce random text and we interpret it? That's no true at all. If that would be true, the output of LLMs would look like this:
The fan box deep is tendency let fedolin are have with no a where I been water can mouse. Big dark wall health we garage child software haben open.
@duco I'm suggesting that we project human qualities like understanding and reasoning on to the output of LLMs where there are none.
@jasongorman if it looks like a duck and it quacks like a duck, it can't possibly be an artificial duck
Pareidolia is fun and cool unless you are fooling yourself that the faces you see are real and they are talking to you.