>If you built an LLM exclusively on the writings and letters of John Steinbeck, you could NOT tell the LLM to solve an integral for you amd expect it to be right.
this shows that you have very less idea on how llm's work.
LLM that is trained only on john steinbeck will not work at all. it simply does not have the generalised reasoning ability. it necessarily needs inputs from every source possible including programming and maths.
You have completely ignored that LLMs have _generalised_ reasoning ability that it derives from disparate sources.
LLMs have the ability to convince you that they've "reasoned". sometimes, an application will loop the output of an LLM to its input to provide a "chain of reasoning"
This is not the same thing as reasoning.
LLMs are pattern matchers. If you trained an llm only to map some input to the output of John Steinbeck, then by golly that's what it'll be able to do. If you give it some input that isn't suitably like any of the input you gave it during training, then you'll get some unpredictable nonsense as output.
> If you trained an llm only to map some input to the output of John Steinbeck
this is literally not possible because the llm does not get generalised reasoning ability. this is not a useful hypothetical because such an llm will simply not work. why do you think you have never seen a domain specific model ever?
if you wanted to falsify this claim: "llm's cant reason" how would one do that? can you come up with some examples that shows that it can't reason? what if we come up with a new board game with some rules and see if it can beat a human at it. just feed the rules of the game to it and nothing else.
there are close to zero domain specific models that beat frontier SOTA models even in their own domain.
(other than few edge cases like token extraction)
why do you think that's the case? lets start from here.
the real answer is that you get benefits from having data from many sources that add up expontentially for intelligence.
> LLMs are pattern matchers
but lets try to falsify this. can you come up with a prompt that clearly shows that LLM's can't reason?
this shows that you have very less idea on how llm's work.
LLM that is trained only on john steinbeck will not work at all. it simply does not have the generalised reasoning ability. it necessarily needs inputs from every source possible including programming and maths.
You have completely ignored that LLMs have _generalised_ reasoning ability that it derives from disparate sources.