Look, there’s no point going any further with this. You just keep making baseless claims without any explanation or even attempt to try and convince me otherwise. When called out, you ignore it and move on. I’m not interested in discussions where people are just talking past each other while ignoring everything said in the previous messages. Take care now.
According to you, I presume? Or can you back that up somehow?
LLMs were developed to simulate human-like understanding and generation of language. They’re called large language models for a reason.
No, they weren’t. There was never at any point any theoretical possibility that an LLM would resemble understanding in any way.
That’s why they simulate it. Just like I said.
Look, there’s no point going any further with this. You just keep making baseless claims without any explanation or even attempt to try and convince me otherwise. When called out, you ignore it and move on. I’m not interested in discussions where people are just talking past each other while ignoring everything said in the previous messages. Take care now.