How far are we from AGI or true logical reasoning?
As I continue to test large language models (LLMs), I'm increasingly convinced they're performing something closer to memorization and interpolation rather than genuine logical reasoning.
Today, I conducted an experiment using a probability puzzle I previously