How far are we from AGI or true logical reasoning? As I continue to test large language models (LLMs), I'm increasingly convinced they're performing something closer to memorization and interpolation rather than genuine logical reasoning. Today, I conducted an experiment using a probability puzzle I previously gave to trading desk