Submitted by New_Computer3619 t3_11hxwsm in MachineLearning
The title of this post is a Tom Scott's video which I watched a while back. I tried the challenges with ChatGPT. Seem like it handle both cases very well.
I wonder how ChatGPT can infer from context like these?
​
​
Edit: I tried the same questions but in separate chats and ChatGPT messed up. Seem like ChatGPT can only analyze sentences grammatically without any "intuition" like us. Is that correct?
​
​
currentscurrents t1_javx4pw wrote
The Winograd Schema is a test of commonsense reasoning. It's hard because it requires not just knowledge of english, but also knowledge of the real world.
But as you found, it's pretty much solved now. As of 2019 LLMs could complete it with better than 90% accuracy, which means it was actually already solved when Tom Scott made his video.