Discussion about this post

User's avatar
Leo Hike's avatar

I love this point "giving 2D ARC tasks to an LLM is like expecting humans to perform reasoning in 4D". I wonder (not really) what human performance on ARC would be if they didn't see the puzzle as a 2d picture, but as a sequence of numbers or a sequence of 1d pictures.

I also could easily see humans failing on hypothetical 3d ARC tasks if their representation is not convenient enough.

Expand full comment
Solim LeGris's avatar

I'm not sure how I missed this post! I totally agree with you. This is something we've been arguing as well, and your example cleanly provides some support for this hypothesis. It's also clear evidence that o3-style models are not solving ARC problems in the same way that people are, if that's something you care about. I think that makes it even more impressive that these models can now solve this kind of problem, while lacking many of the right kinds of inductive biases that people have for solving ARC tasks!

Expand full comment
3 more comments...

No posts

Ready for more?