Announcement_mental_preprint

New preprint: Large Vision Models Can Solve Mental Rotation Problems! We systematically evaluate ViT, CLIP, DINOv2, and DINOv3 on layer-wise mental-rotation tasks and find that self-supervised ViTs capture geometry better than supervised ones, intermediate layers outperform final layers, and difficulty increases with rotation/occlusion—mirroring human reaction-time patterns.