The ARC Prize Foundation has launched the ARC-AGI-3 benchmark, a new tool designed to assess the true intelligence of AI agents. Unlike its predecessors, ARC-AGI-3 operates in an interactive, turn-based 64×64 grid environment where AI agents must independently explore, infer rules, and plan actions without predefined instructions. This benchmark emphasizes "action efficiency," rewarding agents that solve tasks with fewer steps, thus highlighting genuine reasoning over brute-force methods. The benchmark's release follows concerns about previous versions being compromised by AI models' training data. ARC-AGI-3 aims to prevent such issues with its autonomous goal-discovery feature. Current scores for leading AI models include Google Gemini 3.1 Pro Preview at 0.37% and OpenAI GPT-5.4 (High) at 0.26%. The ARC Prize 2026 offers over $2 million in prizes for top-performing AI agents.