After Anthropic's AI model Claude Opus 4.5 outperformed top human candidates, the company redesigned its at-home coding test for job seekers. Led by Tristan Hume, the AI laboratory performance optimization team found that without on-site proctoring, they could not distinguish the best candidates from AI-assisted work. The new test focuses on a novel hardware optimization problem designed to stump existing AI tools. Hume also released the old version of the test and invited anyone who can surpass Claude Opus 4.5 to contact the company.
View Original
This page may contain third-party content, which is provided for information purposes only (not representations/warranties) and should not be considered as an endorsement of its views by Gate, nor as financial or professional advice. See Disclaimer for details.
After Anthropic's AI model Claude Opus 4.5 outperformed top human candidates, the company redesigned its at-home coding test for job seekers. Led by Tristan Hume, the AI laboratory performance optimization team found that without on-site proctoring, they could not distinguish the best candidates from AI-assisted work. The new test focuses on a novel hardware optimization problem designed to stump existing AI tools. Hume also released the old version of the test and invited anyone who can surpass Claude Opus 4.5 to contact the company.