On Thursday, Scale AI and the Center for AI Safety (CAIS) released Humanity's Last Exam (HLE), a new academic benchmark aiming to "test the limits of AI knowledge at the frontiers of human expertise," ...
OpenAI’s newest LLM, o3, is facing scrutiny after independent tests found it solved a far fewer number of tough math problems than the company first claimed. When OpenAI unveiled o3 in December, ...
A discrepancy between first- and third-party benchmark results for OpenAI’s o3 AI model is raising questions about the company’s transparency and model testing practices. When OpenAI unveiled o3 in ...
OpenAI secretly funded and had access to a benchmarking dataset, raising questions about high scores achieved by its new o3 AI model. Revelations that OpenAI secretly funded and had access to the ...
OpenAI’s o3: AI Benchmark Discrepancy Reveals Gaps in Performance Claims Your email has been sent The FrontierMath benchmark from Epoch AI tests generative models on difficult math problems. Find out ...
OpenAI has long been touting the capabilities of its artificial intelligence (AI) developments, especially with their o-series models that are capable of reasoning and more advanced capabilities. The ...
MIT spinout OpenAGI claims its Lux AI agent scores 83.6% on a rigorous computer-use benchmark where OpenAI's Operator hits 61.3% — at one-tenth the cost.
Some results have been hidden because they may be inaccessible to you
Show inaccessible results