Xiaomi Corp. today released MiMo-7B, a new family of reasoning models that it claims can outperform OpenAI’s o1-mini at some tasks. The algorithm series is available under an open-source license. Its ...
Nvidia's Nemotron-Cascade 2 is a 30B MoE model that activates only 3B parameters at inference time, yet achieved gold medal-level performance at the 2025 IMO, IOI, and ICPC World Finals. Nvidia has ...
DeepSeek made waves in early 2025, launching one of the world's first free-to-access thinking models. Now, the Chinese firm has just released DeepSeekMath-V2 with the objective of achieving ...
Mathematics is deemed to be beyond figures. It is described as the foundation of resilience in society. Thus, this made Temitope Comfort Iroko, a PhD candidate in Mathematics at the University of ...
This study introduces MathEval, a comprehensive benchmarking framework designed to systematically evaluate the mathematical reasoning capabilities of large language models (LLMs). Addressing key ...
Top artificial intelligence systems now ace many textbook-style math questions, yet they still fall apart on genuinely new problems. The gap between polished performance on familiar benchmarks and ...
Some results have been hidden because they may be inaccessible to you
Show inaccessible results