Evaluating DeepSeek-R1 and o1: Real-World Performance and Key Insights

In a rapidly evolving landscape of artificial intelligence technology, real-world performance evaluation becomes an integral part of understanding and harnessing these advanced models. The article “Beyond benchmarks: How DeepSeek-R1 and o1 perform on real-world tasks” delves into the in-depth analysis and comparison of two prominent AI models – DeepSeek-R1 and OpenAI’s competing model, o1. By scrutinizing their efficacy in executing real-world tasks, the focus shifts beyond the traditional benchmark tests that typically assess these models, providing a more realistic perspective on their practical applications.

Introduction to DeepSeek-R1 and o1

The primary objective of this comparison was to evaluate the models’ capability to handle ad hoc tasks requiring information gathering from the web, identifying pertinent data, and executing simple yet substantial tasks manually. The experimentation utilized Perplexity Pro Search, a tool supporting both o1 and R1, which ensured a level playing field for both models. A significant observation emerged that both models are prone to certain errors, particularly when the input prompts are not specific enough, which can lead to inaccurate or incomplete outcomes.

Interestingly, although o1 demonstrated a slightly superior aptitude for reasoning tasks, R1 offered an advantage with its transparency in the reasoning process. This transparency proved particularly useful in scenarios where mistakes occurred, which is not uncommon in real-world applications involving complex data sets and multifaceted queries. The ability to understand where and why errors happened allows users to refine their approach and enhance the accuracy of subsequent prompts, making R1 a valuable tool in iterative problem-solving processes.

Real-World Task: Calculating Returns on Investments

To comprehensively assess the models’ abilities, an experiment was designed to test their proficiency in calculating returns on investments (ROI) using web data. The task involved assuming an investment of $140 in seven major companies – Alphabet, Amazon, Apple, Meta, Microsoft, Nvidia, and Tesla – on the first day of every month from January to December 2024. The models needed to pull stock price information for the start of each month, distribute the monthly investment equally among the stocks ($20 per stock), and calculate the current portfolio value.

Despite the seemingly straightforward nature of this task, both models failed to perform it accurately. The o1 model returned a list of stock prices for January 2024 and January 2025, along with an irrelevant formula, and failed to correctly calculate the values. It erroneously concluded that there was no ROI. On the other hand, R1 misplaced the calculation by only investing in January 2024 and then calculating returns for January 2025. However, R1’s transparency in its reasoning process revealed its reliance on Perplexity’s retrieval engine for obtaining the necessary monthly stock prices, pointing directly to the source of its calculation errors.

Addressing Data Retrieval Issues

In a further attempt to mitigate the retrieval issue, additional exploration was performed by providing the models with the required data in a text file. This file included the name of each stock and an HTML table with price data from January to December 2024. Despite this proactive measure to eliminate retrieval problems, both models again failed to deliver accurate results. The o1 model did extract the data but suggested manual calculations using Excel, with its vague reasoning complicating any troubleshooting efforts.

R1, while able to correctly parse the HTML data and perform month-by-month calculations, had its final value lost in the reasoning chain. Additionally, a stock split event for Nvidia further confounded its final output. Nevertheless, the detailed reasoning trace provided by R1 allowed users to understand where the model went wrong and how to refine prompts and data formatting for improved results. This emphasis on transparency provided insightful feedback that can be invaluable for future iterations and refinement of AI models.

Performance Comparison in NBA Statistics Analysis

Another experiment tested the models on comparing the performance improvement of four leading NBA centers in terms of field goal percentage (FG%) from the 2022/2023 to 2023/2024 seasons. Despite this requiring multi-step reasoning, this task was straightforward given the public availability of player stats. Both models successfully identified Giannis Antetokounmpo as having the best improvement, although the statistical figures varied slightly between the two models.

However, a notable error occurred when both models incorrectly included Victor Wembanyama in the comparison. They failed to take into account his rookie status in the NBA, which should have excluded him from this particular analysis. In this instance, R1 stood out by providing a more detailed breakdown, including a comparison table and source links, which helped in refining the prompt. By specifying the need for NBA season stats only, the model was eventually able to exclude Wembanyama from the results, showcasing how user input and model transparency can lead to more accurate outputs.

Insights and Limitations of AI Models

In the rapidly advancing field of artificial intelligence technology, evaluating real-world performance is crucial for truly understanding and utilizing these sophisticated models. The article “Beyond benchmarks: How DeepSeek-R1 and o1 perform on real-world tasks” provides a thorough analysis and comparison of two leading AI models: DeepSeek-R1 and OpenAI’s competing model, o1. By examining their effectiveness in handling real-world tasks, the focus expands beyond conventional benchmark tests that are typically used to evaluate these models. This approach offers a more realistic insight into their practical applications. Assessing AI models through practical use cases reveals how well they perform outside of controlled experimental environments, highlighting the importance of real-world validations in the AI field. This real-world performance assessment is vital in an era where AI technology continues to evolve and become more integrated into various aspects of daily life, demonstrating that practical performance can differ significantly from theoretical benchmarks.

Explore more

Is Data Architecture More Important Than AI Models?

The glistening promise of an autonomous enterprise often shatters against the reality of a fragmented database that cannot distinguish a customer’s lifetime value from a simple transaction code. For several years, the technology sector has remained fixated on the sheer cognitive acrobatics of large language models, treating every incremental update to GPT or Claude as a definitive solution to complex

Six Post-Purchase Moments That Drive Customer Lifetime Value

The instant a digital transaction reaches completion, a profound and often ignored psychological transformation occurs within the mind of the modern consumer as they pivot from excitement to scrutiny. While the majority of contemporary brands commit their entire marketing budgets to the initial pursuit of a sale, they frequently vanish the very second a credit card is authorized. This abrupt

The Future of Marketing Automation: Trends and Growth Through 2026

Aisha Amaira is a leading MarTech strategist with a profound focus on the intersection of customer data platforms and automated innovation. With years of experience helping brands navigate the complexities of CRM integration, she specializes in transforming technical infrastructure into high-growth engines. In this conversation, we explore the evolving landscape of marketing automation, the financial frameworks required to justify large-scale

How Can Autonomous AI Agents Personalize Global Marketing?

Aisha Amaira is a distinguished MarTech strategist who has spent years at the intersection of customer data platforms and automated engagement. With a deep background in CRM technology, she specializes in transforming rigid, manual marketing architectures into fluid, insight-driven ecosystems. Her work focuses on helping brands move past the technical debt of traditional automation to embrace a future where technology

Is It Game Over for Authenticity in Job Interviews?

Ling-yi Tsai has spent decades at the intersection of human capital and technical innovation, helping organizations navigate the messy realities of digital transformation and behavioral change. With a deep focus on HR analytics and talent management systems, she understands that the data behind a hire is often just as important as the cultural “vibe” a manager senses during a first