Evaluating ChatGPT for Software Vulnerability Tasks: A Comparative Analysis

With its impressive 1.7 trillion parameters, ChatGPT has emerged as a powerful language model. However, its applicability to code-oriented tasks, such as software vulnerability analysis and repair, remains relatively unexplored. In this article, we delve into the evaluation of ChatGPT against code-specific models, specifically examining its performance on four vulnerability tasks using the Big-Vul and CVEFixes datasets. This comprehensive analysis sheds light on the potential limitations of using ChatGPT for software vulnerability tasks while emphasizing the need for domain-specific fine-tuning.

Evaluation of ChatGPT against code-specific models

To comprehensively evaluate ChatGPT’s performance, security analysts conducted experiments using the Big-Vul and CVEFixes datasets. These datasets provide a comprehensive set of vulnerability tasks, enabling a thorough comparison of ChatGPT against baseline methods. The evaluation focused on the F1-measure and top-10 accuracy metrics.

The results of the evaluation revealed that ChatGPT achieved an F1-measure of 10% and 29% on the Big-Vul and CVEFixes datasets, respectively. These scores were significantly lower compared to the other baseline methods. Similarly, the top-10 accuracy of ChatGPT was 25% and 65%, which again reflected the lowest performance among the examined models.

Analysis of Multiclass Accuracy

In addition to F1-measure and top-10 accuracy, multiclass accuracy was also considered as a crucial performance indicator. The analysis revealed that ChatGPT achieved the lowest multiclass accuracy of 13%, showcasing a striking 45%-52% difference from the best baseline model. These outcomes underscore the challenges faced by ChatGPT in accurately classifying vulnerability tasks across multiple classes.

Evaluation of Severity Estimation

Severity estimation holds paramount importance in vulnerability analysis to prioritize remediation efforts. However, ChatGPT’s performance in this regard proved to be unsatisfactory. The evaluation indicated that ChatGPT exhibited the highest mean squared error (MSE) of 5.4 and 5.85, implying its inaccurate severity estimation compared to the other baselines. This finding raises concerns about relying on ChatGPT for precise severity estimation in vulnerability assessment.

Assessment of Repair Patch Generation

One vital aspect of vulnerability repair is the generation of correct repair patches. Regrettably, ChatGPT failed to generate accurate repair patches in this evaluation. On the other hand, the baseline models demonstrated success in rectifying vulnerable functions, correctly repairing 7% to 30% of them. This stark contrast highlights the limitations of ChatGPT in generating effective repair solutions.

Limitations of fine-tuning

Fine-tuning is a commonly employed technique to optimize language models for specific tasks. However, in the case of ChatGPT, fine-tuning for vulnerability tasks is not viable due to proprietary parameters. This constraint further underlines the challenges in adapting ChatGPT directly for software vulnerability tasks.

The Importance of Domain-specific Fine-tuning

The analysis of ChatGPT’s performance in vulnerability tasks underscores the significance of domain-specific fine-tuning. The complexity and specificity of software vulnerability tasks necessitate the customization of language models like ChatGPT to better suit the requirements. This suggests the need for further research and work on fine-tuning or adapting ChatGPT specifically for software vulnerability tasks.

Comparison with previous studies

While previous studies have examined the effectiveness of large language models in automated program repair, they have not accounted for the latest versions of ChatGPT. This article bridges that gap by shedding light on the specific performance of ChatGPT in software vulnerability tasks. Additionally, the notable disparities in results indicate the necessity for dedicated exploration of ChatGPT’s potential in this domain.

In conclusion, the evaluation of ChatGPT for software vulnerability tasks reveals its limitations in comparison to code-specific models. The lower F1-measure, top-10 accuracy, multiclass accuracy, inaccurate severity estimation, and inability to generate correct repair patches highlight the challenges faced by ChatGPT in this context. The proprietary nature of its parameters further restricts fine-tuning for vulnerability tasks. As such, this study emphasizes the need for additional research and efforts to fine-tune or tailor ChatGPT specifically for software vulnerability analysis and repair. By addressing these challenges, ChatGPT could potentially be leveraged more effectively in securing software systems in the future.

Explore more

Trend Analysis: Maritime Data Quality and Digitalization

The global shipping industry is currently grappling with a paradox where massive investments in high-end software often result in negligible improvements to the bottom line because the underlying data is essentially unreadable. For years, the narrative around maritime progress has been dominated by the allure of autonomous hulls and hyper-intelligent algorithms, yet the reality on the bridge and in the

Trend Analysis: AI Agents in ERP Workflows

The fundamental nature of enterprise resource planning is undergoing a radical transformation as the age of the passive data repository gives way to a dynamic environment where autonomous agents manage the heaviest administrative burdens. Businesses are no longer content with software that merely records what has happened; they now demand systems that anticipate needs and execute complex tasks with minimal

Why Is Finance Moving Business Central Reporting to Excel?

Finance leaders today are discovering that the rigid architecture of an enterprise resource planning system often acts more as a cage for their data than a springboard for strategic insight. While Microsoft Dynamics 365 Business Central serves as a formidable engine for transaction processing, many organizations are intentionally migrating their primary reporting workflows toward Microsoft Excel. This transition represents a

Dynamics GP to Business Central Migration – Review

Maintaining an aging on-premise ERP system in 2026 feels increasingly like trying to navigate a modern high-speed railway using a vintage steam engine’s schematics. For decades, Microsoft Dynamics GP, formerly known as Great Plains, served as the bedrock for mid-market American enterprises, providing a sturdy, if rigid, framework for accounting and inventory management. However, as the industry moves toward 2029—the

Why Use Statistical Accounts in Dynamics 365 Business Central?

Managing a modern enterprise requires more than just tracking the movement of dollars and cents across various general ledger accounts during a fiscal period. Financial clarity often depends on non-monetary metrics like employee headcount, physical floor space, or the total volume of customer interactions to provide context for the raw numbers. These metrics, known as statistical accounts, allow controllers to