A large-scale empirical analysis of custom GPTs' vulnerabilities in the OpenAI ecosystem

Sunday Ogundoyin, Muhammad Ikram, Hassan Asghar, Benjamin Zhao, Dali Kaafar

Research output: Contribution to conferencePaperpeer-review

Abstract

Millions of users leverage generative pretrained transformer (GPT)-based language models developed by leading model providers for a wide range of tasks. To support enhanced user interaction and customization, many platforms- such as OpenAI- now enable developers to create and publish tailored model instances, known as custom GPTs, via dedicated repositories or application stores. These custom GPTs empower users to browse and interact with specialized applications designed to meet specific needs. However, as custom GPTs see growing adoption, concerns regarding their security vulnerabilities have intensified. Existing research on these vulnerabilities remains largely theoretical, often lacking empirical, large-scale, and statistically rigorous assessments of associated risks.

In this study, we analyze 14,904 custom GPTs to assess their susceptibility to seven exploitable threats, such as roleplay-based attacks, system prompt leakage, phishing content generation, and malicious code synthesis, across various categories and popularity tiers within the OpenAI marketplace. We introduce a multi-metric ranking system to examine the relationship between a custom GPT's popularity and its associated security risks.

Our findings reveal that over 95% of custom GPTs lack adequate security protections. The most prevalent vulnerabilities include roleplay-based vulnerabilities (96.51%), system prompt leakage (92.20%), and phishing (91.22%). Furthermore, we demonstrate that OpenAI's foundational models exhibit inherent security weaknesses, which are often inherited or amplified in custom GPTs. These results highlight the urgent need for enhanced security measures and stricter content moderation to ensure the safe deployment of GPT-based applications
Original languageEnglish
Pages1-17
Number of pages17
Publication statusSubmitted - 15 May 2025
EventInternational Symposium on Research in Attacks, Intrusions, and Defenses - Gold Coast, Australia, Gold Coast, Australia
Duration: 19 Oct 202522 Oct 2025
Conference number: 28
https://raid2025.github.io/

Conference

ConferenceInternational Symposium on Research in Attacks, Intrusions, and Defenses
Abbreviated titleRAID
Country/TerritoryAustralia
CityGold Coast
Period19/10/2522/10/25
Internet address

Cite this