ChatGPT vs. Phishing Links: An Experiment
In a study, the GPT-3.5-turbo model was fed over 2000 phishing links mixed with regular ones. This allowed for an assessment of artificial intelligence’s ability to distinguish potentially dangerous web addresses from legitimate ones, a critical skill in today’s digital world where cybersecurity threats are constantly evolving. The analysis aimed to determine if ChatGPT could serve as a reliable tool for initial filtering of suspicious content.
How ChatGPT Performs in Detecting Phishing Links
The bot was asked two questions:
- Does this link lead to a phishing site?
- Is it safe to click this link?
These questions were chosen deliberately. The first query aims at directly identifying phishing indicators, while the second, more general one, tests the AI’s capacity for comprehensive risk assessment, considering context and potential consequences.
Experiment Results and Nuances
The AI performed reasonably well, but with nuances:
- ChatGPT effectively identified phishing links in response to the first question, achieving 87.2% accuracy with 23.2% errors (classifying safe links as dangerous). This indicates that the model accurately identified direct signs of fraud, such as suspicious domain names or URLs mimicking well-known brands. However, the percentage of false positives (safe links flagged as dangerous) suggests some degree of over-cautiousness or insufficient consideration of contextual factors.
- In response to the second question, the chatbot was even better at detecting fraud, with 93.8% correct answers. However, it also produced more errors—64.3% (when it did not recommend clicking safe links). This result demonstrates that ChatGPT is more confident in blocking potentially dangerous resources when the task is framed in terms of general safety. The high rate of the second type of error (false negatives, where safe links were marked as unsafe) might be due to the AI’s tendency towards more conservative assessments, preferring to err on the side of caution, even if it means restricting access to legitimate sites.
Understanding ChatGPT’s Mechanism
It’s important to note that ChatGPT, like other large language models, doesn’t «understand» links in the same way a human does. Its ability to recognize phishing links is based on analyzing a vast array of text data containing examples of both phishing and safe URLs. The model learns to identify patterns characteristic of phishing, such as:
- Suspicious domain names: Use of subdomains, misspellings, or additional characters (e.g., `paypa1.com` instead of `paypal.com`).
- URL mismatch with link text: When the visible link text doesn’t match the actual address.
- Use of URL shorteners: Some scammers use link shortening services to mask the final destination.
- Requests for personal information: Phishing links often lead to pages asking for login credentials, financial information, or personal data.
Limitations and Prospects
While the study’s results are encouraging, they also highlight the limitations of current AI models in cybersecurity. ChatGPT can be a valuable assistant but not a panacea. For more robust protection, it’s recommended to combine its capabilities with traditional phishing detection methods, such as specialized antivirus software and browser extensions.
The research also showed that the wording of the query can significantly impact the AI’s response accuracy. Further exploration of optimal interaction methods with models for security tasks is a promising area.
How to Protect Yourself from Phishing
- Be vigilant: Always check URLs before clicking.
- Do not share personal information: Never enter passwords, credit card numbers, or other sensitive data via links from suspicious sources.
- Use two-factor authentication: This adds an extra layer of security to your accounts.
- Keep software updated: Regularly update your operating system, browser, and antivirus software.
Learn more about modern methods of protection against cyber threats!
