How AI Platforms Like ChatGPT will Heighten Imposter Attacks—and How to Fight Them

How AI Platforms Like ChatGPT will Heighten Imposter Attacks—and How to Fight Them

Everyone in recent months has been busy either debating the merits of ChatGPT and similar natural language generators—or have been busy using them. The amount of material these AI-driven platforms will produce is mind-boggling, since there seems to be no shortage of businesses and content writers who feel they can benefit from the striking shortcuts these platforms provide. ChatGPT was in fact able to reach 100 million users in its first two months of public existence.

However, legitimate businesses aren’t the only ones who can benefit tremendously from these AI-powered capabilities to generate meaningful passages. Many pundits warn that this ability to create article-length narratives in mere seconds will make it frighteningly simple for criminals to create more persuasive phishing and imposter attacks, and at a far greater volume. This onslaught of new threats will hugely accelerate vulnerability, where even the savviest of network users could be tricked into turning over log-in credentials, healthcare information, or financial data.

It’s not surprising that AI would allow the ranks of cyber criminals to grow. Technology has often opened fields of expertise up to amateurs, making it easier for laymen with minimal skills to master tasks that formerly required much more training and effort. Consider that automated software allows anyone with a CAD program to draft impressive 3-dimensional designs, and WordPress and Wix allow users with even the most basic of abilities to create professional websites. We can view ChatGPT in the same light, as a tool for hackers. It not only allows anyone with an Internet connection to compose believable and supposedly informed text, but it also empowers hackers who start out with even the most rudimentary of skills to swiftly generate scripts and launch language for imposter cyber-attacks.

These imposter events come in various forms. In the corporate community, Business Email Compromise (BEC) occurs when nefarious actors breach the account of a high-level executive, often a CEO. The hacker will send emails from the CEO’s account directing other senior executives to do things like make large wire transfers or reveal sensitive log-in information. These “socially engineered” BEC attacks have increased by 65% since 2019, according to reports from software research site Gitnux, and are expected to spike dramatically along with the new sophistication of language generators.

Brand imposter attacks are when hackers create a credible mock-up of a site that the victim frequents, such as a financial institution, cloud provider, transport company, or healthcare organization. The criminals will send well-composed and convincing emails requesting that the victim click a link to their site due to some matter that needs attention. The user is then brought to the clever look-alike site, and prompted to enter their user names, passwords, banking details, address, or identifying healthcare information.

Here are some ways that ill-intentioned hackers can now produce code more quickly, launch attacks more precisely, and compose phishing content more eloquently than ever:

ChatGPT allows overseas hackers to write grammatically correct, accurately composed language. In the earlier days of phishing, hackers in unregulated foreign countries were often foiled by spelling mistakes, awkward phrasing, and unprofessional grammar that tipped-off readers. Natural language generators will produce well-composed email copy that is completely indistinguishable from ordinary native speech, since the text is not composed by an outsider. It’s composed by an AI algorithm, pulling from existing native sources.

ChatGPT makes it easier for cyber criminals to write effective malware. Not only do AI-based language generators instantly create prose, they can also quickly write code, aiding programmers in developing applications. Researchers have already reported evidence on the dark web of malicious actors abusing ChatGPT to speed the creation of new malware or fine-tune existing malicious programs. As usual, cyber criminals are intelligent and resourceful—they have already found ways to circumvent ChatGPT’s inherent safeguards.

How to Protect Against Heightened Attacks

All this makes it more critical than ever for businesses to use AI-driven email protection. The only way organizations can guard against the power and speed of advanced AI is to leverage the same technologies in their cyber security solutions. The challenge is that even many top-tier software packages don’t utilize best-in-class AI, because they were designed before these sophisticated tools had even been developed.

Many existing security solutions rely on traditional SEG (security email gateway) methods as their legacy technique. This involves the blacklisting of known malicious IP addresses. Yet contextual attacks like the BEC scenarios above simply can’t be detected by these SEG-based solutions. Cyber security solutions must employ powerful AI to interpret the text of ill-intended emails, identifying keywords like “wire transfer” and “credit card” or even recognizing attachments with sensitive images such as healthcare ID cards. Without these intelligent AI-based tools, which include optical character recognition, companies are vulnerable to a ramp-up in breaches now that criminals have access to tools like ChatGPT.

Organizations should consider solutions from new, next generation cybersecurity providers, especially those who specialize in email security, including solutions for anti-malware, anti-virus, and data loss protection. Outbound email protection like best-in-class encryption is also advisable, since hackers can’t exploit emails that they can’t decode. Businesses should also demand email security protection that is easy to use, in order to foster greater adoption across the organization. Technology that doesn’t get used is pointless.

In the end, the only genuine strategy for combatting the increased level of AI-based attacks from these platforms is to use the same AI tools against them. Don’t let your organization be swept up in the watershed of ChatGPT-assisted schemes.

Featured

  • Cost: Reactive vs. Proactive Security

    Security breaches often happen despite the availability of tools to prevent them. To combat this problem, the industry is shifting from reactive correction to proactive protection. This article will examine why so many security leaders have realized they must “lead before the breach” – not after. Read Now

  • Achieving Clear Audio

    In today’s ever-changing world of security and risk management, effective communication via an intercom and door entry communication system is a critical communication tool to keep a facility’s staff, visitors and vendors safe. Read Now

  • Beyond Apps: Access Control for Today’s Residents

    The modern resident lives in an app-saturated world. From banking to grocery delivery, fitness tracking to ridesharing, nearly every service demands another download. But when it comes to accessing the place you live, most people do not want to clutter their phone with yet another app, especially if its only purpose is to open a door. Read Now

  • Survey: 48 Percent of Worshippers Feel Less Safe Attending In-Person Services

    Almost half (48%) of those who attend religious services say they feel less safe attending in-person due to rising acts of violence at places of worship. In fact, 39% report these safety concerns have led them to change how often they attend in-person services, according to new research from Verkada conducted online by The Harris Poll among 1,123 U.S. adults who attend a religious service or event at least once a month. Read Now

  • AI Used as Part of Sophisticated Espionage Campaign

    A cybersecurity inflection point has been reached in which AI models has become genuinely useful in cybersecurity operation. But to no surprise, they can used for both good works and ill will. Systemic evaluations show cyber capabilities double in six months, and they have been tracking real-world cyberattacks showing how malicious actors were using AI capabilities. These capabilities were predicted and are expected to evolve, but what stood out for researchers was how quickly they have done so, at scale. Read Now

New Products

  • ResponderLink

    ResponderLink

    Shooter Detection Systems (SDS), an Alarm.com company and a global leader in gunshot detection solutions, has introduced ResponderLink, a groundbreaking new 911 notification service for gunshot events. ResponderLink completes the circle from detection to 911 notification to first responder awareness, giving law enforcement enhanced situational intelligence they urgently need to save lives. Integrating SDS’s proven gunshot detection system with Noonlight’s SendPolice platform, ResponderLink is the first solution to automatically deliver real-time gunshot detection data to 911 call centers and first responders. When shots are detected, the 911 dispatching center, also known as the Public Safety Answering Point or PSAP, is contacted based on the gunfire location, enabling faster initiation of life-saving emergency protocols.

  • QCS7230 System-on-Chip (SoC)

    QCS7230 System-on-Chip (SoC)

    The latest Qualcomm® Vision Intelligence Platform offers next-generation smart camera IoT solutions to improve safety and security across enterprises, cities and spaces. The Vision Intelligence Platform was expanded in March 2022 with the introduction of the QCS7230 System-on-Chip (SoC), which delivers superior artificial intelligence (AI) inferencing at the edge.

  • 4K Video Decoder

    3xLOGIC’s VH-DECODER-4K is perfect for use in organizations of all sizes in diverse vertical sectors such as retail, leisure and hospitality, education and commercial premises.