#199 Cyber Defenders: Safeguarding GenAI Against Emerging Threats

In this episode, Darren is joined by guest Chris Sistito, CEO of hiddenlayer, as we uncover the vulnerabilities threatening our digital future and explore innovative solutions to protect AI systems from exploitation and misuse.


AI technologies garner significant attention for their transformative potential across multiple industries. However, this rapid technological advance also paves the way for new and unique vulnerabilities. AI models, if unprotected, expose a different kind of security turbulence not covered by traditional cybersecurity measures. Incidences such as the theft of machine learning models showcase the unique threats facing AI systems, escalating the need for developed AI cybersecurity measures.

 The Evolution of Cybersecurity Measures for AI

Conventional cybersecurity focuses predominantly on protecting the infrastructure to safeguard the data. While effective for traditional computer systems, this approach overlooks critical vulnerabilities in AI models, especially generative models and those involving reinforcement learning. AI technologies have been swiftly adopted across various sectors, increasing the urgency for cybersecurity to keep pace.

The free and unchecked exchange of AI models today parallels the early days of the internet. In today’s stringent cybersecurity environment, encryption, strict access permissions, and digital signatures secure our data. However, AI models, which function similarly to code exchange and execution, largely remain overlooked regarding security. AI platforms like Hugging Face, for example, host numerous AI models that are easily downloaded and used, often without serious thought about potential security implications.

 The Emerging Threat Landscape in AI

AI models and machine learning systems are swiftly becoming significant players in the cybersecurity arena. Threats range from malicious code hidden within model weights to simpler tactics like attaching a coin miner. These models have emerged as attractive targets for cyber threat actors, emphasizing the pressing need for an evolved cybersecurity approach.

Prompt Injection is one such technique that poses massive threats to the data community. This technique manipulates an AI model to provide information beyond its designed function. For instance, a model instructed to "write a story" could be tricked into divulging network access. Disclosure of such sensitive data can result in severe consequences, especially in industries like finance, where exposure of customer account data or IP addresses could facilitate illicit activities like data theft.

AI models can ‘hallucinate’ data, but it is not part of their training. If incorrect information is shared, it could lead to reputational harm and legal issues for businesses, particularly in sectors like healthcare. The solution lies in isolating models from the internet to reduce the attack surface and using AI-focused cybersecurity platforms such as Hidden Layer, which can analyze model behavior to identify potential threats during training and runtime operations.

 Staying Vigilant

The advent of AI has increased the potential attack surface for cybersecurity. Recognizing AI-centric threats and including these rapidly evolving technologies is the immediate need of the hour. Comprehensive protection of AI models is crucial, and businesses must consider potential loopholes for unauthorized data access and misuse.

The path forward involves marrying cybersecurity and AI from development to deployment. Failing to do so can expose various organizations and their valuable data to unique threats that traditional cybersecurity measures may not adequately combat. The intersection of AI and cybersecurity is more than a technological necessity—it’s an imperative step to ensure a safer and more secure digital future.


#199 Cyber Defenders: Safeguarding GenAI Against Emerging Threats
Broadcast by