Lakera Finds Lack Of Confidence In AI Security
Written by Sue Gee   
Thursday, 22 August 2024

A survey from Lakera reveals that only 5% of cybersecurity experts have confidence in the security measures protecting their GenAI applications, despite the fact that 90% are actively using or exploring GenAI. Currently only 22% have adopted ​​AI-specific threat modeling to prepare for GenAI specific threats.

Lakera Conc

Lakera is an AI-focused security company based in Zurich that was founded in 2021 to address the fact that traditional security tools were increasingly insufficient against the new GenAI threat landscape and that a more adaptive, AI-driven approach was required.

Lakera Gandalf

To accelerate secure adoption of AI, the company created Gandalf, an educational GenAI hacking platform that challenges users to trick the AI. Playing this games provides a hands-on experience that demonstrates the importance of careful prompting and the potential risks associated with AI misuse and, disturbingly, 200,000 players out of over 1 million have successfully completed the game. Entering commands using their native language and a bit of creativity allowed these players to trick Gandalf’s level seven in only 45 minutes on average.

According to Lakera,

This stark example underscores a troubling truth: everyone is now a potential hacker and businesses require a new approach to security for GenAI.

Commenting on this David Haber, co-founder and CEO at Lakera, points out:

“With just a few well-crafted words, even a novice can manipulate AI systems, leading to unintended actions and data breaches. As businesses increasingly rely on GenAI to accelerate innovation and manage sensitive tasks, they unknowingly expose themselves to new vulnerabilities that traditional cybersecurity measures don’t address. The combination of high adoption and low preparedness may not be that surprising in an emerging area, but the stakes have never been higher.” 

To gain a fuller picture of GenAI security Lakera conducted its GenAI Security Readiness Report survey between May 15-22, 2024. It received 1,000 responses from individuals in a wide range of roles, such as developers, security analysts, and executive-level security roles like CISOs, 60% of whom have more than five years of cybersecurity experience. 

The survey found that 42% of respondents are actively using and implementing GenAI, a further 45% are exploring its use while only 9% have no current plans to adopt LLMs. Of three barriers to LLM adoption, reliability and accuracy comes tops (35%), closely followed by data privacy and security (34%), with lack of skilled personnel in third place (28%).

Lakera Concern

Looking into the perception of  risk associated with the widespread adoption of AI tools, Lakera found that 38% of respondents rated their concern about GenAI/LLM vulnerabilities as "high" and a further 39% opted for "moderate"
The report states:

This highlights a widespread recognition of the escalating risks and the urgent need for comprehensive security frameworks. 

Lakera intends to conduct this survey and produce the GenAI Security Readiness Report annually to track how preparedness changes as teams are more informed about the security risks of GenAI.

You can download the 2024 report here by providing your email address.

 

Lakera Report

More Information

AI Adoption Surges, Security Preparedness Lags Behind

Related Articles

GitHub Reveals Regional Variation In AI Usage

Developers Wary Of The AI Tools They Use

Developers Like Code Assistants Even When They Are Incorrect

Magic Prompts For LLMs?

To be informed about new articles on I Programmer, sign up for our weekly newsletter, subscribe to the RSS feed and follow us on Twitter, Facebook or Linkedin.

 

Banner


The Generation That Can Text But Not Type
08/09/2024

Gen Z-ers, those born between 1997 and 2012 and who have grown up with smartphones, are probably the most tech-savvy generation. But when it comes to working with computers, there is an important skil [ ... ]



OSI Updates Open Source AI Definition
06/09/2024

A new version of the Open Source AI Definition has been released by the Open Source Initiative. The update is part of the drive to have a stable release of the definition ready by the end of October 2 [ ... ]


More News

kotlin book

 

Comments




or email your comment to: comments@i-programmer.info

Last Updated ( Thursday, 22 August 2024 )