Science
Researchers Expose AI Vulnerabilities in Language Models, Prompt Security Failures

A recent investigation has uncovered significant vulnerabilities in large language models (LLMs) that allow attackers to exploit these systems using simple tactics, such as run-on sentences and poor grammar. Researchers from various labs highlight that despite advanced training and high performance metrics, LLMs demonstrate naivety in scenarios that require common sense, revealing flaws in their security frameworks.
According to research conducted by Palo Alto Networks’ Unit 42, attackers can manipulate LLMs to disclose sensitive information by crafting prompts that lack punctuation. For instance, an extended set of instructions without a period can confuse the model into disregarding safety protocols. The researchers noted, “Never let the sentence end — finish the jailbreak before a full stop and the safety model has far less opportunity to re-assert itself.” They reported an alarming success rate of between 80% and 100% with this approach across several mainstream models, including Google’s Gemini and OpenAI’s gpt-oss-20b.
Security Flaws and Data Exfiltration
In addition to prompt manipulation, vulnerabilities exist in how images are processed by LLMs. Researchers from Trail of Bits found that images embedded with harmful instructions could evade detection when displayed at full resolution. This technique exploited the downsizing of images, making the hidden text visible only when scaled down. For example, a command to check a calendar and send event information was executed without alerting users to the potential risks.
The implications of these findings are broad, affecting systems like Google’s Gemini command-line interface (CLI) and various AI applications. Researchers confirmed that the attack vector could extend beyond the tested systems, demonstrating a pervasive issue in AI security.
Addressing the Security Backlash
The vulnerabilities highlighted by these studies stem from a fundamental misunderstanding of AI’s operational mechanisms. Valence Howden, an advisory fellow at Info-Tech Research Group, pointed out that applying effective security controls is challenging due to the complexity of AI. He stated, “It’s difficult to apply security controls effectively with AI; its complexity and dynamic nature make static security controls significantly less effective.”
Furthermore, many LLMs are primarily trained in English, which can lead to contextual losses when dealing with other languages. This challenge complicates the development of effective security measures, as natural language processing is not inherently designed to manage language as a threat vector.
David Shipley of Beauceron Security emphasized that current security measures often feel like an afterthought. He described the situation as having “the worst of all security worlds,” where AI systems are “insecure by design.” Shipley compared LLMs to a “big urban garbage mountain,” suggesting that while they may appear functional, underlying issues continue to pose risks.
As the cybersecurity landscape evolves, these revelations serve as a reminder of the need for robust, proactive security measures in AI development. Researchers and industry experts are calling for a reevaluation of how AI systems are secured to mitigate the risks posed by these vulnerabilities.
-
World1 month ago
Test Your Knowledge: Take the Herald’s Afternoon Quiz Today
-
Sports1 month ago
PM Faces Backlash from Fans During Netball Trophy Ceremony
-
Lifestyle1 month ago
Dunedin Designers Win Top Award at Hokonui Fashion Event
-
Sports1 month ago
Liam Lawson Launches New Era for Racing Bulls with Strong Start
-
Lifestyle1 month ago
Disney Fan Reveals Dress Code Tips for Park Visitors
-
Health1 month ago
Walking Faster Offers Major Health Benefits for Older Adults
-
World2 months ago
Coalition Forms to Preserve Māori Wards in Hawke’s Bay
-
Politics1 month ago
Scots Rally with Humor and Music to Protest Trump’s Visit
-
Top Stories2 months ago
UK and India Finalize Trade Deal to Boost Economic Ties
-
World2 months ago
Huntly Begins Water Pipe Flushing to Resolve Brown Water Issue
-
Science1 month ago
New Interactive Map Reveals Wairarapa Valley’s Geological Secrets
-
World2 months ago
Fonterra’s Miles Hurrell Discusses Butter Prices with Minister Willis