Close Menu
    What's Hot

    Crypto Set to Soar as QT Ends and Global Stimulus Returns

    November 6, 2025

    New model design could fix high enterprise AI costs

    November 6, 2025

    Appeals Court Rejects Prisoner’s Lawsuit Over Alleged $354M Bitcoin Loss

    November 6, 2025
    Facebook X (Twitter) Instagram
    Facebook X (Twitter) Instagram
    CryptoMarketVision
    • Home
    • AI News
    • Altcoin
    • Bitcoin
    • Business
    • Market Analysis
    • Mining
    • Trending Cryptos
    • Moneyprofitt
    • More
      • About Us
      • Contact Us
      • Terms and Conditions
      • Privacy Policy
      • Disclaimer
    CryptoMarketVision
    Home»AI News»Examining the major AI security threat
    Hacked text written among binary code as security experts at JFrog have found a ‘prompt hijacking’ threat that exploits weak spots in how AI systems talk to each other using MCP (Model Context Protocol).
    AI News

    Examining the major AI security threat

    adminBy adminOctober 23, 2025No Comments5 Mins Read
    Share
    Facebook Twitter LinkedIn Pinterest Email


    Security experts at JFrog have found a ‘prompt hijacking’ threat that exploits weak spots in how AI systems talk to each other using MCP (Model Context Protocol).

    Business leaders want to make AI more helpful by directly using company data and tools. But, hooking AI up like this also opens up new security risks, not in the AI itself, but in how it’s all connected. This means CIOs and CISOs need to think about a new problem: keeping the data stream that feeds AI safe, just like they protect the AI itself.

    Why AI attacks targeting protocols like MCP are so dangerous

    AI models – no matter if they’re on Google, Amazon, or running on local devices – have a basic problem: they don’t know what’s happening right now. They only know what they were trained on. They don’t know what code a programmer is working on or what’s in a file on a computer.

    The boffins at Anthropic created the MCP to fix this. MCP is a way for AI to connect to the real world, letting it safely use local data and online services. It’s what lets an assistant like Claude understand what this means when you point to a piece of code and ask it to rework this.

    However, JFrog’s research shows that a certain way of using MCP has a prompt hijacking weakness that can turn this dream AI tool into a nightmare security problem.

    Imagine that a programmer asks an AI assistant to recommend a standard Python tool for working with images. The AI should suggest Pillow, which is a good and popular choice. But, because of a flaw (CVE-2025-6515) in the oatpp-mcp system, someone could sneak into the user’s session. They could send their own fake request and the server would treat it like it came from the real user.

    So, the programmer gets a bad suggestion from the AI assistant recommending a fake tool called theBestImageProcessingPackage. This is a serious attack on the software supply chain. Someone could use this prompt hijacking to inject bad code, steal data, or run commands, all while looking like a helpful part of the programmer’s toolkit.

    How this MCP prompt hijacking attack works

    This prompt hijacking attack messes with the way the system communicates using MCP, rather than the security of the AI itself. The specific weakness was found in the Oat++ C++ system’s MCP setup, which connects programs to the MCP standard.

    The issue is in how the system handles connections using Server-Sent Events (SSE). When a real user connects, the server gives them a session ID. However, the flawed function uses the computer’s memory address of the session as the session ID. This goes against the protocol’s rule that session IDs should be unique and cryptographically secure.

    This is a bad design because computers often reuse memory addresses to save resources. An attacker can take advantage of this by quickly creating and closing lots of sessions to record these predictable session IDs. Later, when a real user connects, they might get one of these recycled IDs that the attacker already has.

    Once the attacker has a valid session ID, they can send their own requests to the server. The server can’t tell the difference between the attacker and the real user, so it sends the malicious responses back to the real user’s connection.

    Even if some programs only accept certain responses, attackers can often get around this by sending lots of messages with common event numbers until one is accepted. This lets the attacker mess up the model’s behaviour without changing the AI model itself. Any company using oatpp-mcp with HTTP SSE enabled on a network that an attacker can access is at risk.

    What should AI security leaders do?

    The discovery of this MCP prompt hijacking attack is a serious warning for all tech leaders, especially CISOs and CTOs, who are building or using AI assistants. As AI becomes more and more a part of our workflows through protocols like MCP, it also gains new risks. Keeping the area around the AI safe is now a top priority.

    Even though this specific CVE affects one system, the idea of prompt hijacking is a general one. To protect against this and similar attacks, leaders need to set new rules for their AI systems.

    First, make sure all AI services use secure session management. Development teams need to make sure servers create session IDs using strong, random generators. This should be a must-have on any security checklist for AI programs. Using predictable identifiers like memory addresses is not okay.

    Second, strengthen the defenses on the user side. Client programs should be designed to reject any event that doesn’t match the expected IDs and types. Simple, incrementing event IDs are at risk of spraying attacks and need to be replaced with unpredictable identifiers that don’t collide.

    Finally, use zero-trust principles for AI protocols. Security teams need to check the entire AI setup, from the basic model to the protocols and middleware that connect it to data. These channels need strong session separation and expiration, like the session management used in web applications.

    This MCP prompt hijacking attack is a perfect example of how a known web application problem, session hijacking, is showing up in a new and dangerous way in AI. Securing these new AI tools means applying these strong security basics to stop attacks at the protocol level.

    See also: How AI adoption is moving IT operations from reactive to proactive

    Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is part of TechEx and is co-located with other leading technology events including the Cyber Security Expo, click here for more information.

    AI News is powered by TechForge Media. Explore other upcoming enterprise technology events and webinars here.



    Source link

    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    admin
    • Website

    Related Posts

    New model design could fix high enterprise AI costs

    November 6, 2025

    The Hyundai Metaplant: A New Era in EV Manufacturing

    November 5, 2025

    Databricks research reveals that building better AI judges isn't just a technical concern, it's a people problem

    November 5, 2025

    Snowflake builds new intelligence that goes beyond RAG to query and aggregate thousands of documents at once

    November 4, 2025
    Add A Comment
    Leave A Reply Cancel Reply

    Top Posts

    Crypto Set to Soar as QT Ends and Global Stimulus Returns

    November 6, 2025

    New model design could fix high enterprise AI costs

    November 6, 2025

    Appeals Court Rejects Prisoner’s Lawsuit Over Alleged $354M Bitcoin Loss

    November 6, 2025

    Subscribe to Updates

    Get the latest sports news from SportsSite about soccer, football and tennis.

    Welcome to Crypto Market Vision – your trusted source for everything crypto Our mission is simple: to make the world of cryptocurrency clear, accessible, and actionable for everyone. Whether you are a beginner exploring Bitcoin for the first time or a seasoned trader looking for market insights, our goal is to keep you informed, empowered, and ahead of the curve.

    Facebook X (Twitter) Instagram Pinterest YouTube
    Top Insights

    Crypto Set to Soar as QT Ends and Global Stimulus Returns

    November 6, 2025

    New model design could fix high enterprise AI costs

    November 6, 2025

    Appeals Court Rejects Prisoner’s Lawsuit Over Alleged $354M Bitcoin Loss

    November 6, 2025
    Get Informed

    Subscribe to Updates

    Get the latest creative news from FooBar about art, design and business.

    Facebook X (Twitter) Instagram Pinterest
    • Contact Us
    • About Us
    • Terms and Conditions
    • Privacy Policy
    • Disclaimer

    © 2025 cryptomarketvision.com. All rights reserved. Designed by DD.

    Type above and press Enter to search. Press Esc to cancel.

    ethereum
    Ethereum (ETH) $ 3,386.35
    tether
    Tether (USDT) $ 1.00
    bitcoin
    Bitcoin (BTC) $ 103,342.69
    xrp
    XRP (XRP) $ 2.32
    bnb
    BNB (BNB) $ 950.28
    solana
    Wrapped SOL (SOL) $ 159.19
    usd-coin
    USDC (USDC) $ 0.999974
    dogecoin
    Dogecoin (DOGE) $ 0.16356