研究人員毒化竊取數據,使AI系統返回錯誤結果

研究人員毒化竊取數據,使AI系統返回錯誤結果

Hacker News·

研究人員開發了一種方法,能夠污染竊取的知識圖譜數據,若未經同意整合到GraphRAG AI系統中,將使其變得無用。此技術旨在透過使AI輸出不可靠,來防止敏感資訊被未經授權地使用。

Image

Image

Image

Image

Image

Topics

Security

Off-Prem

On-Prem

Software

Offbeat

Special Features

Vendor Voice

Vendor Voice

Resources

Image

Researchers poison stolen data to make AI systems return wrong results

Image

Wanted: Chief Disinformation Officer to pollute company knowledge graphs

Image

Image

Researchers affiliated with universities in China and Singapore have devised a technique to make stolen knowledge graph data useless if incorporated into a GraphRAG AI system without consent.

Large language models (LLMs) base their predictions on training data and cannot respond effectively to queries about other data. The AI industry has dealt with that limitation through a process called retrieval-augmented generation (RAG), which gives LLMs access to external datasets. Google's AI Overviews in Search, for example, use RAG to provide the underlying Gemini model with current, though not necessarily accurate, web data.

GraphRAG represents Microsoft's effort to make RAG more effective. By creating semantically related data clusters called knowledge graphs (KGs), GraphRAG outperforms basic RAG when linked to an LLM-based system. The structuring of the data makes it easier for the LLM to make accurate predictions when prompted.

Amazon, Google, and Microsoft all support GraphRAG in their respective cloud services.

In a preprint paper titled Making Theft Useless: Adulteration-Based Protection of Proprietary Knowledge Graphs in GraphRAG Systems, authors Weijie Wang, Peizhuo Lv, et al. observe that enterprise KGs can cost a considerable amount to build, citing a figure of $5.71 per factual statement [PDF] in the KG encompassing 21 million assertions available in Cyc.

Given the potential expense, companies have an incentive to prevent KG assets from being stolen and used to build a competitive AI-oriented product – a concern exhibited by publishers, authors, and other creators of media content. Companies like Pfizer and Siemens have invested in KGs to facilitate drug discovery and assist with manufacturing.

Academics Wang, Lv, and their co-authors propose a KG defense called AURA, which stands for "Active Utility Reduction via Adulteration." The ten authors are affiliated with the Chinese Academy of Sciences, National University of Singapore, Nanyang Technological University, and Beijing University of Technology.

AURA, they explain in their paper, is "a novel framework that makes a stolen KG unusable to an adversary while maintaining minimal performance overhead for the GraphRAG system."

Essentially, it's a mechanism for subtly poisoning or adulterating the data that goes into a KG such that accurate retrieval requires a secret key. Unlike traditional encryption, the goal is not to deny access to cleartext; rather it's to degrade KG responses to LLMs such that predictions made without the key produce reduced accuracy and hallucinations.

Alternative approaches like watermarking may have some utility for making data theft traceable, but they don't address misuse of stolen data in a private setting. And the authors argue that encryption isn't practical.

"Fully encrypting the text and embeddings would require decrypting large portions of the graph for every query," they claim. "This process introduces prohibitive computational overhead and latency, making it unsuitable for real-world use."

The threat model here assumes that the attacker has been able to steal a KG outright but hasn't obtained the secret key. Trade secret lawsuits confirm that companies like Waymo aren't keen to see their IP assets spirited away.

The researchers tested their technique by creating adulterated KGs using datasets MetaQA, WebQSP, FB15K-237, and HotpotQA, then attempted to deploy GraphRAG systems using these poisoned KGs in conjunction with various LLMs (GPT-4o, Gemini-2.5-flash, Llama-2-7b, and Qwen-2.5-7b).

The results indicate that AURA is highly effective. The models retrieved adulterated content 100 percent of the time and emitted incorrect responses to users based on that misinformation 94 percent of the time.

The technique is not perfect, the academics note, because in some cases the KG may contain both correct and incorrect (adulterated) data about a subject and the LLM may choose the correct answer.

There are techniques for detoxifying poisoned data but the authors claim their approach mostly resists checks based on semantic consistency (e.g. Node2Vec), on graph-based anomaly detection (e.g. ODDBALL [PDF]), and on hybrid approaches (e.g. SEKA).

"By degrading the stolen KG's utility, AURA offers a practical solution for protecting intellectual property in GraphRAG," the authors conclude. ®

Image

More about

Narrower topics

Broader topics

Image

Image

More about

Narrower topics

Broader topics

Send us news

Other stories you might like

The Register Biting the hand that feeds IT

Image

Image

Image

Image

Image

Copyright. All rights reserved © 1998–2025

Hacker News

相關文章

  1. AI產業內部人士發起「毒化數據」行動,以阻礙AI發展

    3 個月前

  2. 多數開發者不信任 AI 生成的程式碼,但仍未檢查

    3 個月前

  3. RAG 系統中的文件投毒:攻擊者如何污染 AI 的資料來源

    大約 1 個月前

  4. 林納斯·托瓦茲:停止糾結於內核文檔中的AI生成內容

    3 個月前

  5. 對機器人程式不感興趣,專案點名並羞辱AI生成的開源軟體

    3 個月前