Verdic:AI系統的意圖治理層

Hacker News·

Verdic 推出了一個 AI 系統的意圖治理層,旨在透過評估輸出的預設行為和規範性壓力,而非僅僅過濾關鍵字,來監控和控制 AI 的行為,特別是在代理式工作流程中。

Image

As models become more agentic, outputs often shift quietly from descriptive to prescriptive behavior — without any explicit signal that the system is now effectively taking action. Keyword filters and rule-based guardrails break down quickly in these cases.

Verdic is an intent governance layer that sits between the model and the application. Instead of checking topics or keywords, it evaluates:

whether an output collapses future choices into a specific course of action

whether the response exerts normative pressure (directing behavior vs explaining)

The goal isn’t moderation, but behavioral control: detecting when an AI system is operating outside the intent it was deployed for, especially in regulated or decision-critical workflows.

Verdic currently runs as an API with configurable allow / warn / block outcomes. We’re testing it on agentic workflows and long-running chains where intent drift is hardest to detect.

This is an early release. I’m mainly looking for feedback from people deploying LLMs in production, especially around:

agentic systems

AI governance

risk & compliance

failure modes we might be missing

Happy to answer questions or share more details about the approach.

Image

Hacker News

相關文章

  1. Show HN:Verdic Guard – 為生產環境 AI 提供確定性護欄

    3 個月前

  2. 意圖鴻溝:為何 AI 代理能出色地達成錯誤的目標

    4 個月前

  3. Vercel 的 Agents.md 在評估中表現優於傳統 Agent Skills

    3 個月前

  4. 生產環境中的代理式AI:設計帶有護欄的自主多代理系統 (2026指南)

    Medium · 3 個月前

  5. Show HN:意圖層:AI代理的上下文工程技能

    3 個月前