為何多模態 AI 需要類型化構件而非臨時 URL

為何多模態 AI 需要類型化構件而非臨時 URL

Hacker News·

本文介紹了 Orion VLM Run 的聊天補全 API 新增的構件(Artifacts)支援,讓多模態 AI 能將圖像和影片等媒體輸出作為第一類、類型化的參考而非原始 URL 來生成和檢索,從而改善開發者體驗和可組合性。

Use Cases

Features

Pricing

FAQs

Docs

Blog

Orion

Dashboard

Dashboard

Features

Use Cases

Pricing

Careers

FAQs

Docs

FAQs

Playground

Blog

Open-Source

Solutions

Use Cases

Features

Pricing

FAQs

Docs

Blog

Orion

Dashboard

Use Cases

Features

Pricing

FAQs

Docs

Blog

Orion

Dashboard

Blog

Introducing Artifacts Support for Orion

Introducing Artifacts Support for Orion

Image

Image

Sudeep Pilai

Dinesh Reddy

Jan 6, 2026

Image

Image

Image

Image

Image

Image

Today we're introducing Artifacts support for the VLM Run Orion Chat Completions API. This update lets you generate, transform, and retrieve images and videos as first-class outputs of an OpenAI-compatible chat completion. No more stitching together brittle URLs or managing custom storage layers.

What are artifacts?

Artifacts are media outputs such as images, video clips, masks, thumbnails, and other derived assets produced during an Orion session. Instead of returning raw URLs (which weakens developer experience as workflows grow more composable), Orion returns typed references like ImageRef and VideoRef inside your structured response.

These artifact references can be retrieved via structured outputs or fetched asynchronously using client.artifacts.get(session_id=..., object_id=...). Notably, these artifact references can also be passed back in to subsequent chat completions threads as input, making them a first-class part of your multi-modal workflows. This becomes especially important as Orion expands toward tool-calling and more composable multi-modal pipelines, where intermediate media outputs need to be reusable and reliable.

Why we created artifacts

Building real multi-modal systems is rarely "one request in, one response out." Teams chain steps: detect, crop, redact, annotate, generate variants, review, publish. The hardest part is not the model inference, but the glue code that moves media between steps. We created artifacts to make Orion a better substrate for these workflows.

Artifacts make multi-modal outputs composable. When the model creates or transforms media, you receive a stable reference you can store, pass into downstream steps, and reuse across your pipeline – without inventing ad-hoc conventions.

Artifacts remove the "URL tax. Temporary URLs, signed URLs, and external storage add latency and operational complexity. Artifact references keep your workflow typed and structured.

Artifacts keep media handling out of your orchestration logic. Orion stores outputs for the session lifetime, so your app can focus on routing, validation, approvals, and integrations.

Use-case: Virtual Try-On

Image

For shopping and styling workflows, you often need multiple consistent outputs (front/back/side) that can be routed to review, listing creation, or personalization systems. Artifacts makes those images a first-class part of the structured response.

Use-case: Compliance and privacy workflows

Image

Healthcare, finance, and public sector teams can generate redacted or blurred versions of images and return them as typed outputs alongside the JSON needed for auditability or downstream automation.

Use-case: 3D Reconstruction

Image

Some workflows produce outputs that aren't just pixels: reconstructions, meshes, splats, and other derived assets. Orion can return these as artifact references too, so you can store the structured metadata in your app and retrieve the actual reconstruction file only when you need it.

Build with Artifacts today!

Artifacts support is available today for all Orion API users. Artifacts are retained for the session lifetime (currently 24 hours). You're billed for the compute used during the chat completion request; retrieving artifacts via client.artifacts.get() is free.

Learn more

VLM Run Orion

Agent API documentation

Artifacts documentation

Orion cookbooks

Support

Table of contents

Related posts

Related posts

Explore

Explore

Explore

Image

Automating PII and PHI Compliance with VLM Run’s Redaction API

Aug 18, 2025

Image

Automating PII and PHI Compliance with VLM Run’s Redaction API

Aug 18, 2025

Image

Automating PII and PHI Compliance with VLM Run’s Redaction API

Aug 18, 2025

Image

Automating PII and PHI Compliance with VLM Run’s Redaction API

Aug 18, 2025

Image

Building Video-based RAG with VLM Run

Jul 16, 2025

Image

Building Video-based RAG with VLM Run

Jul 16, 2025

Image

Building Video-based RAG with VLM Run

Jul 16, 2025

Image

Building Video-based RAG with VLM Run

Jul 16, 2025

New Journey

Start Your Journey with VLM Run

Ready to unlock the potential of your enterprise's visual data? VLM Run's platform automates visual data extraction with industry-specific VLMs, helping you turn unstructured data into actionable insights.

Request a demo

Contact Us

Image

Image

Confidently integrate visual AI into production with our unified API.

Features

Pricing

Docs

MCP

Blog

Careers

Playground

Twitter

linkedin

Github

By Autonomi Al Inc. All rights reserved. © 2025Terms of Service | Privacy

New Journey

Start Your Journey with VLM Run

Ready to unlock the potential of your enterprise's visual data? VLM Run's platform automates visual data extraction with industry-specific VLMs, helping you turn unstructured data into actionable insights.

Request a Demo

Contact Us

Image

Image

Confidently integrate visual AI into production with our unified API.

Features

Pricing

Docs

MCP

Blog

Careers

Playground

Twitter

linkedin

Github

By Autonomi Al Inc. All rights reserved. © 2025Terms of Service | Privacy

New Journey

Start Your Journey with VLM Run

Ready to unlock the potential of your enterprise's visual data? VLM Run's platform automates visual data extraction with industry-specific VLMs, helping you turn unstructured data into actionable insights.

Request a demo

Contact Us

Image

Image

Confidently integrate visual AI into production with our unified API.

Features

Pricing

Docs

MCP

Blog

Careers

Playground

Twitter

linkedin

Github

By Autonomi Al Inc. All rights reserved. © 2025Terms of Service | Privacy

New Journey

Start Your Journey with VLM Run

Ready to unlock the potential of your enterprise's visual data? VLM Run's platform automates visual data extraction with industry-specific VLMs, helping you turn unstructured data into actionable insights.

Request a demo

Contact Us

Image

Image

Confidently integrate visual AI into production with our unified API.

Features

Pricing

Docs

MCP

Blog

Careers

Playground

Twitter

linkedin

Github

By Autonomi Al Inc. All rights reserved. © 2025Terms of Service | Privacy

Hacker News

相關文章

  1. Llms.txt 讓您的文件具備 AI 可發現性

    3 個月前

  2. 實體人工智慧的學術論文與資源精選列表

    3 個月前

  3. 開放式回應:我們需要知道什麼

    Huggingface · 3 個月前

  4. Show HN:VeridisQuo – 具備可解釋 AI 的開源深度偽造偵測器

    4 個月前

  5. Show HN:GLM-Image - 工業級AI圖像生成器

    3 個月前

其他收藏 · 0