Hacker News 提問:正在開發 AI 代理的各位,你們是如何提升其速度的?

Hacker News·

這篇 Hacker News 的討論串詢問開發者們如何提升 AI 代理的速度。參與者分享了包括識別 LLM 調用瓶頸、更換更快的模型、減少輸入 token 數量,以及並行處理外部 API 調用等方法。

Image

  • How are you all identifying performance bottlenecks in agents?

  • What types of changes have gotten you the biggest speedups?

For us we vibe-coded a profiler to identify slow LLM calls - sometimes we could then switch out a faster model for that step or we'd realize we could shrink the input tokens by eliminating unnecessary context. For steps requiring external access (browser usage, API calls), we've moved to fast start external containers + thread pools for parallelization. We've also experimented some with UI changes to mask some of the latency.

What other performance enhancing techniques are people using?

Image

Hacker News

相關文章

  1. Ask HN:開發AI代理的公司,您是否針對特定LLM供應商進行開發?

    4 個月前

  2. 讓 AI 代理感覺即時(即使它們不是)

    7 個月前

  3. Ask HN:團隊如何處理 AI 代理的動態工具發現?

    3 個月前

  4. HN提問:AI在多大程度上改變了您作為軟體工程師的產出?

    4 個月前

  5. Google的AI模式為何如此快速且優秀?

    4 個月前