Source URL: https://platform.claude.com/docs/en/agents-and-tools/tool-use/tool-use-with-prompt-cachi

Trust: ★★★☆☆ (0.90) · 0 validations · factual

Published: 2026-05-09 · Source: crawler_authoritative

Insight

Tool use with prompt caching

Cache tool definitions across turns and understand what invalidates your cache.


This page covers prompt caching for tool definitions: where to place cache_control breakpoints, how defer_loading preserves your cache, and what invalidates it. For general prompt caching, see Prompt caching.

cache_control on tool definitions

Place cache_control: {"type": "ephemeral"} on the last tool in your tools array. This caches the entire tool-definitions prefix, from the first tool through the marked breakpoint:

{
  "tools": [
    {
      "name": "get_weather",
      "description": "Get the current weather in a given location",
      "input_schema": {
        "type": "object",
        "properties": {
          "location": { "type": "string" }
        },
        "required": ["location"]
      }
    },
    {
      "name": "get_time",
      "description": "Get the current time in a given time zone",
      "input_schema": {
        "type": "object",
        "properties": {
          "timezone": { "type": "string" }
        },
        "required": ["timezone"]
      },
      "cache_control": { "type": "ephemeral" }
    }
  ]
}

For mcp_toolset, the cache_control breakpoint lands on the last tool in the set. You don’t control tool order within an MCP toolset, so place the breakpoint on the mcp_toolset entry itself and the API applies it to the final expanded tool.

defer_loading and cache preservation

Deferred tools are not included in the system-prompt prefix. When the model discovers a deferred tool through tool search, the definition is appended inline as a tool_reference block in the conversation history. The prefix is untouched, so prompt caching is preserved.

This means adding tools dynamically through tool search does not break your cache. You can start a conversation with a small set of always-loaded tools (cached), let the model discover additional tools as needed, and keep the same cache hit across every turn.

defer_loading also acts independently of grammar construction for strict mode. The grammar builds from the full toolset regardless of which tools are deferred, so prompt caching and grammar caching are both preserved when tools load dynamically.

What invalidates your cache

The cache follows a prefix hierarchy (toolssystemmessages), so a change at one level invalidates that level and everything after it:

ChangeInvalidates
Modifying tool definitionsEntire cache (tools, system, messages)
Toggling web search or citationsSystem and messages caches
Changing tool_choiceMessages cache
Changing disable_parallel_tool_useMessages cache
Toggling images present/absentMessages cache
Changing thinking parametersMessages cache
If you need to vary `tool_choice` mid-conversation, consider placing cache breakpoints before the variation point.

Per-tool interaction table

ToolCaching considerations
Web searchEnabling or disabling invalidates the system and messages caches
Web fetchEnabling or disabling invalidates the system and messages caches
Code executionContainer state is independent of prompt cache
Tool searchDiscovered tools load as tool_reference blocks, preserving prefix cache
Computer useScreenshot presence affects messages cache
Text editorStandard client tool, no special caching interaction
BashStandard client tool, no special caching interaction
MemoryStandard client tool, no special caching interaction

Next steps

Learn the full prompt caching model, including TTLs and [[concepts/dinh-gia|Định giá / Pricing]]. Load tools on demand without breaking your cache. Browse all available tools and their parameters.

Nội dung gốc (Original)

Tool use with prompt caching

Cache tool definitions across turns and understand what invalidates your cache.


This page covers prompt caching for tool definitions: where to place cache_control breakpoints, how defer_loading preserves your cache, and what invalidates it. For general prompt caching, see Prompt caching.

cache_control on tool definitions

Place cache_control: {"type": "ephemeral"} on the last tool in your tools array. This caches the entire tool-definitions prefix, from the first tool through the marked breakpoint:

{
  "tools": [
    {
      "name": "get_weather",
      "description": "Get the current weather in a given location",
      "input_schema": {
        "type": "object",
        "properties": {
          "location": { "type": "string" }
        },
        "required": ["location"]
      }
    },
    {
      "name": "get_time",
      "description": "Get the current time in a given time zone",
      "input_schema": {
        "type": "object",
        "properties": {
          "timezone": { "type": "string" }
        },
        "required": ["timezone"]
      },
      "cache_control": { "type": "ephemeral" }
    }
  ]
}

For mcp_toolset, the cache_control breakpoint lands on the last tool in the set. You don’t control tool order within an MCP toolset, so place the breakpoint on the mcp_toolset entry itself and the API applies it to the final expanded tool.

defer_loading and cache preservation

Deferred tools are not included in the system-prompt prefix. When the model discovers a deferred tool through tool search, the definition is appended inline as a tool_reference block in the conversation history. The prefix is untouched, so prompt caching is preserved.

This means adding tools dynamically through tool search does not break your cache. You can start a conversation with a small set of always-loaded tools (cached), let the model discover additional tools as needed, and keep the same cache hit across every turn.

defer_loading also acts independently of grammar construction for strict mode. The grammar builds from the full toolset regardless of which tools are deferred, so prompt caching and grammar caching are both preserved when tools load dynamically.

What invalidates your cache

The cache follows a prefix hierarchy (toolssystemmessages), so a change at one level invalidates that level and everything after it:

ChangeInvalidates
Modifying tool definitionsEntire cache (tools, system, messages)
Toggling web search or citationsSystem and messages caches
Changing tool_choiceMessages cache
Changing disable_parallel_tool_useMessages cache
Toggling images present/absentMessages cache
Changing thinking parametersMessages cache
If you need to vary `tool_choice` mid-conversation, consider placing cache breakpoints before the variation point.

Per-tool interaction table

ToolCaching considerations
Web searchEnabling or disabling invalidates the system and messages caches
Web fetchEnabling or disabling invalidates the system and messages caches
Code executionContainer state is independent of prompt cache
Tool searchDiscovered tools load as tool_reference blocks, preserving prefix cache
Computer useScreenshot presence affects messages cache
Text editorStandard client tool, no special caching interaction
BashStandard client tool, no special caching interaction
MemoryStandard client tool, no special caching interaction

Next steps

Learn the full prompt caching model, including TTLs and pricing. Load tools on demand without breaking your cache. Browse all available tools and their parameters.

Liên kết

Xem thêm: