The Anthropic Admin API allows programmatic management of organizational resources, including members, workspaces, and API keys.
Create a message completion from Claude. Send a conversation history with user/assistant messages and receive Claude's response. Useful for: chatbots, Q&A systems, text generation, conversational AI, and any task requiring natural language understanding or generation.
Tool to retrieve details of a specific model by its ID. Use after confirming the model ID is valid.
Tool to list available models. Use when you need to see which models are available before selection.
Create messages with prompt caching to reduce costs by up to 90% and latency by up to 85% on repeated API calls. Mark static content (system instructions, documents, context) for reuse with cache_control. Cache reads cost 0.1x base price, writes cost 1.25x (5m TTL) or 2x (1h TTL). Default 5-minute TTL refreshes free on each use. Usage: Add cache_control: {"type": "ephemeral"} to the last block of static content. Up to 4 cache breakpoints supported. For conversations, cache message history by marking final user message. Minimum tokens: Opus 4.5/Haiku 4.5 need 4096, Sonnet 4.5/4 need 1024, Haiku 3.5/3 need 2048. Monitor cache_read_input_tokens and cache_creation_input_tokens in usage response to track effectiveness. Best for: RAG with large documents, agentic tools, long conversations, coding assistants, reference analysis.