Claude Haiku

Family

Anthropic · Claude

Anthropic's fastest Claude line for latency-sensitive, high-volume, and cost-constrained workloads.

fast efficient coding vision tool-use cost-effective model-family
Updated March 6, 2026

Overview

This is a model family overview. For version-specific details, see the individual model entries linked below.

Claude Haiku is the speed-and-cost tier in Anthropic’s lineup. It sacrifices some capability depth compared with Sonnet and Opus but delivers significantly faster responses at a fraction of the price. Haiku is the right choice when throughput, latency, or cost matter more than peak reasoning quality.

Current Latest

Claude Haiku 4.5 remains the efficiency-oriented current reference in this repo snapshot.

Strengths

  • Very fast response times — well suited for real-time applications
  • Lowest per-token cost in the Claude family
  • Surprisingly capable for its tier — handles many tasks that previously required larger models
  • Multimodal support (images and text)
  • Tool use and structured output support

When to Choose Haiku

Choose Haiku when speed or cost is the primary concern:

  • High-volume classification, extraction, and routing tasks
  • Real-time chat applications where latency matters most
  • Bulk content processing pipelines
  • Development prototyping before upgrading to Sonnet/Opus for production
  • Cost-sensitive applications that can tolerate occasional quality drops on hard tasks

Access

  • Anthropic API (direct)
  • AWS Bedrock
  • Google Vertex AI
  • Claude consumer products (often the default for quick interactions)