Skip to content
AI News
AI News2 min read4 views

Infosys Partners with Anthropic to Build Enterprise AI Agents Across Regulated Industries

Infosys and Anthropic announce a collaboration to integrate Claude into enterprise AI deployments across telecom, finance, manufacturing, and software development.

Claude Enters Regulated Enterprise Markets

Anthropic and Infosys announced a strategic collaboration on February 17, 2026, to develop and deliver enterprise AI solutions across telecommunications, financial services, manufacturing, and software development.

Integration Details

The deal integrates Anthropic's Claude models and Claude Code with Infosys' Topaz AI-powered business automation platform. Using the Claude Agent SDK, the partnership will help clients build AI agents that work persistently across long, complex processes — going beyond answering questions to independently handling multi-step tasks.

Industry Applications

  • Telecommunications: A dedicated Anthropic Center of Excellence will build and deploy AI agents tailored to industry-specific operations
  • Financial services: AI agents for risk detection, compliance reporting, and personalized customer interactions
  • Manufacturing: Accelerated product design and simulation, reducing R&D timelines
  • Software development: Claude Code integration for enterprise coding workflows

Strategic Significance

For Anthropic, the partnership opens doors into heavily regulated enterprise sectors requiring industry expertise and governance capabilities. For Infosys, it provides access to the most advanced AI coding and reasoning models available.

Still reading? Stop comparing — try CallSphere live.

CallSphere ships complete AI voice agents per industry — 14 tools for healthcare, 10 agents for real estate, 4 specialists for salons. See how it actually handles a call before you book a demo.

flowchart TD
    HUB(("Claude Enters Regulated<br/>Enterprise Markets"))
    HUB --> L0["Integration Details"]
    style L0 fill:#e0e7ff,stroke:#6366f1,color:#1e293b
    HUB --> L1["Industry Applications"]
    style L1 fill:#e0e7ff,stroke:#6366f1,color:#1e293b
    HUB --> L2["Strategic Significance"]
    style L2 fill:#e0e7ff,stroke:#6366f1,color:#1e293b
    style HUB fill:#4f46e5,stroke:#4338ca,color:#fff
flowchart LR
    IN(["Input prompt"])
    subgraph PRE["Pre processing"]
        TOK["Tokenize"]
        EMB["Embed"]
    end
    subgraph CORE["Model Core"]
        ATTN["Self attention layers"]
        MLP["Feed forward layers"]
    end
    subgraph POST["Post processing"]
        SAMP["Sampling"]
        DETOK["Detokenize"]
    end
    OUT(["Generated text"])
    IN --> TOK --> EMB --> ATTN --> MLP --> SAMP --> DETOK --> OUT
    style IN fill:#f1f5f9,stroke:#64748b,color:#0f172a
    style CORE fill:#ede9fe,stroke:#7c3aed,color:#1e1b4b
    style OUT fill:#059669,stroke:#047857,color:#fff
flowchart TD
    HUB(("Claude Enters Regulated<br/>Enterprise Markets"))
    HUB --> L0["Integration Details"]
    style L0 fill:#e0e7ff,stroke:#6366f1,color:#1e293b
    HUB --> L1["Industry Applications"]
    style L1 fill:#e0e7ff,stroke:#6366f1,color:#1e293b
    HUB --> L2["Strategic Significance"]
    style L2 fill:#e0e7ff,stroke:#6366f1,color:#1e293b
    style HUB fill:#4f46e5,stroke:#4338ca,color:#fff

This follows Anthropic's earlier partnership expansion with Accenture, which is training approximately 30,000 professionals on Claude.

Source: Anthropic | TechCrunch | Bloomberg | Infosys Newsroom

## Infosys Partners with Anthropic to Build Enterprise AI Agents Across Regulated Industries — operator perspective Infosys Partners with Anthropic to Build Enterprise AI Agents Across Regulated Industries is the kind of news that lives or dies on second-week behavior. The first benchmark is marketing. The eval suite a week later is the truth. For CallSphere — Twilio + OpenAI Realtime + ElevenLabs + NestJS + Prisma + Postgres, 37 agents across 6 verticals — the bar for adopting any new model or API is unsentimental: does it shorten the inner loop on a real call, or just on a benchmark? ## What AI news actually moves the needle for SMB call automation Most AI news is noise. A new benchmark score, a leaderboard reshuffle, a leaked memo — none of it changes whether your AI receptionist books appointments without dropping the call. The handful of things that *do* move production AI voice and chat are concrete: realtime API stability (does the WebSocket survive 5+ minutes without a stall?), language coverage (does it handle 57+ languages with usable accents, or is English the only first-class citizen?), tool-use reliability (does the model actually call the right function with the right argument types under load?), multi-agent handoffs (do specialist agents receive structured context, or just transcripts?), and latency under load (p95 first-token under 800ms when 200 concurrent calls hit the same endpoint?). The CallSphere rule on news is: if it doesn't move at least one of those five numbers in a measurable eval, it's a blog post, not a product change. What to track: provider changelogs for realtime endpoints, tool-call schema changes, language-add announcements, and any deprecation that pins your stack to a sunset date. What to ignore: leaderboard wins on tasks that don't map to your call flow, "agentic" benchmarks that don't measure tool latency, and demos that work because the prompt was hand-tuned for the demo. The teams that ship fastest treat AI news the same way ops teams treat CVE feeds — read everything, act on the small fraction that touches your runtime, archive the rest. ## FAQs **Q: How does infosys Partners with Anthropic to Build Enterprise AI Agents Across Regulated Industries change anything for a production AI voice stack?** A: Most of the time it doesn't, and that's the right starting assumption. The relevant test is whether it improves at least one of: p95 first-token latency, tool-call argument accuracy on noisy inputs, multi-turn handoff stability, or per-session cost. CallSphere ships in 57+ languages, is HIPAA and SOC 2 aligned, and runs voice, chat, SMS, and WhatsApp from the same agent stack. **Q: What's the eval gate infosys Partners with Anthropic to Build Enterprise AI Agents Across Regulated Industries would have to pass at CallSphere?** A: The eval gate is unsentimental — a regression suite that simulates real call traffic (noisy ASR, partial inputs, tool-call timeouts) measures four numbers, and a candidate has to win on three of four without losing badly on the fourth. Anything else is treated as a blog post, not a stack change. **Q: Where would infosys Partners with Anthropic to Build Enterprise AI Agents Across Regulated Industries land first in a CallSphere deployment?** A: In a CallSphere deployment, new model and API capabilities land first in the post-call analytics pipeline (lower stakes, async, easy to roll back) and only later in the live realtime path. Today the verticals most likely to absorb new capability first are Sales and Real Estate, which already run the largest share of production traffic. ## See it live Want to see salon agents handle real traffic? Walk through https://salon.callsphere.tech or grab 20 minutes with the founder: https://calendly.com/sagar-callsphere/new-meeting.
Share

Try CallSphere AI Voice Agents

See how AI voice agents work for your industry. Live demo available -- no signup required.

Related Articles You May Like

Agentic AI

Multi-Agent Handoffs with the OpenAI Agents SDK: The Pattern That Actually Scales (2026)

Handoffs done right — when one agent should hand control to another, how to preserve context, and how to evaluate the handoff decision itself.

AI Strategy

AI Agent M&A Activity 2026: Aircall–Vogent, Meta–PlayAI, OpenAI's Six Deals

Q1 2026 saw a record acquisition wave: Aircall bought Vogent (May), Meta acquired Manus and PlayAI, OpenAI closed six deals. The voice AI consolidation phase has begun.

Agentic AI

LangGraph State-Machine Architecture: A Principal-Engineer Deep Dive (2026)

How LangGraph's StateGraph, channels, and reducers actually work — with a working multi-step agent, eval hooks at every node, and the patterns that survive production.

Agentic AI

LangGraph Checkpointers in Production: Durable, Resumable Agents with Eval Replay

Use LangGraph's checkpointer to make agents resumable across crashes and human-in-the-loop pauses, then replay any checkpoint into your eval pipeline.

Agentic AI

Building Your First Agent with the OpenAI Agents SDK in 2026: A Hands-On Walkthrough

Step-by-step build of a working agent with the OpenAI Agents SDK — Agent class, tools, handoffs, tracing — plus an eval pipeline that catches regressions before merge.

Agentic AI

LangGraph Supervisor Pattern: Orchestrating Multi-Agent Teams in 2026

The supervisor pattern in LangGraph for coordinating specialist agents, with full code, an eval pipeline that scores routing accuracy, and the failure modes to watch for.