AI Tool Calling Integrations: Best Infrastructure for LLM-Native Products (2026)
See more here

Ampersand Blog Writings from the founding team

Integration Platforms
18 min read
Apr 28, 2026
Article cover image

AI Tool Calling Integrations: Why Ampersand Is the Right Integration Infrastructure for LLM-Native Products

Why AI tool calling requires more than OAuth, and how deep, schema-aware integration infrastructure powers real LLM-native products

Chris Lopez's profile picture

Chris Lopez

Founding GTM

AI Tool Calling Integrations: Why Ampersand Is the Right Integration Infrastructure for LLM-Native Products (Beyond Nango, Prismatic, and Paragon)

The teams shipping the most ambitious AI agent products in 2026 are running into a conversation that did not exist twelve months ago. They have an LLM that can call tools. They have customers who expect that LLM to read and write into their CRM, their ERP, their ticketing system, sometimes all three at once. And they are choosing integration infrastructure based on the wrong question.

The wrong question: which platform has the easiest OAuth flow for building a Slack tool.

The right question: which platform turns dozens of customer-specific systems of record into reliable, governed, schema-aware tools that an LLM can call without hallucinating field names, without leaking data across tenants, and without the engineering team owning a permanent on-call rotation.

There is a buyer narrative going around that AI tool calling belongs on Nango, Prismatic, or Paragon, while Ampersand is for "traditional" product integrations. That framing is wrong, and it is leading product teams to pick infrastructure that will collapse under their first ten enterprise customers. This post walks through what AI tool calling actually requires from integration infrastructure, where each platform fits, and why Ampersand is built for the LLM-native pattern most agent companies are actually trying to ship.

What AI Tool Calling Actually Requires From Integration Infrastructure

The phrase "tool calling" is doing a lot of heavy lifting right now. In OpenAI's function calling spec, in Anthropic's tool use API, and in the Model Context Protocol that has been gaining traction across the ecosystem, tool calling means the same thing at the surface: the LLM emits a structured request to invoke a named function with typed arguments, the runtime executes that function, and the result is fed back into the model's context.

The hard part is what that named function is actually doing.

In a demo, a tool is a wrapper around a third-party API or a stub that returns canned data. In a real product, a tool is a contract that says: when the user asks the agent to "update the deal stage on the Acme account," the system needs to find the right Salesforce or HubSpot or Pipedrive record across millions of customer-owned records, write to the correct stage value as the customer's instance defines it, respect the customer's permissions, log the change in a way an admin can audit, and recover gracefully when the underlying API throttles or returns a partial success.

That is not OAuth wrapping. That is integration infrastructure.

When engineering leaders we have worked with start mapping out the requirements for AI tool calling integrations, the list is consistent across vertical AI, conversational AI, sales engineering AI, support agents, and ops agents.

They need schema-aware tool definitions. The LLM cannot hardcode field names because every customer has different custom fields. Custom_Pipeline_Stage__c, Stage_Custom2__c, Stage. Same concept, three different field IDs, three different data types depending on which Salesforce org you are looking at. The tool definitions need to reflect each customer's actual schema, not a textbook version of it.

They need bi-directional read and write. Reads are not enough. The agent needs to push state back. That means write paths that respect required fields, validation rules, picklist constraints, workflow triggers, and concurrent edits.

They need multi-tenant isolation. Every customer's connection has to be sealed off from every other. Token bleed, request mis-routing, or cache leakage between tenants is not a bug to fix later, it is a P0 security incident, and it is the first thing an enterprise security review will probe.

They need token and auth management that does not break. OAuth tokens expire. Refresh flows fail. APIs change their auth requirements quietly. Any platform that requires the engineering team to maintain auth state across hundreds of customer connections has already lost the bet, which is exactly the point we made in auth and token management isn't an integration.

They need real-time reads with sane caching. Tool calling latency matters because it sits inside the model's response loop. If a "get account record" tool takes eight seconds, the agent feels broken. Pre-fetched, indexed, cache-warm reads are the difference between an agent that ships and an agent that times out.

They need idempotent writes with full observability. When the agent writes, the engineering team needs to know what was written, when, by which session, and whether it succeeded. Tool calls without an audit trail are unshippable in any regulated vertical.

They need custom object support. Agents do not just touch standard objects. They touch the custom objects each customer has built into their CRM or ERP, which means the integration platform needs to handle dynamic object definitions, not a fixed schema.

They need backfills and historical context. The first time a customer connects, the agent needs context. That context is in years of historical records. Integration infrastructure that cannot bulk backfill on connection is starting every conversation from zero.

This is the actual surface area of AI tool calling integrations. OAuth is one tile in it.

Why the Buyer Narrative Got AI Tool Calling Wrong

The AI tool calling pattern moved from research demo to production in roughly the last eighteen months. OpenAI's function calling, Anthropic's tool use, and the Model Context Protocol have all converged on the same general shape, which means there is now a real market for infrastructure that turns enterprise systems into LLM-callable tools.

The buyers in that market split into two groups, and the buyer narrative collapsed both into one.

Group one is building horizontal AI assistants. ChatGPT-style products that need to talk to whatever the user has connected. For these products, breadth of connector coverage and ease of OAuth is the dominant requirement. Lightweight OAuth proxies fit cleanly here, and there are several reasonable choices.

Group two is building vertical, agent-shaped products that ship to enterprise customers. AI SDR tools, AI customer service agents, AI accounting and finance assistants, AI healthcare workflow automation, AI sales engineering copilots, AI ops tools that touch HRIS and ERP. These products live and die on how deeply they read and write into a small number of customer-critical systems: Salesforce, HubSpot, NetSuite, SAP, Sage, Zendesk, Marketo, Microsoft Dynamics 365. Tool calling for these companies is not "wrap an OAuth flow." It is "be the agent's memory of the enterprise." It demands what we call vertical-specific integrations, deeply scoped to the systems that matter.

The second group is where most AI agent revenue is actually being generated, and it is the group that gets pushed toward the wrong infrastructure when the buyer narrative says "AI tool calling means lightweight OAuth product."

A useful frame we have used with engineering leaders comes from our piece on how AI agents break every integration pattern that worked for traditional SaaS. The read/write loop is so much tighter, the schema dependence so much deeper, and the failure cost so much higher that infrastructure built for batch sync, scheduled jobs, or one-way data pipes simply does not bend to fit the shape of agent work. Tool calling is the most extreme version of this: every model turn might emit one or several integration calls, and each one has to land cleanly.

How Ampersand Handles AI Tool Calling

Ampersand was built around the assumption that integration is part of the product, not a side concern. That orientation is what makes the platform fit AI tool calling so cleanly, even though tool calling was not the original framing.

Ampersand is a deep integration platform for product developers. It lets SaaS companies, including AI agent companies, build native, bi-directional, enterprise-grade integrations across CRM, ERP, GTM, HRIS, Accounting, Life Sciences, and Health Care systems using a declarative, version-controllable YAML framework. The architecture maps onto AI tool calling requirements point for point.

Schema-aware reads and writes. Ampersand exposes each customer's actual schema, including custom fields and custom objects, through a unified interface. When an agent needs a tool definition, the platform can generate it from the customer's live schema instead of a static definition. That eliminates the entire class of "agent wrote to the wrong field" failures.

Real-time read API and on-demand writes. Ampersand provides on-demand read and write API endpoints that map directly onto the patterns LLM tool calls need. The agent can request a record by ID, by external ID, or by query, and write back updates with idempotency keys, all behind a stable interface that does not change when the underlying CRM or ERP API changes.

Bi-directional sync with managed state. Beyond on-demand calls, Ampersand handles scheduled reads, backfills, and bulk write optimization. That matters for agents because the model's context is only as current as the underlying state. Pre-warmed reads keep tool calls fast. Backfills give the agent historical context on day one rather than after weeks of silent accumulation.

Field mapping that adapts per tenant. Ampersand's field mapping is dynamic, declared in YAML, and adapts to each customer's specific schema. We have written about how field mapping is how AI agents learn enterprise reality, because in practice the agent's accuracy depends on mapping the customer's idiosyncratic field names back to the agent's universal concepts. A platform that cannot do this dynamically forces the engineering team to write per-customer mapping logic, which collapses around customer ten.

Managed auth with automatic token refresh. Ampersand handles OAuth, refresh tokens, and re-auth flows across every supported system of record, so the agent's tool calls never fail because a token quietly expired at 3am.

Multi-tenant safety by design. Every customer connection is isolated. Logs, requests, schemas, and field maps are scoped per tenant. There is no shared state that could leak between customers, which is the bar enterprise buyers will hold AI agent products to. We have laid out the architectural details in building multi-tenant CRM integrations at scale.

Observability for tool calls. Every read and write Ampersand executes is logged, alertable, and inspectable. When a customer asks why their agent updated a record, the engineering team can answer in seconds, not days. When a CRM API starts returning 429s, the platform surfaces it before the customer notices.

Customer proof points back this up. 11x cut their AI phone agent's response time from 60 seconds to 5 by routing tool calls through Ampersand instead of an in-house integration layer. That is the difference between an agent that feels real and one that feels broken. The CTO at Hatch, has said Ampersand lets his team focus on building product instead of maintaining integrations, which is the underlying promise that holds given the AI-shaped product surface they own.

Ampersand vs Nango vs Prismatic vs Paragon for AI Tool Calling

A clean comparison helps clarify where each platform actually shines and why the decision usually comes down to the specific shape of the AI agent product being built.

CapabilityAmpersandNangoPrismaticParagon
Primary use caseNative Product Integrations for vertical AI, CRM and ERP-deep productsOAuth proxy and unified API across many third-party APIsEmbedded iPaaS for low-code customer-facing workflowsEmbedded iPaaS for in-app integrations
Bi-directional read/write to CRMs and ERPsFirst-class, with bulk write optimizationRead-focused, writes possible but require custom codeYes, via workflow builderYes, via workflow builder
Custom field and custom object supportDynamic per-tenant schema discoveryLimited, often per-customer codeSupported in workflows, per-customer configSupported in workflows, per-customer config
Field mapping per tenantDeclarative YAML, version-controlled, dynamicManual or app-sideConfigured per workflowConfigured per workflow
Schema-aware tool definitions for LLMsYes, exposes live schema for tool generationNot a built-in patternNot a built-in patternNot a built-in pattern
Managed OAuth and token refreshFull lifecycleCore strengthYesYes
Backfills and scheduled readsNativeLimitedWorkflow-dependentWorkflow-dependent
Integration-as-code (declarative, version-controlled)YAML-first, CI/CD friendlyCode-first SDKLow-code, visual workflowsLow-code, visual workflows
Multi-tenant isolationFirst-class, enterprise-readyPer-connectionPer-customerPer-customer
Observability and audit logsBuilt-in dashboards, logs, alertingBasicBuilt-in workflow logsBuilt-in logs
Best fit forVertical AI agents, CRM and ERP-deep products, multi-tenant SaaSLightweight OAuth-heavy products spanning many APIsVisual workflows owned by ops or implementation teamsVisual workflows for in-app automations

The pattern that emerges is not that one platform is universally better. It is that the platforms have different shapes, and AI tool calling for vertical agent products has a shape that overlaps almost completely with what Ampersand was built for.

Nango is excellent if your tool calling surface is shallow, mostly read-heavy, and spans dozens of small APIs where breadth and OAuth simplicity dominate. Prismatic and Paragon are strong when integrations are visual workflows configured by humans rather than tool calls emitted by a model. The moment your AI agent needs deep, bi-directional, schema-aware behavior across CRMs and ERPs, with multi-tenant safety and the ability to expose live schemas to the model, Native Product Integrations are the right architectural choice, and that is the lane Ampersand is in.

The Ampersand Sell: Native Product Integrations Built for LLM-Native Workloads

If your product is an AI agent that needs to read and write to a customer's CRM or ERP with the precision of a senior admin and the speed of a tool call, you do not have an OAuth problem. You have an integration infrastructure problem. Ampersand solves that problem the same way your platform solves the agent intelligence problem: by treating it as a first-class engineering surface, declaratively defined, multi-tenant from day one, and managed so your team is not the on-call rotation.

The platform supports Direct Integrations across NetSuite, SAP, Sage, Salesforce, HubSpot, Marketo, Microsoft Dynamics 365, Zendesk, Gong, and hundreds more through open-source connectors. It is GDPR compliant and ISO certified, which means it walks through enterprise security review without surprises. It is integration-as-code, which means your CI/CD pipeline owns integration changes the same way it owns product changes. And it is built for the workload AI agents actually generate: many small, high-frequency reads and writes across dozens of customer-specific schemas, with observability that survives a Sunday night incident.

If your team is comparing Native Integrations against embedded iPaaS or OAuth-only platforms for AI tool calling, the technical decision usually comes down to three questions. Does the platform expose live customer schemas in a form an LLM can use as a tool definition? Does it handle custom objects and custom fields without per-customer code? And does it scale to multi-tenant production traffic without forcing the team to babysit auth and rate limits? Ampersand was built to answer yes to all three, and you can see the full architecture in how Ampersand actually works, with reference patterns and SDK details available across the Ampersand documentation.

For teams considering whether to build this layer in-house, our piece on why building integrations in-house breaks at scale walks through the failure modes we see most often in AI agent companies that try to own the integration plane themselves. The TL;DR is that the maintenance cost compounds faster than the product roadmap can absorb it, and the engineering team that started building integrations to ship features ends up running an integration company inside their AI company.

Frequently Asked Questions

Why not just use Nango if all I need is OAuth and a couple of API calls?

If your product really only needs OAuth and a thin wrapper around a few endpoints, an OAuth proxy is fine. The trap is that AI tool calling almost always grows past that surface area. The first enterprise customer wants custom field support. The second wants custom objects. The third wants the agent to write back deal stages and have those updates respect their workflow rules. By the time you are at five enterprise customers, you have rebuilt half of an integration platform on top of an OAuth library, and you still have to maintain it. It is cheaper to start with infrastructure that handles the full surface area of Native Product Integrations than to retrofit it later.

Can Prismatic or Paragon do AI tool calling?

They can route data, and you can build workflows that an agent triggers. What they were not built for is exposing a live, customer-specific schema to an LLM so it can construct accurate tool calls dynamically. Workflow builders are designed for human-defined flows, not model-emitted ones. You can force the pattern, but you end up with a lot of glue code that translates between what the LLM wants to express and what the workflow engine accepts as input. Ampersand exposes the schema directly to the agent layer, which removes that translation step entirely.

How does Ampersand actually plug into an LLM tool calling loop?

The pattern most teams adopt is straightforward. The agent's tool definitions are generated from the customer's Ampersand-managed schema, so they reflect that customer's real CRM or ERP fields. When the LLM emits a tool call, the agent's runtime translates that into an Ampersand read or write API call. The result comes back, gets formatted for the model, and goes into the next turn of the conversation. Because Ampersand handles auth, schema, multi-tenancy, and rate limiting underneath, the agent runtime stays small and focused on reasoning.

What about the Model Context Protocol (MCP)? Where does Ampersand fit?

MCP is an open protocol for letting LLMs and external tools talk, and it is rapidly becoming a standard. Ampersand's API and capabilities map cleanly onto MCP server implementations: the same schema-aware reads and writes that power tool calling in OpenAI's function-calling format work as MCP resources and tools. Teams building with MCP get the same benefit, namely a reliable, governed, multi-tenant Integration Infrastructure layer underneath the protocol. You are not picking between Ampersand and MCP. You are using Ampersand to make MCP servers actually useful in enterprise contexts.

We are early-stage. Is Ampersand overkill for our first AI agent product?

The opposite, usually. Early-stage AI agent companies are the ones who can least afford to spend six months building integration infrastructure that they will then spend the next two years maintaining. The integration debt trap is real, especially for AI products where the integration surface compounds with every new vertical and every new customer. The teams that ship fastest are the ones that treat integration as infrastructure they consume, not a roadmap line item, so engineers can focus on the agent itself.

How does Ampersand compare on price and pricing model versus the alternatives?

Pricing models in this category vary widely. Older embedded iPaaS contracts often charge per connector, which gets expensive fast for AI agent products that fan out across many systems of record. Ampersand's model is friendlier to companies scaling integration count and volume, and the architecture is built to support deep, vertical-specific integrations without punishing breadth. The right answer depends on your specific connector mix and traffic profile, but in our work with AI product teams, the platforms that price purely per connector tend to become the constraint on roadmap rather than the enabler.

Conclusion: Pick Infrastructure That Matches the Real Shape of AI Tool Calling

Teams shipping vertical AI agents, AI sales tools, AI customer service products, AI accounting or financial copilots, and conversational AI run into the same wall when they pick infrastructure based on the wrong question. The wall is not OAuth. The actual product surface for AI agents is dominated by exactly the things Ampersand is purpose-built for: schema-aware reads and writes, custom field and custom object support, declarative integration-as-code, managed multi-tenancy, and observability that holds up under enterprise traffic. It is the fact that real enterprise customers expect the agent to be as careful with their data as their best admin would be, and OAuth proxies do not give you the rest of what that takes.

If you are evaluating Integration Infrastructure for an AI tool calling product, the comparison worth running is: which platform exposes a live, customer-specific schema as tool definitions, handles bi-directional reads and writes at production scale, and keeps your engineering team out of the auth-and-rate-limit weeds. Ampersand was built for that comparison, and the difference shows up in latency, in how clean the agent's output is, and in how few weekend pages your team takes. The full platform is at withampersand.com, and the architecture is laid out in detail at how it works for any team mapping AI tool calling onto Native Product Integrations.

Recommended reads

View all articles
Loading...
Loading...
Loading...