Latest blog post

The content layer problem in Financial AI agents

March 2026Bakul Kakadiya, VP Solutions Engineer
unknown alt

Every investment bank, asset manager, and advisory firm is building GenAI agents. The architecture patterns are converging - ReAct loops, tool-calling LLMs, multi-agent hierarchies. The frameworks are maturing - LangChain, CrewAI, AutoGen, Google A2A.

But if you’ve actually tried to put one of these agents into production at a financial institution, you’ve hit the same wall everyone else has: where does the agent get its market intelligence, and how do you give it access without leaking proprietary context?

This post is about that problem, not a specific vendor pitch, but the architectural decisions you’ll face when connecting an AI agent to both your internal data and external financial content. I’ll walk through three integration patterns we’ve evaluated, the security and compliance tradeoffs that don’t show up in demos, and where the existing data terminal APIs fall short.

The core idea: one agent, two worlds of data

Any production financial agent needs two categories of tools. Internal tools query your proprietary data: SQL against portfolio databases for holdings, positions, and P&L; semantic search over analyst notes and strategy memos. External tools reach out for market intelligence: real-time news, earnings transcripts, SEC filings, entity resolution, and sentiment signals.

The internal side is relatively straightforward - your data, your infrastructure, your access controls. The external side is where the architectural decisions get interesting, because you’re now routing agent queries through a third-party boundary.

Why not just wrap your terminal API?

The obvious question: Known financial data providers already have APIs. Why not wrap those as agent tools and call it done?

You can, and for some use cases you should. But terminal APIs were built for human-driven workflows - a person constructing a specific query with known parameters. AI agents operate differently. They need entity resolution (“which NVIDIA do you mean?”), they issue exploratory queries that may dead-end, and they chain multiple calls in ways that interact poorly with per-seat licensing models. Most terminal APIs also return data shaped for display, not for LLM consumption - HTML tables, proprietary field codes, nested structures that burn tokens without adding signal.

AI-native content layers - platforms like Bigdata.com, among others entering this space - are built differently. They return structured, citation-rich payloads optimized for tool-calling agents. They handle entity resolution upstream. They price for agent-scale query volumes rather than human seat counts. Whether this architectural difference justifies adding another vendor to your stack depends on your specific query patterns and licensing constraints, but it’s a real difference worth evaluating.

Three integration patterns, one platform

Pattern 1: Internal-First Hierarchical Escalation

This is the pattern that will survive your design review, and it should be the default. The agent queries internal sources first - your portfolio database, and your research documents. Only when internal data is genuinely insufficient does it escalate to an external content layer.

Why this matters: most portfolio questions can be answered in milliseconds from your own database. You don’t need a market intelligence API to tell a PM what their current exposure is. The external call - which costs money, adds latency, and crosses a trust boundary - should only fire when the question genuinely requires outside data. This also gives your compliance team a clear audit trail: every external query was triggered by a specific gap in internal knowledge, not a blanket fan-out.

Pattern 2: Direct API Integration

For teams that need precise control over what gets queried and how, direct REST API integration gives you the most explicit contract. You define each tool wrapper yourself - entity lookup, news search, filing retrieval - with your own caching, retry logic, and rate limiting. Entity lookups get cached; queries include exponential backoff for transient failures.

There is more code to maintain, but it’s also more flexible and auditable. Every external call is a function your team wrote, reviewed, and can instrument. For regulated environments, that explicitness has real value.

Pattern 3: Protocol-Level Integration (MCP)

The Model Context Protocol (MCP) offers a single connection that auto-discovers available tools from the provider. Connect once, and your agent can call company lookup, news search, tearsheets, and whatever new capabilities the provider ships in the future - no code changes required.

This sounds great, and for development and prototyping it genuinely is. But in a regulated production environment, “tools appear automatically” is a risk, not just a feature. If your content provider ships a new endpoint on Tuesday, do you want your production agent calling it on Wednesday without any review, testing, or approval? Probably not.

The practical solution: use MCP for rapid prototyping and development, but pin or allowlist specific tool versions in production. Most MCP adapters support this. You get the convenience of protocol-level discovery in dev and the control of explicit tool registration in prod.

What a real workflow looks like

Conference demos tend to show a single mega-query that does everything at once. In practice, portfolio managers don’t work that way. Here’s a more realistic sequence:

A PM asks: “Has there been any negative news on our semiconductor holdings in the last week?”

The agent first queries the internal portfolio database to resolve “our semiconductor holdings” into a list of tickers. It then sends a clean, entity-scoped query to the external content layer - just the tickers, no position sizes or cost basis - with a sentiment filter for negative coverage. The response comes back with cited sources, and the agent cross-references against internal analyst notes to check whether any of the flagged items were already on the research team’s radar.

This is narrow, contextual, and incremental - the way real investment workflows actually operate. The PM might follow up with a deeper question on one specific holding, which triggers a more expensive research query. The internal-first pattern keeps the common case fast and cheap.

The bigger picture

The financial industry is moving past the chatbot phase of GenAI. The real value unlock is agentic systems, that reason across structured and unstructured data and produce audit-ready outputs.

But deploying agents in finance without solving the content layer problem is like building a high-performance engine with no fuel line. The reasoning capability matters. The data access layer (with proper security boundaries, citation provenance, graceful degradation, and cost controls) matters equally.

The question is no longer whether your firm will build GenAI agents. It’s whether you’ve thought hard enough about the boundary between your proprietary data and the external intelligence those agents need to be useful.

Note: This post uses Bigdata.com as a concrete example of an AI-native content layer. A full open-source cookbook demonstrating all three integration patterns is available on GitHub. The architectural considerations discussed here apply regardless of which external content provider you evaluate.

Bakul is the VP Solutions Engineer at RavenPack. He leads technical engagement and solutions architecture, specializing in bridging the gap between complex client needs and advanced technology. Bakul has more than 20 years of experience in the IT industry, with a strong focus on software architecture and financial technology. Prior to RavenPack, he served as a Lead Software Architect at S&P Global for over 15 years, where he managed the architecture of regulatory and compliance products.

Bakul holds the AWS Certified Solutions Architect - Professional and SnowPro Core certifications. Bakul is a trusted technology advisor with deep expertise in Cloud Architecture, Generative AI development, MCP, and Large Language Models (LLM), including the implementation of RAG and model fine-tuning.

unknown alt
shape black left
shape black right