pith. machine review for the scientific record. sign in

arxiv: 2604.21413 · v1 · submitted 2026-04-23 · 💻 cs.DB

Recognition: unknown

An Alternate Agentic AI Architecture (It's About the Data)

Authors on Pith no claims yet

Pith reviewed 2026-05-08 13:16 UTC · model grok-4.3

classification 💻 cs.DB
keywords agentic AIdata integrationquery algebraenterprise systemsLLM agentsdatabase wrappersAQLmediation
0
0 comments X

The pith

Enterprise AI needs explicit query algebras and source wrappers to solve data integration instead of relying on LLM agents.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper argues that enterprises struggle with agentic AI not because of weak reasoning in language models but because critical data lives scattered across governed systems with different schemas and controls. It presents RUBICON as an architecture that replaces opaque agent orchestration with AQL, a compact query algebra of Find, From, and Where operators run through source-specific wrappers. These wrappers align schemas, enforce access rules, normalize results, and expose every intermediate step for inspection. Sympathetic readers would care because the method promises the reach of multi-source agentic search together with the traceability, determinism, and compliance that enterprise environments demand.

Core claim

Enterprises do not suffer from a reasoning deficit but from a data integration problem. By reintroducing explicit query structure, wrapper-based mediation, and cost-based optimization, the architecture obtains the breadth of agentic search while preserving traceability, determinism, and trust in enterprise environments.

What carries the argument

AQL (Agentic Query Language), a small explicit query algebra using Find, From, and Where operators that decomposes questions into structured plans executed through source-specific wrappers enforcing access control, schema alignment, and result normalization.

If this is right

  • Complex questions decompose into structured, auditable query plans instead of hidden chains of LLM calls.
  • All intermediate results stay visible and inspectable at every step.
  • Cost-based optimization techniques from databases can select efficient execution paths across sources.
  • Access controls and governance policies are enforced uniformly by the wrappers.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • This design could let enterprises reuse existing database query optimizers and monitoring tools for AI workloads.
  • It suggests that many reliability issues in current agentic systems trace to data handling rather than model scale.
  • Direct performance comparisons on standardized enterprise query sets would test whether the algebra loses expressiveness relative to unrestricted tool calling.

Load-bearing premise

A small explicit query algebra executed through source-specific wrappers can decompose and resolve complex enterprise questions at the scale and flexibility of current LLM agent systems without loss of capability or introduction of new integration overhead.

What would settle it

A benchmark test in which RUBICON cannot express or resolve a multi-source enterprise question that requires dynamic source selection beyond its fixed algebra, while an LLM-based agent succeeds.

Figures

Figures reproduced from arXiv: 2604.21413 by \c{C}agatay Demiralp, Fabian Wenz, Felix Treutwein, Kai Arenja, Michael Stonebraker.

Figure 1
Figure 1. Figure 1: AQL-Based Query Processing Architecture ( view at source ↗
Figure 2
Figure 2. Figure 2: Tool-call composition per query. Each query shows three stacked bars (LangChain+NL, LangChain+AQL, view at source ↗
read the original abstract

For the last several years, the dominant narrative in "agentic AI" has been that large language models should orchestrate information access by dynamically selecting tools, issuing sub-queries, and synthesizing results. We argue this approach is misguided: enterprises do not suffer from a reasoning deficit, but from a data integration problem. Enterprises are data-centric: critical information is scattered across heterogeneous systems (e.g., databases, documents, and external services), each with its own query language, schema, access controls, and performance constraints. In contrast, contemporary LLM-based architectures are optimized for reasoning over unstructured text and treat enterprise systems as either corpora or external tools invoked by a black-box component. This creates a mismatch between schema-rich, governed, performance-critical data systems and text-centric, probabilistic LLM architectures, leading to limited transparency, weak correctness guarantees, and unpredictable performance. In this paper, we present RUBICON, an alternative architecture grounded in data management principles. Instead of delegating orchestration to an opaque agent, we introduce AQL (Agentic Query Language), a small, explicit query algebra - Find, From, and Where - executed through source-specific wrappers that enforce access control, schema alignment, and result normalization. All intermediate results are visible and inspectable. Complex questions are decomposed into structured, auditable query plans rather than hidden chains of LLM calls. Our thesis is simple: enterprise AI is not a prompt engineering problem; it is a systems problem. By reintroducing explicit query structure, wrapper-based mediation, and cost-based optimization, we obtain the breadth of agentic search while preserving traceability, determinism, and trust in enterprise environments.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

2 major / 2 minor

Summary. The paper argues that enterprise AI challenges arise from data integration problems across heterogeneous systems rather than LLM reasoning deficits. It proposes the RUBICON architecture, which uses AQL—a minimal explicit query algebra consisting of Find, From, and Where operators—executed via source-specific wrappers to enforce access controls, schema alignment, and result normalization, thereby achieving agentic breadth with improved transparency, determinism, and auditability.

Significance. If the proposed architecture can be shown to scale without capability loss, it would represent a meaningful contribution to database systems and enterprise AI by reapplying established principles of query mediation and optimization to agentic workflows. The emphasis on inspectable query plans over opaque LLM chains aligns with core DB strengths in traceability and cost-based planning, but the conceptual nature without any empirical grounding or examples limits its current significance to a position statement.

major comments (2)
  1. [Abstract] The central thesis that AQL plus wrappers can decompose complex enterprise questions while preserving the breadth of agentic search is load-bearing, yet the manuscript provides no worked examples of such decomposition, no argument for the algebra's completeness with respect to typical tasks (e.g., multi-source joins, conditional aggregation, or dynamic access control), and no description of how query planning itself is performed without reintroducing hidden orchestration. This appears in the description of RUBICON and AQL.
  2. The claim of 'cost-based optimization' and 'determinism' is asserted without any formalization of the optimizer, cost model, or execution semantics for the wrapper layer, making it impossible to evaluate whether the architecture actually avoids the unpredictability it criticizes in LLM agents.
minor comments (2)
  1. [Abstract] The abstract and description introduce RUBICON and AQL without defining their scope or relation to existing mediator/wrapper architectures in the database literature (e.g., no citations to classic data integration work).
  2. Terminology such as 'agentic search' and 'breadth of agentic search' is used without precise definition, which reduces clarity when contrasting with LLM-based approaches.

Simulated Author's Rebuttal

2 responses · 0 unresolved

We thank the referee for the constructive and detailed review. We agree that the manuscript is conceptual in nature and would benefit from additional examples and formal details to better support its claims. We address each major comment below and indicate the revisions we will make.

read point-by-point responses
  1. Referee: [Abstract] The central thesis that AQL plus wrappers can decompose complex enterprise questions while preserving the breadth of agentic search is load-bearing, yet the manuscript provides no worked examples of such decomposition, no argument for the algebra's completeness with respect to typical tasks (e.g., multi-source joins, conditional aggregation, or dynamic access control), and no description of how query planning itself is performed without reintroducing hidden orchestration. This appears in the description of RUBICON and AQL.

    Authors: We acknowledge the absence of concrete examples and formal arguments in the current draft. In the revised manuscript we will add a dedicated section with worked examples of query decomposition, including a multi-source compliance query that combines relational joins, document retrieval, and conditional filtering. We will also provide an argument for AQL completeness by showing how the three operators can be composed to express multi-source joins (via multiple From clauses with correlation), conditional aggregation (via Where predicates that invoke source-native aggregate functions), and dynamic access control (enforced inside each wrapper before result normalization). Regarding query planning, the architecture explicitly separates a deterministic planner—which uses source metadata and a cost model to produce an AQL expression—from any LLM involvement; the LLM is used only for initial natural-language to AQL translation when needed, after which all orchestration is explicit and inspectable. We will clarify this separation in the revision. revision: yes

  2. Referee: The claim of 'cost-based optimization' and 'determinism' is asserted without any formalization of the optimizer, cost model, or execution semantics for the wrapper layer, making it impossible to evaluate whether the architecture actually avoids the unpredictability it criticizes in LLM agents.

    Authors: This observation is correct for the present position paper. The manuscript currently states these properties at a high level without formal definitions. In the revision we will add a preliminary formalization of wrapper execution semantics (including result normalization and access-control enforcement) and sketch a basic cost model that accounts for source-specific latency, data volume, and transfer costs. A full optimizer implementation and empirical demonstration of determinism, however, lie beyond the scope of this conceptual work and are noted as future research. revision: partial

Circularity Check

0 steps flagged

No circularity; conceptual proposal without derivations or self-referential reductions

full rationale

The paper presents RUBICON as an architectural alternative grounded in data management principles, introducing AQL (Find/From/Where) executed via wrappers to address integration issues in enterprise settings. No equations, fitted parameters, predictions, or derivation chains exist that could reduce to inputs by construction. The argument contrasts LLM agent practices with explicit query structure and mediation without invoking self-citations in a load-bearing manner or smuggling ansatzes. The central thesis is a design proposal relying on established database concepts (e.g., wrappers, cost-based optimization) rather than any self-referential loop, rendering the presentation self-contained.

Axiom & Free-Parameter Ledger

0 free parameters · 1 axioms · 2 invented entities

The central claim rests on a domain assumption about enterprise problems and the introduction of two new named components without prior empirical grounding or implementation evidence.

axioms (1)
  • domain assumption Enterprises primarily face data integration challenges rather than deficits in AI reasoning capabilities.
    This premise is stated directly as the core thesis in the abstract.
invented entities (2)
  • RUBICON no independent evidence
    purpose: Alternative agentic AI architecture grounded in data management principles.
    Newly proposed system that organizes the described components.
  • AQL no independent evidence
    purpose: Agentic Query Language with Find, From, and Where operations for structured, auditable query plans.
    New query algebra introduced to replace LLM orchestration.

pith-pipeline@v0.9.0 · 5613 in / 1401 out tokens · 40177 ms · 2026-05-08T13:16:15.672672+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

21 extracted references · 2 canonical work pages · 1 internal anchor

  1. [1]

    Spider: A Large-Scale Human-Labeled Dataset for Complex and Cross-Domain Semantic Parsing and Text-to-SQL Task

    Tao Yu, Rui Zhang, Kai Yang, Michihiro Yasunaga, Dongxu Wang, Zifan Li, James Ma, Irene Li, Qingning Yao, Shanelle Roman, et al. Spider: A large-scale human-labeled dataset for complex and cross-domain semantic parsing and text-to-sql task.arXiv preprint arXiv:1809.08887, 2018

  2. [2]

    Can llm already serve as a database interface? a big bench for large-scale database grounded text-to-sqls.Advances in Neural Information Processing Systems, 36, 2024

    Jinyang Li, Binyuan Hui, Ge Qu, Jiaxi Yang, Binhua Li, Bowen Li, Bailin Wang, Bowen Qin, Ruiying Geng, Nan Huo, et al. Can llm already serve as a database interface? a big bench for large-scale database grounded text-to-sqls.Advances in Neural Information Processing Systems, 36, 2024

  3. [3]

    Benchpress: A human-in-the-loop annotation system for rapid text-to-sql benchmark curation

    Fabian Wenz, Omar Bouattour, Devin Yang, Justin Choi, Cecil Gregg, Nesime Tatbul, and Çagatay Demiralp. Benchpress: A human-in-the-loop annotation system for rapid text-to-sql benchmark curation. In16th Conference on Innovative Data Systems Research, CIDR 2026, Chaminade, CA, USA, January 18-21, 2026. www.cidrdb.org, 2026

  4. [4]

    Cafarella, Çagatay Demiralp, and Michael Stonebraker

    Peter Baile Chen, Fabian Wenz, Yi Zhang, Moe Kayali, Nesime Tatbul, Michael J. Cafarella, Çagatay Demiralp, and Michael Stonebraker. BEA VER: an enterprise benchmark for text-to-sql.CoRR, abs/2409.02038, 2024

  5. [5]

    Microsoft 365 copilot: Reinventing productivity with ai

    Microsoft. Microsoft 365 copilot: Reinventing productivity with ai. https://www.microsoft.com/en-us/ microsoft-365/copilot, 2023. Accessed: 2026-02-25

  6. [6]

    Glean: Work ai for the enterprise

    Glean Technologies. Glean: Work ai for the enterprise. https://www.glean.com, 2023. Accessed: 2026-02-25

  7. [7]

    Openclaw: Open source agent framework

    OpenCLAW Contributors. Openclaw: Open source agent framework. https://github.com/openclaw, 2024. Project documentation

  8. [8]

    Data integration: A theoretical perspective

    Maurizio Lenzerini. Data integration: A theoretical perspective. In Lucian Popa, Serge Abiteboul, and Phokion G. Kolaitis, editors,Proceedings of the Twenty-first ACM SIGACT-SIGMOD-SIGART Symposium on Principles of Database Systems, June 3-5, Madison, Wisconsin, USA, pages 233–246. ACM, 2002

  9. [9]

    Magesh Jayapandian and H. V . Jagadish. Automated creation of a forms-based database query interface.Proc. VLDB Endow., 1(1):695–709, 2008

  10. [10]

    Moshé M. Zloof. Query-by-example: the invocation and definition of tables and forms. In Douglas S. Kerr, editor, Proceedings of the International Conference on Very Large Data Bases, September 22-24, 1975, Framingham, Massachusetts, USA, pages 1–24. ACM, 1975

  11. [11]

    You only look once: Unified, real-time object detection, 2016

    Joseph Redmon, Santosh Divvala, Ross Girshick, and Ali Farhadi. You only look once: Unified, real-time object detection, 2016

  12. [12]

    Term-weighting approaches in automatic text retrieval.Inf

    Gerard Salton and Christopher Buckley. Term-weighting approaches in automatic text retrieval.Inf. Process. Manage., 24(5):513–523, August 1988

  13. [13]

    Efficient estimation of word representations in vector space

    Tomás Mikolov, Kai Chen, Greg Corrado, and Jeffrey Dean. Efficient estimation of word representations in vector space. In Yoshua Bengio and Yann LeCun, editors,1st International Conference on Learning Representations, ICLR 2013, Scottsdale, Arizona, USA, May 2-4, 2013, Workshop Track Proceedings, 2013

  14. [14]

    Gmail.https://workspace.google.com/products/gmail/, 2024

    Google. Gmail.https://workspace.google.com/products/gmail/, 2024. Accessed: 2026-02-27

  15. [15]

    Gpt-5 mini

    OpenAI. Gpt-5 mini. https://platform.openai.com/docs/models, 2025. Model documentation, Accessed: 2026-02-27

  16. [16]

    Gemini 3 flash (preview)

    Google DeepMind. Gemini 3 flash (preview). https://ai.google.dev/, 2025. Preview model documentation, Accessed: 2026-02-27. 14 APREPRINT- APRIL24, 2026

  17. [17]

    Claude sonnet 4.6

    Anthropic. Claude sonnet 4.6. https://www.anthropic.com/claude, 2025. Model documentation, Accessed: 2026-02-27

  18. [18]

    Narasimhan, and Yuan Cao

    Shunyu Yao, Jeffrey Zhao, Dian Yu, Nan Du, Izhak Shafran, Karthik R. Narasimhan, and Yuan Cao. React: Synergizing reasoning and acting in language models. InThe Eleventh International Conference on Learning Representations, ICLR 2023, Kigali, Rwanda, May 1-5, 2023. OpenReview.net, 2023

  19. [19]

    Langchain

    Harrison Chase. Langchain. https://github.com/langchain-ai/langchain, 2022. Accessed: 2026-02-27

  20. [20]

    React agent documentation

    LangChain. React agent documentation. https://python.langchain.com/docs/modules/agents/agent_ types/react, 2024. Accessed: 2026-02-27

  21. [21]

    The genai divide: State of ai in business 2025

    Aditya Challapally, Chris Pease, Ramesh Raskar, Pradyumna Chari, and MIT NANDA. The genai divide: State of ai in business 2025. Technical report, Project NANDA, Massachusetts Institute of Technology, July 2025. 15