Architecture

How a request flows
through Infino.

At the core of Infino is an agentic search engine and agent runtime with identity, permissions, and audit. Built in Rust from the ground up.

Fig. 01 / Request path
ARCHITECTURE / REQUEST PATHevery call enters the same boundaryRequestAuthenticationAgentsAuthorizationINGESTED DATACoreDBMetadataEmbedderRerankerObject Store
The path

Same boundary, every call.

Identity first, permissions enforced before data access, every step recorded.

01

Authenticate the caller

A request enters through the UI or API. The Authentication layer verifies identity and establishes the request context.

02

Route through the agent runtime

The Agents service holds session state, scratch memory, and the audit context that follows the request through the rest of the stack.

03

Authorize execution

Authorization enforces RBAC, scope, and policy against the plan before anything can reach storage or remote systems.

04

Resolve against data

Authorized requests resolve against Infino's storage stack — full-text, vector, and SQL on one copy of your data.

Components

What each part does.

Every component shares one identity model, one audit trail, and one storage layer.

01

Authentication

Verifies identity for humans, agents, and services, then establishes the request context for everything downstream.

02

Agents

Agent runtime that holds session state, scratch memory, and the audit context attached to every downstream call.

03

Authorization

Applies RBAC, scope, and policy to the request plan, then dispatches to the storage layer.

04

CoreDB

Append-only, time-aware engine for raw records, embeddings, and context across vector, full-text, relational, and time-series.

05

Object Store

Cost-efficient persistence for raw data and frozen segments, decoupled from compute so retention and query scale independently.

06

Metadata Store

Tracks derived schema, type history, lineage, and segment metadata so every field change remains queryable.

07

Embedding & Reranker

Models running inside the engine for embedding and ranking without external API calls or version drift.

Works with existing stacks

One copy
of your data.

A custom search engine that executes Elasticsearch queries on Parquet files so a single copy of your data runs full-text, vector, and SQL at object-storage scale.

No vector store

Vectors live inside the Parquet files themselves. No Pinecone, no pgvector, no embedding sync job to babysit.

No search cluster

Full-text indexes are embedded in Parquet alongside the data. No Elasticsearch fleet to scale and re-index.

No ETL fan-out

Nothing to mirror, replicate, or reconcile. The Parquet file is the index, the vector store, and the table.

Object-storage scale

S3, GCS, Azure Blob. Storage and compute scale independently — agents stay cheap as data grows.

Open format

Still Parquet. Other tools can read it. You aren't locked into a proprietary store to query your own data.

One governed surface

Identity, permissions, and audit apply once — to the one copy — instead of being re-implemented per system.

Next

See it on your data.