Skip to main content
Products

The private AI operations stack

Start with private model operations, then add the control surfaces around it: Loom for MCP and context coordination, MentatLab for DAG agent workflows over direct model/API calls, and fi-fhir for sensitive healthcare ETL.

Compare

FlexInfer is the runtime layer: private endpoints, model lifecycle, customization hooks, routing, and GPU operations.

Validate

Loom and MentatLab are the operator layers: MCP/context coordination, fleet state, and DAG workflow execution.

Deliver

fi-fhir is the proof path for sensitive data work: profile-driven healthcare integration in an AI-first development loop.

Choose Your Starting Lane

Pick the boundary you need to control

Use Platform for the operating story, then jump to the product page, docs, or playground that matches the boundary in front of you: runtime, context, orchestration, or integration.

Operating Story

From private model to production workflow

The pitch is not four unrelated tools. It is a path for teams that need private inference, governed agent access, repeatable orchestration, and a concrete sensitive-data use case.

01FlexInfer

Operate private model runtimes first

FlexInfer gives teams a Kubernetes-native control plane for model lifecycle, OpenAI-compatible routing, GPU-aware placement, scale-to-zero activation, quantization, adapters, and model artifact delivery.

Can this model run, scale, route, and roll back inside our cluster boundary?

FlexInfer product
02Loom

Govern the agent and tool boundary

Loom and Loom Core turn MCP sprawl into a single governed entrypoint with registry sync, daemon routing, tiered memory, fleet state, sandbox execution, RBAC, audit, and HUD visibility.

Which agents can reach which tools, with what context, and what audit trail?

Loom Core product
03MentatLab

Compose repeatable DAG workflows

MentatLab is the mission-control surface for DAG agent orchestration over direct API model nodes and internal services. In this stack, those model calls can target FlexInfer private endpoints instead of a frontier coding harness.

Can operators design, validate, run, and observe the workflow as a graph?

MentatLab product
04fi-fhir

Apply it to sensitive integration work

fi-fhir ties the platform story to healthcare ETL: Source Profiles, HL7v2/FHIR parsing, semantic events, workflow routing, terminology mapping, and AI-assisted explanations inside a controlled environment.

Can legacy clinical data become validated, explainable, routable events?

fi-fhir product
Inference Control PlaneLayer

FlexInfer

Private inference operations for teams that need model lifecycle, routing, GPU placement, quantization, adapters, and OpenAI-compatible endpoints inside their own boundary.

  • Run private model endpoints behind Kubernetes-native lifecycle and readiness controls
  • Customize with quantization pipelines, OCI model delivery, and LoRA adapter workflows
  • Route chat, completions, embeddings, and image flows through standard OpenAI-compatible APIs
Loom SuiteLayer

Loom

MCP and agent-fleet coordination for private operations: editor sync, single-entry tool access, daemon routing, tiered context, Weaver orchestration, skills, and workflow memory.

  • Single registry across Codex, Claude, Gemini, Zed, VS Code, Kilocode, and Loom Core
  • Agent sessions, tasks, handoffs, recall, compaction, and fleet state stay inspectable
  • RBAC, audit, sandbox execution, remote transport, HUD, and mobile surfaces keep MCP controlled
DAG OrchestrationLayer

MentatLab

Mission Control for DAG-based agent workflows that call direct API model nodes, internal services, MCP-governed tools, and private FlexInfer endpoints.

  • Design workflows as explicit graphs with dependency-aware execution plans
  • Use direct model/API calls instead of assuming a frontier coding harness is the runtime
  • Pair with Loom for context and policy, and FlexInfer for private model endpoints
Healthcare IntegrationLayer

fi-fhir

Healthcare ETL and integration tooling for Source Profiles, semantic events, workflow routing, terminology mapping, and HL7v2/FHIR development in an AI-first IDE loop.

  • Profile-driven ingestion across HL7v2, FHIR, CSV, EDI X12, and CDA/CCDA
  • Semantic event layer, FHIR output, and workflow DSL for routable transformations
  • AI-assisted warning explanations and workflow generation for integration development
Platform Hub

See how the products fit together

Platform walks through the stack as an operating model: private models first, governed agent access second, repeatable DAG workflows third, and healthcare integration as the concrete proof path.

Playground Hub

Validate workflows end-to-end

Use the playground when you need to validate configs, routes, or transformations without reading every doc page first.

Product Comparison

Boundary matrix

FlexInfer
Private inference operations and model customization
Model lifecycle, OpenAI-compatible proxy, GPU-aware placement, quantization, adapters, OCI delivery, and federation
Docs →Open →
Loom Suite
MCP, context, and agent-fleet coordination
Loom Core, Loom, editor sync, proxy + daemon runtime, memory, Weaver, HUD, RBAC, and sandboxed execution
Docs →Open →
MentatLab
DAG agent orchestration over direct model/API calls
Mission Control UI, flow schemas, DAG planning, run visibility, direct model nodes, and internal-service orchestration
Docs →Open →
fi-fhir
Sensitive healthcare ETL and integration development
Source Profiles, HL7v2/FHIR parsing, semantic events, workflow DSL, terminology mapping, and AI-assisted explanations
Docs →Open →
3

Core product surfaces: runtime control, context coordination, DAG orchestration, and sensitive-data integration.

2

Boundary types to validate before rollout: deployment and integration.

OSS + Services

Each lane has docs, playgrounds, implementation notes, or a product page so the pitch stays inspectable.

Architecture

Platform architecture

A tighter visual pass on the runtime and integration surfaces that matter most when evaluating the platform stack.

Loom Core control plane

Registry-driven client sync, loom proxy ingress, daemon routing, and HUD-backed operator surfaces for private MCP operations.

Loom Core control plane

Loom Core context management

Session-start recall, tiered memory, and prompt-pressure compaction keep agent context bounded and queryable.

Loom Core context management

Loom Core Weaver orchestration

FlexInfer-backed multi-domain routing with curated tool bundles and compound tools for reusable agent work.

Loom Core Weaver orchestration

FlexInfer control plane

OpenAI-compatible proxying, model lifecycle control, and GPU-aware execution inside a private cluster boundary.

FlexInfer control plane

FlexInfer runtime lifecycle

Artifact preparation, cache gating, quantization, and preload behavior that turns model intent into a ready endpoint.

FlexInfer runtime lifecycle

FlexInfer federation surface

GlobalProxy and FederatedModel coordinate cross-cluster reach while keeping runtime execution local to each cluster.

FlexInfer federation surface

Healthcare parsing pipeline

Profile-driven transformation flow for sensitive interoperability workloads.

Healthcare parsing pipeline