Skip to main content
Private AI Platform
FlexInfer inference architecture

FlexInfer

Private AI Platform for Sensitive Workloads

On-prem and hybrid AI operations for sensitive workloads: inference control with FlexInfer, context orchestration with Loom Core, and healthcare-aligned integration workflows with fi-fhir.

Start with platform posture, then move into playground and docs for implementation details.

  • On-Prem
  • Hybrid
  • Kubernetes
  • MCP Controls
  • Healthcare Integrations
  • Observability
FlexInfer inference routing architecture
Platform posture

Private AI Platform for Sensitive Workloads

FlexInfer focuses on operational readiness in private environments: repeatable deployment controls, healthcare-aligned integration tooling, and a roadmap for enterprise context governance.

Runtime and inference control plane

FlexInfer

Kubernetes-native routing, deployment controls, and operational workflows for predictable on-prem or hybrid inference.

Deployment: Model runtime placement, rollout safety, and capacity controls inside your cluster boundary.

Integration: Inference APIs and platform operations remain inside your network and observability stack.

Context and orchestration control plane

Loom Core

Registry-driven MCP config generation, daemon routing, and enterprise context controls roadmap (gateway, RBAC, executor).

Deployment: Centralizes MCP server lifecycle and policy boundaries for internal tools and agent access.

Integration: Defines how agents reach internal systems with auditable routing and least-privilege intent.

Sensitive-data integration plane

fi-fhir

Healthcare-focused ingestion and transformation workflows (HL7v2 to FHIR) with profile-driven, testable data handling.

Deployment: Data transformation pipeline runs in your controlled environment and deployment topology.

Integration: Profile-driven mapping and validation isolate source variability while preserving operational traceability.

Core Stack

Loom suite, FlexInfer, fi-fhir, and MentatLab

Product pages cover lane fit. Docs and playground cover execution. MentatLab adds the orchestration UI surface for DAG run control, while Loom Core continues to govern context and policy boundaries.

Operational surface

MentatLab mission control

Loom Core governs context routing and policy boundaries; MentatLab provides the operator UX for DAG design and run visibility.

Core

Loom Core

CLI + daemon for registry-driven MCP configuration, server orchestration, and lifecycle management.

Core

Loom

VS Code extension for multi-platform MCP config sync, skills registry, and agent context workflows.

Core

Loom Zed

Zed integration for Loom workflows: consistent servers, shared registry, and repeatable profiles.

Core

FlexInfer

Kubernetes-native inference routing: ship GPU workloads with predictable configuration and operations.

Core

fi-fhir

Healthcare integration tooling: Source Profiles, parsing pipeline, and HL7v2 to FHIR workflows.

Enterprise capabilities

Roadmap status: currently building gateway, RBAC, and executor

Loom Core enterprise controls are in progress. Operational foundations are available today and already documented.

In progress

MCP Gateway
In progress

Centralized MCP routing layer for enterprise tool aggregation and policy enforcement.

Role-based Access Control (RBAC)
In progress

Role-aware permissions for MCP tool access across teams and environments.

Dev-box Container Executor on K8s
In progress

Containerized execution path for development and automation workloads in Kubernetes.

Available now

Operational Foundations
Available

Production patterns already available: observability, workflow docs, and deployment controls.

Provides the baseline operating model while enterprise controls are being rolled out.

Start Here

Build, validate, and reference the same schemas

Playground pages are designed to match the docs: config editor, schema validation, and entrypoints into Loom Core workflows.

Define Platform Boundaries

Start with deployment and integration boundaries, then map the right product lane.

Validate Operations

Use docs and playground workflows to validate config, routing, and integration behavior.

Need Build Support

Use the consulting path for architecture hardening in private environments.

Consulting

Bring it into production

If you want this stack inside your environment, the fastest path is a scoped audit or build engagement.

Readiness Audit

Fixed-scope diagnostic for sensitive workloads: architecture review, risk register, deployment constraints, and a 90-day plan.

Cluster Build

Build private or hybrid inference infrastructure with observability, GitOps, and safer rollout controls.

Fractional Lead

Hands-on senior guidance to align teams around an executable platform roadmap.