Black, yellow, and pink illustrated graphicBlack, yellow, and pink illustrated graphic

Deliver AI platforms and applications quickly and easily

Mirantis k0rdent AI’s built-in AI PaaS layer integrates with the GPU PaaS and builds on Mirantis k0rdent Enterprise core functionality. It’s a unified platform for defining, deploying, and lifecycle managing AI/ML development, testing, and application hosting environments on Kubernetes, on bare metal, in clouds, and/or out to the edge.

Leveraging AI PaaS, cloud service providers (CSPs) are using k0rdent to swiftly engineer and deliver value-added AI services to customers. Enterprises are leveraging the same functionality to speed innovation: rolling out ready-to-use training and inference platforms to data scientists, data engineers, and developers, so they can innovate quickly and safely, without friction.

Move fast without risk: template-driven operations speed innovation, cut setup from months to days, and get new services online quickly.

Innovate without friction: assemble pre-validated, template-defined open source AI and k0rdent ecosystem partner-provided components into bespoke solutions quickly, with minimum skills required.

Unified lifecycle control: manage Kubernetes clusters and AI services in one platform, across bare metal, private, or hyperscaler clouds.

Secure and compliant: control where data and models reside and how tenants and users connect with them. Easily isolate tenants up and down the full stack. Automatically enforce policies everywhere from a single source of truth.

Operator friendly and self-service ready: configurable web UIs and catalogs for creating and consuming services impose guardrails while eliminating bottlenecks, letting your whole organization move faster with AI.

Observable and billable: built-in observability and fine-grained FinOps help track, allocate, and optimize performance, utilization, cost, and maximize upsides. 

AI PaaS Use Cases

TURNKEY TRAINING
TURNKEY INFERENCE
SELF-SERVICE PORTAL

Turnkey Training for AI Factories on Kubernetes

Stand up governed, reusable training factories fast.

Turnkey Training in Mirantis k0rdent AI lets teams spin up approved stacks for data prep, notebooks, distributed training, evaluation, and promotion—tying model registry, lineage, and live telemetry into a continuous improvement loop. GPU-aware orchestration drives throughput; policy-as-code, audit trails, and multi-tenancy keep work compliant and secure; built-in observability and FinOps connect usage and cost to projects and models.

Neoclouds

Productize training workbenches: Publish curated templates (e.g., KubeRay, Slurm/Soperator, MLflow, model registry) so customers can fine-tune and train quickly.

Close the factory loop: Feed inference telemetry, quality, and cost signals back into data selection and evaluation to improve models each cycle.

Hit performance and cost targets: flexible, GPU-aware orchestration lets you serve more tenants with the same hardware and ensure that SLOs and cost objectives are met.

Monetize with confidence: Quotas/SLAs, per-hour or outcome-aligned pricing, and billing integrations turn commodity GPU rental into higher-margin services.


Enterprises

Accelerate from experiment to production: Self-service, governed environments connect to approved data, track lineage, and promote models through gated stages.

Operate safely at scale: Canary/A-B testing, rollback, and drift/latency telemetry feed targeted retraining; multi-tenancy protects teams on shared clusters.

Unify legacy & modern tooling: Run VM-dependent tools alongside containerized services under one Kubernetes-native framework.

Prove value & ensure compliance: Policy-as-code, audit logging, and per-project cost allocation provide accountability for leaders and regulators.

Stack of documents titled "Mirantis AI Factory Reference Architecture" on a pink background.Stack of documents titled "Mirantis AI Factory Reference Architecture" on a pink background.

EXECUTIVE BRIEF: Mirantis AI Factory Reference Architecture

Understand the role of the AI Factory and what’s inside a production-grade implementation.


VIEW NOW

Turnkey Inference: Configure and lifecycle manage complete inference service stacks

Launch governed, scalable inference in minutes.

Turnkey Inference uses Mirantis k0rdent AI’s PaaS layer to stand up full AI serving platforms across data center, cloud, and edge. 

Platform engineers can assemble inference solutions from a fast-growing catalog of operations frameworks (e.g., Run.ai, KubeRay, Gcore and others), model servers (e.g. vLLM, Triton, KServe, RayServe, etc.), and adjunct components (e.g., vector DBs for RAG). They can wrap in observability and cost/billing analytics, define policies for geolocating data and models and routing traffic (Smart Routing).

Teams can then self-serve, build, and operate AI solutions within a fully-governed, business-ready framework.

Neoclouds

Productize differentiated, value-added services: Innovate quickly. Publish catalog templates (model servers, embeddings, vector stores, caching) as commercial offerings with quotas and SLAs.

Hit performance, latency, and cost targets: GPU-aware orchestration and topology management maps application requirements and traffic to capacity flexibly, ensuring SLOs are met.

Bill with confidence: Built-in metering and tenant attribution enable token/request-based billing and help you tune for profitability.

Keep tenants safe and compliant: k0rdent delivers hard multi-tenancy, policy enforcement, and supports Zero Trust up and down the stack. AI PaaS adds model lineage, promotion gates, MCP-based context governance and other security and compliance features.

Enterprises

Ship faster, safely: Self-service, pre-approved stacks let teams access approved models, document stores, RAG databases, access control and routing schemas, and promote endpoints to production with consistent guardrails.

Operate reliably: Declarative rollouts with canary/A/B and easy rollback standardize MLOps at scale.

See and control spend: Per-model observability and FinOps tie usage, performance, and cost to apps and teams.

Illustration of robotic arms placing circuit-patterned boxes on a conveyor belt with servers in the background, in shades of blue.Illustration of robotic arms placing circuit-patterned boxes on a conveyor belt with servers in the background, in shades of blue.

BLOG: AI Factories: What Are They and Who Needs Them?


VIEW NOW

Self-Service Portal: Productize AI services with click-to-provision marketplaces

Launch branded, governed AI portals in minutes.

Mirantis k0rdent AI’s PaaS layer lets you stand up a branded marketplace (external or internal) where users discover services, view transparent pricing, and provision GPU, storage, and AI components with one click. Metering, billing, and cost controls are built in; policy guardrails, quotas, and approvals keep environments compliant. Unified observability provides real-time GPU utilization, performance, and health to resolve issues proactively and optimize spend.

Neoclouds

Monetize faster: Publish catalog offers (models, embeddings, vector stores, gateways) with tiers, quotas, and SLAs; eliminate sales friction with instant sign-up and automated invoicing.

Operate efficiently: Real-time utilization and health views drive capacity planning; GPU-aware placement protects latency and profitability.

Govern with confidence: Enforce tenant isolation, policy-as-code, and approval workflows across all services.


Enterprises

Unblock teams safely: Internal marketplace enables governed self-service for GPUs, storage, and AI stacks—reducing ticket queues and shadow IT.

Control cost & compliance: Fine-grained metering, budgets, and quotas tie usage to projects; policy guardrails and approvals maintain security and regulatory posture.

Reduce platform toil: Self-service and automation replace repetitive provisioning so platform teams focus on strategic work.

Interface for configuring a DEV AWS Cluster. Includes fields for cluster name, worker nodes, and options for email notification and Grafana registration.Interface for configuring a DEV AWS Cluster. Includes fields for cluster name, worker nodes, and options for email notification and Grafana registration.

Streamline the production of new cloud products with Product Builder — no code needed.


BOOK A DEMO

LET’S TALK

Contact us to learn how Mirantis can accelerate your cloud initiatives.

We see Mirantis as a strategic partner who can help us provide higher performance and greater success as we expand our cloud computing services internationally.

— Aurelio Forese, Head of Cloud, Netsons

image

We see Mirantis as a strategic partner who can help us provide higher performance and greater success as we expand our cloud computing services internationally.

— Aurelio Forese, Head of Cloud, Netsons

image