vCluster Blog

Tech Blog by vCluster

Subscribe to our newsletter

eu1

26230392

15314855-8a4c-4751-ad29-605050b96984

loft-sh

true

By subscribing you agree to our Privacy Policy and provide email consent.

How to: Exploring K8S on vCluster, Deploying an Observability stack - part 1

How to: Exploring K8S on vCluster, Deploying an Observability stack - part 1

Metrics, dashboards, alerting, and long-term storage, deployed on a multi-node Kubernetes cluster that runs entirely in Docker

Observability isn't optional, it's foundational. This guide walks through deploying Prometheus, Grafana, Thanos, and RustFS on a local multi-node Kubernetes cluster powered by vCluster. No cloud account required. No VMs. Just Docker and a few commands.

vCluster Labs Introduces vMetal to Manage Bare Metal AI Infrastructure for Neoclouds and AI Factories

vCluster Labs Introduces vMetal to Manage Bare Metal AI Infrastructure for Neoclouds and AI Factories

New machine management layer joins vCluster to deliver the industry’s first unified AI infrastructure stack, from bare metal to AI-ready environments.

vCluster Labs today introduced vMetal, a new bare metal machine management layer designed to help Neocloud providers and AI factories provision and operate GPU infrastructure at scale — automating everything from initial provisioning to upgrades and repurposing with cloud-like automation.

Polarise and vCluster Labs Announce Strategic Partnership to Accelerate Sovereign AI Infrastructure In Germany & Beyond

Polarise and vCluster Labs Announce Strategic Partnership to Accelerate Sovereign AI Infrastructure In Germany & Beyond

Partnership pairs Polarise's sovereign European GPU cloud with vCluster Labs' Kubernetes multi-tenancy platform to rapidly deliver fully isolated AI compute to enterprise customers.

Polarise and vCluster Labs announce a strategic partnership pairing sovereign European GPU cloud with Kubernetes virtual cluster technology to give mid-market enterprises rapid access to isolated, sovereign AI compute — no vendor lock-in, no data sovereignty trade-offs.

Introducing vMetal: Run Your GPU Data Center Like a Hyperscaler

Introducing vMetal: Run Your GPU Data Center Like a Hyperscaler

Most GPU data centers take months to operationalize. vMetal gets you from rack to running cluster in minutes.

Buying GPUs is the easy part. Operating them like a cloud platform is what separates Neoclouds that launch from ones that stall. vMetal automates the full bare metal lifecycle — discovery, provisioning, cluster attachment — so you can start delivering GPU capacity immediately.

Day 7: The vCluster Platform UI:  Managing vind Clusters Visually

Day 7: The vCluster Platform UI: Managing vind Clusters Visually

A web dashboard for all your local vind clusters with projects, team management, role-based access, and automation through access keys

The CLI is great for daily dev work. But when you need to see all your clusters at a glance, organize them into projects, manage team access, or hand off a demo environment, the vCluster Platform UI takes vind from a better KinD to a real cluster management platform. Day 7 of the 7 Days of vind series.

Day 6: Advanced Features: Sleep/Wake, Registry Proxy, and Custom Networking

Day 6: Advanced Features: Sleep/Wake, Registry Proxy, and Custom Networking

Pause clusters and resume where you left off, pull locally built images in 45ms, install Cilium, and map custom ports to your local cluster

With KinD, you delete your cluster at the end of the day and recreate it tomorrow. With vind, you pause it and resume exactly where you left off. Deployments, services, PVCs, all still there. Day 6 covers sleep/wake, registry proxy, custom CNI, and more.

Day 5: CI/CD with vind: The setup-vind GitHub Action

Day 5: CI/CD with vind: The setup-vind GitHub Action

Drop setup-kind from your GitHub Actions and get built-in registry proxy, automatic log export, and multi-cluster CI workflows with setup-vind

If you're using setup-kind in GitHub Actions, you're still loading images manually and missing automatic log exports. setup-vind is a drop-in replacement with built-in registry proxy, automatic artifact export, and multi-cluster support. Day 5 of the 7 Days of vind series.

Day 4: External Nodes: Joining a GCP Instance to Your Local vind Cluster

Day 4: External Nodes: Joining a GCP Instance to Your Local vind Cluster

Run your control plane locally in Docker, join a real cloud VM as a worker node over VPN, and schedule pods across both

Local Kubernetes tools stop at your laptop. vind doesn't. Join a GCP Compute Engine instance as a real worker node to your local cluster over an encrypted VPN tunnel. Test GPU workloads, mixed architectures, and hybrid setups. Day 4 of the 7 Days of vind series.

Day 3: Multi-Node vind Clusters: Real Scheduling, Real Node Drains

Day 3: Multi-Node vind Clusters: Real Scheduling, Real Node Drains

Create a multi-node local cluster in Docker and test pod distribution, node drains, affinity, and anti-affinity, just like production

Single-node clusters can't test scheduling. With vind, spin up a 4-node cluster in Docker, deploy across workers, drain nodes, and test affinity rules, real Kubernetes behavior on your laptop. Day 3 of the 7 Days of vind series.

Day 2: Getting Started with vind:  Your First Deployment with LoadBalancer

Day 2: Getting Started with vind: Your First Deployment with LoadBalancer

Install vind, create a local Kubernetes cluster, and deploy nginx with a working LoadBalancer — in under 3 minutes

KinD needs MetalLB for LoadBalancer services. vind has it built in. In Day 2 of the 7 Days of vind series, we walk through creating a cluster, deploying nginx, and hitting a real LoadBalancer IP, all running in Docker on your laptop.

Day 1: Introduction to vind:  Why I Replaced KinD with vCluster in Docker [vind]

Day 1: Introduction to vind: Why I Replaced KinD with vCluster in Docker [vind]

KinD works until it doesn't. vind picks up where it leaves off.

KinD works, until you need LoadBalancer services, multi-node setups, or the ability to pause and resume clusters. vind gives you a production-like local Kubernetes experience in Docker with features KinD simply doesn't have. Day 1 of the 7 Days of vind series.

How to Become an NVIDIA Exemplar Cloud - A Guide for NVIDIA Cloud Partners

How to Become an NVIDIA Exemplar Cloud - A Guide for NVIDIA Cloud Partners

Exemplar status isn't about who has the most H100s. It's about who can operate AI infrastructure with measurable, repeatable, production-grade excellence.

GPU capacity alone won't get you Exemplar status. NVIDIA is now rewarding operational maturity, reproducible benchmarking, efficient multi-tenancy, and scalable Kubernetes operations. Here's what NVIDIA Cloud Partners need to architect to qualify.

How to Build an Internal Kubernetes Platform (2026 Guide)

How to Build an Internal Kubernetes Platform (2026 Guide)

Empower your engineering teams with a self-service Kubernetes platform to streamline development workflows and enhance productivity.

Discover how to build an internal Kubernetes platform that enables developers to create and manage their own namespaces, reducing operational overhead and accelerating development cycles. Learn about the benefits of self-service environments, improved resource utilization, and enhanced developer autonomy.

When 37% of Cloud Environments Are Vulnerable, "Just Use VMs" Isn't Good Enough

When 37% of Cloud Environments Are Vulnerable, "Just Use VMs" Isn't Good Enough

How vNode delivers VM-level isolation for containerized AI workloads — without the VM overhead

A three-line Dockerfile broke container security. CVE-2025-23266 exposed 37% of cloud environments running AI workloads, giving attackers full root access to Kubernetes nodes. VMs are too heavy, gVisor can't catch it. vNode offers a third option: container-native isolation that's as strong as VMs but as lightweight as containers.

Reimagining Local Kubernetes: Replacing Kind with vind — A Deep Dive

Reimagining Local Kubernetes: Replacing Kind with vind — A Deep Dive

An open-source alternative to KinD with native LoadBalancer support, free UI, pull-through caching, and the ability to attach external nodes to your local cluster

Kubernetes developers have long relied on tools like KinD (Kubernetes in Docker) to spin up disposable clusters locally for development, testing, and CI/CD workflows. While KinD is a solid tool, it has limitations like not being able to use service type LoadBalancer, accessing homelab clusters from the web, or adding GPU nodes to your local cluster. Introducing vind (vCluster in Docker) - an open source alternative that enables Kubernetes clusters as first-class Docker containers, offering improved performance, modern features, and an enhanced developer experience.

Pragmatic Hybrid AI: Bursting Across Private GPUs and Public Cloud Without Leaking Data or Dollars

Pragmatic Hybrid AI: Bursting Across Private GPUs and Public Cloud Without Leaking Data or Dollars

Hybrid AI That Works: Network Isolation, Data Gravity, and Workload Placement in the Real World

For the past two years, the AI infrastructure debate has been framed as binary: go all-in on on-prem GPU estates or stay all-in on the cloud. Neither approach is sustainable at enterprise scale. The winning pattern is intelligent placement—keep sensitive or data-heavy jobs local, burst elastic workloads into the cloud. Success depends on strict isolation, careful placement, and scheduling that is cost-aware from the start.

Why the nodes/proxy Kubernetes RCE Does Not Apply to vCluster

Why the nodes/proxy Kubernetes RCE Does Not Apply to vCluster

How vCluster provides more security than vanilla Kubernetes when using nodes/proxy permissions for monitoring stacks

A security researcher recently disclosed that Kubernetes nodes/proxy permissions can be exploited for remote code execution. Kubernetes labeled it "working as intended" and issued no CVE. Since vCluster was mentioned in the disclosure, we investigated how this vulnerability affects our users. The conclusion: vCluster is not compromised and actually provides more security than vanilla Kubernetes when using features that require the nodes/proxy permission.

Launching vCluster Free - Get vCluster Enterprise Features at No Cost

Launching vCluster Free - Get vCluster Enterprise Features at No Cost

A free tier that makes advanced Kubernetes multi-tenancy accessible—without trials or sales gates.

We’re launching vCluster Free to make advanced Kubernetes multi-tenancy available to more builders.

Isolating Workloads in a Multi-Tenant GPU Cluster

Isolating Workloads in a Multi-Tenant GPU Cluster

Practical strategies for securing shared GPU environments with Kubernetes-native isolation, hardware partitioning, and operational best practices

Sharing GPU access across teams maximizes hardware ROI, but multitenant environments introduce critical performance and security challenges. This guide explores proven workload isolation strategies, from Kubernetes RBAC and network policies to NVIDIA MIG and time-slicing, that enable you to build secure, scalable GPU clusters. Learn how to prevent resource contention, enforce tenant boundaries, and implement operational safeguards that protect both workloads and data in production AI infrastructure.

Why a Tenancy Layer Belongs in the Kubernetes Tech Stack (in 2026)

Why a Tenancy Layer Belongs in the Kubernetes Tech Stack (in 2026)

Why AI workloads, GPU infrastructure, and platform scale are forcing Kubernetes teams to rethink multi-tenancy

AI workloads and GPU infrastructure are exposing the limits of traditional Kubernetes multi-tenancy. In 2026, platform teams need a new layer.

From GPU Cluster to AI Factory

From GPU Cluster to AI Factory

A 5-stage maturity model for evolving your GPU infrastructure from basic clusters to production-ready AI factories with secure multi-tenancy

Most enterprises start with basic GPU clusters and hit scaling walls fast. This guide walks through the 5-stage journey from manual provisioning to a fully automated AI factory, covering multi-tenancy, cost optimization, and how to achieve cloud-like operations on-prem without vendor lock-in.

Separate Clusters Aren’t as Secure as You Think — Lessons from a Cloud Platform Engineering Director

Separate Clusters Aren’t as Secure as You Think — Lessons from a Cloud Platform Engineering Director

Lessons in Intentional Tenancy and Security at Scale from a Cloud Platform Director

If a workload needs isolation, give it its own cluster. It sounds safe, but at scale, this logic breaks down. Learn why consistency, not separation, is the real security challenge in modern Kubernetes environments.

Deploying vClusters on a GPU-Enabled Kubernetes Cluster

Deploying vClusters on a GPU-Enabled Kubernetes Cluster

Streamline MLOps and Multi-tenancy by Running Isolated GPU Workloads with Virtual Clusters

Scale your AI infrastructure without the overhead. In this hands-on tutorial, we demonstrate how to use vCluster technology to virtualize GPU resources, ensuring secure multi-tenancy and efficient resource sharing for production-grade MLOps.

Solving GPU-Sharing Challenges with Virtual Clusters

Solving GPU-Sharing Challenges with Virtual Clusters

Why MPS and MIG fall short—and how virtual clusters deliver isolation without hardware lock-in

GPUs are expensive, but most organizations only achieve 30-50% utilization. The problem? GPUs weren't designed for sharing. Software solutions like MPS lack isolation. Hardware solutions like MIG lock you into specific vendors. vCluster takes a different approach—solving GPU multitenancy at the Kubernetes orchestration layer.

vCluster Ambassador program

vCluster Ambassador program 

Introducing the first vCluster Ambassadors shaping the future of Kubernetes multi-tenancy and platform engineering

Meet the first vCluster Ambassadors - community leaders and practitioners advancing Kubernetes multi-tenancy, platform engineering, and real-world developer platforms.

DIY GPU Sharing in Kubernetes

DIY GPU Sharing in Kubernetes

Explore time-slicing, MIG, and custom workarounds for sharing GPUs across Kubernetes workloads—plus their trade-offs for isolation and management.

GPUs are expensive and rarely used to full capacity. Learn the standard methods for sharing GPUs in Kubernetes—time-slicing and MIG—along with DIY alternatives, and discover why combining these techniques with vCluster delivers the isolation multi-tenant AI/ML workloads demand.

Architecting a Private Cloud for AI Workloads

Architecting a Private Cloud for AI Workloads

How to design, build, and operate a cost-effective private cloud infrastructure for enterprise AI at scale

Public clouds are convenient for AI experimentation, but production workloads often hit walls. For enterprises running continuous training and inference, a private cloud can deliver better ROI, data sovereignty, and performance. This comprehensive guide walks through architecting a private cloud for AI workloads from the ground up.

Demystifying Karpenter on GCP: The Complete Setup Guide

Demystifying Karpenter on GCP: The Complete Setup Guide

How to deploy Karpenter on Google Cloud Platform (and why you might want vCluster Auto Nodes instead)

Karpenter has become the gold standard for Kubernetes autoscaling on AWS, but what about GCP? This guide shows you how to set it up and introduces a better way.

GPU Multitenancy in Kubernetes: Strategies, Challenges, and Best Practices

GPU Multitenancy in Kubernetes: Strategies, Challenges, and Best Practices

How to safely share expensive GPU infrastructure across teams without sacrificing performance or security

GPUs don't support native sharing between isolated processes. Learn four approaches for running multitenant GPU workloads at scale without performance hits.

AI Infrastructure Isn’t Limited By GPUs. It’s Limited By Multi-Tenancy.

AI Infrastructure Isn’t Limited By GPUs. It’s Limited By Multi-Tenancy.

What the AI Infrastructure 2025 Survey Reveals, And How Platform Teams Can Respond

The latest AI Infrastructure 2025 survey shows that most organizations are struggling not due to GPU scarcity, but because of poor GPU utilization caused by limited multi-tenancy capabilities. Learn how virtual clusters and virtual nodes help platform teams solve high costs, sharing issues, and low operational maturity in Kubernetes environments.

KubeCon + CloudNativeCon North America 2025 Recap

KubeCon + CloudNativeCon North America 2025 Recap

Announcing the Infrastructure Tenancy Platform for NVIDIA DGX—plus what we learned from 100+ conversations at KubeCon about GPU efficiency, isolation, and the future of AI on Kubernetes.

KubeCon Atlanta 2025 was packed with energy, launches, and conversations that shaped the future of AI infrastructure. At Booth #421, we officially launched the Infrastructure Tenancy Platform for NVIDIA DGX—a Kubernetes-native platform designed to maximize GPU efficiency across private AI supercomputers, hyperscalers, and neoclouds. Here's what happened, what we announced, and why it matters for teams scaling AI workloads.

vCluster Labs Introduces Infrastructure Tenancy Platform for AI to Maximize NVIDIA GPU Efficiency on Kubernetes Environments

vCluster Labs Introduces Infrastructure Tenancy Platform for AI to Maximize NVIDIA GPU Efficiency on Kubernetes Environments

New platform combines virtual clusters, dynamic GPU autoscaling, and hybrid networking to maximize NVIDIA infrastructure efficiency—from DGX systems to bare metal and cloud.

vCluster Labs announced its Infrastructure Tenancy Platform for AI at KubeCon North America 2025, delivering a Kubernetes-native foundation for running AI workloads on NVIDIA GPU infrastructure. The platform introduces vCluster Private Nodes, Auto Nodes with Karpenter-based autoscaling, vCluster VPN for hybrid networking, and direct integrations with NVIDIA Base Command Manager, KubeVirt, and Netris—helping organizations maximize GPU utilization while maintaining full workload isolation and security across cloud, on-prem, and bare metal environments.

Virtual Clusters Are More Real Than Virtual

Virtual Clusters Are More Real Than Virtual

Why virtual clusters aren’t “fake” Kubernetes clusters — they’re the foundation of real multi-tenant platforms built for cloud, on-prem, and bare metal.

Many engineers think “virtual” means “not real.” In Kubernetes, the opposite is true. Virtual clusters run full control planes inside namespaces, giving teams real isolation, faster CI/CD, and multi-tenant efficiency across cloud and on-prem infrastructure with vCluster from Loft Labs.

Scaling Without Limits: The What, Why, and How of Cloud Bursting

Scaling Without Limits: The What, Why, and How of Cloud Bursting

A practical guide to implementing cloud bursting using vCluster VPN, Private Nodes, and Auto Nodes for secure, elastic, multi-cloud scalability.

Cloud bursting lets you expand compute capacity on demand without overprovisioning or re-architecting your systems. In this guide, we break down how vCluster VPN connects Private and Auto Nodes securely across environments—so you can scale beyond limits while keeping costs and complexity in check.

vCluster and Netris Partner to Bring Cloud-Grade Kubernetes to AI Factories & GPU Clouds With Strong Network Isolation Requirements

vCluster and Netris Partner to Bring Cloud-Grade Kubernetes to AI Factories & GPU Clouds With Strong Network Isolation Requirements

vCluster Labs and Netris team up to bring cloud-grade Kubernetes automation and network-level multi-tenancy to AI factories and GPU-powered infrastructure.

vCluster Labs has partnered with Netris to revolutionize how AI operators run Kubernetes on GPU infrastructure. By combining vCluster’s Kubernetes-level isolation with Netris’s network automation, the integration delivers a full-stack multi-tenancy solution, simplifying GPU cloud operations, maximizing utilization, and enabling cloud-grade performance anywhere AI runs.

Recapping The Future of Kubernetes Tenancy Launch Series

Recapping The Future of Kubernetes Tenancy Launch Series

How vCluster’s Private Nodes, Auto Nodes, and Standalone releases redefine multi-tenancy for modern Kubernetes platforms.

From hardware-isolated clusters to dynamic autoscaling and fully standalone control planes, vCluster’s latest launch series completes the future of Kubernetes multi-tenancy. Discover how Private Nodes, Auto Nodes, and Standalone unlock new levels of performance, security, and flexibility for platform teams worldwide.

Bootstrapping Kubernetes from Scratch with vCluster Standalone: An End-to-End Walkthrough

Bootstrapping Kubernetes from Scratch with vCluster Standalone: An End-to-End Walkthrough

Bootstrapping Kubernetes from scratch, no host cluster, no external dependencies.

Kubernetes multi-tenancy just got simpler. With vCluster Standalone, you can bootstrap a full Kubernetes control plane directly on bare metal or VMs, no host cluster required. This walkthrough shows how to install, join worker nodes, and run virtual clusters on a single lightweight foundation, reducing vendor dependencies and setup complexity for platform and infrastructure teams.

GPU on Kubernetes: Safe Upgrades, Flexible Multitenancy

GPU on Kubernetes: Safe Upgrades, Flexible Multitenancy

How vCluster and NVIDIA’s KAI Scheduler reshape GPU workload management in Kubernetes - enabling isolation, safety, and maximum utilization.

GPU workloads have become the backbone of modern AI infrastructure, but managing and upgrading GPU schedulers in Kubernetes remains risky and complex. This post explores how vCluster and NVIDIA’s KAI Scheduler together enable fractional GPU allocation, isolated scheduler testing, and multi-team autonomy, helping organizations innovate faster while keeping production safe.

A New Foundation for Multi-Tenancy: Introducing vCluster Standalone

A New Foundation for Multi-Tenancy: Introducing vCluster Standalone

Eliminating the “Cluster 1 problem” with vCluster Standalone v0.29 – the unified foundation for Kubernetes multi-tenancy on bare metal, VMs, and cloud.

vCluster Standalone changes the Kubernetes tenancy spectrum by removing the need for external host clusters. With direct bare metal and VM bootstrapping, teams gain full control, stronger isolation, and vendor-supported simplicity. Explore how vCluster Standalone (v0.29) solves the “Cluster 1 problem” while supporting Shared, Private, and Auto Nodes for any workload.

Introducing vCluster Auto Nodes — Practical deep dive

Introducing vCluster Auto Nodes — Practical deep dive

Auto Nodes extend Private Nodes with provider-agnostic, automated node provisioning and scaling across clouds, on-prem, and bare metal.

Kubernetes makes pods elastic, but node scaling often breaks outside managed clouds. With vCluster Platform 4.4 + v0.28, Auto Nodes fix that gap, combining isolation, elasticity, and portability. Learn how Auto Nodes extend Private Nodes with automated provisioning and dynamic scaling across any environment.

Introducing vCluster Auto Nodes: Karpenter-Based Dynamic Autoscaling Anywhere

Introducing vCluster Auto Nodes: Karpenter-Based Dynamic Autoscaling Anywhere

Dynamic, isolated, and cloud-agnostic autoscaling for every virtual cluster.

vCluster Auto Nodes brings dynamic, Karpenter-powered autoscaling to any environment, public cloud, private cloud, or bare metal. Combined with Private Nodes, it delivers true isolation and elasticity for Kubernetes, letting every virtual cluster scale independently without cloud-specific limits.

How vCluster Auto Nodes Delivers Dynamic Kubernetes Scaling Across Any Infrastructure

How vCluster Auto Nodes Delivers Dynamic Kubernetes Scaling Across Any Infrastructure

Kubernetes pods scale elastically, but node scaling often stops at the provider boundary. Auto Nodes extend Private Nodes to bring elasticity and portability to isolated clusters across clouds, private datacenters, and bare metal.

Pods autoscale in Kubernetes, but nodes don’t. Outside managed services, teams fall back on brittle scripts or costly overprovisioning. With vCluster Platform 4.4 + vCluster v0.28, Auto Nodes close the gap, bringing automated provisioning and elastic scaling to isolated clusters across clouds, private datacenters, and bare metal.

The Case for Portable Autoscaling

The Case for Portable Autoscaling

Kubernetes has pods and deployments covered, but when it comes to nodes, scaling breaks down across clouds, providers, and private infrastructure. Auto Nodes change that.

Kubernetes makes workloads elastic until you hit the node layer. Managed services offer partial fixes, but hybrid and isolated environments still face scaling gaps and wasted resources. vCluster Auto Nodes close this gap by combining isolation, just-in-time elasticity, and environment-agnostic portability.

Running Dedicated Clusters with vCluster: A Technical Deep Dive into Private Nodes

Running Dedicated Clusters with vCluster: A Technical Deep Dive into Private Nodes

A technical walkthrough of Private Nodes in vCluster v0.27 and how they enable true single-tenant Kubernetes clusters.

Private Nodes in vCluster v0.27 take Kubernetes multi-tenancy to the next level by enabling fully isolated, dedicated clusters. In this deep dive, we walk through setup, benefits, and gotchas, from creating a vCluster with Private Nodes to joining worker nodes and deploying workloads. If you need stronger isolation, simpler lifecycle management, or enterprise-grade security, this guide covers how Private Nodes transform vCluster into a powerful single-tenant option without losing the flexibility of virtual clusters.

We’re Now vCluster Labs

We’re Now vCluster Labs

A new name, the same mission, building the best Kubernetes tenancy tools for teams everywhere.

Loft Labs is now vCluster Labs, a name that reflects our focus on building the best Kubernetes multi-tenancy and infrastructure engineering tools. The same team, projects, and mission remain, but with a clearer brand aligned to our product, vCluster.

vCluster v0.27: Introducing Private Nodes for Dedicated Clusters

vCluster v0.27: Introducing Private Nodes for Dedicated Clusters

Dedicated, tenant‑owned nodes with a managed control plane, full isolation without running separate clusters.

Private Nodes complete vCluster’s tenancy spectrum: tenants connect their own nodes to a centrally managed control plane for full isolation, custom runtimes (CRI/CNI/CSI), and consistent performance, ideal for AI/ML, HPC, and regulated environments. Learn how it works and what’s next with Auto Nodes.

How to Scale Kubernetes Without etcd Sharding

How to Scale Kubernetes Without etcd Sharding

Rethinking Kubernetes scale: avoid the risks of etcd sharding with virtual clusters built for performance, stability, and multi-tenant environments.

Is your Kubernetes cluster slowing down under load? etcd doesn’t scale well with multi-tenancy or 30k+ objects. This blog shows how virtual clusters offer an easier, safer way to isolate tenants and scale your control plane, no sharding required.

Three Tenancy Modes, One Platform: Rethinking Flexibility in Kubernetes Multi-Tenancy

Three Tenancy Modes, One Platform: Rethinking Flexibility in Kubernetes Multi-Tenancy

Why covering the full Kubernetes tenancy spectrum is critical, and how Private Nodes bring stronger isolation to vCluster

In this blog, we explore why covering the full Kubernetes tenancy spectrum is essential, and how vCluster’s upcoming Private Nodes feature introduces stronger isolation for teams running production, regulated, or multi-tenant environments without giving up Kubernetes-native workflows.

Scaling Kubernetes Without the Pain of etcd Sharding

Scaling Kubernetes Without the Pain of etcd Sharding

Why sharding etcd doesn’t scale, and how virtual clusters eliminate control plane bottlenecks in large Kubernetes environments.

OpenAI’s outage revealed what happens when etcd breaks at scale. This post explains why sharding isn’t enough, and how vCluster offloads API load with virtual control planes. Benchmark included.

vCluster: The Performance Paradox – How Virtual Clusters Save Millions Without Sacrificing Speed

vCluster: The Performance Paradox – How Virtual Clusters Save Millions Without Sacrificing Speed

How vCluster Balances Kubernetes Cost Reduction With Real-World Performance

Can you really save millions on Kubernetes infrastructure without compromising performance? Yes, with vCluster. In this blog, we break down how virtual clusters reduce control plane overhead, unlock higher node utilization, and simplify multi-tenancy, all while maintaining lightning-fast performance.

5 Must-See KubeCon + CloudNativeCon India 2025 Sessions

5 Must-See KubeCon + CloudNativeCon India 2025 Sessions

A curated list of impactful, technical, and thought-provoking sessions to catch at KubeCon + CloudNativeCon India 2025 in Hyderabad.

KubeCon + CloudNativeCon India 2025 is back in Hyderabad on August 6–7! With so many exciting sessions, it can be hard to choose. Here are 5 standout talks you shouldn't miss, from real-world Kubernetes meltdowns to scaling GitOps at Expedia, and even why Kubernetes is moving to NFTables.

Solving Kubernetes Multi-tenancy Challenges with vCluster

Solving Kubernetes Multi-tenancy Challenges with vCluster

Unlocking Secure and Scalable Multi-Tenancy in Kubernetes with Virtual Clusters

Running multiple tenants on a single Kubernetes cluster can be complex and risky. In this post, Liquid Reply explores how vCluster offers a secure and cost-efficient solution by isolating workloads through lightweight virtual clusters.

NVIDIAScape: How vNode prevents this container breakout without the need for VMs

NVIDIAScape: How vNode prevents this container breakout without the need for VMs

Container breakouts on GPU nodes are real, and just three lines of code can be enough. Discover how vNode neutralizes vulnerabilities like NVIDIAScape without relying on VMs.

NVIDIAScape (CVE-2025-23266) is a critical GPU-related vulnerability that allows attackers to break out of containers and gain root access. While some respond by layering in virtual machines, this blog walks through a better approach, how vNode uses container-native sandboxing to neutralize such attacks at the kernel level without sacrificing performance. Includes a step-by-step replication of the exploit, and a demo of how vNode prevents it.

Building and Testing Kubernetes Controllers: Why Shared Clusters Break Down

Building and Testing Kubernetes Controllers: Why Shared Clusters Break Down

How shared clusters fall short, and why virtual clusters are the future of controller development.

Shared clusters are cost-effective, but when it comes to building and testing Kubernetes controllers, they create bottlenecks, from CRD conflicts to governance issues. This blog breaks down the trade-offs between shared, local, and dedicated clusters and introduces virtual clusters as the scalable solution for platform teams.

What Is GPU Sharing in Kubernetes?

What Is GPU Sharing in Kubernetes?

How Kubernetes can make GPU usage more efficient for AI/ML teams through MPS, MIG, and smart scheduling.

As AI and ML workloads scale rapidly, GPUs have become essential, and expensive resources. But most teams underutilize them. This blog dives into how GPU sharing in Kubernetes can help platform teams increase efficiency, cut costs, and better support AI infrastructure.

Smarter Infrastructure for AI: Why Multi-Tenancy is a Climate Imperative

Smarter Infrastructure for AI: Why Multi-Tenancy is a Climate Imperative

How virtual clusters and smarter tenancy models can reduce carbon impact while scaling AI workloads.

AI’s rapid growth is fueling a silent climate problem: idle infrastructure. This blog explores why multi-tenancy is key to scaling AI sustainably and how vCluster helps teams reduce waste while moving faster.

Automating Kubernetes Cleanup in CI Workflows

Automating Kubernetes Cleanup in CI Workflows

Keep your CI pipelines clean and efficient by automating Kubernetes resource cleanup with vCluster and Loft.

Leftover Kubernetes resources from CI jobs can drive up cloud costs and clutter your clusters. This guide shows how to automate cleanup tasks using vCluster, helping you maintain cleaner, faster CI/CD pipelines.

Automating Kubernetes Cleanup in CI Workflows

Automating Kubernetes Cleanup in CI Workflows

Keep your CI pipelines clean and efficient by automating Kubernetes resource cleanup with vCluster and Loft.

Leftover Kubernetes resources from CI jobs can drive up cloud costs and clutter your clusters. This guide shows how to automate cleanup tasks using vCluster, helping you maintain cleaner, faster CI/CD pipelines.

Bare Metal Kubernetes with GPU: Challenges and Multi-Tenancy Solutions

Bare Metal Kubernetes with GPU: Challenges and Multi-Tenancy Solutions

Why Namespace Isolation Falls Short for GPU Workloads, and How Multi-Tenancy with vCluster Solves It

Managing AI workloads on bare metal Kubernetes with GPUs presents unique challenges, from weak namespace isolation to underutilized resources and operational overhead. This blog explores the pitfalls of namespace-based multi-tenancy, why running a separate cluster per team is expensive, and how vCluster enables secure, efficient, and autonomous GPU sharing for AI teams.

How to Set Up a GPU-Enabled Kubernetes Cluster on GKE: Step-by-Step Guide for AI & ML Workloads

How to Set Up a GPU-Enabled Kubernetes Cluster on GKE: Step-by-Step Guide for AI & ML Workloads

Step-by-step guide to setting up a GPU-enabled Kubernetes cluster on GKE for scalable AI and ML workloads.

Running AI or ML workloads on Kubernetes? This tutorial walks you through setting up a GPU enabled GKE cluster, from configuring GPU quotas and node pools to testing workloads and optimizing for multi-team GPU usage with vCluster.

Ready to take vCluster for a spin?

Deploy your first virtual cluster today.