GPU Multitenancy in Kubernetes: Strategies, Challenges, and Best Practices
How to safely share expensive GPU infrastructure across teams without sacrificing performance or security
GPUs don't support native sharing between isolated processes. Learn four approaches for running multitenant GPU workloads at scale without performance hits.
Introducing vCluster Auto Nodes — Practical deep dive
Auto Nodes extend Private Nodes with provider-agnostic, automated node provisioning and scaling across clouds, on-prem, and bare metal.
Kubernetes makes pods elastic, but node scaling often breaks outside managed clouds. With vCluster Platform 4.4 + v0.28, Auto Nodes fix that gap, combining isolation, elasticity, and portability. Learn how Auto Nodes extend Private Nodes with automated provisioning and dynamic scaling across any environment.
Running Dedicated Clusters with vCluster: A Technical Deep Dive into Private Nodes
A technical walkthrough of Private Nodes in vCluster v0.27 and how they enable true single-tenant Kubernetes clusters.
Private Nodes in vCluster v0.27 take Kubernetes multi-tenancy to the next level by enabling fully isolated, dedicated clusters. In this deep dive, we walk through setup, benefits, and gotchas, from creating a vCluster with Private Nodes to joining worker nodes and deploying workloads. If you need stronger isolation, simpler lifecycle management, or enterprise-grade security, this guide covers how Private Nodes transform vCluster into a powerful single-tenant option without losing the flexibility of virtual clusters.
vCluster v0.27: Introducing Private Nodes for Dedicated Clusters
Dedicated, tenant‑owned nodes with a managed control plane, full isolation without running separate clusters.
Private Nodes complete vCluster’s tenancy spectrum: tenants connect their own nodes to a centrally managed control plane for full isolation, custom runtimes (CRI/CNI/CSI), and consistent performance, ideal for AI/ML, HPC, and regulated environments. Learn how it works and what’s next with Auto Nodes.
Three Tenancy Modes, One Platform: Rethinking Flexibility in Kubernetes Multi-Tenancy
Why covering the full Kubernetes tenancy spectrum is critical, and how Private Nodes bring stronger isolation to vCluster
In this blog, we explore why covering the full Kubernetes tenancy spectrum is essential, and how vCluster’s upcoming Private Nodes feature introduces stronger isolation for teams running production, regulated, or multi-tenant environments without giving up Kubernetes-native workflows.
Solving Kubernetes Multi-tenancy Challenges with vCluster
Unlocking Secure and Scalable Multi-Tenancy in Kubernetes with Virtual Clusters
Running multiple tenants on a single Kubernetes cluster can be complex and risky. In this post, Liquid Reply explores how vCluster offers a secure and cost-efficient solution by isolating workloads through lightweight virtual clusters.
NVIDIAScape: How vNode prevents this container breakout without the need for VMs
Container breakouts on GPU nodes are real, and just three lines of code can be enough. Discover how vNode neutralizes vulnerabilities like NVIDIAScape without relying on VMs.
NVIDIAScape (CVE-2025-23266) is a critical GPU-related vulnerability that allows attackers to break out of containers and gain root access. While some respond by layering in virtual machines, this blog walks through a better approach, how vNode uses container-native sandboxing to neutralize such attacks at the kernel level without sacrificing performance. Includes a step-by-step replication of the exploit, and a demo of how vNode prevents it.
Building and Testing Kubernetes Controllers: Why Shared Clusters Break Down
How shared clusters fall short, and why virtual clusters are the future of controller development.
Shared clusters are cost-effective, but when it comes to building and testing Kubernetes controllers, they create bottlenecks, from CRD conflicts to governance issues. This blog breaks down the trade-offs between shared, local, and dedicated clusters and introduces virtual clusters as the scalable solution for platform teams.
What Is GPU Sharing in Kubernetes?
How Kubernetes can make GPU usage more efficient for AI/ML teams through MPS, MIG, and smart scheduling.
As AI and ML workloads scale rapidly, GPUs have become essential, and expensive resources. But most teams underutilize them. This blog dives into how GPU sharing in Kubernetes can help platform teams increase efficiency, cut costs, and better support AI infrastructure.
Smarter Infrastructure for AI: Why Multi-Tenancy is a Climate Imperative
How virtual clusters and smarter tenancy models can reduce carbon impact while scaling AI workloads.
AI’s rapid growth is fueling a silent climate problem: idle infrastructure. This blog explores why multi-tenancy is key to scaling AI sustainably and how vCluster helps teams reduce waste while moving faster.
Automating Kubernetes Cleanup in CI Workflows
Keep your CI pipelines clean and efficient by automating Kubernetes resource cleanup with vCluster and Loft.
Leftover Kubernetes resources from CI jobs can drive up cloud costs and clutter your clusters. This guide shows how to automate cleanup tasks using vCluster, helping you maintain cleaner, faster CI/CD pipelines.
Automating Kubernetes Cleanup in CI Workflows
Keep your CI pipelines clean and efficient by automating Kubernetes resource cleanup with vCluster and Loft.
Leftover Kubernetes resources from CI jobs can drive up cloud costs and clutter your clusters. This guide shows how to automate cleanup tasks using vCluster, helping you maintain cleaner, faster CI/CD pipelines.
Bare Metal Kubernetes with GPU: Challenges and Multi-Tenancy Solutions
Why Namespace Isolation Falls Short for GPU Workloads, and How Multi-Tenancy with vCluster Solves It
Managing AI workloads on bare metal Kubernetes with GPUs presents unique challenges, from weak namespace isolation to underutilized resources and operational overhead. This blog explores the pitfalls of namespace-based multi-tenancy, why running a separate cluster per team is expensive, and how vCluster enables secure, efficient, and autonomous GPU sharing for AI teams.
How to Set Up a GPU-Enabled Kubernetes Cluster on GKE: Step-by-Step Guide for AI & ML Workloads
Step-by-step guide to setting up a GPU-enabled Kubernetes cluster on GKE for scalable AI and ML workloads.
Running AI or ML workloads on Kubernetes? This tutorial walks you through setting up a GPU enabled GKE cluster, from configuring GPU quotas and node pools to testing workloads and optimizing for multi-team GPU usage with vCluster.
Technical Guide: Using Spot Instances with vCluster for Significant Savings
Cut Kubernetes costs by up to 91% using spot instances and vCluster, without compromising workload stability.
Spot instances offer massive savings but come with unpredictability. In this step-by-step guide, learn how to combine them with vCluster to build resilient, cost-effective Kubernetes environments for CI/CD, AI/ML, and more.
Kubernetes v1.33: Key Features, Updates, and What You Need to Know
Octarine Unleashed: How Kubernetes 1.33 Changes Everything for Devs & Platform Teams
Kubernetes 1.33, codenamed "Octarine: The Color of Magic", lands with 64 feature updates. This blog dives into the most impactful ones and shows how to try them immediately with vCluster. From in-place pod resizing to ClusterTrustBundle, here’s what’s new.
What does your infrastructure look like in 2025 and beyond?
Why Moving from VMware to Kubernetes-native Infrastructure is Critical for Modern Enterprises
Discover why enterprises in 2025 are shifting from traditional VMware based virtual machines to modern, Kubernetes-native architectures. Learn how adopting Kubernetes closer to bare metal simplifies infrastructure, reduces costs, and enhances scalability and efficiency.
Seamless TLS with vCluster v0.22: Cert-Manager Integration Walkthrough
Introduction With the release of vCluster 0.22, managing TLS certificates for virtual clusters has become much easier with the integration of cert-manager. This new built in integration allows you to issue and manage certificates within virtual clusters while leveraging cert-mana...
Put your workloads to sleep with built in Sleep Mode Feature in vCluster
As Kubernetes usage grows, optimizing resource usage is more critical than ever. The Native Sleep Mode introduced in vCluster 0.22 provides a neat and easy solution for cost savings and efficient resource management. In this blog, we’ll dive deep into the functionality of Sleep M...
Kubernetes v1.32: Key Features, Updates, and What You Need to Know
The Kubernetes v1.32 release introduces significant advancements in resource management, security, scheduling, and observability. With a mix of 13 graduating features, 12 beta enhancements, and 19 got added in alpha, Kubernetes continues to refine its capabilities for production-...
Kubernetes HPA - Is it good to be used as is?
One of the most powerful features of Kubernetes is its ability to automatically scale components based on workload demands. This autoscaling capability includes several tools, such as the Horizontal Pod Autoscaler (HPA), Vertical Pod Autoscaler (VPA), and Cluster Autoscaler. In t...
The Complete Guide: What is a Control Plane in Kubernetes?
Managing Kubernetes at scale can quickly become overwhelming without a well-orchestrated control system. Kubernetes is a leading standard in the world of container orchestration. At the center of this platform is a Kubernetes control plane. It orchestrates and regulates the entir...
Kubernetes Ingress Controllers: The Ultimate Guide
Access to external resources is often needed when dealing with Kubernetes workloads. In some cases, you can manually configure a NodePort or LoadBalancer service. However, this approach has drawbacks, including manual overhead and the potential to introduce complex configurations...
Kubernetes v1.31: What’s New with Elli?
Kubernetes 1.31 has been released, bringing with it significant contributions. This large project continually adds new features during each release cycle, with new additions to alpha to reflect the ongoing innovation and optimization within Kubernetes. The 1.31 release team has i...
Kubernetes NGINX Ingress: 10 Useful Configuration Options
Modern-day applications have a lot of functionality that can be difficult to track and manage. Luckily, Ingress Controllers help handle these complexities. Ingress controllers implement ingress, a Kubernetes object that provides routing rules for external access to services in a ...
Kubernetes Manifests: Everything You Need to Know in 2024
Are you struggling to maintain scalability, reliability, and efficiency in your application deployments? Managing containerized applications allows platform engineers and DevOps team members to ensure scalable, reliable, and efficient application deployment, operation, and orches...
Multi-tenancy in Kubernetes: Comparing Isolation and Costs
Having multiple tenants sharing a Kubernetes cluster makes sense from a cost perspective, but what’s the overhead? How much should you invest to keep the tenant isolated, and how does it compare to running several clusters? Before examining the costs, let’s look at the scale of t...
Kubernetes Namespaces Don't Exist
Namespaces are one of the fundamental resources in Kubernetes. But they don’t provide network isolation, are ignored by the scheduler and can’t limit resource usage. Also, they are not real and don’t exist in the infrastructure. The previous statement is at odds with the followin...
Comparing Multi-tenancy Options in Kubernetes
Balancing isolation, management ease, and cost is critical in multi-tenant Kubernetes setups. In this article, we’ll explore how to evaluate these factors to optimize resource utilization and tenant isolation. A key question when planning infrastructure is: how many Kubernetes cl...
How to Host Separate Instances of Your SaaS Software in the Same Kubernetes Cluster
Deploy and manage isolated SaaS instances efficiently within a single Kubernetes cluster using Loft's vCluster technology.
Learn How to Host Separate Instances of Your SaaS Software in the Same Kubernetes Cluster