Skip to content
Skip to documentation content
Browse documentation
In progress Page being written

GPU setup

Verify your cluster's GPU stack and prepare it for LLMKube. Covers GKE, EKS, and self-managed clusters running the NVIDIA GPU Operator.

What this page will cover

  • Confirming GPU nodes are present and the device plugin is healthy.
  • Reading nvidia.com/gpu allocatable counts and matching them to your InferenceService requests.
  • When to set spec.runtimeClassName: nvidia (clusters where NVIDIA isn't the default runtime).
  • Validating the install with a tiny GPU smoke-test workload before deploying a real model.
Read the source on GitHub
LLMKube LLMKube

Kubernetes for Local LLMs. Deploy, manage, and scale AI inference workloads with production-grade orchestration.

© 2026 Defilan Technologies LLC

Community

Built for the Kubernetes and AI communities

LLMKube is not affiliated with or endorsed by the Cloud Native Computing Foundation or the Kubernetes project. Kubernetes® is a registered trademark of The Linux Foundation.