Infrastructure tutorials
Production-grade guides for Linux, servers, security and performance. Copy-paste commands, multi-distro support, written by engineers who run this in production.
Browse by topic
Linux
System administration, shell scripting, package management
Hosting & Servers
Web servers, reverse proxies, SSL, domains
Security
Firewalls, hardening, encryption, access control
Performance
Caching, optimization, profiling, load testing
Databases
MySQL, PostgreSQL, Redis, backups, replication
Networking
DNS, load balancing, VPN, TCP/IP, routing
DevOps
CI/CD, Docker, Kubernetes, automation
Monitoring
Logging, alerting, metrics, observability
Most viewed
Configure Linux system time synchronization with chrony and NTP hardening
linuxInstall and configure CockroachDB cluster with high availability and distributed SQL
databasesInstall and configure PostgreSQL 17 with performance tuning and security hardening
databasesConfigure network interface monitoring with ICMP ping and connectivity testing
networkingInstall and configure ArgoCD for GitOps continuous deployment with RBAC and SSL
devopsRecently published
Implement Kubernetes workload rightsizing with VPA recommendations and cost analysis
devopsConfigure Kubernetes cluster autoscaler with mixed instance types for cost optimization
devopsConfigure Apache Airflow DAG performance optimization best practices
devopsImplement WireGuard multi-site mesh networking with automatic routing and failover
networkingSet up Kubernetes custom metrics autoscaling with Prometheus adapter for application-specific scaling
devopsImplement Kubernetes workload rightsizing with VPA recommendations and cost analysis
Set up Vertical Pod Autoscaler to automatically optimize resource requests and limits for your Kubernetes workloads. Create cost analysis dashboards to track resource utilization and identify opportunities for rightsizing containers in production clusters.
Configure Kubernetes cluster autoscaler with mixed instance types for cost optimization
Set up Kubernetes cluster autoscaler 1.30 with mixed instance types and spot instances to automatically scale nodes based on demand while minimizing infrastructure costs through intelligent instance selection and workload optimization.
Set up Kubernetes custom metrics autoscaling with Prometheus adapter for application-specific scaling
Configure Prometheus adapter to expose custom application metrics to Kubernetes Horizontal Pod Autoscaler for intelligent scaling based on business metrics like queue depth, response time, and user load instead of basic CPU/memory usage.
Configure Kubernetes resource quotas and limit ranges for namespace-level resource management
Set up Kubernetes resource quotas and limit ranges to control CPU, memory, and storage consumption at the namespace level. This tutorial covers implementing resource constraints, monitoring usage, and troubleshooting quota issues for multi-tenant cluster management.
Deploy applications to Kubernetes with Helm charts and production best practices
Learn how to create production-ready Helm charts for Kubernetes deployments with proper templating, values management, security configurations, and environment-specific customizations for scalable application orchestration.
Implement Kubernetes cluster autoscaler for automatic node scaling
Configure Kubernetes cluster autoscaler to automatically add and remove worker nodes based on pod resource demands. This tutorial covers cloud provider integration, scaling policies, and monitoring for production-grade horizontal scaling.
Configure Kubernetes pod disruption budgets for high availability with policy enforcement
Learn to configure PodDisruptionBudget resources in Kubernetes to maintain application availability during voluntary disruptions. This tutorial covers creating disruption budgets, implementing policies for different workload types, and monitoring disruption events with kubectl.
Configure Kubernetes horizontal pod autoscaler for dynamic scaling based on resource metrics
Set up HPA with CPU and memory targets for automatic pod scaling. Configure metrics server and Prometheus adapter for custom metrics monitoring. Enable dynamic workload scaling based on resource utilization.
Implement Kubernetes cluster autoscaling with Helm charts and KEDA for dynamic workload scaling
Configure comprehensive Kubernetes autoscaling with cluster autoscaler for node management, KEDA for event-driven pod scaling, and vertical pod autoscaler for resource optimization. This tutorial covers production-grade deployment using Helm charts with monitoring and optimization strategies.
Need help?
Don't want to manage this yourself?
We handle infrastructure for businesses that depend on uptime. From initial setup to ongoing operations.
Talk to an engineer