Infrastructure tutorials
Production-grade guides for Linux, servers, security and performance. Copy-paste commands, multi-distro support, written by engineers who run this in production.
Browse by topic
Linux
System administration, shell scripting, package management
Hosting & Servers
Web servers, reverse proxies, SSL, domains
Security
Firewalls, hardening, encryption, access control
Performance
Caching, optimization, profiling, load testing
Databases
MySQL, PostgreSQL, Redis, backups, replication
Networking
DNS, load balancing, VPN, TCP/IP, routing
DevOps
CI/CD, Docker, Kubernetes, automation
Monitoring
Logging, alerting, metrics, observability
Most viewed
Install and configure CockroachDB cluster with high availability and distributed SQL
databasesConfigure network interface monitoring with ICMP ping and connectivity testing
networkingConfigure Linux system time synchronization with chrony and NTP hardening
linuxInstall and configure PostgreSQL 17 with performance tuning and security hardening
databasesInstall and configure ArgoCD for GitOps continuous deployment with RBAC and SSL
devopsRecently published
Configure TimescaleDB backup and recovery with pgBackRest for automated PostgreSQL protection
databasesConfigure Apache Airflow DAG security and secrets management with RBAC policies and encryption
securityConfigure Falco runtime security for Kubernetes threat detection with eBPF monitoring
securityImplement Consul Connect mTLS with Vault PKI backend for secure service mesh communication
securityConfigure Podman secrets management with HashiCorp Vault integration
devopsInstall and configure Traefik reverse proxy with SSL automation
Set up Traefik as a reverse proxy with Docker Compose for automatic SSL certificate management, service discovery, and load balancing across multiple backend services.
Install and configure Podman for rootless containers on Linux
Learn to install Podman and configure rootless containers as a secure Docker alternative. Includes Docker Compose migration, systemd integration, and troubleshooting common permission issues.
Install and configure Ollama for local AI models on Linux servers
Set up Ollama to run large language models locally on your Linux server. This tutorial covers installation, GPU acceleration, model deployment, API configuration, and performance optimization.
Need help?
Don't want to manage this yourself?
We handle infrastructure for businesses that depend on uptime. From initial setup to ongoing operations.
Talk to an engineer