Skip to main content

Self-Host Netdata: Real-Time Server Monitoring 2026

·OSSAlt Team
netdatamonitoringself-hostingdockerdevops2026
Share:

TL;DR

Netdata is a real-time server monitoring agent with ~73K GitHub stars. Install in 60 seconds, zero configuration required — it auto-discovers 800+ services and starts collecting per-second metrics immediately. Unlike Prometheus+Grafana, there's no separate database or dashboard to configure. It's the fastest way to get server visibility. Use Netdata for real-time monitoring (per-second resolution); use Prometheus+Grafana for long-term retention and custom queries.

Key Takeaways

  • 73K GitHub stars, one of the most-starred monitoring tools
  • Per-second resolution: Prometheus defaults to 15s scrape intervals; Netdata collects every second
  • 800+ auto-discovered integrations: Nginx, Postgres, Redis, Docker, MySQL, Node.js apps — zero config
  • ~1% CPU, ~70MB RAM agent overhead — minimal impact on monitored server
  • Netdata Parent-Child: Aggregate metrics from multiple servers to one central Netdata instance
  • Machine learning: Anomaly detection built-in — alerts when behavior deviates from baseline

Netdata vs Prometheus + Grafana

FeatureNetdataPrometheus + Grafana
Setup time~1 minute~30 minutes
Configuration requiredNoneModerate
Metric resolutionPer second (1s)15s default
Auto-discovery800+ integrationsManual scrape config
DashboardsAuto-generatedManual (or imported)
Long-term retentionLimited (free cloud)Configurable (local)
Custom PromQL queriesVia integrationNative
Memory usage~70MB~500MB+
ML anomaly detection✅ Built-in❌ (external)
Best forReal-time operationsLong-term analysis

Use both: Many teams run Netdata for immediate visibility (spot problems within seconds) and Prometheus+Grafana for trend analysis and dashboards.


Install in 60 Seconds

Method 1: One-Line Install (Native)

# Installs Netdata agent directly on the OS:
bash <(curl -Ss https://my-netdata.io/kickstart.sh)

# Netdata starts immediately at:
# http://your-server-ip:19999

That's it. No config files. Netdata starts monitoring everything automatically.

Method 2: Docker

docker run -d --name=netdata \
  --pid=host \
  --network=host \
  --restart=unless-stopped \
  -v netdataconfig:/etc/netdata \
  -v netdatalib:/var/lib/netdata \
  -v netdatacache:/var/cache/netdata \
  -v /etc/passwd:/host/etc/passwd:ro \
  -v /etc/group:/host/etc/group:ro \
  -v /proc:/host/proc:ro \
  -v /sys:/host/sys:ro \
  -v /etc/os-release:/host/etc/os-release:ro \
  -e NETDATA_CLAIM_TOKEN=your-cloud-token \  # Optional: connect to Netdata Cloud
  --cap-add SYS_PTRACE \
  --security-opt apparmor=unconfined \
  netdata/netdata

Method 3: Docker Compose

# docker-compose.yml
version: '3.8'

services:
  netdata:
    image: netdata/netdata:latest
    container_name: netdata
    pid: host
    network_mode: host
    restart: unless-stopped
    cap_add:
      - SYS_PTRACE
      - SYS_ADMIN
    security_opt:
      - apparmor=unconfined
    volumes:
      - netdataconfig:/etc/netdata
      - netdatalib:/var/lib/netdata
      - netdatacache:/var/cache/netdata
      - /etc/passwd:/host/etc/passwd:ro
      - /etc/group:/host/etc/group:ro
      - /proc:/host/proc:ro
      - /sys:/host/sys:ro
      - /etc/os-release:/host/etc/os-release:ro
      - /var/run/docker.sock:/var/run/docker.sock:ro
    environment:
      - NETDATA_CLAIM_TOKEN=${NETDATA_CLAIM_TOKEN}  # Optional cloud token
      - NETDATA_CLAIM_URL=https://app.netdata.cloud
      - NETDATA_CLAIM_ROOMS=${NETDATA_CLAIM_ROOMS}

volumes:
  netdataconfig:
  netdatalib:
  netdatacache:

What Netdata Auto-Discovers

After installation, visit http://your-server:19999 — you'll see dashboards for everything running on the server:

System:

  • CPU per core, idle, user, system, iowait, softirq
  • Memory: used, cached, buffers, swap, NUMA
  • Disk: read/write IOPS, utilization, backlog, latency
  • Network: bandwidth per interface, packets, errors

Applications (auto-detected):

  • Web servers: Nginx, Apache, Caddy (request rates, connections, response codes)
  • Databases: PostgreSQL (queries/s, locks, vacuum), MySQL, MongoDB, Redis (ops/s, hit rate, keyspace)
  • Containers: Docker stats per container (CPU, memory, network, I/O)
  • Message queues: RabbitMQ, Kafka
  • Languages: Node.js (eventloop lag), Python, Go runtime metrics
  • Cloud: AWS, GCP, Azure metrics via API

Configuration for auto-discovery (usually not needed):

# /etc/netdata/go.d/nginx.conf (example — Netdata finds this automatically)
jobs:
  - name: local
    url: http://127.0.0.1/nginx_status

Multi-Server Setup: Parent-Child Mode

Monitor multiple servers from one central Netdata instance:

Child Servers (each monitored server)

# /etc/netdata/stream.conf on each child:
[stream]
    enabled = yes
    destination = parent-server-ip:19999
    api key = your-api-key-here

    # How often to stream (default: every second):
    buffer size bytes = 1048576
    reconnect delay seconds = 5

Parent Server (central aggregator)

# /etc/netdata/stream.conf on parent:
[your-api-key-here]
    enabled = yes
    allow from = *   # IP range of children
    default history = 3600
    default memory mode = ram
    health enabled by default = auto
    postpone alarms on connect seconds = 60

Now visit http://parent-server:19999 and see metrics from all connected servers in one UI.


Alerts and Notifications

Netdata comes with hundreds of pre-configured alerts (CPU > 90%, disk > 85%, memory > 80%, etc.).

Configure Slack Notifications

# /etc/netdata/health_alarm_notify.conf
SLACK_WEBHOOK_URL="https://hooks.slack.com/services/xxx/yyy/zzz"
DEFAULT_RECIPIENT_SLACK="#alerts"
SEND_SLACK="YES"

PagerDuty

SEND_PD="YES"
PD_SERVICE_KEY="your-pagerduty-service-key"

Email

SEND_EMAIL="YES"
DEFAULT_RECIPIENT_EMAIL="ops@yourdomain.com"
SENDMAIL="$(which sendmail)"

Custom Alert Rules

# /etc/netdata/health.d/my-alerts.conf
alarm: high_cpu_90s
  on: system.cpu
  lookup: average -90s percentage of user,system
  every: 30s
  warn: $this > 80
  crit: $this > 95
  info: CPU usage over the last 90 seconds
  to: sysadmin

Netdata Cloud (Optional, Free Tier)

Netdata Cloud provides a hosted dashboard that aggregates metrics from all your Netdata agents:

  1. Sign up at app.netdata.cloud (free)
  2. Get your claim token
  3. Connect each agent: netdata-claim.sh -token=YOUR_TOKEN -rooms=YOUR_ROOM_ID

Free tier: Unlimited nodes, 3-day metric retention Pro tier: Longer retention, team features

For full data privacy, skip Netdata Cloud and use the local dashboard or the Parent-Child streaming setup above.


Prometheus + Grafana Integration

If you already run Prometheus, Netdata exposes a Prometheus endpoint:

# Access Netdata's Prometheus metrics:
curl http://your-server:19999/api/v1/allmetrics?format=prometheus

# In prometheus.yml:
scrape_configs:
  - job_name: netdata
    metrics_path: /api/v1/allmetrics
    params:
      format: [prometheus]
    static_configs:
      - targets: ['your-server:19999']

This combines Netdata's 1-second collection with Prometheus's long-term storage and Grafana's query capabilities.


Resource Usage

ConfigurationCPURAM
Agent only (no cloud)~1%~70MB
Agent + Parent streaming~2%~150MB
Parent aggregating 10 nodes~5%~500MB

Netdata is specifically designed to be low-overhead. The 1% CPU figure is on a typical server with 100+ metrics collected per second.


Securing the Dashboard

By default, Netdata listens on port 19999 (all interfaces). Restrict access:

# /etc/netdata/netdata.conf
[web]
    bind to = 127.0.0.1   # Local only
    # Or bind to specific IP: bind to = 192.168.1.10

Then access via SSH tunnel:

ssh -L 19999:localhost:19999 user@your-server
# Open http://localhost:19999 locally

Or put behind Nginx with basic auth:

location / {
    proxy_pass http://127.0.0.1:19999;
    auth_basic "Monitoring";
    auth_basic_user_file /etc/nginx/.htpasswd;
}

Quick Update

# Update native install:
bash <(curl -Ss https://my-netdata.io/kickstart.sh) --reinstall

# Docker:
docker pull netdata/netdata
docker compose up -d

Why Self-Host Netdata?

The case for self-hosting Netdata comes down to three practical factors: data ownership, cost at scale, and operational control.

Data ownership is the fundamental argument. When you use a SaaS version of any tool, your data lives on someone else's infrastructure subject to their terms of service, their security practices, and their business continuity. If the vendor raises prices, gets acquired, changes API limits, or shuts down, you're left scrambling. Self-hosting Netdata means your data and configuration stay on infrastructure you control — whether that's a VPS, a bare metal server, or a home lab.

Cost at scale matters once you move beyond individual use. Most SaaS equivalents charge per user or per data volume. A self-hosted instance on a $10-20/month VPS typically costs less than per-user SaaS pricing for teams of five or more — and the cost doesn't scale linearly with usage. One well-configured server handles dozens of users for a flat monthly fee.

Operational control is the third factor. The Docker Compose configuration above exposes every setting that commercial equivalents often hide behind enterprise plans: custom networking, environment variables, storage backends, and authentication integrations. You decide when to update, how to configure backups, and what access controls to apply.

The honest tradeoff: you're responsible for updates, backups, and availability. For teams running any production workloads, this is familiar territory. For individuals, the learning curve is real but the tooling (Docker, Caddy, automated backups) is well-documented and widely supported.

Server Requirements and Sizing

Before deploying Netdata, assess your server capacity against expected workload.

Minimum viable setup: A 1 vCPU, 1GB RAM VPS with 20GB SSD is sufficient for personal use or small teams. Most consumer VPS providers — Hetzner, DigitalOcean, Linode, Vultr — offer machines in this range for $5-10/month. Hetzner offers excellent price-to-performance for European and US regions.

Recommended production setup: 2 vCPUs with 4GB RAM and 40GB SSD handles most medium deployments without resource contention. This gives Netdata headroom for background tasks, caching, and concurrent users while leaving capacity for other services on the same host.

Storage planning: The Docker volumes in this docker-compose.yml store all persistent Netdata data. Estimate your storage growth rate early — for data-intensive tools, budget for 3-5x your initial estimate. Hetzner Cloud and Vultr both support online volume resizing without stopping your instance.

Operating system: Any modern 64-bit Linux distribution works. Ubuntu 22.04 LTS and Debian 12 are the most commonly tested configurations. Ensure Docker Engine 24.0+ and Docker Compose v2 are installed — verify with docker --version and docker compose version. Avoid Docker Desktop on production Linux servers; it adds virtualization overhead and behaves differently from Docker Engine in ways that cause subtle networking issues.

Network: Only ports 80 and 443 need to be publicly accessible when running behind a reverse proxy. Internal service ports should be bound to localhost only. A minimal UFW firewall that blocks all inbound traffic except SSH, HTTP, and HTTPS is the single most effective security measure for a self-hosted server.

Backup and Disaster Recovery

Running Netdata without a tested backup strategy is an unacceptable availability risk. Docker volumes are not automatically backed up — if you delete a volume or the host fails, data is gone with no recovery path.

What to back up: The named Docker volumes containing Netdata's data (database files, user uploads, application state), your docker-compose.yml and any customized configuration files, and .env files containing secrets.

Backup approach: For simple setups, stop the container, archive the volume contents, then restart. For production environments where stopping causes disruption, use filesystem snapshots or database dump commands (PostgreSQL pg_dump, SQLite .backup, MySQL mysqldump) that produce consistent backups without downtime.

For a complete automated backup workflow that ships snapshots to S3-compatible object storage, see the Restic + Rclone backup guide. Restic handles deduplication and encryption; Rclone handles multi-destination uploads. The same setup works for any Docker volume.

Backup cadence: Daily backups to remote storage are a reasonable baseline for actively used tools. Use a 30-day retention window minimum — long enough to recover from mistakes discovered weeks later. For critical data, extend to 90 days and use a secondary destination.

Restore testing: A backup that has never been restored is a backup you cannot trust. Once a month, restore your Netdata backup to a separate Docker Compose stack on different ports and verify the data is intact. This catches silent backup failures, script errors, and volume permission issues before they matter in a real recovery.

Security Hardening

Self-hosting means you are responsible for Netdata's security posture. The Docker Compose setup provides a functional base; production deployments need additional hardening.

Always use a reverse proxy: Never expose Netdata's internal port directly to the internet. The docker-compose.yml binds to localhost; Caddy or Nginx provides HTTPS termination. Direct HTTP access transmits credentials in plaintext. A reverse proxy also centralizes TLS management, rate limiting, and access logging.

Strong credentials: Change default passwords immediately after first login. For secrets in docker-compose environment variables, generate random values with openssl rand -base64 32 rather than reusing existing passwords.

Firewall configuration:

ufw default deny incoming
ufw allow 22/tcp
ufw allow 80/tcp
ufw allow 443/tcp
ufw enable

Internal service ports (databases, admin panels, internal APIs) should only be reachable from localhost or the Docker network, never directly from the internet.

Network isolation: Docker Compose named networks keep Netdata's services isolated from other containers on the same host. Database containers should not share networks with containers that don't need direct database access.

VPN access for sensitive services: For internal-only tools, restricting access to a VPN adds a strong second layer. Headscale is an open source Tailscale control server that puts your self-hosted stack behind a WireGuard mesh, eliminating public internet exposure for internal tools.

Update discipline: Subscribe to Netdata's GitHub releases page to receive security advisory notifications. Schedule a monthly maintenance window to pull updated images. Running outdated container images is the most common cause of self-hosted service compromises.

Troubleshooting Common Issues

Container exits immediately or won't start

Check logs first — they almost always explain the failure:

docker compose logs -f netdata

Common causes: a missing required environment variable, a port already in use, or a volume permission error. Port conflicts appear as bind: address already in use. Find the conflicting process with ss -tlpn | grep PORT and either stop it or change Netdata's port mapping in docker-compose.yml.

Cannot reach the web interface

Work through this checklist:

  1. Confirm the container is running: docker compose ps
  2. Test locally on the server: curl -I http://localhost:PORT
  3. If local access works but external doesn't, check your firewall: ufw status
  4. If using a reverse proxy, verify it's running and the config is valid: caddy validate --config /etc/caddy/Caddyfile

Permission errors on volume mounts

Some containers run as a non-root user. If the Docker volume is owned by root, the container process cannot write to it. Find the volume's host path with docker volume inspect VOLUME_NAME, check the tool's documentation for its expected UID, and apply correct ownership:

chown -R 1000:1000 /var/lib/docker/volumes/your_volume/_data

High resource usage over time

Memory or CPU growing continuously usually indicates unconfigured log rotation, an unbound cache, or accumulated data needing pruning. Check current usage with docker stats netdata. Add resource limits in docker-compose.yml to prevent one container from starving others. For ongoing visibility into resource trends, deploy Prometheus + Grafana.

Data disappears after container restart

Data stored in the container's writable layer — rather than a named volume — is lost when the container is removed or recreated. This happens when the volume mount path in docker-compose.yml doesn't match where the application writes data. Verify mount paths against the tool's documentation and correct the mapping. Named volumes persist across container removal; only docker compose down -v deletes them.

Keeping Netdata Updated

Netdata follows a regular release cadence. Staying current matters for security patches and compatibility. The update process with Docker Compose is straightforward:

docker compose pull          # Download updated images
docker compose up -d         # Restart with new images
docker image prune -f        # Remove old image layers (optional)

Read the changelog before major version updates. Some releases include database migrations or breaking configuration changes. For major version bumps, test in a staging environment first — run a copy of the service on different ports with the same volume data to validate the migration before touching production.

Version pinning: For stability, pin to a specific image tag in docker-compose.yml instead of latest. Update deliberately after reviewing the changelog. This trades automatic patch delivery for predictable behavior — the right call for business-critical services.

Post-update verification: After updating, confirm Netdata is functioning correctly. Most services expose a /health endpoint that returns HTTP 200 — curl it from the server or monitor it with your uptime tool.


Compare all open source monitoring tools at OSSAlt.com/alternatives/datadog.

The SaaS-to-Self-Hosted Migration Guide (Free PDF)

Step-by-step: infrastructure setup, data migration, backups, and security for 15+ common SaaS replacements. Used by 300+ developers.

Join 300+ self-hosters. Unsubscribe in one click.