Skip to content

Latest commit

 

History

History
192 lines (139 loc) · 6.94 KB

readme.md

File metadata and controls

192 lines (139 loc) · 6.94 KB

vpnham

High-availability monitor for VPN tunnels.

TL;DR

Side A                                               Side B

    +----------+           active          +----------+
    |  Active  | <=======================> |  Active  |
    +----------o                           x----------+
       .           o                   x           .
       .               o           x               .
       .                   o   x                   .
       .                   x   o                   .
       .     backup    x           o    backup     .
       .           x                   o           .
    +----------x                           o----------+
    |  Backup  | < - - - - - - - - - - - > |  Backup  |
    +----------+           backup          +----------+

Details

Typical setup would be:

  • 2x VPN hosts on each side of the bridge.
  • One host on each side configured as active, the other as standby.
  • Each host on each side has VPN tunnels configured to both of the other side hosts.
  • Only active-active tunnel is used, the others are there for backup.

Logic

In the event of the downtime (tunnel is broken, one of active hosts is down), the monitor would:

  • Promote the remaining tunnel to become active.
  • Trigger custom command scripts to adjust for the situation (to re-configure the routes, for example).

In order to achieve that, vpnham does the following:

  • Regularly send UDP-datagrams to the each of the peers on the other side (connections <==>, < - >, < x >, and < o > on the diagram above).

    • The peer adds its bit into the datagram and sends it back.
    • This is how vpnham determines the up/down status of the tunnel (it accounts for the sent probes with their sequence numbers, and expects them to come back).
  • Regularly poll the partner's bridge (i.e. the active bridge polls the standby one, and vice versa; connections < . > above).

    • Failure to poll means the partner is down.
    • If both of the tunnels are down, the bridge marks itself down as well and reports itself accordingly to its partner.
  • If active tunnel is down, the standby is promoted to active.

  • If active bridge is down, the standby is promoted to active.

  • Once the active (by configuration) tunnel gets back up, it will reclaim the active status (in other words, the ties are broken via configuration).

  • Similar approach with the bridges.

Scripts

There are configurable scripts (per bridge, or globally):

  • bridge_activate is triggered when a bridge is promoted to active.

    • Recognised placeholders are:
      • ${proto}
      • ${bridge_peer_cidr}
      • ${bridge_interface}
      • ${bridge_interface_ip}
  • tunnel_activate is triggered when a tunnel is marked active.

    • Recognised placeholders are the same as for bridge_activate, plus:
      • ${tunnel_interface}
      • ${tunnel_interface_ip}
  • tunnel_deactivate is triggered when the tunnel's active mark is removed.

    • Recognised placeholders are the same as for tunnel_activate

Metrics

In addition, there is a metrics endpoint where vpnham reports the following:

  • vpnham_bridge_active is a gauge for count of active bridges.

    • 0 is when no connectivity to the other side (bad).
    • 1 is when all is good (yay).
    • 2 is when both us and the partner consider themselves active (this means bug).
  • vpnham_bridge_up is a gauge for the count of online bridges (from 0 to 2, the more the merrier).

  • vpnham_tunnel_interface_active is a gauge for count of active tunnels.

  • vpnham_tunnel_interface_up is a gauge for count of online tunnels.

Also (since we have that info at our fingertips through probing), the following metrics are exposed:

  • vpnham_probes_sent_total is a counter for probes sent.

  • vpnham_probes_returned_total is a counter for probes returned.

  • vpnham_probes_failed_total is a counter for probes failed to send, or to receive.

  • vpnham_probes_latency_forward_microseconds is a histogram for the probes forward latency (on their trip "there").

  • vpnham_probes_latency_return_microseconds is a histogram for the probe return latency (trip "back")

Example

bridges:
  vpnham-dev-lft:  # bridge name (must match one at the partner's side)
    role: active   # our role (`active` or `standby`)

    bridge_interface: eth0  # interface on which the bridge connects to VPC
    peer_cidr: 10.1.0.0/16  # CIDR range of the VPC we are bridging into

    status_addr: 10.0.0.2:8080                 # address where our partner polls our status
    partner_url: http://10.0.0.3:8080/  # url where we poll the status of the partner

    probe_interval: 1s           # interval between UDP probes or status polls
    probe_location: left/active  # location label for the latency metrics

    tunnel_interfaces:
      eth1:           # interface on which VPN tunnel is running
        role: active  # tunnel role (`active` or `standby`)

        addr: 192.168.255.2:3003        # address where we respond to UDP probes
        probe_addr: 192.168.255.3:3003  # address where we send the UDP probes to

        threshold_down: 5  # count of failed probes/polls to mark peer/partner "down"
        threshold_up: 3    # count of successful probes/polls to mark peer/partner "up"

      eth2:
        role: standby
        addr: 192.168.255.18:3003
        probe_addr: 192.168.255.19:3003
        threshold_down: 7
        threshold_up: 5

    scripts_timeout: 5s  # max amount of time for script commands to finish

metrics:
  listen_addr: 0.0.0.0:8000  # where we expose the metrics (at `/metrics` path)

  latency_buckets_count: 33  # count of histogram buckets for latency metrics

  max_latency_us: 1000000  # max latency bucket in [us]; the buckets are computed
                           # exponentially, so that
                           # max_latency == pow(min_latency, buckets_count)

default_scripts:    # default scripts (complement the `scripts` on bridge config)
  bridge_activate:  # script that we will run when bridge becomes `active`
    - ["sh", "-c", "echo ${bridge_interface} ${bridge_interface_ip} ${bridge_peer_cidr}"]
    - ["sleep", "15"]

  interface_activate:  # script that we will run when tunnel becomes `active`
    - ["sh", "-c", "echo 'activate ${tunnel_interface} ${tunnel_interface_ip}'"]

  interface_deactivate:  # script that we will run when tunnel becomes `inactive`
    - ["sh", "-c", "echo 'deactivate ${tunnel_interface} ${tunnel_interface_ip}'"]

Note

See the following files for the full example:

Also: make docker-compose

CLI

vpnham takes only one cli-parameter --config that should point to the yaml-file with full configuration. By default it will seek .vpnham.yaml file in the working directory.