Hello everyone!
My friend and I have each bought an optiplex server. Our goal is to selfhost a web app (static html) with redundancy. If my server goes down, his takes over and vice versa. I’ve looked into Docker Swarm, but each server has to be totally independent (each runs its own apps, with a few shared ones).
I can’t find a solution that allows each server to take over and manage loadbalancing between the two. Ideally with traefik, because that’s what we’re currently using. To me the real issue is the DNS A record that point to only one IP :(
What your after is DNS fail over or round robin, round robin would be my option as both sites could be independent
This is the correct answer. Healthcheck for each host to remove a dead endpoint from rotation.
Edit: missed your comment about static site
OP said that they have a static website, this eliminates the need for session sync.
Unified state shouldn’t be much of a problem for static web hosting.
Many moons ago I used heartbeat for this, but you’d need both servers in the same cidr range. I assume that’s not the case here.
In your case you could probably use a dynamic DNS service to move the IP around, but the challenge would be knowing when to kick it off.
You could write scripts to determine when the live one goes down, but we’re probably already more complicated than you were looking for.
Your challenge is that you need a loadbalancer. By hosting the loadbalancer yourself (e.g. on a VPS), you could also host your websites directly there…
My approach would be DNS-based. You can have multiple DNS A records, and the client picks one of them. With a little script you could remove one of the A Records of that server goes down. This way, you wouldn’t need a central hardware.
Where would you host the script? If it’s expected that the server that fires it off is always online and performing health checks, why not have it host a load-balancer? Or another local instance of the website? It’s something fun to play around with, but if this is for anything beyond a fun exercise there are much better ways to accomplish this.
I’d host it on both webservers. The script sets the A record to all the servers that are online. Obviously, the script als has to check it’s own service.
It seems a little hacky though, for a business use case I would use another approach.
Essentially you need a load balancer hosted somewhere that the traffic hits before getting routed to one of the 2 servers. That could be a VPS running Traefik if you prefer that.
Alternatively you could both run something like IPFS and run the static site on that, but anyone accessing the site would either need IPFS installed, or use a gateway hosted somewhere (Cloudflare has a public for example).
If you don’t want to mess with another VPS you can use a global server load balancer (GSLB) provider like Akamai, Cloudflare, Azure, etc.
This being a self-host community though it’s unlikely you’d want to pursue something like this, but without knowing more about your specific use case it’s tough to make a recommendation.
If global high-availability is your primary goal then a hosted solution is probably best.
If this is just an exercise you and your friend are working on for giggles and it’s not for a mission-critical Production instance, then presumably self-hosting a load-balancer on each of your servers that includes both nodes in a target group would achieve this, though somewhat counterintuitive; if the website goes down at either location, I would imagine there’s a pretty high likelihood the LB itself would be down as well.