Proxmox NETWORKING: VLANs, Bridges, and Bonds!
I’m sure many of you follow me because you use Proxmox. It’s been a staple of my content for some time now. So, while working on the next episode of the Ceph series, I thought it would be good to do a separate segment on networking. So, here you have it. The basics of VLANs, Bridges, and Bonds in Proxmox VE. I’m only covering the native Linux versions, not Open VSwitch and VXLAN.
Backup Proxmox VE to the CLOUD! Backup Hook Scripts and S3
Proxmox has a pretty good backup scheduler, but it relies on the backup destination being mounted as a storage location. This implies that the backup destination needs to be a protocol that Proxmox supports - SMB (CIFS), NFS, … or Proxmox Backup Server. If you want to push your backups to a cloud service, you probably need something a bit more complicated. Thankfully, Proxmox’s backup scheduler thought about this and has a hook feature we can use for this purpose, and we can use any protocol supported on the Debian base system, including things such as FUSE or s3cmd.
Net Booting the Proxmox VDI Client (feat. Alpine Linux)
This is a continuation of my previous article on the Net Booted Thin Client. The instructions got way too long, so I created a new article for the client setup. You need a functional server setup (TFTP, HTTP, iPXE) which I did in my previous post. I could use something like Linux Terminal Server Project, but that’s a bit overkill for this, and I wanted to learn Alpine anyway, so I’ve chosen to use Alpine Linux for the client operating system.
Not Every Project Works, And That’s Okay (Multiseat USB Dongles)
Sometimes, projects don’t work. Today, I’m going to describe a bit about a few of them. Thank you for coming to my ted talk lol.
I’ve been working on Linux Multiseat for awhile now, it’s a topic that has fascinated me for over a decade now. But, getting it to actually work with cheap hardware has eluded me. So, here’s a bit of an overview of what I’ve learned so far.
Making the $250 Proxmox HA Cluster Hyperconverged
I previously setup a Proxmox high availability cluster on my $35 Dell Wyse 5060 thin clients. Now, I’m improving this cluster to make it hyperconverged. It’s a huge buzzword in the industry now, and basically, it combines storage and compute in the same nodes, with each node having some compute and some storage, and clustering both the storage and compute. In traditional clustering you have a storage system (SAN) and compute system (virtualization cluster / kubernetes / …), so merging the SAN into the compute nodes means all of the nodes are identical and network traffic is, in aggregate, going from all nodes to all nodes without a bottleneck between the compute and SAN nodes.
Hyper-Converged Cluster Megaproject
In this project, I explore using low cost thin clients as cluster nodes, the fundamentals of Proxmox clustering, redundant storage, and hyper-converged infrastructure using Proxmox and Ceph.
Setting up a Proxmox HA Cluster In the first video, I take the Dell Wyse 5060 I bought before and … bought 2 more. Once I had 3, I built a complete high availability cluster using them, demonstrating the very basics of Proxmox clustering, high availability resources, and how Proxmox handles failure.
Setting up a Netboot (PXE) Server for Alpine Linux
I’ve played with network booting before (and I even tried to netboot Windows, what a nightmare that was), but now I want to get serious about it. I want to netboot my VDI Clients. I’ve been working on a thin client series, and the next step is to get rid of the installation entirely. I could use something like Linux Terminal Server Project, but that’s a bit overkill for this, and I wanted to learn Alpine anyway, so I’ve chosen to use Alpine Linux for the client operating system.
Thin Client Megaproject
In this project, I explore thin client software, and set up a Raspberry Pi to act as a thin client. This is a multi-part project, follow along below for each sub-project.
Raspberry Pi SPICE Thin Client The first project in this series creates a Raspberry Pi Thin Client which is permanently bound to a specific VM in Proxmox, and boots directly into the thin client session. You would use this when you have a 1:1 relationship between clients and VMs, such as a computer lab.
2 gamers, 1 cpu, NO Virtualization!
I’ve been using Linux for a long time. The first distribution I ever installed was Ubuntu 6.06 ‘Dapper Drake’, on which I ran a phpBB server for my school friends. Security was poor. But shortly after, as I learned about Linux desktops (and first daily-drove a Linux laptop around 2012), I learned about ‘Multiseat’. The Wikipedia article has a cool picture of a very 90s looking desktop tower with 4x CRT monitors, and that image was published to Wikipedia in 2005, and I saw it in middle/high school years.
Multiuser, Multidesktop, and Multiseat Megaproject
In this project, I explore thin client software, and set up a Raspberry Pi to act as a thin client. This is a multi-part project, follow along below for each sub-project.
A Modern Linux Graphical Terminal Server For the first project, I setup a remote desktop terminal server, allowing many users to connect to a single server and operate indepednent graphical sessions. My test bench was able to handle a dozen simulatneous users, and a proper server should be able to handle many more.