20250514-notes

This commit is contained in:
Petar Cubela
2025-05-14 11:53:43 +02:00
parent cf8c2382ca
commit f6398bf45c
469 changed files with 15210 additions and 29 deletions

View File

@@ -0,0 +1,104 @@
# Proxmox Virtual Environment and Clustering with ZFS and Ceph - Introductory Course Notes
## Introduction
### Basics
* Debian stable + Ubuntu Kernel
* Virtualization with KVM
* LXC Containers
* Supports ZFS and Ceph
### Features
* Snapshots
* KVM Virtual Machines (Windows, Linux, BSD)
* LXC Containers
* High Availability (HA) Clustering
* Live Migration
* Flexible Storage Options
* GUI Management
* Proxmox Data Center Manager
* Proxmox Backup Server
## Virtualization Stack
![Virtualization Stack](/files/proxmox/2025-05-14.png)
## Expansion Options
### Single Node
* **Storage Pool:**
* RAID Controller with LVM Ceph and ZFS dont support a RAID controller; if a RAID controller is used, change to the mode and use ZFS.
### Clustering Without HA
* Several Storage Pools
* Pools are not shared individual
### ZFS Cluster with Asynchronous Replication
* At least two pools
* A quorum device between them replicates the configuration between the nodes.
### Ceph Cluster
* At least 3 nodes
* Pool is combined and shared
## Installation
* It's possible to provide custom TOML files to install Proxmox automatically.
* ZFS RAID1 for the boot drive.
* Define the correct hostname; it's not easy to change later.
* Modify the update repositories as a first step! And update/upgrade the system.
* Hardware: Always better to have as many interface ports as possible.
* 2x 1 Gbit for Management
* 2x 10 Gbit for VMs (redundant, bonded via LACP requires stacked switches or backup mode)
* 2x 25 Gbit for Ceph storage clustering
* Single node: at least 4 ports; for clustering: at least 6 ports.
* Remove the IP address of the virtual bridge and set it directly on the physical interface. This prevents VMs from seeing the Proxmox IP.
* Bond ports for VMs (ideally two 10 Gbit ports) and set the virtual bridge on `bond0` so it can be used by VMs (set 'VLAN aware' at the bridge).
* Its possible to create a bond for the web UI port.
## VM Creation
## Todo
* [ ] Join the forum
* [ ] Research about initial TOML file for Proxmox for unattended configuration of PVE - 'proxmox autoinstaller'
* [ ] Test ZRAID1 on two SSDs at home with a 'turtle' Proxmox host
## Support Subscriptions
### Basic
1 CPU per year / 2 CPUs per year
* Access to enterprise repositories
* Stable software updates
* Support via the customer portal
* Three support tickets
* Response time: 1 business day
### Standard
1 CPU per year / 2 CPUs per year
* All Basic features
**Additionally:**
* 7 support tickets (total 10)
* Remote support (via SSH)
* Offline subscription key activation
### Premium
1 CPU per year / 2 CPUs per year
* All Standard features
### Q&A
- full mesh network for ceph server (directly cable 3 hosts without switches)
- official nvidia support vor vGPUs in proxmox
- virtual tpm can not be saved on nfs share
- one can couple an esxi host in the proxmox ui and directly import VMs from there