Recommeneded partitioning scheme:
- Raid 1 (mirror) 40 000MB ext4 /
- SWAP
-
- HDD less than 130gb = 16GB swap
-
- HDD more than 130GB and RAM less or equal to 64GB = 32GB swap
-
- HDD more than 130GB and RAM more than 64GB = 32GB swap
- Remaining for lv ext4 /var/lib/vz (LVM)
includes and runs the (install-post.sh) script
- Select the Rescue tab for the specific server, via the hetzner robot manager
-
- Operating system=Linux
-
- Architecture=64 bit
-
- Public key=optional
- --> Activate rescue system
- Select the Reset tab for the specific server,
- Check: Execute an automatic hardware reset
- --> Send
- Wait a few mins
- Connect via ssh/terminal to the rescue system running on your server and run the following
wget https://raw.githubusercontent.com/extremeshok/xshok-proxmox/master/hetzner-install.sh -c -O install-hetzner.sh && chmod 777 hetzner-install.sh && ./install-hetzner.sh
- Reboot
- Connect via ssh/terminal to the new Proxmox system running on your server and run the following
wget https://raw.githubusercontent.com/extremeshok/xshok-proxmox/master/lvm2zfs.sh -c -O lvm2zfs.sh && bash lvm2zfs.sh && rm lvm2zfs.sh
- Reboot
- Post Install: Now login via ssh as root and create a password, which will be used for the webinterface when logging in with pam authentication
Select install for the specific server, via the ovh manager
- --INSTALL-->
- Install from an OVH template
- --NEXT-->
- Type of OS: Ready-to-go (graphical user interface)
- VPS Proxmox VE (pick the latest non zfs version)
- Language: EN
- Target disk arrray: (always select the SSD array if you have ssd and hdd arrays)
- Enable/Tick: Customise the partition configuration
- --NEXT-->
- Disks used for this installation: (All of them)
- (Remove all the partitions and do the following)
- Type: Filesystem: Mount Point: LVM Name: RAID: Size:
-
- 1 primary Ext4 / - 1 20.0 GB
-
- 2 primary Swap swap - - 2 x 8.0 GB (minimum 16GB total, set recommended swap size)
-
- 3 LV Ext4 /var/lib/vz data 1 REMAINING GB (use all the remaining space)
- --NEXT-->
- Hostname: server.fqdn.com
- Installation script (URL): https://raw.githubusercontent.com/extremeshok/xshok-proxmox/master/install-post.sh
- Script return value: 0
- SSH keys: (always suggested, however if this value is used a webinterface login will not work without setting a root password in shell)
- --CONFIRM--> After installation, Connect via ssh/terminal to the new Proxmox system running on your server and run the following
wget https://raw.githubusercontent.com/extremeshok/xshok-proxmox/master/lvm2zfs.sh -c -O lvm2zfs.sh && bash lvm2zfs.sh && rm lvm2zfs.sh
- Reboot
- Post Install: login via ssh as root and create a password, which will be used for the webinterface when logging in with pam authentication
not required if server setup with install-hetzner.sh
- Added: 'reboot-quick' command which uses kexec to boot the latest kernel set in the boot loader
- Disables the enterprise repo, enables the public repo
- Adds non-free sources
- Adds the latest ceph
- Fixes known bugs (public key missing, max user watches, etc)
- Updates the system
- Installs openvswitch-switch, zfsutils and common system utilities
- Protects the webinterface with fail2ban
- Increase vzdump backup speed
- Increase max File Discriptor Limits
- Increase max Key limits
- Detect AMD EPYC CPU and install kernel 4.15
- Detect AMD EPYC CPU and Apply EPYC fix to kernel
https://raw.githubusercontent.com/extremeshok/xshok-proxmox/master/install-post.sh
return value is 0
Or run install-post.sh after installation
wget https://raw.githubusercontent.com/extremeshok/xshok-proxmox/master/postinstall.sh -c -O postinstall.sh && bash postinstall.sh && rm postinstall.sh
There can be security implications as the LXC container is running in a higher privileged mode.
curl https://raw.githubusercontent.com/extremeshok/xshok-proxmox/master/pve-enable-lxc-docker.sh --output /usr/sbin/pve-enable-lxc-docker && chmod +x /usr/sbin/pve-enable-lxc-docker
pve-enable-lxc-docker container_id
Converts the storage LVM into a ZFS raid 1 (mirror)
- Uses the LVM with the path/mount of /var/lib/vz
- Will automatically detect the required raid level and optimise
- 1 Drive = zfs (single)
- 2 Drives = mirror (raid1)
- 3-5 Drives = raidz-1 (raid5)
- 6-11 Drives = raidz-2 (raid6)
- 11+ Drives = raidz-3 (raid7)
NOTE: WILL DESTROY ALL DATA ON /var/lib/vz
wget https://raw.githubusercontent.com/extremeshok/xshok-proxmox/master/lvm2zfs.sh -c -O lvm2zfs.sh && bash lvm2zfs.sh && rm lvm2zfs.sh
Creates a zfs pool from specified devices
- Will automatically detect the required raid level and optimise
- 1 Drive = zfs (single)
- 2 Drives = mirror (raid1)
- 3-5 Drives = raidz-1 (raid5)
- 6-11 Drives = raidz-2 (raid6)
- 11+ Drives = raidz-3 (raid7)
NOTE: WILL DESTROY ALL DATA ON SPECIFIED DEVICES
wget https://raw.githubusercontent.com/extremeshok/xshok-proxmox/master/createzfs.sh -c -O createzfs.sh
bash createzfs.sh poolname /dev/device1 /dev/device2
If no interface is specified the default gateway interface will be detected and used.
wget https://raw.githubusercontent.com/extremeshok/xshok-proxmox/master/addiprange.sh -c -O addiprange.sh
bash addiprange.sh ip.xx.xx.xx/cidr interface_optional
tinc private mesh vpn/network which supports multicast, ideal for private cluster communication
wget https://raw.githubusercontent.com/extremeshok/xshok-proxmox/master/tincvpn.sh -c -O tincvpn.sh && bash tincvpn.sh -h
bash tincvpn.sh -i 1 -c host2
bash tincvpn.sh -i 2 -c host3
bash tincvpn.sh -i 3 -c host1