mirror of
https://github.com/gyptazy/ProxLB.git
synced 2026-04-05 20:31:57 +02:00
refactor: Code refactor of ProxLB preparing release 1.1.0
Fixes: #114 Fixes: #132 Fixes: #130 Fixes: #129 Fixes: #128 Fixes: #127 Fixes: #123 Fixes: #102
This commit is contained in:
@@ -1,2 +0,0 @@
|
||||
fixed:
|
||||
- Fix systemd service file missing install target and network requirements (by @thomasfinstad). [#135]
|
||||
12
.changelogs/1.1.0/114_refactor_code_base.yml
Normal file
12
.changelogs/1.1.0/114_refactor_code_base.yml
Normal file
@@ -0,0 +1,12 @@
|
||||
fixed:
|
||||
- Refactored code base for ProxLB [#114]
|
||||
- Renamed package from `proxlb` to `python3-proxlb` to align with Debian packaging guidelines [#114]
|
||||
- Switched to `pycodestyle` for linting [#114]
|
||||
- Package building will be done within GitHub actions pipeline [#114]
|
||||
- ProxLB now only returns a warning when no guests for further balancing are not present (instead of quitting) [132#]
|
||||
- All nodes (according to the free resources) will be used now [#130]
|
||||
- Fixed logging outputs where highest/lowest were mixed-up [#129]
|
||||
- Stop balancing when movement would get worste (new force param to enfoce for affinity rules) [#128]
|
||||
- Added requested documentation regarding Proxmox HA groups [#127]
|
||||
- Rewrite of the whole affinity/anti-affinity rules evaluation and placement [#123]
|
||||
- Fixed the `ignore` parameter for nodes where the node and guests on the node will be untouched [#102]
|
||||
1
.changelogs/1.1.0/release_meta.yml
Normal file
1
.changelogs/1.1.0/release_meta.yml
Normal file
@@ -0,0 +1 @@
|
||||
date: TBD
|
||||
3
.flake8
3
.flake8
@@ -1,3 +0,0 @@
|
||||
[flake8]
|
||||
per-file-ignores =
|
||||
proxlb: E501,E221,E266,E231,E127,E222,E128
|
||||
26
.github/workflows/02-create-package.yml
vendored
26
.github/workflows/02-create-package.yml
vendored
@@ -1,26 +0,0 @@
|
||||
name: Run basic pipeline on push
|
||||
on: [push]
|
||||
jobs:
|
||||
build:
|
||||
runs-on: ubuntu-latest
|
||||
strategy:
|
||||
matrix:
|
||||
python-version: ["3.8"]
|
||||
steps:
|
||||
- uses: actions/checkout@v3
|
||||
- name: Set up Python for ProxLB
|
||||
uses: actions/setup-python@v4
|
||||
with:
|
||||
python-version: ${{ matrix.python-version }}
|
||||
- name: Install dependencies for ProxLB
|
||||
run: |
|
||||
python -m pip install --upgrade pip
|
||||
pip install pytest proxmoxer flake8
|
||||
if [ -f requirements.txt ]; then pip install -r requirements.txt; fi
|
||||
- name: Run Python linting
|
||||
run: |
|
||||
python3 -m flake8 proxlb
|
||||
- name: Create distro packages
|
||||
run: |
|
||||
cd packaging
|
||||
./01_package.sh
|
||||
21
.github/workflows/10-code-liniting.yml
vendored
Normal file
21
.github/workflows/10-code-liniting.yml
vendored
Normal file
@@ -0,0 +1,21 @@
|
||||
name: Code linting
|
||||
on: [push]
|
||||
jobs:
|
||||
build:
|
||||
runs-on: ubuntu-latest
|
||||
strategy:
|
||||
matrix:
|
||||
python-version: ["3.8"]
|
||||
steps:
|
||||
- uses: actions/checkout@v3
|
||||
- name: Setup dependencies for code linting
|
||||
uses: actions/setup-python@v4
|
||||
with:
|
||||
python-version: ${{ matrix.python-version }}
|
||||
- name: Install additional dependencies for code linting
|
||||
run: |
|
||||
sudo apt-get update
|
||||
sudo apt-get -y install python3-pycodestyle pycodestyle
|
||||
- name: Run code linting on ProxLB Python code
|
||||
run: |
|
||||
pycodestyle proxlb/*
|
||||
73
.github/workflows/20-pipeline-build-deb-package.yml
vendored
Normal file
73
.github/workflows/20-pipeline-build-deb-package.yml
vendored
Normal file
@@ -0,0 +1,73 @@
|
||||
name: "Build package: .deb"
|
||||
on: [push]
|
||||
jobs:
|
||||
lint-code-proxlb:
|
||||
runs-on: ubuntu-latest
|
||||
strategy:
|
||||
matrix:
|
||||
python-version: ["3.8"]
|
||||
steps:
|
||||
- uses: actions/checkout@v3
|
||||
- name: Setup dependencies for code linting
|
||||
uses: actions/setup-python@v4
|
||||
with:
|
||||
python-version: ${{ matrix.python-version }}
|
||||
- name: Install additional dependencies for code linting
|
||||
run: |
|
||||
sudo apt-get update
|
||||
sudo apt-get -y install python3-pycodestyle pycodestyle
|
||||
- name: Run code linting on ProxLB Python code
|
||||
run: |
|
||||
pycodestyle proxlb/* && \
|
||||
echo "OK: Code linting successfully performed on ProxLB code."
|
||||
|
||||
build-package-debian:
|
||||
needs: lint-code-proxlb
|
||||
runs-on: ubuntu-latest
|
||||
steps:
|
||||
- name: Check out repository
|
||||
uses: actions/checkout@v3
|
||||
with:
|
||||
ref: 'development'
|
||||
|
||||
- name: Set up Docker with Debian image
|
||||
run: |
|
||||
docker pull debian:latest
|
||||
|
||||
- name: Build DEB package in Docker container
|
||||
run: |
|
||||
docker run --rm -v $(pwd):/workspace -w /workspace debian:latest bash -c "
|
||||
# Install dependencies
|
||||
apt-get update && \
|
||||
apt-get install -y python3 python3-setuptools debhelper dh-python python3-pip python3-stdeb python3-proxmoxer python3-requests python3-urllib3 && \
|
||||
# Build package
|
||||
python3 setup.py --command-packages=stdeb.command bdist_deb && \
|
||||
echo 'OK: Debian package successfully created.'
|
||||
"
|
||||
|
||||
- name: Upload Debian package python3-proxlb as artifact
|
||||
uses: actions/upload-artifact@v4
|
||||
with:
|
||||
name: debian-package
|
||||
path: deb_dist/*.deb
|
||||
|
||||
integration-test-debian:
|
||||
needs: build-package-debian
|
||||
runs-on: ubuntu-latest
|
||||
steps:
|
||||
- name: Download Debian package artifact
|
||||
uses: actions/download-artifact@v4
|
||||
with:
|
||||
name: debian-package
|
||||
path: deb_dist/
|
||||
|
||||
- name: Set up Docker with Debian image
|
||||
run: docker pull debian:latest
|
||||
|
||||
- name: Install and test Debian package in Docker container
|
||||
run: |
|
||||
docker run --rm -v $(pwd)/deb_dist:/deb_dist -w /deb_dist debian:latest bash -c "
|
||||
apt-get update && \
|
||||
apt-get install -y ./python3-proxlb*.deb && \
|
||||
python3 -c 'import proxlb; print(\"OK: Debian package successfully installed.\")'
|
||||
"
|
||||
96
.github/workflows/20-pipeline-build-rpm-package.yml
vendored
Normal file
96
.github/workflows/20-pipeline-build-rpm-package.yml
vendored
Normal file
@@ -0,0 +1,96 @@
|
||||
name: "Build package: .rpm"
|
||||
on: [push]
|
||||
jobs:
|
||||
lint-code-proxlb:
|
||||
runs-on: ubuntu-latest
|
||||
strategy:
|
||||
matrix:
|
||||
python-version: ["3.8"]
|
||||
steps:
|
||||
- uses: actions/checkout@v3
|
||||
- name: Setup dependencies for code linting
|
||||
uses: actions/setup-python@v4
|
||||
with:
|
||||
python-version: ${{ matrix.python-version }}
|
||||
- name: Install additional dependencies for code linting
|
||||
run: |
|
||||
sudo apt-get update
|
||||
sudo apt-get -y install python3-pycodestyle pycodestyle
|
||||
- name: Run code linting on ProxLB Python code
|
||||
run: |
|
||||
pycodestyle proxlb/* && \
|
||||
echo "OK: Code linting successfully performed on ProxLB code."
|
||||
|
||||
build-package-rpm:
|
||||
needs: lint-code-proxlb
|
||||
runs-on: ubuntu-latest
|
||||
steps:
|
||||
- name: Check out repository
|
||||
uses: actions/checkout@v3
|
||||
with:
|
||||
ref: 'development'
|
||||
|
||||
- name: Set up Docker with Debian image
|
||||
run: |
|
||||
docker pull debian:latest
|
||||
|
||||
- name: Build DEB package in Docker container
|
||||
run: |
|
||||
docker run --rm -v $(pwd):/workspace -w /workspace debian:latest bash -c "
|
||||
# Install dependencies
|
||||
apt-get update && \
|
||||
apt-get install -y python3 python3-setuptools rpm debhelper dh-python python3-pip python3-stdeb python3-proxmoxer python3-requests python3-urllib3 && \
|
||||
# Build package
|
||||
python3 setup.py --command-packages=stdeb.command bdist_rpm && \
|
||||
echo 'OK: RPM package successfully created.'
|
||||
"
|
||||
|
||||
- name: Upload RPM package python3-proxlb as artifact
|
||||
uses: actions/upload-artifact@v4
|
||||
with:
|
||||
name: rpm-package
|
||||
path: dist/*.rpm
|
||||
|
||||
# integration-test-rpm-rockylinux-9:
|
||||
# needs: build-package-rpm
|
||||
# runs-on: ubuntu-latest
|
||||
# steps:
|
||||
# - name: Download RPM package artifact
|
||||
# uses: actions/download-artifact@v4
|
||||
# with:
|
||||
# name: rpm-package
|
||||
# path: dist/
|
||||
|
||||
# - name: Set up Docker with RockyLinux 9 image
|
||||
# run: docker pull rockylinux:9
|
||||
|
||||
# - name: Install and test RPM package in Rocky Linux Docker container
|
||||
# run: |
|
||||
# docker run --rm -v $(pwd)/dist:/dist -w /dist rockylinux:9 bash -c "
|
||||
# # DNF does not handle wildcards well
|
||||
# rpm_file=\$(ls proxlb*.noarch.rpm) && \
|
||||
# dnf install -y \$rpm_file && \
|
||||
# python3 -c 'import proxlb; print(\"OK: RPM package successfully installed.\")'
|
||||
# "
|
||||
|
||||
# integration-test-rpm-rockylinux-8:
|
||||
# needs: build-package-rpm
|
||||
# runs-on: ubuntu-latest
|
||||
# steps:
|
||||
# - name: Download RPM package artifact
|
||||
# uses: actions/download-artifact@v4
|
||||
# with:
|
||||
# name: rpm-package
|
||||
# path: dist/
|
||||
|
||||
# - name: Set up Docker with RockyLinux 8 image
|
||||
# run: docker pull rockylinux:8
|
||||
|
||||
# - name: Install and test RPM package in Rocky Linux Docker container
|
||||
# run: |
|
||||
# docker run --rm -v $(pwd)/dist:/dist -w /dist rockylinux:8 bash -c "
|
||||
# # DNF does not handle wildcards well
|
||||
# rpm_file=\$(ls proxlb*.noarch.rpm) && \
|
||||
# dnf install -y \$rpm_file && \
|
||||
# python3 -c 'import proxlb; print(\"OK: RPM package successfully installed.\")'
|
||||
# "
|
||||
9
.gitignore
vendored
9
.gitignore
vendored
@@ -1,2 +1,7 @@
|
||||
packaging/changelog-fragments-creator/
|
||||
dev/
|
||||
__pycache__
|
||||
*.pyc
|
||||
.DS_Store
|
||||
build/
|
||||
dist/
|
||||
*.egg-info/
|
||||
proxlb_dev.yaml
|
||||
|
||||
122
CHANGELOG.md
122
CHANGELOG.md
@@ -1,122 +0,0 @@
|
||||
# Changelog
|
||||
|
||||
All notable changes to this project will be documented in this file.
|
||||
|
||||
The format is based on [Keep a Changelog](https://keepachangelog.com/en/1.0.0/),
|
||||
and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0.html).
|
||||
|
||||
## [1.0.6] - 2024-12-24
|
||||
|
||||
### Fixed
|
||||
|
||||
- Fix maintenance mode when using cli arg and config mode by using the merged list (by @CartCaved). [#119]
|
||||
- Fix that a scheduler time definition of 1 (int) gets wrongly interpreted as a bool (by @gyptazy). [#115]
|
||||
|
||||
## [1.0.5] - 2024-10-30
|
||||
|
||||
### Changed
|
||||
|
||||
- Change docs to make bool usage in configs more clear. [#104]
|
||||
|
||||
### Fixed
|
||||
|
||||
- Fix migration from local disks (by @greenlogles). [#113]
|
||||
- Fix allowed values (add DEBUG, WARNING) for log verbosity. [#98]
|
||||
- Fix node (and its objects) evaluation when not reachable (e.g., maintenance). [#107]
|
||||
- Fix evaluation of maintenance mode where comparing list & string resulted in a crash (by @glitchvern). [#106]
|
||||
|
||||
|
||||
## [1.0.4] - 2024-10-11
|
||||
|
||||
### Added
|
||||
|
||||
- Add feature to make API timeout configureable. [#91]
|
||||
- Add maintenance mode to evacuate a node and move workloads for other nodes in the cluster. [#58]
|
||||
- Add version output cli arg. [#89]
|
||||
|
||||
### Changed
|
||||
|
||||
- Run storage balancing only on supported shared storages. [#79]
|
||||
- Run storage balancing only when needed to save time. [#79]
|
||||
|
||||
### Fixed
|
||||
|
||||
- Fix CPU balancing where calculations are done in float instead of int. (by @glitchvern) [#75]
|
||||
- Fix documentation for the underlying infrastructure. [#81]
|
||||
|
||||
|
||||
## [1.0.3] - 2024-09-12
|
||||
|
||||
### Added
|
||||
|
||||
- Add storage balancing function. [#51]
|
||||
- Add a convert function to cast all bool alike options from configparser to bools. [#53]
|
||||
- Add a config parser options for future features. [#53]
|
||||
- Add a config versio schema that must be supported by ProxLB. [#53]
|
||||
- Add doc how to add dedicated user for authentication. (by @Dulux-Oz)
|
||||
- Add feature to allow the API hosts being provided as a comma separated list. [#60]
|
||||
- Add cli arg `-b` to return the next best node for next VM/CT placement. [#8]
|
||||
|
||||
### Changed
|
||||
|
||||
- Improve the underlying code base for future implementations. [#53]
|
||||
- Provide a more reasonable output when HA services are not active in a Proxmox cluster. [#68]
|
||||
|
||||
### Fixed
|
||||
|
||||
- Fixed `master_only` function by inverting the condition.
|
||||
- Improved the overall validation and error handling. [#64]
|
||||
- Fix bug in the `proxlb.conf` in the vm_balancing section.
|
||||
- Fix anti-affinity rules not evaluating a new and different node. [#67]
|
||||
- Fix documentation for the master_only parameter placed in the wrong config section. [#74]
|
||||
- Fix handling of unset `ignore_nodes` and `ignore_vms` resulted in an attribute error. [#71]
|
||||
|
||||
|
||||
## [1.0.2] - 2024-08-13
|
||||
|
||||
### Added
|
||||
|
||||
- Add option to run ProxLB only on the Proxmox's master node in the cluster (reg. HA feature). [#40]
|
||||
- Add option to run migrations in parallel or sequentially. [#41]
|
||||
|
||||
### Changed
|
||||
|
||||
- Fix daemon timer to use hours instead of minutes. [#45]
|
||||
|
||||
### Fixed
|
||||
|
||||
- Fix CMake packaging for Debian package to avoid overwriting the config file. [#49]
|
||||
|
||||
|
||||
## [1.0.0] - 2024-08-01
|
||||
|
||||
### Added
|
||||
|
||||
- Add option_mode to rebalance by node's free resources in percent (instead of bytes). [#29]
|
||||
- Add LXC/Container integration. [#27]
|
||||
- Add exclude grouping feature to rebalance VMs from being located together to new nodes. [#4]
|
||||
- Add dry-run support to see what kind of rebalancing would be done. [#6]
|
||||
- Add Docker/Podman support. [#10 by @daanbosch]
|
||||
- Add feature to prevent VMs from being relocated by defining a wildcard pattern. [#7]
|
||||
- Add feature to prevent VMs from being relocated by defining the 'plb_ignore_vm' tag. [#7]
|
||||
- Add include grouping feature to rebalance VMs bundled to new nodes. [#3]
|
||||
- Add option to rebalance by assigned VM resources to avoid overprovisioning. [#16]
|
||||
- Add feature to make log verbosity configurable [#17].
|
||||
|
||||
### Changed
|
||||
|
||||
- Adjusted general logging and log more details.
|
||||
|
||||
|
||||
## [0.9.9] - 2024-07-06
|
||||
|
||||
### Added
|
||||
|
||||
- Initial public development release of ProxLB.
|
||||
|
||||
|
||||
## [0.9.0] - 2024-02-01
|
||||
|
||||
### Added
|
||||
|
||||
- Development release of ProxLB.
|
||||
@@ -116,6 +116,6 @@ By participating in this project, you agree to abide by our [Code of Conduct](CO
|
||||
|
||||
## Getting Help
|
||||
|
||||
If you need help or have any questions, feel free to reach out by creating an issue or by joining our [discussion forum](https://github.com/gyptazy/proxlb/discussions). You can also refer to our [documentation](https://github.com/gyptazy/ProxLB/tree/main/docs) for more information about the project or join our [chat room](https://matrix.to/#/#proxlb:gyptazy.ch) in Matrix.
|
||||
If you need help or have any questions, feel free to reach out by creating an issue or by joining our [discussion forum](https://github.com/gyptazy/proxlb/discussions). You can also refer to our [documentation](https://github.com/gyptazy/ProxLB/tree/main/docs) for more information about the project or join our [chat room](https://matrix.to/#/#proxlb:gyptazy.com) in Matrix.
|
||||
|
||||
Thank you for contributing to ProxLB! Together, we can enhance the efficiency and performance of Proxmox clusters.
|
||||
Thank you for contributing to ProxLB! Together, we can enhance the efficiency and performance of Proxmox clusters.
|
||||
22
Dockerfile
22
Dockerfile
@@ -1,22 +0,0 @@
|
||||
# Use the official Python 3.12 image
|
||||
FROM python:3.12
|
||||
|
||||
# Labels
|
||||
LABEL maintainer="gyptazy@gyptazy.ch"
|
||||
LABEL org.label-schema.schema-version="0.9"
|
||||
LABEL org.label-schema.description="ProxLB - Rebalance VM workloads across nodes in a Proxmox cluster."
|
||||
LABEL org.label-schema.url="https://github.com/gyptazy/ProxLB"
|
||||
|
||||
# Create a directory for the app
|
||||
WORKDIR /app
|
||||
|
||||
# Copy the python program from the current directory to /app
|
||||
COPY proxlb /app/proxlb
|
||||
|
||||
# Copy requirements to the container
|
||||
COPY requirements.txt /app/requirements.txt
|
||||
|
||||
RUN pip install -r /app/requirements.txt
|
||||
|
||||
# Set the entry point to use the virtual environment's python
|
||||
ENTRYPOINT ["python3", "/app/proxlb"]
|
||||
2
LICENSE
2
LICENSE
@@ -671,4 +671,4 @@ into proprietary programs. If your program is a subroutine library, you
|
||||
may consider it more useful to permit linking proprietary applications with
|
||||
the library. If this is what you want to do, use the GNU Lesser General
|
||||
Public License instead of this License. But first, please read
|
||||
<https://www.gnu.org/licenses/why-not-lgpl.html>.
|
||||
<https://www.gnu.org/licenses/why-not-lgpl.html>.
|
||||
518
README.md
518
README.md
@@ -5,189 +5,195 @@
|
||||
<p float="center"><img src="https://img.shields.io/github/license/gyptazy/ProxLB"/><img src="https://img.shields.io/github/contributors/gyptazy/ProxLB"/><img src="https://img.shields.io/github/last-commit/gyptazy/ProxLB/main"/><img src="https://img.shields.io/github/issues-raw/gyptazy/ProxLB"/><img src="https://img.shields.io/github/issues-pr/gyptazy/ProxLB"/></p>
|
||||
|
||||
|
||||
# :warning: Important: ProxLB 1.1.x is coming
|
||||
This repository is currently under heavy work and changes. During that time it might come to issues, non working pipelines or wrong documentation. Please select a stable release tag for a suitable version during this time!
|
||||
|
||||
## Table of Contents
|
||||
- [ProxLB - (Re)Balance VM Workloads in Proxmox Clusters](#proxlb---rebalance-vm-workloads-in-proxmox-clusters)
|
||||
- [Table of Contents](#table-of-contents)
|
||||
- [Introduction](#introduction)
|
||||
- [Video of Migration](#video-of-migration)
|
||||
- [Features](#features)
|
||||
- [How does it work?](#how-does-it-work)
|
||||
- [Usage](#usage)
|
||||
- [Dependencies](#dependencies)
|
||||
- [Options](#options)
|
||||
- [Notes](#notes)
|
||||
- [Parameters](#parameters)
|
||||
- [Balancing](#balancing)
|
||||
- [General](#general)
|
||||
- [By Used Memory of VMs/CTs](#by-used-memory-of-vmscts)
|
||||
- [By Assigned Memory of VMs/CTs](#by-assigned-memory-of-vmscts)
|
||||
- [Storage Balancing](#storage-balancing)
|
||||
- [Affinity Rules / Grouping Relationships](#affinity-rules--grouping-relationships)
|
||||
- [Affinity (Stay Together)](#affinity-stay-together)
|
||||
- [Anti-Affinity (Keep Apart)](#anti-affinity-keep-apart)
|
||||
- [Ignore VMs (Tag Style)](#ignore-vms-tag-style)
|
||||
- [Systemd](#systemd)
|
||||
- [Manual](#manual)
|
||||
- [Proxmox GUI Integration](#proxmox-gui-integration)
|
||||
- [Quick Start](#quick-start)
|
||||
- [Container Quick Start (Docker/Podman)](#container-quick-start-dockerpodman)
|
||||
- [Logging](#logging)
|
||||
- [Motivation](#motivation)
|
||||
- [References](#references)
|
||||
- [Downloads](#downloads)
|
||||
- [Packages](#packages)
|
||||
- [Repository](#repository)
|
||||
- [Stable Releases](#stable-releases)
|
||||
- [Beta/Testing Releases](#betatesting-releases)
|
||||
- [Container Images (Docker/Podman)](#container-images-dockerpodman)
|
||||
- [Misc](#misc)
|
||||
- [Bugs](#bugs)
|
||||
- [Contributing](#contributing)
|
||||
- [Documentation](#documentation)
|
||||
- [Support](#support)
|
||||
- [Author(s)](#authors)
|
||||
1. [Introduction](#introduction)
|
||||
2. [Features](#features)
|
||||
3. [How does it work?](#how-does-it-work)
|
||||
4. [Installation](#installation)
|
||||
1. [Requirements / Dependencies](#requirements--dependencies)
|
||||
2. [Debian Package](#debian-package)
|
||||
3. [RedHat Package](#redhat-package)
|
||||
4. [Container / Docker](#container--docker)
|
||||
5. [Source](#source)
|
||||
5. [Upgrading](#upgrading)
|
||||
1. [Upgrading from < 1.1.0](#upgrading-from--110)
|
||||
2. [Upgrading from >= 1.1.0](#upgrading-from--110)
|
||||
6. [Usage / Configuration](#usage--configuration)
|
||||
1. [GUI Integration](#gui-integration)
|
||||
2. [Proxmox HA Integration](#proxmox-ha-integration)
|
||||
3. [Options](#options)
|
||||
7. [Affinity & Anti-Affinity Rules](#affinity--anti-affinity-rules)
|
||||
1. [Affinity Rules](#affinity-rules)
|
||||
2. [Anti-Affinity Rules](#anti-affinity-rules)
|
||||
8. [Maintenance](#maintenance)
|
||||
9. [Misc](#misc)
|
||||
1. [Bugs](#bugs)
|
||||
2. [Contributing](#contributing)
|
||||
3. [Documentation](#documentation)
|
||||
4. [Support](#support)
|
||||
10. [Author(s)](#authors)
|
||||
|
||||
|
||||
## Introduction
|
||||
`ProxLB` (PLB) is an advanced tool designed to enhance the efficiency and performance of Proxmox clusters by optimizing the distribution of virtual machines (VMs) or Containers (CTs) across the cluster nodes by using the Proxmox API. ProxLB meticulously gathers and analyzes a comprehensive set of resource metrics from both the cluster nodes and the running VMs. These metrics include CPU usage, memory consumption, and disk utilization, specifically focusing on local disk resources.
|
||||
ProxLB is an advanced load balancing solution specifically designed for Proxmox clusters, addressing the absence of a Dynamic Resource Scheduler (DRS) that is familiar to VMware users. As a third-party solution, ProxLB enhances the management and efficiency of Proxmox clusters by intelligently distributing workloads across available nodes. Workloads can be balanced by different times like the guest's memory, CPU or disk usage or their assignment to avoid overprovisioning and ensuring resources.
|
||||
|
||||
PLB collects resource usage data from each node in the Proxmox cluster, including CPU, (local) disk and memory utilization. Additionally, it gathers resource usage statistics from all running VMs, ensuring a granular understanding of the cluster's workload distribution.
|
||||
One of the key advantages of ProxLB is that it is fully open-source and free, making it accessible for anyone to use, modify, and contribute to. This ensures transparency and fosters community-driven improvements. ProxLB supports filtering and ignoring specific nodes and guests through configuration files and API calls, providing administrators with the flexibility to tailor the load balancing behavior to their specific needs.
|
||||
|
||||
Intelligent rebalancing is a key feature of ProxLB where it re-balances VMs based on their memory, disk or CPU usage, ensuring that no node is overburdened while others remain underutilized. The rebalancing capabilities of PLB significantly enhance cluster performance and reliability. By ensuring that resources are evenly distributed, PLB helps prevent any single node from becoming a performance bottleneck, improving the reliability and stability of the cluster. Efficient rebalancing leads to better utilization of available resources, potentially reducing the need for additional hardware investments and lowering operational costs.
|
||||
A standout feature of ProxLB is its maintenance mode. When enabled, all guest workloads are automatically moved to other nodes within the cluster, ensuring that a node can be safely updated, rebooted, or undergo hardware maintenance without disrupting the overall cluster operation. Additionally, ProxLB supports both affinity and anti-affinity rules, allowing operators to group multiple guests to run together on the same node or ensure that certain guests do not run on the same node, depending on the cluster's node count. This feature is crucial for optimizing performance and maintaining high availability.
|
||||
|
||||
Automated rebalancing reduces the need for manual actions, allowing operators to focus on other critical tasks, thereby increasing operational efficiency.
|
||||
ProxLB can also return the best next node for guest placement, which can be integrated into CI/CD pipelines using tools like Ansible or Terraform. This capability streamlines the deployment process and ensures efficient resource utilization. Furthermore, ProxLB leverages the Proxmox API, including the entire ACL (Access Control List) system, for secure and efficient operation. Unlike some solutions, it does not require SSH access, enhancing security and simplifying configuration.
|
||||
|
||||
Overall, ProxLB significantly enhances resource management by intelligently distributing workloads, reducing downtime through its maintenance mode, and providing improved flexibility with affinity and anti-affinity rules. Its seamless integration with CI/CD tools and reliance on the Proxmox API make it a robust and secure solution for optimizing Proxmox cluster performance.
|
||||
|
||||
### Video of Migration
|
||||
<img src="https://cdn.gyptazy.com/images/proxlb-rebalancing-demo.gif"/>
|
||||
|
||||
## Features
|
||||
ProxLB's key features are by enabling automatic rebalancing of VMs and CTs across a Proxmox cluster based on memory, CPU, and local disk usage while identifying optimal nodes for automation. It supports maintenance mode, affinity rules, and seamless Proxmox API integration with ACL support, offering flexible usage as a one-time operation, a daemon, or through the Proxmox Web GUI.
|
||||
|
||||
**Features**
|
||||
* Rebalance VMs/CTs in the cluster by:
|
||||
* Memory
|
||||
* Disk (only local storage)
|
||||
* CPU
|
||||
* Rebalance Storage in the cluster
|
||||
* Rebalance VMs/CTs disks to other storage pools
|
||||
* Rebalance by used storage
|
||||
* Get best Node for new VM/CT placement in cluster
|
||||
* Performing
|
||||
* Periodically
|
||||
* One-shot solution
|
||||
* Types
|
||||
* Rebalance only VMs
|
||||
* Rebalance only CTs
|
||||
* Rebalance all (VMs and CTs)
|
||||
* Rebalance VM/CT disks (Storage)
|
||||
* Filter
|
||||
* Exclude nodes
|
||||
* Exclude virtual machines
|
||||
* Grouping
|
||||
* Include groups (VMs that are rebalanced to nodes together)
|
||||
* Exclude groups (VMs that must run on different nodes)
|
||||
* Ignore groups (VMs that should be untouched)
|
||||
* Dry-run support
|
||||
* Human readable output in CLI
|
||||
* JSON output for further parsing
|
||||
* Migrate VM workloads away (e.g. maintenance preparation)
|
||||
* Get best nodes for further automation
|
||||
* Supported Guest Types
|
||||
* VMs
|
||||
* CTs
|
||||
* Maintenance Mode
|
||||
* Set node(s) into maintenance
|
||||
* Move all workloads to different nodes
|
||||
* Affinity / Anti-Affinity Rules
|
||||
* Fully based on Proxmox API
|
||||
* Fully integrated into the Proxmox ACL
|
||||
* No SSH required
|
||||
* Usage
|
||||
* One-Shot (one-shot)
|
||||
* Periodically (daemon)
|
||||
* Proxmox Web GUI Integration (optional)
|
||||
* One-Time
|
||||
* Daemon
|
||||
* Proxmox Web GUI Integration
|
||||
|
||||
## How does it work?
|
||||
ProxLB is a load-balancing system designed to optimize the distribution of virtual machines (VMs) and containers (CTs) across a cluster. It works by first gathering resource usage metrics from all nodes in the cluster through the Proxmox API. This includes detailed resource metrics for each VM and CT on every node. ProxLB then evaluates the difference between the maximum and minimum resource usage of the nodes, referred to as "Balanciness." If this difference exceeds a predefined threshold (which is configurable), the system initiates the rebalancing process.
|
||||
|
||||
Before starting any migrations, ProxLB validates that rebalancing actions are necessary and beneficial. Depending on the selected balancing mode — such as CPU, memory, or disk — it creates a balancing matrix. This matrix sorts the VMs by their maximum used or assigned resources, identifying the VM with the highest usage. ProxLB then places this VM on the node with the most free resources in the selected balancing type. This process runs recursively until the operator-defined Balanciness is achieved. Balancing can be defined for the used or max. assigned resources of VMs/CTs.
|
||||
|
||||
## Usage
|
||||
Running PLB is easy and it runs almost everywhere since it just depends on `Python3` and the `proxmoxer` library. Therefore, it can directly run on a Proxmox node, dedicated systems like Debian, RedHat, or even FreeBSD, as long as the API is reachable by the client running PLB.
|
||||
## Installation
|
||||
|
||||
### Dependencies
|
||||
* Python3
|
||||
* proxmoxer (Python module)
|
||||
### Requirements / Dependencies
|
||||
* Python3.x
|
||||
* proxmoxer
|
||||
* requests
|
||||
* urllib3
|
||||
* pyyaml
|
||||
|
||||
The dependencies can simply be installed with `pip` by running the following command:
|
||||
```
|
||||
pip install -r requirements.txt
|
||||
```
|
||||
|
||||
Distribution packages, such like the provided `.deb` package will automatically resolve and install all required dependencies by using already packaged version from the distribution's repository.
|
||||
|
||||
### Debian Package
|
||||
|
||||
### RedHat Package
|
||||
|
||||
### Container / Docker
|
||||
|
||||
### Source
|
||||
|
||||
## Upgrading
|
||||
|
||||
### Upgrading from < 1.1.0
|
||||
Upgrading ProxLB is not supported due to a fundamental redesign introduced in version 1.1.x. With this update, ProxLB transitioned from a monolithic application to a pure Python-style project, embracing a more modular and flexible architecture. This shift aimed to improve maintainability and extensibility while keeping up with modern development practices. Additionally, ProxLB moved away from traditional ini-style configuration files and adopted YAML for configuration management. This change simplifies configuration handling, reduces the need for extensive validation, and ensures better type casting, ultimately providing a more streamlined and user-friendly experience.
|
||||
|
||||
### Upgrading from >= 1.1.0
|
||||
Uprading within the current stable versions, starting from 1.1.0, will be possible in all supported ways.
|
||||
|
||||
## Usage / Configuration
|
||||
Running ProxLB is straightforward and versatile, as it only requires `Python3` and the `proxmoxer` library. This means ProxLB can be executed directly on a Proxmox node or on dedicated systems such as Debian, RedHat, or even FreeBSD, provided that the Proxmox API is accessible from the client running ProxLB. ProxLB can also run inside a Container - Docker or LXC - and is simply up to you.
|
||||
|
||||
### GUI Integration
|
||||
<img align="left" src="https://cdn.gyptazy.com/images/proxlb-GUI-integration.jpg"/> ProxLB can also be accessed through the Proxmox Web UI by installing the optional `pve-proxmoxlb-service-ui` package, which depends on the proxlb package. For full Web UI integration, this package must be installed on all nodes within the cluster. Once installed, a new menu item - `Rebalancing`, appears in the cluster level under the HA section. Once installed, it offers two key functionalities:
|
||||
* Rebalancing VM workloads
|
||||
* Migrate VM workloads away from a defined node (e.g. maintenance preparation)
|
||||
|
||||
**Note:** This package is currently discontinued and will be readded at a later time. See also: [#44: How to install pve-proxmoxlb-service-ui package](https://github.com/gyptazy/ProxLB/issues/44).
|
||||
|
||||
### Proxmox HA Integration
|
||||
Proxmox HA (High Availability) groups are designed to ensure that virtual machines (VMs) remain running within a Proxmox cluster. HA groups define specific rules for where VMs should be started or migrated in case of node failures, ensuring minimal downtime and automatic recovery.
|
||||
|
||||
However, when used in conjunction with ProxLB, the built-in load balancer for Proxmox, conflicts can arise. ProxLB operates with its own logic for workload distribution, taking into account affinity and anti-affinity rules. While it effectively balances guest workloads, it may re-shift and redistribute VMs in a way that does not align with HA group constraints, potentially leading to unsuitable placements.
|
||||
|
||||
Due to these conflicts, it is currently not recommended to use both HA groups and ProxLB simultaneously. The interaction between the two mechanisms can lead to unexpected behavior, where VMs might not adhere to HA group rules after being moved by ProxLB.
|
||||
|
||||
A solution to improve compatibility between HA groups and ProxLB is under evaluation, aiming to ensure that both features can work together without disrupting VM placement strategies.
|
||||
|
||||
See also: [#65: Host groups: Honour HA groups](https://github.com/gyptazy/ProxLB/issues/65).
|
||||
|
||||
### Options
|
||||
The following options can be set in the `proxlb.conf` file:
|
||||
The following options can be set in the configuration file `proxlb.yaml`:
|
||||
|
||||
| Section | Option | Example | Description |
|
||||
|------|:------:|:------:|:------:|
|
||||
| `proxmox` | api_host | hypervisor01.gyptazy.com | Host or IP address (or comma separated list) of the remote Proxmox API. |
|
||||
| | api_user | root@pam | Username for the API. |
|
||||
| | api_pass | FooBar | Password for the API. |
|
||||
| | verify_ssl | 1 | Validate SSL certificates (1) or ignore (0). (default: 1, type: bool) |
|
||||
| | timeout | 10 | Timeout for the Proxmox API in sec. (default: 10) |
|
||||
| `vm_balancing` | enable | 1 | Enables VM/CT balancing. |
|
||||
| | method | memory | Defines the balancing method (default: memory) where you can use `memory`, `disk` or `cpu`. |
|
||||
| | mode | used | Rebalance by `used` resources (efficiency) or `assigned` (avoid overprovisioning) resources. (default: used)|
|
||||
| | mode_option | byte | Rebalance by node's resources in `bytes` or `percent`. (default: bytes) |
|
||||
| | type | vm | Rebalance only `vm` (virtual machines), `ct` (containers) or `all` (virtual machines & containers). (default: vm)|
|
||||
| | balanciness | 10 | Value of the percentage of lowest and highest resource consumption on nodes may differ before rebalancing. (default: 10) |
|
||||
| | parallel_migrations | 1 | Defines if migrations should be done parallely or sequentially. (default: 1, type: bool) |
|
||||
| | maintenance_nodes | dummynode03,dummynode04 | Defines a comma separated list of nodes to set them into maintenance mode and move VMs/CTs to other nodes. |
|
||||
| | ignore_nodes | dummynode01,dummynode02,test* | Defines a comma separated list of nodes to exclude. |
|
||||
| | ignore_vms | testvm01,testvm02 | Defines a comma separated list of VMs to exclude. (`*` as suffix wildcard or tags are also supported) |
|
||||
| `storage_balancing` | enable | 0 | Enables storage balancing. |
|
||||
| | balanciness | 10 | Value of the percentage of lowest and highest storage consumption may differ before rebalancing. (default: 10) |
|
||||
| | parallel_migrations | 1 | Defines if migrations should be done parallely or sequentially. (default: 1, type: bool) |
|
||||
| `update_service` | enable | 0 | Enables the automated update service (rolling updates). (default: 0, type: bool) |
|
||||
| `api` | enable | 0 | Enables the ProxLB API. |
|
||||
| `service`| daemon | 1 | Run as a daemon (1) or one-shot (0). (default: 1, type: bool) |
|
||||
| | schedule | 24 | Hours to rebalance in hours. (default: 24) |
|
||||
| | master_only | 0 | Defines is this should only be performed (1) on the cluster master node or not (0). (default: 0, type: bool) |
|
||||
| | log_verbosity | INFO | Defines the log level (default: CRITICAL) where you can use `DEBUG`, `INFO`, `WARNING` or `CRITICAL` |
|
||||
| | config_version | 3 | Defines the current config version schema for ProxLB |
|
||||
| Section | Option | Example | Type | Description |
|
||||
|------|:------:|:------:|:------:|:------:|
|
||||
| `proxmox_api` | | | | |
|
||||
| | hosts | ['virt01.example.com', '10.10.10.10', 'fe01::bad:code::cafe'] | `List` | List of Proxmox nodes. Can be IPv4, IPv6 or mixed. |
|
||||
| | user | root@pam | `Str` | Username for the API. |
|
||||
| | pass | FooBar | `Str` | Password for the API. |
|
||||
| | ssl_verification | True | `Bool` | Validate SSL certificates (1) or ignore (0). (default: 1, type: bool) |
|
||||
| | timeout | 10 | `Int` | Timeout for the Proxmox API in sec. (default: 10) |
|
||||
| `proxmox_cluster` | | | | |
|
||||
| | maintenance_nodes | ['virt66.example.com'] | `List` | A list of Proxmox nodes that are defined to be in a maintenance. (default: []) |
|
||||
| | ignore_nodes | [] | `List` | A list of Proxmox nodes that are defined to be ignored. (default: []) |
|
||||
| | overprovisioning | False | `Bool` | Avoids balancing when nodes would become overprovisioned. |
|
||||
| `balancing` | | | | |
|
||||
| | enable | True | `Bool` | Enables the guest balancing. (default: True)|
|
||||
| | force | True | `Bool` | Enforcing affinity/anti-affinity rules but balancing might become worse. (default: False) |
|
||||
| | parallel | False | `Bool` | If guests should be moved in parallel or sequentially. (default: False)|
|
||||
| | live | True | `Bool` | If guests should be moved live or shutdown. (default: True)|
|
||||
| | with_local_disks | True | `Bool` | If balancing of guests should include local disks (default: True)|
|
||||
| | balance_types | ['vm', 'ct'] | `List` | Defined the types of guests that should be honored. (default: ['vm', 'ct']) |
|
||||
| | max_job_validation | 1800 | `Int` | How long a job validation may take in seconds. (default: 1800) |
|
||||
| | balanciness | 10 | `Int` | The maximum delta of resource usage between node with highest and lowest usage. (default: 10) |
|
||||
| | method | memory | `Str` | The balancing method that should be used. (default: memory | choices: memory, cpu, disk)|
|
||||
| | mode | used | `Str` | The balancing mode that should be used. (default: used | choices: used, assigned)|
|
||||
| `service` | | | | |
|
||||
| | daemon | False | `Bool` | If daemon mode should be activated (default: False)|
|
||||
| | schedule | 12 | `Int` | How often rebalancing should occur in hours in daemon mode (default: 12)|
|
||||
| | log_level | INFO | `Str` | Defines the default log level that should be logged. (default: INFO) |
|
||||
|
||||
An example of the configuration file looks like:
|
||||
```
|
||||
[proxmox]
|
||||
api_host: hypervisor01.gyptazy.com
|
||||
api_user: root@pam
|
||||
api_pass: FooBar
|
||||
verify_ssl: 1
|
||||
timeout: 10
|
||||
[vm_balancing]
|
||||
enable: 1
|
||||
method: memory
|
||||
mode: used
|
||||
type: vm
|
||||
# Balanciness defines how much difference may be
|
||||
# between the lowest & highest resource consumption
|
||||
# of nodes before rebalancing will be done.
|
||||
# Examples:
|
||||
# Rebalancing: node01: 41% memory consumption :: node02: 52% consumption
|
||||
# No rebalancing: node01: 43% memory consumption :: node02: 50% consumption
|
||||
balanciness: 10
|
||||
# Enable parallel migrations. If set to 0 it will wait for completed migrations
|
||||
# before starting next migration.
|
||||
parallel_migrations: 1
|
||||
maintenance_nodes: dummynode03,dummynode04
|
||||
ignore_nodes: dummynode01,dummynode02
|
||||
ignore_vms: testvm01,testvm02
|
||||
[storage_balancing]
|
||||
enable: 0
|
||||
[update_service]
|
||||
enable: 0
|
||||
[api]
|
||||
enable: 0
|
||||
[service]
|
||||
# The master_only option might be useful if running ProxLB on all nodes in a cluster
|
||||
# but only a single one should do the balancing. The master node is obtained from the Proxmox
|
||||
# HA status.
|
||||
master_only: 0
|
||||
daemon: 1
|
||||
config_version: 3
|
||||
```
|
||||
proxmox_api:
|
||||
hosts: ['virt01.example.com', '10.10.10.10', 'fe01::bad:code::cafe']
|
||||
user: root@pam
|
||||
pass: crazyPassw0rd!
|
||||
ssl_verification: False
|
||||
timeout: 10
|
||||
|
||||
#### Notes
|
||||
* If running ProxLB on more than one Proxmox node you can set `api_host` to a comma-separated list of each node's IP address or hostname. (Example: `api_host: node01.gyptazy.com,node02.gyptazy.com,node03.gyptazy.com`)
|
||||
* The `verify_ssl` parameter can switch between the mode to verify trusted remote certificates. Keep in mind, that even local ones are **not** trusted by default and need to be imported to the truststore.
|
||||
* Even when using only the `vm_balancing` mode, ensure to have the other sections listed in your config:
|
||||
```
|
||||
[storage_balancing]
|
||||
enable: 0
|
||||
[update_service]
|
||||
enable: 0
|
||||
[api]
|
||||
enable: 0
|
||||
proxmox_cluster:
|
||||
maintenance_nodes: ['virt66.example.com']
|
||||
ignore_nodes: []
|
||||
overprovisioning: True
|
||||
|
||||
balancing:
|
||||
enable: True
|
||||
force: False
|
||||
parallel: False
|
||||
live: True
|
||||
with_local_disks: True
|
||||
balance_types: ['vm', 'ct']
|
||||
max_job_validation: 1800
|
||||
balanciness: 5
|
||||
method: memory
|
||||
mode: assigned
|
||||
|
||||
service:
|
||||
daemon: False
|
||||
schedule: 12
|
||||
log_level: DEBUG
|
||||
```
|
||||
|
||||
### Parameters
|
||||
@@ -196,204 +202,70 @@ The following options and parameters are currently supported:
|
||||
| Option | Long Option | Description | Default |
|
||||
|------|:------:|------:|------:|
|
||||
| -c | --config | Path to a config file. | /etc/proxlb/proxlb.conf (default) |
|
||||
| -d | --dry-run | Performs a dry-run without doing any actions. | Unset |
|
||||
| -j | --json | Returns a JSON of the VM movement. | Unset |
|
||||
| -b | --best-node | Returns the best next node for a VM/CT placement (useful for further usage with Terraform/Ansible). | Unset |
|
||||
| -m | --maintenance | Sets node(s) to maintenance mode & moves workloads away. | Unset |
|
||||
| -v | --version | Returns the ProxLB version on stdout. | Unset |
|
||||
| -d | --dry-run | Performs a dry-run without doing any actions. | False |
|
||||
| -j | --json | Returns a JSON of the VM movement. | False |
|
||||
| -b | --best-node | Returns the best next node for a VM/CT placement (useful for further usage with Terraform/Ansible). | False |
|
||||
| -v | --version | Returns the ProxLB version on stdout. | False |
|
||||
|
||||
### Balancing
|
||||
#### General
|
||||
In general, virtual machines (VMs), containers (CTs) can be rebalanced and moved around nodes or shared storage (storage balancing) in the cluster. Often, this also works without downtime without any further downtimes. However, this does **not** work with containers. LXC based containers will be shutdown, copied and started on the new node. Also to note, live migrations can work fluently without any issues but there are still several things to be considered. This is out of scope for ProxLB and applies in general to Proxmox and your cluster setup. You can find more details about this here: https://pve.proxmox.com/wiki/Migrate_to_Proxmox_VE.
|
||||
## Affinity & Anti-Affinity Rules
|
||||
ProxLB provides an advanced mechanism to define affinity and anti-affinity rules, enabling precise control over virtual machine (VM) placement. These rules help manage resource distribution, improve high availability configurations, and optimize performance within a Proxmox Virtual Environment (PVE) cluster. By leveraging Proxmox’s integrated access management, ProxLB ensures that users can only define and manage rules for guests they have permission to access.
|
||||
|
||||
#### By Used Memory of VMs/CTs
|
||||
By continuously monitoring the current resource usage of VMs, ProxLB intelligently reallocates workloads to prevent any single node from becoming overloaded. This approach ensures that resources are balanced efficiently, providing consistent and optimal performance across the entire cluster at all times. To activate this balancing mode, simply activate the following option in your ProxLB configuration:
|
||||
ProxLB implements affinity and anti-affinity rules through a tag-based system within the Proxmox web interface. Each guest (virtual machine or container) can be assigned specific tags, which then dictate its placement behavior. This method maintains a streamlined and secure approach to managing VM relationships while preserving Proxmox’s inherent permission model.
|
||||
|
||||
### Affinity Rules
|
||||
<img align="left" src="https://cdn.gyptazy.com/images/proxlb-affinity-rules.jpg"/> Affinity rules are used to group certain VMs together, ensuring that they run on the same host whenever possible. This can be beneficial for workloads requiring low-latency communication, such as clustered databases or application servers that frequently exchange data.
|
||||
|
||||
To define an affinity rule which keeps all guests assigned to this tag together on a node, users assign a tag with the prefix `plb_affinity_$TAG`:
|
||||
|
||||
#### Example for Screenshot
|
||||
```
|
||||
mode: used
|
||||
plb_affinity_talos
|
||||
```
|
||||
|
||||
Afterwards, restart the service (if running in daemon mode) to activate this rebalancing mode.
|
||||
As a result, ProxLB will attempt to place all VMs with the `plb_affinity_web` tag on the same host (see also the attached screenshot with the same node).
|
||||
|
||||
#### By Assigned Memory of VMs/CTs
|
||||
By ensuring that resources are always available for each VM, ProxLB prevents over-provisioning and maintains a balanced load across all nodes. This guarantees that users have consistent access to the resources they need. However, if the total assigned resources exceed the combined capacity of the cluster, ProxLB will issue a warning, indicating potential over-provisioning despite its best efforts to balance the load. To activate this balancing mode, simply activate the following option in your ProxLB configuration:
|
||||
### Anti-Affinity Rules
|
||||
<img align="left" src="https://cdn.gyptazy.com/images/proxlb-anti-affinity-rules.jpg"/> Conversely, anti-affinity rules ensure that designated VMs do not run on the same physical host. This is particularly useful for high-availability setups, where redundancy is crucial. Ensuring that critical services are distributed across multiple hosts reduces the risk of a single point of failure.
|
||||
|
||||
To define an anti-affinity rule that ensures to not move systems within this group to the same node, users assign a tag with the prefix:
|
||||
|
||||
#### Example for Screenshot
|
||||
```
|
||||
mode: assigned
|
||||
plb_anti_affinity_ntp
|
||||
```
|
||||
|
||||
Afterwards, restart the service (if running in daemon mode) to activate this rebalancing mode.
|
||||
As a result, ProxLB will try to place the VMs with the `plb_anti_affinity_ntp` tag on different hosts (see also the attached screenshot with the different nodes).
|
||||
|
||||
#### Storage Balancing
|
||||
Starting with ProxLB 1.0.3, ProxLB also supports the balancing of underlying shared storage. In this case, all attached disks (`rootfs` in a context of a CT) of a VM or CT are being fetched and evaluated. If a VM has multiple disks attached, the disks can also be distributed over different storages. As a result, only shared storage is supported. Non shared storage would require to move the whole VM including all attached disks to the parent's node local storage.
|
||||
**Note:** While this ensures that ProxLB tries distribute these VMs across different physical hosts within the Proxmox cluster this may not always work. If you have more guests attached to the group than nodes in the cluster, we still need to run them anywhere. If this case occurs, the next one with the most free resources will be selected.
|
||||
|
||||
Limitations:
|
||||
* Only shared storage
|
||||
* Only supported for the following VM disk types:
|
||||
* ide (only disks, not CD)
|
||||
* nvme
|
||||
* scsi
|
||||
* virtio
|
||||
* sata
|
||||
* rootfs (Container)
|
||||
## Maintenance
|
||||
<img src="https://cdn.gyptazy.com/images/proxlb-rebalancing-demo.gif"/>
|
||||
|
||||
*Note: Storage balancing is currently in beta and should be used carefully.*
|
||||
|
||||
### Affinity Rules / Grouping Relationships
|
||||
#### Affinity (Stay Together)
|
||||
<img align="left" src="https://cdn.gyptazy.com/images/plb-rebalancing-include-balance-group.jpg"/> Access the Proxmox Web UI by opening your web browser and navigating to your Proxmox VE web interface, then log in with your credentials. Navigate to the VM you want to tag by selecting it from the left-hand navigation panel. Click on the "Options" tab to view the VM's options, then select "Edit" or "Add" (depending on whether you are editing an existing tag or adding a new one). In the tag field, enter plb_include_ followed by your unique identifier, for example, plb_include_group1. Save the changes to apply the tag to the VM. Repeat these steps for each VM that should be included in the group.
|
||||
|
||||
#### Anti-Affinity (Keep Apart)
|
||||
<img align="left" src="https://cdn.gyptazy.com/images/plb-rebalancing-exclude-balance-group.jpg"/> Access the Proxmox Web UI by opening your web browser and navigating to your Proxmox VE web interface, then log in with your credentials. Navigate to the VM you want to tag by selecting it from the left-hand navigation panel. Click on the "Options" tab to view the VM's options, then select "Edit" or "Add" (depending on whether you are editing an existing tag or adding a new one). In the tag field, enter plb_exclude_ followed by your unique identifier, for example, plb_exclude_critical. Save the changes to apply the tag to the VM. Repeat these steps for each VM that should be excluded from being on the same node.
|
||||
|
||||
#### Ignore VMs (Tag Style)
|
||||
<img align="left" src="https://cdn.gyptazy.com/images/plb-rebalancing-ignore-vm.jpg"/> In Proxmox, you can ensure that certain VMs are ignored during the rebalancing process by setting a specific tag within the Proxmox Web UI, rather than solely relying on configurations in the ProxLB config file. This can be achieved by adding the tag 'plb_ignore_vm' to the VM. Once this tag is applied, the VM will be excluded from any further rebalancing operations, simplifying the management process.
|
||||
|
||||
### Systemd
|
||||
When installing a Linux distribution (such as .deb or .rpm) file, this will be shipped with a systemd unit file. The default configuration file will be sourced from `/etc/proxlb/proxlb.conf`.
|
||||
|
||||
| Unit Name | Options |
|
||||
|------|:------:|
|
||||
| proxlb | start, stop, status, restart |
|
||||
|
||||
### Manual
|
||||
A manual installation is possible and also supports BSD based systems. Proxmox Rebalancing Service relies on mainly two important files:
|
||||
* proxlb (Python Executable)
|
||||
* proxlb.conf (Config file)
|
||||
|
||||
The executable must be able to read the config file, if no dedicated config file is given by the `-c` argument, PLB tries to read it from `/etc/proxlb/proxlb.conf`.
|
||||
|
||||
### Proxmox GUI Integration
|
||||
<img align="left" src="https://cdn.gyptazy.com/images/proxlb-GUI-integration.jpg"/> PLB can also be directly be used from the Proxmox Web UI by installing the optional package `pve-proxmoxlb-service-ui` package which has a dependency on the `proxlb` package. For the Web UI integration, it requires to be installed (in addition) on the nodes on the cluster. Afterwards, a new menu item is present in the HA chapter called `Rebalancing`. This chapter provides two possibilities:
|
||||
* Rebalancing VM workloads
|
||||
* Migrate VM workloads away from a defined node (e.g. maintenance preparation)
|
||||
|
||||
### Quick Start
|
||||
The easiest way to get started is by using the ready-to-use packages that I provide on my CDN and to run it on a Linux Debian based system. This can also be one of the Proxmox nodes itself.
|
||||
|
||||
```
|
||||
wget https://cdn.gyptazy.com/files/os/debian/proxlb/proxlb_1.0.6_amd64.deb
|
||||
dpkg -i proxlb_1.0.6_amd64.deb
|
||||
# Adjust your config
|
||||
vi /etc/proxlb/proxlb.conf
|
||||
systemctl restart proxlb
|
||||
systemctl status proxlb
|
||||
```
|
||||
|
||||
### Container Quick Start (Docker/Podman)
|
||||
Creating a container image of ProxLB is straightforward using the provided Dockerfile. The Dockerfile simplifies the process by automating the setup and configuration required to get ProxLB running in a container. Simply follow the steps in the Dockerfile to build the image, ensuring all dependencies and configurations are correctly applied. For those looking for an even quicker setup, a ready-to-use ProxLB container image is also available, eliminating the need for manual building and allowing for immediate deployment.
|
||||
|
||||
```bash
|
||||
git clone https://github.com/gyptazy/ProxLB.git
|
||||
cd ProxLB
|
||||
docker build -t proxlb .
|
||||
```
|
||||
|
||||
Afterwards simply adjust the config file to your needs:
|
||||
```
|
||||
vi /etc/proxlb/proxlb.conf
|
||||
```
|
||||
|
||||
Finally, start the created container.
|
||||
```bash
|
||||
docker run -it --rm -v $(pwd)/proxlb.conf:/etc/proxlb/proxlb.conf proxlb
|
||||
```
|
||||
|
||||
### Logging
|
||||
ProxLB uses the `SystemdHandler` for logging. You can find all your logs in your systemd unit log or in the `journalctl`. In default, ProxLB only logs critical events. However, for further understanding of the balancing it might be useful to change this to `INFO` or `DEBUG` which can simply be done in the [proxlb.conf](https://github.com/gyptazy/ProxLB/blob/main/proxlb.conf#L14) file by changing the `log_verbosity` parameter.
|
||||
|
||||
Available logging values:
|
||||
| Verbosity | Description |
|
||||
|------|:------:|
|
||||
| DEBUG | This option logs everything and is needed for debugging the code. |
|
||||
| INFO | This option provides insides behind the scenes. What/why has been something done and with which values. |
|
||||
| WARNING | This option provides only warning messages, which might be a problem in general but not for the application itself. |
|
||||
| CRITICAL | This option logs all critical events that will avoid running ProxLB. |
|
||||
|
||||
## Motivation
|
||||
As a developer managing a cluster of virtual machines for my projects, I often encountered the challenge of resource imbalance. Nodes within the cluster would become unevenly loaded, with some nodes being overburdened while others remained underutilized. This imbalance led to inefficiencies, performance bottlenecks, and increased operational costs. Frustrated by the lack of an adequate solution to address this issue, I decided to develop the ProxLB (PLB) to ensure better resource distribution across my clusters.
|
||||
|
||||
My primary motivation for creating PLB stemmed from my work on my BoxyBSD project, where I consistently faced the difficulty of maintaining balanced nodes while running various VM workloads but also on my personal clusters. The absence of an efficient rebalancing mechanism made it challenging to achieve optimal performance and stability. Recognizing the necessity for a tool that could gather and analyze resource metrics from both the cluster nodes and the running VMs, I embarked on developing ProxLB.
|
||||
|
||||
PLB meticulously collects detailed resource usage data from each node in a Proxmox cluster, including CPU load, memory usage, and local disk space utilization. It also gathers comprehensive statistics from all running VMs, providing a granular understanding of the workload distribution. With this data, PLB intelligently redistributes VMs based on memory usage, local disk usage, and CPU usage. This ensures that no single node is overburdened, storage resources are evenly distributed, and the computational load is balanced, enhancing overall cluster performance.
|
||||
|
||||
As an advocate of the open-source philosophy, I believe in the power of community and collaboration. By sharing solutions like PLB, I aim to contribute to the collective knowledge and tools available to developers facing similar challenges. Open source fosters innovation, transparency, and mutual support, enabling developers to build on each other's work and create better solutions together.
|
||||
|
||||
Developing PLB was driven by a desire to solve a real problem I faced in my projects. However, the spirit behind this effort was to provide a valuable resource to the community. By open-sourcing PLB, I hope to help other developers manage their clusters more efficiently, optimize their resource usage, and reduce operational costs. Sharing this solution aligns with the core principles of open source, where the goal is not only to solve individual problems but also to contribute to the broader ecosystem.
|
||||
|
||||
## References
|
||||
Here you can find some overviews of references for and about the ProxLB (PLB):
|
||||
|
||||
| Description | Link |
|
||||
|------|:------:|
|
||||
| General introduction into ProxLB | https://gyptazy.com/blog/proxlb-rebalancing-vm-workloads-across-nodes-in-proxmox-clusters/ |
|
||||
| Howto install and use ProxLB on Debian to rebalance vm workloads in a Proxmox cluster | https://gyptazy.com/howtos/howto-install-and-use-proxlb-to-rebalance-vm-workloads-across-nodes-in-proxmox-clusters/ |
|
||||
|
||||
## Downloads
|
||||
ProxLB can be obtained in man different ways, depending on which use case you prefer. You can use simply copy the code from GitHub, use created packages for Debian or RedHat based systems, use a Repository to keep ProxLB always up to date or simply use a Container image for Docker/Podman.
|
||||
|
||||
### Packages
|
||||
Ready to use packages can be found at:
|
||||
* https://cdn.gyptazy.com/files/os/debian/proxlb/
|
||||
* https://cdn.gyptazy.com/files/os/ubuntu/proxlb/
|
||||
* https://cdn.gyptazy.com/files/os/redhat/proxlb/
|
||||
|
||||
### Repository
|
||||
Debian based systems can also use the repository by adding the following line to their apt sources:
|
||||
|
||||
#### Stable Releases
|
||||
```
|
||||
deb https://repo.gyptazy.com/stable /
|
||||
```
|
||||
|
||||
#### Beta/Testing Releases
|
||||
```
|
||||
deb https://repo.gyptazy.com/testing /
|
||||
```
|
||||
|
||||
The Repository's GPG key can be found at: `https://repo.gyptazy.com/repository.gpg`
|
||||
|
||||
You can also simply import it by running:
|
||||
|
||||
```
|
||||
# KeyID: 17169F23F9F71A14AD49EDADDB51D3EB01824F4C
|
||||
# UID: gyptazy Solutions Repository <contact@gyptazy.com>
|
||||
# SHA256: 52c267e6f4ec799d40cdbdb29fa518533ac7942dab557fa4c217a76f90d6b0f3 repository.gpg
|
||||
|
||||
wget -O /etc/apt/trusted.gpg.d/proxlb.asc https://repo.gyptazy.com/repository.gpg
|
||||
```
|
||||
|
||||
*Note: The defined repositories `repo.gyptazy.com` and `repo.proxlb.de` are the same!*
|
||||
|
||||
### Container Images (Docker/Podman)
|
||||
Container Images for Podman, Docker etc., can be found at:
|
||||
| Version | Image |
|
||||
|------|:------:|
|
||||
| latest | cr.gyptazy.com/proxlb/proxlb:latest |
|
||||
| v1.0.6 | cr.gyptazy.com/proxlb/proxlb:v1.0.6 |
|
||||
| v1.0.5 | cr.gyptazy.com/proxlb/proxlb:v1.0.5 |
|
||||
| v1.0.4 | cr.gyptazy.com/proxlb/proxlb:v1.0.4 |
|
||||
| v1.0.3 | cr.gyptazy.com/proxlb/proxlb:v1.0.3 |
|
||||
| v1.0.2 | cr.gyptazy.com/proxlb/proxlb:v1.0.2 |
|
||||
| v1.0.0 | cr.gyptazy.com/proxlb/proxlb:v1.0.0 |
|
||||
| v0.9.9 | cr.gyptazy.com/proxlb/proxlb:v0.9.9 |
|
||||
The `maintenance_nodes` option allows operators to designate one or more Proxmox nodes for maintenance mode. When a node is set to maintenance, no new guest workloads will be assigned to it, and all existing workloads will be migrated to other available nodes within the cluster. This process ensures that (anti)-affinity rules and resource availability are respected, preventing disruptions while maintaining optimal performance across the infrastructure.
|
||||
|
||||
## Misc
|
||||
### Bugs
|
||||
Bugs can be reported via the GitHub issue tracker [here](https://github.com/gyptazy/ProxLB/issues). You may also report bugs via email or deliver PRs to fix them on your own. Therefore, you might also see the contributing chapter.
|
||||
|
||||
### Contributing
|
||||
Feel free to add further documentation, to adjust already existing one or to contribute with code. Please take care about the style guide and naming conventions. You can find more in our [CONTRIBUTING.md](https://github.com/gyptazy/ProxLB/blob/main/CONTRIBUTING.md) file.
|
||||
Feel free to add further documentation, to adjust already existing one or to contribute with code. Please take care about the style guide and naming conventions. You can find more in our [CONTRIBUTING.md](https://github.com/gyptazy/ProxLB/blob/development/CONTRIBUTING.md) file.
|
||||
|
||||
### Documentation
|
||||
You can also find additional and more detailed documentation within the [docs/](https://github.com/gyptazy/ProxLB/tree/main/docs) directory.
|
||||
You can also find additional and more detailed documentation within the [docs/](https://github.com/gyptazy/ProxLB/tree/development/docs) directory.
|
||||
|
||||
### Support
|
||||
If you need assistance or have any questions, we offer support through our dedicated [chat room](https://matrix.to/#/#proxlb:gyptazy.com) in Matrix and on Reddit. Join our community for real-time help, advice, and discussions. Connect with us in our dedicated chat room for immediate support and live interaction with other users and developers. You can also visit our [GitHub Community](https://github.com/gyptazy/ProxLB/discussions/) to post your queries, share your experiences, and get support from fellow community members and moderators. You may also just open directly an issue [here](https://github.com/gyptazy/ProxLB/issues) on GitHub. We are here to help and ensure you have the best experience possible.
|
||||
If you need assistance or have any questions, we offer support through our dedicated [chat room](https://matrix.to/#/#proxlb:gyptazy.com) in Matrix or [Discord](https://discord.gg/JemGu7WbfQ). Join our community for real-time help, advice, and discussions. The Matrix and Discord room are bridged to ensure that the communication is not splitted - so simply feel free to join which fits most to you!
|
||||
|
||||
Connect with us in our dedicated chat room for immediate support and live interaction with other users and developers. You can also visit our [GitHub Community](https://github.com/gyptazy/ProxLB/discussions/) to post your queries, share your experiences, and get support from fellow community members and moderators. You may also just open directly an issue [here](https://github.com/gyptazy/ProxLB/issues) on GitHub.
|
||||
|
||||
| Support Channel | Link |
|
||||
|------|:------:|
|
||||
| Matrix | [#proxlb:gyptazy.com](https://matrix.to/#/#proxlb:gyptazy.com) |
|
||||
| Discord | [Discord](https://discord.gg/JemGu7WbfQ) |
|
||||
| GitHub Community | [GitHub Community](https://github.com/gyptazy/ProxLB/discussions/)
|
||||
| GitHub | [ProxLB GitHub](https://github.com/gyptazy/ProxLB/issues) |
|
||||
|
||||
**Note:** Please always keep in mind that this is a one-man show project without any further help. This includes coding, testing, packaging and all the infrastructure around it to keep this project up and running.
|
||||
|
||||
### Author(s)
|
||||
* Florian Paul Azim Hoberg @gyptazy (https://gyptazy.com)
|
||||
* Florian Paul Azim Hoberg @gyptazy (https://gyptazy.com)
|
||||
28
config/proxlb_example.yaml
Normal file
28
config/proxlb_example.yaml
Normal file
@@ -0,0 +1,28 @@
|
||||
proxmox_api:
|
||||
hosts: ['virt01.example.com', '10.10.10.10', 'fe01::bad:code::cafe']
|
||||
user: root@pam
|
||||
pass: crazyPassw0rd!
|
||||
ssl_verification: False
|
||||
timeout: 10
|
||||
|
||||
proxmox_cluster:
|
||||
maintenance_nodes: ['virt66.example.com']
|
||||
ignore_nodes: []
|
||||
overprovisioning: True
|
||||
|
||||
balancing:
|
||||
enable: True
|
||||
force: False
|
||||
parallel: False
|
||||
live: True
|
||||
with_local_disks: True
|
||||
balance_types: ['vm', 'ct']
|
||||
max_job_validation: 1800
|
||||
balanciness: 5
|
||||
method: memory
|
||||
mode: assigned
|
||||
|
||||
service:
|
||||
daemon: False
|
||||
schedule: 12
|
||||
log_level: DEBUG
|
||||
@@ -1,7 +0,0 @@
|
||||
services:
|
||||
proxlb:
|
||||
build: .
|
||||
volumes:
|
||||
- ./proxlb.conf:/etc/proxlb/proxlb.conf
|
||||
restart: unless-stopped
|
||||
container_name: proxlb
|
||||
@@ -1,33 +0,0 @@
|
||||
# Installation
|
||||
|
||||
## Packages
|
||||
The easiest way to get started is by using the ready-to-use packages that I provide on my CDN and to run it on a Linux Debian based system. This can also be one of the Proxmox nodes itself.
|
||||
|
||||
```
|
||||
wget https://cdn.gyptazy.ch/files/amd64/debian/proxlb/proxlb_0.9.9_amd64.deb
|
||||
dpkg -i proxlb_0.9.9_amd64.deb
|
||||
# Adjust your config
|
||||
vi /etc/proxlb/proxlb.conf
|
||||
# Enable and start the service
|
||||
systemctl enable --now proxlb
|
||||
systemctl status proxlb
|
||||
```
|
||||
|
||||
## Container (Docker/Podman)
|
||||
Creating a container image of ProxLB is straightforward using the provided Dockerfile. The Dockerfile simplifies the process by automating the setup and configuration required to get ProxLB running in a container. Simply follow the steps in the Dockerfile to build the image, ensuring all dependencies and configurations are correctly applied. For those looking for an even quicker setup, a ready-to-use ProxLB container image is also available, eliminating the need for manual building and allowing for immediate deployment.
|
||||
|
||||
```bash
|
||||
git clone https://github.com/gyptazy/ProxLB.git
|
||||
cd ProxLB
|
||||
build -t proxlb .
|
||||
```
|
||||
|
||||
Afterwards simply adjust the config file to your needs:
|
||||
```
|
||||
vi /etc/proxlb/proxlb.conf
|
||||
```
|
||||
|
||||
Finally, start the created container.
|
||||
```bash
|
||||
docker run -it --rm -v $(pwd)/proxlb.conf:/etc/proxlb/proxlb.conf proxlb
|
||||
```
|
||||
@@ -1,48 +0,0 @@
|
||||
# Configuration
|
||||
|
||||
## Balancing
|
||||
### By Used Memmory of VMs
|
||||
By continuously monitoring the current resource usage of VMs, ProxLB intelligently reallocates workloads to prevent any single node from becoming overloaded. This approach ensures that resources are balanced efficiently, providing consistent and optimal performance across the entire cluster at all times. To activate this balancing mode, simply activate the following option in your ProxLB configuration:
|
||||
```
|
||||
mode: used
|
||||
```
|
||||
Afterwards, restart the service (if running in daemon mode) to activate this rebalancing mode.
|
||||
|
||||
### By Assigned Memory of VMs
|
||||
By ensuring that resources are always available for each VM, ProxLB prevents over-provisioning and maintains a balanced load across all nodes. This guarantees that users have consistent access to the resources they need. However, if the total assigned resources exceed the combined capacity of the cluster, ProxLB will issue a warning, indicating potential over-provisioning despite its best efforts to balance the load. To activate this balancing mode, simply activate the following option in your ProxLB configuration:
|
||||
```
|
||||
mode: assigned
|
||||
```
|
||||
Afterwards, restart the service (if running in daemon mode) to activate this rebalancing mode.
|
||||
|
||||
## Grouping
|
||||
### Include (Stay Together)
|
||||
<img align="left" src="https://cdn.gyptazy.ch/images/plb-rebalancing-include-balance-group.jpg"/> Access the Proxmox Web UI by opening your web browser and navigating to your Proxmox VE web interface, then log in with your credentials. Navigate to the VM you want to tag by selecting it from the left-hand navigation panel. Click on the "Options" tab to view the VM's options, then select "Edit" or "Add" (depending on whether you are editing an existing tag or adding a new one). In the tag field, enter plb_include_ followed by your unique identifier, for example, plb_include_group1. Save the changes to apply the tag to the VM. Repeat these steps for each VM that should be included in the group.
|
||||
|
||||
### Exclude (Stay Separate)
|
||||
<img align="left" src="https://cdn.gyptazy.ch/images/plb-rebalancing-exclude-balance-group.jpg"/> Access the Proxmox Web UI by opening your web browser and navigating to your Proxmox VE web interface, then log in with your credentials. Navigate to the VM you want to tag by selecting it from the left-hand navigation panel. Click on the "Options" tab to view the VM's options, then select "Edit" or "Add" (depending on whether you are editing an existing tag or adding a new one). In the tag field, enter plb_exclude_ followed by your unique identifier, for example, plb_exclude_critical. Save the changes to apply the tag to the VM. Repeat these steps for each VM that should be excluded from being on the same node.
|
||||
|
||||
### Ignore VMs (tag style)
|
||||
<img align="left" src="https://cdn.gyptazy.ch/images/plb-rebalancing-ignore-vm.jpg"/> In Proxmox, you can ensure that certain VMs are ignored during the rebalancing process by setting a specific tag within the Proxmox Web UI, rather than solely relying on configurations in the ProxLB config file. This can be achieved by adding the tag 'plb_ignore_vm' to the VM. Once this tag is applied, the VM will be excluded from any further rebalancing operations, simplifying the management process.
|
||||
|
||||
## Authentication / User Account / User / Permissions
|
||||
### Authentication
|
||||
ProxLB also supports different accounts in ProxLB. Therefore, you can simply create a new user and group and add the required roles permissions.
|
||||
|
||||
### Creating Dedicated User for Balanciung
|
||||
It is recommended to not use the `root@pam` user for balancing. Therefore, creating a new user might be suitable and is very easy to create.
|
||||
A new user can be created by the gui, api and cli. The required roles are stated in the next chapter, but you can also use the following lines
|
||||
to create a user on the cli with the required roles to balance VMs & CTs.
|
||||
|
||||
```
|
||||
pveum role add ProxLBAdmin --privs Datastore.Audit,Sys.Audit,VM.Audit,VM.Migrate
|
||||
pveum user add proxlb_admin@pve --password <password>
|
||||
pveum acl modify / --roles ProxLBAdmin --users proxlb_admin@pve
|
||||
```
|
||||
|
||||
### Required Roles
|
||||
When using ProxLB with a dedicated account, you might also keep the assigned roles low. Therefore, you need to ensure that the newly created user is at least assigned to the following roles:
|
||||
* Datastore.Audit (Required for storage evaluation)
|
||||
* Sys.Audit (Required to get resource metrics of the nodes)
|
||||
* VM.Audit (Requited to get resource metrics of VMs/CTs)
|
||||
* VM.Migrate (Required for migration of VMs/CTs)
|
||||
@@ -1,87 +0,0 @@
|
||||
## FAQ
|
||||
|
||||
### Could not import all dependencies
|
||||
ProxLB requires the Python library `proxmoxer`. This can simply be installed by the most
|
||||
system repositories. If you encounter this error message you simply need to install it.
|
||||
|
||||
|
||||
```
|
||||
# systemctl status proxlb
|
||||
x proxlb.service - Proxmox Rebalancing Service
|
||||
Loaded: loaded (/etc/systemd/system/proxlb.service; static)
|
||||
Active: failed (Result: exit-code) since Sat 2024-07-06 10:25:16 UTC; 1s ago
|
||||
Duration: 239ms
|
||||
Process: 7285 ExecStart=/usr/bin/proxlb -c /etc/proxlb/proxlb.conf (code=exited, status=2)
|
||||
Main PID: 7285 (code=exited, status=2)
|
||||
CPU: 129ms
|
||||
|
||||
Jul 06 10:25:16 build01 systemd[1]: Started proxlb.service - ProxLB.
|
||||
Jul 06 10:25:16 build01 proxlb[7285]: proxlb: Error: [python-imports]: Could not import all dependencies. Please install "proxmoxer".
|
||||
```
|
||||
|
||||
Debian/Ubuntu: apt-get install python3-proxmoxer
|
||||
If the package is not provided by your systems repository, you can also install it by running `pip3 install proxmoxer`.
|
||||
|
||||
### How does it work?
|
||||
ProxLB is a load-balancing system designed to optimize the distribution of virtual machines (VMs) and containers (CTs) across a cluster. It works by first gathering resource usage metrics from all nodes in the cluster through the Proxmox API. This includes detailed resource metrics for each VM and CT on every node. ProxLB then evaluates the difference between the maximum and minimum resource usage of the nodes, referred to as "Balanciness." If this difference exceeds a predefined threshold (which is configurable), the system initiates the rebalancing process.
|
||||
|
||||
Before starting any migrations, ProxLB validates that rebalancing actions are necessary and beneficial. Depending on the selected balancing mode — such as CPU, memory, or disk — it creates a balancing matrix. This matrix sorts the VMs by their maximum used or assigned resources, identifying the VM with the highest usage. ProxLB then places this VM on the node with the most free resources in the selected balancing type. This process runs recursively until the operator-defined Balanciness is achieved. Balancing can be defined for the used or max. assigned resources of VMs/CTs.
|
||||
|
||||
### ProxLB config version is too low
|
||||
ProxLB may run into an error when the used config schema version is too low. This might happen after major changes that require new config options. Please make sure, to use a supported config version in addition to your running ProxLB config.
|
||||
|
||||
Example Error:
|
||||
```
|
||||
Error: [config-version-validator]: ProxLB config version 2 is too low. Required: 3.
|
||||
```
|
||||
|
||||
The easiest way to solve this, is by taking the minimum required config schema version from a git tag, representing the ProxLB version.
|
||||
|
||||
### Logging
|
||||
ProxLB uses the `SystemdHandler` for logging. You can find all your logs in your systemd unit log or in the `journalctl`. In default, ProxLB only logs critical events. However, for further understanding of the balancing it might be useful to change this to `INFO` or `DEBUG` which can simply be done in the [proxlb.conf](https://github.com/gyptazy/ProxLB/blob/main/proxlb.conf#L14) file by changing the `log_verbosity` parameter.
|
||||
|
||||
Available logging values:
|
||||
| Verbosity | Description |
|
||||
|------|:------:|
|
||||
| DEBUG | This option logs everything and is needed for debugging the code. |
|
||||
| INFO | This option provides insides behind the scenes. What/why has been something done and with which values. |
|
||||
| WARNING | This option provides only warning messages, which might be a problem in general but not for the application itself. |
|
||||
| CRITICAL | This option logs all critical events that will avoid running ProxLB. |
|
||||
|
||||
### Motivation
|
||||
As a developer managing a cluster of virtual machines for my projects, I often encountered the challenge of resource imbalance. Nodes within the cluster would become unevenly loaded, with some nodes being overburdened while others remained underutilized. This imbalance led to inefficiencies, performance bottlenecks, and increased operational costs. Frustrated by the lack of an adequate solution to address this issue, I decided to develop the ProxLB (PLB) to ensure better resource distribution across my clusters.
|
||||
|
||||
My primary motivation for creating PLB stemmed from my work on my BoxyBSD project, where I consistently faced the difficulty of maintaining balanced nodes while running various VM workloads but also on my personal clusters. The absence of an efficient rebalancing mechanism made it challenging to achieve optimal performance and stability. Recognizing the necessity for a tool that could gather and analyze resource metrics from both the cluster nodes and the running VMs, I embarked on developing ProxLB.
|
||||
|
||||
PLB meticulously collects detailed resource usage data from each node in a Proxmox cluster, including CPU load, memory usage, and local disk space utilization. It also gathers comprehensive statistics from all running VMs, providing a granular understanding of the workload distribution. With this data, PLB intelligently redistributes VMs based on memory usage, local disk usage, and CPU usage. This ensures that no single node is overburdened, storage resources are evenly distributed, and the computational load is balanced, enhancing overall cluster performance.
|
||||
|
||||
As an advocate of the open-source philosophy, I believe in the power of community and collaboration. By sharing solutions like PLB, I aim to contribute to the collective knowledge and tools available to developers facing similar challenges. Open source fosters innovation, transparency, and mutual support, enabling developers to build on each other's work and create better solutions together.
|
||||
|
||||
Developing PLB was driven by a desire to solve a real problem I faced in my projects. However, the spirit behind this effort was to provide a valuable resource to the community. By open-sourcing PLB, I hope to help other developers manage their clusters more efficiently, optimize their resource usage, and reduce operational costs. Sharing this solution aligns with the core principles of open source, where the goal is not only to solve individual problems but also to contribute to the broader ecosystem.
|
||||
|
||||
### Packages / Container Images
|
||||
Ready to use packages can be found at:
|
||||
* https://cdn.gyptazy.ch/files/amd64/debian/proxlb/
|
||||
* https://cdn.gyptazy.ch/files/amd64/ubuntu/proxlb/
|
||||
* https://cdn.gyptazy.ch/files/amd64/redhat/proxlb/
|
||||
* https://cdn.gyptazy.ch/files/amd64/freebsd/proxlb/
|
||||
|
||||
Container Images for Podman, Docker etc., can be found at:
|
||||
| Version | Image |
|
||||
|------|:------:|
|
||||
| latest | cr.gyptazy.ch/proxlb/proxlb:latest |
|
||||
|
||||
### Bugs
|
||||
Bugs can be reported via the GitHub issue tracker [here](https://github.com/gyptazy/ProxLB/issues). You may also report bugs via email or deliver PRs to fix them on your own. Therefore, you might also see the contributing chapter.
|
||||
|
||||
### Contributing
|
||||
Feel free to add further documentation, to adjust already existing one or to contribute with code. Please take care about the style guide and naming conventions. You can find more in our [CONTRIBUTING.md](https://github.com/gyptazy/ProxLB/blob/main/CONTRIBUTING.md) file.
|
||||
|
||||
### Support
|
||||
If you need assistance or have any questions, we offer support through our dedicated [chat room](https://matrix.to/#/#proxlb:gyptazy.ch) in Matrix and on Reddit. Join our community for real-time help, advice, and discussions. Connect with us in our dedicated chat room for immediate support and live interaction with other users and developers. You can also visit our [Reddit community](https://www.reddit.com/r/Proxmox/comments/1e78ap3/introducing_proxlb_rebalance_your_vm_workloads/) to post your queries, share your experiences, and get support from fellow community members and moderators. You may also just open directly an issue [here](https://github.com/gyptazy/ProxLB/issues) on GitHub. We are here to help and ensure you have the best experience possible.
|
||||
|
||||
| Support Channel | Link |
|
||||
|------|:------:|
|
||||
| Matrix | [#proxlb:gyptazy.ch](https://matrix.to/#/#proxlb:gyptazy.ch) |
|
||||
| Reddit | [Reddit community](https://www.reddit.com/r/Proxmox/comments/1e78ap3/introducing_proxlb_rebalance_your_vm_workloads/) |
|
||||
| GitHub | [ProxLB GitHub](https://github.com/gyptazy/ProxLB/issues) |
|
||||
6
misc/01-replace-version.sh
Normal file
6
misc/01-replace-version.sh
Normal file
@@ -0,0 +1,6 @@
|
||||
#!/usr/bin/env bash
|
||||
VERSION="1.1.0-alpha"
|
||||
|
||||
sed -i "s/^__version__ = .*/__version__ = \"$VERSION\"/" "proxlb/utils/version.py"
|
||||
sed -i "s/version=\"[0-9]*\.[0-9]*\.[0-9]*\"/version=\"$VERSION\"/" setup.py
|
||||
echo "OK: Versions have been sucessfully set to $VERSION"
|
||||
4
misc/02-create-changelog.sh
Normal file
4
misc/02-create-changelog.sh
Normal file
@@ -0,0 +1,4 @@
|
||||
#!/usr/bin/env bash
|
||||
git clone https://github.com/gyptazy/changelog-fragments-creator.git
|
||||
./changelog-fragments-creator/changelog-creator -f .changelogs/ -o CHANGELOG.md
|
||||
echo "Created changelog file"
|
||||
@@ -1,16 +0,0 @@
|
||||
#!/bin/bash
|
||||
sudo apt-get install rpm cmake git make python3-yaml
|
||||
|
||||
git clone https://github.com/gyptazy/changelog-fragments-creator.git
|
||||
./changelog-fragments-creator/changelog-creator -f ../.changelogs/ -o ../CHANGELOG.md
|
||||
mkdir packages
|
||||
mkdir build
|
||||
cd build
|
||||
cmake ..
|
||||
cpack -G DEB .
|
||||
cpack -G RPM .
|
||||
cp *.deb ../packages
|
||||
cp *.rpm ../packages
|
||||
cd ..
|
||||
rm -rf build
|
||||
echo "Packages created. Packages can be found in directory: packages"
|
||||
@@ -1,4 +0,0 @@
|
||||
#!/bin/bash
|
||||
git clone https://github.com/gyptazy/changelog-fragments-creator.git
|
||||
./changelog-fragments-creator/changelog-creator -f ../.changelogs/ -o ../CHANGELOG.md
|
||||
echo "Created changelog file"
|
||||
@@ -1,40 +0,0 @@
|
||||
cmake_minimum_required(VERSION 3.16)
|
||||
project(proxmox-rebalancing-service VERSION 1.0.6)
|
||||
|
||||
install(PROGRAMS ../proxlb DESTINATION /bin)
|
||||
install(FILES ../proxlb.conf DESTINATION /etc/proxlb)
|
||||
install(FILES proxlb.service DESTINATION /etc/systemd/system)
|
||||
|
||||
# General
|
||||
set(CPACK_PACKAGE_NAME "proxlb")
|
||||
set(CPACK_RESOURCE_FILE_LICENSE "${CMAKE_CURRENT_SOURCE_DIR}/../LICENSE")
|
||||
set(CPACK_RESOURCE_FILE_README "${CMAKE_CURRENT_SOURCE_DIR}/../README.md")
|
||||
set(CPACK_DEBIAN_PACKAGE_MAINTAINER "Florian Paul Azim <gyptazy> Hoberg <gyptazy@gyptazy.com>")
|
||||
set(CPACK_PACKAGE_CONTACT "Florian Paul Azim Hoberg <gyptazy@gyptazy.com>")
|
||||
set(CPACK_PACKAGE_VENDOR "gyptazy")
|
||||
|
||||
# RPM packaging
|
||||
set(CPACK_PACKAGE_VERSION ${CMAKE_PROJECT_VERSION})
|
||||
set(CPACK_GENERATOR "RPM")
|
||||
set(CPACK_RPM_PACKAGE_ARCHITECTURE "amd64")
|
||||
set(CPACK_RPM_PACKAGE_SUMMARY "ProxLB - Rebalance VM workloads across nodes in Proxmox clusters.")
|
||||
set(CPACK_RPM_PACKAGE_DESCRIPTION "ProxLB - Rebalance VM workloads across nodes in Proxmox clusters.")
|
||||
set(CPACK_RPM_CHANGELOG_FILE "${CMAKE_CURRENT_SOURCE_DIR}/changelog_redhat")
|
||||
set(CPACK_PACKAGE_RELEASE 1)
|
||||
set(CPACK_RPM_PACKAGE_LICENSE "GPL 3.0")
|
||||
set(CPACK_RPM_PACKAGE_REQUIRES "python >= 3.2.0")
|
||||
|
||||
# DEB packaging
|
||||
set(CPACK_DEBIAN_FILE_NAME DEB-DEFAULT)
|
||||
set(CPACK_DEBIAN_PACKAGE_ARCHITECTURE "amd64")
|
||||
set(CPACK_DEBIAN_PACKAGE_SUMMARY "ProxLB - Rebalance VM workloads across nodes in Proxmox clusters.")
|
||||
set(CPACK_DEBIAN_PACKAGE_DESCRIPTION "ProxLB - Rebalance VM workloads across nodes in Proxmox clusters.")
|
||||
set(CPACK_DEBIAN_PACKAGE_CONTROL_EXTRA "${CMAKE_CURRENT_SOURCE_DIR}/changelog_debian")
|
||||
set(CPACK_DEBIAN_PACKAGE_DEPENDS "python3, python3-proxmoxer")
|
||||
set(CPACK_DEBIAN_PACKAGE_LICENSE "GPL 3.0")
|
||||
|
||||
# Install
|
||||
set(CPACK_PACKAGING_INSTALL_PREFIX ${CMAKE_INSTALL_PREFIX})
|
||||
set(CPACK_DEBIAN_PACKAGE_CONTROL_EXTRA "${CMAKE_CURRENT_SOURCE_DIR}/postinst;${CMAKE_CURRENT_SOURCE_DIR}/conffiles")
|
||||
set(CPACK_RPM_POST_INSTALL_SCRIPT_FILE "${CMAKE_CURRENT_SOURCE_DIR}/postinst")
|
||||
include(CPack)
|
||||
@@ -1,14 +0,0 @@
|
||||
## Build packages
|
||||
Building the packages requires cmake, deb and rpm.
|
||||
For building packages, simly run the following commands:
|
||||
|
||||
```
|
||||
mkdir build
|
||||
cd build
|
||||
cmake ..
|
||||
cpack -G RPM .
|
||||
cpack -G DEB .
|
||||
```
|
||||
|
||||
When running on Debian/Ubuntu you can directly call `01_package.sh`
|
||||
to create your own packages.
|
||||
@@ -1,63 +0,0 @@
|
||||
proxlb (1.0.5) unstable; urgency=low
|
||||
|
||||
* Fix migration from local disks.
|
||||
* Fix allowed values (add DEBUG, WARNING) for log verbosity.
|
||||
* Fix node (and its objects) evaluation when not reachable (e.g., maintenance).
|
||||
* Fix evaluation of maintenance mode where comparing list & string resulted in a crash.
|
||||
* Change docs to make bool usage in configs more clear.
|
||||
|
||||
-- Florian Paul Azim Hoberg <gyptazy@gyptazy.com> Wed, 30 Oct 2024 17:02:31 +0100
|
||||
|
||||
proxlb (1.0.4) unstable; urgency=low
|
||||
|
||||
* Add feature to make API timeout configureable.
|
||||
* Add maintenance mode to evacuate a node and move workloads for other nodes in the cluster.
|
||||
* Add version output cli arg.
|
||||
* Run storage balancing only on supported shared storages.
|
||||
* Run storage balancing only when needed to save time.
|
||||
* Fix CPU balancing where calculations are done in float instead of int. (by @glitchvern)
|
||||
* Fix documentation for the underlying infrastructure.
|
||||
|
||||
-- Florian Paul Azim Hoberg <gyptazy@gyptazy.com> Fri, 11 Oct 2024 06:14:13 +0200
|
||||
|
||||
proxlb (1.0.3) unstable; urgency=low
|
||||
|
||||
* Add a convert function to cast all bool alike options from configparser to bools.
|
||||
* Add a config parser options for future features.
|
||||
* Add a config versio schema that must be supported by ProxLB.
|
||||
* Add feature to allow the API hosts being provided as a comma separated list.
|
||||
* Add storage balancing function.
|
||||
* Add doc how to add dedicated user for authentication. (by @Dulux-Oz)
|
||||
* Add cli arg `-b` to return the next best node for next VM/CT placement.Fix some wonkey code styles.
|
||||
* Provide a more reasonable output when HA services are not active in a Proxmox cluster.
|
||||
* Improve the underlying code base for future implementations.
|
||||
* Fix documentation for the master_only parameter placed in the wrong config section.
|
||||
* Fixed `master_only` function by inverting the condition.
|
||||
* Improved the overall validation and error handling.
|
||||
* Fix bug in the `proxlb.conf` in the vm_balancing section.
|
||||
* Fix handling of unset `ignore_nodes` and `ignore_vms` resulted in an attribute error.
|
||||
* Fix anti-affinity rules not evaluating a new and different node.
|
||||
|
||||
-- Florian Paul Azim Hoberg <gyptazy@gyptazy.com> Wed, 11 Sep 2024 17:31:03 +0200
|
||||
|
||||
proxlb (1.0.2) unstable; urgency=low
|
||||
|
||||
* Add option to run migration in parallel or sequentially.
|
||||
* Add option to run ProxLB only on a Proxmox cluster master (req. HA feature).
|
||||
* Fix daemon timer to use hours instead of minutes.
|
||||
* Fix CMake packaging for Debian package to avoid overwriting the config file.
|
||||
* Fix some wonkey code styles.
|
||||
|
||||
-- Florian Paul Azim Hoberg <gyptazy@gyptazy.com> Tue, 13 Aug 2024 17:28:14 +0200
|
||||
|
||||
proxlb (1.0.0) unstable; urgency=low
|
||||
|
||||
* Initial release of ProxLB.
|
||||
|
||||
-- Florian Paul Azim Hoberg <gyptazy@gyptazy.com> Thu, 01 Aug 2024 17:04:12 +0200
|
||||
|
||||
proxlb (0.9.0) unstable; urgency=low
|
||||
|
||||
* Initial development release of ProxLB as a tech preview.
|
||||
|
||||
-- Florian Paul Azim Hoberg <gyptazy@gyptazy.com> Sun, 07 Jul 2024 05:38:41 +0200
|
||||
@@ -1,44 +0,0 @@
|
||||
* Wed Oct 30 2024 Florian Paul Azim Hoberg <gyptazy@gyptazy.com>
|
||||
- Fix migration from local disks.
|
||||
- Fix allowed values (add DEBUG, WARNING) for log verbosity.
|
||||
- Fix node (and its objects) evaluation when not reachable (e.g., maintenance).
|
||||
- Fix evaluation of maintenance mode where comparing list & string resulted in a crash.
|
||||
- Change docs to make bool usage in configs more clear.
|
||||
|
||||
* Fri Oct 11 2024 Florian Paul Azim Hoberg <gyptazy@gyptazy.com>
|
||||
- Add feature to make API timeout configureable.
|
||||
- Add maintenance mode to evacuate a node and move workloads for other nodes in the cluster.
|
||||
- Add version output cli arg.
|
||||
- Run storage balancing only on supported shared storages.
|
||||
- Run storage balancing only when needed to save time.
|
||||
- Fix CPU balancing where calculations are done in float instead of int. (by @glitchvern)
|
||||
- Fix documentation for the underlying infrastructure.
|
||||
|
||||
* Wed Sep 12 2024 Florian Paul Azim Hoberg <gyptazy@gyptazy.com>
|
||||
- Add a convert function to cast all bool alike options from configparser to bools.
|
||||
- Add a config parser options for future features.
|
||||
- Add a config versio schema that must be supported by ProxLB.
|
||||
- Add feature to allow the API hosts being provided as a comma separated list.
|
||||
- Add storage balancing function.
|
||||
- Add doc how to add dedicated user for authentication. (by @Dulux-Oz)
|
||||
- Add cli arg `-b` to return the next best node for next VM/CT placement.Fix some wonkey code styles.
|
||||
- Provide a more reasonable output when HA services are not active in a Proxmox cluster.
|
||||
- Improve the underlying code base for future implementations.
|
||||
- Fix documentation for the master_only parameter placed in the wrong config section.
|
||||
- Fixed `master_only` function by inverting the condition.
|
||||
- Improved the overall validation and error handling.
|
||||
- Fix bug in the `proxlb.conf` in the vm_balancing section.
|
||||
- Fix handling of unset `ignore_nodes` and `ignore_vms` resulted in an attribute error.
|
||||
- Fix anti-affinity rules not evaluating a new and different node.
|
||||
|
||||
* Tue Aug 13 2024 Florian Paul Azim Hoberg <gyptazy@gyptazy.com>
|
||||
- Add option to run migration in parallel or sequentially.
|
||||
- Add option to run ProxLB only on a Proxmox cluster master (req. HA feature).
|
||||
- Fixed daemon timer to use hours instead of minutes.
|
||||
- Fixed some wonkey code styles.
|
||||
|
||||
* Thu Aug 01 2024 Florian Paul Azim Hoberg <gyptazy@gyptazy.com>
|
||||
- Initial release of ProxLB.
|
||||
|
||||
* Sun Jul 07 2024 Florian Paul Azim Hoberg <gyptazy@gyptazy.com>
|
||||
- Initial development release of ProxLB as a tech preview.
|
||||
@@ -1 +0,0 @@
|
||||
/etc/proxlb/proxlb.conf
|
||||
@@ -1,5 +0,0 @@
|
||||
#!/bin/bash
|
||||
useradd -m plb
|
||||
chown plb:plb /etc/proxlb/proxlb.conf
|
||||
chmod 600 /etc/proxlb/proxlb.conf
|
||||
systemctl daemon-reload
|
||||
@@ -1,11 +0,0 @@
|
||||
[Unit]
|
||||
Description=ProxLB - Rebalance VM workloads
|
||||
After=network-online.target
|
||||
Wants=network-online.target
|
||||
|
||||
[Service]
|
||||
ExecStart=/usr/bin/proxlb -c /etc/proxlb/proxlb.conf
|
||||
User=plb
|
||||
|
||||
[Install]
|
||||
WantedBy=multi-user.target
|
||||
23
proxlb.conf
23
proxlb.conf
@@ -1,23 +0,0 @@
|
||||
[proxmox]
|
||||
api_host: hypervisor01.gyptazy.ch
|
||||
api_user: root@pam
|
||||
api_pass: FooBar
|
||||
verify_ssl: 1
|
||||
[vm_balancing]
|
||||
enable: 1
|
||||
method: memory
|
||||
mode: used
|
||||
maintenance_nodes: dummynode03,dummynode04
|
||||
ignore_nodes: dummynode01,dummynode02
|
||||
ignore_vms: testvm01,testvm02
|
||||
[storage_balancing]
|
||||
enable: 0
|
||||
[update_service]
|
||||
enable: 0
|
||||
[api]
|
||||
enable: 0
|
||||
[service]
|
||||
daemon: 1
|
||||
schedule: 24
|
||||
log_verbosity: CRITICAL
|
||||
config_version: 3
|
||||
79
proxlb/main.py
Normal file
79
proxlb/main.py
Normal file
@@ -0,0 +1,79 @@
|
||||
"""
|
||||
ProxLB is a load balancing tool for Proxmox Virtual Environment (PVE) clusters.
|
||||
It connects to the Proxmox API, retrieves information about nodes, guests, and groups,
|
||||
and performs calculations to determine the optimal distribution of resources across the
|
||||
cluster. The tool supports daemon mode for continuous operation and can log metrics and
|
||||
perform balancing actions based on the configuration provided. It also includes a CLI
|
||||
parser for handling command-line arguments and a custom logger for systemd integration.
|
||||
"""
|
||||
|
||||
import logging
|
||||
from utils.logger import SystemdLogger
|
||||
from utils.cli_parser import CliParser
|
||||
from utils.config_parser import ConfigParser
|
||||
from utils.proxmox_api import ProxmoxApi
|
||||
from models.nodes import Nodes
|
||||
from models.guests import Guests
|
||||
from models.groups import Groups
|
||||
from models.calculations import Calculations
|
||||
from models.balancing import Balancing
|
||||
from utils.helper import Helper
|
||||
|
||||
|
||||
def main():
|
||||
"""
|
||||
ProxLB main function
|
||||
"""
|
||||
# Initialize logging handler
|
||||
logger = SystemdLogger(level=logging.INFO)
|
||||
|
||||
# Parses arguments passed from the CLI
|
||||
cli_parser = CliParser()
|
||||
cli_args = cli_parser.parse_args()
|
||||
Helper.get_version(cli_args.version)
|
||||
|
||||
# Parse ProxLB config file
|
||||
config_parser = ConfigParser(cli_args.config)
|
||||
proxlb_config = config_parser.get_config()
|
||||
|
||||
# Update log level from config and fallback to INFO if not defined
|
||||
logger.set_log_level(proxlb_config.get('service', {}).get('log_level', 'INFO'))
|
||||
|
||||
# Connect to Proxmox API & create API object
|
||||
proxmox_api = ProxmoxApi(proxlb_config)
|
||||
|
||||
# Overwrite password after creating the API object
|
||||
proxlb_config["proxmox_api"]["pass"] = "********"
|
||||
|
||||
while True:
|
||||
# Get all required objects from the Proxmox cluster
|
||||
meta = {"meta": proxlb_config}
|
||||
nodes = Nodes.get_nodes(proxmox_api, proxlb_config)
|
||||
guests = Guests.get_guests(proxmox_api, nodes)
|
||||
groups = Groups.get_groups(guests, nodes)
|
||||
|
||||
# Merge obtained objects from the Proxmox cluster for further usage
|
||||
proxlb_data = {**meta, **nodes, **guests, **groups}
|
||||
Helper.log_node_metrics(proxlb_data)
|
||||
|
||||
# Update the initial node resource assignments
|
||||
# by the previously created groups.
|
||||
Calculations.set_node_assignments(proxlb_data)
|
||||
Calculations.get_most_free_node(proxlb_data, cli_args.best_node)
|
||||
Calculations.relocate_guests_on_maintenance_nodes(proxlb_data)
|
||||
Calculations.get_balanciness(proxlb_data)
|
||||
Calculations.relocate_guests(proxlb_data)
|
||||
Helper.log_node_metrics(proxlb_data, init=False)
|
||||
|
||||
# Perform balancing actions via Proxmox API
|
||||
if not cli_args.dry_run:
|
||||
Balancing(proxmox_api, proxlb_data)
|
||||
|
||||
# Validate daemon mode
|
||||
Helper.get_daemon_mode(proxlb_config)
|
||||
|
||||
logger.debug(f"Finished: __main__")
|
||||
|
||||
|
||||
if __name__ == "__main__":
|
||||
main()
|
||||
0
proxlb/models/__init__.py
Normal file
0
proxlb/models/__init__.py
Normal file
171
proxlb/models/balancing.py
Normal file
171
proxlb/models/balancing.py
Normal file
@@ -0,0 +1,171 @@
|
||||
"""
|
||||
The balancing class is responsible for processing workloads on Proxmox clusters.
|
||||
The previously generated data (hold in proxlb_data) will processed and guests and
|
||||
other supported types will be moved across Proxmox clusters based on the defined
|
||||
values by an operator.
|
||||
"""
|
||||
|
||||
import proxmoxer
|
||||
import time
|
||||
from utils.logger import SystemdLogger
|
||||
from typing import Dict, Any
|
||||
|
||||
logger = SystemdLogger()
|
||||
|
||||
|
||||
class Balancing:
|
||||
"""
|
||||
The balancing class is responsible for processing workloads on Proxmox clusters.
|
||||
The previously generated data (hold in proxlb_data) will processed and guests and
|
||||
other supported types will be moved across Proxmox clusters based on the defined
|
||||
values by an operator.
|
||||
"""
|
||||
|
||||
def __init__(self, proxmox_api: any, proxlb_data: Dict[str, Any]):
|
||||
"""
|
||||
Initializes the Balancing class with the provided ProxLB data.
|
||||
|
||||
Args:
|
||||
proxlb_data (dict): The data required for balancing VMs and CTs.
|
||||
"""
|
||||
for guest_name, guest_meta in proxlb_data["guests"].items():
|
||||
|
||||
if guest_meta["node_current"] != guest_meta["node_target"]:
|
||||
guest_id = guest_meta["id"]
|
||||
guest_node_current = guest_meta["node_current"]
|
||||
guest_node_target = guest_meta["node_target"]
|
||||
|
||||
# VM Balancing
|
||||
if guest_meta["type"] == "vm":
|
||||
self.exec_rebalancing_vm(proxmox_api, proxlb_data, guest_name)
|
||||
|
||||
# CT Balancing
|
||||
elif guest_meta["type"] == "ct":
|
||||
self.exec_rebalancing_ct(proxmox_api, proxlb_data, guest_name)
|
||||
|
||||
# Hopefully never reaching, but should be catched
|
||||
else:
|
||||
logger.critical(f"Balancing: Got unexpected guest type: {guest_meta['type']}. Cannot proceed guest: {guest_meta['name']}.")
|
||||
|
||||
def exec_rebalancing_vm(self, proxmox_api: any, proxlb_data: Dict[str, Any], guest_name: str) -> None:
|
||||
"""
|
||||
Executes the rebalancing of a virtual machine (VM) to a new node within the cluster.
|
||||
This function initiates the migration of a specified VM to a target node as part of the
|
||||
load balancing process. It logs the migration process and handles any exceptions that
|
||||
may occur during the migration.
|
||||
Args:
|
||||
proxmox_api (object): The Proxmox API client instance used to interact with the Proxmox cluster.
|
||||
proxlb_data (dict): A dictionary containing data related to the ProxLB load balancing configuration.
|
||||
guest_name (str): The name of the guest VM to be migrated.
|
||||
Raises:
|
||||
proxmox_api.core.ResourceException: If an error occurs during the migration process.
|
||||
Returns:
|
||||
None
|
||||
"""
|
||||
logger.debug("Starting: exec_rebalancing_vm.")
|
||||
guest_id = proxlb_data["guests"][guest_name]["id"]
|
||||
guest_node_current = proxlb_data["guests"][guest_name]["node_current"]
|
||||
guest_node_target = proxlb_data["guests"][guest_name]["node_target"]
|
||||
|
||||
if proxlb_data["meta"]["balancing"].get("live", True):
|
||||
online_migration = 1
|
||||
else:
|
||||
online_migration = 0
|
||||
|
||||
if proxlb_data["meta"]["balancing"].get("with_local_disks", True):
|
||||
with_local_disks = 1
|
||||
else:
|
||||
with_local_disks = 0
|
||||
|
||||
migration_options = {
|
||||
'target': {guest_node_target},
|
||||
'online': online_migration,
|
||||
'with-local-disks': with_local_disks
|
||||
}
|
||||
|
||||
try:
|
||||
logger.debug(f"Balancing: Starting to migrate guest {guest_name} of type VM.")
|
||||
job_id = proxmox_api.nodes(guest_node_current).qemu(guest_id).migrate().post(**migration_options)
|
||||
job = self.get_rebalancing_job_status(proxmox_api, proxlb_data, guest_name, guest_node_current, job_id)
|
||||
except proxmoxer.core.ResourceException as proxmox_api_error:
|
||||
logger.critical(f"Balancing: Failed to migrate guest {guest_name} of type CT due to some Proxmox errors. Please check if resource is locked or similar.")
|
||||
|
||||
logger.debug("Finished: exec_rebalancing_vm.")
|
||||
|
||||
def exec_rebalancing_ct(self, proxmox_api: any, proxlb_data: Dict[str, Any], guest_name: str) -> None:
|
||||
"""
|
||||
Executes the rebalancing of a container (CT) to a new node within the cluster.
|
||||
This function initiates the migration of a specified CT to a target node as part of the
|
||||
load balancing process. It logs the migration process and handles any exceptions that
|
||||
may occur during the migration.
|
||||
Args:
|
||||
proxmox_api (object): The Proxmox API client instance used to interact with the Proxmox cluster.
|
||||
proxlb_data (dict): A dictionary containing data related to the ProxLB load balancing configuration.
|
||||
guest_name (str): The name of the guest CT to be migrated.
|
||||
Raises:
|
||||
proxmox_api.core.ResourceException: If an error occurs during the migration process.
|
||||
Returns:
|
||||
None
|
||||
"""
|
||||
logger.debug("Starting: exec_rebalancing_ct.")
|
||||
guest_id = proxlb_data["guests"][guest_name]["id"]
|
||||
guest_node_current = proxlb_data["guests"][guest_name]["node_current"]
|
||||
guest_node_target = proxlb_data["guests"][guest_name]["node_target"]
|
||||
|
||||
try:
|
||||
logger.debug(f"Balancing: Starting to migrate guest {guest_name} of type CT.")
|
||||
job_id = proxmox_api.nodes(guest_node_current).lxc(guest_id).migrate().post(target=guest_node_target, restart=1)
|
||||
job = self.get_rebalancing_job_status(proxmox_api, proxlb_data, guest_name, guest_node_current, job_id)
|
||||
except proxmoxer.core.ResourceException as proxmox_api_error:
|
||||
logger.critical(f"Balancing: Failed to migrate guest {guest_name} of type CT due to some Proxmox errors. Please check if resource is locked or similar.")
|
||||
|
||||
logger.debug("Finished: exec_rebalancing_ct.")
|
||||
|
||||
def get_rebalancing_job_status(self, proxmox_api: any, proxlb_data: Dict[str, Any], guest_name: str, guest_current_node: str, job_id: int, retry_counter: int = 1) -> bool:
|
||||
"""
|
||||
Monitors the status of a rebalancing job on a Proxmox node until it completes or a timeout is reached.
|
||||
|
||||
Args:
|
||||
proxmox_api (object): The Proxmox API client instance.
|
||||
proxlb_data (dict): The ProxLB configuration data.
|
||||
guest_name (str): The name of the guest (virtual machine) being rebalanced.
|
||||
guest_current_node (str): The current node where the guest is running.
|
||||
job_id (str): The ID of the rebalancing job to monitor.
|
||||
retry_counter (int, optional): The current retry count. Defaults to 1.
|
||||
|
||||
Returns:
|
||||
bool: True if the job completed successfully, False otherwise.
|
||||
"""
|
||||
logger.debug("Starting: get_rebalancing_job_status.")
|
||||
# Parallel migrations can take a huge time and create a higher load, if not defined by an
|
||||
# operator we will use a sequential mode by default
|
||||
if not proxlb_data["meta"]["balancing"].get("parallel", False):
|
||||
job = proxmox_api.nodes(guest_current_node).tasks(job_id).status().get()
|
||||
|
||||
# Watch job id until it finalizes
|
||||
if job["status"] == "running":
|
||||
# Do not hammer the API while
|
||||
# watching the job status
|
||||
time.sleep(10)
|
||||
retry_counter += 1
|
||||
|
||||
# Run recursion until we hit the soft-limit of maximum migration time for a guest
|
||||
if retry_counter < proxlb_data["meta"]["balancing"].get("max_job_validation", 1800):
|
||||
logger.debug(f"Balancing: Job ID {job_id} (guest: {guest_name}) for migration is still running... (Run: {retry_counter})")
|
||||
self.get_rebalancing_job_status(proxmox_api, proxlb_data, guest_name, guest_current_node, job_id, retry_counter)
|
||||
else:
|
||||
logger.warning(f"Balancing: Job ID {job_id} (guest: {guest_name}) for migration took too long. Please check manually.")
|
||||
logger.debug("Finished: get_rebalancing_job_status.")
|
||||
return False
|
||||
|
||||
# Validate job output for errors when finished
|
||||
if job["status"] == "stopped":
|
||||
|
||||
if job["exitstatus"] == "OK":
|
||||
logger.debug(f"Balancing: Job ID {job_id} (guest: {guest_name}) was sucessfully.")
|
||||
logger.debug("Finished: get_rebalancing_job_status.")
|
||||
return True
|
||||
else:
|
||||
logger.critical(f"Balancing: Job ID {job_id} (guest: {guest_name}) went into an error! Please check manually.")
|
||||
logger.debug("Finished: get_rebalancing_job_status.")
|
||||
return False
|
||||
308
proxlb/models/calculations.py
Normal file
308
proxlb/models/calculations.py
Normal file
@@ -0,0 +1,308 @@
|
||||
"""
|
||||
The calculation class is responsible for handling the balancing of virtual machines (VMs)
|
||||
and containers (CTs) across all available nodes in a Proxmox cluster. It provides methods
|
||||
to calculate the optimal distribution of VMs and CTs based on the provided data.
|
||||
"""
|
||||
|
||||
import sys
|
||||
from typing import Dict, Any
|
||||
from utils.logger import SystemdLogger
|
||||
|
||||
logger = SystemdLogger()
|
||||
|
||||
|
||||
class Calculations:
|
||||
"""
|
||||
The calculation class is responsible for handling the balancing of virtual machines (VMs)
|
||||
and containers (CTs) across all available nodes in a Proxmox cluster. It provides methods
|
||||
to calculate the optimal distribution of VMs and CTs based on the provided data.
|
||||
"""
|
||||
|
||||
def __init__(self, proxlb_data: Dict[str, Any]):
|
||||
"""
|
||||
Initializes the Calculation class with the provided ProxLB data.
|
||||
|
||||
Args:
|
||||
proxlb_data (Dict[str, Any]): The data required for balancing VMs and CTs.
|
||||
"""
|
||||
|
||||
@staticmethod
|
||||
def set_node_assignments(proxlb_data: Dict[str, Any]) -> Dict[str, Any]:
|
||||
"""
|
||||
Set the assigned ressources of the nodes based on the current assigned
|
||||
guest resources by their created groups as an initial base.
|
||||
|
||||
Args:
|
||||
proxlb_data (Dict[str, Any]): The data holding all current statistics.
|
||||
|
||||
Returns:
|
||||
Dict[str, Any]: Updated ProxLB data of nodes section with updated node assigned values.
|
||||
"""
|
||||
logger.debug("Starting: set_node_assignments.")
|
||||
for group_name, group_meta in proxlb_data["groups"]["affinity"].items():
|
||||
|
||||
for guest_name in group_meta["guests"]:
|
||||
guest_node_current = proxlb_data["guests"][guest_name]["node_current"]
|
||||
# Update Hardware assignments
|
||||
# Update assigned values for the current node
|
||||
proxlb_data["nodes"][guest_node_current]["cpu_assigned"] += proxlb_data["guests"][guest_name]["cpu_total"]
|
||||
proxlb_data["nodes"][guest_node_current]["memory_assigned"] += proxlb_data["guests"][guest_name]["memory_total"]
|
||||
proxlb_data["nodes"][guest_node_current]["disk_assigned"] += proxlb_data["guests"][guest_name]["disk_total"]
|
||||
# Update assigned percentage values for the current node
|
||||
proxlb_data["nodes"][guest_node_current]["cpu_assigned_percent"] = proxlb_data["nodes"][guest_node_current]["cpu_assigned"] / proxlb_data["nodes"][guest_node_current]["cpu_total"] * 100
|
||||
proxlb_data["nodes"][guest_node_current]["memory_assigned_percent"] = proxlb_data["nodes"][guest_node_current]["memory_assigned"] / proxlb_data["nodes"][guest_node_current]["memory_total"] * 100
|
||||
proxlb_data["nodes"][guest_node_current]["disk_assigned_percent"] = proxlb_data["nodes"][guest_node_current]["disk_assigned"] / proxlb_data["nodes"][guest_node_current]["disk_total"] * 100
|
||||
|
||||
logger.debug("Finished: set_node_assignments.")
|
||||
|
||||
@staticmethod
|
||||
def get_balanciness(proxlb_data: Dict[str, Any]) -> Dict[str, Any]:
|
||||
"""
|
||||
Get the blanaciness for further actions where the highest and lowest
|
||||
usage or assignments of Proxmox nodes are compared. Based on the users
|
||||
provided balanciness delta the balancing will be performed.
|
||||
|
||||
Args:
|
||||
proxlb_data (Dict[str, Any]): The data holding all content of all objects.
|
||||
Returns:
|
||||
Dict[str, Any]: Updated meta data section of the balanciness action defined
|
||||
as a bool.
|
||||
"""
|
||||
logger.debug("Starting: get_balanciness.")
|
||||
proxlb_data["meta"]["balancing"]["balance"] = False
|
||||
|
||||
if len(proxlb_data["groups"]) > 0:
|
||||
method = proxlb_data["meta"]["balancing"].get("method", "memory")
|
||||
mode = proxlb_data["meta"]["balancing"].get("mode", "used")
|
||||
balanciness = proxlb_data["meta"]["balancing"].get("balanciness", 10)
|
||||
method_value = [node_meta[f"{method}_{mode}_percent"] for node_meta in proxlb_data["nodes"].values()]
|
||||
method_value_highest = max(method_value)
|
||||
method_value_lowest = min(method_value)
|
||||
|
||||
if method_value_highest - method_value_lowest > balanciness:
|
||||
proxlb_data["meta"]["balancing"]["balance"] = True
|
||||
logger.debug(f"Guest balancing is required. Highest value: {method_value_highest}, lowest value: {method_value_lowest} balanced by {method} and {mode}.")
|
||||
logger.critical(f"Guest balancing is required. Highest value: {method_value_highest}, lowest value: {method_value_lowest} balanced by {method} and {mode}.")
|
||||
else:
|
||||
logger.debug(f"Guest balancing is ok. Highest value: {method_value_highest}, lowest value: {method_value_lowest} balanced by {method} and {mode}.")
|
||||
logger.critical(f"Guest balancing is ok. Highest value: {method_value_highest}, lowest value: {method_value_lowest} balanced by {method} and {mode}.")
|
||||
|
||||
else:
|
||||
logger.warning("No guests for balancing found.")
|
||||
|
||||
logger.debug("Finished: get_balanciness.")
|
||||
|
||||
@staticmethod
|
||||
def get_most_free_node(proxlb_data: Dict[str, Any], return_node: bool = False) -> Dict[str, Any]:
|
||||
"""
|
||||
Get the name of the Proxmox node in the cluster with the most free resources based on
|
||||
the user defined method (e.g.: memory) and mode (e.g.: used).
|
||||
|
||||
Args:
|
||||
proxlb_data (Dict[str, Any]): The data holding all content of all objects.
|
||||
return_node (bool): The indicator to simply return the best node for further
|
||||
assignments.
|
||||
|
||||
Returns:
|
||||
Dict[str, Any]: Updated meta data section of the node with the most free resources that should
|
||||
be used for the next balancing action.
|
||||
"""
|
||||
logger.debug("Starting: get_most_free_node.")
|
||||
proxlb_data["meta"]["balancing"]["balance_next_node"] = ""
|
||||
|
||||
# Do not include nodes that are marked in 'maintenance'
|
||||
filtered_nodes = [node for node in proxlb_data["nodes"].values() if not node["maintenance"]]
|
||||
lowest_usage_node = min(filtered_nodes, key=lambda x: x["memory_used_percent"])
|
||||
proxlb_data["meta"]["balancing"]["balance_reason"] = 'resources'
|
||||
proxlb_data["meta"]["balancing"]["balance_next_node"] = lowest_usage_node["name"]
|
||||
|
||||
# If executed to simply get the best node for further usage, we return
|
||||
# the best node on stdout and gracefully exit here
|
||||
if return_node:
|
||||
print(lowest_usage_node["name"])
|
||||
sys.exit(0)
|
||||
|
||||
logger.debug("Finished: get_most_free_node.")
|
||||
|
||||
@staticmethod
|
||||
def relocate_guests_on_maintenance_nodes(proxlb_data: Dict[str, Any]):
|
||||
"""
|
||||
Relocates guests that are currently on nodes marked for maintenance to
|
||||
nodes with the most available resources.
|
||||
|
||||
This function iterates over all guests on maintenance nodes and attempts
|
||||
to relocate them to nodes with the most free resources that are not in
|
||||
maintenance mode. It updates the node resources accordingly and logs
|
||||
warnings if the balancing may not be perfect due to the maintenance
|
||||
status of the original node.
|
||||
|
||||
Args:
|
||||
proxlb_data (Dict[str, Any]): The data holding all content of all objects.
|
||||
Returns:
|
||||
None
|
||||
"""
|
||||
logger.debug("Starting: get_most_free_node.")
|
||||
proxlb_data["meta"]["balancing"]["balance_next_guest"] = ""
|
||||
|
||||
for guest_name in proxlb_data["groups"]["maintenance"]:
|
||||
# Update the node with the most free nodes which is
|
||||
# not in a maintenance
|
||||
proxlb_data["meta"]["balancing"]["balance_next_guest"] = guest_name
|
||||
Calculations.get_most_free_node(proxlb_data)
|
||||
Calculations.update_node_resources(proxlb_data)
|
||||
logger.warning(f"Warning: Balancing may not be perfect because guest {guest_name} was located on a node which is in maintenance mode.")
|
||||
|
||||
logger.debug("Finished: get_most_free_node.")
|
||||
|
||||
@staticmethod
|
||||
def relocate_guests(proxlb_data: Dict[str, Any]):
|
||||
"""
|
||||
Relocates guests within the provided data structure to ensure affinity groups are
|
||||
placed on nodes with the most free resources.
|
||||
|
||||
This function iterates over each affinity group in the provided data, identifies
|
||||
the node with the most free resources, and migrates all guests within the group
|
||||
to that node. It updates the node resources accordingly.
|
||||
|
||||
Args:
|
||||
proxlb_data (Dict[str, Any]): The data holding all content of all objects.
|
||||
Returns:
|
||||
None
|
||||
"""
|
||||
logger.debug("Starting: relocate_guests.")
|
||||
if proxlb_data["meta"]["balancing"]["balance"] or proxlb_data["meta"]["balancing"]["force"]:
|
||||
|
||||
if proxlb_data["meta"]["balancing"].get("balance", False):
|
||||
logger.debug("Balancing of guests will be performt. Reason: balanciness")
|
||||
|
||||
if proxlb_data["meta"]["balancing"].get("force", False):
|
||||
logger.debug("Balancing of guests will be performt. Reason: force balancing")
|
||||
|
||||
for group_name in proxlb_data["groups"]["affinity"]:
|
||||
|
||||
# We get initially the node with the most free resources and then
|
||||
# migrate all guests within the group to that node to ensure the
|
||||
# affinity.
|
||||
Calculations.get_most_free_node(proxlb_data)
|
||||
|
||||
for guest_name in proxlb_data["groups"]["affinity"][group_name]["guests"]:
|
||||
proxlb_data["meta"]["balancing"]["balance_next_guest"] = guest_name
|
||||
Calculations.val_anti_affinity(proxlb_data, guest_name)
|
||||
Calculations.update_node_resources(proxlb_data)
|
||||
|
||||
logger.debug("Finished: relocate_guests.")
|
||||
|
||||
@staticmethod
|
||||
def val_anti_affinity(proxlb_data: Dict[str, Any], guest_name: str):
|
||||
"""
|
||||
Validates and assigns nodes to guests based on anti-affinity rules.
|
||||
|
||||
This function iterates over all defined anti-affinity groups in the provided
|
||||
`proxlb_data` and checks if the specified `guest_name` is included in any of
|
||||
these groups. If the guest is included and has not been processed yet, it
|
||||
attempts to assign an unused and non-maintenance node to the guest, ensuring
|
||||
that the anti-affinity rules are respected.
|
||||
|
||||
Parameters:
|
||||
proxlb_data (Dict[str, Any]): The data holding all content of all objects.
|
||||
guest_name (str): The name of the guest to be validated and assigned a node.
|
||||
|
||||
Returns:
|
||||
None
|
||||
"""
|
||||
logger.debug("Starting: val_anti_affinity.")
|
||||
# Start by interating over all defined anti-affinity groups
|
||||
for group_name in proxlb_data["groups"]["anti_affinity"].keys():
|
||||
|
||||
# Validate if the provided guest ist included in the anti-affinity group
|
||||
if guest_name in proxlb_data["groups"]["anti_affinity"][group_name]['guests'] and not proxlb_data["guests"][guest_name]["processed"]:
|
||||
logger.debug(f"Anti-Affinity: Guest: {guest_name} is included in anti-affinity group: {group_name}.")
|
||||
|
||||
# Iterate over all available nodes
|
||||
for node_name in proxlb_data["nodes"].keys():
|
||||
|
||||
# Only select node if it was not used before and is not in a
|
||||
# maintenance mode. Afterwards, add it to the list of already
|
||||
# used nodes for the current anti-affinity group
|
||||
if node_name not in proxlb_data["groups"]["anti_affinity"][group_name]["used_nodes"]:
|
||||
|
||||
if not proxlb_data["nodes"][node_name]["maintenance"]:
|
||||
# If the node has not been used yet, we assign this node to the guest
|
||||
proxlb_data["meta"]["balancing"]["balance_next_node"] = node_name
|
||||
proxlb_data["groups"]["anti_affinity"][group_name]["used_nodes"].append(node_name)
|
||||
logger.debug(f"Node: {node_name} marked as used for anti-affinity group: {group_name} with guest {guest_name}")
|
||||
break
|
||||
|
||||
else:
|
||||
logger.critical(f"Node: {node_name} already got used for anti-affinity group:: {group_name}. (Tried for guest: {guest_name})")
|
||||
|
||||
else:
|
||||
logger.debug(f"Guest: {guest_name} is not included in anti-affinity group: {group_name}. Skipping.")
|
||||
|
||||
logger.debug("Finished: val_anti_affinity.")
|
||||
|
||||
@staticmethod
|
||||
def update_node_resources(proxlb_data):
|
||||
"""
|
||||
Updates the resource allocation and usage statistics for nodes when a guest
|
||||
is moved from one node to another.
|
||||
|
||||
Parameters:
|
||||
proxlb_data (dict): A dictionary containing information about the nodes and
|
||||
guests, including their resource allocations and usage.
|
||||
|
||||
The function performs the following steps:
|
||||
1. Retrieves the guest name, current node, and target node from the provided data.
|
||||
2. Updates the resource allocations and usage statistics for the target node by
|
||||
adding the resources of the moved guest.
|
||||
3. Updates the resource allocations and usage statistics for the current node by
|
||||
subtracting the resources of the moved guest.
|
||||
4. Logs the start and end of the resource update process, as well as the movement
|
||||
of the guest from the current node to the target node.
|
||||
"""
|
||||
logger.debug("Starting: update_node_resources.")
|
||||
guest_name = proxlb_data["meta"]["balancing"]["balance_next_guest"]
|
||||
node_current = proxlb_data["guests"][guest_name]["node_current"]
|
||||
node_target = proxlb_data["meta"]["balancing"]["balance_next_node"]
|
||||
|
||||
# Update resources for the target node by the moved guest resources
|
||||
# Add assigned resources to the target node
|
||||
proxlb_data["nodes"][node_target]["cpu_assigned"] += proxlb_data["guests"][guest_name]["cpu_total"]
|
||||
proxlb_data["nodes"][node_target]["memory_assigned"] += proxlb_data["guests"][guest_name]["memory_total"]
|
||||
proxlb_data["nodes"][node_target]["disk_assigned"] += proxlb_data["guests"][guest_name]["disk_total"]
|
||||
# Update the assigned percentages of assigned resources for the target node
|
||||
proxlb_data["nodes"][node_target]["cpu_assigned_percent"] = proxlb_data["nodes"][node_target]["cpu_assigned"] / proxlb_data["nodes"][node_target]["cpu_total"] * 100
|
||||
proxlb_data["nodes"][node_target]["memory_assigned_percent"] = proxlb_data["nodes"][node_target]["memory_assigned"] / proxlb_data["nodes"][node_target]["memory_total"] * 100
|
||||
proxlb_data["nodes"][node_target]["disk_assigned_percent"] = proxlb_data["nodes"][node_target]["disk_assigned"] / proxlb_data["nodes"][node_target]["disk_total"] * 100
|
||||
# Add used resources to the target node
|
||||
proxlb_data["nodes"][node_target]["cpu_used"] += proxlb_data["guests"][guest_name]["cpu_used"]
|
||||
proxlb_data["nodes"][node_target]["memory_used"] += proxlb_data["guests"][guest_name]["memory_used"]
|
||||
proxlb_data["nodes"][node_target]["disk_used"] += proxlb_data["guests"][guest_name]["disk_used"]
|
||||
# Update the used percentages of usage resources for the target node
|
||||
proxlb_data["nodes"][node_target]["cpu_used_percent"] = proxlb_data["nodes"][node_target]["cpu_used"] / proxlb_data["nodes"][node_target]["cpu_total"] * 100
|
||||
proxlb_data["nodes"][node_target]["memory_used_percent"] = proxlb_data["nodes"][node_target]["memory_used"] / proxlb_data["nodes"][node_target]["memory_total"] * 100
|
||||
proxlb_data["nodes"][node_target]["disk_used_percent"] = proxlb_data["nodes"][node_target]["disk_used"] / proxlb_data["nodes"][node_target]["disk_total"] * 100
|
||||
|
||||
# Update resources for the current node by the moved guest resources
|
||||
# Add assigned resources to the target node
|
||||
proxlb_data["nodes"][node_current]["cpu_assigned"] -= proxlb_data["guests"][guest_name]["cpu_total"]
|
||||
proxlb_data["nodes"][node_current]["memory_assigned"] -= proxlb_data["guests"][guest_name]["memory_total"]
|
||||
proxlb_data["nodes"][node_current]["disk_assigned"] -= proxlb_data["guests"][guest_name]["disk_total"]
|
||||
# Update the assigned percentages of assigned resources for the target node
|
||||
proxlb_data["nodes"][node_current]["cpu_assigned_percent"] = proxlb_data["nodes"][node_current]["cpu_assigned"] / proxlb_data["nodes"][node_current]["cpu_total"] * 100
|
||||
proxlb_data["nodes"][node_current]["memory_assigned_percent"] = proxlb_data["nodes"][node_current]["memory_assigned"] / proxlb_data["nodes"][node_current]["memory_total"] * 100
|
||||
proxlb_data["nodes"][node_current]["disk_assigned_percent"] = proxlb_data["nodes"][node_current]["disk_assigned"] / proxlb_data["nodes"][node_current]["disk_total"] * 100
|
||||
# Add used resources to the target node
|
||||
proxlb_data["nodes"][node_current]["cpu_used"] -= proxlb_data["guests"][guest_name]["cpu_used"]
|
||||
proxlb_data["nodes"][node_current]["memory_used"] -= proxlb_data["guests"][guest_name]["memory_used"]
|
||||
proxlb_data["nodes"][node_current]["disk_used"] -= proxlb_data["guests"][guest_name]["disk_used"]
|
||||
# Update the used percentages of usage resources for the target node
|
||||
proxlb_data["nodes"][node_current]["cpu_used_percent"] = proxlb_data["nodes"][node_current]["cpu_used"] / proxlb_data["nodes"][node_current]["cpu_total"] * 100
|
||||
proxlb_data["nodes"][node_current]["memory_used_percent"] = proxlb_data["nodes"][node_current]["memory_used"] / proxlb_data["nodes"][node_current]["memory_total"] * 100
|
||||
proxlb_data["nodes"][node_current]["disk_used_percent"] = proxlb_data["nodes"][node_current]["disk_used"] / proxlb_data["nodes"][node_current]["disk_total"] * 100
|
||||
|
||||
# Assign guest to the new target node
|
||||
proxlb_data["guests"][guest_name]["node_target"] = node_target
|
||||
logger.debug(f"Set guest {guest_name} from node {node_current} to node {node_target}.")
|
||||
|
||||
logger.debug("Finished: update_node_resources.")
|
||||
110
proxlb/models/groups.py
Normal file
110
proxlb/models/groups.py
Normal file
@@ -0,0 +1,110 @@
|
||||
"""
|
||||
The groups class is responsible for handling the correlations between the guests
|
||||
and their groups like affinity and anti-affinity groups. To ensure a proper balancing
|
||||
guests will ge grouped and then evaluated for further balancing.
|
||||
"""
|
||||
|
||||
from typing import Dict, Any
|
||||
from utils.logger import SystemdLogger
|
||||
from utils.helper import Helper
|
||||
|
||||
logger = SystemdLogger()
|
||||
|
||||
|
||||
class Groups:
|
||||
"""
|
||||
The groups class is responsible for handling the correlations between the guests
|
||||
and their groups like affinity and anti-affinity groups. To ensure a proper balancing
|
||||
guests will ge grouped and then evaluated for further balancing.
|
||||
"""
|
||||
|
||||
def __init__(self, proxlb_data: Dict[str, Any]):
|
||||
"""
|
||||
Initializes the Groups class with the provided ProxLB data.
|
||||
|
||||
Args:
|
||||
proxlb_data (Dict[str, Any]): The data required for balancing VMs and CTs.
|
||||
"""
|
||||
|
||||
@staticmethod
|
||||
def get_groups(guests: Dict[str, Any], nodes: Dict[str, Any]) -> Dict[str, Any]:
|
||||
"""
|
||||
Generates and returns a dictionary of affinity and anti-affinity groups based on the provided data.
|
||||
|
||||
Args:
|
||||
guests (Dict[str, Any]): A dictionary containing the guest data.
|
||||
nodes (Dict[str, Any]): A dictionary containing the nodes data.
|
||||
|
||||
Returns:
|
||||
Dict[str, Any]: A dictionary containing the created groups that includes:
|
||||
* Affinity groups (or a randon and uniq group)
|
||||
* Anti-affinity groups
|
||||
* A list of guests that are currently placed on a node which
|
||||
is defined to be in maintenance.
|
||||
"""
|
||||
logger.debug("Starting: get_groups.")
|
||||
groups = {'groups': {'affinity': {}, 'anti_affinity': {}, 'maintenance': []}}
|
||||
|
||||
for guest_name, guest_meta in guests["guests"].items():
|
||||
# Create affinity grouping
|
||||
# Use an affinity group if available for the guest
|
||||
if len(guest_meta["affinity_groups"]) > 0:
|
||||
for affinity_group in guest_meta["affinity_groups"]:
|
||||
group_name = affinity_group
|
||||
logger.debug(f'Affinity group {affinity_group} for {guest_name} will be used.')
|
||||
else:
|
||||
# Generate a random uniq group name for the guest if
|
||||
# the guest does not belong to any affinity group
|
||||
random_group = Helper.get_uuid_string()
|
||||
group_name = random_group
|
||||
logger.debug(f'Random uniq group {random_group} for {guest_name} will be used.')
|
||||
|
||||
if not groups["groups"]["affinity"].get(group_name, False):
|
||||
# Create group template with initial guest meta information
|
||||
groups["groups"]["affinity"][group_name] = {}
|
||||
groups["groups"]["affinity"][group_name]["guests"] = []
|
||||
groups["groups"]["affinity"][group_name]["guests"].append(guest_name)
|
||||
groups["groups"]["affinity"][group_name]["counter"] = 1
|
||||
# Create groups resource template by the guests resources
|
||||
groups["groups"]["affinity"][group_name]["cpu_total"] = guest_meta["cpu_total"]
|
||||
groups["groups"]["affinity"][group_name]["cpu_used"] = guest_meta["cpu_used"]
|
||||
groups["groups"]["affinity"][group_name]["memory_total"] = guest_meta["memory_total"]
|
||||
groups["groups"]["affinity"][group_name]["memory_used"] = guest_meta["cpu_used"]
|
||||
groups["groups"]["affinity"][group_name]["disk_total"] = guest_meta["disk_total"]
|
||||
groups["groups"]["affinity"][group_name]["disk_used"] = guest_meta["cpu_used"]
|
||||
else:
|
||||
# Update group templates by guest meta information
|
||||
groups["groups"]["affinity"][group_name]["guests"].append(guest_name)
|
||||
groups["groups"]["affinity"][group_name]["counter"] += 1
|
||||
# Update group resources by guest resources
|
||||
groups["groups"]["affinity"][group_name]["cpu_total"] += guest_meta["cpu_total"]
|
||||
groups["groups"]["affinity"][group_name]["cpu_used"] += guest_meta["cpu_used"]
|
||||
groups["groups"]["affinity"][group_name]["memory_total"] += guest_meta["memory_total"]
|
||||
groups["groups"]["affinity"][group_name]["memory_used"] += guest_meta["cpu_used"]
|
||||
groups["groups"]["affinity"][group_name]["disk_total"] += guest_meta["disk_total"]
|
||||
groups["groups"]["affinity"][group_name]["disk_used"] += guest_meta["cpu_used"]
|
||||
|
||||
# Create anti-affinity grouping
|
||||
if len(guest_meta["anti_affinity_groups"]) > 0:
|
||||
for anti_affinity_group in guest_meta["anti_affinity_groups"]:
|
||||
anti_affinity_group_name = anti_affinity_group
|
||||
logger.debug(f'Anti-affinity group {anti_affinity_group_name} for {guest_name} will be used.')
|
||||
|
||||
if not groups["groups"]["anti_affinity"].get(anti_affinity_group_name, False):
|
||||
groups["groups"]["anti_affinity"][anti_affinity_group_name] = {}
|
||||
groups["groups"]["anti_affinity"][anti_affinity_group_name]["guests"] = []
|
||||
groups["groups"]["anti_affinity"][anti_affinity_group_name]["guests"].append(guest_name)
|
||||
groups["groups"]["anti_affinity"][anti_affinity_group_name]["counter"] = 1
|
||||
groups["groups"]["anti_affinity"][anti_affinity_group_name]["used_nodes"] = []
|
||||
else:
|
||||
groups["groups"]["anti_affinity"][anti_affinity_group_name]["guests"].append(guest_name)
|
||||
groups["groups"]["anti_affinity"][anti_affinity_group_name]["counter"] += 1
|
||||
|
||||
# Create grouping of guests that are currently located on nodes that are
|
||||
# marked as in maintenance and must be migrated
|
||||
if nodes["nodes"][guest_meta["node_current"]]["maintenance"]:
|
||||
logger.debug(f'{guest_name} will be migrated to another node because the underlying node {guest_meta["node_current"]} is defined to be in maintenance.')
|
||||
groups["groups"]["maintenance"].append(guest_name)
|
||||
|
||||
logger.debug("Finished: get_groups.")
|
||||
return groups
|
||||
97
proxlb/models/guests.py
Normal file
97
proxlb/models/guests.py
Normal file
@@ -0,0 +1,97 @@
|
||||
"""
|
||||
The Guests class retrieves all running guests on the Proxmox cluster across all available nodes.
|
||||
It handles both VM and CT guest types, collecting their resource metrics.
|
||||
"""
|
||||
|
||||
from typing import Dict, Any
|
||||
from utils.logger import SystemdLogger
|
||||
from models.tags import Tags
|
||||
|
||||
logger = SystemdLogger()
|
||||
|
||||
|
||||
class Guests:
|
||||
"""
|
||||
The Guests class retrieves all running guests on the Proxmox cluster across all available nodes.
|
||||
It handles both VM and CT guest types, collecting their resource metrics.
|
||||
"""
|
||||
def __init__(self):
|
||||
"""
|
||||
Initializes the Guests class with the provided ProxLB data.
|
||||
"""
|
||||
|
||||
@staticmethod
|
||||
def get_guests(proxmox_api: any, nodes: Dict[str, Any]) -> Dict[str, Any]:
|
||||
"""
|
||||
Get metrics of all guests in a Proxmox cluster.
|
||||
|
||||
This method retrieves metrics for all running guests (both VMs and CTs) across all nodes in the Proxmox cluster.
|
||||
It iterates over each node and collects resource metrics for each running guest, including CPU, memory, and disk usage.
|
||||
Additionally, it retrieves tags and affinity/anti-affinity groups for each guest.
|
||||
|
||||
Args:
|
||||
proxmox_api (any): The Proxmox API client instance.
|
||||
nodes (Dict[str, Any]): A dictionary containing information about the nodes in the Proxmox cluster.
|
||||
|
||||
Returns:
|
||||
Dict[str, Any]: A dictionary containing metrics and information for all running guests.
|
||||
"""
|
||||
logger.debug("Starting: get_guests.")
|
||||
guests = {"guests": {}}
|
||||
|
||||
# Guest objects are always only in the scope of a node.
|
||||
# Therefore, we need to iterate over all nodes to get all guests.
|
||||
for node in nodes['nodes'].keys():
|
||||
|
||||
# VM objects: Iterate over all VMs on the current node by the qemu API object.
|
||||
# Unlike the nodes we need to keep them even when being ignored to create proper
|
||||
# resource metrics for rebalancing to ensure that we do not overprovisiong the node.
|
||||
for guest in proxmox_api.nodes(node).qemu.get():
|
||||
if guest['status'] == 'running':
|
||||
guests['guests'][guest['name']] = {}
|
||||
guests['guests'][guest['name']]['name'] = guest['name']
|
||||
guests['guests'][guest['name']]['cpu_total'] = guest['cpus']
|
||||
guests['guests'][guest['name']]['cpu_used'] = guest['cpu']
|
||||
guests['guests'][guest['name']]['memory_total'] = guest['maxmem']
|
||||
guests['guests'][guest['name']]['memory_used'] = guest['mem']
|
||||
guests['guests'][guest['name']]['disk_total'] = guest['maxdisk']
|
||||
guests['guests'][guest['name']]['disk_used'] = guest['disk']
|
||||
guests['guests'][guest['name']]['id'] = guest['vmid']
|
||||
guests['guests'][guest['name']]['node_current'] = node
|
||||
guests['guests'][guest['name']]['node_target'] = node
|
||||
guests['guests'][guest['name']]['processed'] = False
|
||||
guests['guests'][guest['name']]['tags'] = Tags.get_tags_from_guests(proxmox_api, node, guest['vmid'], 'vm')
|
||||
guests['guests'][guest['name']]['affinity_groups'] = Tags.get_affinity_groups(guests['guests'][guest['name']]['tags'])
|
||||
guests['guests'][guest['name']]['anti_affinity_groups'] = Tags.get_anti_affinity_groups(guests['guests'][guest['name']]['tags'])
|
||||
guests['guests'][guest['name']]['ignore'] = Tags.get_ignore(guests['guests'][guest['name']]['tags'])
|
||||
guests['guests'][guest['name']]['type'] = 'vm'
|
||||
else:
|
||||
logger.debug(f'Metric for VM {guest["name"]} ignored because VM is not running.')
|
||||
|
||||
# CT objects: Iterate over all VMs on the current node by the lxc API object.
|
||||
# Unlike the nodes we need to keep them even when being ignored to create proper
|
||||
# resource metrics for rebalancing to ensure that we do not overprovisiong the node.
|
||||
for guest in proxmox_api.nodes(node).lxc.get():
|
||||
if guest['status'] == 'running':
|
||||
guests['guests'][guest['name']] = {}
|
||||
guests['guests'][guest['name']]['name'] = guest['name']
|
||||
guests['guests'][guest['name']]['cpu_total'] = guest['cpus']
|
||||
guests['guests'][guest['name']]['cpu_used'] = guest['cpu']
|
||||
guests['guests'][guest['name']]['memory_total'] = guest['maxmem']
|
||||
guests['guests'][guest['name']]['memory_used'] = guest['mem']
|
||||
guests['guests'][guest['name']]['disk_total'] = guest['maxdisk']
|
||||
guests['guests'][guest['name']]['disk_used'] = guest['disk']
|
||||
guests['guests'][guest['name']]['id'] = guest['vmid']
|
||||
guests['guests'][guest['name']]['node_current'] = node
|
||||
guests['guests'][guest['name']]['node_target'] = node
|
||||
guests['guests'][guest['name']]['processed'] = False
|
||||
guests['guests'][guest['name']]['tags'] = Tags.get_tags_from_guests(proxmox_api, node, guest['vmid'], 'ct')
|
||||
guests['guests'][guest['name']]['affinity_groups'] = Tags.get_affinity_groups(guests['guests'][guest['name']]['tags'])
|
||||
guests['guests'][guest['name']]['anti_affinity_groups'] = Tags.get_anti_affinity_groups(guests['guests'][guest['name']]['tags'])
|
||||
guests['guests'][guest['name']]['ignore'] = Tags.get_ignore(guests['guests'][guest['name']]['tags'])
|
||||
guests['guests'][guest['name']]['type'] = 'ct'
|
||||
else:
|
||||
logger.debug(f'Metric for CT {guest["name"]} ignored because CT is not running.')
|
||||
|
||||
logger.debug("Finished: get_guests.")
|
||||
return guests
|
||||
123
proxlb/models/nodes.py
Normal file
123
proxlb/models/nodes.py
Normal file
@@ -0,0 +1,123 @@
|
||||
"""
|
||||
The Nodes class retrieves all running nodes in a Proxmox cluster
|
||||
and collects their resource metrics.
|
||||
"""
|
||||
|
||||
from typing import Dict, Any
|
||||
from utils.logger import SystemdLogger
|
||||
|
||||
logger = SystemdLogger()
|
||||
|
||||
|
||||
class Nodes:
|
||||
"""
|
||||
The Nodes class retrieves all running nodes in a Proxmox cluster
|
||||
and collects their resource metrics.
|
||||
"""
|
||||
def __init__(self):
|
||||
"""
|
||||
Initializes the Nodes class with the provided ProxLB data.
|
||||
"""
|
||||
|
||||
@staticmethod
|
||||
def get_nodes(proxmox_api: any, proxlb_config: Dict[str, Any]) -> Dict[str, Any]:
|
||||
"""
|
||||
Get metrics of all nodes in a Proxmox cluster.
|
||||
|
||||
This method retrieves metrics for all available nodes in the Proxmox cluster.
|
||||
It iterates over each node and collects resource metrics including CPU, memory, and disk usage.
|
||||
|
||||
Args:
|
||||
proxmox_api (any): The Proxmox API client instance.
|
||||
nodes (Dict[str, Any]): A dictionary containing information about the nodes in the Proxmox cluster.
|
||||
|
||||
Returns:
|
||||
Dict[str, Any]: A dictionary containing metrics and information for all running nodes.
|
||||
"""
|
||||
logger.debug("Starting: get_nodes.")
|
||||
nodes = {"nodes": {}}
|
||||
|
||||
for node in proxmox_api.nodes.get():
|
||||
# Ignoring a node results into ignoring all placed guests on the ignored node!
|
||||
if node["status"] == "online" and not Nodes.set_node_ignore(proxlb_config, node["node"]):
|
||||
nodes["nodes"][node["node"]] = {}
|
||||
nodes["nodes"][node["node"]]["name"] = node["node"]
|
||||
nodes["nodes"][node["node"]]["maintenance"] = False
|
||||
nodes["nodes"][node["node"]]["cpu_total"] = node["maxcpu"]
|
||||
nodes["nodes"][node["node"]]["cpu_assigned"] = 0
|
||||
nodes["nodes"][node["node"]]["cpu_used"] = node["cpu"]
|
||||
nodes["nodes"][node["node"]]["cpu_free"] = (node["maxcpu"]) - (node["cpu"] * node["maxcpu"])
|
||||
nodes["nodes"][node["node"]]["cpu_assigned_percent"] = nodes["nodes"][node["node"]]["cpu_assigned"] / nodes["nodes"][node["node"]]["cpu_total"] * 100
|
||||
nodes["nodes"][node["node"]]["cpu_free_percent"] = nodes["nodes"][node["node"]]["cpu_free"] / node["maxcpu"] * 100
|
||||
nodes["nodes"][node["node"]]["cpu_used_percent"] = nodes["nodes"][node["node"]]["cpu_used"] / node["maxcpu"] * 100
|
||||
nodes["nodes"][node["node"]]["memory_total"] = node["maxmem"]
|
||||
nodes["nodes"][node["node"]]["memory_assigned"] = 0
|
||||
nodes["nodes"][node["node"]]["memory_used"] = node["mem"]
|
||||
nodes["nodes"][node["node"]]["memory_free"] = node["maxmem"] - node["mem"]
|
||||
nodes["nodes"][node["node"]]["memory_assigned_percent"] = nodes["nodes"][node["node"]]["memory_assigned"] / nodes["nodes"][node["node"]]["memory_total"] * 100
|
||||
nodes["nodes"][node["node"]]["memory_free_percent"] = nodes["nodes"][node["node"]]["memory_free"] / node["maxmem"] * 100
|
||||
nodes["nodes"][node["node"]]["memory_used_percent"] = nodes["nodes"][node["node"]]["memory_used"] / node["maxmem"] * 100
|
||||
nodes["nodes"][node["node"]]["disk_total"] = node["maxdisk"]
|
||||
nodes["nodes"][node["node"]]["disk_assigned"] = 0
|
||||
nodes["nodes"][node["node"]]["disk_used"] = node["disk"]
|
||||
nodes["nodes"][node["node"]]["disk_free"] = node["maxdisk"] - node["disk"]
|
||||
nodes["nodes"][node["node"]]["disk_assigned_percent"] = nodes["nodes"][node["node"]]["disk_assigned"] / nodes["nodes"][node["node"]]["disk_total"] * 100
|
||||
nodes["nodes"][node["node"]]["disk_free_percent"] = nodes["nodes"][node["node"]]["disk_free"] / node["maxdisk"] * 100
|
||||
nodes["nodes"][node["node"]]["disk_used_percent"] = nodes["nodes"][node["node"]]["disk_used"] / node["maxdisk"] * 100
|
||||
|
||||
# Evaluate if node should be set to maintenance mode
|
||||
if Nodes.set_node_maintenance(proxlb_config, node["node"]):
|
||||
nodes["nodes"][node["node"]]["maintenance"] = True
|
||||
|
||||
logger.debug("Finished: get_nodes.")
|
||||
return nodes
|
||||
|
||||
@staticmethod
|
||||
def set_node_maintenance(proxlb_config: Dict[str, Any], node_name: str) -> Dict[str, Any]:
|
||||
"""
|
||||
Set nodes to maintenance mode based on the provided configuration.
|
||||
|
||||
This method updates the nodes dictionary to mark certain nodes as being in maintenance mode
|
||||
based on the configuration provided in proxlb_config.
|
||||
|
||||
Args:
|
||||
proxlb_config (Dict[str, Any]): A dictionary containing the ProxLB configuration, including maintenance nodes.
|
||||
node_name: (str): The current node name within the outer iteration.
|
||||
|
||||
Returns:
|
||||
Bool: Returns a bool if the provided node name is present in the maintenance section of the config file.
|
||||
"""
|
||||
logger.debug("Starting: set_node_maintenance.")
|
||||
|
||||
if proxlb_config.get("proxmox_cluster", None).get("maintenance_nodes", None) is not None:
|
||||
if len(proxlb_config.get("proxmox_cluster", {}).get("maintenance_nodes", [])) > 0:
|
||||
if node_name in proxlb_config.get("proxmox_cluster", {}).get("maintenance_nodes", []):
|
||||
logger.warning(f"Node: {node_name} has been set to maintenance mode.")
|
||||
return True
|
||||
|
||||
logger.debug("Finished: set_node_maintenance.")
|
||||
|
||||
@staticmethod
|
||||
def set_node_ignore(proxlb_config: Dict[str, Any], node_name: str) -> Dict[str, Any]:
|
||||
"""
|
||||
Set nodes to be ignored based on the provided configuration.
|
||||
|
||||
This method updates the nodes dictionary to mark certain nodes as being ignored
|
||||
based on the configuration provided in proxlb_config.
|
||||
|
||||
Args:
|
||||
proxlb_config (Dict[str, Any]): A dictionary containing the ProxLB configuration, including maintenance nodes.
|
||||
node_name: (str): The current node name within the outer iteration.
|
||||
|
||||
Returns:
|
||||
Bool: Returns a bool if the provided node name is present in the ignore section of the config file.
|
||||
"""
|
||||
logger.debug("Starting: set_node_ignore.")
|
||||
|
||||
if proxlb_config.get("proxmox_cluster", None).get("ignore_nodes", None) is not None:
|
||||
if len(proxlb_config.get("proxmox_cluster", {}).get("ignore_nodes", [])) > 0:
|
||||
if node_name in proxlb_config.get("proxmox_cluster", {}).get("ignore_nodes", []):
|
||||
logger.warning(f"Node: {node_name} has been set to be ignored. Not adding node!")
|
||||
return True
|
||||
|
||||
logger.debug("Finished: set_node_ignore.")
|
||||
130
proxlb/models/tags.py
Normal file
130
proxlb/models/tags.py
Normal file
@@ -0,0 +1,130 @@
|
||||
"""
|
||||
The Tags class retrieves all tags from guests of the type VM or CT running
|
||||
in a Proxmox cluster and validates for affinity, anti-affinity and ignore
|
||||
tags set for the guest in the Proxmox API.
|
||||
"""
|
||||
|
||||
import time
|
||||
from typing import List
|
||||
from utils.logger import SystemdLogger
|
||||
|
||||
logger = SystemdLogger()
|
||||
|
||||
|
||||
class Tags:
|
||||
"""
|
||||
The Tags class retrieves all tags from guests of the type VM or CT running
|
||||
in a Proxmox cluster and validates for affinity, anti-affinity and ignore
|
||||
tags set for the guest in the Proxmox API.
|
||||
"""
|
||||
def __init__(self):
|
||||
"""
|
||||
Initializes the Tags class.
|
||||
"""
|
||||
|
||||
@staticmethod
|
||||
def get_tags_from_guests(proxmox_api: any, node: str, guest_id: int, guest_type: str) -> List[str]:
|
||||
"""
|
||||
Get tags for a guest from the Proxmox cluster by the API.
|
||||
|
||||
This method retrieves all tags for a given guest from the Proxmox API which
|
||||
is held in the guest_config.
|
||||
|
||||
Args:
|
||||
proxmox_api (any): The Proxmox API client instance.
|
||||
node (str): The node name where the given guest is located.
|
||||
guest_id (int): The internal Proxmox ID of the guest.
|
||||
guest_type (str): The type (vm or ct) of the guest.
|
||||
|
||||
Returns:
|
||||
List: A list of all tags assoiciated with the given guest.
|
||||
"""
|
||||
logger.debug("Starting: get_tags_from_guests.")
|
||||
time.sleep(0.1)
|
||||
if guest_type == 'vm':
|
||||
guest_config = proxmox_api.nodes(node).qemu(guest_id).config.get()
|
||||
tags = guest_config.get("tags", [])
|
||||
if guest_type == 'ct':
|
||||
guest_config = proxmox_api.nodes(node).lxc(guest_id).config.get()
|
||||
tags = guest_config.get("tags", [])
|
||||
|
||||
if isinstance(tags, str):
|
||||
tags = tags.split(";")
|
||||
|
||||
logger.debug("Finished: get_tags_from_guests.")
|
||||
return tags
|
||||
|
||||
@staticmethod
|
||||
def get_affinity_groups(tags: List[str]) -> List[str]:
|
||||
"""
|
||||
Get affinity tags for a guest from the Proxmox cluster by the API.
|
||||
|
||||
This method retrieves all tags for a given guest and evaluates the
|
||||
affinity tags which are required during the balancing calculations.
|
||||
|
||||
Args:
|
||||
tags (List): A list holding all defined tags for a given guest.
|
||||
|
||||
Returns:
|
||||
List: A list including all affinity tags for the given guest.
|
||||
"""
|
||||
logger.debug("Starting: get_affinity_groups.")
|
||||
affinity_tags = []
|
||||
|
||||
if len(tags) > 0:
|
||||
for tag in tags:
|
||||
if tag.startswith("plb_affinity"):
|
||||
affinity_tags.append(tag)
|
||||
|
||||
logger.debug("Finished: get_affinity_groups.")
|
||||
return affinity_tags
|
||||
|
||||
@staticmethod
|
||||
def get_anti_affinity_groups(tags: List[str]) -> List[str]:
|
||||
"""
|
||||
Get anti-affinity tags for a guest from the Proxmox cluster by the API.
|
||||
|
||||
This method retrieves all tags for a given guest and evaluates the
|
||||
anti-affinity tags which are required during the balancing calculations.
|
||||
|
||||
Args:
|
||||
tags (List): A list holding all defined tags for a given guest.
|
||||
|
||||
Returns:
|
||||
List: A list including all anti-affinity tags for the given guest..
|
||||
"""
|
||||
logger.debug("Starting: get_anti_affinity_groups.")
|
||||
anti_affinity_tags = []
|
||||
|
||||
if len(tags) > 0:
|
||||
for tag in tags:
|
||||
if tag.startswith("plb_anti_affinity"):
|
||||
anti_affinity_tags.append(tag)
|
||||
|
||||
logger.debug("Finished: get_anti_affinity_groups.")
|
||||
return anti_affinity_tags
|
||||
|
||||
@staticmethod
|
||||
def get_ignore(tags: List[str]) -> bool:
|
||||
"""
|
||||
Validate for ignore tags of a guest from the Proxmox cluster by the API.
|
||||
|
||||
This method retrieves all tags for a given guest and evaluates the
|
||||
ignore tag which are required during the balancing calculations.
|
||||
|
||||
Args:
|
||||
tags (List): A list holding all defined tags for a given guest.
|
||||
|
||||
Returns:
|
||||
Bool: Returns a bool that indicates wether to ignore a guest or not.
|
||||
"""
|
||||
logger.debug("Starting: get_ignore.")
|
||||
ignore_tag = False
|
||||
|
||||
if len(tags) > 0:
|
||||
for tag in tags:
|
||||
if tag.startswith("plb_ignore"):
|
||||
ignore_tag = True
|
||||
|
||||
logger.debug("Finished: get_ignore.")
|
||||
return ignore_tag
|
||||
0
proxlb/utils/__init__.py
Normal file
0
proxlb/utils/__init__.py
Normal file
85
proxlb/utils/cli_parser.py
Normal file
85
proxlb/utils/cli_parser.py
Normal file
@@ -0,0 +1,85 @@
|
||||
"""
|
||||
The CliParser class handles the parsing of command-line interface (CLI) arguments.
|
||||
"""
|
||||
|
||||
import argparse
|
||||
import utils.version
|
||||
from utils.logger import SystemdLogger
|
||||
|
||||
logger = SystemdLogger()
|
||||
|
||||
|
||||
class CliParser:
|
||||
"""
|
||||
The CliParser class handles the parsing of command-line interface (CLI) arguments.
|
||||
"""
|
||||
def __init__(self):
|
||||
"""
|
||||
Initializes the CliParser class.
|
||||
|
||||
This method sets up an argument parser for the command-line interface (CLI) with various options:
|
||||
- `-c` or `--config`: Specifies the path to the configuration file.
|
||||
- `-d` or `--dry-run`: Performs a dry-run without executing any actions.
|
||||
- `-j` or `--json`: Returns a JSON of the VM movement.
|
||||
- `-b` or `--best-node`: Returns the best next node.
|
||||
- `-v` or `--version`: Returns the current ProxLB version.
|
||||
|
||||
Logs the start and end of the initialization process.
|
||||
"""
|
||||
logger.debug("Starting: CliParser.")
|
||||
|
||||
self.parser = argparse.ArgumentParser(
|
||||
description=(
|
||||
f"{utils.version.__app_name__} ({utils.version.__version__}): "
|
||||
f"{utils.version.__app_desc__}"
|
||||
)
|
||||
)
|
||||
|
||||
self.parser.add_argument(
|
||||
"-c", "--config",
|
||||
help="Path to the configuration file",
|
||||
type=str,
|
||||
required=False
|
||||
)
|
||||
self.parser.add_argument(
|
||||
"-d", "--dry-run",
|
||||
help="Perform a dry-run without executing any actions",
|
||||
action="store_true",
|
||||
required=False
|
||||
)
|
||||
self.parser.add_argument(
|
||||
"-j", "--json",
|
||||
help="Return a JSON of the VM movement",
|
||||
action="store_true",
|
||||
required=False
|
||||
)
|
||||
self.parser.add_argument(
|
||||
"-b", "--best-node",
|
||||
help="Returns the best next node",
|
||||
action="store_true",
|
||||
required=False
|
||||
)
|
||||
self.parser.add_argument(
|
||||
"-v", "--version",
|
||||
help="Returns the current ProxLB version",
|
||||
action="store_true",
|
||||
required=False
|
||||
)
|
||||
logger.debug("Finished: CliParser.")
|
||||
|
||||
def parse_args(self) -> argparse.Namespace:
|
||||
"""
|
||||
Parses and returns the parsed command-line interface (CLI) arguments.
|
||||
|
||||
This method uses the argparse library to parse the arguments provided
|
||||
via the command line. It logs the start and end of the parsing process,
|
||||
as well as the parsed arguments for debugging purposes.
|
||||
|
||||
Returns:
|
||||
argparse.Namespace: An object containing the parsed CLI arguments.
|
||||
"""
|
||||
logger.debug("Starting: parse_args.")
|
||||
logger.debug(self.parser.parse_args())
|
||||
|
||||
logger.debug("Finished: parse_args.")
|
||||
return self.parser.parse_args()
|
||||
83
proxlb/utils/config_parser.py
Normal file
83
proxlb/utils/config_parser.py
Normal file
@@ -0,0 +1,83 @@
|
||||
"""
|
||||
The ConfigParser class handles the parsing of configuration file
|
||||
from a given YAML file from any location.
|
||||
"""
|
||||
|
||||
import os
|
||||
import sys
|
||||
try:
|
||||
import yaml
|
||||
PYYAML_PRESENT = True
|
||||
except ImportError:
|
||||
PYYAML_PRESENT = False
|
||||
from typing import Dict, Any
|
||||
from utils.logger import SystemdLogger
|
||||
|
||||
|
||||
if not PYYAML_PRESENT:
|
||||
print("Error: The required library 'pyyaml' is not installed.")
|
||||
sys.exit(1)
|
||||
|
||||
|
||||
logger = SystemdLogger()
|
||||
|
||||
|
||||
class ConfigParser:
|
||||
"""
|
||||
The ConfigParser class handles the parsing of configuration file
|
||||
from a given YAML file from any location.
|
||||
"""
|
||||
def __init__(self, config_path: str):
|
||||
"""
|
||||
Initializes the configuration file parser and validates the config file.
|
||||
"""
|
||||
logger.debug("Starting: ConfigParser.")
|
||||
self.config_path = self.test_config_path(config_path)
|
||||
logger.debug("Finished: ConfigParser.")
|
||||
|
||||
def test_config_path(self, config_path: str) -> None:
|
||||
"""
|
||||
Checks if configuration file is present at given config path.
|
||||
"""
|
||||
logger.debug("Starting: test_config_path.")
|
||||
# Test for config file at given location
|
||||
if config_path is not None:
|
||||
|
||||
if os.path.exists(config_path):
|
||||
logger.debug(f"The file {config_path} exists.")
|
||||
else:
|
||||
logger.error(f"The file {config_path} does not exist.")
|
||||
sys.exit(1)
|
||||
|
||||
# Test for config file at default location as a fallback
|
||||
if config_path is None:
|
||||
default_config_path = "/etc/proxlb/proxlb.yaml"
|
||||
|
||||
if os.path.exists(default_config_path):
|
||||
logger.debug(f"The file {default_config_path} exists.")
|
||||
config_path = default_config_path
|
||||
else:
|
||||
print(f"The config file {default_config_path} does not exist.")
|
||||
logger.critical(f"The config file {default_config_path} does not exist.")
|
||||
sys.exit(1)
|
||||
|
||||
logger.debug("Finished: test_config_path.")
|
||||
return config_path
|
||||
|
||||
def get_config(self) -> Dict[str, Any]:
|
||||
"""
|
||||
Parses and returns CLI arguments.
|
||||
"""
|
||||
logger.debug("Starting: get_config.")
|
||||
logger.info(f"Using config path: {self.config_path}")
|
||||
|
||||
try:
|
||||
with open(self.config_path, "r", encoding="utf-8") as config_file:
|
||||
config_data = yaml.load(config_file, Loader=yaml.FullLoader)
|
||||
return config_data
|
||||
except yaml.YAMLError as exception_error:
|
||||
print(f"Error loading YAML file: {exception_error}")
|
||||
logger.critical(f"Error loading YAML file: {exception_error}")
|
||||
sys.exit(1)
|
||||
|
||||
logger.debug("Finished: get_config.")
|
||||
105
proxlb/utils/helper.py
Normal file
105
proxlb/utils/helper.py
Normal file
@@ -0,0 +1,105 @@
|
||||
"""
|
||||
The Helper class provides some basic helper functions to not mess up the code in other
|
||||
classes.
|
||||
"""
|
||||
|
||||
import uuid
|
||||
import sys
|
||||
import time
|
||||
import utils.version
|
||||
from utils.logger import SystemdLogger
|
||||
from typing import Dict, Any
|
||||
|
||||
logger = SystemdLogger()
|
||||
|
||||
|
||||
class Helper:
|
||||
"""
|
||||
The Helper class provides some basic helper functions to not mess up the code in other
|
||||
classes.
|
||||
"""
|
||||
def __init__(self):
|
||||
"""
|
||||
Initializes the general Helper clas.
|
||||
"""
|
||||
|
||||
@staticmethod
|
||||
def get_uuid_string() -> str:
|
||||
"""
|
||||
Generates a random uuid and returns it as a string.
|
||||
|
||||
Args:
|
||||
None
|
||||
|
||||
Returns:
|
||||
Str: Returns a random uuid as a string.
|
||||
"""
|
||||
logger.debug("Starting: get_uuid_string.")
|
||||
generated_uuid = uuid.uuid4()
|
||||
logger.debug("Finished: get_uuid_string.")
|
||||
return str(generated_uuid)
|
||||
|
||||
@staticmethod
|
||||
def log_node_metrics(proxlb_data: Dict[str, Any], init: bool = True) -> None:
|
||||
"""
|
||||
Logs the memory, CPU, and disk usage metrics of nodes in the provided proxlb_data dictionary.
|
||||
|
||||
This method processes the usage metrics of nodes and logs them. It also updates the
|
||||
'statistics' field in the 'meta' section of the proxlb_data dictionary with the
|
||||
memory, CPU, and disk usage metrics before and after a certain operation.
|
||||
|
||||
proxlb_data (Dict[str, Any]): A dictionary containing node metrics and metadata.
|
||||
init (bool): A flag indicating whether to initialize the 'before' statistics
|
||||
(True) or update the 'after' statistics (False). Default is True.
|
||||
"""
|
||||
logger.debug("Starting: log_node_metrics.")
|
||||
nodes_usage_memory = " | ".join([f"{key}: {value['memory_used_percent']:.2f}%" for key, value in proxlb_data["nodes"].items()])
|
||||
nodes_usage_cpu = " | ".join([f"{key}: {value['cpu_used_percent']:.2f}%" for key, value in proxlb_data["nodes"].items()])
|
||||
nodes_usage_disk = " | ".join([f"{key}: {value['disk_used_percent']:.2f}%" for key, value in proxlb_data["nodes"].items()])
|
||||
|
||||
if init:
|
||||
proxlb_data["meta"]["statistics"] = {"before": {"memory": nodes_usage_memory, "cpu": nodes_usage_cpu, "disk": nodes_usage_disk}, "after": {"memory": "", "cpu": "", "disk": ""}}
|
||||
else:
|
||||
proxlb_data["meta"]["statistics"]["after"] = {"memory": nodes_usage_memory, "cpu": nodes_usage_cpu, "disk": nodes_usage_disk}
|
||||
|
||||
logger.debug(f"Nodes usage memory: {nodes_usage_memory}")
|
||||
logger.debug(f"Nodes usage cpu: {nodes_usage_cpu}")
|
||||
logger.debug(f"Nodes usage disk: {nodes_usage_disk}")
|
||||
logger.debug("Finished: log_node_metrics.")
|
||||
|
||||
@staticmethod
|
||||
def get_version(print_version: bool = False) -> None:
|
||||
"""
|
||||
Returns the current version of ProxLB and optionally prints it to stdout.
|
||||
|
||||
Parameters:
|
||||
print_version (bool): If True, prints the version information to stdout and exits the program.
|
||||
|
||||
Returns:
|
||||
None
|
||||
"""
|
||||
if print_version:
|
||||
print(f"{utils.version.__app_name__} version: {utils.version.__version__}\n(C) 2025 by {utils.version.__author__}\n{utils.version.__url__}")
|
||||
sys.exit(0)
|
||||
|
||||
@staticmethod
|
||||
def get_daemon_mode(proxlb_config: Dict[str, Any]) -> None:
|
||||
"""
|
||||
Checks if the daemon mode is active and handles the scheduling accordingly.
|
||||
|
||||
Parameters:
|
||||
proxlb_config (Dict[str, Any]): A dictionary containing the ProxLB configuration.
|
||||
|
||||
Returns:
|
||||
None
|
||||
"""
|
||||
logger.debug("Starting: get_daemon_mode.")
|
||||
if proxlb_config.get("service", {}).get("daemon", False):
|
||||
sleep_seconds = proxlb_config.get("service", {}).get("schedule", 12) * 3600
|
||||
logger.info(f"Daemon mode active: Next run in: {proxlb_config.get('service', {}).get('schedule', 12)} hours.")
|
||||
time.sleep(sleep_seconds)
|
||||
else:
|
||||
logger.debug("Daemon mode is not active.")
|
||||
sys.exit(0)
|
||||
|
||||
logger.debug("Finished: get_daemon_mode.")
|
||||
109
proxlb/utils/logger.py
Normal file
109
proxlb/utils/logger.py
Normal file
@@ -0,0 +1,109 @@
|
||||
"""
|
||||
The SystemdLogger class provides the root logger support. It dynamically
|
||||
evaluates the further usage and imports of journald and adjusts
|
||||
the logger to the systems functionality where it gets executed
|
||||
"""
|
||||
|
||||
import logging
|
||||
try:
|
||||
from systemd.journal import JournalHandler
|
||||
SYSTEMD_PRESENT = True
|
||||
except ImportError:
|
||||
SYSTEMD_PRESENT = False
|
||||
|
||||
|
||||
class SystemdLogger:
|
||||
"""
|
||||
The SystemdLogger class provides the root logger support. It dynamically
|
||||
evaluates the further usage and imports of journald and adjusts
|
||||
the logger to the systems functionality where it gets executed.
|
||||
"""
|
||||
# Create a singleton instance variable
|
||||
instance = None
|
||||
|
||||
def __new__(cls, name: str = "ProxLB", level: str = logging.INFO) -> 'SystemdLogger':
|
||||
"""
|
||||
Creating a new systemd logger class based on a given logging name
|
||||
and its logging level/verbosity.
|
||||
|
||||
Args:
|
||||
name (str): The application name that is being used for the logger.
|
||||
level (str): The log level defined as a string (e.g.: INFO).
|
||||
|
||||
Returns:
|
||||
SystemdLogger: The systemd logger object.
|
||||
"""
|
||||
# Check if instance already exists, otherwise create a new one
|
||||
if cls.instance is None:
|
||||
cls.instance = super(SystemdLogger, cls).__new__(cls)
|
||||
cls.instance.initialize_logger(name, level)
|
||||
return cls.instance
|
||||
|
||||
def initialize_logger(self, name: str, level: str) -> None:
|
||||
"""
|
||||
Initializing the systemd logger class based on a given logging name
|
||||
and its logging level/verbosity.
|
||||
|
||||
Args:
|
||||
name (str): The application name that is being used for the logger.
|
||||
level (str): The log level defined as a string (e.g.: INFO).
|
||||
"""
|
||||
self.logger = logging.getLogger(name)
|
||||
self.logger.setLevel(level)
|
||||
|
||||
# Create a JournalHandler for systemd integration if this
|
||||
# is supported on the underlying OS.
|
||||
if SYSTEMD_PRESENT:
|
||||
# Add a JournalHandler for systemd integration
|
||||
journal_handler = JournalHandler()
|
||||
journal_handler.setLevel(level)
|
||||
# Set a formatter to include the logger's name and log message
|
||||
formatter = logging.Formatter('%(asctime)s - %(name)s - %(levelname)s - %(message)s')
|
||||
journal_handler.setFormatter(formatter)
|
||||
# Add handler to logger
|
||||
self.logger.addHandler(journal_handler)
|
||||
|
||||
def set_log_level(self, level: str) -> None:
|
||||
"""
|
||||
Modifies and sets the log level on the given log level.
|
||||
|
||||
Args:
|
||||
level (str): The log level defined as a string (e.g.: INFO).
|
||||
"""
|
||||
self.logger.setLevel(level)
|
||||
|
||||
for handler in self.logger.handlers:
|
||||
handler.setLevel(level)
|
||||
|
||||
self.logger.debug("Set to debug level")
|
||||
|
||||
# Handle systemd log levels
|
||||
def debug(self, msg: str) -> str:
|
||||
"""
|
||||
Logger out for messages of type: DEBUG
|
||||
"""
|
||||
self.logger.debug(msg)
|
||||
|
||||
def info(self, msg: str) -> str:
|
||||
"""
|
||||
Logger out for messages of type: INFO
|
||||
"""
|
||||
self.logger.info(msg)
|
||||
|
||||
def warning(self, msg: str) -> str:
|
||||
"""
|
||||
Logger out for messages of type: WARNING
|
||||
"""
|
||||
self.logger.warning(msg)
|
||||
|
||||
def error(self, msg: str) -> str:
|
||||
"""
|
||||
Logger out for messages of type: ERROR
|
||||
"""
|
||||
self.logger.error(msg)
|
||||
|
||||
def critical(self, msg: str) -> str:
|
||||
"""
|
||||
Logger out for messages of type: CRITICAL
|
||||
"""
|
||||
self.logger.critical(msg)
|
||||
292
proxlb/utils/proxmox_api.py
Normal file
292
proxlb/utils/proxmox_api.py
Normal file
@@ -0,0 +1,292 @@
|
||||
"""
|
||||
Module providing a function printing python version.
|
||||
"""
|
||||
|
||||
try:
|
||||
import proxmoxer
|
||||
PROXMOXER_PRESENT = True
|
||||
except ImportError:
|
||||
PROXMOXER_PRESENT = False
|
||||
import random
|
||||
import socket
|
||||
import sys
|
||||
try:
|
||||
import requests
|
||||
REQUESTS_PRESENT = True
|
||||
except ImportError:
|
||||
REQUESTS_PRESENT = False
|
||||
try:
|
||||
import urllib3
|
||||
URLLIB3_PRESENT = True
|
||||
except ImportError:
|
||||
URLLIB3_PRESENT = False
|
||||
from typing import Dict, Any
|
||||
from utils.logger import SystemdLogger
|
||||
|
||||
|
||||
if not PROXMOXER_PRESENT:
|
||||
print("Error: The required library 'proxmoxer' is not installed.")
|
||||
sys.exit(1)
|
||||
|
||||
if not URLLIB3_PRESENT:
|
||||
print("Error: The required library 'urllib3' is not installed.")
|
||||
sys.exit(1)
|
||||
|
||||
if not REQUESTS_PRESENT:
|
||||
print("Error: The required library 'requests' is not installed.")
|
||||
sys.exit(1)
|
||||
|
||||
|
||||
logger = SystemdLogger()
|
||||
|
||||
|
||||
class ProxmoxApi:
|
||||
"""
|
||||
Handles command-line argument parsing for ProxLB.
|
||||
"""
|
||||
def __init__(self, proxlb_config: Dict[str, Any]) -> None:
|
||||
"""
|
||||
Initialize the ProxmoxApi instance.
|
||||
|
||||
This method sets up the ProxmoxApi instance by testing the required module dependencies
|
||||
and establishing a connection to the Proxmox API using the provided configuration.
|
||||
|
||||
Args:
|
||||
proxlb_config (Dict[str, Any]): Configuration dictionary containing Proxmox API connection details.
|
||||
|
||||
Returns:
|
||||
None
|
||||
"""
|
||||
logger.debug("Starting: ProxmoxApi initialization.")
|
||||
self.proxmox_api = self.api_connect(proxlb_config)
|
||||
logger.debug("Finished: ProxmoxApi initialization.")
|
||||
|
||||
def __getattr__(self, name):
|
||||
"""
|
||||
Delegate attribute access to proxmox_api.
|
||||
"""
|
||||
return getattr(self.proxmox_api, name)
|
||||
|
||||
def api_connect_get_hosts(self, proxmox_api_endpoints: list) -> str:
|
||||
"""
|
||||
Perform a connectivity test to determine a working host for the Proxmox API.
|
||||
|
||||
This method takes a list of Proxmox API endpoints and validates their connectivity.
|
||||
It returns a working host from the list. If only one endpoint is provided, it is
|
||||
returned immediately. If multiple endpoints are provided, each one is tested for
|
||||
connectivity. If a valid host is found, it is returned. If multiple valid hosts
|
||||
are found, one is chosen at random to distribute the load across the cluster.
|
||||
|
||||
Args:
|
||||
proxmox_api_endpoints (list): A list of Proxmox API endpoints to test.
|
||||
|
||||
Returns:
|
||||
str: A working Proxmox API host.
|
||||
|
||||
Raises:
|
||||
SystemExit: If the provided endpoints are not a list, if the list is empty,
|
||||
or if no valid hosts are found.
|
||||
"""
|
||||
logger.debug("Starting: api_connect_get_hosts.")
|
||||
# Pre-validate the given API endpoints
|
||||
if not isinstance(proxmox_api_endpoints, list):
|
||||
logger.critical(f"The proxmox_api hosts are not defined as a list type.")
|
||||
sys.exit(1)
|
||||
|
||||
if not proxmox_api_endpoints:
|
||||
logger.critical(f"No proxmox_api hosts are defined.")
|
||||
sys.exit(1)
|
||||
|
||||
if len(proxmox_api_endpoints) == 0:
|
||||
logger.critical(f"No proxmox_api hosts are defined.")
|
||||
sys.exit(1)
|
||||
|
||||
# Get a suitable Proxmox API endpoint. Therefore, we check if we only have
|
||||
# a single Proxmox API endpoint or multiple ones. If only one, we can return
|
||||
# this one immediately. If this one does not work, the urllib will raise an
|
||||
# exception during the connection attempt.
|
||||
if len(proxmox_api_endpoints) == 1:
|
||||
return proxmox_api_endpoints[0]
|
||||
|
||||
# If we have multiple Proxmox API endpoints, we need to check each one by
|
||||
# doing a connection attempt for IPv4 and IPv6. If we find a working one,
|
||||
# we return that one. This allows us to define multiple endpoints in a cluster.
|
||||
validated_api_hosts = []
|
||||
for host in proxmox_api_endpoints:
|
||||
validated = self.test_api_proxmox_host(host)
|
||||
if validated:
|
||||
validated_api_hosts.append(validated)
|
||||
|
||||
if len(validated_api_hosts) > 0:
|
||||
# Choose a random host to distribute the load across the cluster
|
||||
# as a simple load balancing mechanism.
|
||||
return random.choice(validated_api_hosts)
|
||||
|
||||
logger.critical("No valid Proxmox API hosts found.")
|
||||
print("No valid Proxmox API hosts found.")
|
||||
|
||||
logger.debug("Finished: api_connect_get_hosts.")
|
||||
sys.exit(1)
|
||||
|
||||
def test_api_proxmox_host(self, host: str) -> str:
|
||||
"""
|
||||
Tests the connectivity to a Proxmox host by resolving its IP address and
|
||||
checking both IPv4 and IPv6 addresses.
|
||||
|
||||
This function attempts to resolve the given hostname to its IP addresses
|
||||
(both IPv4 and IPv6). It then tests the connectivity to the Proxmox API
|
||||
using the resolved IP addresses. If the host is reachable via either
|
||||
IPv4 or IPv6, the function returns the hostname. If the host is not
|
||||
reachable, the function returns False.
|
||||
|
||||
Args:
|
||||
host (str): The hostname of the Proxmox server to test.
|
||||
|
||||
Returns:
|
||||
str: The hostname if the Proxmox server is reachable.
|
||||
bool: False if the Proxmox server is not reachable.
|
||||
"""
|
||||
logger.debug("Starting: test_api_proxmox_host.")
|
||||
ip = socket.getaddrinfo(host, None, socket.AF_UNSPEC)
|
||||
for address_type in ip:
|
||||
if address_type[0] == socket.AF_INET:
|
||||
logger.debug(f"{host} is type ipv4.")
|
||||
if self.test_api_proxmox_host_ipv4(host):
|
||||
return host
|
||||
elif address_type[0] == socket.AF_INET6:
|
||||
logger.debug(f"{host} is type ipv6.")
|
||||
if self.test_api_proxmox_host_ipv6(host):
|
||||
return host
|
||||
else:
|
||||
return False
|
||||
|
||||
logger.debug("Finished: test_api_proxmox_host.")
|
||||
|
||||
def test_api_proxmox_host_ipv4(self, host: str, port: int = 8006, timeout: int = 1) -> bool:
|
||||
"""
|
||||
Test the reachability of a Proxmox host on its IPv4 management address.
|
||||
|
||||
This method attempts to establish a TCP connection to the specified host and port
|
||||
within a given timeout period. It logs the process and results, indicating whether
|
||||
the host is reachable or not.
|
||||
|
||||
Args:
|
||||
host (str): The IPv4 address or hostname of the Proxmox host to test.
|
||||
port (int, optional): The TCP port to connect to on the host. Defaults to 8006.
|
||||
timeout (int, optional): The timeout duration in seconds for the connection attempt. Defaults to 1.
|
||||
|
||||
Returns:
|
||||
bool: True if the host is reachable on the specified port, False otherwise.
|
||||
"""
|
||||
logger.debug("Starting: test_api_proxmox_host_ipv4.")
|
||||
sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
|
||||
sock.settimeout(timeout)
|
||||
logger.warning(f"Warning: Host {host} ran into a timout when connectoing on IPv4 for tcp/{port}.")
|
||||
result = sock.connect_ex((host, port))
|
||||
|
||||
if result == 0:
|
||||
sock.close()
|
||||
logger.debug(f"Host {host} is reachable on IPv4 for tcp/{port}.")
|
||||
return True
|
||||
|
||||
sock.close()
|
||||
logger.warning(f"Host {host} is unreachable on IPv4 for tcp/{port}.")
|
||||
|
||||
logger.debug("Finished: test_api_proxmox_host_ipv4.")
|
||||
return False
|
||||
|
||||
def test_api_proxmox_host_ipv6(self, host: str, port: int = 8006, timeout: int = 1) -> bool:
|
||||
"""
|
||||
Test the reachability of a Proxmox host on its IPv6 management address.
|
||||
|
||||
This method attempts to establish a TCP connection to the specified host and port
|
||||
within a given timeout period. It logs the process and results, indicating whether
|
||||
the host is reachable or not.
|
||||
|
||||
Args:
|
||||
host (str): The IPv6 address or hostname of the Proxmox host to test.
|
||||
port (int, optional): The TCP port to connect to on the host. Defaults to 8006.
|
||||
timeout (int, optional): The timeout duration in seconds for the connection attempt. Defaults to 1.
|
||||
|
||||
Returns:
|
||||
bool: True if the host is reachable on the specified port, False otherwise.
|
||||
"""
|
||||
logger.debug("Starting: test_api_proxmox_host_ipv6.")
|
||||
sock = socket.socket(socket.AF_INET6, socket.SOCK_STREAM)
|
||||
sock.settimeout(timeout)
|
||||
logger.warning(f"Host {host} ran into a timout when connectoing on IPv6 for tcp/{port}.")
|
||||
result = sock.connect_ex((host, port))
|
||||
|
||||
if result == 0:
|
||||
sock.close()
|
||||
logger.debug(f"Host {host} is reachable on IPv6 for tcp/{port}.")
|
||||
return True
|
||||
|
||||
sock.close()
|
||||
logger.warning(f"Host {host} is unreachable on IPv6 for tcp/{port}.")
|
||||
|
||||
logger.debug("Finished: test_api_proxmox_host_ipv4.")
|
||||
return False
|
||||
|
||||
def api_connect(self, proxlb_config: Dict[str, Any]) -> proxmoxer.ProxmoxAPI:
|
||||
"""
|
||||
Establishes a connection to the Proxmox API using the provided configuration.
|
||||
|
||||
This function retrieves the Proxmox API endpoint from the configuration, optionally disables SSL certificate
|
||||
validation warnings, and attempts to authenticate and create a ProxmoxAPI object. It handles various exceptions
|
||||
related to authentication, connection timeouts, SSL errors, and connection refusals, logging appropriate error
|
||||
messages and exiting the program if necessary.
|
||||
|
||||
Args:
|
||||
proxlb_config (Dict[str, Any]): A dictionary containing the Proxmox API configuration. Expected keys include:
|
||||
- "proxmox_api": A dictionary with the following keys:
|
||||
- "hosts" (List[str]): A list of Proxmox API host addresses.
|
||||
- "user" (str): The username for Proxmox API authentication.
|
||||
- "pass" (str): The password for Proxmox API authentication.
|
||||
- "ssl_verification" (bool): Whether to verify SSL certificates (default is True).
|
||||
- "timeout" (int): The timeout duration for API requests.
|
||||
|
||||
Returns:
|
||||
proxmoxer.ProxmoxAPI: An authenticated ProxmoxAPI object.
|
||||
|
||||
Raises:
|
||||
proxmoxer.backends.https.AuthenticationError: If authentication fails.
|
||||
requests.exceptions.ConnectTimeout: If the connection to the Proxmox API times out.
|
||||
requests.exceptions.SSLError: If SSL certificate validation fails.
|
||||
requests.exceptions.ConnectionError: If the connection to the Proxmox API is refused.
|
||||
"""
|
||||
logger.debug("Starting: api_connect.")
|
||||
# Get a valid Proxmox API endpoint
|
||||
proxmox_api_endpoint = self.api_connect_get_hosts(proxlb_config.get("proxmox_api", {}).get("hosts", []))
|
||||
|
||||
# Disable warnings for SSL certificate validation
|
||||
if not proxlb_config.get("proxmox_api").get("ssl_verification", True):
|
||||
logger.warning(f"SSL certificate validation to host {proxmox_api_endpoint} is deactivated.")
|
||||
urllib3.disable_warnings(category=urllib3.exceptions.InsecureRequestWarning)
|
||||
requests.packages.urllib3.disable_warnings()
|
||||
|
||||
# Login into Proxmox API and create API object
|
||||
try:
|
||||
proxmox_api = proxmoxer.ProxmoxAPI(
|
||||
proxmox_api_endpoint,
|
||||
user=proxlb_config.get("proxmox_api").get("user", True),
|
||||
password=proxlb_config.get("proxmox_api").get("pass", True),
|
||||
verify_ssl=proxlb_config.get("proxmox_api").get("ssl_verification", True),
|
||||
timeout=proxlb_config.get("proxmox_api").get("timeout", True))
|
||||
except proxmoxer.backends.https.AuthenticationError as proxmox_api_error:
|
||||
logger.critical(f"Authentication failed. Please check the defined credentials: {proxmox_api_error}")
|
||||
sys.exit(2)
|
||||
except requests.exceptions.ConnectTimeout:
|
||||
logger.critical(f"Connection timeout to host {proxmox_api_endpoint}")
|
||||
sys.exit(2)
|
||||
except requests.exceptions.SSLError as proxmox_api_error:
|
||||
logger.critical(f"SSL certificate validation failed: {proxmox_api_error}")
|
||||
sys.exit(2)
|
||||
except requests.exceptions.ConnectionError:
|
||||
logger.critical(f"Connection refused by host {proxmox_api_endpoint}")
|
||||
sys.exit(2)
|
||||
|
||||
logger.info(f"API connection to host {proxmox_api_endpoint} succeeded.")
|
||||
|
||||
logger.debug("Finished: api_connect.")
|
||||
return proxmox_api
|
||||
5
proxlb/utils/version.py
Normal file
5
proxlb/utils/version.py
Normal file
@@ -0,0 +1,5 @@
|
||||
__app_name__ = "ProxLB"
|
||||
__app_desc__ = "A DRS alike loadbalancer for Proxmox clusters."
|
||||
__author__ = "Florian Paul Azim Hoberg <gyptazy>"
|
||||
__version__ = "1.1.0-alpha"
|
||||
__url__ = "https://github.com/gyptazy/ProxLB"
|
||||
@@ -1,5 +1,4 @@
|
||||
argparse
|
||||
configparser
|
||||
proxmoxer
|
||||
requests
|
||||
urllib3
|
||||
urllib3
|
||||
PyYAML
|
||||
11
service/proxlb.service
Normal file
11
service/proxlb.service
Normal file
@@ -0,0 +1,11 @@
|
||||
[Unit]
|
||||
Description=ProxLB - A loadbalancer for Proxmox clusters
|
||||
After=network-online.target
|
||||
Wants=network-online.target
|
||||
|
||||
[Service]
|
||||
ExecStart=python3 /usr/lib/python3/dist-packages/proxlb/main.py -c /etc/proxlb/proxlb.yaml
|
||||
User=plb
|
||||
|
||||
[Install]
|
||||
WantedBy=multi-user.target
|
||||
21
setup.py
Normal file
21
setup.py
Normal file
@@ -0,0 +1,21 @@
|
||||
from setuptools import setup
|
||||
|
||||
setup(
|
||||
name="proxlb",
|
||||
version="1.1.0-alpha",
|
||||
description="A DRS alike loadbalancer for Proxmox clusters.",
|
||||
long_description="An advanced DRS alike loadbalancer for Proxmox clusters that also supports maintenance modes and affinity/anti-affinity rules.",
|
||||
author="Florian Paul Azim Hoberg",
|
||||
author_email="gyptazy@gyptazy.com",
|
||||
maintainer="Florian Paul Azim Hoberg",
|
||||
maintainer_email="gyptazy@gyptazy.com",
|
||||
url="https://github.com/gyptazy/ProxLB",
|
||||
packages=["proxlb", "proxlb.utils", "proxlb.models"],
|
||||
install_requires=[
|
||||
"requests",
|
||||
"urllib3",
|
||||
"proxmoxer",
|
||||
"pyyaml",
|
||||
],
|
||||
data_files=[('/etc/systemd/system', ['service/proxlb.service']), ('/etc/proxlb/', ['config/proxlb_example.yaml'])],
|
||||
)
|
||||
@@ -1 +0,0 @@
|
||||
## Unit Tests
|
||||
175
tests/tests.py
175
tests/tests.py
@@ -1,175 +0,0 @@
|
||||
import unittest
|
||||
from unittest.mock import patch, MagicMock
|
||||
import logging
|
||||
import sys
|
||||
import os
|
||||
import configparser
|
||||
|
||||
from proxlb import (
|
||||
initialize_logger,
|
||||
pre_validations,
|
||||
post_validations,
|
||||
validate_daemon,
|
||||
__validate_imports,
|
||||
__validate_config_file,
|
||||
initialize_args,
|
||||
initialize_config_path,
|
||||
initialize_config_options,
|
||||
api_connect,
|
||||
get_node_statistics,
|
||||
get_vm_statistics,
|
||||
balancing_calculations,
|
||||
__get_node_most_free_values,
|
||||
run_vm_rebalancing,
|
||||
SystemdHandler,
|
||||
__errors__
|
||||
)
|
||||
|
||||
class TestProxLB(unittest.TestCase):
|
||||
|
||||
def test_initialize_logger(self):
|
||||
with patch('logging.getLogger') as mock_get_logger, patch('logging.Handler'):
|
||||
mock_logger = MagicMock()
|
||||
mock_get_logger.return_value = mock_logger
|
||||
initialize_logger(logging.DEBUG, SystemdHandler())
|
||||
mock_logger.setLevel.assert_called_with(logging.DEBUG)
|
||||
self.assertTrue(mock_logger.addHandler.called)
|
||||
|
||||
def test_pre_validations(self):
|
||||
with patch('proxlb.__validate_imports') as mock_validate_imports, patch('proxlb.__validate_config_file') as mock_validate_config_file:
|
||||
pre_validations('/path/to/config')
|
||||
self.assertTrue(mock_validate_imports.called)
|
||||
mock_validate_config_file.assert_called_with('/path/to/config')
|
||||
|
||||
def test_post_validations(self):
|
||||
global __errors__
|
||||
__errors__ = False
|
||||
with patch('logging.critical') as mock_critical, patch('logging.info') as mock_info:
|
||||
post_validations()
|
||||
self.assertTrue(mock_info.called)
|
||||
self.assertFalse(mock_critical.called)
|
||||
|
||||
__errors__ = True
|
||||
with patch('logging.critical') as mock_critical, patch('logging.info'):
|
||||
post_validations()
|
||||
self.assertTrue(mock_critical.called)
|
||||
|
||||
def test_validate_daemon(self):
|
||||
with patch('logging.info') as mock_info, patch('time.sleep') as mock_sleep, patch('sys.exit') as mock_exit:
|
||||
validate_daemon(1, 1)
|
||||
self.assertTrue(mock_info.called)
|
||||
self.assertTrue(mock_sleep.called)
|
||||
|
||||
validate_daemon(0, 1)
|
||||
self.assertTrue(mock_exit.called)
|
||||
|
||||
def test_validate_imports(self):
|
||||
global _imports
|
||||
_imports = True
|
||||
with patch('logging.critical') as mock_critical, patch('logging.info') as mock_info, patch('sys.exit') as mock_exit:
|
||||
__validate_imports()
|
||||
self.assertTrue(mock_info.called)
|
||||
self.assertFalse(mock_exit.called)
|
||||
self.assertFalse(mock_critical.called)
|
||||
|
||||
_imports = False
|
||||
with patch('logging.critical') as mock_critical, patch('logging.info'), patch('sys.exit') as mock_exit:
|
||||
__validate_imports()
|
||||
self.assertTrue(mock_critical.called)
|
||||
self.assertTrue(mock_exit.called)
|
||||
|
||||
def test_validate_config_file(self):
|
||||
with patch('os.path.isfile', return_value=True), patch('logging.critical') as mock_critical, patch('logging.info') as mock_info, patch('sys.exit') as mock_exit:
|
||||
__validate_config_file('/path/to/config')
|
||||
self.assertTrue(mock_info.called)
|
||||
self.assertFalse(mock_exit.called)
|
||||
self.assertFalse(mock_critical.called)
|
||||
|
||||
with patch('os.path.isfile', return_value=False), patch('logging.critical') as mock_critical, patch('logging.info'), patch('sys.exit') as mock_exit:
|
||||
__validate_config_file('/path/to/config')
|
||||
self.assertTrue(mock_critical.called)
|
||||
self.assertTrue(mock_exit.called)
|
||||
|
||||
@patch('argparse.ArgumentParser.parse_args', return_value=argparse.Namespace(config='/path/to/config'))
|
||||
def test_initialize_args(self, mock_parse_args):
|
||||
args = initialize_args()
|
||||
self.assertEqual(args.config, '/path/to/config')
|
||||
|
||||
def test_initialize_config_path(self):
|
||||
app_args = MagicMock(config='/path/to/config')
|
||||
with patch('logging.info') as mock_info:
|
||||
config_path = initialize_config_path(app_args)
|
||||
self.assertEqual(config_path, '/path/to/config')
|
||||
self.assertTrue(mock_info.called)
|
||||
|
||||
app_args.config = None
|
||||
with patch('logging.info') as mock_info:
|
||||
config_path = initialize_config_path(app_args)
|
||||
self.assertEqual(config_path, '/etc/proxlb/proxlb.conf')
|
||||
self.assertTrue(mock_info.called)
|
||||
|
||||
@patch('configparser.ConfigParser.read', side_effect=lambda x: setattr(configparser.ConfigParser(), 'proxmox', {'api_host': 'host', 'api_user': 'user', 'api_pass': 'pass', 'verify_ssl': '0'}))
|
||||
def test_initialize_config_options(self, mock_read):
|
||||
with patch('logging.info') as mock_info, patch('sys.exit') as mock_exit:
|
||||
config_path = '/path/to/config'
|
||||
proxmox_api_host, proxmox_api_user, proxmox_api_pass, proxmox_api_ssl_v, balancing_method, ignore_nodes, ignore_vms, daemon, schedule = initialize_config_options(config_path)
|
||||
self.assertEqual(proxmox_api_host, 'host')
|
||||
self.assertEqual(proxmox_api_user, 'user')
|
||||
self.assertEqual(proxmox_api_pass, 'pass')
|
||||
self.assertEqual(proxmox_api_ssl_v, '0')
|
||||
self.assertTrue(mock_info.called)
|
||||
self.assertFalse(mock_exit.called)
|
||||
|
||||
@patch('proxmoxer.ProxmoxAPI')
|
||||
def test_api_connect(self, mock_proxmox_api):
|
||||
with patch('requests.packages.urllib3.disable_warnings') as mock_disable_warnings, patch('logging.warning') as mock_warning, patch('logging.info') as mock_info:
|
||||
proxmox_api_ssl_v = 0
|
||||
api_object = api_connect('host', 'user', 'pass', proxmox_api_ssl_v)
|
||||
self.assertTrue(mock_disable_warnings.called)
|
||||
self.assertTrue(mock_warning.called)
|
||||
self.assertTrue(mock_info.called)
|
||||
self.assertTrue(mock_proxmox_api.called)
|
||||
|
||||
def test_get_node_statistics(self):
|
||||
mock_api_object = MagicMock()
|
||||
mock_api_object.nodes.get.return_value = [{'node': 'node1', 'status': 'online', 'maxcpu': 100, 'cpu': 50, 'maxmem': 1000, 'mem': 500, 'maxdisk': 10000, 'disk': 5000}]
|
||||
node_statistics = get_node_statistics(mock_api_object, '')
|
||||
self.assertIn('node1', node_statistics)
|
||||
self.assertEqual(node_statistics['node1']['cpu_total'], 100)
|
||||
self.assertEqual(node_statistics['node1']['cpu_used'], 50)
|
||||
self.assertEqual(node_statistics['node1']['memory_total'], 1000)
|
||||
self.assertEqual(node_statistics['node1']['memory_used'], 500)
|
||||
self.assertEqual(node_statistics['node1']['disk_total'], 10000)
|
||||
self.assertEqual(node_statistics['node1']['disk_used'], 5000)
|
||||
|
||||
def test_get_vm_statistics(self):
|
||||
mock_api_object = MagicMock()
|
||||
mock_api_object.nodes.get.return_value = [{'node': 'node1', 'status': 'online'}]
|
||||
mock_api_object.nodes().qemu.get.return_value = [{'name': 'vm1', 'status': 'running', 'cpus': 4, 'cpu': 2, 'maxmem': 8000, 'mem': 4000, 'maxdisk': 20000, 'disk': 10000, 'vmid': 101}]
|
||||
vm_statistics = get_vm_statistics(mock_api_object, '')
|
||||
self.assertIn('vm1', vm_statistics)
|
||||
self.assertEqual(vm_statistics['vm1']['cpu_total'], 4)
|
||||
self.assertEqual(vm_statistics['vm1']['cpu_used'], 2)
|
||||
self.assertEqual(vm_statistics['vm1']['memory_total'], 8000)
|
||||
self.assertEqual(vm_statistics['vm1']['memory_used'], 4000)
|
||||
self.assertEqual(vm_statistics['vm1']['disk_total'], 20000)
|
||||
self.assertEqual(vm_statistics['vm1']['disk_used'], 10000)
|
||||
self.assertEqual(vm_statistics['vm1']['vmid'], 101)
|
||||
self.assertEqual(vm_statistics['vm1']['node_parent'], 'node1')
|
||||
|
||||
def test_balancing_calculations(self):
|
||||
node_statistics = {
|
||||
'node1': {'cpu_free': 80, 'memory_free': 8000, 'disk_free': 80000},
|
||||
'node2': {'cpu_free': 70, 'memory_free': 7000, 'disk_free': 70000}
|
||||
}
|
||||
vm_statistics = {
|
||||
'vm1': {'cpu_used': 20, 'memory_used': 2000, 'disk_used': 20000, 'node_parent': 'node1'},
|
||||
'vm2': {'cpu_used': 30, 'memory_used': 3000, 'disk_used': 30000, 'node_parent': 'node1'}
|
||||
}
|
||||
with patch('logging.info') as mock_info, patch('logging.error') as mock_error:
|
||||
node_statistics_rebalanced, vm_statistics_rebalanced = balancing_calculations('memory', node_statistics, vm_statistics)
|
||||
self.assertTrue(mock_info.called)
|
||||
self.assertFalse(mock_error.called)
|
||||
self.assertEqual(vm_statistics_rebalanced['vm1']['node_rebalance'], 'node2')
|
||||
self.assertEqual(vm_statistics_rebalanced['vm2']['node
|
||||
|
||||
Reference in New Issue
Block a user