Compare commits
60 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
| e930178e9c | |||
| 5afc940df8 | |||
| 3961aa3448 | |||
| 8fd57032e7 | |||
| 5ede7dea07 | |||
| 5ea007c3f6 | |||
| 7cb0f7ad40 | |||
| 83fa12ec54 | |||
| 98b74d9aed | |||
| 45953848e2 | |||
| 6116f4f885 | |||
| 24d1214855 | |||
| 3c39d8aa97 | |||
| ebb084aa9d | |||
| 86d5e8917b | |||
| be84e2b89a | |||
| fd0a9e1230 | |||
| b934a30124 | |||
| ea397ef889 | |||
| def6bab7f4 | |||
| ecb616f6d9 | |||
| b7f731c6ac | |||
| 4fba81dc2c | |||
| 9225e745b0 | |||
| b981a9ad36 | |||
| c0185d4b86 | |||
| d34dfc5bf7 | |||
| 7c395a80fe | |||
| 8fe81e3ab3 | |||
| cfe01eb63e | |||
| 55b9b16deb | |||
| 6a22597f1f | |||
| 295d1d6310 | |||
| 8b804f3912 | |||
| bc9e039cb3 | |||
| 54cf117591 | |||
| 8e32e01c20 | |||
| af3bbf7515 | |||
| 180fa31d14 | |||
| 39a6a9d23e | |||
| f716569aa7 | |||
| 22ae634a87 | |||
| 375bf6a982 | |||
| 487386a7cc | |||
| 15097dcba4 | |||
| 1d642d41b2 | |||
| 80304f40d1 | |||
| c0e1531083 | |||
| 529bef9679 | |||
| 8928bba337 | |||
| c872fc6b90 | |||
| 030afafc20 | |||
| 2dfc336288 | |||
| 27e1c517bc | |||
| e6166d1fe3 | |||
| bffabd9c8f | |||
| 98ac25743b | |||
| 7fc23961b0 | |||
| 8ace133c23 | |||
| 2afeb827ed |
@ -3,7 +3,7 @@ Version: 1.0
|
||||
Section: utils
|
||||
Priority: optional
|
||||
Architecture: all
|
||||
Depends: logrotate, python3 (>= 3.6), python3-psycopg2, python3-yaml, systemd
|
||||
Depends: logrotate, python3 (>= 3.6), python3-psycopg2, python3-requests, python3-yaml, systemd
|
||||
Maintainer: James Campbell <james@commandprompt.com>
|
||||
Homepage: https://www.commandprompt.com
|
||||
Description: A bridge to sit between monitoring tools and PostgreSQL
|
||||
|
||||
@ -1,52 +0,0 @@
|
||||
# Copyright 2024 Gentoo Authors
|
||||
# Distributed under the terms of the GNU General Public License v2
|
||||
|
||||
EAPI=8
|
||||
|
||||
PYTHON_COMPAT=( python3_{6..12} )
|
||||
|
||||
inherit git-r3 python-r1
|
||||
|
||||
DESCRIPTION="PostgreSQL monitoring bridge"
|
||||
HOMEPAGE="None"
|
||||
|
||||
LICENSE="BSD"
|
||||
SLOT="0"
|
||||
KEYWORDS="amd64"
|
||||
|
||||
EGIT_REPO_URI="https://code2.shh-dot-com.org/james/pgmon.git"
|
||||
#EGIT_COMMIT=""
|
||||
|
||||
DEPEND="
|
||||
${PYTHON_DEPS}
|
||||
dev-python/psycopg:3
|
||||
dev-python/pyyaml
|
||||
acct-user/zabbix
|
||||
acct-group/zabbix
|
||||
agent? ( net-analyzer/zabbix[agent] )
|
||||
agent2? ( net-analyzer/zabbix[agent2] )
|
||||
app-admin/logrotate
|
||||
"
|
||||
RDEPEND="${DEPEND}"
|
||||
BDEPEND=""
|
||||
|
||||
src_install() {
|
||||
default
|
||||
|
||||
# Install init script
|
||||
newinitd "${FILESDIR}/pgmon.openrc" pgmon
|
||||
|
||||
# Install script
|
||||
exeinto /usr/bin
|
||||
newexe "${S}/pgmon.py" pgmon
|
||||
|
||||
# Install default config
|
||||
diropts -o root -g zabbix -m 0755
|
||||
insinto /etc/pgmon
|
||||
doins "${FILESDIR}/pgmon.yml"
|
||||
doins "${S}/pgmon-metrics.yml"
|
||||
|
||||
# Install logrotate config
|
||||
insinto /etc/logrotate.d
|
||||
newins "${FILESDIR}/pgmon.logrotate" pgmon
|
||||
}
|
||||
@ -5,7 +5,7 @@ EAPI=8
|
||||
|
||||
PYTHON_COMPAT=( python3_{6..13} )
|
||||
|
||||
inherit git-r3 python-r1
|
||||
inherit python-r1
|
||||
|
||||
DESCRIPTION="PostgreSQL monitoring bridge"
|
||||
HOMEPAGE="None"
|
||||
@ -14,7 +14,9 @@ LICENSE="BSD"
|
||||
SLOT="0"
|
||||
KEYWORDS="amd64"
|
||||
|
||||
SRC_URI="https://code2.shh-dot-com.org/james/${PN}/archive/v${PV}.tar.gz -> ${P}.tar.gz"
|
||||
SRC_URI="https://code2.shh-dot-com.org/james/${PN}/archive/v${PV}.tar.bz2 -> ${P}.tar.bz2"
|
||||
|
||||
IUSE="-systemd"
|
||||
|
||||
DEPEND="
|
||||
${PYTHON_DEPS}
|
||||
@ -25,21 +27,36 @@ DEPEND="
|
||||
RDEPEND="${DEPEND}"
|
||||
BDEPEND=""
|
||||
|
||||
S="${WORKDIR}/${PN}"
|
||||
RESTRICT="fetch"
|
||||
|
||||
#S="${WORKDIR}/${PN}"
|
||||
|
||||
pkg_nofetch() {
|
||||
einfo "Please download"
|
||||
einfo " - ${P}.tar.bz2"
|
||||
einfo "from ${HOMEPAGE} and place it in your DISTDIR directory."
|
||||
einfo "The file should be owned by portage:portage."
|
||||
}
|
||||
|
||||
src_compile() {
|
||||
true
|
||||
}
|
||||
|
||||
src_install() {
|
||||
default
|
||||
|
||||
# Install init script
|
||||
newinitd "openrc/pgmon.initd" pgmon
|
||||
newconfd "openrc/pgmon.confd" pgmon
|
||||
if ! use systemd ; then
|
||||
newinitd "openrc/pgmon.initd" pgmon
|
||||
newconfd "openrc/pgmon.confd" pgmon
|
||||
fi
|
||||
|
||||
# Install systemd unit
|
||||
systemd_dounit "systemd/pgmon.service"
|
||||
if use systemd ; then
|
||||
systemd_dounit "systemd/pgmon.service"
|
||||
fi
|
||||
|
||||
# Install script
|
||||
exeinto /usr/bin
|
||||
newexe "pgmon.py" pgmon
|
||||
newexe "src/pgmon.py" pgmon
|
||||
|
||||
# Install default config
|
||||
diropts -o root -g root -m 0755
|
||||
73
GENTOO/pgmon-1.0.2.ebuild
Normal file
73
GENTOO/pgmon-1.0.2.ebuild
Normal file
@ -0,0 +1,73 @@
|
||||
# Copyright 2024 Gentoo Authors
|
||||
# Distributed under the terms of the GNU General Public License v2
|
||||
|
||||
EAPI=8
|
||||
|
||||
PYTHON_COMPAT=( python3_{6..13} )
|
||||
|
||||
inherit python-r1
|
||||
|
||||
DESCRIPTION="PostgreSQL monitoring bridge"
|
||||
HOMEPAGE="None"
|
||||
|
||||
LICENSE="BSD"
|
||||
SLOT="0"
|
||||
KEYWORDS="amd64"
|
||||
|
||||
SRC_URI="https://code2.shh-dot-com.org/james/${PN}/archive/v${PV}.tar.bz2 -> ${P}.tar.bz2"
|
||||
|
||||
IUSE="-systemd"
|
||||
|
||||
DEPEND="
|
||||
${PYTHON_DEPS}
|
||||
dev-python/psycopg:2
|
||||
dev-python/pyyaml
|
||||
app-admin/logrotate
|
||||
"
|
||||
RDEPEND="${DEPEND}"
|
||||
BDEPEND=""
|
||||
|
||||
RESTRICT="fetch"
|
||||
|
||||
#S="${WORKDIR}/${PN}"
|
||||
|
||||
pkg_nofetch() {
|
||||
einfo "Please download"
|
||||
einfo " - ${P}.tar.bz2"
|
||||
einfo "from ${HOMEPAGE} and place it in your DISTDIR directory."
|
||||
einfo "The file should be owned by portage:portage."
|
||||
}
|
||||
|
||||
src_compile() {
|
||||
true
|
||||
}
|
||||
|
||||
src_install() {
|
||||
# Install init script
|
||||
if ! use systemd ; then
|
||||
newinitd "openrc/pgmon.initd" pgmon
|
||||
newconfd "openrc/pgmon.confd" pgmon
|
||||
fi
|
||||
|
||||
# Install systemd unit
|
||||
if use systemd ; then
|
||||
systemd_dounit "systemd/pgmon.service"
|
||||
fi
|
||||
|
||||
# Install script
|
||||
exeinto /usr/bin
|
||||
newexe "src/pgmon.py" pgmon
|
||||
|
||||
# Install default config
|
||||
diropts -o root -g root -m 0755
|
||||
insinto /etc/pgmon
|
||||
doins "sample-config/pgmon.yml"
|
||||
doins "sample-config/pgmon-metrics.yml"
|
||||
|
||||
# Install logrotate config
|
||||
insinto /etc/logrotate.d
|
||||
newins "logrotate/pgmon.logrotate" pgmon
|
||||
|
||||
# Install man page
|
||||
doman manpages/pgmon.1
|
||||
}
|
||||
74
GENTOO/pgmon-1.0.3.ebuild
Normal file
74
GENTOO/pgmon-1.0.3.ebuild
Normal file
@ -0,0 +1,74 @@
|
||||
# Copyright 2024 Gentoo Authors
|
||||
# Distributed under the terms of the GNU General Public License v2
|
||||
|
||||
EAPI=8
|
||||
|
||||
PYTHON_COMPAT=( python3_{6..13} )
|
||||
|
||||
inherit python-r1 systemd
|
||||
|
||||
DESCRIPTION="PostgreSQL monitoring bridge"
|
||||
HOMEPAGE="None"
|
||||
|
||||
LICENSE="BSD"
|
||||
SLOT="0"
|
||||
KEYWORDS="amd64"
|
||||
|
||||
SRC_URI="https://code2.shh-dot-com.org/james/${PN}/releases/download/v${PV}/${P}.tar.bz2"
|
||||
|
||||
IUSE="-systemd"
|
||||
|
||||
DEPEND="
|
||||
${PYTHON_DEPS}
|
||||
dev-python/psycopg:2
|
||||
dev-python/pyyaml
|
||||
dev-python/requests
|
||||
app-admin/logrotate
|
||||
"
|
||||
RDEPEND="${DEPEND}"
|
||||
BDEPEND=""
|
||||
|
||||
#RESTRICT="fetch"
|
||||
|
||||
#S="${WORKDIR}/${PN}"
|
||||
|
||||
#pkg_nofetch() {
|
||||
# einfo "Please download"
|
||||
# einfo " - ${P}.tar.bz2"
|
||||
# einfo "from ${HOMEPAGE} and place it in your DISTDIR directory."
|
||||
# einfo "The file should be owned by portage:portage."
|
||||
#}
|
||||
|
||||
src_compile() {
|
||||
true
|
||||
}
|
||||
|
||||
src_install() {
|
||||
# Install init script
|
||||
if ! use systemd ; then
|
||||
newinitd "openrc/pgmon.initd" pgmon
|
||||
newconfd "openrc/pgmon.confd" pgmon
|
||||
fi
|
||||
|
||||
# Install systemd unit
|
||||
if use systemd ; then
|
||||
systemd_dounit "systemd/pgmon.service"
|
||||
fi
|
||||
|
||||
# Install script
|
||||
exeinto /usr/bin
|
||||
newexe "src/pgmon.py" pgmon
|
||||
|
||||
# Install default config
|
||||
diropts -o root -g root -m 0755
|
||||
insinto /etc/pgmon
|
||||
doins "sample-config/pgmon.yml"
|
||||
doins "sample-config/pgmon-metrics.yml"
|
||||
|
||||
# Install logrotate config
|
||||
insinto /etc/logrotate.d
|
||||
newins "logrotate/pgmon.logrotate" pgmon
|
||||
|
||||
# Install man page
|
||||
doman manpages/pgmon.1
|
||||
}
|
||||
74
GENTOO/pgmon-1.0.4.ebuild
Normal file
74
GENTOO/pgmon-1.0.4.ebuild
Normal file
@ -0,0 +1,74 @@
|
||||
# Copyright 2024 Gentoo Authors
|
||||
# Distributed under the terms of the GNU General Public License v2
|
||||
|
||||
EAPI=8
|
||||
|
||||
PYTHON_COMPAT=( python3_{6..13} )
|
||||
|
||||
inherit python-r1 systemd
|
||||
|
||||
DESCRIPTION="PostgreSQL monitoring bridge"
|
||||
HOMEPAGE="None"
|
||||
|
||||
LICENSE="BSD"
|
||||
SLOT="0"
|
||||
KEYWORDS="amd64"
|
||||
|
||||
SRC_URI="https://code2.shh-dot-com.org/james/${PN}/releases/download/v${PV}/${P}.tar.bz2"
|
||||
|
||||
IUSE="-systemd"
|
||||
|
||||
DEPEND="
|
||||
${PYTHON_DEPS}
|
||||
dev-python/psycopg:2
|
||||
dev-python/pyyaml
|
||||
dev-python/requests
|
||||
app-admin/logrotate
|
||||
"
|
||||
RDEPEND="${DEPEND}"
|
||||
BDEPEND=""
|
||||
|
||||
#RESTRICT="fetch"
|
||||
|
||||
#S="${WORKDIR}/${PN}"
|
||||
|
||||
#pkg_nofetch() {
|
||||
# einfo "Please download"
|
||||
# einfo " - ${P}.tar.bz2"
|
||||
# einfo "from ${HOMEPAGE} and place it in your DISTDIR directory."
|
||||
# einfo "The file should be owned by portage:portage."
|
||||
#}
|
||||
|
||||
src_compile() {
|
||||
true
|
||||
}
|
||||
|
||||
src_install() {
|
||||
# Install init script
|
||||
if ! use systemd ; then
|
||||
newinitd "openrc/pgmon.initd" pgmon
|
||||
newconfd "openrc/pgmon.confd" pgmon
|
||||
fi
|
||||
|
||||
# Install systemd unit
|
||||
if use systemd ; then
|
||||
systemd_dounit "systemd/pgmon.service"
|
||||
fi
|
||||
|
||||
# Install script
|
||||
exeinto /usr/bin
|
||||
newexe "src/pgmon.py" pgmon
|
||||
|
||||
# Install default config
|
||||
diropts -o root -g root -m 0755
|
||||
insinto /etc/pgmon
|
||||
doins "sample-config/pgmon.yml"
|
||||
doins "sample-config/pgmon-metrics.yml"
|
||||
|
||||
# Install logrotate config
|
||||
insinto /etc/logrotate.d
|
||||
newins "logrotate/pgmon.logrotate" pgmon
|
||||
|
||||
# Install man page
|
||||
doman manpages/pgmon.1
|
||||
}
|
||||
120
Makefile
120
Makefile
@ -1,9 +1,25 @@
|
||||
# Package details
|
||||
PACKAGE_NAME := pgmon
|
||||
VERSION := 1.0
|
||||
|
||||
SCRIPT := src/$(PACKAGE_NAME).py
|
||||
|
||||
# Figure out the version components
|
||||
# Note: The release is for RPM packages, where prerelease releases are written as 0.<release>
|
||||
FULL_VERSION := $(shell grep -m 1 '^VERSION = ' "$(SCRIPT)" | sed -ne 's/.*"\(.*\)".*/\1/p')
|
||||
VERSION := $(shell echo $(FULL_VERSION) | sed -n 's/\(.*\)\(-rc.*\|$$\)/\1/p')
|
||||
RELEASE := $(shell echo $(FULL_VERSION) | sed -n 's/.*-rc\([0-9]\+\)$$/\1/p')
|
||||
|
||||
ifeq ($(RELEASE),)
|
||||
RPM_RELEASE := 1
|
||||
RPM_VERSION := $(VERSION)-$(RPM_RELEASE)
|
||||
DEB_VERSION := $(VERSION)
|
||||
else
|
||||
RPM_RELEASE := 0.$(RELEASE)
|
||||
RPM_VERSION := $(VERSION)-$(RPM_RELEASE)
|
||||
DEB_VERSION := $(VERSION)~rc$(RELEASE)
|
||||
endif
|
||||
|
||||
|
||||
# Where packages are built
|
||||
BUILD_DIR := build
|
||||
|
||||
@ -15,18 +31,29 @@ SUPPORTED := ubuntu-20.04 \
|
||||
debian-11 \
|
||||
rockylinux-8 \
|
||||
rockylinux-9 \
|
||||
oraclelinux-7
|
||||
oraclelinux-7 \
|
||||
gentoo
|
||||
|
||||
##
|
||||
# These targets are the main ones to use for most things.
|
||||
##
|
||||
|
||||
.PHONY: all clean tgz test install
|
||||
.PHONY: all clean tgz test query-tests install-common install-openrc install-systemd
|
||||
|
||||
all: package-all
|
||||
|
||||
version:
|
||||
@echo "full version=$(FULL_VERSION) version=$(VERSION) rel=$(RELEASE) rpm=$(RPM_VERSION) deb=$(DEB_VERSION)"
|
||||
|
||||
# Build all packages
|
||||
.PHONY: package-all
|
||||
all: $(foreach distro_release, $(SUPPORTED), package-$(distro_release))
|
||||
package-all: $(foreach distro_release, $(SUPPORTED), package-$(distro_release))
|
||||
|
||||
# Gentoo package (tar.gz) creation
|
||||
.PHONY: package-gentoo
|
||||
package-gentoo:
|
||||
mkdir -p $(BUILD_DIR)/gentoo
|
||||
tar --transform "s,^,$(PACKAGE_NAME)-$(FULL_VERSION)/," -acjf $(BUILD_DIR)/gentoo/$(PACKAGE_NAME)-$(FULL_VERSION).tar.bz2 --exclude .gitignore $(shell git ls-tree --full-tree --name-only -r HEAD)
|
||||
|
||||
|
||||
# Create a deb package
|
||||
@ -42,13 +69,12 @@ package-%:
|
||||
--user $(shell id -u):$(shell id -g) \
|
||||
"$(DISTRO)-packager:$(RELEASE)"
|
||||
|
||||
|
||||
# Create a tarball
|
||||
tgz:
|
||||
rm -rf $(BUILD_DIR)/tgz/root
|
||||
mkdir -p $(BUILD_DIR)/tgz/root
|
||||
$(MAKE) install DESTDIR=$(BUILD_DIR)/tgz/root
|
||||
tar -cz -f $(BUILD_DIR)/tgz/$(PACKAGE_NAME)-$(VERSION).tgz -C $(BUILD_DIR)/tgz/root .
|
||||
$(MAKE) install-openrc DESTDIR=$(BUILD_DIR)/tgz/root
|
||||
tar -cz -f $(BUILD_DIR)/tgz/$(PACKAGE_NAME)-$(FULL_VERSION).tgz -C $(BUILD_DIR)/tgz/root .
|
||||
|
||||
# Clean up the build directory
|
||||
clean:
|
||||
@ -58,18 +84,21 @@ clean:
|
||||
test:
|
||||
cd src ; python3 -m unittest
|
||||
|
||||
# Install the script at the specified base directory
|
||||
install:
|
||||
# Run query tests
|
||||
query-tests:
|
||||
cd tests ; ./run-tests.sh
|
||||
|
||||
# Install the script at the specified base directory (common components)
|
||||
install-common:
|
||||
# Set up directories
|
||||
mkdir -p $(DESTDIR)/etc/$(PACKAGE_NAME)
|
||||
mkdir -p ${DESTDIR}/etc/logrotate.d
|
||||
mkdir -p $(DESTDIR)/lib/systemd/system
|
||||
mkdir -p $(DESTDIR)/usr/local/bin
|
||||
mkdir -p $(DESTDIR)/usr/bin
|
||||
mkdir -p $(DESTDIR)/usr/share/man/man1
|
||||
|
||||
# Install script
|
||||
cp $(SCRIPT) $(DESTDIR)/usr/local/bin/$(PACKAGE_NAME)
|
||||
chmod 755 $(DESTDIR)/usr/local/bin/$(PACKAGE_NAME)
|
||||
cp $(SCRIPT) $(DESTDIR)/usr/bin/$(PACKAGE_NAME)
|
||||
chmod 755 $(DESTDIR)/usr/bin/$(PACKAGE_NAME)
|
||||
|
||||
# Install manpage
|
||||
cp manpages/* $(DESTDIR)/usr/share/man/man1/
|
||||
@ -78,15 +107,39 @@ install:
|
||||
# Install sample config
|
||||
cp sample-config/* $(DESTDIR)/etc/$(PACKAGE_NAME)/
|
||||
|
||||
# Install systemd unit files
|
||||
cp systemd/* $(DESTDIR)/lib/systemd/system/
|
||||
|
||||
# Install logrotate config
|
||||
cp logrotate/${PACKAGE_NAME}.logrotate ${DESTDIR}/etc/logrotate.d/${PACKAGE_NAME}
|
||||
|
||||
# Install for systemd
|
||||
install-systemd:
|
||||
# Install the common stuff
|
||||
$(MAKE) install-common
|
||||
|
||||
# Set up directories
|
||||
mkdir -p $(DESTDIR)/lib/systemd/system
|
||||
|
||||
# Install systemd unit files
|
||||
cp systemd/* $(DESTDIR)/lib/systemd/system/
|
||||
|
||||
# Install for open-rc
|
||||
install-openrc:
|
||||
# Install the common stuff
|
||||
$(MAKE) install-common
|
||||
|
||||
# Set up directories
|
||||
mkdir -p $(DESTDIR)/etc/init.d
|
||||
mkdir -p $(DESTDIR)/etc/conf.d
|
||||
|
||||
# Install init script
|
||||
cp openrc/pgmon.initd $(DESTDIR)/etc/init.d/pgmon
|
||||
chmod 755 $(DESTDIR)/etc/init.d/pgmon
|
||||
|
||||
# Install init script config file
|
||||
cp openrc/pgmon.confd $(DESTDIR)/etc/conf.d/pgmon
|
||||
|
||||
|
||||
# Run all of the install tests
|
||||
.PHONY: install-tests debian-%-install-test rockylinux-%-install-test ubuntu-%-install-test
|
||||
.PHONY: install-tests debian-%-install-test rockylinux-%-install-test ubuntu-%-install-test gentoo-install-test
|
||||
install-tests: $(foreach distro_release, $(SUPPORTED), $(distro_release)-install-test)
|
||||
|
||||
|
||||
@ -95,28 +148,33 @@ debian-%-install-test:
|
||||
docker run --rm \
|
||||
-v ./$(BUILD_DIR):/output \
|
||||
debian:$* \
|
||||
bash -c 'apt-get update && apt-get install -y /output/$(PACKAGE_NAME)-$(VERSION)-debian-$*.deb'
|
||||
bash -c 'apt-get update && apt-get install -y /output/$(PACKAGE_NAME)-$(DEB_VERSION)-debian-$*.deb'
|
||||
|
||||
# Run a RedHat install test
|
||||
rockylinux-%-install-test:
|
||||
docker run --rm \
|
||||
-v ./$(BUILD_DIR):/output \
|
||||
rockylinux:$* \
|
||||
bash -c 'dnf makecache && dnf install -y /output/$(PACKAGE_NAME)-$(VERSION)-1.el$*.noarch.rpm'
|
||||
bash -c 'dnf makecache && dnf install -y /output/$(PACKAGE_NAME)-$(RPM_VERSION).el$*.noarch.rpm'
|
||||
|
||||
# Run an Ubuntu install test
|
||||
ubuntu-%-install-test:
|
||||
docker run --rm \
|
||||
-v ./$(BUILD_DIR):/output \
|
||||
ubuntu:$* \
|
||||
bash -c 'apt-get update && apt-get install -y /output/$(PACKAGE_NAME)-$(VERSION)-ubuntu-$*.deb'
|
||||
bash -c 'apt-get update && apt-get install -y /output/$(PACKAGE_NAME)-$(DEB_VERSION)-ubuntu-$*.deb'
|
||||
|
||||
# Run an OracleLinux install test (this is for EL7 since CentOS7 images no longer exist)
|
||||
oraclelinux-%-install-test:
|
||||
docker run --rm \
|
||||
-v ./$(BUILD_DIR):/output \
|
||||
oraclelinux:7 \
|
||||
bash -c 'yum makecache && yum install -y /output/$(PACKAGE_NAME)-$(VERSION)-1.el7.noarch.rpm'
|
||||
bash -c 'yum makecache && yum install -y /output/$(PACKAGE_NAME)-$(RPM_VERSION).el7.noarch.rpm'
|
||||
|
||||
# Run a Gentoo install test
|
||||
gentoo-install-test:
|
||||
# May impliment this in the future, but would require additional headaches to set up a repo
|
||||
true
|
||||
|
||||
##
|
||||
# Container targets
|
||||
@ -151,30 +209,30 @@ package-image-%:
|
||||
|
||||
# Debian package creation
|
||||
actually-package-debian-%:
|
||||
$(MAKE) install DESTDIR=/output/debian-$*
|
||||
$(MAKE) install-systemd DESTDIR=/output/debian-$*
|
||||
cp -r --preserve=mode DEBIAN /output/debian-$*/
|
||||
dpkg-deb -Zgzip --build /output/debian-$* "/output/$(PACKAGE_NAME)-$(VERSION)-debian-$*.deb"
|
||||
dpkg-deb -Zgzip --build /output/debian-$* "/output/$(PACKAGE_NAME)-$(DEB_VERSION)-debian-$*.deb"
|
||||
|
||||
# RedHat package creation
|
||||
actually-package-rockylinux-%:
|
||||
mkdir -p /output/rockylinux-$*/{BUILD,RPMS,SOURCES,SPECS,SRPMS}
|
||||
cp RPM/$(PACKAGE_NAME).spec /output/rockylinux-$*/SPECS/
|
||||
sed -e "s/@@VERSION@@/$(VERSION)/g" -e "s/@@RELEASE@@/$(RPM_RELEASE)/g" RPM/$(PACKAGE_NAME).spec > /output/rockylinux-$*/SPECS/$(PACKAGE_NAME).spec
|
||||
rpmbuild --define '_topdir /output/rockylinux-$*' \
|
||||
--define 'version $(VERSION)' \
|
||||
--define 'version $(RPM_VERSION)' \
|
||||
-bb /output/rockylinux-$*/SPECS/$(PACKAGE_NAME).spec
|
||||
cp /output/rockylinux-$*/RPMS/noarch/$(PACKAGE_NAME)-$(VERSION)-1.el$*.noarch.rpm /output/
|
||||
cp /output/rockylinux-$*/RPMS/noarch/$(PACKAGE_NAME)-$(RPM_VERSION).el$*.noarch.rpm /output/
|
||||
|
||||
# Ubuntu package creation
|
||||
actually-package-ubuntu-%:
|
||||
$(MAKE) install DESTDIR=/output/ubuntu-$*
|
||||
$(MAKE) install-systemd DESTDIR=/output/ubuntu-$*
|
||||
cp -r --preserve=mode DEBIAN /output/ubuntu-$*/
|
||||
dpkg-deb -Zgzip --build /output/ubuntu-$* "/output/$(PACKAGE_NAME)-$(VERSION)-ubuntu-$*.deb"
|
||||
dpkg-deb -Zgzip --build /output/ubuntu-$* "/output/$(PACKAGE_NAME)-$(DEB_VERSION)-ubuntu-$*.deb"
|
||||
|
||||
# OracleLinux package creation
|
||||
actually-package-oraclelinux-%:
|
||||
mkdir -p /output/oraclelinux-$*/{BUILD,RPMS,SOURCES,SPECS,SRPMS}
|
||||
cp RPM/$(PACKAGE_NAME)-el7.spec /output/oraclelinux-$*/SPECS/$(PACKAGE_NAME).spec
|
||||
sed -e "s/@@VERSION@@/$(VERSION)/g" -e "s/@@RELEASE@@/$(RPM_RELEASE)/g" RPM/$(PACKAGE_NAME)-el7.spec > /output/oraclelinux-$*/SPECS/$(PACKAGE_NAME).spec
|
||||
rpmbuild --define '_topdir /output/oraclelinux-$*' \
|
||||
--define 'version $(VERSION)' \
|
||||
--define 'version $(RPM_VERSION)' \
|
||||
-bb /output/oraclelinux-$*/SPECS/$(PACKAGE_NAME).spec
|
||||
cp /output/oraclelinux-$*/RPMS/noarch/$(PACKAGE_NAME)-$(VERSION)-1.el$*.noarch.rpm /output/
|
||||
cp /output/oraclelinux-$*/RPMS/noarch/$(PACKAGE_NAME)-$(RPM_VERSION).el$*.noarch.rpm /output/
|
||||
|
||||
@ -1,13 +1,13 @@
|
||||
Name: pgmon
|
||||
Version: 1.0
|
||||
Release: 1%{?dist}
|
||||
Version: @@VERSION@@
|
||||
Release: @@RELEASE@@%{?dist}
|
||||
Summary: A bridge to sit between monitoring tools and PostgreSQL
|
||||
|
||||
License: MIT
|
||||
URL: https://www.commandprompt.com
|
||||
|
||||
BuildArch: noarch
|
||||
Requires: logrotate, python, python-psycopg2, PyYAML, systemd
|
||||
Requires: logrotate, python, python-psycopg2, PyYAML, python-requests, systemd
|
||||
|
||||
%description
|
||||
A bridge to sit between monitoring tools and PostgreSQL
|
||||
@ -19,7 +19,7 @@ A bridge to sit between monitoring tools and PostgreSQL
|
||||
# Do nothing since we have nothing to build
|
||||
|
||||
%install
|
||||
make -C /src install DESTDIR=%{buildroot}
|
||||
make -C /src install-systemd DESTDIR=%{buildroot}
|
||||
|
||||
%files
|
||||
/etc/logrotate.d/pgmon
|
||||
@ -28,7 +28,7 @@ make -C /src install DESTDIR=%{buildroot}
|
||||
/etc/pgmon/pgmon-service.conf
|
||||
/lib/systemd/system/pgmon.service
|
||||
/lib/systemd/system/pgmon@.service
|
||||
/usr/local/bin/pgmon
|
||||
/usr/bin/pgmon
|
||||
/usr/share/man/man1/pgmon.1.gz
|
||||
|
||||
%post
|
||||
|
||||
@ -1,13 +1,13 @@
|
||||
Name: pgmon
|
||||
Version: 1.0
|
||||
Release: 1%{?dist}
|
||||
Version: @@VERSION@@
|
||||
Release: @@RELEASE@@%{?dist}
|
||||
Summary: A bridge to sit between monitoring tools and PostgreSQL
|
||||
|
||||
License: MIT
|
||||
URL: https://www.commandprompt.com
|
||||
|
||||
BuildArch: noarch
|
||||
Requires: logrotate, python3, python3-psycopg2, python3-pyyaml, systemd
|
||||
Requires: logrotate, python3, python3-psycopg2, python3-pyyaml, python3-requests, systemd
|
||||
|
||||
%description
|
||||
A bridge to sit between monitoring tools and PostgreSQL
|
||||
@ -19,7 +19,7 @@ A bridge to sit between monitoring tools and PostgreSQL
|
||||
# Do nothing since we have nothing to build
|
||||
|
||||
%install
|
||||
make -C /src install DESTDIR=%{buildroot}
|
||||
make -C /src install-systemd DESTDIR=%{buildroot}
|
||||
|
||||
%files
|
||||
/etc/logrotate.d/pgmon
|
||||
@ -28,7 +28,7 @@ make -C /src install DESTDIR=%{buildroot}
|
||||
/etc/pgmon/pgmon-service.conf
|
||||
/lib/systemd/system/pgmon.service
|
||||
/lib/systemd/system/pgmon@.service
|
||||
/usr/local/bin/pgmon
|
||||
/usr/bin/pgmon
|
||||
/usr/share/man/man1/pgmon.1.gz
|
||||
|
||||
%post
|
||||
|
||||
@ -11,7 +11,14 @@ PGMON_USER="${PGMON_USER:-postgres}"
|
||||
PGMON_GROUP="${PGMON_GROUP:-$PGMON_USER}"
|
||||
CONFIG_FILE="/etc/pgmon/${agent_name}.yml"
|
||||
|
||||
output_log=/var/log/pgmon/${SVCNAME}.log
|
||||
error_log=/var/log/pgmon/${SVCNAME}.err
|
||||
|
||||
start_pre() {
|
||||
checkpath -f -m 0644 -o "${PGMON_USER}:${PGMON_GROUP}" "${output_log}" "${error_log}"
|
||||
}
|
||||
|
||||
command="/usr/bin/pgmon"
|
||||
command_args="'$CONFIG_FILE'"
|
||||
command_args="-c '$CONFIG_FILE'"
|
||||
command_background="true"
|
||||
command_user="${PGMON_USER}:${PGMON_GROUP}"
|
||||
|
||||
4
requirements-dev.yml
Normal file
4
requirements-dev.yml
Normal file
@ -0,0 +1,4 @@
|
||||
-r requirements.txt
|
||||
testcontainers[postgresql]
|
||||
pytest
|
||||
black
|
||||
@ -1,45 +1,307 @@
|
||||
metrics:
|
||||
##
|
||||
# Discovery metrics
|
||||
##
|
||||
discover_dbs:
|
||||
type: set
|
||||
query:
|
||||
0: SELECT datname AS dbname FROM pg_database
|
||||
0: >
|
||||
SELECT datname AS dbname
|
||||
FROM pg_database
|
||||
|
||||
# Note: If the user lacks sufficient privileges, these fields will be NULL.
|
||||
# The WHERE clause is intended to prevent Zabbix from discovering a
|
||||
# connection it cannot monitor. Ideally this would generate an error
|
||||
# instead.
|
||||
discover_rep:
|
||||
type: set
|
||||
query:
|
||||
0: SELECT client_addr || '_' || regexp_replace(application_name, '[ ,]', '_', 'g') AS repid, client_addr, state FROM pg_stat_replication
|
||||
0: >
|
||||
SELECT host(client_addr) || '_' || regexp_replace(application_name, '[ ,]', '_', 'g') AS repid,
|
||||
client_addr,
|
||||
state
|
||||
FROM pg_stat_replication
|
||||
WHERE state IS NOT NULL
|
||||
|
||||
discover_slots:
|
||||
type: set
|
||||
query:
|
||||
90400: >
|
||||
SELECT slot_name,
|
||||
plugin,
|
||||
slot_type,
|
||||
database,
|
||||
false as temporary,
|
||||
active
|
||||
FROM pg_replication_slots
|
||||
100000: >
|
||||
SELECT slot_name,
|
||||
plugin,
|
||||
slot_type,
|
||||
database,
|
||||
temporary,
|
||||
active
|
||||
FROM pg_replication_slots
|
||||
|
||||
|
||||
##
|
||||
# cluster-wide metrics
|
||||
##
|
||||
version:
|
||||
type: value
|
||||
query:
|
||||
0: SHOW server_version_num
|
||||
max_frozen_age:
|
||||
type: value
|
||||
query:
|
||||
0: SELECT max(age(datfrozenxid)) FROM pg_database
|
||||
|
||||
max_frozen_age:
|
||||
type: row
|
||||
query:
|
||||
0: >
|
||||
SELECT max(age(datfrozenxid)) AS xid_age,
|
||||
NULL AS mxid_age
|
||||
FROM pg_database
|
||||
90600: >
|
||||
SELECT max(age(datfrozenxid)) AS xid_age,
|
||||
max(mxid_age(datminmxid)) AS mxid_age
|
||||
FROM pg_database
|
||||
|
||||
bgwriter:
|
||||
type: row
|
||||
query:
|
||||
0: >
|
||||
SELECT checkpoints_timed,
|
||||
checkpoints_req,
|
||||
checkpoint_write_time,
|
||||
checkpoint_sync_time,
|
||||
buffers_checkpoint,
|
||||
buffers_clean,
|
||||
maxwritten_clean,
|
||||
buffers_backend,
|
||||
buffers_backend_fsync,
|
||||
buffers_alloc
|
||||
FROM pg_stat_bgwriter
|
||||
170000: >
|
||||
SELECT cp.num_timed AS checkpoints_timed,
|
||||
cp.num_requested AS checkpoints_req,
|
||||
cp.write_time AS checkpoint_write_time,
|
||||
cp.sync_time AS checkpoint_sync_time,
|
||||
cp.buffers_written AS buffers_checkpoint,
|
||||
bg.buffers_clean AS buffers_clean,
|
||||
bg.maxwritten_clean AS maxwritten_clean,
|
||||
NULL AS buffers_backend,
|
||||
NULL AS buffers_backend_fsync,
|
||||
bg.buffers_alloc AS buffers_alloc
|
||||
FROM pg_stat_bgwriter bg
|
||||
CROSS JOIN pg_stat_checkpointer cp
|
||||
|
||||
io_per_backend:
|
||||
type: set
|
||||
query:
|
||||
160000: >
|
||||
SELECT backend_type,
|
||||
COALESCE(SUM(reads * op_bytes), 0) AS reads,
|
||||
COALESCE(SUM(read_time), 0) AS read_time,
|
||||
COALESCE(SUM(writes * op_bytes), 0) AS writes,
|
||||
COALESCE(SUM(write_time), 0) AS write_time,
|
||||
COALESCE(SUM(writebacks * op_bytes), 0) AS writebacks,
|
||||
COALESCE(SUM(writeback_time), 0) AS writeback_time,
|
||||
COALESCE(SUM(extends * op_bytes), 0) AS extends,
|
||||
COALESCE(SUM(extend_time), 0) AS extend_time,
|
||||
COALESCE(SUM(op_bytes), 0) AS op_bytes,
|
||||
COALESCE(SUM(hits), 0) AS hits,
|
||||
COALESCE(SUM(evictions), 0) AS evictions,
|
||||
COALESCE(SUM(reuses), 0) AS reuses,
|
||||
COALESCE(SUM(fsyncs), 0) AS fsyncs,
|
||||
COALESCE(SUM(fsync_time), 0) AS fsync_time
|
||||
FROM pg_stat_io
|
||||
GROUP BY backend_type
|
||||
|
||||
|
||||
##
|
||||
# Per-database metrics
|
||||
##
|
||||
db_stats:
|
||||
type: row
|
||||
query:
|
||||
0: SELECT numbackends, xact_commit, xact_rollback, blks_read, blks_hit, tup_returned, tup_fetched, tup_inserted, tup_updated, tup_deleted, conflicts, temp_files, temp_bytes, deadlocks, blk_read_time, blk_write_time, extract('epoch' from stats_reset)::float FROM pg_stat_database WHERE datname = %(dbname)s
|
||||
140000: SELECT numbackends, xact_commit, xact_rollback, blks_read, blks_hit, tup_returned, tup_fetched, tup_inserted, tup_updated, tup_deleted, conflicts, temp_files, temp_bytes, deadlocks, COALESCE(checksum_failures, 0) AS checksum_failures, blk_read_time, blk_write_time, session_time, active_time, idle_in_transaction_time, sessions, sessions_abandoned, sessions_fatal, sessions_killed, extract('epoch' from stats_reset)::float FROM pg_stat_database WHERE datname = %(dbname)s
|
||||
0: >
|
||||
SELECT numbackends,
|
||||
xact_commit,
|
||||
xact_rollback,
|
||||
blks_read,
|
||||
blks_hit,
|
||||
tup_returned,
|
||||
tup_fetched,
|
||||
tup_inserted,
|
||||
tup_updated,
|
||||
tup_deleted,
|
||||
conflicts,
|
||||
temp_files,
|
||||
temp_bytes,
|
||||
deadlocks,
|
||||
NULL AS checksum_failures,
|
||||
blk_read_time,
|
||||
blk_write_time,
|
||||
NULL AS session_time,
|
||||
NULL AS active_time,
|
||||
NULL AS idle_in_transaction_time,
|
||||
NULL AS sessions,
|
||||
NULL AS sessions_abandoned,
|
||||
NULL AS sessions_fatal,
|
||||
NULL AS sessions_killed,
|
||||
extract('epoch' from stats_reset) AS stats_reset
|
||||
FROM pg_stat_database WHERE datname = %(dbname)s
|
||||
140000: >
|
||||
SELECT numbackends,
|
||||
xact_commit,
|
||||
xact_rollback,
|
||||
blks_read,
|
||||
blks_hit,
|
||||
tup_returned,
|
||||
tup_fetched,
|
||||
tup_inserted,
|
||||
tup_updated,
|
||||
tup_deleted,
|
||||
conflicts,
|
||||
temp_files,
|
||||
temp_bytes,
|
||||
deadlocks,
|
||||
COALESCE(checksum_failures, 0) AS checksum_failures,
|
||||
blk_read_time,
|
||||
blk_write_time,
|
||||
session_time,
|
||||
active_time,
|
||||
idle_in_transaction_time,
|
||||
sessions,
|
||||
sessions_abandoned,
|
||||
sessions_fatal,
|
||||
sessions_killed,
|
||||
extract('epoch' from stats_reset) AS stats_reset
|
||||
FROM pg_stat_database WHERE datname = %(dbname)s
|
||||
test_args:
|
||||
dbname: postgres
|
||||
|
||||
hit_ratios:
|
||||
type: row
|
||||
query:
|
||||
0: >
|
||||
SELECT sum(heap_blks_read)::float / NULLIF(sum(heap_blks_read + heap_blks_hit), 0) AS avg_heap_hit_ratio,
|
||||
sum(idx_blks_hit)::float / NULLIF(sum(idx_blks_read + idx_blks_hit), 0) AS avg_idx_hit_ratio,
|
||||
sum(toast_blks_hit)::float / NULLIF(sum(toast_blks_read + toast_blks_hit), 0) AS avg_toast_hit_ratio,
|
||||
sum(tidx_blks_hit)::float / NULLIF(sum(tidx_blks_read + tidx_blks_hit), 0) AS avg_tidx_hit_ratio
|
||||
FROM pg_statio_all_tables
|
||||
test_args:
|
||||
dbname: postgres
|
||||
|
||||
activity:
|
||||
type: set
|
||||
query:
|
||||
0: >
|
||||
SELECT state,
|
||||
count(*) AS backend_count,
|
||||
COALESCE(EXTRACT(EPOCH FROM max(now() - state_change)), 0) AS max_state_time
|
||||
FROM pg_stat_activity
|
||||
WHERE datname = %(dbname)s
|
||||
GROUP BY state
|
||||
test_args:
|
||||
dbname: postgres
|
||||
|
||||
sequence_usage:
|
||||
type: value
|
||||
query:
|
||||
# 9.2 lacks lateral joins, the pg_sequence_last_value function, and the pg_sequences view
|
||||
# 0: >
|
||||
# SELECT COALESCE(MAX(pg_sequence_last_value(c.oid)::float / (pg_sequence_parameters(oid)).maximum_value), 0) AS max_usage
|
||||
# FROM pg_class c
|
||||
# WHERE c.relkind = 'S'
|
||||
# 9.3 - 9.6 lacks the pg_sequence_last_value function, and pg_sequences view
|
||||
# 90300: >
|
||||
# SELECT COALESCE(MAX(pg_sequence_last_value(c.oid)::float / s.maximum_value), 0) AS max_usage
|
||||
# FROM pg_class c
|
||||
# CROSS JOIN LATERAL pg_sequence_parameters(c.oid) AS s
|
||||
# WHERE c.relkind = 'S'
|
||||
100000: SELECT COALESCE(MAX(last_value::float / max_value), 0) AS max_usage FROM pg_sequences;
|
||||
test_args:
|
||||
dbname: postgres
|
||||
|
||||
sequence_visibility:
|
||||
type: row
|
||||
query:
|
||||
100000: >
|
||||
SELECT COUNT(*) FILTER (WHERE has_sequence_privilege(c.oid, 'SELECT,USAGE')) AS visible_sequences,
|
||||
COUNT(*) AS total_sequences
|
||||
FROM pg_class AS c
|
||||
WHERE relkind = 'S';
|
||||
|
||||
|
||||
##
|
||||
# Per-replication metrics
|
||||
##
|
||||
rep_stats:
|
||||
type: row
|
||||
query:
|
||||
90400: >
|
||||
SELECT pid, usename,
|
||||
EXTRACT(EPOCH FROM backend_start) AS backend_start,
|
||||
state,
|
||||
pg_xlog_location_diff(pg_current_xlog_location(), sent_location) AS sent_lsn,
|
||||
pg_xlog_location_diff(pg_current_xlog_location(), write_location) AS write_lsn,
|
||||
pg_xlog_location_diff(pg_current_xlog_location(), flush_location) AS flush_lsn,
|
||||
pg_xlog_location_diff(pg_current_xlog_location(), replay_location) AS replay_lsn,
|
||||
NULL AS write_lag,
|
||||
NULL AS flush_lag,
|
||||
NULL AS replay_lag,
|
||||
sync_state
|
||||
FROM pg_stat_replication
|
||||
WHERE host(client_addr) || '_' || regexp_replace(application_name, '[ ,]', '_', 'g') = %(repid)s
|
||||
100000: >
|
||||
SELECT pid, usename,
|
||||
EXTRACT(EPOCH FROM backend_start) AS backend_start,
|
||||
state,
|
||||
pg_wal_lsn_diff(pg_current_wal_lsn(), sent_lsn) AS sent_lsn,
|
||||
pg_wal_lsn_diff(pg_current_wal_lsn(), write_lsn) AS write_lsn,
|
||||
pg_wal_lsn_diff(pg_current_wal_lsn(), flush_lsn) AS flush_lsn,
|
||||
pg_wal_lsn_diff(pg_current_wal_lsn(), replay_lsn) AS replay_lsn,
|
||||
COALESCE(EXTRACT(EPOCH FROM write_lag), 0) AS write_lag,
|
||||
COALESCE(EXTRACT(EPOCH FROM flush_lag), 0) AS flush_lag,
|
||||
COALESCE(EXTRACT(EPOCH FROM replay_lag), 0) AS replay_lag,
|
||||
sync_state
|
||||
FROM pg_stat_replication
|
||||
WHERE host(client_addr) || '_' || regexp_replace(application_name, '[ ,]', '_', 'g') = %(repid)s
|
||||
test_args:
|
||||
repid: 127.0.0.1_test_rep
|
||||
|
||||
|
||||
##
|
||||
# Per-slot metrics
|
||||
##
|
||||
slot_stats:
|
||||
type: row
|
||||
query:
|
||||
90400: >
|
||||
SELECT NULL as active_pid,
|
||||
xmin,
|
||||
pg_xlog_location_diff(pg_current_xlog_location(), restart_lsn) AS restart_bytes,
|
||||
NULL AS confirmed_flush_bytes
|
||||
FROM pg_replication_slots WHERE slot_name = %(slot)s
|
||||
90600: >
|
||||
SELECT active_pid,
|
||||
xmin,
|
||||
pg_xlog_location_diff(pg_current_xlog_location(), restart_lsn) AS restart_bytes,
|
||||
pg_xlog_location_diff(pg_current_xlog_location(), confirmed_flush_lsn) AS confirmed_flush_bytes
|
||||
FROM pg_replication_slots WHERE slot_name = %(slot)s
|
||||
100000: >
|
||||
SELECT active_pid,
|
||||
xmin,
|
||||
pg_wal_lsn_diff(pg_current_wal_lsn(), restart_lsn) AS restart_bytes,
|
||||
pg_wal_lsn_diff(pg_current_wal_lsn(), confirmed_flush_lsn) AS confirmed_flush_bytes
|
||||
FROM pg_replication_slots WHERE slot_name = %(slot)s
|
||||
test_args:
|
||||
slot: test_slot
|
||||
|
||||
|
||||
##
|
||||
# Debugging
|
||||
##
|
||||
ntables:
|
||||
type: value
|
||||
query:
|
||||
0: SELECT count(*) AS ntables FROM pg_stat_user_tables
|
||||
|
||||
# Per-replication metrics
|
||||
rep_stats:
|
||||
type: row
|
||||
query:
|
||||
0: SELECT * FROM pg_stat_database WHERE client_addr || '_' || regexp_replace(application_name, '[ ,]', '_', 'g') = '{repid}'
|
||||
|
||||
# Debugging
|
||||
sleep:
|
||||
type: value
|
||||
query:
|
||||
0: SELECT now(), pg_sleep(5);
|
||||
|
||||
@ -1,3 +1,9 @@
|
||||
# The address the agent binds to
|
||||
#address: 127.0.0.1
|
||||
|
||||
# The port the agent listens on for requests
|
||||
#port: 5400
|
||||
|
||||
# Min PostgreSQL connection pool size (per database)
|
||||
#min_pool_size: 0
|
||||
|
||||
@ -23,6 +29,9 @@
|
||||
# Default database to connect to when none is specified for a metric
|
||||
#dbname: 'postgres'
|
||||
|
||||
# SSL connection mode
|
||||
#ssl_mode: require
|
||||
|
||||
# Timeout for getting a connection slot from a pool
|
||||
#pool_slot_timeout: 5
|
||||
|
||||
|
||||
587
src/pgmon.py
587
src/pgmon.py
@ -4,6 +4,7 @@ import yaml
|
||||
import json
|
||||
import time
|
||||
import os
|
||||
import sys
|
||||
|
||||
import argparse
|
||||
import logging
|
||||
@ -11,7 +12,7 @@ import logging
|
||||
from datetime import datetime, timedelta
|
||||
|
||||
import psycopg2
|
||||
from psycopg2.extras import DictCursor
|
||||
from psycopg2.extras import RealDictCursor
|
||||
from psycopg2.pool import ThreadedConnectionPool
|
||||
|
||||
from contextlib import contextmanager
|
||||
@ -23,7 +24,12 @@ from http.server import BaseHTTPRequestHandler, HTTPServer
|
||||
from http.server import ThreadingHTTPServer
|
||||
from urllib.parse import urlparse, parse_qs
|
||||
|
||||
VERSION = '0.1.0'
|
||||
import requests
|
||||
import re
|
||||
|
||||
from decimal import Decimal
|
||||
|
||||
VERSION = "1.0.4"
|
||||
|
||||
# Configuration
|
||||
config = {}
|
||||
@ -42,6 +48,12 @@ cluster_version = None
|
||||
cluster_version_next_check = None
|
||||
cluster_version_lock = Lock()
|
||||
|
||||
# PostgreSQL latest version information
|
||||
latest_version = None
|
||||
latest_version_next_check = None
|
||||
latest_version_lock = Lock()
|
||||
release_supported = None
|
||||
|
||||
# Running state (used to gracefully shut down)
|
||||
running = True
|
||||
|
||||
@ -53,64 +65,79 @@ config_file = None
|
||||
|
||||
# Configure logging
|
||||
log = logging.getLogger(__name__)
|
||||
formatter = logging.Formatter('%(asctime)s - %(levelname)s - %(filename)s: %(funcName)s() line %(lineno)d: %(message)s')
|
||||
formatter = logging.Formatter(
|
||||
"%(asctime)s - %(levelname)s - %(filename)s: %(funcName)s() line %(lineno)d: %(message)s"
|
||||
)
|
||||
console_log_handler = logging.StreamHandler()
|
||||
console_log_handler.setFormatter(formatter)
|
||||
log.addHandler(console_log_handler)
|
||||
|
||||
|
||||
# Error types
|
||||
class ConfigError(Exception):
|
||||
pass
|
||||
|
||||
|
||||
class DisconnectedError(Exception):
|
||||
pass
|
||||
|
||||
|
||||
class UnhappyDBError(Exception):
|
||||
pass
|
||||
|
||||
|
||||
class UnknownMetricError(Exception):
|
||||
pass
|
||||
|
||||
|
||||
class MetricVersionError(Exception):
|
||||
pass
|
||||
|
||||
|
||||
class LatestVersionCheckError(Exception):
|
||||
pass
|
||||
|
||||
|
||||
# Default config settings
|
||||
default_config = {
|
||||
# The address the agent binds to
|
||||
"address": "127.0.0.1",
|
||||
# The port the agent listens on for requests
|
||||
"port": 5400,
|
||||
# Min PostgreSQL connection pool size (per database)
|
||||
'min_pool_size': 0,
|
||||
|
||||
"min_pool_size": 0,
|
||||
# Max PostgreSQL connection pool size (per database)
|
||||
'max_pool_size': 4,
|
||||
|
||||
"max_pool_size": 4,
|
||||
# How long a connection can sit idle in the pool before it's removed (seconds)
|
||||
'max_idle_time': 30,
|
||||
|
||||
"max_idle_time": 30,
|
||||
# Log level for stderr logging
|
||||
'log_level': 'error',
|
||||
|
||||
"log_level": "error",
|
||||
# Database user to connect as
|
||||
'dbuser': 'postgres',
|
||||
|
||||
"dbuser": "postgres",
|
||||
# Database host
|
||||
'dbhost': '/var/run/postgresql',
|
||||
|
||||
"dbhost": "/var/run/postgresql",
|
||||
# Database port
|
||||
'dbport': 5432,
|
||||
|
||||
"dbport": 5432,
|
||||
# Default database to connect to when none is specified for a metric
|
||||
'dbname': 'postgres',
|
||||
|
||||
"dbname": "postgres",
|
||||
# SSL connection mode
|
||||
"ssl_mode": "require",
|
||||
# Timeout for getting a connection slot from a pool
|
||||
'pool_slot_timeout': 5,
|
||||
|
||||
"pool_slot_timeout": 5,
|
||||
# PostgreSQL connection timeout (seconds)
|
||||
# Note: It can actually be double this because of retries
|
||||
'connect_timeout': 5,
|
||||
|
||||
"connect_timeout": 5,
|
||||
# Time to wait before trying to reconnect again after a reconnect failure (seconds)
|
||||
'reconnect_cooldown': 30,
|
||||
|
||||
"reconnect_cooldown": 30,
|
||||
# How often to check the version of PostgreSQL (seconds)
|
||||
'version_check_period': 300,
|
||||
|
||||
"version_check_period": 300,
|
||||
# How often to check the latest supported version of PostgreSQL (seconds)
|
||||
"latest_version_check_period": 86400,
|
||||
# Metrics
|
||||
'metrics': {}
|
||||
"metrics": {},
|
||||
}
|
||||
|
||||
|
||||
def update_deep(d1, d2):
|
||||
"""
|
||||
Recursively update a dict, adding keys to dictionaries and appending to
|
||||
@ -124,24 +151,33 @@ def update_deep(d1, d2):
|
||||
The new d1
|
||||
"""
|
||||
if not isinstance(d1, dict) or not isinstance(d2, dict):
|
||||
raise TypeError('Both arguments to update_deep need to be dictionaries')
|
||||
raise TypeError("Both arguments to update_deep need to be dictionaries")
|
||||
|
||||
for k, v2 in d2.items():
|
||||
if isinstance(v2, dict):
|
||||
v1 = d1.get(k, {})
|
||||
if not isinstance(v1, dict):
|
||||
raise TypeError('Type mismatch between dictionaries: {} is not a dict'.format(type(v1).__name__))
|
||||
raise TypeError(
|
||||
"Type mismatch between dictionaries: {} is not a dict".format(
|
||||
type(v1).__name__
|
||||
)
|
||||
)
|
||||
d1[k] = update_deep(v1, v2)
|
||||
elif isinstance(v2, list):
|
||||
v1 = d1.get(k, [])
|
||||
if not isinstance(v1, list):
|
||||
raise TypeError('Type mismatch between dictionaries: {} is not a list'.format(type(v1).__name__))
|
||||
raise TypeError(
|
||||
"Type mismatch between dictionaries: {} is not a list".format(
|
||||
type(v1).__name__
|
||||
)
|
||||
)
|
||||
d1[k] = v1 + v2
|
||||
else:
|
||||
d1[k] = v2
|
||||
return d1
|
||||
|
||||
def read_config(path, included = False):
|
||||
|
||||
def read_config(path, included=False):
|
||||
"""
|
||||
Read a config file.
|
||||
|
||||
@ -151,7 +187,7 @@ def read_config(path, included = False):
|
||||
"""
|
||||
# Read config file
|
||||
log.info("Reading log file: {}".format(path))
|
||||
with open(path, 'r') as f:
|
||||
with open(path, "r") as f:
|
||||
try:
|
||||
cfg = yaml.safe_load(f)
|
||||
except yaml.parser.ParserError as e:
|
||||
@ -161,42 +197,53 @@ def read_config(path, included = False):
|
||||
config_base = os.path.dirname(path)
|
||||
|
||||
# Read any external queries and validate metric definitions
|
||||
for name, metric in cfg.get('metrics', {}).items():
|
||||
for name, metric in cfg.get("metrics", {}).items():
|
||||
# Validate return types
|
||||
try:
|
||||
if metric['type'] not in ['value', 'row', 'column', 'set']:
|
||||
raise ConfigError("Invalid return type: {} for metric {} in {}".format(metric['type'], name, path))
|
||||
if metric["type"] not in ["value", "row", "column", "set"]:
|
||||
raise ConfigError(
|
||||
"Invalid return type: {} for metric {} in {}".format(
|
||||
metric["type"], name, path
|
||||
)
|
||||
)
|
||||
except KeyError:
|
||||
raise ConfigError("No type specified for metric {} in {}".format(name, path))
|
||||
raise ConfigError(
|
||||
"No type specified for metric {} in {}".format(name, path)
|
||||
)
|
||||
|
||||
# Ensure queries exist
|
||||
query_dict = metric.get('query', {})
|
||||
query_dict = metric.get("query", {})
|
||||
if type(query_dict) is not dict:
|
||||
raise ConfigError("Query definition should be a dictionary, got: {} for metric {} in {}".format(query_dict, name, path))
|
||||
raise ConfigError(
|
||||
"Query definition should be a dictionary, got: {} for metric {} in {}".format(
|
||||
query_dict, name, path
|
||||
)
|
||||
)
|
||||
|
||||
if len(query_dict) == 0:
|
||||
raise ConfigError("Missing queries for metric {} in {}".format(name, path))
|
||||
|
||||
# Read external sql files and validate version keys
|
||||
for vers, query in metric['query'].items():
|
||||
for vers, query in metric["query"].items():
|
||||
try:
|
||||
int(vers)
|
||||
except:
|
||||
raise ConfigError("Invalid version: {} for metric {} in {}".format(vers, name, path))
|
||||
raise ConfigError(
|
||||
"Invalid version: {} for metric {} in {}".format(vers, name, path)
|
||||
)
|
||||
|
||||
if query.startswith('file:'):
|
||||
if query.startswith("file:"):
|
||||
query_path = query[5:]
|
||||
if not query_path.startswith('/'):
|
||||
query_path = os.path.join(config_base, query_path)
|
||||
with open(query_path, 'r') as f:
|
||||
metric['query'][vers] = f.read()
|
||||
|
||||
if not query_path.startswith("/"):
|
||||
query_path = os.path.join(config_base, query_path)
|
||||
with open(query_path, "r") as f:
|
||||
metric["query"][vers] = f.read()
|
||||
|
||||
# Read any included config files
|
||||
for inc in cfg.get('include', []):
|
||||
for inc in cfg.get("include", []):
|
||||
# Prefix relative paths with the directory from the current config
|
||||
if not inc.startswith('/'):
|
||||
inc = os.path.join(config_base, inc)
|
||||
if not inc.startswith("/"):
|
||||
inc = os.path.join(config_base, inc)
|
||||
update_deep(cfg, read_config(inc, included=True))
|
||||
|
||||
# Return the config we read if this is an include, otherwise set the final
|
||||
@ -209,19 +256,26 @@ def read_config(path, included = False):
|
||||
update_deep(new_config, cfg)
|
||||
|
||||
# Minor sanity checks
|
||||
if len(new_config['metrics']) == 0:
|
||||
if len(new_config["metrics"]) == 0:
|
||||
log.error("No metrics are defined")
|
||||
raise ConfigError("No metrics defined")
|
||||
|
||||
# Validate the new log level before changing the config
|
||||
if new_config['log_level'].upper() not in ['DEBUG', 'INFO', 'WARNING', 'ERROR', 'CRITICAL']:
|
||||
raise ConfigError("Invalid log level: {}".format(new_config['log_level']))
|
||||
if new_config["log_level"].upper() not in [
|
||||
"DEBUG",
|
||||
"INFO",
|
||||
"WARNING",
|
||||
"ERROR",
|
||||
"CRITICAL",
|
||||
]:
|
||||
raise ConfigError("Invalid log level: {}".format(new_config["log_level"]))
|
||||
|
||||
global config
|
||||
config = new_config
|
||||
|
||||
# Apply changes to log level
|
||||
log.setLevel(logging.getLevelName(config['log_level'].upper()))
|
||||
log.setLevel(logging.getLevelName(config["log_level"].upper()))
|
||||
|
||||
|
||||
def signal_handler(sig, frame):
|
||||
"""
|
||||
@ -233,7 +287,7 @@ def signal_handler(sig, frame):
|
||||
signal.signal(signal.SIGINT, signal.default_int_handler)
|
||||
|
||||
# Signal everything to shut down
|
||||
if sig in [ signal.SIGINT, signal.SIGTERM, signal.SIGQUIT ]:
|
||||
if sig in [signal.SIGINT, signal.SIGTERM, signal.SIGQUIT]:
|
||||
log.info("Shutting down ...")
|
||||
global running
|
||||
running = False
|
||||
@ -245,11 +299,12 @@ def signal_handler(sig, frame):
|
||||
log.warning("Received config reload signal")
|
||||
read_config(config_file)
|
||||
|
||||
|
||||
class ConnectionPool(ThreadedConnectionPool):
|
||||
def __init__(self, dbname, minconn, maxconn, *args, **kwargs):
|
||||
# Make sure dbname isn't different in the kwargs
|
||||
kwargs['dbname'] = dbname
|
||||
|
||||
kwargs["dbname"] = dbname
|
||||
|
||||
super().__init__(minconn, maxconn, *args, **kwargs)
|
||||
self.name = dbname
|
||||
|
||||
@ -270,7 +325,10 @@ class ConnectionPool(ThreadedConnectionPool):
|
||||
except psycopg2.pool.PoolError:
|
||||
# If we failed to get the connection slot, wait a bit and try again
|
||||
time.sleep(0.1)
|
||||
raise TimeoutError("Timed out waiting for an available connection to {}".format(self.name))
|
||||
raise TimeoutError(
|
||||
"Timed out waiting for an available connection to {}".format(self.name)
|
||||
)
|
||||
|
||||
|
||||
def get_pool(dbname):
|
||||
"""
|
||||
@ -288,26 +346,32 @@ def get_pool(dbname):
|
||||
# lock
|
||||
if dbname not in connections:
|
||||
log.info("Creating connection pool for: {}".format(dbname))
|
||||
# Actually create the connection pool
|
||||
connections[dbname] = ConnectionPool(
|
||||
dbname,
|
||||
int(config['min_pool_size']),
|
||||
int(config['max_pool_size']),
|
||||
application_name='pgmon',
|
||||
host=config['dbhost'],
|
||||
port=config['dbport'],
|
||||
user=config['dbuser'],
|
||||
connect_timeout=float(config['connect_timeout']),
|
||||
sslmode='require')
|
||||
int(config["min_pool_size"]),
|
||||
int(config["max_pool_size"]),
|
||||
application_name="pgmon",
|
||||
host=config["dbhost"],
|
||||
port=config["dbport"],
|
||||
user=config["dbuser"],
|
||||
connect_timeout=int(config["connect_timeout"]),
|
||||
sslmode=config["ssl_mode"],
|
||||
)
|
||||
# Clear the unhappy indicator if present
|
||||
unhappy_cooldown.pop(dbname, None)
|
||||
return connections[dbname]
|
||||
|
||||
|
||||
def handle_connect_failure(pool):
|
||||
"""
|
||||
Mark the database as being unhappy so we can leave it alone for a while
|
||||
"""
|
||||
dbname = pool.name
|
||||
unhappy_cooldown[dbname] = datetime.now() + timedelta(seconds=int(config['reconnect_cooldown']))
|
||||
unhappy_cooldown[dbname] = datetime.now() + timedelta(
|
||||
seconds=int(config["reconnect_cooldown"])
|
||||
)
|
||||
|
||||
|
||||
def get_query(metric, version):
|
||||
"""
|
||||
@ -318,42 +382,61 @@ def get_query(metric, version):
|
||||
version: The PostgreSQL version number, as given by server_version_num
|
||||
"""
|
||||
# Select the correct query
|
||||
for v in reversed(sorted(metric['query'].keys())):
|
||||
for v in reversed(sorted(metric["query"].keys())):
|
||||
if version >= v:
|
||||
if len(metric['query'][v].strip()) == 0:
|
||||
raise MetricVersionError("Metric no longer applies to PostgreSQL {}".format(version))
|
||||
return metric['query'][v]
|
||||
if len(metric["query"][v].strip()) == 0:
|
||||
raise MetricVersionError(
|
||||
"Metric no longer applies to PostgreSQL {}".format(version)
|
||||
)
|
||||
return metric["query"][v]
|
||||
|
||||
raise MetricVersionError('Missing metric query for PostgreSQL {}'.format(version))
|
||||
raise MetricVersionError("Missing metric query for PostgreSQL {}".format(version))
|
||||
|
||||
|
||||
def json_encode_special(obj):
|
||||
"""
|
||||
Encoder function to handle types the standard JSON package doesn't know what
|
||||
to do with
|
||||
"""
|
||||
if isinstance(obj, Decimal):
|
||||
return float(obj)
|
||||
raise TypeError(f'Cannot serialize object of {type(obj)}')
|
||||
|
||||
|
||||
def run_query_no_retry(pool, return_type, query, args):
|
||||
"""
|
||||
Run the query with no explicit retry code
|
||||
"""
|
||||
with pool.connection(float(config['connect_timeout'])) as conn:
|
||||
with pool.connection(float(config["connect_timeout"])) as conn:
|
||||
try:
|
||||
with conn.cursor(cursor_factory=DictCursor) as curs:
|
||||
with conn.cursor(cursor_factory=RealDictCursor) as curs:
|
||||
curs.execute(query, args)
|
||||
res = curs.fetchall()
|
||||
|
||||
if return_type == 'value':
|
||||
if return_type == "value":
|
||||
if len(res) == 0:
|
||||
return ""
|
||||
return str(list(res[0].values())[0])
|
||||
elif return_type == 'row':
|
||||
return json.dumps(res[0])
|
||||
elif return_type == 'column':
|
||||
return json.dumps([list(r.values())[0] for r in res])
|
||||
elif return_type == 'set':
|
||||
return json.dumps(res)
|
||||
elif return_type == "row":
|
||||
if len(res) == 0:
|
||||
return "[]"
|
||||
return json.dumps(res[0], default=json_encode_special)
|
||||
elif return_type == "column":
|
||||
if len(res) == 0:
|
||||
return "[]"
|
||||
return json.dumps([list(r.values())[0] for r in res], default=json_encode_special)
|
||||
elif return_type == "set":
|
||||
return json.dumps(res, default=json_encode_special)
|
||||
except:
|
||||
dbname = pool.name
|
||||
if dbname in unhappy_cooldown:
|
||||
raise UnhappyDBError()
|
||||
elif conn.broken:
|
||||
elif conn.closed != 0:
|
||||
raise DisconnectedError()
|
||||
else:
|
||||
raise
|
||||
|
||||
|
||||
def run_query(pool, return_type, query, args):
|
||||
"""
|
||||
Run the query, and if we find upon the first attempt that the connection
|
||||
@ -384,6 +467,7 @@ def run_query(pool, return_type, query, args):
|
||||
handle_connect_failure(pool)
|
||||
raise UnhappyDBError()
|
||||
|
||||
|
||||
def get_cluster_version():
|
||||
"""
|
||||
Get the PostgreSQL version if we don't already know it, or if it's been
|
||||
@ -395,26 +479,228 @@ def get_cluster_version():
|
||||
# If we don't know the version or it's past the recheck time, get the
|
||||
# version from the database. Only one thread needs to do this, so they all
|
||||
# try to grab the lock, and then make sure nobody else beat them to it.
|
||||
if cluster_version is None or cluster_version_next_check is None or cluster_version_next_check < datetime.now():
|
||||
if (
|
||||
cluster_version is None
|
||||
or cluster_version_next_check is None
|
||||
or cluster_version_next_check < datetime.now()
|
||||
):
|
||||
with cluster_version_lock:
|
||||
# Only check if nobody already got the version before us
|
||||
if cluster_version is None or cluster_version_next_check is None or cluster_version_next_check < datetime.now():
|
||||
log.info('Checking PostgreSQL cluster version')
|
||||
pool = get_pool(config['dbname'])
|
||||
cluster_version = int(run_query(pool, 'value', 'SHOW server_version_num', None))
|
||||
cluster_version_next_check = datetime.now() + timedelta(seconds=int(config['version_check_period']))
|
||||
if (
|
||||
cluster_version is None
|
||||
or cluster_version_next_check is None
|
||||
or cluster_version_next_check < datetime.now()
|
||||
):
|
||||
log.info("Checking PostgreSQL cluster version")
|
||||
pool = get_pool(config["dbname"])
|
||||
cluster_version = int(
|
||||
run_query(pool, "value", "SHOW server_version_num", None)
|
||||
)
|
||||
cluster_version_next_check = datetime.now() + timedelta(
|
||||
seconds=int(config["version_check_period"])
|
||||
)
|
||||
log.info("Got PostgreSQL cluster version: {}".format(cluster_version))
|
||||
log.debug("Next PostgreSQL cluster version check will be after: {}".format(cluster_version_next_check))
|
||||
log.debug(
|
||||
"Next PostgreSQL cluster version check will be after: {}".format(
|
||||
cluster_version_next_check
|
||||
)
|
||||
)
|
||||
|
||||
return cluster_version
|
||||
|
||||
|
||||
def version_num_to_release(version_num):
|
||||
"""
|
||||
Extract the revease from a version_num.
|
||||
|
||||
In other words, this converts things like:
|
||||
90603 => 9.6
|
||||
130010 => 13
|
||||
"""
|
||||
if version_num // 10000 < 10:
|
||||
return version_num // 10000 + (version_num % 10000 // 100 / 10)
|
||||
else:
|
||||
return version_num // 10000
|
||||
|
||||
|
||||
def parse_version_rss(raw_rss, release):
|
||||
"""
|
||||
Parse the raw RSS from the versions.rss feed to extract the latest version of
|
||||
PostgreSQL that's availabe for the cluster being monitored.
|
||||
|
||||
This sets these global variables:
|
||||
latest_version
|
||||
release_supported
|
||||
|
||||
It is expected that the caller already holds the latest_version_lock lock.
|
||||
|
||||
params:
|
||||
raw_rss: The raw rss text from versions.rss
|
||||
release: The PostgreSQL release we care about (ex: 9.2, 14)
|
||||
"""
|
||||
global latest_version
|
||||
global release_supported
|
||||
|
||||
# Regular expressions for parsing the RSS document
|
||||
version_line = re.compile(
|
||||
r".*?([0-9][0-9.]+) is the latest release in the {} series.*".format(release)
|
||||
)
|
||||
unsupported_line = re.compile(r"^This version is unsupported")
|
||||
|
||||
# Loop through the RSS until we find the current release
|
||||
release_found = False
|
||||
for line in raw_rss.splitlines():
|
||||
m = version_line.match(line)
|
||||
if m:
|
||||
# Note that we found the version we were looking for
|
||||
release_found = True
|
||||
|
||||
# Convert the version to version_num format
|
||||
version = m.group(1)
|
||||
parts = list(map(int, version.split(".")))
|
||||
if parts[0] < 10:
|
||||
latest_version = int(
|
||||
"{}{:02}{:02}".format(parts[0], parts[1], parts[2])
|
||||
)
|
||||
else:
|
||||
latest_version = int("{}00{:02}".format(parts[0], parts[1]))
|
||||
elif release_found:
|
||||
# The next line after the version tells if the version is supported
|
||||
if unsupported_line.match(line):
|
||||
release_supported = False
|
||||
else:
|
||||
release_supported = True
|
||||
break
|
||||
|
||||
# Make sure we actually found it
|
||||
if not release_found:
|
||||
raise LatestVersionCheckError("Current release ({}) not found".format(release))
|
||||
|
||||
log.info(
|
||||
"Got latest PostgreSQL version: {} supported={}".format(
|
||||
latest_version, release_supported
|
||||
)
|
||||
)
|
||||
log.debug(
|
||||
"Next latest PostgreSQL version check will be after: {}".format(
|
||||
latest_version_next_check
|
||||
)
|
||||
)
|
||||
|
||||
|
||||
def get_latest_version():
|
||||
"""
|
||||
Get the latest supported version of the major PostgreSQL release running on the server being monitored.
|
||||
"""
|
||||
|
||||
global latest_version_next_check
|
||||
|
||||
# If we don't know the latest version or it's past the recheck time, get the
|
||||
# version from the PostgreSQL RSS feed. Only one thread needs to do this, so
|
||||
# they all try to grab the lock, and then make sure nobody else beat them to it.
|
||||
if (
|
||||
latest_version is None
|
||||
or latest_version_next_check is None
|
||||
or latest_version_next_check < datetime.now()
|
||||
):
|
||||
# Note: we get the cluster version here before grabbing the latest_version_lock
|
||||
# lock so it's not held while trying to talk with the DB.
|
||||
release = version_num_to_release(get_cluster_version())
|
||||
|
||||
with latest_version_lock:
|
||||
# Only check if nobody already got the version before us
|
||||
if (
|
||||
latest_version is None
|
||||
or latest_version_next_check is None
|
||||
or latest_version_next_check < datetime.now()
|
||||
):
|
||||
log.info("Checking latest PostgreSQL version")
|
||||
latest_version_next_check = datetime.now() + timedelta(
|
||||
seconds=int(config["latest_version_check_period"])
|
||||
)
|
||||
|
||||
# Grab the RSS feed
|
||||
raw_rss = requests.get("https://www.postgresql.org/versions.rss")
|
||||
if raw_rss.status_code != 200:
|
||||
raise LatestVersionCheckError("code={}".format(r.status_code))
|
||||
|
||||
# Parse the RSS body and set global variables
|
||||
parse_version_rss(raw_rss.text, release)
|
||||
|
||||
return latest_version
|
||||
|
||||
|
||||
def sample_metric(dbname, metric_name, args, retry=True):
|
||||
"""
|
||||
Run the appropriate query for the named metric against the specified database
|
||||
"""
|
||||
# Get the metric definition
|
||||
try:
|
||||
metric = config["metrics"][metric_name]
|
||||
except KeyError:
|
||||
raise UnknownMetricError("Unknown metric: {}".format(metric_name))
|
||||
|
||||
# Get the connection pool for the database, or create one if it doesn't
|
||||
# already exist.
|
||||
pool = get_pool(dbname)
|
||||
|
||||
# Identify the PostgreSQL version
|
||||
version = get_cluster_version()
|
||||
|
||||
# Get the query version
|
||||
query = get_query(metric, version)
|
||||
|
||||
# Execute the quert
|
||||
if retry:
|
||||
return run_query(pool, metric["type"], query, args)
|
||||
else:
|
||||
return run_query_no_retry(pool, metric["type"], query, args)
|
||||
|
||||
|
||||
def test_queries():
|
||||
"""
|
||||
Run all of the metric queries against a database and check the results
|
||||
"""
|
||||
# We just use the default db for tests
|
||||
dbname = config["dbname"]
|
||||
# Loop through all defined metrics.
|
||||
for name, metric in config["metrics"].items():
|
||||
# If the metric has arguments to use while testing, grab those
|
||||
args = metric.get("test_args", {})
|
||||
print("Testing {} [{}]".format(name, ", ".join(["{}={}".format(key, value) for key, value in args.items()])))
|
||||
# When testing against a docker container, we may end up connecting
|
||||
# before the service is truly up (it restarts during the initialization
|
||||
# phase). To cope with this, we'll allow a few connection failures.
|
||||
tries = 5
|
||||
while True:
|
||||
# Run the query without the ability to retry
|
||||
try:
|
||||
res = sample_metric(dbname, name, args, retry=False)
|
||||
break
|
||||
except MetricVersionError:
|
||||
res = "Unsupported for this version"
|
||||
break
|
||||
except psycopg2.OperationalError as e:
|
||||
print("Error encountered, {} tries left: {}".format(tries, e))
|
||||
if tries <= 0:
|
||||
raise
|
||||
time.sleep(1)
|
||||
tries -= 1
|
||||
# Compare the result to the provided sample results
|
||||
# TODO
|
||||
print("{} -> {}".format(name, res))
|
||||
# Return the number of errors
|
||||
# TODO
|
||||
return 0
|
||||
|
||||
|
||||
class SimpleHTTPRequestHandler(BaseHTTPRequestHandler):
|
||||
"""
|
||||
This is our request handling server. It is responsible for listening for
|
||||
requests, processing them, and responding.
|
||||
"""
|
||||
|
||||
def log_request(self, code='-', size='-'):
|
||||
def log_request(self, code="-", size="-"):
|
||||
"""
|
||||
Override to suppress standard request logging
|
||||
"""
|
||||
@ -436,71 +722,58 @@ class SimpleHTTPRequestHandler(BaseHTTPRequestHandler):
|
||||
"""
|
||||
# Parse the URL
|
||||
parsed_path = urlparse(self.path)
|
||||
name = parsed_path.path.strip('/')
|
||||
metric_name = parsed_path.path.strip("/")
|
||||
parsed_query = parse_qs(parsed_path.query)
|
||||
|
||||
if name == 'agent_version':
|
||||
if metric_name == "agent_version":
|
||||
self._reply(200, VERSION)
|
||||
return
|
||||
elif metric_name == "latest_version_info":
|
||||
try:
|
||||
get_latest_version()
|
||||
self._reply(
|
||||
200,
|
||||
json.dumps(
|
||||
{
|
||||
"latest": latest_version,
|
||||
"supported": 1 if release_supported else 0,
|
||||
}
|
||||
),
|
||||
)
|
||||
except LatestVersionCheckError as e:
|
||||
log.error("Failed to retrieve latest version information: {}".format(e))
|
||||
self._reply(503, "Failed to retrieve latest version info")
|
||||
return
|
||||
|
||||
# Note: parse_qs returns the values as a list. Since we always expect
|
||||
# single values, just grab the first from each.
|
||||
args = {key: values[0] for key, values in parsed_query.items()}
|
||||
|
||||
# Get the metric definition
|
||||
try:
|
||||
metric = config['metrics'][name]
|
||||
except KeyError:
|
||||
log.error("Unknown metric: {}".format(name))
|
||||
self._reply(404, 'Unknown metric')
|
||||
return
|
||||
|
||||
# Get the dbname. If none was provided, use the default from the
|
||||
# config.
|
||||
dbname = args.get('dbname', config['dbname'])
|
||||
dbname = args.get("dbname", config["dbname"])
|
||||
|
||||
# Get the connection pool for the database, or create one if it doesn't
|
||||
# already exist.
|
||||
# Sample the metric
|
||||
try:
|
||||
pool = get_pool(dbname)
|
||||
except UnhappyDBError:
|
||||
self._reply(200, sample_metric(dbname, metric_name, args))
|
||||
return
|
||||
except UnknownMetricError as e:
|
||||
log.error("Unknown metric: {}".format(metric_name))
|
||||
self._reply(404, "Unknown metric")
|
||||
return
|
||||
except MetricVersionError as e:
|
||||
log.error(
|
||||
"Failed to find a version of {} for {}".format(metric_name, version)
|
||||
)
|
||||
self._reply(404, "Unsupported version")
|
||||
return
|
||||
except UnhappyDBError as e:
|
||||
log.info("Database {} is unhappy, please be patient".format(dbname))
|
||||
self._reply(503, 'Database unavailable')
|
||||
return
|
||||
|
||||
# Identify the PostgreSQL version
|
||||
try:
|
||||
version = get_cluster_version()
|
||||
except UnhappyDBError:
|
||||
self._reply(503, "Database unavailable")
|
||||
return
|
||||
except Exception as e:
|
||||
if dbname in unhappy_cooldown:
|
||||
log.info("Database {} is unhappy, please be patient".format(dbname))
|
||||
self._reply(503, 'Database unavailable')
|
||||
else:
|
||||
log.error("Failed to get PostgreSQL version: {}".format(e))
|
||||
self._reply(500, 'Error getting DB version')
|
||||
return
|
||||
|
||||
# Get the query version
|
||||
try:
|
||||
query = get_query(metric, version)
|
||||
except KeyError:
|
||||
log.error("Failed to find a version of {} for {}".format(name, version))
|
||||
self._reply(404, 'Unsupported version')
|
||||
return
|
||||
|
||||
# Execute the quert
|
||||
try:
|
||||
self._reply(200, run_query(pool, metric['type'], query, args))
|
||||
return
|
||||
except Exception as e:
|
||||
if dbname in unhappy_cooldown:
|
||||
log.info("Database {} is unhappy, please be patient".format(dbname))
|
||||
self._reply(503, 'Database unavailable')
|
||||
else:
|
||||
log.error("Error running query: {}".format(e))
|
||||
self._reply(500, "Error running query")
|
||||
log.error("Error running query: {}".format(e))
|
||||
self._reply(500, "Unexpected error: {}".format(e))
|
||||
return
|
||||
|
||||
def _reply(self, code, content):
|
||||
@ -508,19 +781,29 @@ class SimpleHTTPRequestHandler(BaseHTTPRequestHandler):
|
||||
Send a reply to the client
|
||||
"""
|
||||
self.send_response(code)
|
||||
self.send_header('Content-type', 'application/json')
|
||||
self.send_header("Content-type", "application/json")
|
||||
self.end_headers()
|
||||
|
||||
self.wfile.write(bytes(content, 'utf-8'))
|
||||
self.wfile.write(bytes(content, "utf-8"))
|
||||
|
||||
if __name__ == '__main__':
|
||||
|
||||
if __name__ == "__main__":
|
||||
# Handle cli args
|
||||
parser = argparse.ArgumentParser(
|
||||
prog = 'pgmon',
|
||||
description='A PostgreSQL monitoring agent')
|
||||
prog="pgmon", description="A PostgreSQL monitoring agent"
|
||||
)
|
||||
|
||||
parser.add_argument('config_file', default='pgmon.yml', nargs='?',
|
||||
help='The config file to read (default: %(default)s)')
|
||||
parser.add_argument(
|
||||
"-c",
|
||||
"--config_file",
|
||||
default="pgmon.yml",
|
||||
nargs="?",
|
||||
help="The config file to read (default: %(default)s)",
|
||||
)
|
||||
|
||||
parser.add_argument(
|
||||
"-t", "--test", action="store_true", help="Run query tests and exit"
|
||||
)
|
||||
|
||||
args = parser.parse_args()
|
||||
|
||||
@ -530,8 +813,16 @@ if __name__ == '__main__':
|
||||
# Read the config file
|
||||
read_config(config_file)
|
||||
|
||||
# Run query tests and exit if test mode is enabled
|
||||
if args.test:
|
||||
errors = test_queries()
|
||||
if errors > 0:
|
||||
sys.exit(1)
|
||||
else:
|
||||
sys.exit(0)
|
||||
|
||||
# Set up the http server to receive requests
|
||||
server_address = ('127.0.0.1', config['port'])
|
||||
server_address = (config["address"], config["port"])
|
||||
httpd = ThreadingHTTPServer(server_address, SimpleHTTPRequestHandler)
|
||||
|
||||
# Set up the signal handler
|
||||
@ -539,7 +830,7 @@ if __name__ == '__main__':
|
||||
signal.signal(signal.SIGHUP, signal_handler)
|
||||
|
||||
# Handle requests.
|
||||
log.info("Listening on port {}...".format(config['port']))
|
||||
log.info("Listening on port {}...".format(config["port"]))
|
||||
while running:
|
||||
httpd.handle_request()
|
||||
|
||||
|
||||
File diff suppressed because it is too large
Load Diff
@ -7,7 +7,7 @@ After=network.target
|
||||
[Service]
|
||||
EnvironmentFile=/etc/pgmon/%i-service.conf
|
||||
User=${SERVICE_USER:-postgres}
|
||||
ExecStart=/usr/local/bin/pgmon /etc/pgmon/%i.yml
|
||||
ExecStart=/usr/bin/pgmon -c /etc/pgmon/%i.yml
|
||||
ExecReload=kill -HUP $MAINPID
|
||||
Restart=on-failure
|
||||
Type=exec
|
||||
|
||||
23
tests/Dockerfile
Normal file
23
tests/Dockerfile
Normal file
@ -0,0 +1,23 @@
|
||||
FROM alpine:3.21
|
||||
|
||||
RUN apk update && \
|
||||
apk add py3-psycopg2 \
|
||||
py3-requests \
|
||||
py3-yaml \
|
||||
tini
|
||||
|
||||
WORKDIR /app
|
||||
|
||||
COPY src/pgmon.py /app/
|
||||
|
||||
COPY sample-config/pgmon-metrics.yml /app/
|
||||
|
||||
COPY tests/test-config.yml /app/
|
||||
|
||||
COPY --chmod=0600 --chown=postgres:postgres tests/pgpass /root/.pgpass
|
||||
|
||||
ENTRYPOINT ["tini", "--"]
|
||||
|
||||
EXPOSE 5400
|
||||
|
||||
CMD ["/app/pgmon.py", "-c", "/app/test-config.yml", "--test"]
|
||||
32
tests/docker-compose.yml
Normal file
32
tests/docker-compose.yml
Normal file
@ -0,0 +1,32 @@
|
||||
---
|
||||
|
||||
services:
|
||||
agent:
|
||||
image: pgmon
|
||||
build:
|
||||
context: ..
|
||||
dockerfile: tests/Dockerfile
|
||||
ports:
|
||||
- :5400
|
||||
depends_on:
|
||||
db:
|
||||
condition: service_healthy
|
||||
|
||||
db:
|
||||
image: "postgres:${PGTAG:-17-bookworm}"
|
||||
ports:
|
||||
- :5432
|
||||
environment:
|
||||
POSTGRES_PASSWORD: secret
|
||||
healthcheck:
|
||||
#test: [ "CMD", "pg_isready", "-U", "postgres" ]
|
||||
test: [ "CMD-SHELL", "pg_controldata /var/lib/postgresql/data/ | grep -q 'in production'" ]
|
||||
interval: 5s
|
||||
timeout: 2s
|
||||
retries: 40
|
||||
command: >
|
||||
postgres -c ssl=on
|
||||
-c ssl_cert_file='/etc/ssl/certs/ssl-cert-snakeoil.pem'
|
||||
-c ssl_key_file='/etc/ssl/private/ssl-cert-snakeoil.key'
|
||||
-c listen_addresses='*'
|
||||
|
||||
1
tests/pgpass
Normal file
1
tests/pgpass
Normal file
@ -0,0 +1 @@
|
||||
db:5432:*:postgres:secret
|
||||
65
tests/run-tests.sh
Executable file
65
tests/run-tests.sh
Executable file
@ -0,0 +1,65 @@
|
||||
#!/bin/bash
|
||||
|
||||
# Versions to test
|
||||
versions=( $@ )
|
||||
|
||||
# If we weren't given any versions, test them all
|
||||
if [ ${#versions[@]} -eq 0 ]
|
||||
then
|
||||
versions=( 9.2 9.4 9.6 10 11 12 13 14 15 16 17 )
|
||||
fi
|
||||
|
||||
# Image tags to use
|
||||
declare -A images=()
|
||||
images["9.2"]='9.2'
|
||||
images["9.3"]='9.3'
|
||||
images["9.4"]='9.4'
|
||||
images["9.5"]='9.5'
|
||||
images["9.6"]='9.6-bullseye'
|
||||
images["10"]='10-bullseye'
|
||||
images["11"]='11-bookworm'
|
||||
images["12"]='12-bookworm'
|
||||
images["13"]='13-bookworm'
|
||||
images["14"]='14-bookworm'
|
||||
images["15"]='15-bookworm'
|
||||
images["16"]='16-bookworm'
|
||||
images["17"]='17-bookworm'
|
||||
|
||||
declare -A results=()
|
||||
|
||||
# Make sure everything's down to start with
|
||||
docker compose down
|
||||
|
||||
# Make sure our agent container is up to date
|
||||
docker compose build agent
|
||||
|
||||
for version in "${versions[@]}"
|
||||
do
|
||||
echo
|
||||
echo "Testing: PostgreSQL ${version}"
|
||||
|
||||
# Specify the version we're testing against
|
||||
export PGTAG="${images["$version"]}"
|
||||
|
||||
# Start the containers
|
||||
docker compose up --exit-code-from=agent agent
|
||||
rc=$?
|
||||
|
||||
results["$version"]=$rc
|
||||
|
||||
# Destroy the containers
|
||||
docker compose down
|
||||
done
|
||||
|
||||
echo
|
||||
echo
|
||||
for v in "${versions[@]}"
|
||||
do
|
||||
case "${results["$v"]}" in
|
||||
0) msg="OK" ;;
|
||||
1) msg="Query failure detected" ;;
|
||||
18) msg="Docker image error: 18" ;;
|
||||
*) msg="Unexpected error: ${results["$v"]}" ;;
|
||||
esac
|
||||
echo "$v -> $msg"
|
||||
done
|
||||
17
tests/test-config.yml
Normal file
17
tests/test-config.yml
Normal file
@ -0,0 +1,17 @@
|
||||
---
|
||||
|
||||
# Bind to all interfaces so we can submit requests from outside the test container
|
||||
address: 0.0.0.0
|
||||
|
||||
# We always just connect to the db container
|
||||
dbhost: db
|
||||
dbport: 5432
|
||||
dbuser: postgres
|
||||
|
||||
# The SSL cipher parameters are too old in the 9.2 container, so we allow the tests
|
||||
# to be run without encryption
|
||||
ssl_mode: prefer
|
||||
|
||||
# Pull in the standard metrics
|
||||
include:
|
||||
- pgmon-metrics.yml
|
||||
File diff suppressed because it is too large
Load Diff
Loading…
Reference in New Issue
Block a user