mirror of
https://github.com/mag37/dockcheck.git
synced 2026-04-17 18:07:46 +00:00
Compare commits
85 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
bf1e78d2ff | ||
|
|
9fa398e553 | ||
|
|
9ef2ea7135 | ||
|
|
8c3b899332 | ||
|
|
8110cd8892 | ||
|
|
502a167919 | ||
|
|
61f90893ef | ||
|
|
229cde0efb | ||
|
|
0c6674ac8e | ||
|
|
3e079e2ec5 | ||
|
|
b2d67c9f52 | ||
|
|
3aeee837f9 | ||
|
|
408a8b14dd | ||
|
|
a2868ea505 | ||
|
|
62a3d10b4f | ||
|
|
210c076968 | ||
|
|
65e875e860 | ||
|
|
3655f5ae8a | ||
|
|
78a7e1137f | ||
|
|
76e6a5c38b | ||
|
|
41029f628d | ||
|
|
b918844336 | ||
|
|
27896c18ba | ||
|
|
e68adb34d0 | ||
|
|
6bc896b193 | ||
|
|
6cba140522 | ||
|
|
73050abf10 | ||
|
|
d4d89c305c | ||
|
|
bc6a3529c7 | ||
|
|
01b9b33751 | ||
|
|
9ed2a0bad0 | ||
|
|
43307350ca | ||
|
|
b3600f26ac | ||
|
|
79def47754 | ||
|
|
e6ff634394 | ||
|
|
6444e18a4f | ||
|
|
14872b0471 | ||
|
|
07ad241e9f | ||
|
|
58d53d3aaf | ||
|
|
15ce226a0a | ||
|
|
cce5438aca | ||
|
|
be6f5edf52 | ||
|
|
99befd6938 | ||
|
|
8a63fd360c | ||
|
|
d3786d6f75 | ||
|
|
487cfb2822 | ||
|
|
b4943df46c | ||
|
|
8603c8d4b6 | ||
|
|
ae66a6f0fd | ||
|
|
3ac0521a9b | ||
|
|
00ae250511 | ||
|
|
704387a7fe | ||
|
|
1a6826e2ac | ||
|
|
a28b9e555f | ||
|
|
8309b80dc2 | ||
|
|
bbe26a0ac2 | ||
|
|
d98d052af7 | ||
|
|
fbba77dc1f | ||
|
|
09c9faa008 | ||
|
|
17ed46d157 | ||
|
|
d744d51473 | ||
|
|
3821c22660 | ||
|
|
5dd42b7336 | ||
|
|
8da2cbe611 | ||
|
|
9937d6bc3f | ||
|
|
5517abb090 | ||
|
|
cb8c23e62d | ||
|
|
0009adc2cf | ||
|
|
8d39e1b66b | ||
|
|
ec068dd19a | ||
|
|
119932aa19 | ||
|
|
a9f2433fab | ||
|
|
18e467274e | ||
|
|
6db92be627 | ||
|
|
fc47b1313b | ||
|
|
76d2921719 | ||
|
|
2844e54fbd | ||
|
|
e0d66cc773 | ||
|
|
071fa4b245 | ||
|
|
cb73025a61 | ||
|
|
87f1886d28 | ||
|
|
ac10a96cc8 | ||
|
|
e5a7a715c6 | ||
|
|
8a6a1fbec2 | ||
|
|
04b0c46366 |
73
README.md
73
README.md
@@ -10,19 +10,30 @@
|
||||
<a href="https://github.com/sponsors/mag37"><img src="https://img.shields.io/badge/-Sponsor-grey?style=flat-square&logo=github" alt="Github Sponsor"></a>
|
||||
</p>
|
||||
|
||||
<h3 align="center">CLI tool to automate docker image updates. <br>No <b>pre-pull</b>, selective, optional notifications and prune when done.</h3>
|
||||
<h2 align="center">Now with simple notification integrations!</h2>
|
||||
<h4 align="center">With features like excluding specific containers, custom container labels, auto-prune when done and more.</h4>
|
||||
<h2 align="center">CLI tool to automate docker image updates or notifying when updates are available.</h2>
|
||||
<h3 align="center">Features:</h3>
|
||||
<h3 align="center">selective updates, exclude containers, custom labels, notification plugins, prune when done and more.</h3>
|
||||
|
||||
<h4 align="center">For Podman - see the fork <a href="https://github.com/sudo-kraken/podcheck">sudo-kraken/podcheck</a>!</h4>
|
||||
|
||||
<h4 align="center">:whale: Docker Hub pull limit :chart_with_downwards_trend: not an issue for checks but for actual pulls - <a href="#whale-docker-hub-pull-limit-chart_with_downwards_trend-not-an-issue-for-checks-but-for-actual-pulls">read more</a></h4>
|
||||
|
||||
___
|
||||
## :bell: Changelog
|
||||
|
||||
- **v0.4.9**: Added a function to enrich the notify-message with release note URLs. See [Release notes addon](https://github.com/mag37/dockcheck#date-release-notes-addon-to-notifications)
|
||||
- **v0.4.8**: Rewrote prune logic to not prompt with options `-a|-y` or `-n`. Auto prune with `-p`.
|
||||
- **v0.4.7**: Notification Template changes to gotify(new!), DSM(improved), SMTP(deprecation alternative).
|
||||
- **v0.4.6**: Compatibility changes to timeout, due to busybox.
|
||||
- **v0.4.5**: Bugfixes, compatibility changes to timeout and arrays.
|
||||
- **v0.4.3**: Added timeout option to skip container if registry check takes too long (10s default).
|
||||
Made MaxAsync=1 the default - edit to change.
|
||||
Added -x option to pass a MaxAsync value on runtime.
|
||||
Made it possible to disable xargs -P-flag by setting MaxAsync=0 or passing -x 0 option.
|
||||
|
||||
- **v0.5.6.1**: Async xargs hotfix - due to errors `failed to request manifest head ... context canceled`
|
||||
- Defaulted subprocess to 1 with `MaxAsync=1`, increase to find a stable value in your environment.
|
||||
- Added `-x N` option to pass `MaxAsync` value at runtime.
|
||||
- To disable xargs `-P` flag (max processes) all together, set `MaxAsync` to 0.
|
||||
- **v0.5.6.0**: Heavily improved performance due to async checking for updates.
|
||||
- **v0.5.5.0**: osx and bsd compatibility changes + rewrite of dependency installer
|
||||
- **v0.5.4.0**: Added support for a Prometheus+node_exporter metric collection through a file collector.
|
||||
- **v0.5.3.0**: Local image check changed (use imageId instead of name) and Gotify-template fixed (whale icon removed).
|
||||
- **v0.5.2.1**: Rewrite of dependency downloads, jq can be installed with package manager or static binary.
|
||||
___
|
||||
|
||||
|
||||
@@ -36,6 +47,7 @@ Example: dockcheck.sh -y -d 10 -e nextcloud,heimdall
|
||||
|
||||
Options:"
|
||||
-a|y Automatic updates, without interaction.
|
||||
-c D Exports metrics as prom file for the prometheus node_exporter. Provide the collector textfile directory.
|
||||
-d N Only update to new images that are N+ days old. Lists too recent with +prefix and age. 2xSlower.
|
||||
-e X Exclude containers, separated by comma.
|
||||
-f Force stack restart after update. Caution: restarts once for every updated container within stack.
|
||||
@@ -51,7 +63,6 @@ Options:"
|
||||
-v Prints current version.
|
||||
```
|
||||
|
||||
|
||||
### Basic example:
|
||||
```
|
||||
$ ./dockcheck.sh
|
||||
@@ -68,20 +79,24 @@ Containers with updates available:
|
||||
Choose what containers to update:
|
||||
Enter number(s) separated by comma, [a] for all - [q] to quit:
|
||||
```
|
||||
Then it proceedes to run `pull` and `up -d` on every container with updates.
|
||||
Then it proceeds to run `pull` and `up -d` on every container with updates.
|
||||
After the updates are complete, you'll get prompted if you'd like to prune dangling images.
|
||||
|
||||
___
|
||||
|
||||
## :nut_and_bolt: Dependencies
|
||||
- Running docker (duh) and compose, either standalone or plugin.
|
||||
- Running docker (duh) and compose, either standalone or plugin. (see [Podman fork](https://github.com/sudo-kraken/podcheck)
|
||||
- Bash shell or compatible shell of at least v4.3
|
||||
- POSIX `xargs`, usually default but can be installed with the `findutils` package - to enable async.
|
||||
- [jq](https://github.com/jqlang/jq)
|
||||
- User will be prompted to install with package manager or download static binary.
|
||||
- [regclient/regctl](https://github.com/regclient/regclient) (Licensed under [Apache-2.0 License](http://www.apache.org/licenses/LICENSE-2.0))
|
||||
- User will be prompted to download `regctl` if not in `PATH` or `PWD`.
|
||||
- regctl requires `amd64/arm64` - see [workaround](#roller_coaster-workaround-for-non-amd64--arm64) if other architecture is used.
|
||||
|
||||
## :tent: Install Instructions
|
||||
Download the script to a directory in **PATH**, I'd suggest using `~/.local/bin` as that's usually in **PATH**.
|
||||
Download the script to a directory in **PATH**, I'd suggest using `~/.local/bin` as that's usually in **PATH**.
|
||||
For OSX/macOS preferably use `/usr/local/bin`.
|
||||
```sh
|
||||
# basic example with curl:
|
||||
curl -L https://raw.githubusercontent.com/mag37/dockcheck/main/dockcheck.sh -o ~/.local/bin/dockcheck.sh
|
||||
@@ -89,6 +104,9 @@ chmod +x ~/.local/bin/dockcheck.sh
|
||||
|
||||
# or oneliner with wget:
|
||||
wget -O ~/.local/bin/dockcheck.sh "https://raw.githubusercontent.com/mag37/dockcheck/main/dockcheck.sh" && chmod +x ~/.local/bin/dockcheck.sh
|
||||
|
||||
# OSX or macOS version with curl:
|
||||
curl -L https://raw.githubusercontent.com/mag37/dockcheck/main/dockcheck.sh -o /usr/local/bin/dockcheck.sh && chmod +x /usr/local/bin/dockcheck.sh
|
||||
```
|
||||
Then call the script anywhere with just `dockcheck.sh`.
|
||||
Add preferred `notify.sh`-template to the same directory - this will not be touched by the scripts self-update function.
|
||||
@@ -111,13 +129,14 @@ Use a `notify_X.sh` template file from the **notify_templates** directory, copy
|
||||
- [Telegram](https://telegram.org/) - Telegram chat API.
|
||||
- [Matrix-Synapse](https://github.com/element-hq/synapse) - [Matrix](https://matrix.org/), open, secure, decentralised communication.
|
||||
- [Pushover](https://pushover.net/) - Simple Notifications (to your phone, wearables, desktops)
|
||||
- [Discord](https://support.discord.com/hc/en-us/articles/228383668-Intro-to-Webhooks) - Discord webhooks.
|
||||
|
||||
Further additions are welcome - suggestions or PR!
|
||||
<sub><sup>Initiated and first contributed by [yoyoma2](https://github.com/yoyoma2).</sup></sub>
|
||||
|
||||
### :date: Release notes addon to Notifications
|
||||
There's a function to use a lookup-file to add release note URL's to the notification message.
|
||||
Copy the notify_templates/`urls.list` file to the script directory and modify it as necessary, it will be used automatically if it's there.
|
||||
There's a function to use a lookup-file to add release note URL's to the notification message.
|
||||
Copy the notify_templates/`urls.list` file to the script directory, it will be used automatically if it's there. Modify it as necessary, the names of interest in the left column needs to match your container names.
|
||||
The output of the notification will look something like this:
|
||||
```
|
||||
Containers on hostname with updates available:
|
||||
@@ -128,6 +147,17 @@ nginx -> https://github.com/docker-library/official-images/blob/master/library
|
||||
```
|
||||
The `urls.list` file is just an example and I'd gladly see that people contribute back when they add their preferred URLs to their lists.
|
||||
|
||||
## :chart_with_upwards_trend: Prometheus and node_exporter
|
||||
Dockcheck can be used together with [Prometheus](https://github.com/prometheus/prometheus) and [node_exporter](https://github.com/prometheus/node_exporter) to export metrics via the file collector, scheduled with cron or likely.
|
||||
This is done with the `-c` option, like this:
|
||||
```
|
||||
dockcheck.sh -c /path/to/exporter/directory
|
||||
```
|
||||
|
||||
See the [README.md](./addons/prometheus/README.md) for more detailed information on how to set it up!
|
||||
|
||||
<sub><sup>Contributed by [tdralle](https://github.com/tdralle).</sup></sub>
|
||||
|
||||
## :bookmark: Labels
|
||||
Optionally add labels to compose-files. Currently these are the usable labels:
|
||||
```
|
||||
@@ -158,7 +188,15 @@ chmod 755 regctl
|
||||
```
|
||||
Test it with `./regctl --help` and then either add the file to the same path as *dockcheck.sh* or in your path (eg. `~/.local/bin/regctl`).
|
||||
|
||||
## :guardsman: Function to auth with docker hub before running
|
||||
## :whale: Docker Hub pull limit :chart_with_downwards_trend: not an issue for checks but for actual pulls
|
||||
Due to recent changes in [Docker Hub usage and limits](https://docs.docker.com/docker-hub/usage/)
|
||||
>Unauthenticated users: 10 pulls/hour
|
||||
>Authenticated users with a free account: 100 pulls/hour
|
||||
|
||||
This is not an issue for registry checks. But if you have a large stack and pull more than 10 updates at once consider updating more often or to create a free account.
|
||||
You could use/modify the login-wrapper function in the example below to automate the login prior to running `dockcheck.sh`.
|
||||
|
||||
### :guardsman: Function to auth with docker hub before running
|
||||
**Example** - Change names, paths, and remove cat+password flag if you rather get prompted:
|
||||
```sh
|
||||
function dchk {
|
||||
@@ -179,7 +217,7 @@ function dchk {
|
||||
Containers need to be manually stopped, removed and created again to run on the new image.
|
||||
|
||||
## :wrench: Debugging
|
||||
If you hit issues, you could check the output of the `extras/errorCheck.sh` script for clues.
|
||||
If you hit issues, you could check the output of the `extras/errorCheck.sh` script for clues.
|
||||
Another option is to run the main script with debugging in a subshell `bash -x dockcheck.sh` - if there's a particular container/image that's causing issues you can filter for just that through `bash -x dockcheck.sh nginx`.
|
||||
|
||||
## :scroll: License
|
||||
@@ -192,4 +230,3 @@ dockcheck is created and released under the [GNU GPL v3.0](https://www.gnu.org/l
|
||||
___
|
||||
|
||||
### :floppy_disk: The [story](https://mag37.org/posts/project_dockcheck/) behind it. 1 year in retrospect.
|
||||
|
||||
|
||||
61
addons/prometheus/README.md
Normal file
61
addons/prometheus/README.md
Normal file
@@ -0,0 +1,61 @@
|
||||
## [Prometheus](https://github.com/prometheus/prometheus) and [node_exporter](https://github.com/prometheus/node_exporter)
|
||||
Dockcheck is capable to export metrics to prometheus via the text file collector provided by the node_exporter.
|
||||
In order to do so the -c flag has to be specified followed by the file path that is configured in the text file collector of the node_exporter.
|
||||
A simple cron job can be configured to export these metrics on a regular interval as shown in the sample below:
|
||||
|
||||
```
|
||||
0 1 * * * /root/dockcheck.sh -n -c /var/lib/node_exporter/textfile_collector
|
||||
```
|
||||
|
||||
The following metrics are exported to prometheus
|
||||
|
||||
```
|
||||
# HELP dockcheck_images_analyzed Docker images that have been analyzed
|
||||
# TYPE dockcheck_images_analyzed gauge
|
||||
dockcheck_images_analyzed 22
|
||||
# HELP dockcheck_images_outdated Docker images that are outdated
|
||||
# TYPE dockcheck_images_outdated gauge
|
||||
dockcheck_images_outdated 7
|
||||
# HELP dockcheck_images_latest Docker images that are outdated
|
||||
# TYPE dockcheck_images_latest gauge
|
||||
dockcheck_images_latest 14
|
||||
# HELP dockcheck_images_error Docker images with analysis errors
|
||||
# TYPE dockcheck_images_error gauge
|
||||
dockcheck_images_error 1
|
||||
# HELP dockcheck_images_analyze_timestamp_seconds Last dockercheck run time
|
||||
# TYPE dockcheck_images_analyze_timestamp_seconds gauge
|
||||
dockcheck_images_analyze_timestamp_seconds 1737924029
|
||||
```
|
||||
|
||||
Once those metrics are exported they can be used to define alarms as shown below
|
||||
|
||||
```
|
||||
- alert: dockcheck_images_outdated
|
||||
expr: sum by(instance) (dockcheck_images_outdated) > 0
|
||||
for: 15s
|
||||
labels:
|
||||
severity: warning
|
||||
annotations:
|
||||
summary: "{{ $labels.instance }} has {{ $value }} outdated docker images."
|
||||
description: "{{ $labels.instance }} has {{ $value }} outdated docker images."
|
||||
- alert: dockcheck_images_error
|
||||
expr: sum by(instance) (dockcheck_images_error) > 0
|
||||
for: 15s
|
||||
labels:
|
||||
severity: warning
|
||||
annotations:
|
||||
summary: "{{ $labels.instance }} has {{ $value }} docker images having an error."
|
||||
description: "{{ $labels.instance }} has {{ $value }} docker images having an error."
|
||||
- alert: dockercheck_image_last_analyze
|
||||
expr: (time() - dockcheck_images_analyze_timestamp_seconds) > (3600 * 24 * 3)
|
||||
for: 15s
|
||||
labels:
|
||||
severity: warning
|
||||
annotations:
|
||||
summary: "{{ $labels.instance }} has not updated the dockcheck statistics for more than 3 days."
|
||||
description: "{{ $labels.instance }} has not updated the dockcheck statistics for more than 3 days."
|
||||
```
|
||||
|
||||
There is a reference Grafana dashboard in [grafana/grafana_dashboard.json](./grafana/grafana_dashboard.json).
|
||||
|
||||

|
||||
382
addons/prometheus/grafana/grafana_dashboard.json
Normal file
382
addons/prometheus/grafana/grafana_dashboard.json
Normal file
@@ -0,0 +1,382 @@
|
||||
{
|
||||
"__inputs": [
|
||||
{
|
||||
"name": "DS_PROMETHEUS",
|
||||
"label": "prometheus",
|
||||
"description": "",
|
||||
"type": "datasource",
|
||||
"pluginId": "prometheus",
|
||||
"pluginName": "Prometheus"
|
||||
}
|
||||
],
|
||||
"__elements": {},
|
||||
"__requires": [
|
||||
{
|
||||
"type": "grafana",
|
||||
"id": "grafana",
|
||||
"name": "Grafana",
|
||||
"version": "11.4.0"
|
||||
},
|
||||
{
|
||||
"type": "datasource",
|
||||
"id": "prometheus",
|
||||
"name": "Prometheus",
|
||||
"version": "1.0.0"
|
||||
},
|
||||
{
|
||||
"type": "panel",
|
||||
"id": "table",
|
||||
"name": "Table",
|
||||
"version": ""
|
||||
}
|
||||
],
|
||||
"annotations": {
|
||||
"list": [
|
||||
{
|
||||
"builtIn": 1,
|
||||
"datasource": {
|
||||
"type": "grafana",
|
||||
"uid": "-- Grafana --"
|
||||
},
|
||||
"enable": true,
|
||||
"hide": true,
|
||||
"iconColor": "rgba(0, 211, 255, 1)",
|
||||
"name": "Annotations & Alerts",
|
||||
"type": "dashboard"
|
||||
}
|
||||
]
|
||||
},
|
||||
"editable": true,
|
||||
"fiscalYearStartMonth": 0,
|
||||
"graphTooltip": 0,
|
||||
"id": null,
|
||||
"links": [],
|
||||
"panels": [
|
||||
{
|
||||
"datasource": {
|
||||
"type": "prometheus",
|
||||
"uid": "${DS_PROMETHEUS}"
|
||||
},
|
||||
"fieldConfig": {
|
||||
"defaults": {
|
||||
"color": {
|
||||
"mode": "thresholds"
|
||||
},
|
||||
"custom": {
|
||||
"align": "auto",
|
||||
"cellOptions": {
|
||||
"type": "auto"
|
||||
},
|
||||
"inspect": false
|
||||
},
|
||||
"mappings": [],
|
||||
"thresholds": {
|
||||
"mode": "absolute",
|
||||
"steps": [
|
||||
{
|
||||
"color": "green",
|
||||
"value": null
|
||||
},
|
||||
{
|
||||
"color": "red",
|
||||
"value": 80
|
||||
}
|
||||
]
|
||||
}
|
||||
},
|
||||
"overrides": [
|
||||
{
|
||||
"matcher": {
|
||||
"id": "byName",
|
||||
"options": "last_analyze_timestamp"
|
||||
},
|
||||
"properties": [
|
||||
{
|
||||
"id": "unit",
|
||||
"value": "dateTimeAsIso"
|
||||
}
|
||||
]
|
||||
},
|
||||
{
|
||||
"matcher": {
|
||||
"id": "byName",
|
||||
"options": "last_analyze_since"
|
||||
},
|
||||
"properties": [
|
||||
{
|
||||
"id": "unit",
|
||||
"value": "s"
|
||||
},
|
||||
{
|
||||
"id": "custom.cellOptions",
|
||||
"value": {
|
||||
"mode": "gradient",
|
||||
"type": "color-background"
|
||||
}
|
||||
},
|
||||
{
|
||||
"id": "thresholds",
|
||||
"value": {
|
||||
"mode": "absolute",
|
||||
"steps": [
|
||||
{
|
||||
"color": "green",
|
||||
"value": null
|
||||
},
|
||||
{
|
||||
"color": "red",
|
||||
"value": 259200
|
||||
}
|
||||
]
|
||||
}
|
||||
}
|
||||
]
|
||||
},
|
||||
{
|
||||
"matcher": {
|
||||
"id": "byName",
|
||||
"options": "images_outdated"
|
||||
},
|
||||
"properties": [
|
||||
{
|
||||
"id": "custom.cellOptions",
|
||||
"value": {
|
||||
"mode": "gradient",
|
||||
"type": "color-background"
|
||||
}
|
||||
},
|
||||
{
|
||||
"id": "thresholds",
|
||||
"value": {
|
||||
"mode": "absolute",
|
||||
"steps": [
|
||||
{
|
||||
"color": "green",
|
||||
"value": null
|
||||
},
|
||||
{
|
||||
"color": "red",
|
||||
"value": 1
|
||||
}
|
||||
]
|
||||
}
|
||||
}
|
||||
]
|
||||
},
|
||||
{
|
||||
"matcher": {
|
||||
"id": "byName",
|
||||
"options": "images_error"
|
||||
},
|
||||
"properties": [
|
||||
{
|
||||
"id": "custom.cellOptions",
|
||||
"value": {
|
||||
"mode": "gradient",
|
||||
"type": "color-background"
|
||||
}
|
||||
},
|
||||
{
|
||||
"id": "thresholds",
|
||||
"value": {
|
||||
"mode": "absolute",
|
||||
"steps": [
|
||||
{
|
||||
"color": "green",
|
||||
"value": null
|
||||
},
|
||||
{
|
||||
"color": "red",
|
||||
"value": 1
|
||||
}
|
||||
]
|
||||
}
|
||||
}
|
||||
]
|
||||
}
|
||||
]
|
||||
},
|
||||
"gridPos": {
|
||||
"h": 14,
|
||||
"w": 24,
|
||||
"x": 0,
|
||||
"y": 0
|
||||
},
|
||||
"id": 2,
|
||||
"options": {
|
||||
"cellHeight": "sm",
|
||||
"footer": {
|
||||
"countRows": false,
|
||||
"fields": "",
|
||||
"reducer": [
|
||||
"sum"
|
||||
],
|
||||
"show": false
|
||||
},
|
||||
"frameIndex": 1,
|
||||
"showHeader": true,
|
||||
"sortBy": []
|
||||
},
|
||||
"pluginVersion": "11.4.0",
|
||||
"targets": [
|
||||
{
|
||||
"disableTextWrap": false,
|
||||
"editorMode": "code",
|
||||
"exemplar": false,
|
||||
"expr": "sum by(instance) (dockcheck_images_analyzed)",
|
||||
"format": "table",
|
||||
"fullMetaSearch": false,
|
||||
"hide": false,
|
||||
"includeNullMetadata": true,
|
||||
"instant": true,
|
||||
"interval": "",
|
||||
"legendFormat": "{{instance}}",
|
||||
"range": false,
|
||||
"refId": "dockcheck_images_analyzed",
|
||||
"useBackend": false,
|
||||
"datasource": {
|
||||
"type": "prometheus",
|
||||
"uid": "${DS_PROMETHEUS}"
|
||||
}
|
||||
},
|
||||
{
|
||||
"datasource": {
|
||||
"type": "prometheus",
|
||||
"uid": "${DS_PROMETHEUS}"
|
||||
},
|
||||
"disableTextWrap": false,
|
||||
"editorMode": "code",
|
||||
"exemplar": false,
|
||||
"expr": "sum by(instance) (dockcheck_images_outdated)",
|
||||
"format": "table",
|
||||
"fullMetaSearch": false,
|
||||
"hide": false,
|
||||
"includeNullMetadata": true,
|
||||
"instant": true,
|
||||
"legendFormat": "{{instance}}",
|
||||
"range": false,
|
||||
"refId": "dockcheck_images_outdated",
|
||||
"useBackend": false
|
||||
},
|
||||
{
|
||||
"datasource": {
|
||||
"type": "prometheus",
|
||||
"uid": "${DS_PROMETHEUS}"
|
||||
},
|
||||
"disableTextWrap": false,
|
||||
"editorMode": "code",
|
||||
"exemplar": false,
|
||||
"expr": "sum by(instance) (dockcheck_images_latest)",
|
||||
"format": "table",
|
||||
"fullMetaSearch": false,
|
||||
"hide": false,
|
||||
"includeNullMetadata": true,
|
||||
"instant": true,
|
||||
"legendFormat": "{{instance}}",
|
||||
"range": false,
|
||||
"refId": "dockcheck_images_latest",
|
||||
"useBackend": false
|
||||
},
|
||||
{
|
||||
"datasource": {
|
||||
"type": "prometheus",
|
||||
"uid": "${DS_PROMETHEUS}"
|
||||
},
|
||||
"editorMode": "code",
|
||||
"exemplar": false,
|
||||
"expr": "sum by(instance) (dockcheck_images_error)",
|
||||
"format": "table",
|
||||
"hide": false,
|
||||
"instant": true,
|
||||
"legendFormat": "{{instance}}",
|
||||
"range": false,
|
||||
"refId": "dockcheck_images_error"
|
||||
},
|
||||
{
|
||||
"datasource": {
|
||||
"type": "prometheus",
|
||||
"uid": "${DS_PROMETHEUS}"
|
||||
},
|
||||
"editorMode": "code",
|
||||
"exemplar": false,
|
||||
"expr": "dockcheck_images_analyze_timestamp_seconds * 1000",
|
||||
"format": "table",
|
||||
"hide": false,
|
||||
"instant": true,
|
||||
"legendFormat": "{{instance}}",
|
||||
"range": false,
|
||||
"refId": "dockcheck_images_analyze_timestamp_seconds"
|
||||
},
|
||||
{
|
||||
"datasource": {
|
||||
"type": "prometheus",
|
||||
"uid": "${DS_PROMETHEUS}"
|
||||
},
|
||||
"editorMode": "code",
|
||||
"exemplar": false,
|
||||
"expr": "time() - dockcheck_images_analyze_timestamp_seconds",
|
||||
"format": "table",
|
||||
"hide": false,
|
||||
"instant": true,
|
||||
"legendFormat": "{{instance}}",
|
||||
"range": false,
|
||||
"refId": "dockcheck_images_last_analyze"
|
||||
}
|
||||
],
|
||||
"title": "Dockcheck Status",
|
||||
"transformations": [
|
||||
{
|
||||
"id": "merge",
|
||||
"options": {}
|
||||
},
|
||||
{
|
||||
"id": "organize",
|
||||
"options": {
|
||||
"excludeByName": {
|
||||
"Time": true,
|
||||
"__name__": true,
|
||||
"job": true
|
||||
},
|
||||
"includeByName": {},
|
||||
"indexByName": {
|
||||
"Time": 0,
|
||||
"Value #dockcheck_images_analyze_timestamp_seconds": 2,
|
||||
"Value #dockcheck_images_analyzed": 4,
|
||||
"Value #dockcheck_images_error": 7,
|
||||
"Value #dockcheck_images_last_analyze": 3,
|
||||
"Value #dockcheck_images_latest": 5,
|
||||
"Value #dockcheck_images_outdated": 6,
|
||||
"instance": 1,
|
||||
"job": 8
|
||||
},
|
||||
"renameByName": {
|
||||
"Value #A": "analyze_timestamp",
|
||||
"Value #dockcheck_images_analyze_timestamp_seconds": "last_analyze_timestamp",
|
||||
"Value #dockcheck_images_analyzed": "images_analyzed",
|
||||
"Value #dockcheck_images_error": "images_error",
|
||||
"Value #dockcheck_images_last_analyze": "last_analyze_since",
|
||||
"Value #dockcheck_images_latest": "images_latest",
|
||||
"Value #dockcheck_images_outdated": "images_outdated"
|
||||
}
|
||||
}
|
||||
}
|
||||
],
|
||||
"type": "table"
|
||||
}
|
||||
],
|
||||
"schemaVersion": 40,
|
||||
"tags": [],
|
||||
"templating": {
|
||||
"list": []
|
||||
},
|
||||
"time": {
|
||||
"from": "now-6h",
|
||||
"to": "now"
|
||||
},
|
||||
"timepicker": {},
|
||||
"timezone": "browser",
|
||||
"title": "Dockcheck Status",
|
||||
"uid": "feb4pv3kv1hxca",
|
||||
"version": 17,
|
||||
"weekStart": ""
|
||||
}
|
||||
BIN
addons/prometheus/grafana/grafana_dashboard.png
Normal file
BIN
addons/prometheus/grafana/grafana_dashboard.png
Normal file
Binary file not shown.
|
After Width: | Height: | Size: 50 KiB |
28
addons/prometheus/prometheus_collector.sh
Normal file
28
addons/prometheus/prometheus_collector.sh
Normal file
@@ -0,0 +1,28 @@
|
||||
prometheus_exporter() {
|
||||
checkedImages=$(($1 + $2 + $3))
|
||||
checkTimestamp=$(date +%s)
|
||||
|
||||
promFileContent=()
|
||||
promFileContent+=("# HELP dockcheck_images_analyzed Docker images that have been analyzed")
|
||||
promFileContent+=("# TYPE dockcheck_images_analyzed gauge")
|
||||
promFileContent+=("dockcheck_images_analyzed $checkedImages")
|
||||
|
||||
promFileContent+=("# HELP dockcheck_images_outdated Docker images that are outdated")
|
||||
promFileContent+=("# TYPE dockcheck_images_outdated gauge")
|
||||
promFileContent+=("dockcheck_images_outdated ${#GotUpdates[@]}")
|
||||
|
||||
promFileContent+=("# HELP dockcheck_images_latest Docker images that are outdated")
|
||||
promFileContent+=("# TYPE dockcheck_images_latest gauge")
|
||||
promFileContent+=("dockcheck_images_latest ${#NoUpdates[@]}")
|
||||
|
||||
promFileContent+=("# HELP dockcheck_images_error Docker images with analysis errors")
|
||||
promFileContent+=("# TYPE dockcheck_images_error gauge")
|
||||
promFileContent+=("dockcheck_images_error ${#GotErrors[@]}")
|
||||
|
||||
promFileContent+=("# HELP dockcheck_images_analyze_timestamp_seconds Last dockercheck run time")
|
||||
promFileContent+=("# TYPE dockcheck_images_analyze_timestamp_seconds gauge")
|
||||
promFileContent+=("dockcheck_images_analyze_timestamp_seconds $checkTimestamp")
|
||||
|
||||
printf "%s\n" "${promFileContent[@]}" > "$CollectorTextFileDirectory/dockcheck_info.prom\$\$"
|
||||
mv -f "$CollectorTextFileDirectory/dockcheck_info.prom\$\$" "$CollectorTextFileDirectory/dockcheck_info.prom"
|
||||
}
|
||||
289
dockcheck.sh
289
dockcheck.sh
@@ -1,25 +1,30 @@
|
||||
#!/usr/bin/env bash
|
||||
VERSION="v0.4.9"
|
||||
### ChangeNotes: Added a function to enrich the notify-message with release note URLs. See README.
|
||||
VERSION="v0.5.6.1"
|
||||
### ChangeNotes: Async hotfix, 1 subprocess default, modify MaxAsync variable or pass -x N option to increase.
|
||||
Github="https://github.com/mag37/dockcheck"
|
||||
RawUrl="https://raw.githubusercontent.com/mag37/dockcheck/main/dockcheck.sh"
|
||||
|
||||
### Variables for self updating
|
||||
# Variables for self updating
|
||||
ScriptArgs=( "$@" )
|
||||
ScriptPath="$(readlink -f "$0")"
|
||||
ScriptWorkDir="$(dirname "$ScriptPath")"
|
||||
|
||||
### Check if there's a new release of the script:
|
||||
# Check if there's a new release of the script
|
||||
LatestRelease="$(curl -s -r 0-50 $RawUrl | sed -n "/VERSION/s/VERSION=//p" | tr -d '"')"
|
||||
LatestChanges="$(curl -s -r 0-200 $RawUrl | sed -n "/ChangeNotes/s/### ChangeNotes: //p")"
|
||||
LatestChanges="$(curl -s -r 0-200 $RawUrl | sed -n "/ChangeNotes/s/# ChangeNotes: //p")"
|
||||
|
||||
### Help Function:
|
||||
# User customizable defaults
|
||||
MaxAsync=1
|
||||
Timeout=10
|
||||
|
||||
# Help Function
|
||||
Help() {
|
||||
echo "Syntax: dockcheck.sh [OPTION] [part of name to filter]"
|
||||
echo "Example: dockcheck.sh -y -d 10 -e nextcloud,heimdall"
|
||||
echo
|
||||
echo "Options:"
|
||||
echo "-a|y Automatic updates, without interaction."
|
||||
echo "-c Exports metrics as prom file for the prometheus node_exporter. Provide the collector textfile directory."
|
||||
echo "-d N Only update to new images that are N+ days old. Lists too recent with +prefix and age. 2xSlower."
|
||||
echo "-e X Exclude containers, separated by comma."
|
||||
echo "-f Force stack restart after update. Caution: restarts once for every updated container within stack."
|
||||
@@ -27,17 +32,18 @@ Help() {
|
||||
echo "-i Inform - send a preconfigured notification."
|
||||
echo "-l Only update if label is set. See readme."
|
||||
echo "-m Monochrome mode, no printf color codes."
|
||||
echo "-n No updates, only checking availability."
|
||||
echo "-p Auto-Prune dangling images after update."
|
||||
echo "-r Allow updating images for docker run, wont update the container."
|
||||
echo "-n No updates; only checking availability."
|
||||
echo "-p Auto-prune dangling images after update."
|
||||
echo "-r Allow updating images for docker run; won't update the container."
|
||||
echo "-s Include stopped containers in the check. (Logic: docker ps -a)."
|
||||
echo "-t Set a timeout (in seconds) per container for registry checkups, 10 is default."
|
||||
echo "-v Prints current version."
|
||||
echo "-x N Set max asynchronous subprocesses, 1 default, 0 to disable, 32+ tested."
|
||||
echo
|
||||
echo "Project source: $Github"
|
||||
}
|
||||
|
||||
### Colors:
|
||||
# Colors
|
||||
c_red="\033[0;31m"
|
||||
c_green="\033[0;32m"
|
||||
c_yellow="\033[0;33m"
|
||||
@@ -45,11 +51,12 @@ c_blue="\033[0;34m"
|
||||
c_teal="\033[0;36m"
|
||||
c_reset="\033[0m"
|
||||
|
||||
Timeout=10
|
||||
Stopped=""
|
||||
while getopts "aynpfrhlisvme:d:t:" options; do
|
||||
while getopts "aynpfrhlisvmc:e:d:t:x:" options; do
|
||||
case "${options}" in
|
||||
a|y) AutoUp="yes" ;;
|
||||
c) CollectorTextFileDirectory="${OPTARG}"
|
||||
if ! [[ -d $CollectorTextFileDirectory ]] ; then { printf "The directory (%s) does not exist.\n" "${CollectorTextFileDirectory}" ; exit 2; } fi ;;
|
||||
n) AutoUp="no" ;;
|
||||
r) DRunUp="yes" ;;
|
||||
p) AutoPrune="yes" ;;
|
||||
@@ -61,6 +68,7 @@ while getopts "aynpfrhlisvme:d:t:" options; do
|
||||
s) Stopped="-a" ;;
|
||||
t) Timeout="${OPTARG}" ;;
|
||||
v) printf "%s\n" "$VERSION" ; exit 0 ;;
|
||||
x) MaxAsync=${OPTARG} ;;
|
||||
d) DaysOld=${OPTARG}
|
||||
if ! [[ $DaysOld =~ ^[0-9]+$ ]] ; then { printf "Days -d argument given (%s) is not a number.\n" "${DaysOld}" ; exit 2 ; } ; fi ;;
|
||||
h|*) Help ; exit 2 ;;
|
||||
@@ -68,18 +76,19 @@ while getopts "aynpfrhlisvme:d:t:" options; do
|
||||
done
|
||||
shift "$((OPTIND-1))"
|
||||
|
||||
# Self-update function
|
||||
self_update_curl() {
|
||||
cp "$ScriptPath" "$ScriptPath".bak
|
||||
if [[ $(builtin type -P curl) ]]; then
|
||||
if [[ $(command -v curl) ]]; then
|
||||
curl -L $RawUrl > "$ScriptPath" ; chmod +x "$ScriptPath"
|
||||
printf "\n%s\n" "--- starting over with the updated version ---"
|
||||
exec "$ScriptPath" "${ScriptArgs[@]}" # run the new script with old arguments
|
||||
exit 1 # exit the old instance
|
||||
elif [[ $(builtin type -P wget) ]]; then
|
||||
exit 1 # Exit the old instance
|
||||
elif [[ $(command -v wget) ]]; then
|
||||
wget $RawUrl -O "$ScriptPath" ; chmod +x "$ScriptPath"
|
||||
printf "\n%s\n" "--- starting over with the updated version ---"
|
||||
exec "$ScriptPath" "${ScriptArgs[@]}" # run the new script with old arguments
|
||||
exit 1 # exit the old instance
|
||||
exit 1 # Exit the old instance
|
||||
else
|
||||
printf "curl/wget not available - download the update manually: %s \n" "$Github"
|
||||
fi
|
||||
@@ -87,7 +96,7 @@ self_update_curl() {
|
||||
|
||||
self_update() {
|
||||
cd "$ScriptWorkDir" || { printf "Path error, skipping update.\n" ; return ; }
|
||||
if [[ $(builtin type -P git) ]] && [[ "$(git ls-remote --get-url 2>/dev/null)" =~ .*"mag37/dockcheck".* ]] ; then
|
||||
if [[ $(command -v git) ]] && [[ "$(git ls-remote --get-url 2>/dev/null)" =~ .*"mag37/dockcheck".* ]] ; then
|
||||
printf "\n%s\n" "Pulling the latest version."
|
||||
git pull --force || { printf "Git error, manually pull/clone.\n" ; return ; }
|
||||
printf "\n%s\n" "--- starting over with the updated version ---"
|
||||
@@ -100,7 +109,7 @@ self_update() {
|
||||
fi
|
||||
}
|
||||
|
||||
### Choose from list -function:
|
||||
# Choose from list function
|
||||
choosecontainers() {
|
||||
while [[ -z "$ChoiceClean" ]]; do
|
||||
read -r -p "Enter number(s) separated by comma, [a] for all - [q] to quit: " Choice
|
||||
@@ -112,7 +121,7 @@ choosecontainers() {
|
||||
else
|
||||
ChoiceClean=${Choice//[,.:;]/ }
|
||||
for CC in $ChoiceClean ; do
|
||||
if [[ "$CC" -lt 1 || "$CC" -gt $UpdCount ]] ; then # reset choice if out of bounds
|
||||
if [[ "$CC" -lt 1 || "$CC" -gt $UpdCount ]] ; then # Reset choice if out of bounds
|
||||
echo "Number not in list: $CC" ; unset ChoiceClean ; break 1
|
||||
else
|
||||
SelectedUpdates+=( "${GotUpdates[$CC-1]}" )
|
||||
@@ -126,8 +135,9 @@ choosecontainers() {
|
||||
}
|
||||
|
||||
datecheck() {
|
||||
ImageDate=$($regbin image inspect "$RepoUrl" --format='{{.Created}}' | cut -d" " -f1 )
|
||||
ImageAge=$(( ( $(date +%s) - $(date -d "$ImageDate" +%s) )/86400 ))
|
||||
ImageDate=$($regbin -v error image inspect "$RepoUrl" --format='{{.Created}}' | cut -d" " -f1 )
|
||||
ImageEpoch=$(date -d "$ImageDate" +%s 2>/dev/null) || ImageEpoch=$(date -f "%Y-%m-%d" -j "$ImageDate" +%s)
|
||||
ImageAge=$(( ( $(date +%s) - $ImageEpoch )/86400 ))
|
||||
if [ "$ImageAge" -gt "$DaysOld" ] ; then
|
||||
return 0
|
||||
else
|
||||
@@ -139,26 +149,26 @@ progress_bar() {
|
||||
QueCurrent="$1"
|
||||
QueTotal="$2"
|
||||
((Percent=100*QueCurrent/QueTotal))
|
||||
((Complete=50*Percent/100)) # change first number for width (50)
|
||||
((Left=50-Complete)) # change first number for width (50)
|
||||
((Complete=50*Percent/100)) # Change first number for width (50)
|
||||
((Left=50-Complete)) # Change first number for width (50)
|
||||
BarComplete=$(printf "%${Complete}s" | tr " " "#")
|
||||
BarLeft=$(printf "%${Left}s" | tr " " "-")
|
||||
[[ "$QueTotal" == "$QueCurrent" ]] || printf "\r[%s%s] %s/%s " "$BarComplete" "$BarLeft" "$QueCurrent" "$QueTotal"
|
||||
[[ "$QueTotal" == "$QueCurrent" ]] && printf "\r[%b%s%b] %s/%s \n" "$c_teal" "$BarComplete" "$c_reset" "$QueCurrent" "$QueTotal"
|
||||
}
|
||||
|
||||
### Function to add user-provided urls to releasenotes
|
||||
releasenotes() {
|
||||
for update in ${Updates[@]}; do
|
||||
# Function to add user-provided urls to releasenotes
|
||||
releasenotes() {
|
||||
for update in ${GotUpdates[@]}; do
|
||||
found=false
|
||||
while read -r container url; do
|
||||
[[ $update == $container ]] && printf "%s -> %s\n" "$update" "$url" && found=true
|
||||
[[ $update == $container ]] && Updates+=("$update -> $url") && found=true
|
||||
done < "$ScriptWorkDir"/urls.list
|
||||
[[ $found == false ]] && printf "%s -> url missing\n" "$update" || continue
|
||||
[[ $found == false ]] && Updates+=("$update -> url missing") || continue
|
||||
done
|
||||
}
|
||||
|
||||
### Version check & initiate self update
|
||||
# Version check & initiate self update
|
||||
if [[ "$VERSION" != "$LatestRelease" ]] ; then
|
||||
printf "New version available! %b%s%b ⇒ %b%s%b \n Change Notes: %s \n" "$c_yellow" "$VERSION" "$c_reset" "$c_green" "$LatestRelease" "$c_reset" "$LatestChanges"
|
||||
if [[ -z "$AutoUp" ]] ; then
|
||||
@@ -167,38 +177,75 @@ if [[ "$VERSION" != "$LatestRelease" ]] ; then
|
||||
fi
|
||||
fi
|
||||
|
||||
### Set $1 to a variable for name filtering later.
|
||||
# Set $1 to a variable for name filtering later
|
||||
SearchName="$1"
|
||||
### Create array of excludes
|
||||
# Create array of excludes
|
||||
IFS=',' read -r -a Excludes <<< "$Exclude" ; unset IFS
|
||||
|
||||
### Check if required binary exists in PATH or directory:
|
||||
if [[ $(builtin type -P "regctl") ]]; then regbin="regctl" ;
|
||||
elif [[ -f "$ScriptWorkDir/regctl" ]]; then regbin="$ScriptWorkDir/regctl" ;
|
||||
else
|
||||
read -r -p "Required dependency 'regctl' missing, do you want it downloaded? y/[n] " GetDep
|
||||
if [[ "$GetDep" =~ [yY] ]] ; then
|
||||
### Check arch:
|
||||
case "$(uname --machine)" in
|
||||
x86_64|amd64) architecture="amd64" ;;
|
||||
arm64|aarch64) architecture="arm64";;
|
||||
*) echo "Architecture not supported, exiting." ; exit 1;;
|
||||
esac
|
||||
RegUrl="https://github.com/regclient/regclient/releases/latest/download/regctl-linux-$architecture"
|
||||
if [[ $(builtin type -P curl) ]]; then curl -L $RegUrl > "$ScriptWorkDir/regctl" ; chmod +x "$ScriptWorkDir/regctl" ; regbin="$ScriptWorkDir/regctl" ;
|
||||
elif [[ $(builtin type -P wget) ]]; then wget $RegUrl -O "$ScriptWorkDir/regctl" ; chmod +x "$ScriptWorkDir/regctl" ; regbin="$ScriptWorkDir/regctl" ;
|
||||
else
|
||||
printf "%s\n" "curl/wget not available - get regctl manually from the repo link, quitting."
|
||||
fi
|
||||
else
|
||||
printf "%s\n" "Dependency missing, quitting."
|
||||
exit 1
|
||||
# Static binary downloader for dependencies
|
||||
binary_downloader() {
|
||||
BinaryName="$1"
|
||||
BinaryUrl="$2"
|
||||
case "$(uname -m)" in
|
||||
x86_64|amd64) architecture="amd64" ;;
|
||||
arm64|aarch64) architecture="arm64";;
|
||||
*) printf "\n%bArchitecture not supported, exiting.%b\n" "$c_red" "$c_reset" ; exit 1;;
|
||||
esac
|
||||
GetUrl="${BinaryUrl/TEMP/"$architecture"}"
|
||||
if [[ $(command -v curl) ]]; then curl -L $GetUrl > "$ScriptWorkDir/$BinaryName" ;
|
||||
elif [[ $(command -v wget) ]]; then wget $GetUrl -O "$ScriptWorkDir/$BinaryName" ;
|
||||
else printf "%s\n" "curl/wget not available - get $BinaryName manually from the repo link, exiting."; exit 1;
|
||||
fi
|
||||
fi
|
||||
### final check if binary is correct
|
||||
$regbin version &> /dev/null || { printf "%s\n" "regctl is not working - try to remove it and re-download it, exiting."; exit 1; }
|
||||
[[ -f "$ScriptWorkDir/$BinaryName" ]] && chmod +x "$ScriptWorkDir/$BinaryName"
|
||||
}
|
||||
|
||||
### Check docker compose binary:
|
||||
distro_checker() {
|
||||
if [[ -f /etc/arch-release ]] ; then PkgInstaller="pacman -S"
|
||||
elif [[ -f /etc/redhat-release ]] ; then PkgInstaller="sudo dnf install"
|
||||
elif [[ -f /etc/SuSE-release ]] ; then PkgInstaller="sudo zypper install"
|
||||
elif [[ -f /etc/debian_version ]] ; then PkgInstaller="sudo apt-get install"
|
||||
elif [[ $(uname -s) == "Darwin" ]] ; then PkgInstaller="brew install"
|
||||
else PkgInstaller="ERROR" ; printf "\n%bNo distribution could be determined%b, falling back to static binary.\n" "$c_yellow" "$c_reset"
|
||||
fi
|
||||
}
|
||||
|
||||
# Dependency check + installer function
|
||||
dependency_check() {
|
||||
AppName="$1"
|
||||
AppVar="$2"
|
||||
AppUrl="$3"
|
||||
if [[ $(command -v $AppName) ]]; then export $AppVar="$AppName" ;
|
||||
elif [[ -f "$ScriptWorkDir/$AppName" ]]; then export $AppVar="$ScriptWorkDir/$AppName" ;
|
||||
else
|
||||
printf "%s\n" "Required dependency '$AppName' missing, do you want to install it?"
|
||||
read -r -p "y: With packagemanager (sudo). / s: Download static binary. y/s/[n] " GetBin
|
||||
GetBin=${GetBin:-no} # set default to no if nothing is given
|
||||
if [[ "$GetBin" =~ [yYsS] ]] ; then
|
||||
[[ "$GetBin" =~ [yY] ]] && distro_checker
|
||||
if [[ -n "$PkgInstaller" && "$PkgInstaller" != "ERROR" ]] ; then
|
||||
[[ $(uname -s) == "Darwin" && "$AppName" == "regctl" ]] && AppName="regclient"
|
||||
($PkgInstaller $AppName) ; PkgExitcode="$?" && AppName="$1"
|
||||
if [[ "$PkgExitcode" == 0 ]] ; then { export $AppVar="$AppName" && printf "\n%b$AppName installed.%b\n" "$c_green" "$c_reset"; }
|
||||
else printf "\n%bPackagemanager install failed%b, falling back to static binary.\n" "$c_yellow" "$c_reset"
|
||||
fi
|
||||
fi
|
||||
if [[ "$GetBin" =~ [sS] || "$PkgInstaller" == "ERROR" || "$PkgExitcode" != 0 ]] ; then
|
||||
binary_downloader "$AppName" "$AppUrl"
|
||||
[[ -f "$ScriptWorkDir/$AppName" ]] && { export $AppVar="$ScriptWorkDir/$1" && printf "\n%b$AppName downloaded.%b\n" "$c_green" "$c_reset"; }
|
||||
fi
|
||||
else printf "\n%bDependency missing, exiting.%b\n" "$c_red" "$c_reset" ; exit 1 ;
|
||||
fi
|
||||
fi
|
||||
# Final check if binary is correct
|
||||
[[ "$1" == "jq" ]] && VerFlag="--version"
|
||||
[[ "$1" == "regctl" ]] && VerFlag="version"
|
||||
${!AppVar} $VerFlag &> /dev/null || { printf "%s\n" "$AppName is not working - try to remove it and re-download it, exiting."; exit 1; }
|
||||
}
|
||||
|
||||
dependency_check "regctl" "regbin" "https://github.com/regclient/regclient/releases/latest/download/regctl-linux-TEMP"
|
||||
dependency_check "jq" "jqbin" "https://github.com/jqlang/jq/releases/latest/download/jq-linux-TEMP"
|
||||
|
||||
# Check docker compose binary
|
||||
if docker compose version &> /dev/null ; then DockerBin="docker compose" ;
|
||||
elif docker-compose -v &> /dev/null; then DockerBin="docker-compose" ;
|
||||
elif docker -v &> /dev/null; then
|
||||
@@ -209,7 +256,7 @@ else
|
||||
exit 1
|
||||
fi
|
||||
|
||||
### Numbered List -function:
|
||||
# Numbered List function
|
||||
options() {
|
||||
num=1
|
||||
for i in "${GotUpdates[@]}"; do
|
||||
@@ -218,7 +265,7 @@ for i in "${GotUpdates[@]}"; do
|
||||
done
|
||||
}
|
||||
|
||||
### Listing typed exclusions:
|
||||
# Listing typed exclusions
|
||||
if [[ -n ${Excludes[*]} ]] ; then
|
||||
printf "\n%bExcluding these names:%b\n" "$c_blue" "$c_reset"
|
||||
printf "%s\n" "${Excludes[@]}"
|
||||
@@ -226,11 +273,11 @@ if [[ -n ${Excludes[*]} ]] ; then
|
||||
fi
|
||||
|
||||
# Variables for progress_bar function
|
||||
DocCount=$(docker ps $Stopped --filter "name=$SearchName" --format '{{.Names}}' | wc -l)
|
||||
ContCount=$(docker ps $Stopped --filter "name=$SearchName" --format '{{.Names}}' | wc -l)
|
||||
RegCheckQue=0
|
||||
|
||||
### Testing and setting timeout binary
|
||||
t_out=$(type -P "timeout")
|
||||
# Testing and setting timeout binary
|
||||
t_out=$(command -v timeout)
|
||||
if [[ $t_out ]]; then
|
||||
t_out=$(realpath $t_out 2>/dev/null || readlink -f $t_out)
|
||||
if [[ $t_out =~ "busybox" ]]; then
|
||||
@@ -240,47 +287,92 @@ if [[ $t_out ]]; then
|
||||
else t_out=""
|
||||
fi
|
||||
|
||||
### Check the image-hash of every running container VS the registry
|
||||
for i in $(docker ps $Stopped --filter "name=$SearchName" --format '{{.Names}}') ; do
|
||||
((RegCheckQue+=1))
|
||||
progress_bar "$RegCheckQue" "$DocCount"
|
||||
### Looping every item over the list of excluded names and skipping:
|
||||
for e in "${Excludes[@]}" ; do [[ "$i" == "$e" ]] && continue 2 ; done
|
||||
check_image() {
|
||||
i="$1"
|
||||
local Excludes=($Excludes_string)
|
||||
for e in "${Excludes[@]}" ; do
|
||||
if [[ "$i" == "$e" ]]; then
|
||||
echo Skip $i
|
||||
return
|
||||
fi
|
||||
done
|
||||
|
||||
local NoUpdates GotUpdates GotErrors
|
||||
ImageId=$(docker inspect "$i" --format='{{.Image}}')
|
||||
RepoUrl=$(docker inspect "$i" --format='{{.Config.Image}}')
|
||||
LocalHash=$(docker image inspect "$RepoUrl" --format '{{.RepoDigests}}')
|
||||
# Checking for errors while setting the variable:
|
||||
if RegHash=$(${t_out} $regbin image digest --list "$RepoUrl" 2>&1) ; then
|
||||
LocalHash=$(docker image inspect "$ImageId" --format '{{.RepoDigests}}')
|
||||
|
||||
# Checking for errors while setting the variable
|
||||
if RegHash=$(${t_out} $regbin -v error image digest --list "$RepoUrl" 2>&1) ; then
|
||||
if [[ "$LocalHash" = *"$RegHash"* ]] ; then
|
||||
NoUpdates+=("$i")
|
||||
echo NoUpdates "$i"
|
||||
else
|
||||
if [[ -n "$DaysOld" ]] && ! datecheck ; then
|
||||
NoUpdates+=("+$i ${ImageAge}d")
|
||||
echo NoUpdates "+$i ${ImageAge}d"
|
||||
else
|
||||
GotUpdates+=("$i")
|
||||
echo GotUpdates "$i"
|
||||
fi
|
||||
fi
|
||||
else
|
||||
# Here the RegHash is the result of an error code.
|
||||
GotErrors+=("$i - ${RegHash}")
|
||||
# Here the RegHash is the result of an error code
|
||||
echo GotErrors "$i - ${RegHash}"
|
||||
fi
|
||||
done
|
||||
}
|
||||
|
||||
### Sort arrays alphabetically
|
||||
# Make required functions and variables available to subprocesses
|
||||
export -f check_image datecheck
|
||||
export Excludes_string="${Excludes[@]}" # Can only export scalar variables
|
||||
export t_out regbin RepoUrl DaysOld
|
||||
|
||||
# Check for POSIX xargs with -P option, fallback without async
|
||||
if (echo "test" | xargs -P 2 >/dev/null 2>&1) && [[ "$MaxAsync" != 0 ]]; then
|
||||
XargsAsync="-P $MaxAsync"
|
||||
else
|
||||
XargsAsync=""
|
||||
[[ "$MaxAsync" != 0 ]] && printf "%bMissing POSIX xargs, consider installing 'findutils' for asynchronous lookups.%b\n" "$c_red" "$c_reset"
|
||||
fi
|
||||
|
||||
# Asynchronously check the image-hash of every running container VS the registry
|
||||
while read -r line; do
|
||||
((RegCheckQue+=1))
|
||||
progress_bar "$RegCheckQue" "$ContCount"
|
||||
|
||||
Got=${line%% *} # Extracts the first word (NoUpdates, GotUpdates, GotErrors)
|
||||
item=${line#* }
|
||||
|
||||
case "$Got" in
|
||||
NoUpdates) NoUpdates+=("$item") ;;
|
||||
GotUpdates) GotUpdates+=("$item") ;;
|
||||
GotErrors) GotErrors+=("$item") ;;
|
||||
Skip) ;;
|
||||
*) echo "Error! Unexpected output from subprocess: ${line}" ;;
|
||||
esac
|
||||
done < <( \
|
||||
docker ps $Stopped --filter "name=$SearchName" --format '{{.Names}}' | \
|
||||
xargs ${XargsAsync} -I {} bash -c 'check_image "{}"' \
|
||||
)
|
||||
|
||||
# Sort arrays alphabetically
|
||||
IFS=$'\n'
|
||||
NoUpdates=($(sort <<<"${NoUpdates[*]}"))
|
||||
GotUpdates=($(sort <<<"${GotUpdates[*]}"))
|
||||
unset IFS
|
||||
|
||||
### Define how many updates are available
|
||||
# Run the prometheus exporter function
|
||||
if [ -n "$CollectorTextFileDirectory" ] ; then
|
||||
source "$ScriptWorkDir"/addons/prometheus/prometheus_collector.sh && prometheus_exporter ${#NoUpdates[@]} ${#GotUpdates[@]} ${#GotErrors[@]}
|
||||
fi
|
||||
|
||||
# Define how many updates are available
|
||||
UpdCount="${#GotUpdates[@]}"
|
||||
|
||||
### List what containers got updates or not
|
||||
# List what containers got updates or not
|
||||
if [[ -n ${NoUpdates[*]} ]] ; then
|
||||
printf "\n%bContainers on latest version:%b\n" "$c_green" "$c_reset"
|
||||
printf "%s\n" "${NoUpdates[@]}"
|
||||
fi
|
||||
if [[ -n ${GotErrors[*]} ]] ; then
|
||||
printf "\n%bContainers with errors, wont get updated:%b\n" "$c_red" "$c_reset"
|
||||
printf "\n%bContainers with errors, won't get updated:%b\n" "$c_red" "$c_reset"
|
||||
printf "%s\n" "${GotErrors[@]}"
|
||||
printf "%binfo:%b 'unauthorized' often means not found in a public registry.\n" "$c_blue" "$c_reset"
|
||||
fi
|
||||
@@ -290,11 +382,11 @@ if [[ -n ${GotUpdates[*]} ]] ; then
|
||||
[[ -n "$Notify" ]] && { [[ $(type -t send_notification) == function ]] && send_notification "${GotUpdates[@]}" || printf "Could not source notification function.\n" ; }
|
||||
fi
|
||||
|
||||
### Optionally get updates if there's any
|
||||
# Optionally get updates if there's any
|
||||
if [ -n "$GotUpdates" ] ; then
|
||||
if [ -z "$AutoUp" ] ; then
|
||||
printf "\n%bChoose what containers to update.%b\n" "$c_teal" "$c_reset"
|
||||
choosecontainers
|
||||
printf "\n%bChoose what containers to update.%b\n" "$c_teal" "$c_reset"
|
||||
choosecontainers
|
||||
else
|
||||
SelectedUpdates=( "${GotUpdates[@]}" )
|
||||
fi
|
||||
@@ -305,14 +397,23 @@ if [ -n "$GotUpdates" ] ; then
|
||||
do
|
||||
((CurrentQue+=1))
|
||||
unset CompleteConfs
|
||||
ContPath=$(docker inspect "$i" --format '{{ index .Config.Labels "com.docker.compose.project.working_dir" }}')
|
||||
ContConfigFile=$(docker inspect "$i" --format '{{ index .Config.Labels "com.docker.compose.project.config_files" }}')
|
||||
ContName=$(docker inspect "$i" --format '{{ index .Config.Labels "com.docker.compose.service" }}')
|
||||
ContEnv=$(docker inspect "$i" --format '{{index .Config.Labels "com.docker.compose.project.environment_file" }}')
|
||||
# Extract labels and metadata
|
||||
ContLabels=$(docker inspect "$i" --format '{{json .Config.Labels}}')
|
||||
ContImage=$(docker inspect "$i" --format='{{.Config.Image}}')
|
||||
ContUpdateLabel=$(docker inspect "$i" --format '{{ index .Config.Labels "mag37.dockcheck.update" }}')
|
||||
ContRestartStack=$(docker inspect "$i" --format '{{ index .Config.Labels "mag37.dockcheck.restart-stack" }}')
|
||||
### Checking if compose-values are empty - hence started with docker run:
|
||||
ContPath=$($jqbin -r '."com.docker.compose.project.working_dir"' <<< "$ContLabels")
|
||||
[ "$ContPath" == "null" ] && ContPath=""
|
||||
ContConfigFile=$($jqbin -r '."com.docker.compose.project.config_files"' <<< "$ContLabels")
|
||||
[ "$ContConfigFile" == "null" ] && ContConfigFile=""
|
||||
ContName=$($jqbin -r '."com.docker.compose.service"' <<< "$ContLabels")
|
||||
[ "$ContName" == "null" ] && ContName=""
|
||||
ContEnv=$($jqbin -r '."com.docker.compose.project.environment_file"' <<< "$ContLabels")
|
||||
[ "$ContEnv" == "null" ] && ContEnv=""
|
||||
ContUpdateLabel=$($jqbin -r '."mag37.dockcheck.update"' <<< "$ContLabels")
|
||||
[ "$ContUpdateLabel" == "null" ] && ContUpdateLabel=""
|
||||
ContRestartStack=$($jqbin -r '."mag37.dockcheck.restart-stack"' <<< "$ContLabels")
|
||||
[ "$ContRestartStack" == "null" ] && ContRestartStack=""
|
||||
|
||||
# Checking if compose-values are empty - hence started with docker run
|
||||
if [ -z "$ContPath" ] ; then
|
||||
if [ "$DRunUp" == "yes" ] ; then
|
||||
docker pull "$ContImage"
|
||||
@@ -322,7 +423,7 @@ if [ -n "$GotUpdates" ] ; then
|
||||
fi
|
||||
continue
|
||||
fi
|
||||
### cd to the compose-file directory to account for people who use relative volumes, eg - ${PWD}/data:data
|
||||
# cd to the compose-file directory to account for people who use relative volumes
|
||||
cd "$ContPath" || { echo "Path error - skipping $i" ; continue ; }
|
||||
## Reformatting path + multi compose
|
||||
if [[ $ContConfigFile = '/'* ]] ; then
|
||||
@@ -331,12 +432,12 @@ if [ -n "$GotUpdates" ] ; then
|
||||
CompleteConfs=$(for conf in ${ContConfigFile//,/ } ; do printf -- "-f %s/%s " "$ContPath" "$conf"; done)
|
||||
fi
|
||||
printf "\n%bNow updating (%s/%s): %b%s%b\n" "$c_teal" "$CurrentQue" "$NumberofUpdates" "$c_blue" "$i" "$c_reset"
|
||||
### Checking if Label Only -option is set, and if container got the label
|
||||
# Checking if Label Only -option is set, and if container got the label
|
||||
[[ "$OnlyLabel" == true ]] && { [[ "$ContUpdateLabel" != true ]] && { echo "No update label, skipping." ; continue ; } }
|
||||
docker pull "$ContImage"
|
||||
### Check if the container got an environment file set and reformat it
|
||||
# Check if the container got an environment file set and reformat it
|
||||
if [ -n "$ContEnv" ]; then ContEnvs=$(for env in ${ContEnv//,/ } ; do printf -- "--env-file %s " "$env"; done) ; fi
|
||||
### Check if the whole stack should be restarted
|
||||
# Check if the whole stack should be restarted
|
||||
if [[ "$ContRestartStack" == true ]] || [[ "$ForceRestartStacks" == true ]] ; then
|
||||
$DockerBin ${CompleteConfs} stop ; $DockerBin ${CompleteConfs} ${ContEnvs} up -d
|
||||
else
|
||||
|
||||
@@ -2,12 +2,21 @@
|
||||
SearchName="$1"
|
||||
for i in $(docker ps --filter "name=$SearchName" --format '{{.Names}}') ; do
|
||||
echo "------------ $i ------------"
|
||||
ContPath=$(docker inspect "$i" --format '{{ index .Config.Labels "com.docker.compose.project.working_dir" }}')
|
||||
[ -z "$ContPath" ] && { "$i has no compose labels - skipping" ; continue ; }
|
||||
ContConfigFile=$(docker inspect "$i" --format '{{ index .Config.Labels "com.docker.compose.project.config_files" }}')
|
||||
ContName=$(docker inspect "$i" --format '{{ index .Config.Labels "com.docker.compose.service" }}')
|
||||
ContEnv=$(docker inspect "$i" --format '{{index .Config.Labels "com.docker.compose.project.environment_file" }}')
|
||||
ContLabels=$(docker inspect "$i" --format '{{json .Config.Labels}}')
|
||||
ContImage=$(docker inspect "$i" --format='{{.Config.Image}}')
|
||||
ContPath=$(jq -r '."com.docker.compose.project.working_dir"' <<< "$ContLabels")
|
||||
[ "$ContPath" == "null" ] && ContPath=""
|
||||
[ -z "$ContPath" ] && { "$i has no compose labels - skipping" ; continue ; }
|
||||
ContConfigFile=$(jq -r '."com.docker.compose.project.config_files"' <<< "$ContLabels")
|
||||
[ "$ContConfigFile" == "null" ] && ContConfigFile=""
|
||||
ContName=$(jq -r '."com.docker.compose.service"' <<< "$ContLabels")
|
||||
[ "$ContName" == "null" ] && ContName=""
|
||||
ContEnv=$(jq -r '."com.docker.compose.project.environment_file"' <<< "$ContLabels")
|
||||
[ "$ContEnv" == "null" ] && ContEnv=""
|
||||
ContUpdateLabel=$(jq -r '."mag37.dockcheck.update"' <<< "$ContLabels")
|
||||
[ "$ContUpdateLabel" == "null" ] && ContUpdateLabel=""
|
||||
ContRestartStack=$(jq -r '."mag37.dockcheck.restart-stack"' <<< "$ContLabels")
|
||||
[ "$ContRestartStack" == "null" ] && ContRestartStack=""
|
||||
|
||||
if [[ $ContConfigFile = '/'* ]] ; then
|
||||
ComposeFile="$ContConfigFile"
|
||||
@@ -20,6 +29,8 @@ for i in $(docker ps --filter "name=$SearchName" --format '{{.Names}}') ; do
|
||||
echo -e "Compose files:\t\t$ComposeFile"
|
||||
echo -e "Environment files:\t$ContEnv"
|
||||
echo -e "Container image:\t$ContImage"
|
||||
echo -e "Update label:\t$ContUpdateLabel"
|
||||
echo -e "Restart Stack label:\t$ContRestartStack"
|
||||
echo
|
||||
echo "Mounts:"
|
||||
docker inspect -f '{{ range .Mounts }}{{ .Source }}:{{ .Destination }}{{ printf "\n" }}{{ end }}' "$i"
|
||||
|
||||
Binary file not shown.
|
Before Width: | Height: | Size: 404 KiB After Width: | Height: | Size: 472 KiB |
BIN
extras/example_old.gif
Normal file
BIN
extras/example_old.gif
Normal file
Binary file not shown.
|
After Width: | Height: | Size: 404 KiB |
@@ -9,7 +9,7 @@
|
||||
MSMTP=$(which msmtp)
|
||||
SSMTP=$(which ssmtp)
|
||||
|
||||
if [ -n "$MSMPT" ] ; then
|
||||
if [ -n "$MSMTP" ] ; then
|
||||
MailPkg=$MSMTP
|
||||
elif [ -n "$SSMTP" ] ; then
|
||||
MailPkg=$SSMTP
|
||||
@@ -18,8 +18,8 @@ else
|
||||
fi
|
||||
|
||||
send_notification() {
|
||||
Updates=("$@")
|
||||
[ -s "$ScriptWorkDir"/urls.list ] && UpdToString=$( releasenotes ) || UpdToString=$( printf "%s\n" "${Updates[@]}" )
|
||||
[ -s "$ScriptWorkDir"/urls.list ] && releasenotes || Updates=("$@")
|
||||
UpdToString=$( printf '%s\\n' "${Updates[@]}" )
|
||||
FromHost=$(hostname)
|
||||
CfgFile="/usr/syno/etc/synosmtp.conf"
|
||||
|
||||
@@ -36,6 +36,8 @@ SenderMail=${SenderMail:-$(grep 'eventmail1' $CfgFile | sed -n 's/.*"\([^"]*\)".
|
||||
|
||||
printf "\nSending email notification.\n"
|
||||
|
||||
printf -v MessageBody "🐋 Containers on $FromHost with updates available:\n\n$UpdToString"
|
||||
|
||||
$MailPkg $SendMailTo << __EOF
|
||||
From: "$SenderName" <$SenderMail>
|
||||
date:$(date -R)
|
||||
@@ -44,10 +46,9 @@ Subject: $SubjectTag Updates available on $FromHost
|
||||
Content-Type: text/plain; charset=UTF-8; format=flowed
|
||||
Content-Transfer-Encoding: 7bit
|
||||
|
||||
The following containers on $FromHost have updates available:
|
||||
|
||||
$UpdToString
|
||||
|
||||
$MessageBody
|
||||
From $SenderName
|
||||
__EOF
|
||||
# This ensures DSM's container manager will also see the update
|
||||
/var/packages/ContainerManager/target/tool/image_upgradable_checker
|
||||
}
|
||||
|
||||
@@ -5,20 +5,15 @@
|
||||
# Modify to fit your setup - if API, set AppriseURL to your Apprise ip/domain.
|
||||
|
||||
send_notification() {
|
||||
Updates=("$@")
|
||||
[ -s "$ScriptWorkDir"/urls.list ] && UpdToString=$( releasenotes ) || UpdToString=$( printf "%s\n" "${Updates[@]}" )
|
||||
[ -s "$ScriptWorkDir"/urls.list ] && releasenotes || Updates=("$@")
|
||||
UpdToString=$( printf '%s\\n' "${Updates[@]}" )
|
||||
FromHost=$(hostname)
|
||||
|
||||
printf "\nSending Apprise notification\n"
|
||||
|
||||
MessageTitle="$FromHost - updates available."
|
||||
# Setting the MessageBody variable here.
|
||||
read -d '\n' MessageBody << __EOF
|
||||
Containers on $FromHost with updates available:
|
||||
|
||||
$UpdToString
|
||||
|
||||
__EOF
|
||||
printf -v MessageBody "🐋 Containers on $FromHost with updates available:\n$UpdToString"
|
||||
|
||||
# Modify to fit your setup:
|
||||
apprise -vv -t "$MessageTitle" -b "$MessageBody" \
|
||||
|
||||
28
notify_templates/notify_discord.sh
Normal file
28
notify_templates/notify_discord.sh
Normal file
@@ -0,0 +1,28 @@
|
||||
### DISCLAIMER: This is a third party addition to dockcheck - best effort testing.
|
||||
#
|
||||
# Copy/rename this file to notify.sh to enable the notification snippet.
|
||||
# Required receiving services must already be set up.
|
||||
# Modify to fit your setup - set DiscordWebhookUrl
|
||||
|
||||
send_notification() {
|
||||
[ -s "$ScriptWorkDir"/urls.list ] && releasenotes || Updates=("$@")
|
||||
UpdToString=$( printf '%s\\n' "${Updates[@]}" )
|
||||
|
||||
echo "$UpdToString"
|
||||
FromHost=$(hostname)
|
||||
|
||||
# platform specific notification code would go here
|
||||
printf "\nSending Discord notification\n"
|
||||
|
||||
# Setting the MessageBody variable here.
|
||||
MessageBody="🐋 Containers on $FromHost with updates available: \n$UpdToString"
|
||||
|
||||
# Modify to fit your setup:
|
||||
DiscordWebhookUrl="PasteYourFullDiscordWebhookURL"
|
||||
|
||||
MsgBody="{\"username\":\"$FromHost\",\"content\":\"$MessageBody\"}"
|
||||
|
||||
curl -sS -o /dev/null --fail -X POST -H "Content-Type: application/json" -d "$MsgBody" "$DiscordWebhookUrl"
|
||||
|
||||
}
|
||||
|
||||
@@ -4,11 +4,13 @@
|
||||
# generic sample, the "Hello World" of notification addons
|
||||
|
||||
send_notification() {
|
||||
Updates=("$@")
|
||||
[ -s "$ScriptWorkDir"/urls.list ] && UpdToString=$( releasenotes ) || UpdToString=$( printf "%s\n" "${Updates[@]}" )
|
||||
[ -s "$ScriptWorkDir"/urls.list ] && releasenotes || Updates=("$@")
|
||||
UpdToString=$( printf '%s\\n' "${Updates[@]}" )
|
||||
|
||||
FromHost=$(hostname)
|
||||
|
||||
# platform specific notification code would go here
|
||||
printf "\n%bGeneric notification addon:%b" "$c_green" "$c_reset"
|
||||
printf "\nThe following docker containers on %s need to be updated:\n%s\n" "$FromHost" "$UpdToString"
|
||||
printf "\nThe following docker containers on %s need to be updated:\n" "$FromHost"
|
||||
printf "$UpdToString"
|
||||
}
|
||||
|
||||
@@ -5,8 +5,8 @@
|
||||
# Modify to fit your setup - set GotifyUrl and GotifyToken.
|
||||
|
||||
send_notification() {
|
||||
Updates=("$@")
|
||||
[ -s "$ScriptWorkDir"/urls.list ] && UpdToString=$( releasenotes ) || UpdToString=$( printf "%s\n" "${Updates[@]}" )
|
||||
[ -s "$ScriptWorkDir"/urls.list ] && releasenotes || Updates=("$@")
|
||||
UpdToString=$( printf '%s\\n' "${Updates[@]}" )
|
||||
FromHost=$(hostname)
|
||||
|
||||
# platform specific notification code would go here
|
||||
@@ -14,7 +14,7 @@ send_notification() {
|
||||
|
||||
# Setting the MessageTitle and MessageBody variable here.
|
||||
MessageTitle="${FromHost} - updates available."
|
||||
MessageBody="Containers on ${FromHost} with updates available: ${UpdToString}"
|
||||
printf -v MessageBody "Containers on $FromHost with updates available:\n$UpdToString"
|
||||
|
||||
# Modify to fit your setup:
|
||||
GotifyToken="Your Gotify token here"
|
||||
@@ -24,6 +24,6 @@ send_notification() {
|
||||
-F "title=${MessageTitle}" \
|
||||
-F "message=${MessageBody}" \
|
||||
-F "priority=5" \
|
||||
-X POST "${GotifyUrl}" &> /dev/null
|
||||
-X POST "${GotifyUrl}" 1> /dev/null
|
||||
|
||||
}
|
||||
|
||||
@@ -5,8 +5,8 @@
|
||||
# Modify to fit your setup - set MatrixServer, Room_id and AccessToken
|
||||
|
||||
send_notification() {
|
||||
Updates=("$@")
|
||||
[ -s "$ScriptWorkDir"/urls.list ] && UpdToString=$( releasenotes ) || UpdToString=$( printf "%s\n" "${Updates[@]}" )
|
||||
[ -s "$ScriptWorkDir"/urls.list ] && releasenotes || Updates=("$@")
|
||||
UpdToString=$( printf '%s\\n' "${Updates[@]}" )
|
||||
FromHost=$(hostname)
|
||||
|
||||
# platform specific notification code would go here
|
||||
|
||||
@@ -5,20 +5,15 @@
|
||||
# Use your unique Topic Name in the URL below.
|
||||
|
||||
send_notification() {
|
||||
Updates=("$@")
|
||||
[ -s "$ScriptWorkDir"/urls.list ] && UpdToString=$( releasenotes ) || UpdToString=$( printf "%s\n" "${Updates[@]}" )
|
||||
[ -s "$ScriptWorkDir"/urls.list ] && releasenotes || Updates=("$@")
|
||||
UpdToString=$( printf '%s\\n' "${Updates[@]}" )
|
||||
FromHost=$(hostname)
|
||||
|
||||
printf "\nSending ntfy.sh notification\n"
|
||||
|
||||
MessageTitle="$FromHost - updates available."
|
||||
# Setting the MessageBody variable here.
|
||||
read -d '\n' MessageBody << __EOF
|
||||
Containers on $FromHost with updates available:
|
||||
|
||||
$UpdToString
|
||||
|
||||
__EOF
|
||||
printf -v MessageBody "🐋 Containers on $FromHost with updates available:\n$UpdToString"
|
||||
|
||||
# Modify to fit your setup:
|
||||
NtfyUrl="ntfy.sh/YourUniqueTopicName"
|
||||
|
||||
@@ -6,8 +6,8 @@
|
||||
# Modify to fit your setup - set Url and Token.
|
||||
|
||||
send_notification() {
|
||||
Updates=("$@")
|
||||
[ -s "$ScriptWorkDir"/urls.list ] && UpdToString=$( releasenotes ) || UpdToString=$( printf "%s\n" "${Updates[@]}" )
|
||||
[ -s "$ScriptWorkDir"/urls.list ] && releasenotes || Updates=("$@")
|
||||
UpdToString=$( printf '%s\\n' "${Updates[@]}" )
|
||||
FromHost=$(hostname)
|
||||
|
||||
# platform specific notification code would go here
|
||||
@@ -15,7 +15,7 @@ printf "\nSending pushbullet notification\n"
|
||||
|
||||
MessageTitle="$FromHost - updates available."
|
||||
# Setting the MessageBody variable here.
|
||||
MessageBody="Containers on $FromHost with updates available: $UpdToString"
|
||||
printf -v MessageBody "🐋 Containers on $FromHost with updates available:\n$UpdToString"
|
||||
|
||||
# Modify to fit your setup:
|
||||
PushUrl="https://api.pushbullet.com/v2/pushes"
|
||||
|
||||
@@ -6,8 +6,8 @@
|
||||
# Modify to fit your setup - set Url and Token.
|
||||
|
||||
send_notification() {
|
||||
Updates=("$@")
|
||||
[ -s "$ScriptWorkDir"/urls.list ] && UpdToString=$( releasenotes ) || UpdToString=$( printf "%s\n" "${Updates[@]}" )
|
||||
[ -s "$ScriptWorkDir"/urls.list ] && releasenotes || Updates=("$@")
|
||||
UpdToString=$( printf '%s\\n' "${Updates[@]}" )
|
||||
FromHost=$(hostname)
|
||||
|
||||
# platform specific notification code would go here
|
||||
@@ -15,7 +15,7 @@ send_notification() {
|
||||
|
||||
MessageTitle="$FromHost - updates available."
|
||||
# Setting the MessageBody variable here.
|
||||
MessageBody="Containers on $FromHost with updates available: $UpdToString"
|
||||
printf -v MessageBody "🐋 Containers on $FromHost with updates available:\n$UpdToString"
|
||||
|
||||
# Modify to fit your setup:
|
||||
PushoverUrl="https://api.pushover.net/1/messages.json"
|
||||
|
||||
@@ -8,7 +8,7 @@
|
||||
MSMTP=$(which msmtp)
|
||||
SSMTP=$(which ssmtp)
|
||||
|
||||
if [ -n "$MSMPT" ] ; then
|
||||
if [ -n "$MSMTP" ] ; then
|
||||
MailPkg=$MSMTP
|
||||
elif [ -n "$SSMTP" ] ; then
|
||||
MailPkg=$SSMTP
|
||||
@@ -17,8 +17,8 @@ else
|
||||
fi
|
||||
|
||||
send_notification() {
|
||||
Updates=("$@")
|
||||
[ -s "$ScriptWorkDir"/urls.list ] && UpdToString=$( releasenotes ) || UpdToString=$( printf "%s\n" "${Updates[@]}" )
|
||||
[ -s "$ScriptWorkDir"/urls.list ] && releasenotes || Updates=("$@")
|
||||
UpdToString=$( printf '%s\\n' "${Updates[@]}" )
|
||||
FromHost=$(hostname)
|
||||
|
||||
# User variables:
|
||||
@@ -28,6 +28,8 @@ SubjectTag="dockcheck"
|
||||
|
||||
printf "\nSending email notification.\n"
|
||||
|
||||
printf -v MessageBody "🐋 Containers on $FromHost with updates available:\n\n$UpdToString"
|
||||
|
||||
$MailPkg $SendMailTo << __EOF
|
||||
From: "$FromHost" <$SendMailFrom>
|
||||
date:$(date -R)
|
||||
@@ -36,9 +38,7 @@ Subject: [$SubjectTag] Updates available on $FromHost
|
||||
Content-Type: text/plain; charset=UTF-8; format=flowed
|
||||
Content-Transfer-Encoding: 7bit
|
||||
|
||||
The following containers on $FromHost have updates available:
|
||||
|
||||
$UpdToString
|
||||
$MessageBody
|
||||
|
||||
__EOF
|
||||
}
|
||||
|
||||
@@ -5,8 +5,8 @@
|
||||
# Modify to fit your setup - set TelegramChatId and TelegramToken.
|
||||
|
||||
send_notification() {
|
||||
Updates=("$@")
|
||||
[ -s "$ScriptWorkDir"/urls.list ] && UpdToString=$( releasenotes ) || UpdToString=$( printf "%s\n" "${Updates[@]}" )
|
||||
[ -s "$ScriptWorkDir"/urls.list ] && releasenotes || Updates=("$@")
|
||||
UpdToString=$( printf '%s\\n' "${Updates[@]}" )
|
||||
FromHost=$(hostname)
|
||||
|
||||
# platform specific notification code would go here
|
||||
|
||||
@@ -6,3 +6,30 @@ apprise-api https://github.com/linuxserver/docker-apprise-api/releases
|
||||
homer https://github.com/bastienwirtz/homer/releases
|
||||
nginx https://github.com/docker-library/official-images/blob/master/library/nginx
|
||||
vaultwarden-server https://github.com/dani-garcia/vaultwarden/releases
|
||||
bruceforce-vaultwarden-backup https://github.com/Bruceforce/vaultwarden-backup/blob/main/CHANGELOG.md
|
||||
actual_server https://actualbudget.org/blog
|
||||
gotify https://github.com/gotify/server/releases
|
||||
traefik https://github.com/traefik/traefik/releases
|
||||
caddy https://github.com/caddyserver/caddy/releases
|
||||
|
||||
calibre https://github.com/linuxserver/docker-calibre/releases
|
||||
calibre-web https://github.com/linuxserver/docker-calibre-web/releases
|
||||
readarr https://github.com/Readarr/Readarr/releases
|
||||
audiobookshelf https://github.com/advplyr/audiobookshelf/releases
|
||||
|
||||
gluetun https://github.com/qdm12/gluetun/releases
|
||||
bazarr https://github.com/morpheus65535/bazarr/releases
|
||||
bazarr-ls https://github.com/linuxserver/docker-bazarr/releases
|
||||
prowlarr https://github.com/Prowlarr/Prowlarr/releases
|
||||
prowlarr-ls https://github.com/linuxserver/docker-prowlarr/releases
|
||||
recyclarr https://github.com/recyclarr/recyclarr/releases
|
||||
sabnzbd https://github.com/linuxserver/docker-sabnzbd/releases
|
||||
sonarr https://github.com/linuxserver/docker-sonarr/releases
|
||||
radarr https://github.com/linuxserver/docker-radarr/releases
|
||||
lidarr https://github.com/linuxserver/docker-lidarr/releases
|
||||
jellyseerr https://github.com/Fallenbagel/jellyseerr/releases
|
||||
jellyfin https://github.com/jellyfin/jellyfin/releases
|
||||
|
||||
home-assistant https://github.com/home-assistant/docker/releases
|
||||
zigbee2mqtt https://github.com/Koenkk/zigbee2mqtt/releases
|
||||
mqtt https://github.com/eclipse/mosquitto/tags
|
||||
|
||||
Reference in New Issue
Block a user