Compare commits
271 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
61f90893ef | ||
|
|
229cde0efb | ||
|
|
0c6674ac8e | ||
|
|
3e079e2ec5 | ||
|
|
b2d67c9f52 | ||
|
|
3aeee837f9 | ||
|
|
408a8b14dd | ||
|
|
a2868ea505 | ||
|
|
62a3d10b4f | ||
|
|
210c076968 | ||
|
|
65e875e860 | ||
|
|
3655f5ae8a | ||
|
|
78a7e1137f | ||
|
|
76e6a5c38b | ||
|
|
41029f628d | ||
|
|
b918844336 | ||
|
|
27896c18ba | ||
|
|
e68adb34d0 | ||
|
|
6bc896b193 | ||
|
|
6cba140522 | ||
|
|
73050abf10 | ||
|
|
d4d89c305c | ||
|
|
bc6a3529c7 | ||
|
|
01b9b33751 | ||
|
|
9ed2a0bad0 | ||
|
|
43307350ca | ||
|
|
b3600f26ac | ||
|
|
79def47754 | ||
|
|
e6ff634394 | ||
|
|
6444e18a4f | ||
|
|
14872b0471 | ||
|
|
07ad241e9f | ||
|
|
58d53d3aaf | ||
|
|
15ce226a0a | ||
|
|
cce5438aca | ||
|
|
be6f5edf52 | ||
|
|
99befd6938 | ||
|
|
8a63fd360c | ||
|
|
d3786d6f75 | ||
|
|
487cfb2822 | ||
|
|
b4943df46c | ||
|
|
8603c8d4b6 | ||
|
|
ae66a6f0fd | ||
|
|
3ac0521a9b | ||
|
|
00ae250511 | ||
|
|
704387a7fe | ||
|
|
1a6826e2ac | ||
|
|
a28b9e555f | ||
|
|
8309b80dc2 | ||
|
|
bbe26a0ac2 | ||
|
|
d98d052af7 | ||
|
|
fbba77dc1f | ||
|
|
09c9faa008 | ||
|
|
17ed46d157 | ||
|
|
d744d51473 | ||
|
|
3821c22660 | ||
|
|
5dd42b7336 | ||
|
|
8da2cbe611 | ||
|
|
9937d6bc3f | ||
|
|
5517abb090 | ||
|
|
cb8c23e62d | ||
|
|
0009adc2cf | ||
|
|
8d39e1b66b | ||
|
|
ec068dd19a | ||
|
|
119932aa19 | ||
|
|
a9f2433fab | ||
|
|
18e467274e | ||
|
|
6db92be627 | ||
|
|
fc47b1313b | ||
|
|
76d2921719 | ||
|
|
2844e54fbd | ||
|
|
e0d66cc773 | ||
|
|
071fa4b245 | ||
|
|
cb73025a61 | ||
|
|
87f1886d28 | ||
|
|
ac10a96cc8 | ||
|
|
e5a7a715c6 | ||
|
|
8a6a1fbec2 | ||
|
|
04b0c46366 | ||
|
|
319b43bc1b | ||
|
|
8b40dce2cd | ||
|
|
62b52796f2 | ||
|
|
85ee721cd1 | ||
|
|
6de00a2b90 | ||
|
|
a550994659 | ||
|
|
0e3210b561 | ||
|
|
db4431d40a | ||
|
|
9f8c62ed96 | ||
|
|
827bb7e14a | ||
|
|
91ff4aaaa7 | ||
|
|
999694985a | ||
|
|
640046668c | ||
|
|
a79de24bcc | ||
|
|
f31a886424 | ||
|
|
5b31e68f73 | ||
|
|
923736c413 | ||
|
|
7d825f36c9 | ||
|
|
29dffbb9b3 | ||
|
|
588c22afc9 | ||
|
|
39c69b7648 | ||
|
|
35e33b3997 | ||
|
|
a615d759ad | ||
|
|
df310c8f50 | ||
|
|
b1448fbb95 | ||
|
|
de67cab998 | ||
|
|
608c218281 | ||
|
|
3d122f2b86 | ||
|
|
8e280145b9 | ||
|
|
da389a04e7 | ||
|
|
25563d16b6 | ||
|
|
4917c97076 | ||
|
|
33fc647cb1 | ||
|
|
abec27c38a | ||
|
|
082b0f4c7c | ||
|
|
8521280070 | ||
|
|
0aac94f27c | ||
|
|
730dba5c3d | ||
|
|
f8455d3787 | ||
|
|
bbe9164554 | ||
|
|
5d86c79710 | ||
|
|
b0f8431dbf | ||
|
|
331d2cc5a7 | ||
|
|
35f460ed61 | ||
|
|
aab63901d0 | ||
|
|
5e53fabbf6 | ||
|
|
8b3926fd5d | ||
|
|
049413cf5a | ||
|
|
b1b7f048a3 | ||
|
|
361e98c378 | ||
|
|
d1ad0c68f6 | ||
|
|
76697bc281 | ||
|
|
86649fb0fe | ||
|
|
80f7dd9b82 | ||
|
|
959f636152 | ||
|
|
d1bf9256fa | ||
|
|
5513e13109 | ||
|
|
f0b2f7eab7 | ||
|
|
44a90d5e62 | ||
|
|
c63643dde8 | ||
|
|
b2767864ab | ||
|
|
c09e813aad | ||
|
|
0b52dd50b6 | ||
|
|
b86df76fb6 | ||
|
|
b8450aa774 | ||
|
|
aa9421adda | ||
|
|
f853bfd318 | ||
|
|
292b3e1a32 | ||
|
|
4c26768374 | ||
|
|
b7f8b67232 | ||
|
|
a63a509bd2 | ||
|
|
5ed3b90ca4 | ||
|
|
78d11d1b37 | ||
|
|
dfe6cf3e08 | ||
|
|
3bc7352b54 | ||
|
|
1470055329 | ||
|
|
904cd508d1 | ||
|
|
8ee1366505 | ||
|
|
c063b09003 | ||
|
|
ea096b5dbc | ||
|
|
038a30b628 | ||
|
|
24d82d74b3 | ||
|
|
8ecc354663 | ||
|
|
4a2daa9899 | ||
|
|
ae8a232a1a | ||
|
|
abbb924867 | ||
|
|
05d21a707a | ||
|
|
4817d75d94 | ||
|
|
32f6d428d3 | ||
|
|
4a8afde790 | ||
|
|
ae53fe2cc7 | ||
|
|
722fb90ce2 | ||
|
|
18947a462d | ||
|
|
0887025412 | ||
|
|
e331dda080 | ||
|
|
5b87072b11 | ||
|
|
6b2af36841 | ||
|
|
bad7bfdf86 | ||
|
|
2a604ff8bd | ||
|
|
e1b26fb772 | ||
|
|
9e4b445c50 | ||
|
|
55323b7562 | ||
|
|
51c8a7a05d | ||
|
|
6a3c78d907 | ||
|
|
4b6173683b | ||
|
|
fb4e9fd521 | ||
|
|
e00bfb1d21 | ||
|
|
77d7340e5b | ||
|
|
82f6442933 | ||
|
|
6c37e255fb | ||
|
|
61e247f0ea | ||
|
|
b540ffa119 | ||
|
|
edf6d5bd72 | ||
|
|
92ba4a184e | ||
|
|
47ebb75691 | ||
|
|
3053304356 | ||
|
|
63b35ab2c3 | ||
|
|
45dbe65c28 | ||
|
|
ae5cd4d86e | ||
|
|
58bb55c79a | ||
|
|
fe31aed33c | ||
|
|
5485a3b2cf | ||
|
|
94fd34478b | ||
|
|
8e92a3a4cb | ||
|
|
9ead0feb9f | ||
|
|
2e620cd2c0 | ||
|
|
fc7b2888f1 | ||
|
|
e15a8e9dbf | ||
|
|
342ff271ca | ||
|
|
3f56af0c8a | ||
|
|
ab3aaf1140 | ||
|
|
80a81f59d6 | ||
|
|
ba3f436ac9 | ||
|
|
24438124eb | ||
|
|
e39ea0748f | ||
|
|
ab30f22343 | ||
|
|
d39badea36 | ||
|
|
9a1f2aa7c7 | ||
|
|
15d3a96b2c | ||
|
|
fbd27d6e37 | ||
|
|
7bd5f87ed1 | ||
|
|
e24cba0078 | ||
|
|
886379dc7d | ||
|
|
488669b99a | ||
|
|
e20fdd9e85 | ||
|
|
3d1c424d23 | ||
|
|
132fda3388 | ||
|
|
cc9b3b0828 | ||
|
|
29c2c5e961 | ||
|
|
5dc9af9874 | ||
|
|
f91ca08d54 | ||
|
|
9a246b52c4 | ||
|
|
0c33f581f7 | ||
|
|
b492aacf0f | ||
|
|
f498138d4b | ||
|
|
6a3198eb08 | ||
|
|
48bfea9fbf | ||
|
|
f2fc5a4d34 | ||
|
|
f0940e81ec | ||
|
|
a25e648557 | ||
|
|
0e2bdc94ad | ||
|
|
19eb57077f | ||
|
|
1f28c9c144 | ||
|
|
0d6da201b7 | ||
|
|
ecb1256fbb | ||
|
|
afe97c9de5 | ||
|
|
9b6396f6de | ||
|
|
5e58abd8e9 | ||
|
|
3d7340c91c | ||
|
|
1decbd77b5 | ||
|
|
bcc0c63413 | ||
|
|
4a2bc2dded | ||
|
|
fd67e0dff1 | ||
|
|
0fb95ea616 | ||
|
|
e4cf54c166 | ||
|
|
e677d8513b | ||
|
|
d366d4680c | ||
|
|
76337ee326 | ||
|
|
6cd90e4597 | ||
|
|
08ae7a8d23 | ||
|
|
062c62d071 | ||
|
|
13549c2018 | ||
|
|
9094bedfd3 | ||
|
|
0d4eb4365c | ||
|
|
ac77470dee | ||
|
|
ea7a1cd171 | ||
|
|
40910d83b6 | ||
|
|
d7cc363931 | ||
|
|
4f8d6289c2 | ||
|
|
79febc7356 | ||
|
|
8acff25f0d | ||
|
|
2391d6916b |
5
.gitignore
vendored
Normal file
@@ -0,0 +1,5 @@
|
||||
# ignore users custom notify.sh
|
||||
/notify.sh
|
||||
/urls.list
|
||||
# ignore the auto-installed regctl
|
||||
regctl
|
||||
11
.pre-commit-config.yaml
Normal file
@@ -0,0 +1,11 @@
|
||||
repos:
|
||||
- repo: https://github.com/pre-commit/pre-commit-hooks
|
||||
rev: v4.6.0
|
||||
hooks:
|
||||
- id: check-yaml
|
||||
- id: check-added-large-files
|
||||
- id: trailing-whitespace
|
||||
args: [--markdown-linebreak-ext=md]
|
||||
- id: end-of-file-fixer
|
||||
- id: mixed-line-ending
|
||||
args: ['--fix=lf']
|
||||
224
README.md
@@ -1,45 +1,61 @@
|
||||
# dockcheck
|
||||
### A script checking updates for docker images **without pulling** - then selectively auto-update some/all containers.
|
||||
<p align="center">
|
||||
<img src="extras/dockcheck_logo_by_booYah187.png" width="160" title="dockcheck">
|
||||
</p>
|
||||
<p align="center">
|
||||
<img src="https://img.shields.io/badge/bash-4.3-green?style=flat-square&logo=gnubash" alt="bash">
|
||||
<a href="https://www.gnu.org/licenses/gpl-3.0.html"><img src="https://img.shields.io/badge/license-GPLv3-red?style=flat-square" alt="GPLv3"></a>
|
||||
<img src="https://img.shields.io/github/v/tag/mag37/dockcheck?style=flat-square&label=release" alt="release">
|
||||
<a href="https://ko-fi.com/mag37"><img src="https://img.shields.io/badge/-Ko--fi-grey?style=flat-square&logo=Ko-fi" alt="Buy me a Coffee"></a>
|
||||
<a href="https://liberapay.com/user-bin-rob/donate"><img src="https://img.shields.io/badge/-LiberaPay-grey?style=flat-square&logo=liberapay" alt="LiberaPay"></a>
|
||||
<a href="https://github.com/sponsors/mag37"><img src="https://img.shields.io/badge/-Sponsor-grey?style=flat-square&logo=github" alt="Github Sponsor"></a>
|
||||
</p>
|
||||
|
||||
### :warning: URGENT!
|
||||
The 2.0 change had a breaking error - compose-recreation might have included previous containers compose-file.
|
||||
If you've had odd errors, you can use the [errorCheck.sh](https://github.com/mag37/dockcheck/blob/main/errorCheck.sh) script to list current running container configs in a readable way. Look especially for **Compose files** listed in wrong places.
|
||||
Recreate the suspicious container(s) manually with `docker compose down && docker compose up -d`.
|
||||
<h2 align="center">CLI tool to automate docker image updates or notifying when updates are available.</h2>
|
||||
<h3 align="center">Features:</h3>
|
||||
<h3 align="center">selective updates, exclude containers, custom labels, notification plugins, prune when done and more.</h3>
|
||||
|
||||
<h4 align="center">For Podman - see the fork <a href="https://github.com/sudo-kraken/podcheck">sudo-kraken/podcheck</a>!</h4>
|
||||
|
||||
<h4 align="center">:whale: Docker Hub pull limit :chart_with_downwards_trend: not an issue for checks but for actual pulls - <a href="#whale-docker-hub-pull-limit-chart_with_downwards_trend-not-an-issue-for-checks-but-for-actual-pulls">read more</a></h4>
|
||||
|
||||
### :pushpin: Recent changes:
|
||||
- **v0.2.2**: Fixed breaking errors with multi-compose, odd breakage and working dir error.
|
||||
- **v0.2.1**: Added option to exclude a list of containers.
|
||||
- **v0.2.1**: Added multi-compose support (eg. override).
|
||||
- **v0.2.0**: Fixed error with container:tag definition.
|
||||
- **v0.1.9:** Fixed custom env-support.
|
||||
- **v0.1.8:** Added option to prune dangling images.
|
||||
___
|
||||
## :bell: Changelog
|
||||
|
||||
## Dependencies:
|
||||
Running docker (duh) and compose, either standalone or plugin.
|
||||
[`regclient/regctl`](https://github.com/regclient/regclient) (Licensed under [Apache-2.0 License](http://www.apache.org/licenses/LICENSE-2.0))
|
||||
User will be prompted to download `regctl` if not in `PATH` or `PWD`
|
||||
- **v0.5.6.0**: Heavily improved performance due to async checking for updates.
|
||||
- **v0.5.5.0**: osx and bsd compatibility changes + rewrite of dependency installer
|
||||
- **v0.5.4.0**: Added support for a Prometheus+node_exporter metric collection through a file collector.
|
||||
- **v0.5.3.0**: Local image check changed (use imageId instead of name) and Gotify-template fixed (whale icon removed).
|
||||
- **v0.5.2.1**: Rewrite of dependency downloads, jq can be installed with package manager or static binary.
|
||||
___
|
||||
|
||||
|
||||

|
||||

|
||||
|
||||
## `dockcheck.sh`
|
||||
## :mag_right: `dockcheck.sh`
|
||||
```
|
||||
$ ./dockcheck.sh -h
|
||||
Syntax: dockcheck.sh [OPTION] [part of name to filter]
|
||||
Example: dockcheck.sh -a -e nextcloud,heimdall
|
||||
Example: dockcheck.sh -y -d 10 -e nextcloud,heimdall
|
||||
|
||||
Options:
|
||||
-h Print this Help.
|
||||
Options:"
|
||||
-a|y Automatic updates, without interaction.
|
||||
-c D Exports metrics as prom file for the prometheus node_exporter. Provide the collector textfile directory.
|
||||
-d N Only update to new images that are N+ days old. Lists too recent with +prefix and age. 2xSlower.
|
||||
-e X Exclude containers, separated by comma.
|
||||
-f Force stack restart after update. Caution: restarts once for every updated container within stack.
|
||||
-h Print this Help.
|
||||
-i Inform - send a preconfigured notification.
|
||||
-l Only update if label is set. See readme.
|
||||
-m Monochrome mode, no printf color codes.
|
||||
-n No updates, only checking availability.
|
||||
-p Auto-Prune dangling images after update.
|
||||
-e Exclude containers, separated by comma.
|
||||
-r Allow updating images for docker run, wont update the container.
|
||||
-s Include stopped containers in the check. (Logic: docker ps -a).
|
||||
-t Set a timeout (in seconds) per container for registry checkups, 10 is default.
|
||||
-v Prints current version.
|
||||
```
|
||||
|
||||
Basic example:
|
||||
### Basic example:
|
||||
```
|
||||
$ ./dockcheck.sh
|
||||
. . .
|
||||
@@ -52,33 +68,157 @@ Containers with updates available:
|
||||
2) syncthing
|
||||
3) whoogle-search
|
||||
|
||||
|
||||
Choose what containers to update:
|
||||
Enter number(s) separated by comma, [a] for all - [q] to quit:
|
||||
|
||||
```
|
||||
Then it proceedes to run `pull` and `up -d` on every container with updates.
|
||||
Then it proceeds to run `pull` and `up -d` on every container with updates.
|
||||
After the updates are complete, you'll get prompted if you'd like to prune dangling images.
|
||||
|
||||
### `-r flag` :warning: disclaimer and warning:
|
||||
**Wont auto-update the containers, only their images. (compose is recommended)**
|
||||
___
|
||||
|
||||
## :nut_and_bolt: Dependencies
|
||||
- Running docker (duh) and compose, either standalone or plugin. (see [Podman fork](https://github.com/sudo-kraken/podcheck)
|
||||
- Bash shell or compatible shell of at least v4.3
|
||||
- POSIX `xargs`, usually default but can be installed with the `findutils` package - to enable async.
|
||||
- [jq](https://github.com/jqlang/jq)
|
||||
- User will be prompted to install with package manager or download static binary.
|
||||
- [regclient/regctl](https://github.com/regclient/regclient) (Licensed under [Apache-2.0 License](http://www.apache.org/licenses/LICENSE-2.0))
|
||||
- User will be prompted to download `regctl` if not in `PATH` or `PWD`.
|
||||
- regctl requires `amd64/arm64` - see [workaround](#roller_coaster-workaround-for-non-amd64--arm64) if other architecture is used.
|
||||
|
||||
## :tent: Install Instructions
|
||||
Download the script to a directory in **PATH**, I'd suggest using `~/.local/bin` as that's usually in **PATH**.
|
||||
For OSX/macOS preferably use `/usr/local/bin`.
|
||||
```sh
|
||||
# basic example with curl:
|
||||
curl -L https://raw.githubusercontent.com/mag37/dockcheck/main/dockcheck.sh -o ~/.local/bin/dockcheck.sh
|
||||
chmod +x ~/.local/bin/dockcheck.sh
|
||||
|
||||
# or oneliner with wget:
|
||||
wget -O ~/.local/bin/dockcheck.sh "https://raw.githubusercontent.com/mag37/dockcheck/main/dockcheck.sh" && chmod +x ~/.local/bin/dockcheck.sh
|
||||
|
||||
# OSX or macOS version with curl:
|
||||
curl -L https://raw.githubusercontent.com/mag37/dockcheck/main/dockcheck.sh -o /usr/local/bin/dockcheck.sh && chmod +x /usr/local/bin/dockcheck.sh
|
||||
```
|
||||
Then call the script anywhere with just `dockcheck.sh`.
|
||||
Add preferred `notify.sh`-template to the same directory - this will not be touched by the scripts self-update function.
|
||||
|
||||
|
||||
## :loudspeaker: Notifications
|
||||
Trigger with the `-i` flag.
|
||||
Run it scheduled with `-ni` to only get notified when there's updates available!
|
||||
|
||||
Use a `notify_X.sh` template file from the **notify_templates** directory, copy it to `notify.sh` alongside the script, modify it to your needs! (notify.sh is added to .gitignore)
|
||||
**Current templates:**
|
||||
- Synology [DSM](https://www.synology.com/en-global/dsm)
|
||||
- Email with [mSMTP](https://wiki.debian.org/msmtp) (or deprecated alternative [sSMTP](https://wiki.debian.org/sSMTP))
|
||||
- Apprise (with it's [multitude](https://github.com/caronc/apprise#supported-notifications) of notifications)
|
||||
- both native [caronc/apprise](https://github.com/caronc/apprise) and the standalone [linuxserver/docker-apprise-api](https://github.com/linuxserver/docker-apprise-api)
|
||||
- Read the [QuickStart](extras/apprise_quickstart.md)
|
||||
- [ntfy.sh](https://ntfy.sh/) - HTTP-based pub-sub notifications.
|
||||
- [Gotify](https://gotify.net/) - a simple server for sending and receiving messages.
|
||||
- [Pushbullet](https://www.pushbullet.com/) - connecting different devices with cross-platform features.
|
||||
- [Telegram](https://telegram.org/) - Telegram chat API.
|
||||
- [Matrix-Synapse](https://github.com/element-hq/synapse) - [Matrix](https://matrix.org/), open, secure, decentralised communication.
|
||||
- [Pushover](https://pushover.net/) - Simple Notifications (to your phone, wearables, desktops)
|
||||
- [Discord](https://support.discord.com/hc/en-us/articles/228383668-Intro-to-Webhooks) - Discord webhooks.
|
||||
|
||||
Further additions are welcome - suggestions or PR!
|
||||
<sub><sup>Initiated and first contributed by [yoyoma2](https://github.com/yoyoma2).</sup></sub>
|
||||
|
||||
### :date: Release notes addon to Notifications
|
||||
There's a function to use a lookup-file to add release note URL's to the notification message.
|
||||
Copy the notify_templates/`urls.list` file to the script directory, it will be used automatically if it's there. Modify it as necessary, the names of interest in the left column needs to match your container names.
|
||||
The output of the notification will look something like this:
|
||||
```
|
||||
Containers on hostname with updates available:
|
||||
apprise-api -> https://github.com/linuxserver/docker-apprise-api/releases
|
||||
homer -> https://github.com/bastienwirtz/homer/releases
|
||||
nginx -> https://github.com/docker-library/official-images/blob/master/library/nginx
|
||||
...
|
||||
```
|
||||
The `urls.list` file is just an example and I'd gladly see that people contribute back when they add their preferred URLs to their lists.
|
||||
|
||||
## :chart_with_upwards_trend: Prometheus and node_exporter
|
||||
Dockcheck can be used together with [Prometheus](https://github.com/prometheus/prometheus) and [node_exporter](https://github.com/prometheus/node_exporter) to export metrics via the file collector, scheduled with cron or likely.
|
||||
This is done with the `-c` option, like this:
|
||||
```
|
||||
dockcheck.sh -c /path/to/exporter/directory
|
||||
```
|
||||
|
||||
See the [README.md](./addons/prometheus/README.md) for more detailed information on how to set it up!
|
||||
|
||||
<sub><sup>Contributed by [tdralle](https://github.com/tdralle).</sup></sub>
|
||||
|
||||
## :bookmark: Labels
|
||||
Optionally add labels to compose-files. Currently these are the usable labels:
|
||||
```
|
||||
labels:
|
||||
mag37.dockcheck.restart-stack: true
|
||||
mag37.dockcheck.update: true
|
||||
```
|
||||
- `mag37.dockcheck.restart-stack: true` works instead of the `-f` option, forcing stop+restart on the whole compose-stack (Caution: Will restart on every updated container within stack).
|
||||
- `mag37.dockcheck.update: true` will when used with the `-l` option only update containers with this label and skip the rest. Will still list updates as usual.
|
||||
|
||||
## :roller_coaster: Workaround for non **amd64** / **arm64**
|
||||
`regctl` provides binaries for amd64/arm64, to use on other architecture you could try this workaround.
|
||||
Run regctl in a container wrapped in a shell script. Copied from [regclient/docs/install.md](https://github.com/regclient/regclient/blob/main/docs/install.md):
|
||||
|
||||
```sh
|
||||
cat >regctl <<EOF
|
||||
#!/bin/sh
|
||||
opts=""
|
||||
case "\$*" in
|
||||
"registry login"*) opts="-t";;
|
||||
esac
|
||||
docker container run \$opts -i --rm --net host \\
|
||||
-u "\$(id -u):\$(id -g)" -e HOME -v \$HOME:\$HOME \\
|
||||
-v /etc/docker/certs.d:/etc/docker/certs.d:ro \\
|
||||
ghcr.io/regclient/regctl:latest "\$@"
|
||||
EOF
|
||||
chmod 755 regctl
|
||||
```
|
||||
Test it with `./regctl --help` and then either add the file to the same path as *dockcheck.sh* or in your path (eg. `~/.local/bin/regctl`).
|
||||
|
||||
## :whale: Docker Hub pull limit :chart_with_downwards_trend: not an issue for checks but for actual pulls
|
||||
Due to recent changes in [Docker Hub usage and limits](https://docs.docker.com/docker-hub/usage/)
|
||||
>Unauthenticated users: 10 pulls/hour
|
||||
>Authenticated users with a free account: 100 pulls/hour
|
||||
|
||||
This is not an issue for registry checks. But if you have a large stack and pull more than 10 updates at once consider updating more often or to create a free account.
|
||||
You could use/modify the login-wrapper function in the example below to automate the login prior to running `dockcheck.sh`.
|
||||
|
||||
### :guardsman: Function to auth with docker hub before running
|
||||
**Example** - Change names, paths, and remove cat+password flag if you rather get prompted:
|
||||
```sh
|
||||
function dchk {
|
||||
cat ~/pwd.txt | docker login --username YourUser --password-stdin
|
||||
~/dockcheck.sh "$@"
|
||||
}
|
||||
```
|
||||
|
||||
## :hammer: Known issues
|
||||
- No detailed error feedback (just skip + list what's skipped).
|
||||
- Not respecting `--profile` options when re-creating the container.
|
||||
- Not working well with containers created by **Portainer**.
|
||||
- **Watchtower** might cause issues due to retagging images when checking for updates (and thereby pulling new images).
|
||||
|
||||
## :warning: `-r flag` disclaimer and warning
|
||||
**Wont auto-update the containers, only their images. (compose is recommended)**
|
||||
`docker run` dont support using new images just by restarting a container.
|
||||
Containers need to be manually stopped, removed and created again to run on the new image.
|
||||
|
||||
### :hammer: Known issues
|
||||
- No detailed error feedback (just skip + list what's skipped) .
|
||||
- Not respecting `--profile` options when re-creating the container.
|
||||
## :wrench: Debugging
|
||||
If you hit issues, you could check the output of the `extras/errorCheck.sh` script for clues.
|
||||
Another option is to run the main script with debugging in a subshell `bash -x dockcheck.sh` - if there's a particular container/image that's causing issues you can filter for just that through `bash -x dockcheck.sh nginx`.
|
||||
|
||||
## `dc_brief.sh`
|
||||
Just a brief, slimmed down version of the script to only print what containers got updates, no updates or errors.
|
||||
|
||||
# License
|
||||
## :scroll: License
|
||||
dockcheck is created and released under the [GNU GPL v3.0](https://www.gnu.org/licenses/gpl-3.0-standalone.html) license.
|
||||
|
||||
## :heartpulse: Sponsorlist
|
||||
|
||||
- [avegy](https://github.com/avegy)
|
||||
|
||||
___
|
||||
|
||||
## Check out a spinoff brother-project:
|
||||
### [Palleri/dockcheck-web](https://github.com/Palleri/dockcheck-web) for a WebUI-front!
|
||||
|
||||
## Special Thanks:
|
||||
- :bison: [t0rnis](https://github.com/t0rnis)
|
||||
- :leopard: [Palleri](https://github.com/Palleri)
|
||||
### :floppy_disk: The [story](https://mag37.org/posts/project_dockcheck/) behind it. 1 year in retrospect.
|
||||
|
||||
61
addons/prometheus/README.md
Normal file
@@ -0,0 +1,61 @@
|
||||
## [Prometheus](https://github.com/prometheus/prometheus) and [node_exporter](https://github.com/prometheus/node_exporter)
|
||||
Dockcheck is capable to export metrics to prometheus via the text file collector provided by the node_exporter.
|
||||
In order to do so the -c flag has to be specified followed by the file path that is configured in the text file collector of the node_exporter.
|
||||
A simple cron job can be configured to export these metrics on a regular interval as shown in the sample below:
|
||||
|
||||
```
|
||||
0 1 * * * /root/dockcheck.sh -n -c /var/lib/node_exporter/textfile_collector
|
||||
```
|
||||
|
||||
The following metrics are exported to prometheus
|
||||
|
||||
```
|
||||
# HELP dockcheck_images_analyzed Docker images that have been analyzed
|
||||
# TYPE dockcheck_images_analyzed gauge
|
||||
dockcheck_images_analyzed 22
|
||||
# HELP dockcheck_images_outdated Docker images that are outdated
|
||||
# TYPE dockcheck_images_outdated gauge
|
||||
dockcheck_images_outdated 7
|
||||
# HELP dockcheck_images_latest Docker images that are outdated
|
||||
# TYPE dockcheck_images_latest gauge
|
||||
dockcheck_images_latest 14
|
||||
# HELP dockcheck_images_error Docker images with analysis errors
|
||||
# TYPE dockcheck_images_error gauge
|
||||
dockcheck_images_error 1
|
||||
# HELP dockcheck_images_analyze_timestamp_seconds Last dockercheck run time
|
||||
# TYPE dockcheck_images_analyze_timestamp_seconds gauge
|
||||
dockcheck_images_analyze_timestamp_seconds 1737924029
|
||||
```
|
||||
|
||||
Once those metrics are exported they can be used to define alarms as shown below
|
||||
|
||||
```
|
||||
- alert: dockcheck_images_outdated
|
||||
expr: sum by(instance) (dockcheck_images_outdated) > 0
|
||||
for: 15s
|
||||
labels:
|
||||
severity: warning
|
||||
annotations:
|
||||
summary: "{{ $labels.instance }} has {{ $value }} outdated docker images."
|
||||
description: "{{ $labels.instance }} has {{ $value }} outdated docker images."
|
||||
- alert: dockcheck_images_error
|
||||
expr: sum by(instance) (dockcheck_images_error) > 0
|
||||
for: 15s
|
||||
labels:
|
||||
severity: warning
|
||||
annotations:
|
||||
summary: "{{ $labels.instance }} has {{ $value }} docker images having an error."
|
||||
description: "{{ $labels.instance }} has {{ $value }} docker images having an error."
|
||||
- alert: dockercheck_image_last_analyze
|
||||
expr: (time() - dockcheck_images_analyze_timestamp_seconds) > (3600 * 24 * 3)
|
||||
for: 15s
|
||||
labels:
|
||||
severity: warning
|
||||
annotations:
|
||||
summary: "{{ $labels.instance }} has not updated the dockcheck statistics for more than 3 days."
|
||||
description: "{{ $labels.instance }} has not updated the dockcheck statistics for more than 3 days."
|
||||
```
|
||||
|
||||
There is a reference Grafana dashboard in [grafana/grafana_dashboard.json](./grafana/grafana_dashboard.json).
|
||||
|
||||

|
||||
382
addons/prometheus/grafana/grafana_dashboard.json
Normal file
@@ -0,0 +1,382 @@
|
||||
{
|
||||
"__inputs": [
|
||||
{
|
||||
"name": "DS_PROMETHEUS",
|
||||
"label": "prometheus",
|
||||
"description": "",
|
||||
"type": "datasource",
|
||||
"pluginId": "prometheus",
|
||||
"pluginName": "Prometheus"
|
||||
}
|
||||
],
|
||||
"__elements": {},
|
||||
"__requires": [
|
||||
{
|
||||
"type": "grafana",
|
||||
"id": "grafana",
|
||||
"name": "Grafana",
|
||||
"version": "11.4.0"
|
||||
},
|
||||
{
|
||||
"type": "datasource",
|
||||
"id": "prometheus",
|
||||
"name": "Prometheus",
|
||||
"version": "1.0.0"
|
||||
},
|
||||
{
|
||||
"type": "panel",
|
||||
"id": "table",
|
||||
"name": "Table",
|
||||
"version": ""
|
||||
}
|
||||
],
|
||||
"annotations": {
|
||||
"list": [
|
||||
{
|
||||
"builtIn": 1,
|
||||
"datasource": {
|
||||
"type": "grafana",
|
||||
"uid": "-- Grafana --"
|
||||
},
|
||||
"enable": true,
|
||||
"hide": true,
|
||||
"iconColor": "rgba(0, 211, 255, 1)",
|
||||
"name": "Annotations & Alerts",
|
||||
"type": "dashboard"
|
||||
}
|
||||
]
|
||||
},
|
||||
"editable": true,
|
||||
"fiscalYearStartMonth": 0,
|
||||
"graphTooltip": 0,
|
||||
"id": null,
|
||||
"links": [],
|
||||
"panels": [
|
||||
{
|
||||
"datasource": {
|
||||
"type": "prometheus",
|
||||
"uid": "${DS_PROMETHEUS}"
|
||||
},
|
||||
"fieldConfig": {
|
||||
"defaults": {
|
||||
"color": {
|
||||
"mode": "thresholds"
|
||||
},
|
||||
"custom": {
|
||||
"align": "auto",
|
||||
"cellOptions": {
|
||||
"type": "auto"
|
||||
},
|
||||
"inspect": false
|
||||
},
|
||||
"mappings": [],
|
||||
"thresholds": {
|
||||
"mode": "absolute",
|
||||
"steps": [
|
||||
{
|
||||
"color": "green",
|
||||
"value": null
|
||||
},
|
||||
{
|
||||
"color": "red",
|
||||
"value": 80
|
||||
}
|
||||
]
|
||||
}
|
||||
},
|
||||
"overrides": [
|
||||
{
|
||||
"matcher": {
|
||||
"id": "byName",
|
||||
"options": "last_analyze_timestamp"
|
||||
},
|
||||
"properties": [
|
||||
{
|
||||
"id": "unit",
|
||||
"value": "dateTimeAsIso"
|
||||
}
|
||||
]
|
||||
},
|
||||
{
|
||||
"matcher": {
|
||||
"id": "byName",
|
||||
"options": "last_analyze_since"
|
||||
},
|
||||
"properties": [
|
||||
{
|
||||
"id": "unit",
|
||||
"value": "s"
|
||||
},
|
||||
{
|
||||
"id": "custom.cellOptions",
|
||||
"value": {
|
||||
"mode": "gradient",
|
||||
"type": "color-background"
|
||||
}
|
||||
},
|
||||
{
|
||||
"id": "thresholds",
|
||||
"value": {
|
||||
"mode": "absolute",
|
||||
"steps": [
|
||||
{
|
||||
"color": "green",
|
||||
"value": null
|
||||
},
|
||||
{
|
||||
"color": "red",
|
||||
"value": 259200
|
||||
}
|
||||
]
|
||||
}
|
||||
}
|
||||
]
|
||||
},
|
||||
{
|
||||
"matcher": {
|
||||
"id": "byName",
|
||||
"options": "images_outdated"
|
||||
},
|
||||
"properties": [
|
||||
{
|
||||
"id": "custom.cellOptions",
|
||||
"value": {
|
||||
"mode": "gradient",
|
||||
"type": "color-background"
|
||||
}
|
||||
},
|
||||
{
|
||||
"id": "thresholds",
|
||||
"value": {
|
||||
"mode": "absolute",
|
||||
"steps": [
|
||||
{
|
||||
"color": "green",
|
||||
"value": null
|
||||
},
|
||||
{
|
||||
"color": "red",
|
||||
"value": 1
|
||||
}
|
||||
]
|
||||
}
|
||||
}
|
||||
]
|
||||
},
|
||||
{
|
||||
"matcher": {
|
||||
"id": "byName",
|
||||
"options": "images_error"
|
||||
},
|
||||
"properties": [
|
||||
{
|
||||
"id": "custom.cellOptions",
|
||||
"value": {
|
||||
"mode": "gradient",
|
||||
"type": "color-background"
|
||||
}
|
||||
},
|
||||
{
|
||||
"id": "thresholds",
|
||||
"value": {
|
||||
"mode": "absolute",
|
||||
"steps": [
|
||||
{
|
||||
"color": "green",
|
||||
"value": null
|
||||
},
|
||||
{
|
||||
"color": "red",
|
||||
"value": 1
|
||||
}
|
||||
]
|
||||
}
|
||||
}
|
||||
]
|
||||
}
|
||||
]
|
||||
},
|
||||
"gridPos": {
|
||||
"h": 14,
|
||||
"w": 24,
|
||||
"x": 0,
|
||||
"y": 0
|
||||
},
|
||||
"id": 2,
|
||||
"options": {
|
||||
"cellHeight": "sm",
|
||||
"footer": {
|
||||
"countRows": false,
|
||||
"fields": "",
|
||||
"reducer": [
|
||||
"sum"
|
||||
],
|
||||
"show": false
|
||||
},
|
||||
"frameIndex": 1,
|
||||
"showHeader": true,
|
||||
"sortBy": []
|
||||
},
|
||||
"pluginVersion": "11.4.0",
|
||||
"targets": [
|
||||
{
|
||||
"disableTextWrap": false,
|
||||
"editorMode": "code",
|
||||
"exemplar": false,
|
||||
"expr": "sum by(instance) (dockcheck_images_analyzed)",
|
||||
"format": "table",
|
||||
"fullMetaSearch": false,
|
||||
"hide": false,
|
||||
"includeNullMetadata": true,
|
||||
"instant": true,
|
||||
"interval": "",
|
||||
"legendFormat": "{{instance}}",
|
||||
"range": false,
|
||||
"refId": "dockcheck_images_analyzed",
|
||||
"useBackend": false,
|
||||
"datasource": {
|
||||
"type": "prometheus",
|
||||
"uid": "${DS_PROMETHEUS}"
|
||||
}
|
||||
},
|
||||
{
|
||||
"datasource": {
|
||||
"type": "prometheus",
|
||||
"uid": "${DS_PROMETHEUS}"
|
||||
},
|
||||
"disableTextWrap": false,
|
||||
"editorMode": "code",
|
||||
"exemplar": false,
|
||||
"expr": "sum by(instance) (dockcheck_images_outdated)",
|
||||
"format": "table",
|
||||
"fullMetaSearch": false,
|
||||
"hide": false,
|
||||
"includeNullMetadata": true,
|
||||
"instant": true,
|
||||
"legendFormat": "{{instance}}",
|
||||
"range": false,
|
||||
"refId": "dockcheck_images_outdated",
|
||||
"useBackend": false
|
||||
},
|
||||
{
|
||||
"datasource": {
|
||||
"type": "prometheus",
|
||||
"uid": "${DS_PROMETHEUS}"
|
||||
},
|
||||
"disableTextWrap": false,
|
||||
"editorMode": "code",
|
||||
"exemplar": false,
|
||||
"expr": "sum by(instance) (dockcheck_images_latest)",
|
||||
"format": "table",
|
||||
"fullMetaSearch": false,
|
||||
"hide": false,
|
||||
"includeNullMetadata": true,
|
||||
"instant": true,
|
||||
"legendFormat": "{{instance}}",
|
||||
"range": false,
|
||||
"refId": "dockcheck_images_latest",
|
||||
"useBackend": false
|
||||
},
|
||||
{
|
||||
"datasource": {
|
||||
"type": "prometheus",
|
||||
"uid": "${DS_PROMETHEUS}"
|
||||
},
|
||||
"editorMode": "code",
|
||||
"exemplar": false,
|
||||
"expr": "sum by(instance) (dockcheck_images_error)",
|
||||
"format": "table",
|
||||
"hide": false,
|
||||
"instant": true,
|
||||
"legendFormat": "{{instance}}",
|
||||
"range": false,
|
||||
"refId": "dockcheck_images_error"
|
||||
},
|
||||
{
|
||||
"datasource": {
|
||||
"type": "prometheus",
|
||||
"uid": "${DS_PROMETHEUS}"
|
||||
},
|
||||
"editorMode": "code",
|
||||
"exemplar": false,
|
||||
"expr": "dockcheck_images_analyze_timestamp_seconds * 1000",
|
||||
"format": "table",
|
||||
"hide": false,
|
||||
"instant": true,
|
||||
"legendFormat": "{{instance}}",
|
||||
"range": false,
|
||||
"refId": "dockcheck_images_analyze_timestamp_seconds"
|
||||
},
|
||||
{
|
||||
"datasource": {
|
||||
"type": "prometheus",
|
||||
"uid": "${DS_PROMETHEUS}"
|
||||
},
|
||||
"editorMode": "code",
|
||||
"exemplar": false,
|
||||
"expr": "time() - dockcheck_images_analyze_timestamp_seconds",
|
||||
"format": "table",
|
||||
"hide": false,
|
||||
"instant": true,
|
||||
"legendFormat": "{{instance}}",
|
||||
"range": false,
|
||||
"refId": "dockcheck_images_last_analyze"
|
||||
}
|
||||
],
|
||||
"title": "Dockcheck Status",
|
||||
"transformations": [
|
||||
{
|
||||
"id": "merge",
|
||||
"options": {}
|
||||
},
|
||||
{
|
||||
"id": "organize",
|
||||
"options": {
|
||||
"excludeByName": {
|
||||
"Time": true,
|
||||
"__name__": true,
|
||||
"job": true
|
||||
},
|
||||
"includeByName": {},
|
||||
"indexByName": {
|
||||
"Time": 0,
|
||||
"Value #dockcheck_images_analyze_timestamp_seconds": 2,
|
||||
"Value #dockcheck_images_analyzed": 4,
|
||||
"Value #dockcheck_images_error": 7,
|
||||
"Value #dockcheck_images_last_analyze": 3,
|
||||
"Value #dockcheck_images_latest": 5,
|
||||
"Value #dockcheck_images_outdated": 6,
|
||||
"instance": 1,
|
||||
"job": 8
|
||||
},
|
||||
"renameByName": {
|
||||
"Value #A": "analyze_timestamp",
|
||||
"Value #dockcheck_images_analyze_timestamp_seconds": "last_analyze_timestamp",
|
||||
"Value #dockcheck_images_analyzed": "images_analyzed",
|
||||
"Value #dockcheck_images_error": "images_error",
|
||||
"Value #dockcheck_images_last_analyze": "last_analyze_since",
|
||||
"Value #dockcheck_images_latest": "images_latest",
|
||||
"Value #dockcheck_images_outdated": "images_outdated"
|
||||
}
|
||||
}
|
||||
}
|
||||
],
|
||||
"type": "table"
|
||||
}
|
||||
],
|
||||
"schemaVersion": 40,
|
||||
"tags": [],
|
||||
"templating": {
|
||||
"list": []
|
||||
},
|
||||
"time": {
|
||||
"from": "now-6h",
|
||||
"to": "now"
|
||||
},
|
||||
"timepicker": {},
|
||||
"timezone": "browser",
|
||||
"title": "Dockcheck Status",
|
||||
"uid": "feb4pv3kv1hxca",
|
||||
"version": 17,
|
||||
"weekStart": ""
|
||||
}
|
||||
BIN
addons/prometheus/grafana/grafana_dashboard.png
Normal file
|
After Width: | Height: | Size: 50 KiB |
28
addons/prometheus/prometheus_collector.sh
Normal file
@@ -0,0 +1,28 @@
|
||||
prometheus_exporter() {
|
||||
checkedImages=$(($1 + $2 + $3))
|
||||
checkTimestamp=$(date +%s)
|
||||
|
||||
promFileContent=()
|
||||
promFileContent+=("# HELP dockcheck_images_analyzed Docker images that have been analyzed")
|
||||
promFileContent+=("# TYPE dockcheck_images_analyzed gauge")
|
||||
promFileContent+=("dockcheck_images_analyzed $checkedImages")
|
||||
|
||||
promFileContent+=("# HELP dockcheck_images_outdated Docker images that are outdated")
|
||||
promFileContent+=("# TYPE dockcheck_images_outdated gauge")
|
||||
promFileContent+=("dockcheck_images_outdated ${#GotUpdates[@]}")
|
||||
|
||||
promFileContent+=("# HELP dockcheck_images_latest Docker images that are outdated")
|
||||
promFileContent+=("# TYPE dockcheck_images_latest gauge")
|
||||
promFileContent+=("dockcheck_images_latest ${#NoUpdates[@]}")
|
||||
|
||||
promFileContent+=("# HELP dockcheck_images_error Docker images with analysis errors")
|
||||
promFileContent+=("# TYPE dockcheck_images_error gauge")
|
||||
promFileContent+=("dockcheck_images_error ${#GotErrors[@]}")
|
||||
|
||||
promFileContent+=("# HELP dockcheck_images_analyze_timestamp_seconds Last dockercheck run time")
|
||||
promFileContent+=("# TYPE dockcheck_images_analyze_timestamp_seconds gauge")
|
||||
promFileContent+=("dockcheck_images_analyze_timestamp_seconds $checkTimestamp")
|
||||
|
||||
printf "%s\n" "${promFileContent[@]}" > "$CollectorTextFileDirectory/dockcheck_info.prom\$\$"
|
||||
mv -f "$CollectorTextFileDirectory/dockcheck_info.prom\$\$" "$CollectorTextFileDirectory/dockcheck_info.prom"
|
||||
}
|
||||
496
dockcheck.sh
@@ -1,95 +1,115 @@
|
||||
#!/usr/bin/env bash
|
||||
VERSION="v0.2.2-URGENT"
|
||||
VERSION="v0.5.6.0"
|
||||
### ChangeNotes: Heavily improved performance due to asynchronous update checks.
|
||||
Github="https://github.com/mag37/dockcheck"
|
||||
RawUrl="https://raw.githubusercontent.com/mag37/dockcheck/main/dockcheck.sh"
|
||||
|
||||
### Check if there's a new release of the script:
|
||||
LatestRelease="$(curl -s -r 0-50 https://raw.githubusercontent.com/mag37/dockcheck/main/dockcheck.sh | sed -n "/VERSION/s/VERSION=//p" | tr -d '"')"
|
||||
[ "$VERSION" != "$LatestRelease" ] && printf "New version available! Latest: %s - Local: %s \nGrab it here: %s \n\n" "$LatestRelease" "$VERSION" "$Github"
|
||||
# Variables for self updating
|
||||
ScriptArgs=( "$@" )
|
||||
ScriptPath="$(readlink -f "$0")"
|
||||
ScriptWorkDir="$(dirname "$ScriptPath")"
|
||||
|
||||
### Help Function:
|
||||
# Check if there's a new release of the script
|
||||
LatestRelease="$(curl -s -r 0-50 $RawUrl | sed -n "/VERSION/s/VERSION=//p" | tr -d '"')"
|
||||
LatestChanges="$(curl -s -r 0-200 $RawUrl | sed -n "/ChangeNotes/s/# ChangeNotes: //p")"
|
||||
|
||||
# Help Function
|
||||
Help() {
|
||||
echo "Syntax: dockcheck.sh [OPTION] [part of name to filter]"
|
||||
echo "Example: dockcheck.sh -a -e nextcloud,heimdall"
|
||||
echo "Syntax: dockcheck.sh [OPTION] [part of name to filter]"
|
||||
echo "Example: dockcheck.sh -y -d 10 -e nextcloud,heimdall"
|
||||
echo
|
||||
echo "Options:"
|
||||
echo "-h Print this Help."
|
||||
echo "-a|y Automatic updates, without interaction."
|
||||
echo "-n No updates, only checking availability."
|
||||
echo "-e Exclude containers, separated by comma."
|
||||
echo "-p Auto-Prune dangling images after update."
|
||||
echo "-r Allow updating images for docker run, wont update the container"
|
||||
echo "-c Exports metrics as prom file for the prometheus node_exporter. Provide the collector textfile directory."
|
||||
echo "-d N Only update to new images that are N+ days old. Lists too recent with +prefix and age. 2xSlower."
|
||||
echo "-e X Exclude containers, separated by comma."
|
||||
echo "-f Force stack restart after update. Caution: restarts once for every updated container within stack."
|
||||
echo "-h Print this Help."
|
||||
echo "-i Inform - send a preconfigured notification."
|
||||
echo "-l Only update if label is set. See readme."
|
||||
echo "-m Monochrome mode, no printf color codes."
|
||||
echo "-n No updates; only checking availability."
|
||||
echo "-p Auto-prune dangling images after update."
|
||||
echo "-r Allow updating images for docker run; won't update the container."
|
||||
echo "-s Include stopped containers in the check. (Logic: docker ps -a)."
|
||||
echo "-t Set a timeout (in seconds) per container for registry checkups, 10 is default."
|
||||
echo "-v Prints current version."
|
||||
echo
|
||||
echo "Project source: $Github"
|
||||
}
|
||||
|
||||
while getopts "aynprhe:" options; do
|
||||
# Colors
|
||||
c_red="\033[0;31m"
|
||||
c_green="\033[0;32m"
|
||||
c_yellow="\033[0;33m"
|
||||
c_blue="\033[0;34m"
|
||||
c_teal="\033[0;36m"
|
||||
c_reset="\033[0m"
|
||||
|
||||
MaxAsync=32
|
||||
Timeout=10
|
||||
Stopped=""
|
||||
while getopts "aynpfrhlisvmc:e:d:t:" options; do
|
||||
case "${options}" in
|
||||
a|y) UpdYes="yes" ;;
|
||||
n) UpdYes="no" ;;
|
||||
r) DrUp="yes" ;;
|
||||
p) PruneQ="yes" ;;
|
||||
e) Exclude=${OPTARG} ;;
|
||||
h|*) Help ; exit 0 ;;
|
||||
a|y) AutoUp="yes" ;;
|
||||
c) CollectorTextFileDirectory="${OPTARG}"
|
||||
if ! [[ -d $CollectorTextFileDirectory ]] ; then { printf "The directory (%s) does not exist.\n" "${CollectorTextFileDirectory}" ; exit 2; } fi ;;
|
||||
n) AutoUp="no" ;;
|
||||
r) DRunUp="yes" ;;
|
||||
p) AutoPrune="yes" ;;
|
||||
l) OnlyLabel=true ;;
|
||||
f) ForceRestartStacks=true ;;
|
||||
i) [ -s "$ScriptWorkDir"/notify.sh ] && { source "$ScriptWorkDir"/notify.sh ; Notify="yes" ; } ;;
|
||||
e) Exclude=${OPTARG} ;;
|
||||
m) declare c_{red,green,yellow,blue,teal,reset}="" ;;
|
||||
s) Stopped="-a" ;;
|
||||
t) Timeout="${OPTARG}" ;;
|
||||
v) printf "%s\n" "$VERSION" ; exit 0 ;;
|
||||
d) DaysOld=${OPTARG}
|
||||
if ! [[ $DaysOld =~ ^[0-9]+$ ]] ; then { printf "Days -d argument given (%s) is not a number.\n" "${DaysOld}" ; exit 2 ; } ; fi ;;
|
||||
h|*) Help ; exit 2 ;;
|
||||
esac
|
||||
done
|
||||
shift "$((OPTIND-1))"
|
||||
|
||||
### Set $1 to a variable for name filtering later.
|
||||
SearchName="$1"
|
||||
### Create array of excludes
|
||||
IFS=',' read -r -a Excludes <<< "$Exclude" ; unset IFS
|
||||
|
||||
### Check if required binary exists in PATH or directory:
|
||||
if [[ $(builtin type -P "regctl") ]]; then regbin="regctl" ;
|
||||
elif [[ -f "./regctl" ]]; then regbin="./regctl" ;
|
||||
else
|
||||
read -r -p "Required dependency 'regctl' missing, do you want it downloaded? y/[n] " GetDep
|
||||
if [[ "$GetDep" =~ [yY] ]] ; then
|
||||
### Check arch:
|
||||
case "$(uname --machine)" in
|
||||
x86_64|amd64) architecture="amd64" ;;
|
||||
arm64|aarch64) architecture="arm64";;
|
||||
*) echo "Architecture not supported, exiting." ; exit 1;;
|
||||
esac
|
||||
RegUrl="https://github.com/regclient/regclient/releases/latest/download/regctl-linux-$architecture"
|
||||
if [[ $(builtin type -P curl) ]]; then curl -L $RegUrl > ./regctl ; chmod +x ./regctl ; regbin="./regctl" ;
|
||||
elif [[ $(builtin type -P wget) ]]; then wget $RegUrl -O ./regctl ; chmod +x ./regctl ; regbin="./regctl" ;
|
||||
else
|
||||
printf "%s\n" "curl/wget not available - get regctl manually from the repo link, quitting."
|
||||
fi
|
||||
# Self-update function
|
||||
self_update_curl() {
|
||||
cp "$ScriptPath" "$ScriptPath".bak
|
||||
if [[ $(command -v curl) ]]; then
|
||||
curl -L $RawUrl > "$ScriptPath" ; chmod +x "$ScriptPath"
|
||||
printf "\n%s\n" "--- starting over with the updated version ---"
|
||||
exec "$ScriptPath" "${ScriptArgs[@]}" # run the new script with old arguments
|
||||
exit 1 # Exit the old instance
|
||||
elif [[ $(command -v wget) ]]; then
|
||||
wget $RawUrl -O "$ScriptPath" ; chmod +x "$ScriptPath"
|
||||
printf "\n%s\n" "--- starting over with the updated version ---"
|
||||
exec "$ScriptPath" "${ScriptArgs[@]}" # run the new script with old arguments
|
||||
exit 1 # Exit the old instance
|
||||
else
|
||||
printf "%s\n" "Dependency missing, quitting."
|
||||
exit 1
|
||||
printf "curl/wget not available - download the update manually: %s \n" "$Github"
|
||||
fi
|
||||
fi
|
||||
### final check if binary is correct
|
||||
$regbin version &> /dev/null || { printf "%s\n" "regctl is not working - try to remove it and re-download it, exiting."; exit 1; }
|
||||
|
||||
### Check docker compose binary:
|
||||
if docker compose version &> /dev/null ; then
|
||||
DockerBin="docker compose"
|
||||
elif docker-compose -v &> /dev/null; then
|
||||
DockerBin="docker-compose"
|
||||
elif docker -v &> /dev/null; then
|
||||
printf "%s\n" "No docker compose binary available, using plain docker (Not recommended!)"
|
||||
printf "%s\n" "'docker run' will ONLY update images, not the container itself."
|
||||
else
|
||||
printf "%s\n" "No docker binaries available, exiting."
|
||||
exit 1
|
||||
fi
|
||||
|
||||
### Numbered List -function:
|
||||
options() {
|
||||
num=1
|
||||
for i in "${GotUpdates[@]}"; do
|
||||
echo "$num) $i"
|
||||
((num++))
|
||||
done
|
||||
}
|
||||
|
||||
### Choose from list -function:
|
||||
self_update() {
|
||||
cd "$ScriptWorkDir" || { printf "Path error, skipping update.\n" ; return ; }
|
||||
if [[ $(command -v git) ]] && [[ "$(git ls-remote --get-url 2>/dev/null)" =~ .*"mag37/dockcheck".* ]] ; then
|
||||
printf "\n%s\n" "Pulling the latest version."
|
||||
git pull --force || { printf "Git error, manually pull/clone.\n" ; return ; }
|
||||
printf "\n%s\n" "--- starting over with the updated version ---"
|
||||
cd - || { printf "Path error.\n" ; return ; }
|
||||
exec "$ScriptPath" "${ScriptArgs[@]}" # run the new script with old arguments
|
||||
exit 1 # exit the old instance
|
||||
else
|
||||
cd - || { printf "Path error.\n" ; return ; }
|
||||
self_update_curl
|
||||
fi
|
||||
}
|
||||
|
||||
# Choose from list function
|
||||
choosecontainers() {
|
||||
while [[ -z "$ChoiceClean" ]]; do
|
||||
read -r -p "Enter number(s) separated by comma, [a] for all - [q] to quit: " Choice
|
||||
if [[ "$Choice" =~ [qQnN] ]] ; then
|
||||
if [[ "$Choice" =~ [qQnN] ]] ; then
|
||||
exit 0
|
||||
elif [[ "$Choice" =~ [aAyY] ]] ; then
|
||||
SelectedUpdates=( "${GotUpdates[@]}" )
|
||||
@@ -97,7 +117,7 @@ choosecontainers() {
|
||||
else
|
||||
ChoiceClean=${Choice//[,.:;]/ }
|
||||
for CC in $ChoiceClean ; do
|
||||
if [[ "$CC" -lt 1 || "$CC" -gt $UpdCount ]] ; then # reset choice if out of bounds
|
||||
if [[ "$CC" -lt 1 || "$CC" -gt $UpdCount ]] ; then # Reset choice if out of bounds
|
||||
echo "Number not in list: $CC" ; unset ChoiceClean ; break 1
|
||||
else
|
||||
SelectedUpdates+=( "${GotUpdates[$CC-1]}" )
|
||||
@@ -110,93 +130,319 @@ choosecontainers() {
|
||||
printf "\n"
|
||||
}
|
||||
|
||||
### Check the image-hash of every running container VS the registry
|
||||
for i in $(docker ps --filter "name=$SearchName" --format '{{.Names}}') ; do
|
||||
[[ " ${Excludes[*]} " =~ ${i} ]] && continue; # Skip if the container is excluded
|
||||
printf ". "
|
||||
RepoUrl=$(docker inspect "$i" --format='{{.Config.Image}}')
|
||||
LocalHash=$(docker image inspect "$RepoUrl" --format '{{.RepoDigests}}')
|
||||
### Checking for errors while setting the variable:
|
||||
if RegHash=$($regbin image digest --list "$RepoUrl" 2>/dev/null) ; then
|
||||
if [[ "$LocalHash" = *"$RegHash"* ]] ; then NoUpdates+=("$i"); else GotUpdates+=("$i"); fi
|
||||
datecheck() {
|
||||
ImageDate=$($regbin -v error image inspect "$RepoUrl" --format='{{.Created}}' | cut -d" " -f1 )
|
||||
ImageEpoch=$(date -d "$ImageDate" +%s 2>/dev/null) || ImageEpoch=$(date -f "%Y-%m-%d" -j "$ImageDate" +%s)
|
||||
ImageAge=$(( ( $(date +%s) - $ImageEpoch )/86400 ))
|
||||
if [ "$ImageAge" -gt "$DaysOld" ] ; then
|
||||
return 0
|
||||
else
|
||||
GotErrors+=("$i")
|
||||
return 1
|
||||
fi
|
||||
done
|
||||
}
|
||||
|
||||
### Sort arrays alphabetically
|
||||
IFS=$'\n'
|
||||
progress_bar() {
|
||||
QueCurrent="$1"
|
||||
QueTotal="$2"
|
||||
((Percent=100*QueCurrent/QueTotal))
|
||||
((Complete=50*Percent/100)) # Change first number for width (50)
|
||||
((Left=50-Complete)) # Change first number for width (50)
|
||||
BarComplete=$(printf "%${Complete}s" | tr " " "#")
|
||||
BarLeft=$(printf "%${Left}s" | tr " " "-")
|
||||
[[ "$QueTotal" == "$QueCurrent" ]] || printf "\r[%s%s] %s/%s " "$BarComplete" "$BarLeft" "$QueCurrent" "$QueTotal"
|
||||
[[ "$QueTotal" == "$QueCurrent" ]] && printf "\r[%b%s%b] %s/%s \n" "$c_teal" "$BarComplete" "$c_reset" "$QueCurrent" "$QueTotal"
|
||||
}
|
||||
|
||||
# Function to add user-provided urls to releasenotes
|
||||
releasenotes() {
|
||||
for update in ${GotUpdates[@]}; do
|
||||
found=false
|
||||
while read -r container url; do
|
||||
[[ $update == $container ]] && Updates+=("$update -> $url") && found=true
|
||||
done < "$ScriptWorkDir"/urls.list
|
||||
[[ $found == false ]] && Updates+=("$update -> url missing") || continue
|
||||
done
|
||||
}
|
||||
|
||||
# Version check & initiate self update
|
||||
if [[ "$VERSION" != "$LatestRelease" ]] ; then
|
||||
printf "New version available! %b%s%b ⇒ %b%s%b \n Change Notes: %s \n" "$c_yellow" "$VERSION" "$c_reset" "$c_green" "$LatestRelease" "$c_reset" "$LatestChanges"
|
||||
if [[ -z "$AutoUp" ]] ; then
|
||||
read -r -p "Would you like to update? y/[n]: " SelfUpdate
|
||||
[[ "$SelfUpdate" =~ [yY] ]] && self_update
|
||||
fi
|
||||
fi
|
||||
|
||||
# Set $1 to a variable for name filtering later
|
||||
SearchName="$1"
|
||||
# Create array of excludes
|
||||
IFS=',' read -r -a Excludes <<< "$Exclude" ; unset IFS
|
||||
|
||||
# Static binary downloader for dependencies
|
||||
binary_downloader() {
|
||||
BinaryName="$1"
|
||||
BinaryUrl="$2"
|
||||
case "$(uname -m)" in
|
||||
x86_64|amd64) architecture="amd64" ;;
|
||||
arm64|aarch64) architecture="arm64";;
|
||||
*) printf "\n%bArchitecture not supported, exiting.%b\n" "$c_red" "$c_reset" ; exit 1;;
|
||||
esac
|
||||
GetUrl="${BinaryUrl/TEMP/"$architecture"}"
|
||||
if [[ $(command -v curl) ]]; then curl -L $GetUrl > "$ScriptWorkDir/$BinaryName" ;
|
||||
elif [[ $(command -v wget) ]]; then wget $GetUrl -O "$ScriptWorkDir/$BinaryName" ;
|
||||
else printf "%s\n" "curl/wget not available - get $BinaryName manually from the repo link, exiting."; exit 1;
|
||||
fi
|
||||
[[ -f "$ScriptWorkDir/$BinaryName" ]] && chmod +x "$ScriptWorkDir/$BinaryName"
|
||||
}
|
||||
|
||||
distro_checker() {
|
||||
if [[ -f /etc/arch-release ]] ; then PkgInstaller="pacman -S"
|
||||
elif [[ -f /etc/redhat-release ]] ; then PkgInstaller="sudo dnf install"
|
||||
elif [[ -f /etc/SuSE-release ]] ; then PkgInstaller="sudo zypper install"
|
||||
elif [[ -f /etc/debian_version ]] ; then PkgInstaller="sudo apt-get install"
|
||||
elif [[ $(uname -s) == "Darwin" ]] ; then PkgInstaller="brew install"
|
||||
else PkgInstaller="ERROR" ; printf "\n%bNo distribution could be determined%b, falling back to static binary.\n" "$c_yellow" "$c_reset"
|
||||
fi
|
||||
}
|
||||
|
||||
# Dependency check + installer function
|
||||
dependency_check() {
|
||||
AppName="$1"
|
||||
AppVar="$2"
|
||||
AppUrl="$3"
|
||||
if [[ $(command -v $AppName) ]]; then export $AppVar="$AppName" ;
|
||||
elif [[ -f "$ScriptWorkDir/$AppName" ]]; then export $AppVar="$ScriptWorkDir/$AppName" ;
|
||||
else
|
||||
printf "%s\n" "Required dependency '$AppName' missing, do you want to install it?"
|
||||
read -r -p "y: With packagemanager (sudo). / s: Download static binary. y/s/[n] " GetBin
|
||||
GetBin=${GetBin:-no} # set default to no if nothing is given
|
||||
if [[ "$GetBin" =~ [yYsS] ]] ; then
|
||||
[[ "$GetBin" =~ [yY] ]] && distro_checker
|
||||
if [[ -n "$PkgInstaller" && "$PkgInstaller" != "ERROR" ]] ; then
|
||||
[[ $(uname -s) == "Darwin" && "$AppName" == "regctl" ]] && AppName="regclient"
|
||||
($PkgInstaller $AppName) ; PkgExitcode="$?" && AppName="$1"
|
||||
if [[ "$PkgExitcode" == 0 ]] ; then { export $AppVar="$AppName" && printf "\n%b$AppName installed.%b\n" "$c_green" "$c_reset"; }
|
||||
else printf "\n%bPackagemanager install failed%b, falling back to static binary.\n" "$c_yellow" "$c_reset"
|
||||
fi
|
||||
fi
|
||||
if [[ "$GetBin" =~ [sS] || "$PkgInstaller" == "ERROR" || "$PkgExitcode" != 0 ]] ; then
|
||||
binary_downloader "$AppName" "$AppUrl"
|
||||
[[ -f "$ScriptWorkDir/$AppName" ]] && { export $AppVar="$ScriptWorkDir/$1" && printf "\n%b$AppName downloaded.%b\n" "$c_green" "$c_reset"; }
|
||||
fi
|
||||
else printf "\n%bDependency missing, exiting.%b\n" "$c_red" "$c_reset" ; exit 1 ;
|
||||
fi
|
||||
fi
|
||||
# Final check if binary is correct
|
||||
[[ "$1" == "jq" ]] && VerFlag="--version"
|
||||
[[ "$1" == "regctl" ]] && VerFlag="version"
|
||||
${!AppVar} $VerFlag &> /dev/null || { printf "%s\n" "$AppName is not working - try to remove it and re-download it, exiting."; exit 1; }
|
||||
}
|
||||
|
||||
dependency_check "regctl" "regbin" "https://github.com/regclient/regclient/releases/latest/download/regctl-linux-TEMP"
|
||||
dependency_check "jq" "jqbin" "https://github.com/jqlang/jq/releases/latest/download/jq-linux-TEMP"
|
||||
|
||||
# Check docker compose binary
|
||||
if docker compose version &> /dev/null ; then DockerBin="docker compose" ;
|
||||
elif docker-compose -v &> /dev/null; then DockerBin="docker-compose" ;
|
||||
elif docker -v &> /dev/null; then
|
||||
printf "%s\n" "No docker compose binary available, using plain docker (Not recommended!)"
|
||||
printf "%s\n" "'docker run' will ONLY update images, not the container itself."
|
||||
else
|
||||
printf "%s\n" "No docker binaries available, exiting."
|
||||
exit 1
|
||||
fi
|
||||
|
||||
# Numbered List function
|
||||
options() {
|
||||
num=1
|
||||
for i in "${GotUpdates[@]}"; do
|
||||
echo "$num) $i"
|
||||
((num++))
|
||||
done
|
||||
}
|
||||
|
||||
# Listing typed exclusions
|
||||
if [[ -n ${Excludes[*]} ]] ; then
|
||||
printf "\n%bExcluding these names:%b\n" "$c_blue" "$c_reset"
|
||||
printf "%s\n" "${Excludes[@]}"
|
||||
printf "\n"
|
||||
fi
|
||||
|
||||
# Variables for progress_bar function
|
||||
ContCount=$(docker ps $Stopped --filter "name=$SearchName" --format '{{.Names}}' | wc -l)
|
||||
RegCheckQue=0
|
||||
|
||||
# Testing and setting timeout binary
|
||||
t_out=$(command -v timeout)
|
||||
if [[ $t_out ]]; then
|
||||
t_out=$(realpath $t_out 2>/dev/null || readlink -f $t_out)
|
||||
if [[ $t_out =~ "busybox" ]]; then
|
||||
t_out="timeout ${Timeout}"
|
||||
else t_out="timeout --foreground ${Timeout}"
|
||||
fi
|
||||
else t_out=""
|
||||
fi
|
||||
|
||||
check_image() {
|
||||
i="$1"
|
||||
local Excludes=($Excludes_string)
|
||||
for e in "${Excludes[@]}" ; do
|
||||
if [[ "$i" == "$e" ]]; then
|
||||
echo Skip $i
|
||||
return
|
||||
fi
|
||||
done
|
||||
|
||||
local NoUpdates GotUpdates GotErrors
|
||||
ImageId=$(docker inspect "$i" --format='{{.Image}}')
|
||||
RepoUrl=$(docker inspect "$i" --format='{{.Config.Image}}')
|
||||
LocalHash=$(docker image inspect "$ImageId" --format '{{.RepoDigests}}')
|
||||
|
||||
# Checking for errors while setting the variable
|
||||
if RegHash=$(${t_out} $regbin -v error image digest --list "$RepoUrl" 2>&1) ; then
|
||||
if [[ "$LocalHash" = *"$RegHash"* ]] ; then
|
||||
echo NoUpdates "$i"
|
||||
else
|
||||
if [[ -n "$DaysOld" ]] && ! datecheck ; then
|
||||
echo NoUpdates "+$i ${ImageAge}d"
|
||||
else
|
||||
echo GotUpdates "$i"
|
||||
fi
|
||||
fi
|
||||
else
|
||||
# Here the RegHash is the result of an error code
|
||||
echo GotErrors "$i - ${RegHash}"
|
||||
fi
|
||||
}
|
||||
|
||||
# Make required functions and variables available to subprocesses
|
||||
export -f check_image datecheck
|
||||
export Excludes_string="${Excludes[@]}" # Can only export scalar variables
|
||||
export t_out regbin RepoUrl DaysOld
|
||||
|
||||
# Check for POSIX xargs with -P option, fallback without async
|
||||
if (echo "test" | xargs -P 10 >/dev/null 2>&1) ; then
|
||||
XargsAsync="-P $MaxAsync"
|
||||
else
|
||||
XargsAsync=""
|
||||
printf "%bMissing POSIX xargs, consider installing 'findutils' for asynchronous lookups.%b\n" "$c_red" "$c_reset"
|
||||
fi
|
||||
|
||||
# Asynchronously check the image-hash of every running container VS the registry
|
||||
while read -r line; do
|
||||
((RegCheckQue+=1))
|
||||
progress_bar "$RegCheckQue" "$ContCount"
|
||||
|
||||
Got=${line%% *} # Extracts the first word (NoUpdates, GotUpdates, GotErrors)
|
||||
item=${line#* }
|
||||
|
||||
case "$Got" in
|
||||
NoUpdates) NoUpdates+=("$item") ;;
|
||||
GotUpdates) GotUpdates+=("$item") ;;
|
||||
GotErrors) GotErrors+=("$item") ;;
|
||||
Skip) ;;
|
||||
*) echo "Error! Unexpected output from subprocess: ${line}" ;;
|
||||
esac
|
||||
done < <( \
|
||||
docker ps $Stopped --filter "name=$SearchName" --format '{{.Names}}' | \
|
||||
xargs ${XargsAsync} -I {} bash -c 'check_image "{}"' \
|
||||
)
|
||||
|
||||
# Sort arrays alphabetically
|
||||
IFS=$'\n'
|
||||
NoUpdates=($(sort <<<"${NoUpdates[*]}"))
|
||||
GotUpdates=($(sort <<<"${GotUpdates[*]}"))
|
||||
GotErrors=($(sort <<<"${GotErrors[*]}"))
|
||||
unset IFS
|
||||
### Define how many updates are available
|
||||
|
||||
# Run the prometheus exporter function
|
||||
if [ -n "$CollectorTextFileDirectory" ] ; then
|
||||
source "$ScriptWorkDir"/addons/prometheus/prometheus_collector.sh && prometheus_exporter ${#NoUpdates[@]} ${#GotUpdates[@]} ${#GotErrors[@]}
|
||||
fi
|
||||
|
||||
# Define how many updates are available
|
||||
UpdCount="${#GotUpdates[@]}"
|
||||
|
||||
### List what containers got updates or not
|
||||
# List what containers got updates or not
|
||||
if [[ -n ${NoUpdates[*]} ]] ; then
|
||||
printf "\n\033[0;32mContainers on latest version:\033[0m\n"
|
||||
printf "\n%bContainers on latest version:%b\n" "$c_green" "$c_reset"
|
||||
printf "%s\n" "${NoUpdates[@]}"
|
||||
fi
|
||||
if [[ -n ${GotErrors[*]} ]] ; then
|
||||
printf "\n\033[0;31mContainers with errors, wont get updated:\033[0m\n"
|
||||
printf "\n%bContainers with errors, won't get updated:%b\n" "$c_red" "$c_reset"
|
||||
printf "%s\n" "${GotErrors[@]}"
|
||||
printf "%binfo:%b 'unauthorized' often means not found in a public registry.\n" "$c_blue" "$c_reset"
|
||||
fi
|
||||
if [[ -n ${GotUpdates[*]} ]] ; then
|
||||
printf "\n\033[0;33mContainers with updates available:\033[0m\n"
|
||||
[[ -z "$UpdYes" ]] && options || printf "%s\n" "${GotUpdates[@]}"
|
||||
if [[ -n ${GotUpdates[*]} ]] ; then
|
||||
printf "\n%bContainers with updates available:%b\n" "$c_yellow" "$c_reset"
|
||||
[[ -z "$AutoUp" ]] && options || printf "%s\n" "${GotUpdates[@]}"
|
||||
[[ -n "$Notify" ]] && { [[ $(type -t send_notification) == function ]] && send_notification "${GotUpdates[@]}" || printf "Could not source notification function.\n" ; }
|
||||
fi
|
||||
|
||||
### Optionally get updates if there's any
|
||||
# Optionally get updates if there's any
|
||||
if [ -n "$GotUpdates" ] ; then
|
||||
if [ -z "$UpdYes" ] ; then
|
||||
printf "\n\033[0;36mChoose what containers to update.\033[0m\n"
|
||||
choosecontainers
|
||||
if [ -z "$AutoUp" ] ; then
|
||||
printf "\n%bChoose what containers to update.%b\n" "$c_teal" "$c_reset"
|
||||
choosecontainers
|
||||
else
|
||||
SelectedUpdates=( "${GotUpdates[@]}" )
|
||||
fi
|
||||
if [ "$UpdYes" == "${UpdYes#[Nn]}" ] ; then
|
||||
if [ "$AutoUp" == "${AutoUp#[Nn]}" ] ; then
|
||||
NumberofUpdates="${#SelectedUpdates[@]}"
|
||||
CurrentQue=0
|
||||
for i in "${SelectedUpdates[@]}"
|
||||
do
|
||||
((CurrentQue+=1))
|
||||
unset CompleteConfs
|
||||
ContPath=$(docker inspect "$i" --format '{{ index .Config.Labels "com.docker.compose.project.working_dir" }}')
|
||||
ContConfigFile=$(docker inspect "$i" --format '{{ index .Config.Labels "com.docker.compose.project.config_files" }}')
|
||||
ContName=$(docker inspect "$i" --format '{{ index .Config.Labels "com.docker.compose.service" }}')
|
||||
ContEnv=$(docker inspect "$i" --format '{{index .Config.Labels "com.docker.compose.project.environment_file" }}')
|
||||
# Extract labels and metadata
|
||||
ContLabels=$(docker inspect "$i" --format '{{json .Config.Labels}}')
|
||||
ContImage=$(docker inspect "$i" --format='{{.Config.Image}}')
|
||||
### Checking if compose-values are empty - hence started with docker run:
|
||||
if [ -z "$ContPath" ] ; then
|
||||
if [ "$DrUp" == "yes" ] ; then
|
||||
ContPath=$($jqbin -r '."com.docker.compose.project.working_dir"' <<< "$ContLabels")
|
||||
[ "$ContPath" == "null" ] && ContPath=""
|
||||
ContConfigFile=$($jqbin -r '."com.docker.compose.project.config_files"' <<< "$ContLabels")
|
||||
[ "$ContConfigFile" == "null" ] && ContConfigFile=""
|
||||
ContName=$($jqbin -r '."com.docker.compose.service"' <<< "$ContLabels")
|
||||
[ "$ContName" == "null" ] && ContName=""
|
||||
ContEnv=$($jqbin -r '."com.docker.compose.project.environment_file"' <<< "$ContLabels")
|
||||
[ "$ContEnv" == "null" ] && ContEnv=""
|
||||
ContUpdateLabel=$($jqbin -r '."mag37.dockcheck.update"' <<< "$ContLabels")
|
||||
[ "$ContUpdateLabel" == "null" ] && ContUpdateLabel=""
|
||||
ContRestartStack=$($jqbin -r '."mag37.dockcheck.restart-stack"' <<< "$ContLabels")
|
||||
[ "$ContRestartStack" == "null" ] && ContRestartStack=""
|
||||
|
||||
# Checking if compose-values are empty - hence started with docker run
|
||||
if [ -z "$ContPath" ] ; then
|
||||
if [ "$DRunUp" == "yes" ] ; then
|
||||
docker pull "$ContImage"
|
||||
printf "%s\n" "$i got a new image downloaded, rebuild manually with preferred 'docker run'-parameters"
|
||||
else
|
||||
printf "\n\033[33;1m%s\033[0m has no compose labels, probably started with docker run - \033[33;1mskipping\033[0m\n\n" "$i"
|
||||
printf "\n%b%s%b has no compose labels, probably started with docker run - %bskipping%b\n\n" "$c_yellow" "$i" "$c_reset" "$c_yellow" "$c_reset"
|
||||
fi
|
||||
continue
|
||||
continue
|
||||
fi
|
||||
### Checking if "com.docker.compose.project.config_files" returns the full path to the config file or just the file name
|
||||
if [[ $ContConfigFile = '/'* ]] ; then
|
||||
ComposeFile="$ContConfigFile"
|
||||
else
|
||||
ComposeFile="$ContPath/$ContConfigFile"
|
||||
fi
|
||||
### cd to the compose-file directory to account for people who use relative volumes, eg - ${PWD}/data:data
|
||||
# cd to the compose-file directory to account for people who use relative volumes
|
||||
cd "$ContPath" || { echo "Path error - skipping $i" ; continue ; }
|
||||
docker pull "$ContImage"
|
||||
### Reformat for multi-compose:
|
||||
IFS=',' read -r -a Confs <<< "$ContConfigFile" ; unset IFS
|
||||
for conf in "${Confs[@]}"; do CompleteConfs+="-f $conf " ; done
|
||||
|
||||
### Check if the container got an environment file set, use it if so:
|
||||
if [ -n "$ContEnv" ]; then
|
||||
$DockerBin ${CompleteConfs[@]} --env-file "$ContEnv" up -d "$ContName" # unquoted array to allow split - rework?
|
||||
## Reformatting path + multi compose
|
||||
if [[ $ContConfigFile = '/'* ]] ; then
|
||||
CompleteConfs=$(for conf in ${ContConfigFile//,/ } ; do printf -- "-f %s " "$conf"; done)
|
||||
else
|
||||
$DockerBin ${CompleteConfs[@]} up -d "$ContName" # unquoted array to allow split - rework?
|
||||
CompleteConfs=$(for conf in ${ContConfigFile//,/ } ; do printf -- "-f %s/%s " "$ContPath" "$conf"; done)
|
||||
fi
|
||||
printf "\n%bNow updating (%s/%s): %b%s%b\n" "$c_teal" "$CurrentQue" "$NumberofUpdates" "$c_blue" "$i" "$c_reset"
|
||||
# Checking if Label Only -option is set, and if container got the label
|
||||
[[ "$OnlyLabel" == true ]] && { [[ "$ContUpdateLabel" != true ]] && { echo "No update label, skipping." ; continue ; } }
|
||||
docker pull "$ContImage"
|
||||
# Check if the container got an environment file set and reformat it
|
||||
if [ -n "$ContEnv" ]; then ContEnvs=$(for env in ${ContEnv//,/ } ; do printf -- "--env-file %s " "$env"; done) ; fi
|
||||
# Check if the whole stack should be restarted
|
||||
if [[ "$ContRestartStack" == true ]] || [[ "$ForceRestartStacks" == true ]] ; then
|
||||
$DockerBin ${CompleteConfs} stop ; $DockerBin ${CompleteConfs} ${ContEnvs} up -d
|
||||
else
|
||||
$DockerBin ${CompleteConfs} ${ContEnvs} up -d ${ContName}
|
||||
fi
|
||||
done
|
||||
printf "\033[0;32mAll done!\033[0m\n"
|
||||
[[ -z "$PruneQ" ]] && read -r -p "Would you like to prune dangling images? y/[n]: " PruneQ
|
||||
[[ "$PruneQ" =~ [yY] ]] && docker image prune -f
|
||||
printf "\n%bAll done!%b\n" "$c_green" "$c_reset"
|
||||
if [[ -z "$AutoPrune" ]] && [[ -z "$AutoUp" ]]; then read -r -p "Would you like to prune dangling images? y/[n]: " AutoPrune ; fi
|
||||
[[ "$AutoPrune" =~ [yY] ]] && docker image prune -f
|
||||
else
|
||||
printf "\nNo updates installed, exiting.\n"
|
||||
fi
|
||||
|
||||
@@ -1,26 +0,0 @@
|
||||
#!/usr/bin/env bash
|
||||
SearchName="$1"
|
||||
for i in $(docker ps --filter "name=$SearchName" --format '{{.Names}}') ; do
|
||||
echo "------------ $i ------------"
|
||||
ContPath=$(docker inspect "$i" --format '{{ index .Config.Labels "com.docker.compose.project.working_dir" }}')
|
||||
[ -z "$ContPath" ] && { "$i has no compose labels - skipping" ; continue ; }
|
||||
ContConfigFile=$(docker inspect "$i" --format '{{ index .Config.Labels "com.docker.compose.project.config_files" }}')
|
||||
ContName=$(docker inspect "$i" --format '{{ index .Config.Labels "com.docker.compose.service" }}')
|
||||
ContEnv=$(docker inspect "$i" --format '{{index .Config.Labels "com.docker.compose.project.environment_file" }}')
|
||||
ContImage=$(docker inspect "$i" --format='{{.Config.Image}}')
|
||||
|
||||
if [[ $ContConfigFile = '/'* ]] ; then
|
||||
ComposeFile="$ContConfigFile"
|
||||
else
|
||||
ComposeFile="$ContPath/$ContConfigFile"
|
||||
fi
|
||||
|
||||
echo -e "Service name:\t\t$ContName"
|
||||
echo -e "Project working dir:\t$ContPath"
|
||||
echo -e "Compose files:\t\t$ComposeFile"
|
||||
echo -e "Container image:\t$ContImage"
|
||||
echo
|
||||
echo "Mounts:"
|
||||
docker inspect -f '{{ range .Mounts }}{{ .Source }}:{{ .Destination }}{{ printf "\n" }}{{ end }}' $i
|
||||
echo
|
||||
done
|
||||
BIN
example.gif
|
Before Width: | Height: | Size: 333 KiB |
BIN
extras/apprise-ex1.png
Normal file
|
After Width: | Height: | Size: 155 KiB |
BIN
extras/apprise-ex2.png
Normal file
|
After Width: | Height: | Size: 114 KiB |
130
extras/apprise_quickstart.md
Normal file
@@ -0,0 +1,130 @@
|
||||
# A small guide on getting started with Apprise notifications.
|
||||
|
||||
|
||||
## Standalone docker container: [linuxserver/apprise-api](https://hub.docker.com/r/linuxserver/apprise-api)
|
||||
|
||||
Set up the docker compose as preferred:
|
||||
```yaml
|
||||
---
|
||||
version: "2.1"
|
||||
services:
|
||||
apprise-api:
|
||||
image: lscr.io/linuxserver/apprise-api:latest
|
||||
container_name: apprise-api
|
||||
environment:
|
||||
- PUID=1000
|
||||
- PGID=1000
|
||||
- TZ=Etc/UTC
|
||||
volumes:
|
||||
- /path/to/apprise-api/config:/config
|
||||
ports:
|
||||
- 8000:8000
|
||||
restart: unless-stopped
|
||||
```
|
||||
|
||||
Then browse to the webui.
|
||||

|
||||
Here you'll click **Configuration Manager**, read the overview and then click on **Configuration**.
|
||||
Under **Configuration** you'll craft/paste your notification config.
|
||||
|
||||

|
||||
The simplest way is just paste the url's as is (like in the example above).
|
||||
There are many ways to customize with tags, groups, json and more. Read [caronc/apprise-api](https://github.com/caronc/apprise-api) for more info!
|
||||
|
||||
Look at the [apprise wiki: Notification Services](https://github.com/caronc/apprise/wiki) for more info about how the url syntax for different services works.
|
||||
|
||||
|
||||
You can also use the [caronc/apprise-api](https://github.com/caronc/apprise-api) to host the api as a frontend to an already existing **Apprise**-setup on the host.
|
||||
|
||||
|
||||
### Customize the **notify.sh** file.
|
||||
After you're done with the setup of the container and tried your notifications, you can copy the `notify_apprise.sh` file to `notify.sh` and start editing it.
|
||||
|
||||
Comment out/remove the bare metal apprise-command (starting with `apprise -vv -t...`).
|
||||
Uncomment and edit the `AppriseURL` variable and *curl* line
|
||||
It should look something like this when curling the API:
|
||||
```bash
|
||||
send_notification() {
|
||||
Updates=("$@")
|
||||
UpdToString=$( printf "%s\n" "${Updates[@]}" )
|
||||
FromHost=$(hostname)
|
||||
|
||||
printf "\nSending Apprise notification\n"
|
||||
|
||||
MessageTitle="$FromHost - updates available."
|
||||
# Setting the MessageBody variable here.
|
||||
read -d '\n' MessageBody << __EOF
|
||||
Containers on $FromHost with updates available:
|
||||
|
||||
$UpdToString
|
||||
|
||||
__EOF
|
||||
|
||||
AppriseURL="http://IP.or.mydomain.tld:8000/notify/apprise"
|
||||
curl -X POST -F "title=$MessageTitle" -F "body=$MessageBody" -F "tags=all" $AppriseURL
|
||||
|
||||
}
|
||||
```
|
||||
|
||||
That's all!
|
||||
___
|
||||
___
|
||||
|
||||
|
||||
## On host installed **Apprise**
|
||||
Follow the official guide on [caronc/apprise](https://github.com/caronc/apprise)!
|
||||
|
||||
### A brief, basic "get started"
|
||||
|
||||
- Install **apprise**
|
||||
- python package `pip install apprise`
|
||||
- packaged in EPEL/Fedora `dnf install apprise`
|
||||
- packaged in AUR `[yay/pikaur/paru/other] apprise`
|
||||
|
||||
- Create a config file with your notification credentials (source of notifications):
|
||||
```ini
|
||||
mailto://user:password@yahoo.com
|
||||
slack://token_a/token_b/token_c
|
||||
kodi://example.com
|
||||
```
|
||||
Then either source the notifications with `-c=/path/to/config/apprise` or store them in *PATH* to skip referencing (`~/.apprise` or `~/.config/apprise`).
|
||||
|
||||
- Test apprise with a single notification:
|
||||
- `apprise -vv -t 'test title' -b 'test notification body' 'mailto://myemail:mypass@gmail.com'`
|
||||
- Set up your notification URL's and test them.
|
||||
- Look at the [apprise wiki: Notification Services](https://github.com/caronc/apprise/wiki) for more info about how the url syntax for different services works.
|
||||
|
||||
### When done, customize the **notify.sh** file.
|
||||
After you're done with the setup of the container and tried your notifications, you can copy the `notify_apprise.sh` file to `notify.sh` and start editing it.
|
||||
|
||||
Replace the url's corresponding to the services you've configured.
|
||||
```bash
|
||||
send_notification() {
|
||||
Updates=("$@")
|
||||
UpdToString=$( printf "%s\n" "${Updates[@]}" )
|
||||
FromHost=$(hostname)
|
||||
|
||||
printf "\nSending Apprise notification\n"
|
||||
|
||||
MessageTitle="$FromHost - updates available."
|
||||
# Setting the MessageBody variable here.
|
||||
read -d '\n' MessageBody << __EOF
|
||||
Containers on $FromHost with updates available:
|
||||
|
||||
$UpdToString
|
||||
|
||||
__EOF
|
||||
|
||||
# Modify to fit your setup:
|
||||
apprise -vv -t "$MessageTitle" -b "$MessageBody" \
|
||||
mailto://myemail:mypass@gmail.com \
|
||||
mastodons://{token}@{host} \
|
||||
pbul://o.gn5kj6nfhv736I7jC3cj3QLRiyhgl98b \
|
||||
tgram://{bot_token}/{chat_id}/
|
||||
|
||||
}
|
||||
```
|
||||
|
||||
That's all!
|
||||
___
|
||||
___
|
||||
@@ -16,7 +16,7 @@ IFS=',' read -r -a Excludes <<< "$Exclude" ; unset IFS
|
||||
SearchName="$1"
|
||||
|
||||
for i in $(docker ps --filter "name=$SearchName" --format '{{.Names}}') ; do
|
||||
[[ " ${Excludes[*]} " =~ ${i} ]] && continue; # Skip if the container is excluded
|
||||
for e in "${Excludes[@]}" ; do [[ "$i" == "$e" ]] && continue 2 ; done
|
||||
printf ". "
|
||||
RepoUrl=$(docker inspect "$i" --format='{{.Config.Image}}')
|
||||
LocalHash=$(docker image inspect "$RepoUrl" --format '{{.RepoDigests}}')
|
||||
@@ -29,7 +29,7 @@ for i in $(docker ps --filter "name=$SearchName" --format '{{.Names}}') ; do
|
||||
done
|
||||
|
||||
### Sort arrays alphabetically
|
||||
IFS=$'\n'
|
||||
IFS=$'\n'
|
||||
NoUpdates=($(sort <<<"${NoUpdates[*]}"))
|
||||
GotUpdates=($(sort <<<"${GotUpdates[*]}"))
|
||||
GotErrors=($(sort <<<"${GotErrors[*]}"))
|
||||
@@ -44,7 +44,7 @@ if [[ -n ${GotErrors[*]} ]] ; then
|
||||
printf "\n\033[0;31mContainers with errors, wont get updated:\033[0m\n"
|
||||
printf "%s\n" "${GotErrors[@]}"
|
||||
fi
|
||||
if [[ -n ${GotUpdates[*]} ]] ; then
|
||||
if [[ -n ${GotUpdates[*]} ]] ; then
|
||||
printf "\n\033[0;33mContainers with updates available:\033[0m\n"
|
||||
printf "%s\n" "${GotUpdates[@]}"
|
||||
fi
|
||||
BIN
extras/dockcheck_logo_by_booYah187.png
Normal file
|
After Width: | Height: | Size: 46 KiB |
38
extras/errorCheck.sh
Executable file
@@ -0,0 +1,38 @@
|
||||
#!/usr/bin/env bash
|
||||
SearchName="$1"
|
||||
for i in $(docker ps --filter "name=$SearchName" --format '{{.Names}}') ; do
|
||||
echo "------------ $i ------------"
|
||||
ContLabels=$(docker inspect "$i" --format '{{json .Config.Labels}}')
|
||||
ContImage=$(docker inspect "$i" --format='{{.Config.Image}}')
|
||||
ContPath=$(jq -r '."com.docker.compose.project.working_dir"' <<< "$ContLabels")
|
||||
[ "$ContPath" == "null" ] && ContPath=""
|
||||
[ -z "$ContPath" ] && { "$i has no compose labels - skipping" ; continue ; }
|
||||
ContConfigFile=$(jq -r '."com.docker.compose.project.config_files"' <<< "$ContLabels")
|
||||
[ "$ContConfigFile" == "null" ] && ContConfigFile=""
|
||||
ContName=$(jq -r '."com.docker.compose.service"' <<< "$ContLabels")
|
||||
[ "$ContName" == "null" ] && ContName=""
|
||||
ContEnv=$(jq -r '."com.docker.compose.project.environment_file"' <<< "$ContLabels")
|
||||
[ "$ContEnv" == "null" ] && ContEnv=""
|
||||
ContUpdateLabel=$(jq -r '."mag37.dockcheck.update"' <<< "$ContLabels")
|
||||
[ "$ContUpdateLabel" == "null" ] && ContUpdateLabel=""
|
||||
ContRestartStack=$(jq -r '."mag37.dockcheck.restart-stack"' <<< "$ContLabels")
|
||||
[ "$ContRestartStack" == "null" ] && ContRestartStack=""
|
||||
|
||||
if [[ $ContConfigFile = '/'* ]] ; then
|
||||
ComposeFile="$ContConfigFile"
|
||||
else
|
||||
ComposeFile="$ContPath/$ContConfigFile"
|
||||
fi
|
||||
|
||||
echo -e "Service name:\t\t$ContName"
|
||||
echo -e "Project working dir:\t$ContPath"
|
||||
echo -e "Compose files:\t\t$ComposeFile"
|
||||
echo -e "Environment files:\t$ContEnv"
|
||||
echo -e "Container image:\t$ContImage"
|
||||
echo -e "Update label:\t$ContUpdateLabel"
|
||||
echo -e "Restart Stack label:\t$ContRestartStack"
|
||||
echo
|
||||
echo "Mounts:"
|
||||
docker inspect -f '{{ range .Mounts }}{{ .Source }}:{{ .Destination }}{{ printf "\n" }}{{ end }}' "$i"
|
||||
echo
|
||||
done
|
||||
BIN
extras/example.gif
Normal file
|
After Width: | Height: | Size: 472 KiB |
BIN
extras/example_old.gif
Normal file
|
After Width: | Height: | Size: 404 KiB |
1
extras/temp
Normal file
@@ -0,0 +1 @@
|
||||
temp
|
||||
54
notify_templates/notify_DSM.sh
Normal file
@@ -0,0 +1,54 @@
|
||||
### DISCLAIMER: This is a third party addition to dockcheck - best effort testing.
|
||||
# INFO: ssmtp is depcerated - consider to use msmtp instead.
|
||||
#
|
||||
# Copy/rename this file to notify.sh to enable the notification snipppet.
|
||||
# mSMTP/sSMTP has to be installed and configured manually.
|
||||
# The existing DSM Notification Email configuration will be used automatically.
|
||||
# Modify to your liking - changing SendMailTo and Subject and content.
|
||||
|
||||
MSMTP=$(which msmtp)
|
||||
SSMTP=$(which ssmtp)
|
||||
|
||||
if [ -n "$MSMTP" ] ; then
|
||||
MailPkg=$MSMTP
|
||||
elif [ -n "$SSMTP" ] ; then
|
||||
MailPkg=$SSMTP
|
||||
else
|
||||
echo "No msmtp or ssmtp binary found in PATH: $PATH" ; exit 1
|
||||
fi
|
||||
|
||||
send_notification() {
|
||||
[ -s "$ScriptWorkDir"/urls.list ] && releasenotes || Updates=("$@")
|
||||
UpdToString=$( printf '%s\\n' "${Updates[@]}" )
|
||||
FromHost=$(hostname)
|
||||
CfgFile="/usr/syno/etc/synosmtp.conf"
|
||||
|
||||
# User variables:
|
||||
# Automatically sends to your usual destination for synology DSM notification emails.
|
||||
# You can also manually override by assigning something else to SendMailTo below.
|
||||
SendMailTo=$(grep 'eventmail1' $CfgFile | sed -n 's/.*"\([^"]*\)".*/\1/p')
|
||||
#SendMailTo="me@mydomain.com"
|
||||
|
||||
SubjectTag=$(grep 'eventsubjectprefix' $CfgFile | sed -n 's/.*"\([^"]*\)".*/\1/p')
|
||||
SenderName=$(grep 'smtp_from_name' $CfgFile | sed -n 's/.*"\([^"]*\)".*/\1/p')
|
||||
SenderMail=$(grep 'smtp_from_mail' $CfgFile | sed -n 's/.*"\([^"]*\)".*/\1/p')
|
||||
SenderMail=${SenderMail:-$(grep 'eventmail1' $CfgFile | sed -n 's/.*"\([^"]*\)".*/\1/p')}
|
||||
|
||||
printf "\nSending email notification.\n"
|
||||
|
||||
printf -v MessageBody "🐋 Containers on $FromHost with updates available:\n\n$UpdToString"
|
||||
|
||||
$MailPkg $SendMailTo << __EOF
|
||||
From: "$SenderName" <$SenderMail>
|
||||
date:$(date -R)
|
||||
To: <$SendMailTo>
|
||||
Subject: $SubjectTag Updates available on $FromHost
|
||||
Content-Type: text/plain; charset=UTF-8; format=flowed
|
||||
Content-Transfer-Encoding: 7bit
|
||||
|
||||
$MessageBody
|
||||
From $SenderName
|
||||
__EOF
|
||||
# This ensures DSM's container manager will also see the update
|
||||
/var/packages/ContainerManager/target/tool/image_upgradable_checker
|
||||
}
|
||||
30
notify_templates/notify_apprise.sh
Normal file
@@ -0,0 +1,30 @@
|
||||
### DISCLAIMER: This is a third party addition to dockcheck - best effort testing.
|
||||
#
|
||||
# Copy/rename this file to notify.sh to enable the notification snippet.
|
||||
# Required receiving services must already be set up.
|
||||
# Modify to fit your setup - if API, set AppriseURL to your Apprise ip/domain.
|
||||
|
||||
send_notification() {
|
||||
[ -s "$ScriptWorkDir"/urls.list ] && releasenotes || Updates=("$@")
|
||||
UpdToString=$( printf '%s\\n' "${Updates[@]}" )
|
||||
FromHost=$(hostname)
|
||||
|
||||
printf "\nSending Apprise notification\n"
|
||||
|
||||
MessageTitle="$FromHost - updates available."
|
||||
# Setting the MessageBody variable here.
|
||||
printf -v MessageBody "🐋 Containers on $FromHost with updates available:\n$UpdToString"
|
||||
|
||||
# Modify to fit your setup:
|
||||
apprise -vv -t "$MessageTitle" -b "$MessageBody" \
|
||||
mailto://myemail:mypass@gmail.com \
|
||||
mastodons://{token}@{host} \
|
||||
pbul://o.gn5kj6nfhv736I7jC3cj3QLRiyhgl98b \
|
||||
tgram://{bot_token}/{chat_id}/
|
||||
|
||||
### If you use the Apprise-API - Comment out the apprise command above.
|
||||
### Uncomment the AppriseURL and the curl-line below:
|
||||
# AppriseURL="http://apprise.mydomain.tld:1234/notify/apprise"
|
||||
# curl -X POST -F "title=$MessageTitle" -F "body=$MessageBody" -F "tags=all" $AppriseURL
|
||||
|
||||
}
|
||||
28
notify_templates/notify_discord.sh
Normal file
@@ -0,0 +1,28 @@
|
||||
### DISCLAIMER: This is a third party addition to dockcheck - best effort testing.
|
||||
#
|
||||
# Copy/rename this file to notify.sh to enable the notification snippet.
|
||||
# Required receiving services must already be set up.
|
||||
# Modify to fit your setup - set DiscordWebhookUrl
|
||||
|
||||
send_notification() {
|
||||
[ -s "$ScriptWorkDir"/urls.list ] && releasenotes || Updates=("$@")
|
||||
UpdToString=$( printf '%s\\n' "${Updates[@]}" )
|
||||
|
||||
echo "$UpdToString"
|
||||
FromHost=$(hostname)
|
||||
|
||||
# platform specific notification code would go here
|
||||
printf "\nSending Discord notification\n"
|
||||
|
||||
# Setting the MessageBody variable here.
|
||||
MessageBody="🐋 Containers on $FromHost with updates available: \n$UpdToString"
|
||||
|
||||
# Modify to fit your setup:
|
||||
DiscordWebhookUrl="PasteYourFullDiscordWebhookURL"
|
||||
|
||||
MsgBody="{\"username\":\"$FromHost\",\"content\":\"$MessageBody\"}"
|
||||
|
||||
curl -sS -o /dev/null --fail -X POST -H "Content-Type: application/json" -d "$MsgBody" "$DiscordWebhookUrl"
|
||||
|
||||
}
|
||||
|
||||
16
notify_templates/notify_generic.sh
Normal file
@@ -0,0 +1,16 @@
|
||||
### DISCLAIMER: This is a third party addition to dockcheck - best effort testing.
|
||||
#
|
||||
# Copy/rename this file to notify.sh to enable the notification snippet.
|
||||
# generic sample, the "Hello World" of notification addons
|
||||
|
||||
send_notification() {
|
||||
[ -s "$ScriptWorkDir"/urls.list ] && releasenotes || Updates=("$@")
|
||||
UpdToString=$( printf '%s\\n' "${Updates[@]}" )
|
||||
|
||||
FromHost=$(hostname)
|
||||
|
||||
# platform specific notification code would go here
|
||||
printf "\n%bGeneric notification addon:%b" "$c_green" "$c_reset"
|
||||
printf "\nThe following docker containers on %s need to be updated:\n" "$FromHost"
|
||||
printf "$UpdToString"
|
||||
}
|
||||
29
notify_templates/notify_gotify.sh
Normal file
@@ -0,0 +1,29 @@
|
||||
### DISCLAIMER: This is a third party addition to dockcheck - best effort testing.
|
||||
#
|
||||
# Copy/rename this file to notify.sh to enable the notification snippet.
|
||||
# Required receiving services must already be set up.
|
||||
# Modify to fit your setup - set GotifyUrl and GotifyToken.
|
||||
|
||||
send_notification() {
|
||||
[ -s "$ScriptWorkDir"/urls.list ] && releasenotes || Updates=("$@")
|
||||
UpdToString=$( printf '%s\\n' "${Updates[@]}" )
|
||||
FromHost=$(hostname)
|
||||
|
||||
# platform specific notification code would go here
|
||||
printf "\nSending Gotify notification\n"
|
||||
|
||||
# Setting the MessageTitle and MessageBody variable here.
|
||||
MessageTitle="${FromHost} - updates available."
|
||||
printf -v MessageBody "Containers on $FromHost with updates available:\n$UpdToString"
|
||||
|
||||
# Modify to fit your setup:
|
||||
GotifyToken="Your Gotify token here"
|
||||
GotifyUrl="https://api.gotify/message?token=${GotifyToken}"
|
||||
|
||||
curl \
|
||||
-F "title=${MessageTitle}" \
|
||||
-F "message=${MessageBody}" \
|
||||
-F "priority=5" \
|
||||
-X POST "${GotifyUrl}" 1> /dev/null
|
||||
|
||||
}
|
||||
28
notify_templates/notify_matrix.sh
Normal file
@@ -0,0 +1,28 @@
|
||||
### DISCLAIMER: This is a third party addition to dockcheck - best effort testing.
|
||||
#
|
||||
# Copy/rename this file to notify.sh to enable the notification snippet.
|
||||
# Required receiving services must already be set up.
|
||||
# Modify to fit your setup - set MatrixServer, Room_id and AccessToken
|
||||
|
||||
send_notification() {
|
||||
[ -s "$ScriptWorkDir"/urls.list ] && releasenotes || Updates=("$@")
|
||||
UpdToString=$( printf '%s\\n' "${Updates[@]}" )
|
||||
FromHost=$(hostname)
|
||||
|
||||
# platform specific notification code would go here
|
||||
printf "\nSending Matrix notification\n"
|
||||
|
||||
# Setting the MessageBody variable here.
|
||||
MessageBody="🐋 Containers on $FromHost with updates available: \n$UpdToString"
|
||||
|
||||
# Modify to fit your setup:
|
||||
AccessToken="Your Matrix token here"
|
||||
Room_id="Enter Room_id here"
|
||||
MatrixServer="Enter Your HomeServer URL"
|
||||
MsgBody="{\"msgtype\":\"m.text\",\"body\":\"$MessageBody\"}"
|
||||
|
||||
# URL Example: https://matrix.org/_matrix/client/r0/rooms/!xxxxxx:example.com/send/m.room.message?access_token=xxxxxxxx
|
||||
|
||||
curl -sS -o /dev/null --fail -X POST "$MatrixServer/_matrix/client/r0/rooms/$Room_id/send/m.room.message?access_token=$AccessToken" -H 'Content-Type: application/json' -d "$MsgBody"
|
||||
|
||||
}
|
||||
26
notify_templates/notify_ntfy-sh.sh
Normal file
@@ -0,0 +1,26 @@
|
||||
### DISCLAIMER: This is a third party addition to dockcheck - best effort testing.
|
||||
#
|
||||
# Copy/rename this file to notify.sh to enable the notification snippet.
|
||||
# Setup app and subscription at https://ntfy.sh
|
||||
# Use your unique Topic Name in the URL below.
|
||||
|
||||
send_notification() {
|
||||
[ -s "$ScriptWorkDir"/urls.list ] && releasenotes || Updates=("$@")
|
||||
UpdToString=$( printf '%s\\n' "${Updates[@]}" )
|
||||
FromHost=$(hostname)
|
||||
|
||||
printf "\nSending ntfy.sh notification\n"
|
||||
|
||||
MessageTitle="$FromHost - updates available."
|
||||
# Setting the MessageBody variable here.
|
||||
printf -v MessageBody "🐋 Containers on $FromHost with updates available:\n$UpdToString"
|
||||
|
||||
# Modify to fit your setup:
|
||||
NtfyUrl="ntfy.sh/YourUniqueTopicName"
|
||||
|
||||
curl -sS -o /dev/null --show-error --fail \
|
||||
-H "Title: $MessageTitle" \
|
||||
-d "$MessageBody" \
|
||||
$NtfyUrl
|
||||
|
||||
}
|
||||
27
notify_templates/notify_pushbullet.sh
Normal file
@@ -0,0 +1,27 @@
|
||||
### DISCLAIMER: This is a third party addition to dockcheck - best effort testing.
|
||||
#
|
||||
# Copy/rename this file to notify.sh to enable the notification snippet.
|
||||
# Required receiving services must already be set up.
|
||||
# Requires jq installed and in PATH.
|
||||
# Modify to fit your setup - set Url and Token.
|
||||
|
||||
send_notification() {
|
||||
[ -s "$ScriptWorkDir"/urls.list ] && releasenotes || Updates=("$@")
|
||||
UpdToString=$( printf '%s\\n' "${Updates[@]}" )
|
||||
FromHost=$(hostname)
|
||||
|
||||
# platform specific notification code would go here
|
||||
printf "\nSending pushbullet notification\n"
|
||||
|
||||
MessageTitle="$FromHost - updates available."
|
||||
# Setting the MessageBody variable here.
|
||||
printf -v MessageBody "🐋 Containers on $FromHost with updates available:\n$UpdToString"
|
||||
|
||||
# Modify to fit your setup:
|
||||
PushUrl="https://api.pushbullet.com/v2/pushes"
|
||||
PushToken="Your Pushbullet token here"
|
||||
|
||||
# Requires jq to process json data
|
||||
jq -n --arg title "$MessageTitle" --arg body "$MessageBody" '{body: $body, title: $title, type: "note"}' | curl -sS -o /dev/null --show-error --fail -X POST -H "Access-Token: $PushToken" -H "Content-type: application/json" $PushUrl -d @-
|
||||
|
||||
}
|
||||
32
notify_templates/notify_pushover.sh
Normal file
@@ -0,0 +1,32 @@
|
||||
### DISCLAIMER: This is a third party addition to dockcheck - best effort testing.
|
||||
#
|
||||
# Copy/rename this file to notify.sh to enable the notification snippet.
|
||||
# Required receiving services must already be set up.
|
||||
# Requires jq installed and in PATH.
|
||||
# Modify to fit your setup - set Url and Token.
|
||||
|
||||
send_notification() {
|
||||
[ -s "$ScriptWorkDir"/urls.list ] && releasenotes || Updates=("$@")
|
||||
UpdToString=$( printf '%s\\n' "${Updates[@]}" )
|
||||
FromHost=$(hostname)
|
||||
|
||||
# platform specific notification code would go here
|
||||
printf "\nSending pushover notification\n"
|
||||
|
||||
MessageTitle="$FromHost - updates available."
|
||||
# Setting the MessageBody variable here.
|
||||
printf -v MessageBody "🐋 Containers on $FromHost with updates available:\n$UpdToString"
|
||||
|
||||
# Modify to fit your setup:
|
||||
PushoverUrl="https://api.pushover.net/1/messages.json"
|
||||
PushoverUserKey="Your Pushover User Key Here"
|
||||
PushoverToken="Your Pushover API Token Here"
|
||||
|
||||
# Sending the notification via Pushover
|
||||
curl -sS -o /dev/null --show-error --fail -X POST \
|
||||
-F "token=$PushoverToken" \
|
||||
-F "user=$PushoverUserKey" \
|
||||
-F "title=$MessageTitle" \
|
||||
-F "message=$MessageBody" \
|
||||
$PushoverUrl
|
||||
}
|
||||
44
notify_templates/notify_smtp.sh
Normal file
@@ -0,0 +1,44 @@
|
||||
### DISCLAIMER: This is a third party addition to dockcheck - best effort testing.
|
||||
# INFO: ssmtp is depcerated - consider to use msmtp instead.
|
||||
#
|
||||
# Copy/rename this file to notify.sh to enable the notification snipppet.
|
||||
# mSMTP/sSMTP has to be installed and configured manually.
|
||||
# Modify to fit your setup - changing SendMailFrom, SendMailTo, SubjectTag
|
||||
|
||||
MSMTP=$(which msmtp)
|
||||
SSMTP=$(which ssmtp)
|
||||
|
||||
if [ -n "$MSMTP" ] ; then
|
||||
MailPkg=$MSMTP
|
||||
elif [ -n "$SSMTP" ] ; then
|
||||
MailPkg=$SSMTP
|
||||
else
|
||||
echo "No msmtp or ssmtp binary found in PATH: $PATH" ; exit 1
|
||||
fi
|
||||
|
||||
send_notification() {
|
||||
[ -s "$ScriptWorkDir"/urls.list ] && releasenotes || Updates=("$@")
|
||||
UpdToString=$( printf '%s\\n' "${Updates[@]}" )
|
||||
FromHost=$(hostname)
|
||||
|
||||
# User variables:
|
||||
SendMailFrom="me@mydomain.tld"
|
||||
SendMailTo="me@mydomain.tld"
|
||||
SubjectTag="dockcheck"
|
||||
|
||||
printf "\nSending email notification.\n"
|
||||
|
||||
printf -v MessageBody "🐋 Containers on $FromHost with updates available:\n\n$UpdToString"
|
||||
|
||||
$MailPkg $SendMailTo << __EOF
|
||||
From: "$FromHost" <$SendMailFrom>
|
||||
date:$(date -R)
|
||||
To: <$SendMailTo>
|
||||
Subject: [$SubjectTag] Updates available on $FromHost
|
||||
Content-Type: text/plain; charset=UTF-8; format=flowed
|
||||
Content-Transfer-Encoding: 7bit
|
||||
|
||||
$MessageBody
|
||||
|
||||
__EOF
|
||||
}
|
||||
27
notify_templates/notify_telegram.sh
Normal file
@@ -0,0 +1,27 @@
|
||||
### DISCLAIMER: This is a third party addition to dockcheck - best effort testing.
|
||||
#
|
||||
# Copy/rename this file to notify.sh to enable the notification snippet.
|
||||
# Required receiving services must already be set up.
|
||||
# Modify to fit your setup - set TelegramChatId and TelegramToken.
|
||||
|
||||
send_notification() {
|
||||
[ -s "$ScriptWorkDir"/urls.list ] && releasenotes || Updates=("$@")
|
||||
UpdToString=$( printf '%s\\n' "${Updates[@]}" )
|
||||
FromHost=$(hostname)
|
||||
|
||||
# platform specific notification code would go here
|
||||
printf "\nSending Telegram notification\n"
|
||||
|
||||
# Setting the MessageBody variable here.
|
||||
MessageBody="🐋 Containers on $FromHost with updates available: \n$UpdToString"
|
||||
|
||||
# Modify to fit your setup:
|
||||
TelegramToken="Your Telegram token here"
|
||||
TelegramChatId="Your Telegram ChatId here"
|
||||
TelegramUrl="https://api.telegram.org/bot$TelegramToken"
|
||||
TelegramTopicID=12345678 ## Set to 0 if not using specific topic within chat
|
||||
TelegramData="{\"chat_id\":\"$TelegramChatId\",\"text\":\"$MessageBody\",\"message_thread_id\":\"$TelegramTopicID\",\"disable_notification\": false}"
|
||||
|
||||
curl -sS -o /dev/null --fail -X POST "$TelegramUrl/sendMessage" -H 'Content-Type: application/json' -d "$TelegramData"
|
||||
|
||||
}
|
||||
35
notify_templates/urls.list
Normal file
@@ -0,0 +1,35 @@
|
||||
# This is a list of container names and releasenote urls, separated by space.
|
||||
# Modify, add and (if necessary) remove to fit your needs.
|
||||
# Additions are welcome! Append your list to the git-repo, use generic names and sensible urls.
|
||||
|
||||
apprise-api https://github.com/linuxserver/docker-apprise-api/releases
|
||||
homer https://github.com/bastienwirtz/homer/releases
|
||||
nginx https://github.com/docker-library/official-images/blob/master/library/nginx
|
||||
vaultwarden-server https://github.com/dani-garcia/vaultwarden/releases
|
||||
bruceforce-vaultwarden-backup https://github.com/Bruceforce/vaultwarden-backup/blob/main/CHANGELOG.md
|
||||
actual_server https://actualbudget.org/blog
|
||||
gotify https://github.com/gotify/server/releases
|
||||
traefik https://github.com/traefik/traefik/releases
|
||||
caddy https://github.com/caddyserver/caddy/releases
|
||||
|
||||
calibre https://github.com/linuxserver/docker-calibre/releases
|
||||
calibre-web https://github.com/linuxserver/docker-calibre-web/releases
|
||||
readarr https://github.com/Readarr/Readarr/releases
|
||||
audiobookshelf https://github.com/advplyr/audiobookshelf/releases
|
||||
|
||||
gluetun https://github.com/qdm12/gluetun/releases
|
||||
bazarr https://github.com/morpheus65535/bazarr/releases
|
||||
bazarr-ls https://github.com/linuxserver/docker-bazarr/releases
|
||||
prowlarr https://github.com/Prowlarr/Prowlarr/releases
|
||||
prowlarr-ls https://github.com/linuxserver/docker-prowlarr/releases
|
||||
recyclarr https://github.com/recyclarr/recyclarr/releases
|
||||
sabnzbd https://github.com/linuxserver/docker-sabnzbd/releases
|
||||
sonarr https://github.com/linuxserver/docker-sonarr/releases
|
||||
radarr https://github.com/linuxserver/docker-radarr/releases
|
||||
lidarr https://github.com/linuxserver/docker-lidarr/releases
|
||||
jellyseerr https://github.com/Fallenbagel/jellyseerr/releases
|
||||
jellyfin https://github.com/jellyfin/jellyfin/releases
|
||||
|
||||
home-assistant https://github.com/home-assistant/docker/releases
|
||||
zigbee2mqtt https://github.com/Koenkk/zigbee2mqtt/releases
|
||||
mqtt https://github.com/eclipse/mosquitto/tags
|
||||