mirror of
https://github.com/mag37/dockcheck.git
synced 2026-04-17 09:57:49 +00:00
Compare commits
43 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
bc6a3529c7 | ||
|
|
01b9b33751 | ||
|
|
9ed2a0bad0 | ||
|
|
43307350ca | ||
|
|
b3600f26ac | ||
|
|
79def47754 | ||
|
|
e6ff634394 | ||
|
|
6444e18a4f | ||
|
|
14872b0471 | ||
|
|
07ad241e9f | ||
|
|
58d53d3aaf | ||
|
|
15ce226a0a | ||
|
|
cce5438aca | ||
|
|
be6f5edf52 | ||
|
|
99befd6938 | ||
|
|
8a63fd360c | ||
|
|
d3786d6f75 | ||
|
|
487cfb2822 | ||
|
|
b4943df46c | ||
|
|
8603c8d4b6 | ||
|
|
ae66a6f0fd | ||
|
|
3ac0521a9b | ||
|
|
00ae250511 | ||
|
|
704387a7fe | ||
|
|
1a6826e2ac | ||
|
|
a28b9e555f | ||
|
|
8309b80dc2 | ||
|
|
bbe26a0ac2 | ||
|
|
d98d052af7 | ||
|
|
fbba77dc1f | ||
|
|
09c9faa008 | ||
|
|
17ed46d157 | ||
|
|
d744d51473 | ||
|
|
3821c22660 | ||
|
|
5dd42b7336 | ||
|
|
8da2cbe611 | ||
|
|
9937d6bc3f | ||
|
|
5517abb090 | ||
|
|
cb8c23e62d | ||
|
|
0009adc2cf | ||
|
|
8d39e1b66b | ||
|
|
ec068dd19a | ||
|
|
119932aa19 |
46
README.md
46
README.md
@@ -10,25 +10,26 @@
|
||||
<a href="https://github.com/sponsors/mag37"><img src="https://img.shields.io/badge/-Sponsor-grey?style=flat-square&logo=github" alt="Github Sponsor"></a>
|
||||
</p>
|
||||
|
||||
<h3 align="center">CLI tool to automate docker image updates. <br>No <b>pre-pull</b>, selective, optional notifications and prune when done.</h3>
|
||||
<h2 align="center">Now with simple notification integrations!</h2>
|
||||
<h4 align="center">With features like excluding specific containers, custom container labels, auto-prune when done and more.</h4>
|
||||
<h2 align="center">CLI tool to automate docker image updates or notifying when updates are available.</h2>
|
||||
<h3 align="center">Features:</h3>
|
||||
<h3 align="center">selective updates, exclude containers, custom labels, notification plugins, prune when done and more.</h3>
|
||||
|
||||
<h4 align="center">For Podman - see the fork <a href="https://github.com/sudo-kraken/podcheck">sudo-kraken/podcheck</a>!</h4>
|
||||
|
||||
___
|
||||
## :bell: Changelog
|
||||
|
||||
- **v0.5.4.0**: Added support for a Prometheus+node_exporter metric collection through a file collector.
|
||||
- **v0.5.3.0**: Local image check changed (use imageId instead of name) and Gotify-template fixed (whale icon removed).
|
||||
- **v0.5.2.1**: Rewrite of dependency downloads, jq can be installed with package manager or static binary.
|
||||
- **v0.5.1**: DEPENDENCY WARNING: now requires **jq**. + Upstreaming changes from [sudo-kraken/podcheck](https://github.com/sudo-kraken/podcheck)
|
||||
- **v0.5.0**: Rewritten notify logic - all templates are adjusted and should be migrated!
|
||||
- Copy the custom settings from your current template to the new version of the same template.
|
||||
- Look into, copy and customize the `urls.list` file if that's of interest.
|
||||
- Other changes:
|
||||
- Other changes:
|
||||
- Added Discord notify template.
|
||||
- Verbosity changed of `regctl`.
|
||||
- **v0.4.9**: Added a function to enrich the notify-message with release note URLs. See [Release notes addon](https://github.com/mag37/dockcheck#date-release-notes-addon-to-notifications)
|
||||
- **v0.4.8**: Rewrote prune logic to not prompt with options `-a|-y` or `-n`. Auto prune with `-p`.
|
||||
- **v0.4.7**: Notification Template changes to gotify(new!), DSM(improved), SMTP(deprecation alternative).
|
||||
- **v0.4.6**: Compatibility changes to timeout, due to busybox.
|
||||
- **v0.4.5**: Bugfixes, compatibility changes to timeout and arrays.
|
||||
- **v0.4.3**: Added timeout option to skip container if registry check takes too long (10s default).
|
||||
___
|
||||
|
||||
|
||||
@@ -42,6 +43,7 @@ Example: dockcheck.sh -y -d 10 -e nextcloud,heimdall
|
||||
|
||||
Options:"
|
||||
-a|y Automatic updates, without interaction.
|
||||
-c D Exports metrics as prom file for the prometheus node_exporter. Provide the collector textfile directory.
|
||||
-d N Only update to new images that are N+ days old. Lists too recent with +prefix and age. 2xSlower.
|
||||
-e X Exclude containers, separated by comma.
|
||||
-f Force stack restart after update. Caution: restarts once for every updated container within stack.
|
||||
@@ -57,7 +59,6 @@ Options:"
|
||||
-v Prints current version.
|
||||
```
|
||||
|
||||
|
||||
### Basic example:
|
||||
```
|
||||
$ ./dockcheck.sh
|
||||
@@ -74,14 +75,16 @@ Containers with updates available:
|
||||
Choose what containers to update:
|
||||
Enter number(s) separated by comma, [a] for all - [q] to quit:
|
||||
```
|
||||
Then it proceedes to run `pull` and `up -d` on every container with updates.
|
||||
Then it proceeds to run `pull` and `up -d` on every container with updates.
|
||||
After the updates are complete, you'll get prompted if you'd like to prune dangling images.
|
||||
|
||||
___
|
||||
|
||||
## :nut_and_bolt: Dependencies
|
||||
- Running docker (duh) and compose, either standalone or plugin.
|
||||
- Running docker (duh) and compose, either standalone or plugin. (see [Podman fork](https://github.com/sudo-kraken/podcheck)
|
||||
- Bash shell or compatible shell of at least v4.3
|
||||
- [jq](https://github.com/jqlang/jq)
|
||||
- User will be prompted to install with package manager or download static binary.
|
||||
- [regclient/regctl](https://github.com/regclient/regclient) (Licensed under [Apache-2.0 License](http://www.apache.org/licenses/LICENSE-2.0))
|
||||
- User will be prompted to download `regctl` if not in `PATH` or `PWD`.
|
||||
- regctl requires `amd64/arm64` - see [workaround](#roller_coaster-workaround-for-non-amd64--arm64) if other architecture is used.
|
||||
@@ -117,13 +120,14 @@ Use a `notify_X.sh` template file from the **notify_templates** directory, copy
|
||||
- [Telegram](https://telegram.org/) - Telegram chat API.
|
||||
- [Matrix-Synapse](https://github.com/element-hq/synapse) - [Matrix](https://matrix.org/), open, secure, decentralised communication.
|
||||
- [Pushover](https://pushover.net/) - Simple Notifications (to your phone, wearables, desktops)
|
||||
- [Discord](https://support.discord.com/hc/en-us/articles/228383668-Intro-to-Webhooks) - Discord webhooks.
|
||||
|
||||
Further additions are welcome - suggestions or PR!
|
||||
<sub><sup>Initiated and first contributed by [yoyoma2](https://github.com/yoyoma2).</sup></sub>
|
||||
|
||||
### :date: Release notes addon to Notifications
|
||||
There's a function to use a lookup-file to add release note URL's to the notification message.
|
||||
Copy the notify_templates/`urls.list` file to the script directory, it will be used automatically if it's there. Modify it as necessary, the names of interest in the left column needs to match your container names.
|
||||
There's a function to use a lookup-file to add release note URL's to the notification message.
|
||||
Copy the notify_templates/`urls.list` file to the script directory, it will be used automatically if it's there. Modify it as necessary, the names of interest in the left column needs to match your container names.
|
||||
The output of the notification will look something like this:
|
||||
```
|
||||
Containers on hostname with updates available:
|
||||
@@ -134,6 +138,17 @@ nginx -> https://github.com/docker-library/official-images/blob/master/library
|
||||
```
|
||||
The `urls.list` file is just an example and I'd gladly see that people contribute back when they add their preferred URLs to their lists.
|
||||
|
||||
## :chart_with_upwards_trend: Prometheus and node_exporter
|
||||
Dockcheck can be used together with [Prometheus](https://github.com/prometheus/prometheus) and [node_exporter](https://github.com/prometheus/node_exporter) to export metrics via the file collector, scheduled with cron or likely.
|
||||
This is done with the `-c` option, like this:
|
||||
```
|
||||
dockcheck.sh -c /path/to/exporter/directory
|
||||
```
|
||||
|
||||
See the [README.md](./addons/prometheus/README.md) for more detailed information on how to set it up!
|
||||
|
||||
<sub><sup>Contributed by [tdralle](https://github.com/tdralle).</sup></sub>
|
||||
|
||||
## :bookmark: Labels
|
||||
Optionally add labels to compose-files. Currently these are the usable labels:
|
||||
```
|
||||
@@ -185,7 +200,7 @@ function dchk {
|
||||
Containers need to be manually stopped, removed and created again to run on the new image.
|
||||
|
||||
## :wrench: Debugging
|
||||
If you hit issues, you could check the output of the `extras/errorCheck.sh` script for clues.
|
||||
If you hit issues, you could check the output of the `extras/errorCheck.sh` script for clues.
|
||||
Another option is to run the main script with debugging in a subshell `bash -x dockcheck.sh` - if there's a particular container/image that's causing issues you can filter for just that through `bash -x dockcheck.sh nginx`.
|
||||
|
||||
## :scroll: License
|
||||
@@ -198,4 +213,3 @@ dockcheck is created and released under the [GNU GPL v3.0](https://www.gnu.org/l
|
||||
___
|
||||
|
||||
### :floppy_disk: The [story](https://mag37.org/posts/project_dockcheck/) behind it. 1 year in retrospect.
|
||||
|
||||
|
||||
61
addons/prometheus/README.md
Normal file
61
addons/prometheus/README.md
Normal file
@@ -0,0 +1,61 @@
|
||||
## [Prometheus](https://github.com/prometheus/prometheus) and [node_exporter](https://github.com/prometheus/node_exporter)
|
||||
Dockcheck is capable to export metrics to prometheus via the text file collector provided by the node_exporter.
|
||||
In order to do so the -c flag has to be specified followed by the file path that is configured in the text file collector of the node_exporter.
|
||||
A simple cron job can be configured to export these metrics on a regular interval as shown in the sample below:
|
||||
|
||||
```
|
||||
0 1 * * * /root/dockcheck.sh -n -c /var/lib/node_exporter/textfile_collector
|
||||
```
|
||||
|
||||
The following metrics are exported to prometheus
|
||||
|
||||
```
|
||||
# HELP dockcheck_images_analyzed Docker images that have been analyzed
|
||||
# TYPE dockcheck_images_analyzed gauge
|
||||
dockcheck_images_analyzed 22
|
||||
# HELP dockcheck_images_outdated Docker images that are outdated
|
||||
# TYPE dockcheck_images_outdated gauge
|
||||
dockcheck_images_outdated 7
|
||||
# HELP dockcheck_images_latest Docker images that are outdated
|
||||
# TYPE dockcheck_images_latest gauge
|
||||
dockcheck_images_latest 14
|
||||
# HELP dockcheck_images_error Docker images with analysis errors
|
||||
# TYPE dockcheck_images_error gauge
|
||||
dockcheck_images_error 1
|
||||
# HELP dockcheck_images_analyze_timestamp_seconds Last dockercheck run time
|
||||
# TYPE dockcheck_images_analyze_timestamp_seconds gauge
|
||||
dockcheck_images_analyze_timestamp_seconds 1737924029
|
||||
```
|
||||
|
||||
Once those metrics are exported they can be used to define alarms as shown below
|
||||
|
||||
```
|
||||
- alert: dockcheck_images_outdated
|
||||
expr: sum by(instance) (dockcheck_images_outdated) > 0
|
||||
for: 15s
|
||||
labels:
|
||||
severity: warning
|
||||
annotations:
|
||||
summary: "{{ $labels.instance }} has {{ $value }} outdated docker images."
|
||||
description: "{{ $labels.instance }} has {{ $value }} outdated docker images."
|
||||
- alert: dockcheck_images_error
|
||||
expr: sum by(instance) (dockcheck_images_error) > 0
|
||||
for: 15s
|
||||
labels:
|
||||
severity: warning
|
||||
annotations:
|
||||
summary: "{{ $labels.instance }} has {{ $value }} docker images having an error."
|
||||
description: "{{ $labels.instance }} has {{ $value }} docker images having an error."
|
||||
- alert: dockercheck_image_last_analyze
|
||||
expr: (time() - dockcheck_images_analyze_timestamp_seconds) > (3600 * 24 * 3)
|
||||
for: 15s
|
||||
labels:
|
||||
severity: warning
|
||||
annotations:
|
||||
summary: "{{ $labels.instance }} has not updated the dockcheck statistics for more than 3 days."
|
||||
description: "{{ $labels.instance }} has not updated the dockcheck statistics for more than 3 days."
|
||||
```
|
||||
|
||||
There is a reference Grafana dashboard in [grafana/grafana_dashboard.json](./grafana/grafana_dashboard.json).
|
||||
|
||||

|
||||
382
addons/prometheus/grafana/grafana_dashboard.json
Normal file
382
addons/prometheus/grafana/grafana_dashboard.json
Normal file
@@ -0,0 +1,382 @@
|
||||
{
|
||||
"__inputs": [
|
||||
{
|
||||
"name": "DS_PROMETHEUS",
|
||||
"label": "prometheus",
|
||||
"description": "",
|
||||
"type": "datasource",
|
||||
"pluginId": "prometheus",
|
||||
"pluginName": "Prometheus"
|
||||
}
|
||||
],
|
||||
"__elements": {},
|
||||
"__requires": [
|
||||
{
|
||||
"type": "grafana",
|
||||
"id": "grafana",
|
||||
"name": "Grafana",
|
||||
"version": "11.4.0"
|
||||
},
|
||||
{
|
||||
"type": "datasource",
|
||||
"id": "prometheus",
|
||||
"name": "Prometheus",
|
||||
"version": "1.0.0"
|
||||
},
|
||||
{
|
||||
"type": "panel",
|
||||
"id": "table",
|
||||
"name": "Table",
|
||||
"version": ""
|
||||
}
|
||||
],
|
||||
"annotations": {
|
||||
"list": [
|
||||
{
|
||||
"builtIn": 1,
|
||||
"datasource": {
|
||||
"type": "grafana",
|
||||
"uid": "-- Grafana --"
|
||||
},
|
||||
"enable": true,
|
||||
"hide": true,
|
||||
"iconColor": "rgba(0, 211, 255, 1)",
|
||||
"name": "Annotations & Alerts",
|
||||
"type": "dashboard"
|
||||
}
|
||||
]
|
||||
},
|
||||
"editable": true,
|
||||
"fiscalYearStartMonth": 0,
|
||||
"graphTooltip": 0,
|
||||
"id": null,
|
||||
"links": [],
|
||||
"panels": [
|
||||
{
|
||||
"datasource": {
|
||||
"type": "prometheus",
|
||||
"uid": "${DS_PROMETHEUS}"
|
||||
},
|
||||
"fieldConfig": {
|
||||
"defaults": {
|
||||
"color": {
|
||||
"mode": "thresholds"
|
||||
},
|
||||
"custom": {
|
||||
"align": "auto",
|
||||
"cellOptions": {
|
||||
"type": "auto"
|
||||
},
|
||||
"inspect": false
|
||||
},
|
||||
"mappings": [],
|
||||
"thresholds": {
|
||||
"mode": "absolute",
|
||||
"steps": [
|
||||
{
|
||||
"color": "green",
|
||||
"value": null
|
||||
},
|
||||
{
|
||||
"color": "red",
|
||||
"value": 80
|
||||
}
|
||||
]
|
||||
}
|
||||
},
|
||||
"overrides": [
|
||||
{
|
||||
"matcher": {
|
||||
"id": "byName",
|
||||
"options": "last_analyze_timestamp"
|
||||
},
|
||||
"properties": [
|
||||
{
|
||||
"id": "unit",
|
||||
"value": "dateTimeAsIso"
|
||||
}
|
||||
]
|
||||
},
|
||||
{
|
||||
"matcher": {
|
||||
"id": "byName",
|
||||
"options": "last_analyze_since"
|
||||
},
|
||||
"properties": [
|
||||
{
|
||||
"id": "unit",
|
||||
"value": "s"
|
||||
},
|
||||
{
|
||||
"id": "custom.cellOptions",
|
||||
"value": {
|
||||
"mode": "gradient",
|
||||
"type": "color-background"
|
||||
}
|
||||
},
|
||||
{
|
||||
"id": "thresholds",
|
||||
"value": {
|
||||
"mode": "absolute",
|
||||
"steps": [
|
||||
{
|
||||
"color": "green",
|
||||
"value": null
|
||||
},
|
||||
{
|
||||
"color": "red",
|
||||
"value": 259200
|
||||
}
|
||||
]
|
||||
}
|
||||
}
|
||||
]
|
||||
},
|
||||
{
|
||||
"matcher": {
|
||||
"id": "byName",
|
||||
"options": "images_outdated"
|
||||
},
|
||||
"properties": [
|
||||
{
|
||||
"id": "custom.cellOptions",
|
||||
"value": {
|
||||
"mode": "gradient",
|
||||
"type": "color-background"
|
||||
}
|
||||
},
|
||||
{
|
||||
"id": "thresholds",
|
||||
"value": {
|
||||
"mode": "absolute",
|
||||
"steps": [
|
||||
{
|
||||
"color": "green",
|
||||
"value": null
|
||||
},
|
||||
{
|
||||
"color": "red",
|
||||
"value": 1
|
||||
}
|
||||
]
|
||||
}
|
||||
}
|
||||
]
|
||||
},
|
||||
{
|
||||
"matcher": {
|
||||
"id": "byName",
|
||||
"options": "images_error"
|
||||
},
|
||||
"properties": [
|
||||
{
|
||||
"id": "custom.cellOptions",
|
||||
"value": {
|
||||
"mode": "gradient",
|
||||
"type": "color-background"
|
||||
}
|
||||
},
|
||||
{
|
||||
"id": "thresholds",
|
||||
"value": {
|
||||
"mode": "absolute",
|
||||
"steps": [
|
||||
{
|
||||
"color": "green",
|
||||
"value": null
|
||||
},
|
||||
{
|
||||
"color": "red",
|
||||
"value": 1
|
||||
}
|
||||
]
|
||||
}
|
||||
}
|
||||
]
|
||||
}
|
||||
]
|
||||
},
|
||||
"gridPos": {
|
||||
"h": 14,
|
||||
"w": 24,
|
||||
"x": 0,
|
||||
"y": 0
|
||||
},
|
||||
"id": 2,
|
||||
"options": {
|
||||
"cellHeight": "sm",
|
||||
"footer": {
|
||||
"countRows": false,
|
||||
"fields": "",
|
||||
"reducer": [
|
||||
"sum"
|
||||
],
|
||||
"show": false
|
||||
},
|
||||
"frameIndex": 1,
|
||||
"showHeader": true,
|
||||
"sortBy": []
|
||||
},
|
||||
"pluginVersion": "11.4.0",
|
||||
"targets": [
|
||||
{
|
||||
"disableTextWrap": false,
|
||||
"editorMode": "code",
|
||||
"exemplar": false,
|
||||
"expr": "sum by(instance) (dockcheck_images_analyzed)",
|
||||
"format": "table",
|
||||
"fullMetaSearch": false,
|
||||
"hide": false,
|
||||
"includeNullMetadata": true,
|
||||
"instant": true,
|
||||
"interval": "",
|
||||
"legendFormat": "{{instance}}",
|
||||
"range": false,
|
||||
"refId": "dockcheck_images_analyzed",
|
||||
"useBackend": false,
|
||||
"datasource": {
|
||||
"type": "prometheus",
|
||||
"uid": "${DS_PROMETHEUS}"
|
||||
}
|
||||
},
|
||||
{
|
||||
"datasource": {
|
||||
"type": "prometheus",
|
||||
"uid": "${DS_PROMETHEUS}"
|
||||
},
|
||||
"disableTextWrap": false,
|
||||
"editorMode": "code",
|
||||
"exemplar": false,
|
||||
"expr": "sum by(instance) (dockcheck_images_outdated)",
|
||||
"format": "table",
|
||||
"fullMetaSearch": false,
|
||||
"hide": false,
|
||||
"includeNullMetadata": true,
|
||||
"instant": true,
|
||||
"legendFormat": "{{instance}}",
|
||||
"range": false,
|
||||
"refId": "dockcheck_images_outdated",
|
||||
"useBackend": false
|
||||
},
|
||||
{
|
||||
"datasource": {
|
||||
"type": "prometheus",
|
||||
"uid": "${DS_PROMETHEUS}"
|
||||
},
|
||||
"disableTextWrap": false,
|
||||
"editorMode": "code",
|
||||
"exemplar": false,
|
||||
"expr": "sum by(instance) (dockcheck_images_latest)",
|
||||
"format": "table",
|
||||
"fullMetaSearch": false,
|
||||
"hide": false,
|
||||
"includeNullMetadata": true,
|
||||
"instant": true,
|
||||
"legendFormat": "{{instance}}",
|
||||
"range": false,
|
||||
"refId": "dockcheck_images_latest",
|
||||
"useBackend": false
|
||||
},
|
||||
{
|
||||
"datasource": {
|
||||
"type": "prometheus",
|
||||
"uid": "${DS_PROMETHEUS}"
|
||||
},
|
||||
"editorMode": "code",
|
||||
"exemplar": false,
|
||||
"expr": "sum by(instance) (dockcheck_images_error)",
|
||||
"format": "table",
|
||||
"hide": false,
|
||||
"instant": true,
|
||||
"legendFormat": "{{instance}}",
|
||||
"range": false,
|
||||
"refId": "dockcheck_images_error"
|
||||
},
|
||||
{
|
||||
"datasource": {
|
||||
"type": "prometheus",
|
||||
"uid": "${DS_PROMETHEUS}"
|
||||
},
|
||||
"editorMode": "code",
|
||||
"exemplar": false,
|
||||
"expr": "dockcheck_images_analyze_timestamp_seconds * 1000",
|
||||
"format": "table",
|
||||
"hide": false,
|
||||
"instant": true,
|
||||
"legendFormat": "{{instance}}",
|
||||
"range": false,
|
||||
"refId": "dockcheck_images_analyze_timestamp_seconds"
|
||||
},
|
||||
{
|
||||
"datasource": {
|
||||
"type": "prometheus",
|
||||
"uid": "${DS_PROMETHEUS}"
|
||||
},
|
||||
"editorMode": "code",
|
||||
"exemplar": false,
|
||||
"expr": "time() - dockcheck_images_analyze_timestamp_seconds",
|
||||
"format": "table",
|
||||
"hide": false,
|
||||
"instant": true,
|
||||
"legendFormat": "{{instance}}",
|
||||
"range": false,
|
||||
"refId": "dockcheck_images_last_analyze"
|
||||
}
|
||||
],
|
||||
"title": "Dockcheck Status",
|
||||
"transformations": [
|
||||
{
|
||||
"id": "merge",
|
||||
"options": {}
|
||||
},
|
||||
{
|
||||
"id": "organize",
|
||||
"options": {
|
||||
"excludeByName": {
|
||||
"Time": true,
|
||||
"__name__": true,
|
||||
"job": true
|
||||
},
|
||||
"includeByName": {},
|
||||
"indexByName": {
|
||||
"Time": 0,
|
||||
"Value #dockcheck_images_analyze_timestamp_seconds": 2,
|
||||
"Value #dockcheck_images_analyzed": 4,
|
||||
"Value #dockcheck_images_error": 7,
|
||||
"Value #dockcheck_images_last_analyze": 3,
|
||||
"Value #dockcheck_images_latest": 5,
|
||||
"Value #dockcheck_images_outdated": 6,
|
||||
"instance": 1,
|
||||
"job": 8
|
||||
},
|
||||
"renameByName": {
|
||||
"Value #A": "analyze_timestamp",
|
||||
"Value #dockcheck_images_analyze_timestamp_seconds": "last_analyze_timestamp",
|
||||
"Value #dockcheck_images_analyzed": "images_analyzed",
|
||||
"Value #dockcheck_images_error": "images_error",
|
||||
"Value #dockcheck_images_last_analyze": "last_analyze_since",
|
||||
"Value #dockcheck_images_latest": "images_latest",
|
||||
"Value #dockcheck_images_outdated": "images_outdated"
|
||||
}
|
||||
}
|
||||
}
|
||||
],
|
||||
"type": "table"
|
||||
}
|
||||
],
|
||||
"schemaVersion": 40,
|
||||
"tags": [],
|
||||
"templating": {
|
||||
"list": []
|
||||
},
|
||||
"time": {
|
||||
"from": "now-6h",
|
||||
"to": "now"
|
||||
},
|
||||
"timepicker": {},
|
||||
"timezone": "browser",
|
||||
"title": "Dockcheck Status",
|
||||
"uid": "feb4pv3kv1hxca",
|
||||
"version": 17,
|
||||
"weekStart": ""
|
||||
}
|
||||
BIN
addons/prometheus/grafana/grafana_dashboard.png
Normal file
BIN
addons/prometheus/grafana/grafana_dashboard.png
Normal file
Binary file not shown.
|
After Width: | Height: | Size: 50 KiB |
28
addons/prometheus/prometheus_collector.sh
Normal file
28
addons/prometheus/prometheus_collector.sh
Normal file
@@ -0,0 +1,28 @@
|
||||
prometheus_exporter() {
|
||||
checkedImages=$(($1 + $2 + $3))
|
||||
checkTimestamp=$(date +%s)
|
||||
|
||||
promFileContent=()
|
||||
promFileContent+=("# HELP dockcheck_images_analyzed Docker images that have been analyzed")
|
||||
promFileContent+=("# TYPE dockcheck_images_analyzed gauge")
|
||||
promFileContent+=("dockcheck_images_analyzed $checkedImages")
|
||||
|
||||
promFileContent+=("# HELP dockcheck_images_outdated Docker images that are outdated")
|
||||
promFileContent+=("# TYPE dockcheck_images_outdated gauge")
|
||||
promFileContent+=("dockcheck_images_outdated ${#GotUpdates[@]}")
|
||||
|
||||
promFileContent+=("# HELP dockcheck_images_latest Docker images that are outdated")
|
||||
promFileContent+=("# TYPE dockcheck_images_latest gauge")
|
||||
promFileContent+=("dockcheck_images_latest ${#NoUpdates[@]}")
|
||||
|
||||
promFileContent+=("# HELP dockcheck_images_error Docker images with analysis errors")
|
||||
promFileContent+=("# TYPE dockcheck_images_error gauge")
|
||||
promFileContent+=("dockcheck_images_error ${#GotErrors[@]}")
|
||||
|
||||
promFileContent+=("# HELP dockcheck_images_analyze_timestamp_seconds Last dockercheck run time")
|
||||
promFileContent+=("# TYPE dockcheck_images_analyze_timestamp_seconds gauge")
|
||||
promFileContent+=("dockcheck_images_analyze_timestamp_seconds $checkTimestamp")
|
||||
|
||||
printf "%s\n" "${promFileContent[@]}" > "$CollectorTextFileDirectory/dockcheck_info.prom\$\$"
|
||||
mv -f "$CollectorTextFileDirectory/dockcheck_info.prom\$\$" "$CollectorTextFileDirectory/dockcheck_info.prom"
|
||||
}
|
||||
204
dockcheck.sh
204
dockcheck.sh
@@ -1,25 +1,26 @@
|
||||
#!/usr/bin/env bash
|
||||
VERSION="v0.5.0"
|
||||
### ChangeNotes: Rewritten notify logic - all templates adjusted, transfer your current settings to a new template! See README.
|
||||
VERSION="v0.5.4.0"
|
||||
### ChangeNotes: Added support for a Prometheus+node_exporter metric collection through a file collector.
|
||||
Github="https://github.com/mag37/dockcheck"
|
||||
RawUrl="https://raw.githubusercontent.com/mag37/dockcheck/main/dockcheck.sh"
|
||||
|
||||
### Variables for self updating
|
||||
# Variables for self updating
|
||||
ScriptArgs=( "$@" )
|
||||
ScriptPath="$(readlink -f "$0")"
|
||||
ScriptWorkDir="$(dirname "$ScriptPath")"
|
||||
|
||||
### Check if there's a new release of the script:
|
||||
# Check if there's a new release of the script
|
||||
LatestRelease="$(curl -s -r 0-50 $RawUrl | sed -n "/VERSION/s/VERSION=//p" | tr -d '"')"
|
||||
LatestChanges="$(curl -s -r 0-200 $RawUrl | sed -n "/ChangeNotes/s/### ChangeNotes: //p")"
|
||||
LatestChanges="$(curl -s -r 0-200 $RawUrl | sed -n "/ChangeNotes/s/# ChangeNotes: //p")"
|
||||
|
||||
### Help Function:
|
||||
# Help Function
|
||||
Help() {
|
||||
echo "Syntax: dockcheck.sh [OPTION] [part of name to filter]"
|
||||
echo "Example: dockcheck.sh -y -d 10 -e nextcloud,heimdall"
|
||||
echo
|
||||
echo "Options:"
|
||||
echo "-a|y Automatic updates, without interaction."
|
||||
echo "-c Exports metrics as prom file for the prometheus node_exporter. Provide the collector textfile directory."
|
||||
echo "-d N Only update to new images that are N+ days old. Lists too recent with +prefix and age. 2xSlower."
|
||||
echo "-e X Exclude containers, separated by comma."
|
||||
echo "-f Force stack restart after update. Caution: restarts once for every updated container within stack."
|
||||
@@ -27,9 +28,9 @@ Help() {
|
||||
echo "-i Inform - send a preconfigured notification."
|
||||
echo "-l Only update if label is set. See readme."
|
||||
echo "-m Monochrome mode, no printf color codes."
|
||||
echo "-n No updates, only checking availability."
|
||||
echo "-p Auto-Prune dangling images after update."
|
||||
echo "-r Allow updating images for docker run, wont update the container."
|
||||
echo "-n No updates; only checking availability."
|
||||
echo "-p Auto-prune dangling images after update."
|
||||
echo "-r Allow updating images for docker run; won't update the container."
|
||||
echo "-s Include stopped containers in the check. (Logic: docker ps -a)."
|
||||
echo "-t Set a timeout (in seconds) per container for registry checkups, 10 is default."
|
||||
echo "-v Prints current version."
|
||||
@@ -37,7 +38,7 @@ Help() {
|
||||
echo "Project source: $Github"
|
||||
}
|
||||
|
||||
### Colors:
|
||||
# Colors
|
||||
c_red="\033[0;31m"
|
||||
c_green="\033[0;32m"
|
||||
c_yellow="\033[0;33m"
|
||||
@@ -47,9 +48,11 @@ c_reset="\033[0m"
|
||||
|
||||
Timeout=10
|
||||
Stopped=""
|
||||
while getopts "aynpfrhlisvme:d:t:" options; do
|
||||
while getopts "aynpfrhlisvmc:e:d:t:" options; do
|
||||
case "${options}" in
|
||||
a|y) AutoUp="yes" ;;
|
||||
c) CollectorTextFileDirectory="${OPTARG}"
|
||||
if ! [[ -d $CollectorTextFileDirectory ]] ; then { printf "The directory (%s) does not exist.\n" "${CollectorTextFileDirectory}" ; exit 2; } fi ;;
|
||||
n) AutoUp="no" ;;
|
||||
r) DRunUp="yes" ;;
|
||||
p) AutoPrune="yes" ;;
|
||||
@@ -68,18 +71,19 @@ while getopts "aynpfrhlisvme:d:t:" options; do
|
||||
done
|
||||
shift "$((OPTIND-1))"
|
||||
|
||||
# Self-update function
|
||||
self_update_curl() {
|
||||
cp "$ScriptPath" "$ScriptPath".bak
|
||||
if [[ $(builtin type -P curl) ]]; then
|
||||
if [[ $(command -v curl) ]]; then
|
||||
curl -L $RawUrl > "$ScriptPath" ; chmod +x "$ScriptPath"
|
||||
printf "\n%s\n" "--- starting over with the updated version ---"
|
||||
exec "$ScriptPath" "${ScriptArgs[@]}" # run the new script with old arguments
|
||||
exit 1 # exit the old instance
|
||||
elif [[ $(builtin type -P wget) ]]; then
|
||||
exit 1 # Exit the old instance
|
||||
elif [[ $(command -v wget) ]]; then
|
||||
wget $RawUrl -O "$ScriptPath" ; chmod +x "$ScriptPath"
|
||||
printf "\n%s\n" "--- starting over with the updated version ---"
|
||||
exec "$ScriptPath" "${ScriptArgs[@]}" # run the new script with old arguments
|
||||
exit 1 # exit the old instance
|
||||
exit 1 # Exit the old instance
|
||||
else
|
||||
printf "curl/wget not available - download the update manually: %s \n" "$Github"
|
||||
fi
|
||||
@@ -87,7 +91,7 @@ self_update_curl() {
|
||||
|
||||
self_update() {
|
||||
cd "$ScriptWorkDir" || { printf "Path error, skipping update.\n" ; return ; }
|
||||
if [[ $(builtin type -P git) ]] && [[ "$(git ls-remote --get-url 2>/dev/null)" =~ .*"mag37/dockcheck".* ]] ; then
|
||||
if [[ $(command -v git) ]] && [[ "$(git ls-remote --get-url 2>/dev/null)" =~ .*"mag37/dockcheck".* ]] ; then
|
||||
printf "\n%s\n" "Pulling the latest version."
|
||||
git pull --force || { printf "Git error, manually pull/clone.\n" ; return ; }
|
||||
printf "\n%s\n" "--- starting over with the updated version ---"
|
||||
@@ -100,7 +104,7 @@ self_update() {
|
||||
fi
|
||||
}
|
||||
|
||||
### Choose from list -function:
|
||||
# Choose from list function
|
||||
choosecontainers() {
|
||||
while [[ -z "$ChoiceClean" ]]; do
|
||||
read -r -p "Enter number(s) separated by comma, [a] for all - [q] to quit: " Choice
|
||||
@@ -112,7 +116,7 @@ choosecontainers() {
|
||||
else
|
||||
ChoiceClean=${Choice//[,.:;]/ }
|
||||
for CC in $ChoiceClean ; do
|
||||
if [[ "$CC" -lt 1 || "$CC" -gt $UpdCount ]] ; then # reset choice if out of bounds
|
||||
if [[ "$CC" -lt 1 || "$CC" -gt $UpdCount ]] ; then # Reset choice if out of bounds
|
||||
echo "Number not in list: $CC" ; unset ChoiceClean ; break 1
|
||||
else
|
||||
SelectedUpdates+=( "${GotUpdates[$CC-1]}" )
|
||||
@@ -139,15 +143,15 @@ progress_bar() {
|
||||
QueCurrent="$1"
|
||||
QueTotal="$2"
|
||||
((Percent=100*QueCurrent/QueTotal))
|
||||
((Complete=50*Percent/100)) # change first number for width (50)
|
||||
((Left=50-Complete)) # change first number for width (50)
|
||||
((Complete=50*Percent/100)) # Change first number for width (50)
|
||||
((Left=50-Complete)) # Change first number for width (50)
|
||||
BarComplete=$(printf "%${Complete}s" | tr " " "#")
|
||||
BarLeft=$(printf "%${Left}s" | tr " " "-")
|
||||
[[ "$QueTotal" == "$QueCurrent" ]] || printf "\r[%s%s] %s/%s " "$BarComplete" "$BarLeft" "$QueCurrent" "$QueTotal"
|
||||
[[ "$QueTotal" == "$QueCurrent" ]] && printf "\r[%b%s%b] %s/%s \n" "$c_teal" "$BarComplete" "$c_reset" "$QueCurrent" "$QueTotal"
|
||||
}
|
||||
|
||||
### Function to add user-provided urls to releasenotes
|
||||
# Function to add user-provided urls to releasenotes
|
||||
releasenotes() {
|
||||
for update in ${GotUpdates[@]}; do
|
||||
found=false
|
||||
@@ -158,7 +162,7 @@ releasenotes() {
|
||||
done
|
||||
}
|
||||
|
||||
### Version check & initiate self update
|
||||
# Version check & initiate self update
|
||||
if [[ "$VERSION" != "$LatestRelease" ]] ; then
|
||||
printf "New version available! %b%s%b ⇒ %b%s%b \n Change Notes: %s \n" "$c_yellow" "$VERSION" "$c_reset" "$c_green" "$LatestRelease" "$c_reset" "$LatestChanges"
|
||||
if [[ -z "$AutoUp" ]] ; then
|
||||
@@ -167,38 +171,75 @@ if [[ "$VERSION" != "$LatestRelease" ]] ; then
|
||||
fi
|
||||
fi
|
||||
|
||||
### Set $1 to a variable for name filtering later.
|
||||
# Set $1 to a variable for name filtering later
|
||||
SearchName="$1"
|
||||
### Create array of excludes
|
||||
# Create array of excludes
|
||||
IFS=',' read -r -a Excludes <<< "$Exclude" ; unset IFS
|
||||
|
||||
### Check if required binary exists in PATH or directory:
|
||||
if [[ $(builtin type -P "regctl") ]]; then regbin="regctl" ;
|
||||
elif [[ -f "$ScriptWorkDir/regctl" ]]; then regbin="$ScriptWorkDir/regctl" ;
|
||||
# Static binary downloader for dependencies
|
||||
binary_downloader() {
|
||||
BinaryName="$1"
|
||||
BinaryUrl="$2"
|
||||
case "$(uname --machine)" in
|
||||
x86_64|amd64) architecture="amd64" ;;
|
||||
arm64|aarch64) architecture="arm64";;
|
||||
*) printf "\n%bArchitecture not supported, exiting.%b\n" "$c_red" "$c_reset" ; exit 1;;
|
||||
esac
|
||||
GetUrl="${BinaryUrl/TEMP/"$architecture"}"
|
||||
if [[ $(command -v curl) ]]; then curl -L $GetUrl > "$ScriptWorkDir/$BinaryName" ;
|
||||
elif [[ $(command -v wget) ]]; then wget $GetUrl -O "$ScriptWorkDir/$BinaryName" ;
|
||||
else printf "%s\n" "curl/wget not available - get $BinaryName manually from the repo link, exiting."; exit 1;
|
||||
fi
|
||||
[[ -f "$ScriptWorkDir/$BinaryName" ]] && chmod +x "$ScriptWorkDir/$BinaryName"
|
||||
}
|
||||
|
||||
distro_checker() {
|
||||
if [[ -f /etc/arch-release ]] ; then PkgInstaller="pacman -S"
|
||||
elif [[ -f /etc/redhat-release ]] ; then PkgInstaller="dnf install"
|
||||
elif [[ -f /etc/SuSE-release ]] ; then PkgInstaller="zypper install"
|
||||
elif [[ -f /etc/debian_version ]] ; then PkgInstaller="apt-get install"
|
||||
else PkgInstaller="ERROR" ; printf "\n%bNo distribution could be determined%b, falling back to static binary.\n" "$c_yellow" "$c_reset"
|
||||
fi
|
||||
}
|
||||
|
||||
# Dependency check for jq in PATH or directory
|
||||
if [[ $(command -v jq) ]]; then jqbin="jq" ;
|
||||
elif [[ -f "$ScriptWorkDir/jq" ]]; then jqbin="$ScriptWorkDir/jq" ;
|
||||
else
|
||||
read -r -p "Required dependency 'regctl' missing, do you want it downloaded? y/[n] " GetDep
|
||||
if [[ "$GetDep" =~ [yY] ]] ; then
|
||||
### Check arch:
|
||||
case "$(uname --machine)" in
|
||||
x86_64|amd64) architecture="amd64" ;;
|
||||
arm64|aarch64) architecture="arm64";;
|
||||
*) echo "Architecture not supported, exiting." ; exit 1;;
|
||||
esac
|
||||
RegUrl="https://github.com/regclient/regclient/releases/latest/download/regctl-linux-$architecture"
|
||||
if [[ $(builtin type -P curl) ]]; then curl -L $RegUrl > "$ScriptWorkDir/regctl" ; chmod +x "$ScriptWorkDir/regctl" ; regbin="$ScriptWorkDir/regctl" ;
|
||||
elif [[ $(builtin type -P wget) ]]; then wget $RegUrl -O "$ScriptWorkDir/regctl" ; chmod +x "$ScriptWorkDir/regctl" ; regbin="$ScriptWorkDir/regctl" ;
|
||||
else
|
||||
printf "%s\n" "curl/wget not available - get regctl manually from the repo link, quitting."
|
||||
printf "%s\n" "Required dependency 'jq' missing, do you want to install it?"
|
||||
read -r -p "y: With packagemanager (sudo). / s: Download static binary. y/s/[n] " GetJq
|
||||
GetJq=${GetJq:-no} # set default to no if nothing is given
|
||||
if [[ "$GetJq" =~ [yYsS] ]] ; then
|
||||
[[ "$GetJq" =~ [yY] ]] && distro_checker
|
||||
if [[ -n "$PkgInstaller" && "$PkgInstaller" != "ERROR" ]] ; then
|
||||
(sudo $PkgInstaller jq) ; PkgExitcode="$?"
|
||||
[[ "$PkgExitcode" == 0 ]] && jqbin="jq" || printf "\n%bPackagemanager install failed%b, falling back to static binary.\n" "$c_yellow" "$c_reset"
|
||||
fi
|
||||
else
|
||||
printf "%s\n" "Dependency missing, quitting."
|
||||
exit 1
|
||||
if [[ "$GetJq" =~ [nN] || "$PkgInstaller" == "ERROR" || "$PkgExitcode" != 0 ]] ; then
|
||||
binary_downloader "jq" "https://github.com/jqlang/jq/releases/latest/download/jq-linux-TEMP"
|
||||
[[ -f "$ScriptWorkDir/jq" ]] && jqbin="$ScriptWorkDir/jq"
|
||||
fi
|
||||
else printf "\n%bDependency missing, exiting.%b\n" "$c_red" "$c_reset" ; exit 1 ;
|
||||
fi
|
||||
fi
|
||||
### final check if binary is correct
|
||||
# Final check if binary is correct
|
||||
$jqbin --version &> /dev/null || { printf "%s\n" "jq is not working - try to remove it and re-download it, exiting."; exit 1; }
|
||||
|
||||
# Dependency check for regctl in PATH or directory
|
||||
if [[ $(command -v regctl) ]]; then regbin="regctl" ;
|
||||
elif [[ -f "$ScriptWorkDir/regctl" ]]; then regbin="$ScriptWorkDir/regctl" ;
|
||||
else
|
||||
read -r -p "Required dependency 'regctl' missing, do you want it downloaded? y/[n] " GetRegctl
|
||||
if [[ "$GetRegctl" =~ [yY] ]] ; then
|
||||
binary_downloader "regctl" "https://github.com/regclient/regclient/releases/latest/download/regctl-linux-TEMP"
|
||||
[[ -f "$ScriptWorkDir/regctl" ]] && regbin="$ScriptWorkDir/regctl"
|
||||
else printf "\n%bDependency missing, exiting.%b\n" "$c_red" "$c_reset" ; exit 1 ;
|
||||
fi
|
||||
fi
|
||||
# Final check if binary is correct
|
||||
$regbin version &> /dev/null || { printf "%s\n" "regctl is not working - try to remove it and re-download it, exiting."; exit 1; }
|
||||
|
||||
### Check docker compose binary:
|
||||
# Check docker compose binary
|
||||
if docker compose version &> /dev/null ; then DockerBin="docker compose" ;
|
||||
elif docker-compose -v &> /dev/null; then DockerBin="docker-compose" ;
|
||||
elif docker -v &> /dev/null; then
|
||||
@@ -209,7 +250,7 @@ else
|
||||
exit 1
|
||||
fi
|
||||
|
||||
### Numbered List -function:
|
||||
# Numbered List function
|
||||
options() {
|
||||
num=1
|
||||
for i in "${GotUpdates[@]}"; do
|
||||
@@ -218,7 +259,7 @@ for i in "${GotUpdates[@]}"; do
|
||||
done
|
||||
}
|
||||
|
||||
### Listing typed exclusions:
|
||||
# Listing typed exclusions
|
||||
if [[ -n ${Excludes[*]} ]] ; then
|
||||
printf "\n%bExcluding these names:%b\n" "$c_blue" "$c_reset"
|
||||
printf "%s\n" "${Excludes[@]}"
|
||||
@@ -226,11 +267,11 @@ if [[ -n ${Excludes[*]} ]] ; then
|
||||
fi
|
||||
|
||||
# Variables for progress_bar function
|
||||
DocCount=$(docker ps $Stopped --filter "name=$SearchName" --format '{{.Names}}' | wc -l)
|
||||
ContCount=$(docker ps $Stopped --filter "name=$SearchName" --format '{{.Names}}' | wc -l)
|
||||
RegCheckQue=0
|
||||
|
||||
### Testing and setting timeout binary
|
||||
t_out=$(type -P "timeout")
|
||||
# Testing and setting timeout binary
|
||||
t_out=$(command -v timeout)
|
||||
if [[ $t_out ]]; then
|
||||
t_out=$(realpath $t_out 2>/dev/null || readlink -f $t_out)
|
||||
if [[ $t_out =~ "busybox" ]]; then
|
||||
@@ -240,15 +281,16 @@ if [[ $t_out ]]; then
|
||||
else t_out=""
|
||||
fi
|
||||
|
||||
### Check the image-hash of every running container VS the registry
|
||||
# Check the image-hash of every running container VS the registry
|
||||
for i in $(docker ps $Stopped --filter "name=$SearchName" --format '{{.Names}}') ; do
|
||||
((RegCheckQue+=1))
|
||||
progress_bar "$RegCheckQue" "$DocCount"
|
||||
### Looping every item over the list of excluded names and skipping:
|
||||
progress_bar "$RegCheckQue" "$ContCount"
|
||||
# Looping every item over the list of excluded names and skipping
|
||||
for e in "${Excludes[@]}" ; do [[ "$i" == "$e" ]] && continue 2 ; done
|
||||
ImageId=$(docker inspect "$i" --format='{{.Image}}')
|
||||
RepoUrl=$(docker inspect "$i" --format='{{.Config.Image}}')
|
||||
LocalHash=$(docker image inspect "$RepoUrl" --format '{{.RepoDigests}}')
|
||||
# Checking for errors while setting the variable:
|
||||
LocalHash=$(docker image inspect "$ImageId" --format '{{.RepoDigests}}')
|
||||
# Checking for errors while setting the variable
|
||||
if RegHash=$(${t_out} $regbin -v error image digest --list "$RepoUrl" 2>&1) ; then
|
||||
if [[ "$LocalHash" = *"$RegHash"* ]] ; then
|
||||
NoUpdates+=("$i")
|
||||
@@ -260,27 +302,32 @@ for i in $(docker ps $Stopped --filter "name=$SearchName" --format '{{.Names}}')
|
||||
fi
|
||||
fi
|
||||
else
|
||||
# Here the RegHash is the result of an error code.
|
||||
# Here the RegHash is the result of an error code
|
||||
GotErrors+=("$i - ${RegHash}")
|
||||
fi
|
||||
done
|
||||
|
||||
### Sort arrays alphabetically
|
||||
# Sort arrays alphabetically
|
||||
IFS=$'\n'
|
||||
NoUpdates=($(sort <<<"${NoUpdates[*]}"))
|
||||
GotUpdates=($(sort <<<"${GotUpdates[*]}"))
|
||||
unset IFS
|
||||
|
||||
### Define how many updates are available
|
||||
# Run the prometheus exporter function
|
||||
if [ -n "$CollectorTextFileDirectory" ] ; then
|
||||
source "$ScriptWorkDir"/addons/prometheus/prometheus_collector.sh && prometheus_exporter ${#NoUpdates[@]} ${#GotUpdates[@]} ${#GotError[@]}
|
||||
fi
|
||||
|
||||
# Define how many updates are available
|
||||
UpdCount="${#GotUpdates[@]}"
|
||||
|
||||
### List what containers got updates or not
|
||||
# List what containers got updates or not
|
||||
if [[ -n ${NoUpdates[*]} ]] ; then
|
||||
printf "\n%bContainers on latest version:%b\n" "$c_green" "$c_reset"
|
||||
printf "%s\n" "${NoUpdates[@]}"
|
||||
fi
|
||||
if [[ -n ${GotErrors[*]} ]] ; then
|
||||
printf "\n%bContainers with errors, wont get updated:%b\n" "$c_red" "$c_reset"
|
||||
printf "\n%bContainers with errors, won't get updated:%b\n" "$c_red" "$c_reset"
|
||||
printf "%s\n" "${GotErrors[@]}"
|
||||
printf "%binfo:%b 'unauthorized' often means not found in a public registry.\n" "$c_blue" "$c_reset"
|
||||
fi
|
||||
@@ -290,11 +337,11 @@ if [[ -n ${GotUpdates[*]} ]] ; then
|
||||
[[ -n "$Notify" ]] && { [[ $(type -t send_notification) == function ]] && send_notification "${GotUpdates[@]}" || printf "Could not source notification function.\n" ; }
|
||||
fi
|
||||
|
||||
### Optionally get updates if there's any
|
||||
# Optionally get updates if there's any
|
||||
if [ -n "$GotUpdates" ] ; then
|
||||
if [ -z "$AutoUp" ] ; then
|
||||
printf "\n%bChoose what containers to update.%b\n" "$c_teal" "$c_reset"
|
||||
choosecontainers
|
||||
printf "\n%bChoose what containers to update.%b\n" "$c_teal" "$c_reset"
|
||||
choosecontainers
|
||||
else
|
||||
SelectedUpdates=( "${GotUpdates[@]}" )
|
||||
fi
|
||||
@@ -305,14 +352,23 @@ if [ -n "$GotUpdates" ] ; then
|
||||
do
|
||||
((CurrentQue+=1))
|
||||
unset CompleteConfs
|
||||
ContPath=$(docker inspect "$i" --format '{{ index .Config.Labels "com.docker.compose.project.working_dir" }}')
|
||||
ContConfigFile=$(docker inspect "$i" --format '{{ index .Config.Labels "com.docker.compose.project.config_files" }}')
|
||||
ContName=$(docker inspect "$i" --format '{{ index .Config.Labels "com.docker.compose.service" }}')
|
||||
ContEnv=$(docker inspect "$i" --format '{{index .Config.Labels "com.docker.compose.project.environment_file" }}')
|
||||
# Extract labels and metadata
|
||||
ContLabels=$(docker inspect "$i" --format '{{json .Config.Labels}}')
|
||||
ContImage=$(docker inspect "$i" --format='{{.Config.Image}}')
|
||||
ContUpdateLabel=$(docker inspect "$i" --format '{{ index .Config.Labels "mag37.dockcheck.update" }}')
|
||||
ContRestartStack=$(docker inspect "$i" --format '{{ index .Config.Labels "mag37.dockcheck.restart-stack" }}')
|
||||
### Checking if compose-values are empty - hence started with docker run:
|
||||
ContPath=$($jqbin -r '."com.docker.compose.project.working_dir"' <<< "$ContLabels")
|
||||
[ "$ContPath" == "null" ] && ContPath=""
|
||||
ContConfigFile=$($jqbin -r '."com.docker.compose.project.config_files"' <<< "$ContLabels")
|
||||
[ "$ContConfigFile" == "null" ] && ContConfigFile=""
|
||||
ContName=$($jqbin -r '."com.docker.compose.service"' <<< "$ContLabels")
|
||||
[ "$ContName" == "null" ] && ContName=""
|
||||
ContEnv=$($jqbin -r '."com.docker.compose.project.environment_file"' <<< "$ContLabels")
|
||||
[ "$ContEnv" == "null" ] && ContEnv=""
|
||||
ContUpdateLabel=$($jqbin -r '."mag37.dockcheck.update"' <<< "$ContLabels")
|
||||
[ "$ContUpdateLabel" == "null" ] && ContUpdateLabel=""
|
||||
ContRestartStack=$($jqbin -r '."mag37.dockcheck.restart-stack"' <<< "$ContLabels")
|
||||
[ "$ContRestartStack" == "null" ] && ContRestartStack=""
|
||||
|
||||
# Checking if compose-values are empty - hence started with docker run
|
||||
if [ -z "$ContPath" ] ; then
|
||||
if [ "$DRunUp" == "yes" ] ; then
|
||||
docker pull "$ContImage"
|
||||
@@ -322,7 +378,7 @@ if [ -n "$GotUpdates" ] ; then
|
||||
fi
|
||||
continue
|
||||
fi
|
||||
### cd to the compose-file directory to account for people who use relative volumes, eg - ${PWD}/data:data
|
||||
# cd to the compose-file directory to account for people who use relative volumes
|
||||
cd "$ContPath" || { echo "Path error - skipping $i" ; continue ; }
|
||||
## Reformatting path + multi compose
|
||||
if [[ $ContConfigFile = '/'* ]] ; then
|
||||
@@ -331,12 +387,12 @@ if [ -n "$GotUpdates" ] ; then
|
||||
CompleteConfs=$(for conf in ${ContConfigFile//,/ } ; do printf -- "-f %s/%s " "$ContPath" "$conf"; done)
|
||||
fi
|
||||
printf "\n%bNow updating (%s/%s): %b%s%b\n" "$c_teal" "$CurrentQue" "$NumberofUpdates" "$c_blue" "$i" "$c_reset"
|
||||
### Checking if Label Only -option is set, and if container got the label
|
||||
# Checking if Label Only -option is set, and if container got the label
|
||||
[[ "$OnlyLabel" == true ]] && { [[ "$ContUpdateLabel" != true ]] && { echo "No update label, skipping." ; continue ; } }
|
||||
docker pull "$ContImage"
|
||||
### Check if the container got an environment file set and reformat it
|
||||
# Check if the container got an environment file set and reformat it
|
||||
if [ -n "$ContEnv" ]; then ContEnvs=$(for env in ${ContEnv//,/ } ; do printf -- "--env-file %s " "$env"; done) ; fi
|
||||
### Check if the whole stack should be restarted
|
||||
# Check if the whole stack should be restarted
|
||||
if [[ "$ContRestartStack" == true ]] || [[ "$ForceRestartStacks" == true ]] ; then
|
||||
$DockerBin ${CompleteConfs} stop ; $DockerBin ${CompleteConfs} ${ContEnvs} up -d
|
||||
else
|
||||
|
||||
@@ -2,12 +2,21 @@
|
||||
SearchName="$1"
|
||||
for i in $(docker ps --filter "name=$SearchName" --format '{{.Names}}') ; do
|
||||
echo "------------ $i ------------"
|
||||
ContPath=$(docker inspect "$i" --format '{{ index .Config.Labels "com.docker.compose.project.working_dir" }}')
|
||||
[ -z "$ContPath" ] && { "$i has no compose labels - skipping" ; continue ; }
|
||||
ContConfigFile=$(docker inspect "$i" --format '{{ index .Config.Labels "com.docker.compose.project.config_files" }}')
|
||||
ContName=$(docker inspect "$i" --format '{{ index .Config.Labels "com.docker.compose.service" }}')
|
||||
ContEnv=$(docker inspect "$i" --format '{{index .Config.Labels "com.docker.compose.project.environment_file" }}')
|
||||
ContLabels=$(docker inspect "$i" --format '{{json .Config.Labels}}')
|
||||
ContImage=$(docker inspect "$i" --format='{{.Config.Image}}')
|
||||
ContPath=$(jq -r '."com.docker.compose.project.working_dir"' <<< "$ContLabels")
|
||||
[ "$ContPath" == "null" ] && ContPath=""
|
||||
[ -z "$ContPath" ] && { "$i has no compose labels - skipping" ; continue ; }
|
||||
ContConfigFile=$(jq -r '."com.docker.compose.project.config_files"' <<< "$ContLabels")
|
||||
[ "$ContConfigFile" == "null" ] && ContConfigFile=""
|
||||
ContName=$(jq -r '."com.docker.compose.service"' <<< "$ContLabels")
|
||||
[ "$ContName" == "null" ] && ContName=""
|
||||
ContEnv=$(jq -r '."com.docker.compose.project.environment_file"' <<< "$ContLabels")
|
||||
[ "$ContEnv" == "null" ] && ContEnv=""
|
||||
ContUpdateLabel=$(jq -r '."mag37.dockcheck.update"' <<< "$ContLabels")
|
||||
[ "$ContUpdateLabel" == "null" ] && ContUpdateLabel=""
|
||||
ContRestartStack=$(jq -r '."mag37.dockcheck.restart-stack"' <<< "$ContLabels")
|
||||
[ "$ContRestartStack" == "null" ] && ContRestartStack=""
|
||||
|
||||
if [[ $ContConfigFile = '/'* ]] ; then
|
||||
ComposeFile="$ContConfigFile"
|
||||
@@ -20,6 +29,8 @@ for i in $(docker ps --filter "name=$SearchName" --format '{{.Names}}') ; do
|
||||
echo -e "Compose files:\t\t$ComposeFile"
|
||||
echo -e "Environment files:\t$ContEnv"
|
||||
echo -e "Container image:\t$ContImage"
|
||||
echo -e "Update label:\t$ContUpdateLabel"
|
||||
echo -e "Restart Stack label:\t$ContRestartStack"
|
||||
echo
|
||||
echo "Mounts:"
|
||||
docker inspect -f '{{ range .Mounts }}{{ .Source }}:{{ .Destination }}{{ printf "\n" }}{{ end }}' "$i"
|
||||
|
||||
Binary file not shown.
|
Before Width: | Height: | Size: 404 KiB After Width: | Height: | Size: 472 KiB |
BIN
extras/example_old.gif
Normal file
BIN
extras/example_old.gif
Normal file
Binary file not shown.
|
After Width: | Height: | Size: 404 KiB |
@@ -49,4 +49,6 @@ Content-Transfer-Encoding: 7bit
|
||||
$MessageBody
|
||||
From $SenderName
|
||||
__EOF
|
||||
# This ensures DSM's container manager will also see the update
|
||||
/var/packages/ContainerManager/target/tool/image_upgradable_checker
|
||||
}
|
||||
|
||||
@@ -21,5 +21,8 @@ send_notification() {
|
||||
DiscordWebhookUrl="PasteYourFullDiscordWebhookURL"
|
||||
|
||||
MsgBody="{\"username\":\"$FromHost\",\"content\":\"$MessageBody\"}"
|
||||
|
||||
curl -sS -o /dev/null --fail -X POST -H "Content-Type: application/json" -d "$MsgBody" "$DiscordWebhookUrl"
|
||||
|
||||
}
|
||||
|
||||
|
||||
@@ -14,7 +14,7 @@ send_notification() {
|
||||
|
||||
# Setting the MessageTitle and MessageBody variable here.
|
||||
MessageTitle="${FromHost} - updates available."
|
||||
printf -v MessageBody "🐋 Containers on $FromHost with updates available:\n$UpdToString"
|
||||
printf -v MessageBody "Containers on $FromHost with updates available:\n$UpdToString"
|
||||
|
||||
# Modify to fit your setup:
|
||||
GotifyToken="Your Gotify token here"
|
||||
@@ -24,6 +24,6 @@ send_notification() {
|
||||
-F "title=${MessageTitle}" \
|
||||
-F "message=${MessageBody}" \
|
||||
-F "priority=5" \
|
||||
-X POST "${GotifyUrl}" &> /dev/null
|
||||
-X POST "${GotifyUrl}" 1> /dev/null
|
||||
|
||||
}
|
||||
|
||||
@@ -6,6 +6,7 @@ apprise-api https://github.com/linuxserver/docker-apprise-api/releases
|
||||
homer https://github.com/bastienwirtz/homer/releases
|
||||
nginx https://github.com/docker-library/official-images/blob/master/library/nginx
|
||||
vaultwarden-server https://github.com/dani-garcia/vaultwarden/releases
|
||||
bruceforce-vaultwarden-backup https://github.com/Bruceforce/vaultwarden-backup/blob/main/CHANGELOG.md
|
||||
actual_server https://actualbudget.org/blog
|
||||
gotify https://github.com/gotify/server/releases
|
||||
traefik https://github.com/traefik/traefik/releases
|
||||
|
||||
Reference in New Issue
Block a user