Compare commits

...

123 Commits

Author SHA1 Message Date
auricom
a75e75f2b5 Bump to re-trigger repo upload (#338) 2020-12-11 05:25:14 -05:00
ᗪєνιη ᗷυнʟ
731b70fd98 [appdaemon] remove mountPath (#334) 2020-12-10 15:07:20 -05:00
ᗪєνιη ᗷυнʟ
b928e1ea44 [appdaemon] bump version due to release issue (#332) 2020-12-09 19:20:49 -05:00
ᗪєνιη ᗷυнʟ
61dbea3bad [home-assistant] deprecate appdaemon and configurator sidecars (#327) 2020-12-09 16:25:28 -05:00
renovate[bot]
1aa31e3bd9 [ombi] Update helm chart mariadb to v9.1.2 (#324)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-12-09 22:16:56 +01:00
renovate[bot]
b974051a0a [monica] Update helm chart mariadb to v9.1.2 (#323)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-12-09 22:15:35 +01:00
renovate[bot]
da761a5c4c [traccar] Update helm chart mariadb to v9.1.2 (#325)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-12-09 22:14:52 +01:00
ᗪєνιη ᗷυнʟ
a968ee3082 [calibre-web] convert to common library (#329) 2020-12-09 15:55:16 -05:00
ᗪєνιη ᗷυнʟ
ea2b63d930 [appdaemon] create separate chart outside of home-assistant (#326) 2020-12-09 15:54:56 -05:00
auricom
c139284fc3 [navidrome] new chart (#313) 2020-12-08 14:52:36 -05:00
ᗪєνιη ᗷυнʟ
5bb1c5242d [sabnzbd] add documentation on hostname-check (#321) 2020-12-08 13:15:55 -05:00
renovate[bot]
11e76c9f15 [home-assistant] Update helm chart postgresql to v10.1.3 (#317)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-12-08 08:04:29 -05:00
renovate[bot]
66d14d9d7d [statping] Update helm chart postgresql to v10.1.3 (#319)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-12-08 07:57:12 -05:00
renovate[bot]
43fe354c35 [powerdns] Update helm chart postgresql to v10.1.3 (#318)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-12-08 07:53:53 -05:00
renovate[bot]
9604ece2a5 [teslamate] Update helm chart postgresql to v10.1.3 (#320)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-12-08 07:52:13 -05:00
renovate[bot]
f051bb16cf [home-assistant] Update helm chart influxdb to v1.1.2 (#308)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-12-07 14:55:01 -05:00
Mike Wilson
166c4b3cd9 [home-assistant] Tiny typo fix in values.yaml (#315) 2020-12-07 14:16:41 -05:00
Aleksey Sviridkin
351d46296f [home-assistant] Fix env for appdaemon (#305) 2020-12-07 10:54:56 -05:00
auricom
601341ca9b [lychee] new chart (#307) 2020-12-07 10:53:37 -05:00
renovate[bot]
39f3555873 [unifi-poller] Update helm chart influxdb to v1.1.2 (#309)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-12-07 10:29:24 -05:00
dza89
44fbca1fb8 [traefik-forward-auth] Add envFrom option (#304) 2020-12-05 09:50:30 -05:00
ᗪєνιη ᗷυнʟ
6ad3ba062a [esphome] use common (#303)
Co-authored-by: Ryan Holt <ryan@ryanholt.net>
2020-12-04 18:20:33 -05:00
renovate[bot]
ab9a0fda18 [home-assistant] Update helm chart postgresql to v10.1.2 (#297)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
Co-authored-by: ᗪєνιη ᗷυнʟ <onedr0p@users.noreply.github.com>
2020-12-04 17:23:22 -05:00
renovate[bot]
0a3e9eb93c [teslamate] Update helm chart postgresql to v10.1.2 (#300)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-12-04 17:21:48 -05:00
renovate[bot]
7801f2e626 [statping] Update helm chart postgresql to v10.1.2 (#299)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-12-04 17:19:53 -05:00
renovate[bot]
fe64832570 [powerdns] Update helm chart postgresql to v10.1.2 (#298)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-12-04 17:17:55 -05:00
renovate[bot]
0099a56e38 [unifi-poller] Update helm chart influxdb to v1.1.1 (#302)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-12-04 17:16:28 -05:00
Ryan Holt
cfec6a9cc2 [monica] initial add for monica (#296) 2020-12-04 13:13:25 -05:00
Bᴇʀɴᴅ Sᴄʜᴏʀɢᴇʀs
13dedb876b [pyload] Bump to re-trigger repo upload (#295) 2020-12-04 12:01:55 +01:00
Thomas John Wesolowski
d0fba6f031 [booksonic-air][Add] Add booksonic-air (#294)
* Add booksonic-air

Signed-off-by: TJ Wesolowski <wojoinc@pm.me>

* Clarify description in values.yaml

Signed-off-by: TJ Wesolowski <wojoinc@pm.me>

* Add storage notes

Signed-off-by: TJ Wesolowski <wojoinc@pm.me>
2020-12-03 20:28:58 -05:00
ᗪєνιη ᗷυнʟ
d310100ce7 [unifi-poller] switch to environment variables (#293) 2020-12-02 14:35:57 -05:00
Ryan Holt
c2d6c1c4f1 Merge pull request #290 from k8s-at-home/add-photoprism-chart
[photoprism] Add PhotoPrism Chart Leveraging Library
2020-12-01 09:48:23 -05:00
Ryan Holt
9e2a21995a add default values for photoprism's storage locations to env
Signed-off-by: Ryan Holt <ryan@ryanholt.net>
2020-12-01 09:06:58 -05:00
Ryan Holt
643db3e8cb make appversion a string to pass CT tests
Signed-off-by: Ryan Holt <ryan@ryanholt.net>
2020-12-01 08:43:02 -05:00
Ryan Holt
6822d3c4b4 add a link to photoprism documentation within values
Signed-off-by: Ryan Holt <ryan@ryanholt.net>
2020-12-01 08:41:31 -05:00
Ryan Holt
df1a3f48e0 various changes per @onedr0p comments
Signed-off-by: Ryan Holt <ryan@ryanholt.net>
2020-12-01 08:38:22 -05:00
Ryan Holt
6cf224d0b0 add more readme details
Signed-off-by: Ryan Holt <ryan@ryanholt.net>
2020-11-30 17:48:06 -05:00
Ryan Holt
bca5078de2 add newline
Signed-off-by: Ryan Holt <ryan@ryanholt.net>
2020-11-30 17:45:41 -05:00
Ryan Holt
31e992eb4a add sample originals mount
Signed-off-by: Ryan Holt <ryan@ryanholt.net>
2020-11-30 17:43:47 -05:00
Ryan Holt
a4c913aef6 revert from originals to config pv name
Signed-off-by: Ryan Holt <ryan@ryanholt.net>
2020-11-30 17:23:45 -05:00
Ryan Holt
d4dd34a24b add some additional photoprism default values and mounts
Signed-off-by: Ryan Holt <ryan@ryanholt.net>
2020-11-30 17:02:16 -05:00
Ryan Holt
19cb90701e initial commit for base photoprism files
Signed-off-by: Ryan Holt <ryan@ryanholt.net>
2020-11-30 16:42:46 -05:00
ᗪєνιη ᗷυнʟ
f0eb5542a4 [traccar] new chart (#289) 2020-11-30 15:46:23 -05:00
Thomas John Wesolowski
f2336eedaa [bitwardenrs] Fix HPA target (#287)
* Add persistence type check to hpa target

Signed-off-by: TJ Wesolowski <wojoinc@pm.me>

* Bump chart version

Signed-off-by: TJ Wesolowski <wojoinc@pm.me>

Co-authored-by: ᗪєνιη ᗷυнʟ <onedr0p@users.noreply.github.com>
2020-11-30 06:06:00 -05:00
Bᴇʀɴᴅ Sᴄʜᴏʀɢᴇʀs
915c94887d Fix persistence mountPath (#288)
* Fix persistence mountPath

* Update Chart.yaml
2020-11-30 05:54:24 -05:00
renovate[bot]
f84ac3006b [powerdns] Update helm chart postgresql to v10.1.1 (#285)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-28 18:36:58 -05:00
Ryan Holt
dde8baec08 [powerdns] flip default to postgres, fix postgres dependent chart (#78) 2020-11-28 16:43:54 -05:00
auricom
f2452edfb9 [pyload] new chart (#284) 2020-11-28 16:42:47 -05:00
ᗪєνιη ᗷυнʟ
db1260e482 [grocy/heimdall] use common lib (#283) 2020-11-27 11:30:36 -05:00
Mike K
b1859c895d [paperless] New chart (#252) 2020-11-27 13:28:29 +01:00
renovate[bot]
fed0c7a8f1 [nzbhydra2] Update helm chart common to v1.7.0 (#268)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
Co-authored-by: ᗪєνιη ᗷυнʟ <onedr0p@users.noreply.github.com>
2020-11-26 13:02:05 -05:00
renovate[bot]
40e3179aaf [nzbget] Update helm chart common to v1.7.0 (#267)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-26 12:55:49 -05:00
renovate[bot]
535f78770f [organizr] Update helm chart common to v1.7.0 (#270)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-26 12:52:58 -05:00
renovate[bot]
645f8d4c6e [piaware] Update helm chart common to v1.7.0 (#271)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-26 12:52:11 -05:00
renovate[bot]
7f326fc846 [plex-media-server] Update helm chart common to v1.7.0 (#272)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-26 12:50:23 -05:00
renovate[bot]
df26482e05 [qbittorrent] Update helm chart common to v1.7.0 (#273)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-26 12:44:31 -05:00
renovate[bot]
07fa996d70 [ombi] Update helm chart common to v1.7.0 (#269)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-26 12:42:31 -05:00
renovate[bot]
2a441c9315 [node-red] Update helm chart common to v1.7.0 (#266)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-26 11:40:25 -05:00
renovate[bot]
a16d5ef3a6 [radarr] Update helm chart common to v1.7.0 (#274)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-26 09:52:02 -05:00
renovate[bot]
55c30ff34c [sabnzbd] Update helm chart common to v1.7.0 (#275)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-26 09:49:29 -05:00
renovate[bot]
3281996a22 [sonarr] Update helm chart common to v1.7.0 (#276)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-26 09:47:45 -05:00
renovate[bot]
5ebaccedb3 [tautulli] Update helm chart common to v1.7.0 (#277)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-26 09:46:43 -05:00
renovate[bot]
111c993fea [unifi-poller] Update helm chart common to v1.7.0 (#278)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-26 09:43:47 -05:00
renovate[bot]
95a84496db [xteve] Update helm chart common to v1.7.0 (#279)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-26 09:42:51 -05:00
renovate[bot]
cba3f59391 [zwave2mqtt] Update helm chart common to v1.7.0 (#281)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-26 09:36:52 -05:00
renovate[bot]
9d0a7efc7d [lidarr] Update helm chart common to v1.7.0 (#265)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-26 09:34:24 -05:00
renovate[bot]
d0bad9dfb0 [jellyfin] Update helm chart common to v1.7.0 (#264)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-26 09:28:57 -05:00
renovate[bot]
67eed62c99 [jackett] Update helm chart common to v1.7.0 (#263)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-26 09:15:35 -05:00
renovate[bot]
ae38a94541 [zigbee2mqtt] Update helm chart common to v1.7.0 (#280)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-26 09:07:18 -05:00
renovate[bot]
40a2bb2d26 [homer] Update helm chart common to v1.7.0 (#262)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-26 09:01:33 -05:00
renovate[bot]
fa75854137 [freshrss] Update helm chart common to v1.7.0 (#261)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-26 09:01:25 -05:00
renovate[bot]
c379eb9efd [flood] Update helm chart common to v1.7.0 (#260)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-26 09:01:20 -05:00
Bᴇʀɴᴅ Sᴄʜᴏʀɢᴇʀs
5b64342920 [octoprint] New chart (#282) 2020-11-26 09:00:25 -05:00
renovate[bot]
ab746638e8 [ddclient] Update helm chart common to v1.7.0 (#259)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-26 08:56:31 -05:00
renovate[bot]
643d60fed7 [dashmachine] Update helm chart common to v1.7.0 (#258)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-26 08:56:23 -05:00
renovate[bot]
d5377e032d [couchpotato] Update helm chart common to v1.7.0 (#257)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-26 08:56:15 -05:00
renovate[bot]
d6e87293c1 [bazarr] Update helm chart common to v1.7.0 (#256)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-26 08:51:29 -05:00
renovate[bot]
5ae5ab007d [alertmanager-bot] Update helm chart common to v1.7.0 (#255)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-26 08:51:05 -05:00
ᗪєνιη ᗷυнʟ
a077a803d4 add separate pre-release job with turnstyle (#254) 2020-11-26 14:19:28 +01:00
Mike K
45b6282a15 [common] Add args option to mainContainer (#253)
Co-authored-by: Bᴇʀɴᴅ Sᴄʜᴏʀɢᴇʀs <6213398+bjw-s@users.noreply.github.com>
2020-11-26 08:11:01 -05:00
ᗪєνιη ᗷυнʟ
e6a9d994f9 [multi-chart] bump patch (#251) 2020-11-25 19:40:41 -05:00
renovate[bot]
e94ef45b70 [alertmanager-bot] Update helm chart common to v1.6.2 (#223)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-25 16:19:01 -05:00
renovate[bot]
86907798e0 [jellyfin] Update helm chart common to v1.6.2 (#232)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-25 16:18:56 -05:00
renovate[bot]
1218e56529 [nzbhydra2] Update helm chart common to v1.6.2 (#236)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-25 16:18:49 -05:00
renovate[bot]
902574734b [organizr] Update helm chart common to v1.6.2 (#238)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-25 16:18:40 -05:00
renovate[bot]
f276356baa [piaware] Update helm chart common to v1.6.2 (#239)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-25 16:18:35 -05:00
renovate[bot]
2fdce70485 [plex-media-server] Update helm chart common to v1.6.2 (#240)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-25 16:18:25 -05:00
renovate[bot]
56b03ff1f1 [qbittorrent] Update helm chart common to v1.6.2 (#241)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-25 16:18:20 -05:00
renovate[bot]
6c1e3f0e23 [unifi-poller] Update helm chart common to v1.6.2 (#246)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-25 16:18:11 -05:00
renovate[bot]
25e129b57c [xteve] Update helm chart common to v1.6.2 (#247)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-25 16:17:46 -05:00
renovate[bot]
e0bae04364 [zigbee2mqtt] Update helm chart common to v1.6.2 (#248)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-25 16:17:39 -05:00
renovate[bot]
03473ff13f [zwave2mqtt] Update helm chart common to v1.6.2 (#249)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-25 16:17:30 -05:00
renovate[bot]
a024f2c9b4 [node-red] Update helm chart common to v1.6.2 (#234)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-25 16:16:27 -05:00
renovate[bot]
9a43343d0e [ombi] Update helm chart common to v1.6.2 (#237)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-25 16:16:00 -05:00
renovate[bot]
ec4525d0c1 [homer] Update helm chart common to v1.6.2 (#230)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-25 16:15:35 -05:00
renovate[bot]
85a391e902 [sabnzbd] Update helm chart common to v1.6.2 (#243)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-25 16:15:16 -05:00
renovate[bot]
e301c08620 [nzbget] Update helm chart common to v1.6.2 (#235)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-25 16:15:02 -05:00
renovate[bot]
429db07f64 [sonarr] Update helm chart common to v1.6.2 (#244)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-25 16:14:45 -05:00
renovate[bot]
6a281a2828 [lidarr] Update helm chart common to v1.6.2 (#233)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-25 16:14:32 -05:00
renovate[bot]
d021b87405 [jackett] Update helm chart common to v1.6.2 (#231)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-25 16:14:11 -05:00
renovate[bot]
6174dfc8ef [radarr] Update helm chart common to v1.6.2 (#242)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-25 16:13:53 -05:00
renovate[bot]
5e9897d8cc [tautulli] Update helm chart common to v1.6.2 (#245)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-25 16:13:12 -05:00
renovate[bot]
819e3a6fce [freshrss] Update helm chart common to v1.6.2 (#229)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-25 16:12:48 -05:00
renovate[bot]
b278fc51b5 [flood] Update helm chart common to v1.6.2 (#228)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-25 16:12:24 -05:00
renovate[bot]
94521bcdda [ddclient] Update helm chart common to v1.6.2 (#227)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-25 16:12:04 -05:00
renovate[bot]
fa5cc7171c [bazarr] Update helm chart common to v1.6.2 (#224)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-25 16:11:49 -05:00
ᗪєνιη ᗷυнʟ
58af006816 [renovate] only rebase when conflicted (#250) 2020-11-25 16:11:29 -05:00
renovate[bot]
fba5edf834 [dashmachine] Update helm chart common to v1.6.2 (#226)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-25 15:45:40 -05:00
renovate[bot]
250d308995 [couchpotato] Update helm chart common to v1.6.2 (#225)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-25 15:45:13 -05:00
ᗪєνιη ᗷυнʟ
dc90788f78 remove gitAuthor from renovate config (#222) 2020-11-25 21:40:04 +01:00
ᗪєνιη ᗷυнʟ
e93bf613f1 [common] set dnspolicy default to ClusterFirst (#221) 2020-11-25 21:10:25 +01:00
renovate[bot]
5edd9b8b00 [teslamate] Update helm chart postgresql to v10.1.1 (#219)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-25 20:27:44 +01:00
renovate[bot]
70990295db [statping] Update helm chart postgresql to v10.1.1 (#218)
Co-authored-by: Renovate Bot <bot@renovateapp.com>
2020-11-25 20:22:30 +01:00
ᗪєνιη ᗷυнʟ
66feb4ca7e add suppressNotifications (#220) 2020-11-25 17:25:48 +01:00
ᗪєνιη ᗷυнʟ
693902005a [home-assistant] update postgresql to use right options (#215) 2020-11-25 11:19:46 -05:00
ᗪєνιη ᗷυнʟ
10ac770b40 [homer] add new chart (#216) 2020-11-24 19:50:40 -05:00
Ryan Holt
8d6ec42996 Merge pull request #214 from k8s-at-home/alertmanager-bot-common
[alertmanager-bot] new chart
2020-11-24 19:05:10 -05:00
Devin Buhl
2632a248a4 [alertmanager-bot] fix bci errors 2020-11-24 17:12:18 -05:00
Devin Buhl
c78292c8e7 [alertmanager-bot] fix blank lines 2020-11-24 17:07:49 -05:00
ᗪєνιη ᗷυнʟ
226c0d19c1 remove tabs 2020-11-24 17:03:56 -05:00
ᗪєνιη ᗷυнʟ
c46da9968d Update ct-values.yaml 2020-11-24 17:03:37 -05:00
Devin Buhl
0bf211c8e9 [alertmanager-bot] new chart 2020-11-24 17:02:04 -05:00
Ryan Holt
4ef7b2b312 [dashmachine] convert dashmachine to common library and bump to 2.0.0 (#213)
Signed-off-by: Ryan Holt <ryan@ryanholt.net>
2020-11-24 13:48:07 -05:00
214 changed files with 2635 additions and 3094 deletions

View File

@@ -1,10 +1,11 @@
{
"enabled": true,
"gitAuthor": "Renovate Bot <renovate@whitesourcesoftware.com>",
"commitMessagePrefix": "[{{{parentDir}}}]",
"branchTopic": "{{{parentDir}}}-{{{depNameSanitized}}}-{{{newMajor}}}{{#if isPatch}}.{{{newMinor}}}{{/if}}.x{{#if isLockfileUpdate}}-lockfile{{/if}}",
"assigneesFromCodeOwners": true,
"reviewersFromCodeOwners": true,
"suppressNotifications": ["prIgnoreNotification"],
"rebaseWhen": "conflicted",
"helm-values": {
"enabled": false
},

View File

@@ -8,8 +8,19 @@ on:
- "charts/**"
jobs:
release:
pre-release:
runs-on: ubuntu-latest
timeout-minutes: 5
steps:
- name: Block concurrent releases
uses: softprops/turnstyle@v1
with:
continue-after-seconds: 180
env:
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
release:
needs: pre-release
runs-on: ubuntu-latest
steps:
- name: Checkout
@@ -17,13 +28,6 @@ jobs:
with:
fetch-depth: 0
- name: Turnstyle
uses: softprops/turnstyle@v1
with:
continue-after-seconds: 180
env:
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
- name: Configure Git
run: |
git config user.name "$GITHUB_ACTOR"
@@ -45,8 +49,8 @@ jobs:
# needed until https://github.com/helm/chart-releaser/issues/90
# or helm/chart-releaser-action supports this
post-release:
runs-on: ubuntu-latest
needs: release
runs-on: ubuntu-latest
steps:
- name: Checkout
uses: actions/checkout@v2

View File

@@ -0,0 +1,24 @@
# Patterns to ignore when building packages.
# This supports shell glob matching, relative path matching, and
# negation (prefixed with !). Only one pattern per line.
.DS_Store
# Common VCS dirs
.git/
.gitignore
.bzr/
.bzrignore
.hg/
.hgignore
.svn/
# Common backup files
*.swp
*.bak
*.tmp
*~
# Various IDEs
.project
.idea/
*.tmproj
.vscode/
# OWNERS file for Kubernetes
OWNERS

View File

@@ -0,0 +1,21 @@
apiVersion: v2
appVersion: 0.4.2
description: Bot for Prometheus Alertmanager
name: alertmanager-bot
version: 1.2.0
keywords:
- alertmanager
- telegram
- bot
- alerting
home: https://github.com/k8s-at-home/charts/tree/master/charts/alertmanager-bot
sources:
- https://hub.docker.com/r/metalmatze/alertmanager-bot
- https://github.com/metalmatze/alertmanager-bot
maintainers:
- name: billimek
email: jeff@billimek.com
dependencies:
- name: common
repository: https://k8s-at-home.com/charts/
version: 1.7.0

View File

@@ -0,0 +1,8 @@
approvers:
- billimek
- onedr0p
- bjw-s
reviewers:
- billimek
- onedr0p
- bjw-s

View File

@@ -0,0 +1,61 @@
# alertmanager-bot
This is a helm chart for [alertmanager-bot](https://github.com/metalmatze/alertmanager-bot).
**This chart is not maintained by the upstream project and any issues with the chart should be raised [here](https://github.com/k8s-at-home/charts/issues/new/choose)**
## TL;DR;
```shell
$ helm repo add k8s-at-home https://k8s-at-home.com/charts/
$ helm install k8s-at-home/alertmanager-bot
```
## Installing the Chart
To install the chart with the release name `my-release`:
```console
helm install --name my-release k8s-at-home/alertmanager-bot
```
## Uninstalling the Chart
To uninstall/delete the `my-release` deployment:
```console
helm delete my-release --purge
```
The command removes all the Kubernetes components associated with the chart and deletes the release.
## Configuration
Read through the charts [values.yaml](https://github.com/k8s-at-home/charts/blob/master/charts/alertmanager-bot/values.yaml)
file. It has several commented out suggested values.
Additionally you can take a look at the common library [values.yaml](https://github.com/k8s-at-home/charts/blob/master/charts/common/values.yaml) for more (advanced) configuration options.
Specify each parameter using the `--set key=value[,key=value]` argument to `helm install`. For example,
```console
helm install alertmanager-bot \
--set env.TZ="America/New_York" \
k8s-at-home/alertmanager-bot
```
Alternatively, a YAML file that specifies the values for the above parameters can be provided while installing the
chart. For example,
```console
helm install alertmanager-bot k8s-at-home/alertmanager-bot --values values.yaml
```
```yaml
image:
tag: ...
```
---
**NOTE**
If you get
```console
Error: rendered manifests contain a resource that already exists. Unable to continue with install: existing resource conflict: ...`
```
it may be because you uninstalled the chart with `skipuninstall` enabled, you need to manually delete the pvc or use `existingClaim`.

View File

@@ -0,0 +1,7 @@
service:
enabled: false
probes:
liveness:
enabled: false
readiness:
enabled: false

View File

@@ -0,0 +1 @@
{{- include "common.notes.defaultNotes" . -}}

View File

@@ -0,0 +1 @@
{{ include "common.all" . }}

View File

@@ -0,0 +1,26 @@
# Default values for alertmanager-bot.
image:
repository: metalmatze/alertmanager-bot
pullPolicy: IfNotPresent
tag: 0.4.2
strategy:
type: Recreate
# See more environment varaibles in the alertmanager-botdocumentation
# https://github.com/metalmatze/alertmanager-bot
env:
STORE: bolt
BOLT_PATH: /data/bot.db
# ALERTMANAGER_URL:
# TELEGRAM_ADMIN:
# TELEGRAM_TOKEN:
service:
enabled: false
persistence:
data:
enabled: false
emptyDir: false

View File

@@ -0,0 +1,24 @@
# Patterns to ignore when building packages.
# This supports shell glob matching, relative path matching, and
# negation (prefixed with !). Only one pattern per line.
.DS_Store
# Common VCS dirs
.git/
.gitignore
.bzr/
.bzrignore
.hg/
.hgignore
.svn/
# Common backup files
*.swp
*.bak
*.tmp
*~
# Various IDEs
.project
.idea/
*.tmproj
.vscode/
# OWNERS file for Kubernetes
OWNERS

View File

@@ -0,0 +1,20 @@
apiVersion: v2
appVersion: 4.0.5
description: AppDaemon is a loosely coupled, multi-threaded, sandboxed python execution environment for writing automation apps for various types of Home Automation Software including Home Assistant and MQTT.
name: appdaemon
version: 1.0.2
keywords:
- appdaemon
- home-automation
- home-assistant
- mqtt
home: https://github.com/k8s-at-home/charts/tree/master/charts/appdaemon
sources:
- https://github.com/AppDaemon/appdaemon
maintainers:
- name: billimek
email: jeff@billimek.com
dependencies:
- name: common
repository: https://k8s-at-home.com/charts/
version: 1.7.0

8
charts/appdaemon/OWNERS Normal file
View File

@@ -0,0 +1,8 @@
approvers:
- billimek
- onedr0p
- bjw-s
reviewers:
- billimek
- onedr0p
- bjw-s

View File

@@ -0,0 +1,67 @@
# appdaemon
This is a helm chart for [appdaemon](https://github.com/AppDaemon/appdaemon).
**This chart is not maintained by the upstream project and any issues with the chart should be raised [here](https://github.com/k8s-at-home/charts/issues/new/choose)**
## TL;DR;
```shell
$ helm repo add k8s-at-home https://k8s-at-home.com/charts/
$ helm install k8s-at-home/appdaemon
```
## Installing the Chart
To install the chart with the release name `my-release`:
```console
helm install --name my-release k8s-at-home/appdaemon
```
## Uninstalling the Chart
To uninstall/delete the `my-release` deployment:
```console
helm delete my-release --purge
```
The command removes all the Kubernetes components associated with the chart and deletes the release.
## Configuration
Read through the charts [values.yaml](https://github.com/k8s-at-home/charts/blob/master/charts/appdaemon/values.yaml)
file. It has several commented out suggested values.
Additionally you can take a look at the common library [values.yaml](https://github.com/k8s-at-home/charts/blob/master/charts/common/values.yaml) for more (advanced) configuration options.
Specify each parameter using the `--set key=value[,key=value]` argument to `helm install`. For example,
```console
helm install appdaemon \
--set env.TZ="America/New_York" \
k8s-at-home/appdaemon
```
Alternatively, a YAML file that specifies the values for the above parameters can be provided while installing the
chart. For example,
```console
helm install appdaemon k8s-at-home/appdaemon --values values.yaml
```
```yaml
image:
tag: ...
```
---
**NOTE**
If you get
```console
Error: rendered manifests contain a resource that already exists. Unable to continue with install: existing resource conflict: ...`
```
it may be because you uninstalled the chart with `skipuninstall` enabled, you need to manually delete the pvc or use `existingClaim`.
---
## Upgrading an existing Release to a new major version
A major chart version change (like 4.0.1 -> 5.0.0) indicates that there is an incompatible breaking change potentially needing manual actions.

View File

@@ -0,0 +1,8 @@
env:
TZ: "Africa/Cairo"
LATITUDE: "29.9792"
LONGITUDE: "31.1342"
ELEVATION: "64"
DASH_URL: "http://0.0.0.0:5050"
ingress:
enabled: true

View File

@@ -0,0 +1 @@
{{- include "common.notes.defaultNotes" . -}}

View File

@@ -0,0 +1 @@
{{ include "common.all" . }}

View File

@@ -0,0 +1,93 @@
# Default values for Appdaemon.
# Use our own image at https://github.com/k8s-at-home/container-images/tree/main/appdaemon
# until https://github.com/AppDaemon/appdaemon/pull/1002 is merged
image:
repository: k8sathome/appdaemon
pullPolicy: IfNotPresent
tag: v4.0.5
strategy:
type: Recreate
service:
port:
port: 5050
# # Boilerplate config for code-server functionality, be sure to make changes to suit your environment
# additionalServices:
# - enabled: true
# nameSuffix: vscode
# type: ClusterIP
# port:
# port: 12321
# name: vscode
# protocol: TCP
# targetPort: 12321
# Environment variables that can be configure can be seen at
# https://appdaemon.readthedocs.io/en/latest/DOCKER_TUTORIAL.html
env: {}
# TZ:
# LATITUDE:
# LONGITUDE:
# ELEVATION:
# HA_URL:
# DASH_URL:
# TOKEN:
# # Boilerplate config for code-server functionality, be sure to make changes to suit your environment
# additionalContainers:
# - name: vscode
# image: codercom/code-server:3.7.2
# imagePullPolicy: IfNotPresent
# workingDir: /www/assets
# args:
# - "--port"
# - "12321"
# - "--auth"
# - "none"
# ports:
# - name: vscode
# containerPort: 12321
# protocol: TCP
# volumeMounts:
# - name: config
# mountPath: /www/assets
# securityContext:
# runAsUser: 0
ingress:
enabled: false
# # Boilerplate config for code-server functionality, be sure to make changes to suit your environment
# additionalIngresses:
# - enabled: true
# nameSuffix: "vscode"
# serviceName: appdaemon-vscode
# servicePort: 12321
# annotations:
# kubernetes.io/ingress.class: "nginx"
# hosts:
# - host: appdaemon-editor.k8s-at-home.com
# paths:
# - path: /
# pathType: Prefix
# tls:
# - hosts:
# - appdaemon-editor.k8s-at-home.com
persistence:
config:
enabled: false
emptyDir: false
## Persistent Volume Storage Class
## If defined, storageClassName: <storageClass>
## If set to "-", storageClassName: "", which disables dynamic provisioning
## If undefined (the default) or set to null, no storageClassName spec is
## set, choosing the default provisioner. (gp2 on AWS, standard on
## GKE, AWS & OpenStack)
# storageClass: "-"
# accessMode: ReadWriteOnce
# size: 1Gi
## Set to true to retain the PVC upon helm uninstall
# skipuninstall: false
# existingClaim: ""

View File

@@ -2,7 +2,7 @@ apiVersion: v2
appVersion: v0.9.0.5
description: Bazarr is a companion application to Sonarr and Radarr. It manages and downloads subtitles based on your requirements
name: bazarr
version: 4.6.1
version: 4.8.0
keywords:
- bazarr
- radarr
@@ -21,4 +21,4 @@ maintainers:
dependencies:
- name: common
repository: https://k8s-at-home.com/charts/
version: 1.6.1
version: 1.7.0

View File

@@ -2,7 +2,7 @@ apiVersion: v2
name: bitwardenrs
description: Unofficial Bitwarden compatible server written in Rust
type: application
version: 1.0.0
version: 1.0.1
appVersion: 1.16.3
keywords:
- bitwarden

View File

@@ -8,7 +8,11 @@ metadata:
spec:
scaleTargetRef:
apiVersion: apps/v1
{{- if eq .Values.persistence.type "statefulset" }}
kind: StatefulSet
{{- else }}
kind: Deployment
{{- end }}
name: {{ include "bitwardenrs.fullname" . }}
minReplicas: {{ .Values.autoscaling.minReplicas }}
maxReplicas: {{ .Values.autoscaling.maxReplicas }}

View File

@@ -0,0 +1,20 @@
apiVersion: v2
appVersion: v2009.1.0
description: Booksonic is a platform for accessing the audibooks you own wherever you are
name: booksonic-air
version: 1.0.0
keywords:
- booksonic
- audiobook
home: https://github.com/k8s-at-home/charts/tree/master/charts/booksonic-air
icon: https://raw.githubusercontent.com/linuxserver/docker-templates/master/linuxserver.io/img/booksonic-air.png
sources:
- https://github.com/popeen/Booksonic-Air
- https://hub.docker.com/r/linuxserver/booksonic-air
maintainers:
- name: wojoinc
email: wojoinc@pm.me
dependencies:
- name: common
repository: https://k8s-at-home.com/charts/
version: 1.7.0

View File

@@ -0,0 +1,8 @@
approvers:
- billimek
- onedr0p
- bjw-s
reviewers:
- billimek
- onedr0p
- bjw-s

View File

@@ -0,0 +1,78 @@
# Booksonic-air
This is a helm chart for [Booksonic-air](https://github.com/popeen/Booksonic-Air).
Booksonic-air is the successor to [Booksonic](https://github.com/popeen/Booksonic-Air#history).
**This chart is not maintained by the upstream project and any issues with the chart should be raised [here](https://github.com/k8s-at-home/charts/issues/new/choose)**
This chart leverages the [linuxserver.io](https://hub.docker.com/r/linuxserver/booksonic-air) image. To specify a specific version, or to use a different image (not recommended),
see [configuration](#configuration)
## TL;DR;
```shell
$ helm repo add k8s-at-home https://k8s-at-home.com/charts/
$ helm install k8s-at-home/booksonic-air
```
## Storage
If you plan to use networked storage to store your media or config for Booksonic, (NFS, etc.) please take a look at the
Fast Access option in the Booksonic settings. This will help improve the perfomance of the application
by not constantly monitoring media folders.
## Installing the Chart
To install the chart with the release name `my-release`:
```console
helm install --name my-release k8s-at-home/booksonic-air
```
## Uninstalling the Chart
To uninstall/delete the `my-release` deployment:
```console
helm delete my-release --purge
```
The command removes all the Kubernetes components associated with the chart and deletes the release.
## Configuration
Read through the charts [values.yaml](https://github.com/k8s-at-home/charts/blob/master/charts/booksonic-air/values.yaml)
file. It has several commented out suggested values. Most notably, these include several environment variables used to
customize the container.
Additionally you can take a look at the common library [values.yaml](https://github.com/k8s-at-home/charts/blob/master/charts/common/values.yaml) for more (advanced) configuration options.
Specify each parameter using the `--set key=value[,key=value]` argument to `helm install`. For example,
```console
helm install booksonic \
--set env.TZ="America/New_York" \
k8s-at-home/booksonic-air
```
Alternatively, a YAML file that specifies the values for the above parameters can be provided while installing the
chart. For example,
```console
helm install booksonic k8s-at-home/booksonic-air -f values.yaml
```
```yaml
image:
tag: ...
```
---
**NOTE**
If you get
```console
Error: rendered manifests contain a resource that already exists. Unable to continue with install: existing resource conflict: ...`
```
it may be because you uninstalled the chart with `skipuninstall` enabled, you need to manually delete the pvc or use `existingClaim`.
---
## Upgrading an existing Release to a new major version
A major chart version change (like 4.0.1 -> 5.0.0) indicates that there is an incompatible breaking change potentially needing manual actions.

View File

@@ -0,0 +1,2 @@
ingress:
enabled: true

View File

@@ -0,0 +1 @@
{{- include "common.notes.defaultNotes" . -}}

View File

@@ -0,0 +1 @@
{{ include "common.all" . }}

View File

@@ -0,0 +1,49 @@
# Default values for Booksonic-air.
image:
repository: linuxserver/booksonic-air
pullPolicy: IfNotPresent
tag: v2009.1.0-ls2
strategy:
type: Recreate
service:
port:
port: 4040
env: {}
# TZ: UTC
# PUID: "1001"
# PGID: "1001"
# CONTEXT_PATH=url-base # Used to set the base path for reverse proxies eg. /booksonic, /books, etc.
persistence:
config:
enabled: false
emptyDir: false
audiobooks:
enabled: false
emptyDir: false
mountPath: /audiobooks
## Persistent Volume Storage Class
## If defined, storageClassName: <storageClass>
## If set to "-", storageClassName: "", which disables dynamic provisioning
## If undefined (the default) or set to null, no storageClassName spec is
## set, choosing the default provisioner. (gp2 on AWS, standard on
## GKE, AWS & OpenStack)
# storageClass: "-"
# accessMode: ReadWriteOnce
# size: 1Gi
## Do not delete the pvc upon helm uninstall
# skipuninstall: false
# existingClaim: ""
podcasts:
enabled: false
emptyDir: false
mountPath: /podcasts
othermedia:
enabled: false
emptyDir: false
mountPath: /othermedia

View File

@@ -14,10 +14,11 @@
*.swp
*.bak
*.tmp
*.orig
*~
# Various IDEs
.project
.idea/
*.tmproj
.vscode/
# OWNERS file for Kubernetes
OWNERS

View File

@@ -1,9 +1,8 @@
type: application
apiVersion: v2
appVersion: 0.6.8
appVersion: 0.6.9
description: Calibre-Web is a web app providing a clean interface for browsing, reading and downloading eBooks using an existing Calibre database.
name: calibre-web
version: 2.0.1
version: 3.0.0
keywords:
- calibre
- ebook
@@ -15,3 +14,7 @@ sources:
maintainers:
- name: wojoinc
email: wojoinc@pm.me
dependencies:
- name: common
repository: https://k8s-at-home.com/charts/
version: 1.7.0

View File

@@ -1,4 +1,8 @@
approvers:
- billimek
- onedr0p
- bjw-s
reviewers:
- billimek
- billimek
- onedr0p
- bjw-s

View File

@@ -1,14 +1,22 @@
# calibre-web
# Calibre-Web
This is a helm chart for [calibre-web](https://github.com/janeczku/calibre-web) utilizing the [Linuxserver.io calibre-web image](https://hub.docker.com/r/linuxserver/calibre-web)
This is a helm chart for [Calibre-Web](https://github.com/janeczku/calibre-web).
**This chart is not maintained by the upstream project and any issues with the chart should be raised [here](https://github.com/k8s-at-home/charts/issues/new/choose)**
## TL;DR;
```shell
helm repo add k8s-at-home https://k8s-at-home.com/charts/
helm install k8s-at-home/calibre-web
$ helm repo add k8s-at-home https://k8s-at-home.com/charts/
$ helm install k8s-at-home/calibre-web
```
## Storage
If you plan to use networked storage to store your media or config for Booksonic, (NFS, etc.) please take a look at the
Fast Access option in the Booksonic settings. This will help improve the perfomance of the application
by not constantly monitoring media folders.
## Installing the Chart
To install the chart with the release name `my-release`:
@@ -28,80 +36,50 @@ helm delete my-release --purge
The command removes all the Kubernetes components associated with the chart and deletes the release.
## Configuration
The following tables lists the configurable parameters of the `calibre-web` chart and their default values.
| Parameter | Description | Default |
| --------------------------------------------- | ----------------------------------------------------------------------------------------------------------------------------------------------------------------------- | ------------------------- |
| `image.repository` | Image repository | `linuxserver/calibre-web` |
| `image.tag` | Image tag. Possible values listed [here](https://hub.docker.com/r/linuxserver/calibre-web/tags/). | `0.6.8-ls74` |
| `image.pullPolicy` | Image pull policy | `IfNotPresent` |
| `dockerMods.ebookConversion.enabled` | Enable optional ebook conversion feature. (x86_64 only) refer [here](https://github.com/linuxserver/docker-calibre-web#application-setup) | `false` |
| `dockerMods.ebookConversion.image.repository` | DockerMod image repository | `linuxserver/calibre-web` |
| `dockerMods.ebookConversion.image.tag` | DockerMod image tag. These typically start with `v`, with the "stable" tag being `calibre`. Can be found [here](https://hub.docker.com/r/linuxserver/calibre-web/tags/) | `calibre` |
| `strategyType` | Specifies the strategy used to replace old Pods by new ones | `Recreate` |
| `timezone` | Timezone the calibre-web instance should run as, e.g. 'America/New_York' | `UTC` |
| `puid` | process userID the calibre-web instance should run as | `1001` |
| `pgid` | process groupID the calibre-web instance should run as | `1001` |
| `umask` | Default umask to use when creating new files | `22` |
| `probes.liveness.initialDelaySeconds` | Specify liveness `initialDelaySeconds` parameter for the deployment | `60` |
| `probes.liveness.failureThreshold` | Specify liveness `failureThreshold` parameter for the deployment | `5` |
| `probes.liveness.timeoutSeconds` | Specify liveness `timeoutSeconds` parameter for the deployment | `10` |
| `probes.readiness.initialDelaySeconds` | Specify readiness `initialDelaySeconds` parameter for the deployment | `60` |
| `probes.readiness.failureThreshold` | Specify readiness `failureThreshold` parameter for the deployment | `5` |
| `probes.readiness.timeoutSeconds` | Specify readiness `timeoutSeconds` parameter for the deployment | `10` |
| `Service.type` | Kubernetes service type for the calibre-web GUI | `ClusterIP` |
| `Service.port` | Kubernetes port where the calibre-web GUI is exposed | `8083` |
| `Service.annotations` | Service annotations for the calibre-web GUI | `{}` |
| `Service.labels` | Custom labels | `{}` |
| `Service.loadBalancerIP` | Loadbalance IP for the calibre-web GUI | `{}` |
| `Service.loadBalancerSourceRanges` | List of IP CIDRs allowed access to load balancer (if supported) | None |
| `ingress.enabled` | Enables Ingress | `false` |
| `ingress.annotations` | Ingress annotations | `{}` |
| `ingress.labels` | Custom labels | `{}` |
| `ingress.path` | Ingress path | `/` |
| `ingress.hosts` | Ingress accepted hostnames | `chart-example.local` |
| `ingress.tls` | Ingress TLS configuration | `[]` |
| `persistence.config.enabled` | Use persistent volume to store configuration data | `true` |
| `persistence.config.size` | Size of persistent volume claim | `1Gi` |
| `persistence.config.existingClaim` | Use an existing PVC to persist data | `nil` |
| `persistence.config.storageClass` | Type of persistent volume claim | `-` |
| `persistence.config.accessMode` | Persistence access mode | `ReadWriteOnce` |
| `persistence.config.skipuninstall` | Do not delete the pvc upon helm uninstall | `false` |
| `persistence.books.enabled` | Use persistent volume to store books data | `true` |
| `persistence.books.size` | Size of persistent volume claim | `10Gi` |
| `persistence.books.existingClaim` | Use an existing PVC to persist data | `nil` |
| `persistence.books.storageClass` | Type of persistent volume claim | `-` |
| `persistence.books.accessMode` | Persistence access mode | `ReadWriteOnce` |
| `persistence.books.skipuninstall` | Do not delete the pvc upon helm uninstall | `false` |
| `persistence.extraExistingClaimMounts` | Optionally add multiple existing claims | `[]` |
| `resources` | CPU/Memory resource requests/limits | `{}` |
| `nodeSelector` | Node labels for pod assignment | `{}` |
| `tolerations` | Toleration labels for pod assignment | `[]` |
| `affinity` | Affinity settings for pod assignment | `{}` |
| `podAnnotations` | Key-value pairs to add as pod annotations | `{}` |
| `deploymentAnnotations` | Key-value pairs to add as deployment annotations | `{}` |
Read through the charts [values.yaml](https://github.com/k8s-at-home/charts/blob/master/charts/calibre-web/values.yaml)
file. It has several commented out suggested values. Most notably, these include several environment variables used to
customize the container.
Additionally you can take a look at the common library [values.yaml](https://github.com/k8s-at-home/charts/blob/master/charts/common/values.yaml) for more (advanced) configuration options.
Specify each parameter using the `--set key=value[,key=value]` argument to `helm install`. For example,
```console
helm install --name my-release \
--set timezone="America/New York" \
helm install calibre-web \
--set env.TZ="America/New_York" \
k8s-at-home/calibre-web
```
Alternatively, a YAML file that specifies the values for the above parameters can be provided while installing the chart. For example,
Alternatively, a YAML file that specifies the values for the above parameters can be provided while installing the
chart. For example,
```console
helm install --name my-release -f values.yaml stable/calibre-web
helm install calibre-web k8s-at-home/calibre-web -f values.yaml
```
```yaml
image:
tag: ...
```
---
**NOTE**
If you get `Error: rendered manifests contain a resource that already exists. Unable to continue with install: existing resource conflict: ...` it may be because you uninstalled the chart with `skipuninstall` enabled, you need to manually delete the pvc or use `existingClaim`.
If you get
```console
Error: rendered manifests contain a resource that already exists. Unable to continue with install: existing resource conflict: ...`
```
it may be because you uninstalled the chart with `skipuninstall` enabled, you need to manually delete the pvc or use `existingClaim`.
---
Read through the [values.yaml](https://github.com/k8s-at-home/charts/blob/master/charts/calibre-web/values.yaml) file. It has several commented out suggested values.
## Upgrading an existing Release to a new major version
A major chart version change (like 4.0.1 -> 5.0.0) indicates that there is an incompatible breaking change potentially needing manual actions.
### Upgrading from 2.x.x to 3.x.x
Due to migrating to a centralized common library some values in `values.yaml` have changed.
Examples:
* `service.port` has been moved to `service.port.port`.
* `persistence.type` has been moved to `controllerType`.
Refer to the library values.yaml for more configuration options.

View File

@@ -0,0 +1,2 @@
ingress:
enabled: true

View File

@@ -1,25 +1 @@
1. Get the application URL by running these commands:
{{- if .Values.ingress.enabled }}
{{- range $host := .Values.ingress.hosts }}
http{{ if $.Values.ingress.tls }}s{{ end }}://{{ $host }}{{ $.Values.ingress.path }}
{{- end }}
{{- else if contains "NodePort" .Values.service.type }}
export NODE_PORT=$(kubectl get --namespace {{ .Release.Namespace }} -o jsonpath="{.spec.ports[0].nodePort}" services {{ include "calibre-web.fullname" . }})
export NODE_IP=$(kubectl get nodes --namespace {{ .Release.Namespace }} -o jsonpath="{.items[0].status.addresses[0].address}")
echo http://$NODE_IP:$NODE_PORT
{{- else if contains "LoadBalancer" .Values.service.type }}
NOTE: It may take a few minutes for the LoadBalancer IP to be available.
You can watch the status of by running 'kubectl get --namespace {{ .Release.Namespace }} svc -w {{ include "calibre-web.fullname" . }}'
export SERVICE_IP=$(kubectl get svc --namespace {{ .Release.Namespace }} {{ include "calibre-web.fullname" . }} --template "{{"{{ range (index .status.loadBalancer.ingress 0) }}{{.}}{{ end }}"}}")
echo http://$SERVICE_IP:{{ .Values.service.port }}
{{- else if contains "ClusterIP" .Values.service.type }}
export POD_NAME=$(kubectl get pods --namespace {{ .Release.Namespace }} -l "app.kubernetes.io/name={{ include "calibre-web.name" . }},app.kubernetes.io/instance={{ .Release.Name }}" -o jsonpath="{.items[0].metadata.name}")
echo "Visit http://127.0.0.1:8080 to use your application"
kubectl --namespace {{ .Release.Namespace }} port-forward $POD_NAME 8080:{{ .Values.service.port }}
{{- end }}
{{- if .Values.dockerMods.ebookConversion.enabled }}
2. DockerMods:
You have enabled the EbookConversion Docker Mod using the {{ .Values.dockerMods.ebookConversion.image.repository }}:{{ .Values.dockerMods.ebookConversion.image.tag }} image.
Remember to visit https://github.com/linuxserver/docker-calibre-web#application-setup for instructions on configuring this feature in the calibre-web settings.
{{- end }}
{{- include "common.notes.defaultNotes" . -}}

View File

@@ -1,63 +0,0 @@
{{/* vim: set filetype=mustache: */}}
{{/*
Expand the name of the chart.
*/}}
{{- define "calibre-web.name" -}}
{{- default .Chart.Name .Values.nameOverride | trunc 63 | trimSuffix "-" }}
{{- end }}
{{/*
Create a default fully qualified app name.
We truncate at 63 chars because some Kubernetes name fields are limited to this (by the DNS naming spec).
If release name contains chart name it will be used as a full name.
*/}}
{{- define "calibre-web.fullname" -}}
{{- if .Values.fullnameOverride }}
{{- .Values.fullnameOverride | trunc 63 | trimSuffix "-" }}
{{- else }}
{{- $name := default .Chart.Name .Values.nameOverride }}
{{- if contains $name .Release.Name }}
{{- .Release.Name | trunc 63 | trimSuffix "-" }}
{{- else }}
{{- printf "%s-%s" .Release.Name $name | trunc 63 | trimSuffix "-" }}
{{- end }}
{{- end }}
{{- end }}
{{/*
Create chart name and version as used by the chart label.
*/}}
{{- define "calibre-web.chart" -}}
{{- printf "%s-%s" .Chart.Name .Chart.Version | replace "+" "_" | trunc 63 | trimSuffix "-" }}
{{- end }}
{{/*
Common labels
*/}}
{{- define "calibre-web.labels" -}}
helm.sh/chart: {{ include "calibre-web.chart" . }}
{{ include "calibre-web.selectorLabels" . }}
{{- if .Chart.AppVersion }}
app.kubernetes.io/version: {{ .Chart.AppVersion | quote }}
{{- end }}
app.kubernetes.io/managed-by: {{ .Release.Service }}
{{- end }}
{{/*
Selector labels
*/}}
{{- define "calibre-web.selectorLabels" -}}
app.kubernetes.io/name: {{ include "calibre-web.name" . }}
app.kubernetes.io/instance: {{ .Release.Name }}
{{- end }}
{{/*
Create the name of the service account to use
*/}}
{{- define "calibre-web.serviceAccountName" -}}
{{- if .Values.serviceAccount.create }}
{{- default (include "calibre-web.fullname" .) .Values.serviceAccount.name }}
{{- else }}
{{- default "default" .Values.serviceAccount.name }}
{{- end }}
{{- end }}

View File

@@ -1,25 +0,0 @@
{{- if and .Values.persistence.books.enabled (not .Values.persistence.books.existingClaim) }}
kind: PersistentVolumeClaim
apiVersion: v1
metadata:
name: {{ template "calibre-web.fullname" . }}-books
{{- if .Values.persistence.books.skipuninstall }}
annotations:
"helm.sh/resource-policy": keep
{{- end }}
labels: {{ include "calibre-web.labels" . | nindent 4}}
spec:
accessModes:
- {{ .Values.persistence.books.accessMode | quote }}
resources:
requests:
storage: {{ .Values.persistence.books.size | quote }}
{{- if .Values.persistence.books.storageClass }}
{{- if (eq "-" .Values.persistence.books.storageClass) }}
storageClassName: ""
{{- else }}
storageClassName: "{{ .Values.persistence.books.storageClass }}"
{{- end }}
{{- end }}
{{- end -}}

View File

@@ -0,0 +1 @@
{{ include "common.all" . }}

View File

@@ -1,25 +0,0 @@
{{- if and .Values.persistence.config.enabled (not .Values.persistence.config.existingClaim) }}
kind: PersistentVolumeClaim
apiVersion: v1
metadata:
name: {{ template "calibre-web.fullname" . }}-config
{{- if .Values.persistence.config.skipuninstall }}
annotations:
"helm.sh/resource-policy": keep
{{- end }}
labels: {{ include "calibre-web.labels" . | nindent 4 }}
spec:
accessModes:
- {{ .Values.persistence.config.accessMode | quote }}
resources:
requests:
storage: {{ .Values.persistence.config.size | quote }}
{{- if .Values.persistence.config.storageClass }}
{{- if (eq "-" .Values.persistence.config.storageClass) }}
storageClassName: ""
{{- else }}
storageClassName: "{{ .Values.persistence.config.storageClass }}"
{{- end }}
{{- end }}
{{- end -}}

View File

@@ -1,112 +0,0 @@
apiVersion: apps/v1
kind: Deployment
metadata:
name: {{ include "calibre-web.fullname" . }}
{{- if .Values.deploymentAnnotations }}
annotations:
{{- range $key, $value := .Values.deploymentAnnotations }}
{{ $key }}: {{ $value | quote }}
{{- end }}
{{- end }}
labels: {{ include "calibre-web.labels" . | nindent 4 }}
spec:
replicas: 1
revisionHistoryLimit: 3
strategy:
type: {{ .Values.strategyType }}
selector:
matchLabels:
app.kubernetes.io/name: {{ include "calibre-web.name" . }}
app.kubernetes.io/instance: {{ .Release.Name }}
template:
metadata:
labels:
app.kubernetes.io/name: {{ include "calibre-web.name" . }}
app.kubernetes.io/instance: {{ .Release.Name }}
{{- if .Values.podAnnotations }}
annotations:
{{- range $key, $value := .Values.podAnnotations }}
{{ $key }}: {{ $value | quote }}
{{- end }}
{{- end }}
spec:
containers:
- name: {{ .Chart.Name }}
image: "{{ .Values.image.repository }}:{{ .Values.image.tag }}"
imagePullPolicy: {{ .Values.image.pullPolicy }}
ports:
- name: http
containerPort: 8083
protocol: TCP
livenessProbe:
tcpSocket:
port: http
initialDelaySeconds: {{ .Values.probes.liveness.initialDelaySeconds }}
failureThreshold: {{ .Values.probes.liveness.failureThreshold }}
timeoutSeconds: {{ .Values.probes.liveness.timeoutSeconds }}
readinessProbe:
tcpSocket:
port: http
initialDelaySeconds: {{ .Values.probes.readiness.initialDelaySeconds }}
failureThreshold: {{ .Values.probes.readiness.failureThreshold }}
timeoutSeconds: {{ .Values.probes.readiness.timeoutSeconds }}
env:
- name: TZ
value: "{{ .Values.timezone }}"
- name: PUID
value: "{{ .Values.puid }}"
- name: PGID
value: "{{ .Values.pgid }}"
- name: UMASK
value: "{{ .Values.umask }}"
{{- if eq .Values.dockerMods.ebookConversion.enabled true }}
- name: DOCKER_MODS
value: "{{ .Values.dockerMods.ebookConversion.image.repository }}:{{ .Values.dockerMods.ebookConversion.image.tag }}"
{{- end }}
volumeMounts:
- mountPath: /config
name: config
- mountPath: /books
name: books
{{- if .Values.persistence.books.subPath }}
subPath: {{ .Values.persistence.books.subPath }}
{{- end }}
{{- range .Values.persistence.extraExistingClaimMounts }}
- name: {{ .name }}
mountPath: {{ .mountPath }}
readOnly: {{ .readOnly }}
{{- end }}
resources:
{{ toYaml .Values.resources | indent 12 }}
volumes:
- name: config
{{- if .Values.persistence.config.enabled }}
persistentVolumeClaim:
claimName: {{ if .Values.persistence.config.existingClaim }}{{ .Values.persistence.config.existingClaim }}{{- else }}{{ template "calibre-web.fullname" . }}-config{{- end }}
{{- else }}
emptyDir: {}
{{- end }}
- name: books
{{- if .Values.persistence.books.enabled }}
persistentVolumeClaim:
claimName: {{ if .Values.persistence.books.existingClaim }}{{ .Values.persistence.books.existingClaim }}{{- else }}{{ template "calibre-web.fullname" . }}-books{{- end }}
{{- else }}
emptyDir: {}
{{- end }}
{{- range .Values.persistence.extraExistingClaimMounts }}
- name: {{ .name }}
persistentVolumeClaim:
claimName: {{ .existingClaim }}
{{- end }}
{{- with .Values.nodeSelector }}
nodeSelector:
{{ toYaml . | indent 8 }}
{{- end }}
{{- with .Values.affinity }}
affinity:
{{ toYaml . | indent 8 }}
{{- end }}
{{- with .Values.tolerations }}
tolerations:
{{ toYaml . | indent 8 }}
{{- end }}

View File

@@ -1,34 +0,0 @@
{{- if .Values.ingress.enabled -}}
{{- $fullName := include "calibre-web.fullname" . -}}
{{- $ingressPath := .Values.ingress.path -}}
apiVersion: extensions/v1beta1
kind: Ingress
metadata:
name: {{ $fullName }}
labels: {{ include "calibre-web.labels" . | nindent 4 }}
{{- with .Values.ingress.annotations }}
annotations:
{{ toYaml . | indent 4 }}
{{- end }}
spec:
{{- if .Values.ingress.tls }}
tls:
{{- range .Values.ingress.tls }}
- hosts:
{{- range .hosts }}
- {{ . | quote }}
{{- end }}
secretName: {{ .secretName }}
{{- end }}
{{- end }}
rules:
{{- range .Values.ingress.hosts }}
- host: {{ . | quote }}
http:
paths:
- path: {{ $ingressPath }}
backend:
serviceName: {{ $fullName }}
servicePort: http
{{- end }}
{{- end }}

View File

@@ -1,48 +0,0 @@
apiVersion: v1
kind: Service
metadata:
name: {{ template "calibre-web.fullname" . }}
labels: {{ include "calibre-web.labels" . | nindent 4 }}
{{- if .Values.service.labels }}
{{ toYaml .Values.service.labels | indent 4 }}
{{- end }}
{{- with .Values.service.annotations }}
annotations:
{{ toYaml . | indent 4 }}
{{- end }}
spec:
{{- if (or (eq .Values.service.type "ClusterIP") (empty .Values.service.type)) }}
type: ClusterIP
{{- if .Values.service.clusterIP }}
clusterIP: {{ .Values.service.clusterIP }}
{{end}}
{{- else if eq .Values.service.type "LoadBalancer" }}
type: {{ .Values.service.type }}
{{- if .Values.service.loadBalancerIP }}
loadBalancerIP: {{ .Values.service.loadBalancerIP }}
{{- end }}
{{- if .Values.service.loadBalancerSourceRanges }}
loadBalancerSourceRanges:
{{ toYaml .Values.service.loadBalancerSourceRanges | indent 4 }}
{{- end -}}
{{- else }}
type: {{ .Values.service.type }}
{{- end }}
{{- if .Values.service.externalIPs }}
externalIPs:
{{ toYaml .Values.service.externalIPs | indent 4 }}
{{- end }}
{{- if .Values.service.externalTrafficPolicy }}
externalTrafficPolicy: {{ .Values.service.externalTrafficPolicy }}
{{- end }}
ports:
- name: http
port: {{ .Values.service.port }}
protocol: TCP
targetPort: http
{{ if (and (eq .Values.service.type "NodePort") (not (empty .Values.service.nodePort))) }}
nodePort: {{.Values.service.nodePort}}
{{ end }}
selector:
app.kubernetes.io/name: {{ include "calibre-web.name" . }}
app.kubernetes.io/instance: {{ .Release.Name }}

View File

@@ -1,15 +0,0 @@
apiVersion: v1
kind: Pod
metadata:
name: "{{ include "calibre-web.fullname" . }}-test-connection"
labels:
{{- include "calibre-web.labels" . | nindent 4 }}
annotations:
"helm.sh/hook": test-success
spec:
containers:
- name: wget
image: busybox
command: ['wget']
args: ['{{ include "calibre-web.fullname" . }}:{{ .Values.service.port }}']
restartPolicy: Never

View File

@@ -1,136 +1,41 @@
# Default values for calibre-web.
# This is a YAML-formatted file.
# Declare variables to be passed into your templates.
# Default values for Calibre-Web.
image:
repository: linuxserver/calibre-web
tag: 0.6.8-ls74
pullPolicy: IfNotPresent
tag: version-0.6.9
# upgrade strategy type (e.g. Recreate or RollingUpdate)
strategyType: Recreate
# Probes configuration
probes:
liveness:
initialDelaySeconds: 60
failureThreshold: 5
timeoutSeconds: 10
readiness:
initialDelaySeconds: 60
failureThreshold: 5
timeoutSeconds: 10
nameOverride: ""
fullnameOverride: ""
# Linuxserver.io additional layers.
# Enables additional features for the image, at the cost of increased size and
# possible incompatabilities with certain architectures. Disabled by default.
dockerMods:
# Enable ebook conversion. Compatible with x86_64 image only.
# To enable, set enabled to true, and follow the guide at: https://github.com/linuxserver/docker-calibre-web#application-setup
# to configure the application.
ebookConversion:
enabled: false
image:
repository: linuxserver/calibre-web
# To use a different tag for the DockerMod, specify the tag below
tag: calibre
timezone: UTC
puid: 1001
pgid: 1001
umask: 22
strategy:
type: Recreate
service:
type: ClusterIP
port: 8083
## Specify the nodePort value for the LoadBalancer and NodePort service types.
## ref: https://kubernetes.io/docs/concepts/services-networking/service/#type-nodeport
##
# nodePort:
## Provide any additional annotations which may be required. This can be used to
## set the LoadBalancer service type to internal only.
## ref: https://kubernetes.io/docs/concepts/services-networking/service/#internal-load-balancer
##
annotations: {}
labels: {}
## Use loadBalancerIP to request a specific static IP,
## otherwise leave blank
##
loadBalancerIP:
# loadBalancerSourceRanges: []
## Set the externalTrafficPolicy in the Service to either Cluster or Local
# externalTrafficPolicy: Cluster
port:
port: 8083
ingress:
enabled: false
annotations: {}
# kubernetes.io/ingress.class: nginx
# kubernetes.io/tls-acme: "true"
path: /
hosts:
- chart-example.local
tls: []
# - secretName: chart-example-tls
# hosts:
# - chart-example.local
env: {}
# TZ:
# PUID:
# PGID:
# UMASK:
# DOCKER_MODS:
persistence:
config:
enabled: true
## calibre-web configuration data Persistent Volume Storage Class
enabled: false
emptyDir: false
books:
enabled: false
emptyDir: false
## Persistent Volume Storage Class
## If defined, storageClassName: <storageClass>
## If set to "-", storageClassName: "", which disables dynamic provisioning
## If undefined (the default) or set to null, no storageClassName spec is
## set, choosing the default provisioner. (gp2 on AWS, standard on
## GKE, AWS & OpenStack)
##
# storageClass: "-"
##
## If you want to reuse an existing claim, you can pass the name of the PVC using
## the existingClaim variable
# existingClaim: your-claim
accessMode: ReadWriteOnce
size: 1Gi
## Do not delete the pvc upon helm uninstall
skipuninstall: false
books:
enabled: true
# storageClass: "-"
# existingClaim: your-claim
# subPath: some-subpath
accessMode: ReadWriteOnce
size: 10Gi
## Do not delete the pvc upon helm uninstall
skipuninstall: false
extraExistingClaimMounts: []
# - name: external-mount
# mountPath: /srv/external-mount
## A manually managed Persistent Volume and Claim
## If defined, PVC must be created manually before volume will be bound
# existingClaim:
# readOnly: true
resources: {}
# We usually recommend not to specify default resources and to leave this as a conscious
# choice for the user. This also increases chances charts run on environments with little
# resources, such as Minikube. If you do want to specify resources, uncomment the following
# lines, adjust them as necessary, and remove the curly braces after 'resources:'.
# limits:
# cpu: 100m
# memory: 128Mi
# requests:
# cpu: 100m
# memory: 128Mi
nodeSelector: {}
tolerations: []
affinity: {}
podAnnotations: {}
deploymentAnnotations: {}
# accessMode: ReadWriteOnce
# size: 1Gi
## Set to true to retain the PVC upon helm uninstall
# skipuninstall: false
# existingClaim: ""

View File

@@ -2,7 +2,7 @@ apiVersion: v2
name: common
description: Function library for k8s-at-home charts
type: library
version: 1.6.1
version: 1.7.0
keywords:
- k8s-at-home
- common

View File

@@ -5,6 +5,9 @@ The main container included in the controller.
- name: {{ include "common.names.fullname" . }}
image: "{{ .Values.image.repository }}:{{ .Values.image.tag | default .Chart.AppVersion }}"
imagePullPolicy: {{ .Values.image.pullPolicy }}
{{- with .Values.args }}
args: {{ . }}
{{- end }}
{{- with .Values.securityContext }}
securityContext:
{{- toYaml . | nindent 4 }}

View File

@@ -12,6 +12,9 @@ strategy:
## DaemonSets ignore this
type: RollingUpdate
# Override the default args
args: []
# Set annotations on the pod
podAnnotations: {}
@@ -29,7 +32,7 @@ env: {}
# When using hostNetwork make sure you set dnsPolicy to ClusterFirstWithHostNet
hostNetwork: false
dnsPolicy: Default
dnsPolicy: ClusterFirst
initContainers: []

View File

@@ -2,7 +2,7 @@ apiVersion: v2
appVersion: latest
description: CouchPotato (CP) is an automatic NZB and torrent downloader.
name: couchpotato
version: 3.1.1
version: 3.3.0
keywords:
- couchpotato
- usenet
@@ -18,4 +18,4 @@ maintainers:
dependencies:
- name: common
repository: https://k8s-at-home.com/charts/
version: 1.6.1
version: 1.7.0

View File

@@ -0,0 +1,24 @@
# Patterns to ignore when building packages.
# This supports shell glob matching, relative path matching, and
# negation (prefixed with !). Only one pattern per line.
.DS_Store
# Common VCS dirs
.git/
.gitignore
.bzr/
.bzrignore
.hg/
.hgignore
.svn/
# Common backup files
*.swp
*.bak
*.tmp
*~
# Various IDEs
.project
.idea/
*.tmproj
.vscode
# OWNERS file for Kubernetes
OWNERS

View File

@@ -4,9 +4,13 @@ description: DashMachine is another web application bookmark dashboard, with fun
icon: https://github.com/rmountjoy92/DashMachine/raw/master/dashmachine/static/images/logo/logo.png
home: https://github.com/rmountjoy92/DashMachine
name: dashmachine
version: 1.0.0
version: 2.2.0
sources:
- https://github.com/rmountjoy92/DashMachine
maintainers:
- name: carpenike
email: ryan@ryanholt.net
dependencies:
- name: common
version: 1.7.0
repository: https://k8s-at-home.com/charts/

View File

@@ -0,0 +1,4 @@
approvers:
- carpenike
reviewers:
- carpenike

View File

@@ -1,31 +1,69 @@
dashmachine
===========
DashMachine is another web application bookmark dashboard, with fun features.
# Dashmachine
## Chart Values
This is a helm chart for [DashMachine](https://github.com/rmountjoy92/DashMachine).
| Key | Type | Default | Description |
|-----|------|---------|-------------|
| affinity | object | `{}` | |
| deploymentAnnotations | object | `{}` | |
| fullnameOverride | string | `""` | |
| image.pullPolicy | string | `"IfNotPresent"` | |
| image.repository | string | `"rmountjoy/dashmachine"` | |
| image.tag | string | `"latest"` | |
| ingress.annotations | object | `{}` | |
| ingress.enabled | bool | `false` | |
| ingress.hosts[0] | string | `"chart-example.local"` | |
| ingress.paths[0] | string | `"/"` | |
| ingress.tls | list | `[]` | |
| nameOverride | string | `""` | |
| nodeSelector | object | `{}` | |
| persistence.accessModes[0] | string | `"ReadWriteOnce"` | |
| persistence.enabled | bool | `false` | |
| persistence.size | string | `"1Gi"` | |
| persistence.storageClassName | string | `""` | |
| podAnnotations | object | `{}` | |
| replicaCount | int | `1` | |
| resources | object | `{}` | |
| service.port | int | `5000` | |
| service.type | string | `"ClusterIP"` | |
| tolerations | list | `[]` | |
**This chart is not maintained by the upstream project and any issues with the chart should be raised [here](https://github.com/k8s-at-home/charts/issues/new/choose)**
## TL;DR;
```shell
$ helm repo add k8s-at-home https://k8s-at-home.com/charts/
$ helm install k8s-at-home/dashmachine
```
## Installing the Chart
To install the chart with the release name `my-release`:
```console
helm install --name my-release k8s-at-home/dashmachine
```
## Uninstalling the Chart
To uninstall/delete the `my-release` deployment:
```console
helm delete my-release --purge
```
The command removes all the Kubernetes components associated with the chart and deletes the release.
## Configuration
Read through the charts [values.yaml](https://github.com/k8s-at-home/charts/blob/master/charts/dashmachine/values.yaml)
file. It has several commented out suggested values.
Additionally you can take a look at the common library [values.yaml](https://github.com/k8s-at-home/charts/blob/master/charts/common/values.yaml) for more (advanced) configuration options.
Specify each parameter using the `--set key=value[,key=value]` argument to `helm install`. For example,
```console
helm install dashmachine \
--set env.TZ="America/New_York" \
k8s-at-home/dashmachine
```
Alternatively, a YAML file that specifies the values for the above parameters can be provided while installing the
chart. For example,
```console
helm install dashmachine k8s-at-home/dashmachine --values values.yaml
```
```yaml
image:
tag: ...
```
---
**NOTE**
If you get
```console
Error: rendered manifests contain a resource that already exists. Unable to continue with install: existing resource conflict: ...`
```
it may be because you uninstalled the chart with `skipuninstall` enabled, you need to manually delete the pvc or use `existingClaim`.
---
## Upgrading an existing Release to a new major version
A major chart version change (like 4.0.1 -> 5.0.0) indicates that there is an incompatible breaking change potentially needing manual actions.
Refer to the library values.yaml for more configuration options.

View File

@@ -0,0 +1,2 @@
ingress:
enabled: true

View File

@@ -1,21 +0,0 @@
1. Get the application URL by running these commands:
{{- if .Values.ingress.enabled }}
{{- range $host := .Values.ingress.hosts }}
{{- range $.Values.ingress.paths }}
http{{ if $.Values.ingress.tls }}s{{ end }}://{{ $host }}{{ . }}
{{- end }}
{{- end }}
{{- else if contains "NodePort" .Values.service.type }}
export NODE_PORT=$(kubectl get --namespace {{ .Release.Namespace }} -o jsonpath="{.spec.ports[0].nodePort}" services {{ include "dashmachine.fullname" . }})
export NODE_IP=$(kubectl get nodes --namespace {{ .Release.Namespace }} -o jsonpath="{.items[0].status.addresses[0].address}")
echo http://$NODE_IP:$NODE_PORT
{{- else if contains "LoadBalancer" .Values.service.type }}
NOTE: It may take a few minutes for the LoadBalancer IP to be available.
You can watch the status of by running 'kubectl get svc -w {{ include "dashmachine.fullname" . }}'
export SERVICE_IP=$(kubectl get svc --namespace {{ .Release.Namespace }} {{ include "dashmachine.fullname" . }} -o jsonpath='{.status.loadBalancer.ingress[0].ip}')
echo http://$SERVICE_IP:{{ .Values.service.port }}
{{- else if contains "ClusterIP" .Values.service.type }}
export POD_NAME=$(kubectl get pods --namespace {{ .Release.Namespace }} -l "app.kubernetes.io/name={{ include "dashmachine.name" . }},app.kubernetes.io/instance={{ .Release.Name }}" -o jsonpath="{.items[0].metadata.name}")
echo "Visit http://127.0.0.1:8080 to use your application"
kubectl port-forward $POD_NAME 8080:80
{{- end }}

View File

@@ -1,32 +0,0 @@
{{/* vim: set filetype=mustache: */}}
{{/*
Expand the name of the chart.
*/}}
{{- define "dashmachine.name" -}}
{{- default .Chart.Name .Values.nameOverride | trunc 63 | trimSuffix "-" -}}
{{- end -}}
{{/*
Create a default fully qualified app name.
We truncate at 63 chars because some Kubernetes name fields are limited to this (by the DNS naming spec).
If release name contains chart name it will be used as a full name.
*/}}
{{- define "dashmachine.fullname" -}}
{{- if .Values.fullnameOverride -}}
{{- .Values.fullnameOverride | trunc 63 | trimSuffix "-" -}}
{{- else -}}
{{- $name := default .Chart.Name .Values.nameOverride -}}
{{- if contains $name .Release.Name -}}
{{- .Release.Name | trunc 63 | trimSuffix "-" -}}
{{- else -}}
{{- printf "%s-%s" .Release.Name $name | trunc 63 | trimSuffix "-" -}}
{{- end -}}
{{- end -}}
{{- end -}}
{{/*
Create chart name and version as used by the chart label.
*/}}
{{- define "dashmachine.chart" -}}
{{- printf "%s-%s" .Chart.Name .Chart.Version | replace "+" "_" | trunc 63 | trimSuffix "-" -}}
{{- end -}}

View File

@@ -0,0 +1 @@
{{ include "common.all" . }}

View File

@@ -1,78 +0,0 @@
apiVersion: apps/v1
kind: Deployment
metadata:
name: {{ include "dashmachine.fullname" . }}
{{- if .Values.deploymentAnnotations }}
annotations:
{{- range $key, $value := .Values.deploymentAnnotations }}
{{ $key }}: {{ $value | quote }}
{{- end }}
{{- end }}
labels:
app.kubernetes.io/name: {{ include "dashmachine.name" . }}
helm.sh/chart: {{ include "dashmachine.chart" . }}
app.kubernetes.io/instance: {{ .Release.Name }}
app.kubernetes.io/managed-by: {{ .Release.Service }}
spec:
replicas: {{ .Values.replicaCount }}
selector:
matchLabels:
app.kubernetes.io/name: {{ include "dashmachine.name" . }}
app.kubernetes.io/instance: {{ .Release.Name }}
template:
metadata:
labels:
app.kubernetes.io/name: {{ include "dashmachine.name" . }}
app.kubernetes.io/instance: {{ .Release.Name }}
{{- if .Values.podAnnotations }}
annotations:
{{- range $key, $value := .Values.podAnnotations }}
{{ $key }}: {{ $value | quote }}
{{- end }}
{{- end }}
spec:
{{- if .Values.dnsConfig }}
dnsConfig:
{{- toYaml .Values.dnsConfig | nindent 8 }}
{{- end }}
containers:
- name: {{ .Chart.Name }}
image: "{{ .Values.image.repository }}:{{ .Values.image.tag }}"
imagePullPolicy: {{ .Values.image.pullPolicy }}
ports:
- name: http
containerPort: 5000
protocol: TCP
# livenessProbe:
# httpGet:
# path: /notifications
# port: http
# readinessProbe:
# httpGet:
# path: /notifications
# port: http
resources:
{{- toYaml .Values.resources | nindent 12 }}
volumeMounts:
- name: config
mountPath: /dashmachine/dashmachine/user_data
volumes:
- name: config
{{- if .Values.persistence.enabled }}
persistentVolumeClaim:
claimName: {{ template "dashmachine.fullname" . }}
{{- else }}
emptyDir: {}
{{- end }}
{{- with .Values.nodeSelector }}
nodeSelector:
{{- toYaml . | nindent 8 }}
{{- end }}
{{- with .Values.affinity }}
affinity:
{{- toYaml . | nindent 8 }}
{{- end }}
{{- with .Values.tolerations }}
tolerations:
{{- toYaml . | nindent 8 }}
{{- end }}

View File

@@ -1,40 +0,0 @@
{{- if .Values.ingress.enabled -}}
{{- $fullName := include "dashmachine.fullname" . -}}
{{- $ingressPaths := .Values.ingress.paths -}}
apiVersion: extensions/v1beta1
kind: Ingress
metadata:
name: {{ $fullName }}
labels:
app.kubernetes.io/name: {{ include "dashmachine.name" . }}
helm.sh/chart: {{ include "dashmachine.chart" . }}
app.kubernetes.io/instance: {{ .Release.Name }}
app.kubernetes.io/managed-by: {{ .Release.Service }}
{{- with .Values.ingress.annotations }}
annotations:
{{- toYaml . | nindent 4 }}
{{- end }}
spec:
{{- if .Values.ingress.tls }}
tls:
{{- range .Values.ingress.tls }}
- hosts:
{{- range .hosts }}
- {{ . | quote }}
{{- end }}
secretName: {{ .secretName }}
{{- end }}
{{- end }}
rules:
{{- range .Values.ingress.hosts }}
- host: {{ . | quote }}
http:
paths:
{{- range $ingressPaths }}
- path: {{ . }}
backend:
serviceName: {{ $fullName }}
servicePort: http
{{- end }}
{{- end }}
{{- end }}

View File

@@ -1,24 +0,0 @@
{{- if and .Values.persistence.enabled (not .Values.persistence.existingClaim) }}
apiVersion: v1
kind: PersistentVolumeClaim
metadata:
name: {{ template "dashmachine.fullname" . }}
labels:
app.kubernetes.io/name: {{ include "dashmachine.name" . }}
helm.sh/chart: {{ include "dashmachine.chart" . }}
app.kubernetes.io/instance: {{ .Release.Name }}
app.kubernetes.io/managed-by: {{ .Release.Service }}
{{- with .Values.persistence.annotations }}
annotations:
{{ toYaml . | indent 4 }}
{{- end }}
spec:
accessModes:
{{- range .Values.persistence.accessModes }}
- {{ . | quote }}
{{- end }}
resources:
requests:
storage: {{ .Values.persistence.size | quote }}
storageClassName: {{ .Values.persistence.storageClass }}
{{- end -}}

View File

@@ -1,19 +0,0 @@
apiVersion: v1
kind: Service
metadata:
name: {{ include "dashmachine.fullname" . }}
labels:
app.kubernetes.io/name: {{ include "dashmachine.name" . }}
helm.sh/chart: {{ include "dashmachine.chart" . }}
app.kubernetes.io/instance: {{ .Release.Name }}
app.kubernetes.io/managed-by: {{ .Release.Service }}
spec:
type: {{ .Values.service.type }}
ports:
- port: {{ .Values.service.port }}
targetPort: http
protocol: TCP
name: http
selector:
app.kubernetes.io/name: {{ include "dashmachine.name" . }}
app.kubernetes.io/instance: {{ .Release.Name }}

View File

@@ -1,18 +0,0 @@
apiVersion: v1
kind: Pod
metadata:
name: "{{ include "dashmachine.fullname" . }}-test-connection"
labels:
app.kubernetes.io/name: {{ include "dashmachine.name" . }}
helm.sh/chart: {{ include "dashmachine.chart" . }}
app.kubernetes.io/instance: {{ .Release.Name }}
app.kubernetes.io/managed-by: {{ .Release.Service }}
annotations:
"helm.sh/hook": test-success
spec:
containers:
- name: wget
image: busybox
command: ['wget']
args: ['{{ include "dashmachine.fullname" . }}:{{ .Values.service.port }}']
restartPolicy: Never

View File

@@ -1,65 +1,22 @@
# Default values for dashmachine.
# This is a YAML-formatted file.
# Declare variables to be passed into your templates.
replicaCount: 1
image:
repository: rmountjoy/dashmachine
tag: v0.5-4
pullPolicy: IfNotPresent
nameOverride: ""
fullnameOverride: ""
strategy:
type: Recreate
service:
type: ClusterIP
port: 5000
port:
port: 5000
ingress:
enabled: false
annotations: {}
# kubernetes.io/ingress.class: nginx
# kubernetes.io/tls-acme: "true"
paths: ["/"]
hosts:
- chart-example.local
tls: []
# - secretName: chart-example-tls
# hosts:
# - chart-example.local
resources: {}
# We usually recommend not to specify default resources and to leave this as a conscious
# choice for the user. This also increases chances charts run on environments with little
# resources, such as Minikube. If you do want to specify resources, uncomment the following
# lines, adjust them as necessary, and remove the curly braces after 'resources:'.
# limits:
# cpu: 100m
# memory: 128Mi
# requests:
# cpu: 100m
# memory: 128Mi
nodeSelector: {}
tolerations: []
affinity: {}
env: {}
# TZ: UTC
persistence:
enabled: false
## If defined, storageClassName: <storageClass>
## If set to "-", storageClassName: "", which disables dynamic provisioning
## If undefined (the default) or set to null, no storageClassName spec is
## set, choosing the default provisioner. (gp2 on AWS, standard on
## GKE, AWS & OpenStack)
##
storageClass: ""
accessModes:
- ReadWriteOnce
size: 1Gi
podAnnotations: {}
deploymentAnnotations: {}
config:
enabled: false
emptyDir: false
mountPath: /dashmachine/dashmachine/user_data

View File

@@ -2,7 +2,7 @@ apiVersion: v2
appVersion: 3.9.1
description: Perl client used to update dynamic DNS entries for accounts on Dynamic DNS Network Service Providers
name: ddclient
version: 1.1.1
version: 1.3.0
keywords:
- ddclient
- dns
@@ -17,4 +17,4 @@ maintainers:
dependencies:
- name: common
repository: https://k8s-at-home.com/charts/
version: 1.6.1
version: 1.7.0

View File

@@ -19,5 +19,6 @@
.project
.idea/
*.tmproj
.vscode/
# OWNERS file for Kubernetes
OWNERS

View File

@@ -1,14 +1,19 @@
apiVersion: v2
appVersion: 1.15.3
description: ESPHome
description: ESPHome is a system to control your ESP8266/ESP32 by simple yet powerful configuration files and control them remotely through Home Automation systems.
name: esphome
version: 2.3.0
version: 3.0.0
keywords:
- esphome
- esphome
- home-automation
home: https://github.com/k8s-at-home/charts/tree/master/charts/esphome
icon: https://esphome.io/_images/logo-text.svg
sources:
- https://github.com/esphome/esphome
maintainers:
- name: billimek
email: jeff@billimek.com
- name: billimek
email: jeff@billimek.com
dependencies:
- name: common
repository: https://k8s-at-home.com/charts/
version: 1.7.0

View File

@@ -1,4 +1,8 @@
approvers:
- billimek
- onedr0p
- bjw-s
reviewers:
- billimek
- onedr0p
- bjw-s

View File

@@ -1,120 +1,78 @@
# ESPHome
# esphome
This is a helm chart for [ESPHome](https://esphome.io)
This is a helm chart for [esphome](https://github.com/esphome/esphome).
**This chart is not maintained by the upstream project and any issues with the chart should be raised [here](https://github.com/k8s-at-home/charts/issues/new/choose)**
## TL;DR;
```shell
helm repo add k8s-at-home https://k8s-at-home.com/charts/
helm install k8s-at-home/esphome
$ helm repo add k8s-at-home https://k8s-at-home.com/charts/
$ helm install k8s-at-home/esphome
```
## Introduction
This code is adapted for [the official esphome docker image](https://hub.docker.com/r/esphome/esphome/)
## Installing the Chart
To install the chart with the release name `my-release`:
```shell
```console
helm install --name my-release k8s-at-home/esphome
```
## Uninstalling the Chart
To uninstall/delete the `my-release` deployment:
```shell
```console
helm delete my-release --purge
```
The command removes all the Kubernetes components associated with the chart and deletes the release.
## Configuration
The following tables lists the configurable parameters of the ESPHome chart and their default values.
| Parameter | Description | Default |
|----------------------------------------|--------------------------------------------------------------------------------------------------------------------|-----------------------|
| `image.repository` | Image repository | `esphome/esphome` |
| `image.tag` | Image tag. Possible values listed [here](https://hub.docker.com/r/esphome/esphome/tags/). | `0.14.5` |
| `image.pullPolicy` | Image pull policy | `IfNotPresent` |
| `image.pullSecrets` | Secrets to use when pulling the image | `[]` |
| `strategyType` | Specifies the strategy used to replace old Pods by new ones | `Recreate` |
| `probes.liveness.enabled` | Use the livenessProbe? | `true` |
| `probes.liveness.scheme ` | Specify liveness `scheme` parameter for the deployment | `HTTP` |
| `probes.liveness.initialDelaySeconds` | Specify liveness `initialDelaySeconds` parameter for the deployment | `60` |
| `probes.liveness.failureThreshold` | Specify liveness `failureThreshold` parameter for the deployment | `5` |
| `probes.liveness.timeoutSeconds` | Specify liveness `timeoutSeconds` parameter for the deployment | `10` |
| `probes.readiness.enabled` | Use the readinessProbe? | `true` |
| `probes.readiness.scheme ` | Specify readiness `scheme` parameter for the deployment | `HTTP` |
| `probes.readiness.initialDelaySeconds` | Specify readiness `initialDelaySeconds` parameter for the deployment | `60` |
| `probes.readiness.failureThreshold` | Specify readiness `failureThreshold` parameter for the deployment | `5` |
| `probes.readiness.timeoutSeconds` | Specify readiness `timeoutSeconds` parameter for the deployment | `10` |
| `probes.startup.enabled` | Use the startupProbe? (new in kubernetes 1.16) | `false` |
| `probes.startup.scheme ` | Specify startup `scheme` parameter for the deployment | `HTTP` |
| `probes.startup.failureThreshold` | Specify startup `failureThreshold` parameter for the deployment | `5` |
| `probes.startup.periodSeconds` | Specify startup `periodSeconds` parameter for the deployment | `10` |
| `service.type` | Kubernetes service type for the esphome GUI | `ClusterIP` |
| `service.port` | Kubernetes port where the esphome GUI is exposed | `6052` |
| `service.portName` | Kubernetes port name where the esphome GUI is exposed | `api` |
| `service.additionalPorts` | Add additional ports exposed by the esphome container integrations. Example homematic needs to expose a proxy port | `{}` |
| `service.annotations` | Service annotations for the esphome GUI | `{}` |
| `service.clusterIP` | Cluster IP for the esphom GUI | `` |
| `service.externalIPs` | External IPs for the esphome GUI | `[]` |
| `service.loadBalancerIP` | Loadbalancer IP for the esphome GUI | `` |
| `service.loadBalancerSourceRanges` | Loadbalancer client IP restriction range for the esphome GUI | `[]` |
| `service.publishNotReadyAddresses` | Set to true if the notReadyAddresses should be published | `false` |
| `service.externalTrafficPolicy` | Loadbalancer externalTrafficPolicy | `` |
| `hostNetwork` | Enable hostNetwork - might be needed for discovery to work | `false` |
| `service.nodePort` | nodePort to listen on for the esphome GUI | `` |
| `ingress.enabled` | Enables Ingress | `false` |
| `ingress.annotations` | Ingress annotations | `{}` |
| `ingress.path` | Ingress path | `/` |
| `ingress.hosts` | Ingress accepted hostnames | `chart-example.local` |
| `ingress.tls` | Ingress TLS configuration | `[]` |
| `persistence.enabled` | Use persistent volume to store data | `true` |
| `persistence.size` | Size of persistent volume claim | `5Gi` |
| `persistence.existingClaim` | Use an existing PVC to persist data | `nil` |
| `persistence.hostPath` | The path to the config directory on the host, instead of a PVC | `nil` |
| `persistence.storageClass` | Type of persistent volume claim | `-` |
| `persistence.accessMode` | Persistence access modes | `ReadWriteMany` |
| `git.enabled` | Use git-sync in init container | `false` |
| `git.secret` | Git secret to use for git-sync | `git-creds` |
| `git.syncPath` | Git sync path | `/config` |
| `git.keyPath` | Git ssh key path | `/root/.ssh` |
| `git.user.name` | Human-readable name in the “committer” and “author” fields | `` |
| `git.user.email` | Email address for the “committer” and “author” fields | `` |
| `hostMounts` | Array of host directories to mount; can be used for devices | [] |
| `hostMounts.name` | Name of the volume | `nil` |
| `hostMounts.hostPath` | The path on the host machine | `nil` |
| `hostMounts.mountPath` | The path at which to mount (optional; assumed same as hostPath) | `nil` |
| `hostMounts.type` | The type to mount (optional, i.e., `Directory`) | `nil` |
| `extraEnv` | Extra ENV vars to pass to the esphome container | `{}` |
| `extraEnvSecrets` | Extra env vars to pass to the esphome container from k8s secrets - see `values.yaml` for an example | `{}` |
| `resources` | CPU/Memory resource requests/limits or the esphome GUI | `{}` |
| `nodeSelector` | Node labels for pod assignment or the esphome GUI | `{}` |
| `tolerations` | Toleration labels for pod assignment or the esphome GUI | `[]` |
| `affinity` | Affinity settings for pod assignment or the esphome GUI | `{}` |
| `podAnnotations` | Key-value pairs to add as pod annotations | `{}` |
| `extraVolumes` | Any extra volumes to define for the pod | `{}` |
| `extraVolumeMounts` | Any extra volumes mounts to define for each container of the pod | `{}` |
Read through the charts [values.yaml](https://github.com/k8s-at-home/charts/blob/master/charts/esphome/values.yaml)
file. It has several commented out suggested values.
Additionally you can take a look at the common library [values.yaml](https://github.com/k8s-at-home/charts/blob/master/charts/common/values.yaml) for more (advanced) configuration options.
Specify each parameter using the `--set key=value[,key=value]` argument to `helm install`. For example,
```shell
helm install --name my-release \
--set image.tag=latest \
```console
helm install esphome \
--set env.TZ="America/New_York" \
k8s-at-home/esphome
```
Alternatively, a YAML file that specifies the values for the above parameters can be provided while installing the chart. For example,
```shell
helm install --name my-release -f values.yaml k8s-at-home/esphome
Alternatively, a YAML file that specifies the values for the above parameters can be provided while installing the
chart. For example,
```console
helm install esphome k8s-at-home/esphome --values values.yaml
```
Read through the [values.yaml](values.yaml) file. It has several commented out suggested values.
```yaml
image:
tag: ...
```
---
**NOTE**
If you get
```console
Error: rendered manifests contain a resource that already exists. Unable to continue with install: existing resource conflict: ...`
```
it may be because you uninstalled the chart with `skipuninstall` enabled, you need to manually delete the pvc or use `existingClaim`.
---
## Upgrading an existing Release to a new major version
A major chart version change (like 4.0.1 -> 5.0.0) indicates that there is an incompatible breaking change potentially needing manual actions.
### Upgrading from 2.x.x to 3.x.x
Due to migrating to a centralized common library some values in `values.yaml` have changed.
Examples:
* `service.port` has been moved to `service.port.port`.
* `persistence.type` has been moved to `controllerType`.
Refer to the library values.yaml for more configuration options.

View File

@@ -0,0 +1,4 @@
env:
ESPHOME_DASHBOARD_USE_PING: true
ingress:
enabled: true

View File

@@ -1,19 +1 @@
1. Get the application URL by running these commands:
{{- if .Values.ingress.enabled }}
{{- range .Values.ingress.hosts }}
http{{ if $.Values.ingress.tls }}s{{ end }}://{{ . }}{{ $.Values.ingress.path }}
{{- end }}
{{- else if contains "NodePort" .Values.service.type }}
export NODE_PORT=$(kubectl get --namespace {{ .Release.Namespace }} -o jsonpath="{.spec.ports[0].nodePort}" services {{ template "esphome.fullname" . }})
export NODE_IP=$(kubectl get nodes --namespace {{ .Release.Namespace }} -o jsonpath="{.items[0].status.addresses[0].address}")
echo http://$NODE_IP:$NODE_PORT
{{- else if contains "LoadBalancer" .Values.service.type }}
NOTE: It may take a few minutes for the LoadBalancer IP to be available.
You can watch the status of by running 'kubectl get svc -w {{ template "esphome.fullname" . }}'
export SERVICE_IP=$(kubectl get svc --namespace {{ .Release.Namespace }} {{ template "esphome.fullname" . }} -o jsonpath='{.status.loadBalancer.ingress[0].ip}')
echo http://$SERVICE_IP:{{ .Values.service.port }}
{{- else if contains "ClusterIP" .Values.service.type }}
export POD_NAME=$(kubectl get pods --namespace {{ .Release.Namespace }} -l "app={{ template "esphome.name" . }},release={{ .Release.Name }}" -o jsonpath="{.items[0].metadata.name}")
echo "Visit http://127.0.0.1:8080 to use your application"
kubectl port-forward $POD_NAME 8080:80
{{- end }}
{{- include "common.notes.defaultNotes" . -}}

View File

@@ -1,32 +0,0 @@
{{/* vim: set filetype=mustache: */}}
{{/*
Expand the name of the chart.
*/}}
{{- define "esphome.name" -}}
{{- default .Chart.Name .Values.nameOverride | trunc 63 | trimSuffix "-" -}}
{{- end -}}
{{/*
Create a default fully qualified app name.
We truncate at 63 chars because some Kubernetes name fields are limited to this (by the DNS naming spec).
If release name contains chart name it will be used as a full name.
*/}}
{{- define "esphome.fullname" -}}
{{- if .Values.fullnameOverride -}}
{{- .Values.fullnameOverride | trunc 63 | trimSuffix "-" -}}
{{- else -}}
{{- $name := default .Chart.Name .Values.nameOverride -}}
{{- if contains $name .Release.Name -}}
{{- .Release.Name | trunc 63 | trimSuffix "-" -}}
{{- else -}}
{{- printf "%s-%s" .Release.Name $name | trunc 63 | trimSuffix "-" -}}
{{- end -}}
{{- end -}}
{{- end -}}
{{/*
Create chart name and version as used by the chart label.
*/}}
{{- define "esphome.chart" -}}
{{- printf "%s-%s" .Chart.Name .Chart.Version | replace "+" "_" | trunc 63 | trimSuffix "-" -}}
{{- end -}}

View File

@@ -0,0 +1 @@
{{ include "common.all" . }}

View File

@@ -1,216 +0,0 @@
apiVersion: apps/v1
kind: Deployment
metadata:
name: {{ template "esphome.fullname" . }}
labels:
app.kubernetes.io/name: {{ include "esphome.name" . }}
helm.sh/chart: {{ include "esphome.chart" . }}
app.kubernetes.io/instance: {{ .Release.Name }}
app.kubernetes.io/managed-by: {{ .Release.Service }}
spec:
replicas: 1
strategy:
type: {{ .Values.strategyType }}
selector:
matchLabels:
app.kubernetes.io/name: {{ include "esphome.name" . }}
app.kubernetes.io/instance: {{ .Release.Name }}
template:
metadata:
labels:
app.kubernetes.io/name: {{ include "esphome.name" . }}
app.kubernetes.io/instance: {{ .Release.Name }}
{{- if .Values.podAnnotations }}
annotations:
{{- range $key, $value := .Values.podAnnotations }}
{{ $key }}: {{ $value | quote }}
{{- end }}
{{- end }}
spec:
{{- with .Values.image.pullSecrets }}
imagePullSecrets:
{{- range . }}
- name: {{ . }}
{{- end }}
{{- end }}
{{- if .Values.hostNetwork }}
hostNetwork: {{ .Values.hostNetwork }}
dnsPolicy: ClusterFirstWithHostNet
{{- end }}
initContainers:
{{- if .Values.git.enabled }}
- name: git-sync
image: "{{ .Values.git.image.repository }}:{{ .Values.git.image.tag }}"
imagePullPolicy: {{ .Values.git.image.pullPolicy }}
{{- if .Values.git.command }}
command:
{{- range .Values.git.command }}
- {{ . | quote }}
{{- end }}
{{- else }}
command: ["/bin/sh", "-c"]
args:
- set -e;
if [ -d "{{ .Values.git.syncPath }}/.git" ];
then
git -C "{{ .Values.git.syncPath }}" pull || true;
else
if [ "$(ls -A {{ .Values.git.syncPath }})" ];
then
git clone --depth 2 "{{ .Values.git.repo }}" /tmp/repo;
cp -rf /tmp/repo/.git "{{ .Values.git.syncPath }}";
cd "{{ .Values.git.syncPath }}";
git checkout -f;
else
git clone --depth 2 "{{ .Values.git.repo }}" "{{ .Values.git.syncPath }}";
fi;
fi;
if [ -f "{{ .Values.git.keyPath }}/git-crypt-key" ];
then
cd {{ .Values.git.syncPath }};
git-crypt unlock "{{ .Values.git.keyPath }}/git-crypt-key";
fi;
{{- end }}
volumeMounts:
- mountPath: /config
name: config
- mountPath: {{ .Values.git.keyPath }}
name: git-secret
{{- if .Values.extraVolumeMounts }}{{ toYaml .Values.extraVolumeMounts | trim | nindent 8 }}{{ end }}
{{- if .Values.usePodSecurityContext }}
securityContext:
runAsUser: {{ default 0 .Values.runAsUser }}
{{- if and (.Values.runAsUser) (.Values.fsGroup) }}
{{- if not (eq .Values.runAsUser 0.0) }}
fsGroup: {{ .Values.fsGroup }}
{{- end }}
{{- end }}
{{- end }}
{{- end }}
containers:
- name: {{ .Chart.Name }}
image: "{{ .Values.image.repository }}:{{ .Values.image.tag }}"
imagePullPolicy: {{ .Values.image.pullPolicy }}
ports:
- name: api
containerPort: {{ .Values.service.port }}
protocol: TCP
{{- range .Values.service.additionalPorts }}
- name: {{ .name }}
containerPort: {{ .targetPort }}
{{- end }}
{{- if .Values.probes.liveness.enabled }}
livenessProbe:
httpGet:
path: /ping
port: api
scheme: {{ .Values.probes.liveness.scheme }}
initialDelaySeconds: {{ .Values.probes.liveness.initialDelaySeconds }}
failureThreshold: {{ .Values.probes.liveness.failureThreshold }}
timeoutSeconds: {{ .Values.probes.liveness.timeoutSeconds }}
{{- end }}
{{- if .Values.probes.readiness.enabled }}
readinessProbe:
httpGet:
path: /ping
port: api
scheme: {{ .Values.probes.readiness.scheme }}
initialDelaySeconds: {{ .Values.probes.readiness.initialDelaySeconds }}
failureThreshold: {{ .Values.probes.readiness.failureThreshold }}
timeoutSeconds: {{ .Values.probes.readiness.timeoutSeconds }}
{{- end }}
{{- if .Values.probes.startup.enabled }}
startupProbe:
httpGet:
path: /ping
port: api
scheme: {{ .Values.probes.startup.scheme }}
failureThreshold: {{ .Values.probes.startup.failureThreshold }}
periodSeconds: {{ .Values.probes.startup.periodSeconds }}
{{- end }}
env:
{{- range $key, $value := .Values.extraEnv }}
- name: {{ $key | quote }}
value: {{ $value | quote }}
{{- end }}
{{- range $name, $opts := .Values.extraEnvSecrets }}
- name: {{ $name }}
valueFrom:
secretKeyRef:
name: {{ $opts.secret }}
key: {{ $opts.key }}
{{- end }}
envFrom:
{{- range .Values.extraSecretForEnvFrom }}
- secretRef:
name: {{ . }}
{{- end }}
volumeMounts:
- mountPath: /config
name: config
{{- range .Values.hostMounts }}
{{- if .mountPath }}
- mountPath: {{ .mountPath }}
{{- else }}
- mountPath: {{ .hostPath }}
{{- end }}
name: {{ .name }}
{{- end }}
{{- if .Values.extraVolumeMounts }}{{ toYaml .Values.extraVolumeMounts | trim | nindent 10 }}{{ end }}
{{- if .Values.usePodSecurityContext }}
securityContext:
runAsUser: {{ default 0 .Values.runAsUser }}
{{- if and (.Values.runAsUser) (.Values.fsGroup) }}
{{- if not (eq .Values.runAsUser 0.0) }}
fsGroup: {{ .Values.fsGroup }}
{{- end }}
{{- end }}
{{- end }}
resources:
{{ toYaml .Values.resources | indent 12 }}
{{- if .Values.hostMounts }}
securityContext:
privileged: true
{{- end }}
volumes:
- name: config
{{- if .Values.persistence.enabled }}
{{- if .Values.persistence.hostPath }}
hostPath:
path: {{.Values.persistence.hostPath}}
type: Directory
{{- else }}
persistentVolumeClaim:
claimName: {{ if .Values.persistence.existingClaim }}{{ .Values.persistence.existingClaim }}{{- else }}{{ template "esphome.fullname" . }}{{- end }}
{{- end }}
{{- else }}
emptyDir: {}
{{ end }}
{{- range .Values.hostMounts }}
- name: {{ .name }}
hostPath:
path: {{.hostPath}}
{{- if .type }}
type: {{ .type }}
{{- end }}
{{- end }}
{{- if .Values.git.enabled }}
- name: git-secret
secret:
defaultMode: 256
secretName: {{ .Values.git.secret }}
optional: true
{{ end }}
{{- if .Values.extraVolumes }}{{ toYaml .Values.extraVolumes | trim | nindent 6 }}{{ end }}
{{- with .Values.nodeSelector }}
nodeSelector:
{{ toYaml . | indent 8 }}
{{- end }}
{{- with .Values.affinity }}
affinity:
{{ toYaml . | indent 8 }}
{{- end }}
{{- with .Values.tolerations }}
tolerations:
{{ toYaml . | indent 8 }}
{{- end }}

View File

@@ -1,39 +0,0 @@
{{- if .Values.ingress.enabled -}}
{{- $fullName := include "esphome.fullname" . -}}
{{- $servicePort := .Values.service.port -}}
{{- $ingressPath := .Values.ingress.path -}}
apiVersion: extensions/v1beta1
kind: Ingress
metadata:
name: {{ $fullName }}
labels:
app.kubernetes.io/name: {{ include "esphome.name" . }}
helm.sh/chart: {{ include "esphome.chart" . }}
app.kubernetes.io/instance: {{ .Release.Name }}
app.kubernetes.io/managed-by: {{ .Release.Service }}
{{- with .Values.ingress.annotations }}
annotations:
{{ toYaml . | indent 4 }}
{{- end }}
spec:
{{- if .Values.ingress.tls }}
tls:
{{- range .Values.ingress.tls }}
- hosts:
{{- range .hosts }}
- {{ . | quote }}
{{- end }}
secretName: {{ .secretName }}
{{- end }}
{{- end }}
rules:
{{- range .Values.ingress.hosts }}
- host: {{ . }}
http:
paths:
- path: {{ $ingressPath }}
backend:
serviceName: {{ $fullName }}
servicePort: {{ $servicePort }}
{{- end }}
{{- end }}

View File

@@ -1,26 +0,0 @@
{{- if .Values.persistence.enabled -}}
{{- if not .Values.persistence.existingClaim -}}
kind: PersistentVolumeClaim
apiVersion: v1
metadata:
name: {{ template "esphome.fullname" . }}
labels:
app.kubernetes.io/name: {{ include "esphome.name" . }}
helm.sh/chart: {{ include "esphome.chart" . }}
app.kubernetes.io/instance: {{ .Release.Name }}
app.kubernetes.io/managed-by: {{ .Release.Service }}
spec:
accessModes:
- {{ .Values.persistence.accessMode | quote }}
resources:
requests:
storage: {{ .Values.persistence.size | quote }}
{{- if .Values.persistence.storageClass }}
{{- if (eq "-" .Values.persistence.storageClass) }}
storageClassName: ""
{{- else }}
storageClassName: "{{ .Values.persistence.storageClass }}"
{{- end }}
{{- end }}
{{- end -}}
{{- end -}}

View File

@@ -1,56 +0,0 @@
apiVersion: v1
kind: Service
metadata:
name: {{ template "esphome.fullname" . }}
labels:
app.kubernetes.io/name: {{ include "esphome.name" . }}
helm.sh/chart: {{ include "esphome.chart" . }}
app.kubernetes.io/instance: {{ .Release.Name }}
app.kubernetes.io/managed-by: {{ .Release.Service }}
{{- if .Values.service.labels }}
{{ toYaml .Values.service.labels | indent 4 }}
{{- end }}
{{- with .Values.service.annotations }}
annotations:
{{ toYaml . | indent 4 }}
{{- end }}
spec:
{{- if (or (eq .Values.service.type "ClusterIP") (empty .Values.service.type)) }}
type: ClusterIP
{{- if .Values.service.clusterIP }}
clusterIP: {{ .Values.service.clusterIP }}
{{end}}
{{- else if eq .Values.service.type "LoadBalancer" }}
type: {{ .Values.service.type }}
{{- if .Values.service.loadBalancerIP }}
loadBalancerIP: {{ .Values.service.loadBalancerIP }}
{{- end }}
{{- if .Values.service.externalTrafficPolicy }}
externalTrafficPolicy: {{ .Values.service.externalTrafficPolicy }}
{{- end }}
{{- if .Values.service.loadBalancerSourceRanges }}
loadBalancerSourceRanges:
{{ toYaml .Values.service.loadBalancerSourceRanges | indent 4 }}
{{- end -}}
{{- else }}
type: {{ .Values.service.type }}
{{- end }}
{{- if .Values.service.externalIPs }}
externalIPs:
{{ toYaml .Values.service.externalIPs | indent 4 }}
{{- end }}
publishNotReadyAddresses: {{ .Values.service.publishNotReadyAddresses }}
ports:
- name: {{ .Values.service.portName }}
port: {{ .Values.service.port }}
protocol: TCP
targetPort: 6052
{{ if (and (eq .Values.service.type "NodePort") (not (empty .Values.service.nodePort))) }}
nodePort: {{.Values.service.nodePort}}
{{ end }}
{{- if .Values.service.additionalPorts }}
{{- .Values.service.additionalPorts | toYaml | indent 4 }}
{{- end }}
selector:
app.kubernetes.io/name: {{ include "esphome.name" . }}
app.kubernetes.io/instance: {{ .Release.Name }}

View File

@@ -1,176 +1,39 @@
# Default values for esphome.
# This is a YAML-formatted file.
# Declare variables to be passed into your templates.
# Default values for Bazarr.
image:
repository: esphome/esphome
tag: 1.15.3
pullPolicy: IfNotPresent
pullSecrets: []
tag: 1.15.3
# upgrade strategy type (e.g. Recreate or RollingUpdate)
strategyType: Recreate
strategy:
type: Recreate
# Probes configuration
probes:
liveness:
enabled: true
scheme: HTTP
initialDelaySeconds: 60
failureThreshold: 5
timeoutSeconds: 10
readiness:
enabled: true
scheme: HTTP
initialDelaySeconds: 60
failureThreshold: 5
timeoutSeconds: 10
startup:
enabled: false
scheme: HTTP
failureThreshold: 30
periodSeconds: 10
service:
type: ClusterIP
port: 6052
portName: api
additionalPorts: []
# - name: homematicproxy
# port: 2001
# targetPort: 2001
annotations: {}
labels: {}
clusterIP: ""
## List of IP addresses at which the hass-configurator service is available
## Ref: https://kubernetes.io/docs/user-guide/services/#external-ips
##
externalIPs: []
loadBalancerIP: ""
loadBalancerSourceRanges: []
# nodePort: 30000
publishNotReadyAddresses: false
port:
port: 6052
ingress:
enabled: false
annotations: {}
# kubernetes.io/ingress.class: nginx
# kubernetes.io/tls-acme: "true"
path: /
hosts:
- esphome.local
tls: []
# - secretName: esphome-tls
# hosts:
# - esphome.local
hostNetwork: false
env: {}
# ESPHOME_DASHBOARD_USE_PING: true
# ESPHOME_DASHBOARD_RELATIVE_URL: "/"
# ESPHOME_QUICKWIZARD:
# ESPHOME_IS_HASSIO:
# DISABLE_HA_AUTHENTICATION:
# USERNAME:
# PASSWORD:
persistence:
enabled: true
## esphome data Persistent Volume Storage Class
## If defined, storageClassName: <storageClass>
## If set to "-", storageClassName: "", which disables dynamic provisioning
## If undefined (the default) or set to null, no storageClassName spec is
## set, choosing the default provisioner. (gp2 on AWS, standard on
## GKE, AWS & OpenStack)
##
# storageClass: "-"
##
## If you want to reuse an existing claim, you can pass the name of the PVC using
## the existingClaim variable
# existingClaim: your-claim
##
## If you want to use a volume on the host machine instead of a PVC:
# hostPath: /path/to/the/config/folder
accessMode: ReadWriteOnce
size: 1Gi
## Additional hass container environment variable
## For instance to add a http_proxy
##
extraEnv: {}
## Additional hass container environment variable from k8s secrets
## For instance to add a password
## can use `!env_var` in the esphome configuration to reference these variables
extraEnvSecrets:
# Example
# This will set ${MQTT_PASSWORD} to the 'password' key from the 'mqtt' secret
# MQTT_PASSWORD:
# secret: mqtt
# key: password
## If you'd like to provide your own Kubernetes Secret object instead of passing your values
## individually, pass in the name of a created + populated Secret.
## All secrets will be mounted as environment variables, with each key/value mapping to a
## corresponding environment variable.
##
extraSecretForEnvFrom: []
# - esphome-secrets
# Enable pod security context (must be `true` if runAsUser or fsGroup are set)
usePodSecurityContext: true
# Set runAsUser to 1000 to let esphome run as non-root user.
# When setting runAsUser to a different value than 0 also set fsGroup to the same value:
# runAsUser: <defaults to 0>
# fsGroup: <will be omitted in deployment if runAsUser is 0>
# Mount devices or folders from the host machine. Can be used for USB device mounting.
hostMounts: []
# Example
# - name: zha
# hostPath: /dev/serial/by-id/usb-Silicon_Labs_HubZ_Smart_Home_Controller_6120245D-if01-port0
resources: {}
# We usually recommend not to specify default resources and to leave this as a conscious
# choice for the user. This also increases chances charts run on environments with little
# resources, such as Minikube. If you do want to specify resources, uncomment the following
# lines, adjust them as necessary, and remove the curly braces after 'resources:'.
# limits:
# cpu: 100m
# memory: 128Mi
# requests:
# cpu: 100m
# memory: 128Mi
nodeSelector: {}
tolerations: []
affinity: {}
podAnnotations: {}
# Any extra volumes to define for the pod
extraVolumes: []
# - name: example-name
# hostPath:
# path: /path/on/host
# type: DirectoryOrCreate
# Any extra volume mounts to define for the containers
extraVolumeMounts: []
# - name: example-name
# mountPath: /path/in/container
git:
enabled: false
image:
repository: k8sathome/git-crypt
tag: 2020.09.28
pullPolicy: IfNotPresent
## Specify the command that runs in the git-sync container to pull in configuration.
# command: []
# Committer settings
user:
name: ""
email: ""
repo: ""
secret: git-creds
syncPath: /config
keyPath: /root/.ssh
config:
enabled: false
emptyDir: false
## Persistent Volume Storage Class
## If defined, storageClassName: <storageClass>
## If set to "-", storageClassName: "", which disables dynamic provisioning
## If undefined (the default) or set to null, no storageClassName spec is
## set, choosing the default provisioner. (gp2 on AWS, standard on
## GKE, AWS & OpenStack)
# storageClass: "-"
# accessMode: ReadWriteOnce
# size: 1Gi
## Do not delete the pvc upon helm uninstall
# skipuninstall: false
# existingClaim: ""

View File

@@ -2,7 +2,7 @@ apiVersion: v2
appVersion: 4.1.1
description: Flood is a monitoring service for various torrent clients
name: flood
version: 1.1.1
version: 1.3.0
keywords:
- flood
- rtorrent
@@ -19,4 +19,4 @@ maintainers:
dependencies:
- name: common
repository: https://k8s-at-home.com/charts/
version: 1.6.1
version: 1.7.0

View File

@@ -2,7 +2,7 @@ apiVersion: v2
appVersion: 1.17.0
description: FreshRSS is a self-hosted RSS feed aggregator
name: freshrss
version: 1.0.0
version: 1.2.0
keywords:
- freshrss
- rss
@@ -17,4 +17,4 @@ maintainers:
dependencies:
- name: common
repository: https://k8s-at-home.com/charts/
version: 1.6.1
version: 1.7.0

View File

@@ -19,5 +19,6 @@
.project
.idea/
*.tmproj
.vscode
# OWNERS file for Kubernetes
OWNERS

View File

@@ -1,8 +1,8 @@
apiVersion: v2
appVersion: v2.6.1
appVersion: v2.7.1
description: ERP beyond your fridge - grocy is a web-based self-hosted groceries & household management solution for your home
name: grocy
version: 2.0.0
version: 3.0.0
keywords:
- grocy
home: https://github.com/k8s-at-home/charts/tree/master/charts/grocy
@@ -10,5 +10,9 @@ icon: https://github.com/grocy/grocy/raw/master/public/img/appicons/mstile-150x1
sources:
- https://github.com/grocy/grocy
maintainers:
- name: carpenike
email: ryan@ryanholt.net
- name: billimek
email: jeff@billimek.com
dependencies:
- name: common
version: 1.7.0
repository: https://k8s-at-home.com/charts/

View File

@@ -1,4 +1,8 @@
approvers:
- billimek
- onedr0p
- bjw-s
reviewers:
- billimek
- billimek
- onedr0p
- bjw-s

View File

@@ -1,6 +1,8 @@
# Grocy
# grocy
This is a helm chart for [Grocy](https://grocy.info/) leveraging the [Linuxserver.io image](https://hub.docker.com/r/linuxserver/grocy/)
This is a helm chart for [grocy](https://grocy.info/).
**This chart is not maintained by the upstream project and any issues with the chart should be raised [here](https://github.com/k8s-at-home/charts/issues/new/choose)**
## TL;DR;
@@ -28,71 +30,40 @@ helm delete my-release --purge
The command removes all the Kubernetes components associated with the chart and deletes the release.
## Configuration
The following tables lists the configurable parameters of the Sentry chart and their default values.
| Parameter | Description | Default |
|----------------------------|-------------------------------------|---------------------------------------------------------|
| `image.repository` | Image repository | `linuxserver/grocy` |
| `image.tag` | Image tag. Possible values listed [here](https://hub.docker.com/r/linuxserver/grocy/tags/).| `v2.6.1-ls51`|
| `image.pullPolicy` | Image pull policy | `IfNotPresent` |
| `strategyType` | Specifies the strategy used to replace old Pods by new ones | `Recreate` |
| `timezone` | Timezone the grocy instance should run as, e.g. 'America/New_York' | `UTC` |
| `puid` | process userID the grocy instance should run as | `1000` |
| `pgid` | process groupID the grocy instance should run as | `1000` |
| `probes.liveness.initialDelaySeconds` | Specify liveness `initialDelaySeconds` parameter for the deployment | `60` |
| `probes.liveness.failureThreshold` | Specify liveness `failureThreshold` parameter for the deployment | `5` |
| `probes.liveness.timeoutSeconds` | Specify liveness `timeoutSeconds` parameter for the deployment | `10` |
| `probes.liveness.periodSeconds` | Specify liveness `periodSeconds` parameter for the deployment | `10` |
| `probes.readiness.initialDelaySeconds` | Specify readiness `initialDelaySeconds` parameter for the deployment | `60` |
| `probes.readiness.failureThreshold` | Specify readiness `failureThreshold` parameter for the deployment | `5` |
| `probes.readiness.timeoutSeconds` | Specify readiness `timeoutSeconds` parameter for the deployment | `10` |
| `probes.readiness.periodSeconds` | Specify readiness `periodSeconds` parameter for the deployment | `10` |
| `Service.type` | Kubernetes service type for the grocy GUI | `ClusterIP` |
| `Service.port` | Kubernetes port where the grocy GUI is exposed| `9283` |
| `Service.annotations` | Service annotations for the grocy GUI | `{}` |
| `Service.labels` | Custom labels | `{}` |
| `Service.loadBalancerIP` | Loadbalance IP for the grocy GUI | `{}` |
| `Service.loadBalancerSourceRanges` | List of IP CIDRs allowed access to load balancer (if supported) | None
| `ingress.enabled` | Enables Ingress | `false` |
| `ingress.annotations` | Ingress annotations | `{}` |
| `ingress.labels` | Custom labels | `{}`
| `ingress.path` | Ingress path | `/` |
| `ingress.hosts` | Ingress accepted hostnames | `chart-example.local` |
| `ingress.tls` | Ingress TLS configuration | `[]` |
| `persistence.config.enabled` | Use persistent volume to store configuration data | `true` |
| `persistence.config.size` | Size of persistent volume claim | `1Gi` |
| `persistence.config.existingClaim`| Use an existing PVC to persist data | `nil` |
| `persistence.config.subPath` | Mount a sub directory of the persistent volume if set | `""` |
| `persistence.config.storageClass` | Type of persistent volume claim | `-` |
| `persistence.config.accessMode` | Persistence access mode | `ReadWriteOnce` |
| `persistence.config.skipuninstall` | Do not delete the pvc upon helm uninstall | `false` |
| `resources` | CPU/Memory resource requests/limits | `{}` |
| `nodeSelector` | Node labels for pod assignment | `{}` |
| `tolerations` | Toleration labels for pod assignment | `[]` |
| `affinity` | Affinity settings for pod assignment | `{}` |
| `podAnnotations` | Key-value pairs to add as pod annotations | `{}` |
| `deploymentAnnotations` | Key-value pairs to add as deployment annotations | `{}` |
Read through the charts [values.yaml](https://github.com/k8s-at-home/charts/blob/master/charts/grocy/values.yaml)
file. It has several commented out suggested values.
Additionally you can take a look at the common library [values.yaml](https://github.com/k8s-at-home/charts/blob/master/charts/common/values.yaml) for more (advanced) configuration options.
Specify each parameter using the `--set key=value[,key=value]` argument to `helm install`. For example,
```console
helm install --name my-release \
--set timezone="America/New York" \
helm install grocy \
--set env.TZ="America/New_York" \
k8s-at-home/grocy
```
Alternatively, a YAML file that specifies the values for the above parameters can be provided while installing the chart. For example,
Alternatively, a YAML file that specifies the values for the above parameters can be provided while installing the
chart. For example,
```console
helm install --name my-release -f values.yaml k8s-at-home/grocy
helm install grocy k8s-at-home/grocy --values values.yaml
```
```yaml
image:
tag: ...
```
---
**NOTE**
If you get `Error: rendered manifests contain a resource that already exists. Unable to continue with install: existing resource conflict: ...` it may be because you uninstalled the chart with `skipuninstall` enabled, you need to manually delete the pvc or use `existingClaim`.
If you get
```console
Error: rendered manifests contain a resource that already exists. Unable to continue with install: existing resource conflict: ...`
```
it may be because you uninstalled the chart with `skipuninstall` enabled, you need to manually delete the pvc or use `existingClaim`.
---
Read through the [values.yaml](https://github.com/k8s-at-home/charts/blob/master/charts/grocy/values.yaml) file. It has several commented out suggested values.
## Upgrading an existing Release to a new major version
A major chart version change (like 4.0.1 -> 5.0.0) indicates that there is an incompatible breaking change potentially needing manual actions.
Refer to the library values.yaml for more configuration options.

View File

@@ -0,0 +1,2 @@
ingress:
enabled: true

View File

@@ -1,19 +0,0 @@
1. Get the application URL by running these commands:
{{- if .Values.ingress.enabled }}
{{- range .Values.ingress.hosts }}
http{{ if $.Values.ingress.tls }}s{{ end }}://{{ . }}{{ $.Values.ingress.path }}
{{- end }}
{{- else if contains "NodePort" .Values.service.type }}
export NODE_PORT=$(kubectl get --namespace {{ .Release.Namespace }} -o jsonpath="{.spec.ports[0].nodePort}" services {{ include "grocy.fullname" . }})
export NODE_IP=$(kubectl get nodes --namespace {{ .Release.Namespace }} -o jsonpath="{.items[0].status.addresses[0].address}")
echo http://$NODE_IP:$NODE_PORT
{{- else if contains "LoadBalancer" .Values.service.type }}
NOTE: It may take a few minutes for the LoadBalancer IP to be available.
You can watch the status of by running 'kubectl get svc -w {{ include "grocy.fullname" . }}'
export SERVICE_IP=$(kubectl get svc --namespace {{ .Release.Namespace }} {{ include "grocy.fullname" . }} -o jsonpath='{.status.loadBalancer.ingress[0].ip}')
echo http://$SERVICE_IP:{{ .Values.service.port }}
{{- else if contains "ClusterIP" .Values.service.type }}
export POD_NAME=$(kubectl get pods --namespace {{ .Release.Namespace }} -l "app.kubernetes.io/name={{ include "grocy.name" . }},app.kubernetes.io/instance={{ .Release.Name }}" -o jsonpath="{.items[0].metadata.name}")
echo "Visit http://127.0.0.1:3579 to use your application"
kubectl port-forward $POD_NAME 3579:80
{{- end }}

View File

@@ -1,32 +0,0 @@
{{/* vim: set filetype=mustache: */}}
{{/*
Expand the name of the chart.
*/}}
{{- define "grocy.name" -}}
{{- default .Chart.Name .Values.nameOverride | trunc 63 | trimSuffix "-" -}}
{{- end -}}
{{/*
Create a default fully qualified app name.
We truncate at 63 chars because some Kubernetes name fields are limited to this (by the DNS naming spec).
If release name contains chart name it will be used as a full name.
*/}}
{{- define "grocy.fullname" -}}
{{- if .Values.fullnameOverride -}}
{{- .Values.fullnameOverride | trunc 63 | trimSuffix "-" -}}
{{- else -}}
{{- $name := default .Chart.Name .Values.nameOverride -}}
{{- if contains $name .Release.Name -}}
{{- .Release.Name | trunc 63 | trimSuffix "-" -}}
{{- else -}}
{{- printf "%s-%s" .Release.Name $name | trunc 63 | trimSuffix "-" -}}
{{- end -}}
{{- end -}}
{{- end -}}
{{/*
Create chart name and version as used by the chart label.
*/}}
{{- define "grocy.chart" -}}
{{- printf "%s-%s" .Chart.Name .Chart.Version | replace "+" "_" | trunc 63 | trimSuffix "-" -}}
{{- end -}}

View File

@@ -0,0 +1 @@
{{ include "common.all" . }}

View File

@@ -1,29 +0,0 @@
{{- if and .Values.persistence.config.enabled (not .Values.persistence.config.existingClaim) }}
kind: PersistentVolumeClaim
apiVersion: v1
metadata:
name: {{ template "grocy.fullname" . }}-config
{{- if .Values.persistence.config.skipuninstall }}
annotations:
"helm.sh/resource-policy": keep
{{- end }}
labels:
app.kubernetes.io/name: {{ include "grocy.name" . }}
helm.sh/chart: {{ include "grocy.chart" . }}
app.kubernetes.io/instance: {{ .Release.Name }}
app.kubernetes.io/managed-by: {{ .Release.Service }}
spec:
accessModes:
- {{ .Values.persistence.config.accessMode | quote }}
resources:
requests:
storage: {{ .Values.persistence.config.size | quote }}
{{- if .Values.persistence.config.storageClass }}
{{- if (eq "-" .Values.persistence.config.storageClass) }}
storageClassName: ""
{{- else }}
storageClassName: "{{ .Values.persistence.config.storageClass }}"
{{- end }}
{{- end }}
{{- end -}}

View File

@@ -1,95 +0,0 @@
apiVersion: apps/v1
kind: Deployment
metadata:
name: {{ include "grocy.fullname" . }}
{{- if .Values.deploymentAnnotations }}
annotations:
{{- range $key, $value := .Values.deploymentAnnotations }}
{{ $key }}: {{ $value | quote }}
{{- end }}
{{- end }}
labels:
app.kubernetes.io/name: {{ include "grocy.name" . }}
helm.sh/chart: {{ include "grocy.chart" . }}
app.kubernetes.io/instance: {{ .Release.Name }}
app.kubernetes.io/managed-by: {{ .Release.Service }}
spec:
replicas: 1
revisionHistoryLimit: 3
strategy:
type: {{ .Values.strategyType }}
selector:
matchLabels:
app.kubernetes.io/name: {{ include "grocy.name" . }}
app.kubernetes.io/instance: {{ .Release.Name }}
template:
metadata:
labels:
app.kubernetes.io/name: {{ include "grocy.name" . }}
app.kubernetes.io/instance: {{ .Release.Name }}
{{- if .Values.podAnnotations }}
annotations:
{{- range $key, $value := .Values.podAnnotations }}
{{ $key }}: {{ $value | quote }}
{{- end }}
{{- end }}
spec:
containers:
- name: {{ .Chart.Name }}
image: "{{ .Values.image.repository }}:{{ .Values.image.tag }}"
imagePullPolicy: {{ .Values.image.pullPolicy }}
ports:
- name: http
containerPort: 80
protocol: TCP
livenessProbe:
httpGet:
path: /
port: http
initialDelaySeconds: {{ .Values.probes.liveness.initialDelaySeconds }}
failureThreshold: {{ .Values.probes.liveness.failureThreshold }}
periodSeconds: {{ .Values.probes.liveness.periodSeconds }}
timeoutSeconds: {{ .Values.probes.liveness.timeoutSeconds }}
readinessProbe:
httpGet:
path: /
port: http
initialDelaySeconds: {{ .Values.probes.readiness.initialDelaySeconds }}
failureThreshold: {{ .Values.probes.readiness.failureThreshold }}
periodSeconds: {{ .Values.probes.readiness.periodSeconds }}
timeoutSeconds: {{ .Values.probes.readiness.timeoutSeconds }}
env:
- name: TZ
value: "{{ .Values.timezone }}"
- name: PUID
value: "{{ .Values.puid }}"
- name: PGID
value: "{{ .Values.pgid }}"
volumeMounts:
- mountPath: /config
name: config
{{- if .Values.persistence.config.subPath }}
subPath: "{{ .Values.persistence.config.subPath }}"
{{- end }}
resources:
{{ toYaml .Values.resources | indent 12 }}
volumes:
- name: config
{{- if .Values.persistence.config.enabled }}
persistentVolumeClaim:
claimName: {{ if .Values.persistence.config.existingClaim }}{{ .Values.persistence.config.existingClaim }}{{- else }}{{ template "grocy.fullname" . }}-config{{- end }}
{{- else }}
emptyDir: {}
{{ end }}
{{- with .Values.nodeSelector }}
nodeSelector:
{{ toYaml . | indent 8 }}
{{- end }}
{{- with .Values.affinity }}
affinity:
{{ toYaml . | indent 8 }}
{{- end }}
{{- with .Values.tolerations }}
tolerations:
{{ toYaml . | indent 8 }}
{{- end }}

View File

@@ -1,41 +0,0 @@
{{- if .Values.ingress.enabled -}}
{{- $fullName := include "grocy.fullname" . -}}
{{- $ingressPath := .Values.ingress.path -}}
apiVersion: extensions/v1beta1
kind: Ingress
metadata:
name: {{ $fullName }}
labels:
app.kubernetes.io/name: {{ include "grocy.name" . }}
helm.sh/chart: {{ include "grocy.chart" . }}
app.kubernetes.io/instance: {{ .Release.Name }}
app.kubernetes.io/managed-by: {{ .Release.Service }}
{{- with .Values.ingress.labels -}}
{{ toYaml . | nindent 4 }}
{{- end -}}
{{- with .Values.ingress.annotations }}
annotations:
{{ toYaml . | indent 4 }}
{{- end }}
spec:
{{- if .Values.ingress.tls }}
tls:
{{- range .Values.ingress.tls }}
- hosts:
{{- range .hosts }}
- {{ . | quote }}
{{- end }}
secretName: {{ .secretName }}
{{- end }}
{{- end }}
rules:
{{- range .Values.ingress.hosts }}
- host: {{ . | quote }}
http:
paths:
- path: {{ $ingressPath }}
backend:
serviceName: {{ $fullName }}
servicePort: http
{{- end }}
{{- end }}

View File

@@ -1,53 +0,0 @@
apiVersion: v1
kind: Service
metadata:
name: {{ template "grocy.fullname" . }}
labels:
app.kubernetes.io/name: {{ include "grocy.name" . }}
helm.sh/chart: {{ include "grocy.chart" . }}
app.kubernetes.io/instance: {{ .Release.Name }}
app.kubernetes.io/managed-by: {{ .Release.Service }}
{{- if .Values.service.labels }}
{{ toYaml .Values.service.labels | indent 4 }}
{{- end }}
{{- with .Values.service.annotations }}
annotations:
{{ toYaml . | indent 4 }}
{{- end }}
spec:
{{- if (or (eq .Values.service.type "ClusterIP") (empty .Values.service.type)) }}
type: ClusterIP
{{- if .Values.service.clusterIP }}
clusterIP: {{ .Values.service.clusterIP }}
{{end}}
{{- else if eq .Values.service.type "LoadBalancer" }}
type: {{ .Values.service.type }}
{{- if .Values.service.loadBalancerIP }}
loadBalancerIP: {{ .Values.service.loadBalancerIP }}
{{- end }}
{{- if .Values.service.loadBalancerSourceRanges }}
loadBalancerSourceRanges:
{{ toYaml .Values.service.loadBalancerSourceRanges | indent 4 }}
{{- end -}}
{{- else }}
type: {{ .Values.service.type }}
{{- end }}
{{- if .Values.service.externalIPs }}
externalIPs:
{{ toYaml .Values.service.externalIPs | indent 4 }}
{{- end }}
{{- if .Values.service.externalTrafficPolicy }}
externalTrafficPolicy: {{ .Values.service.externalTrafficPolicy }}
{{- end }}
ports:
- name: http
port: {{ .Values.service.port }}
protocol: TCP
targetPort: http
{{ if (and (eq .Values.service.type "NodePort") (not (empty .Values.service.nodePort))) }}
nodePort: {{.Values.service.nodePort}}
{{ end }}
selector:
app.kubernetes.io/name: {{ include "grocy.name" . }}
app.kubernetes.io/instance: {{ .Release.Name }}

View File

@@ -1,113 +1,23 @@
# Default values for Ombi.
# This is a YAML-formatted file.
# Declare variables to be passed into your templates.
# Default values for grocy.
image:
repository: linuxserver/grocy
tag: v2.6.1-ls51
tag: version-v2.7.1
pullPolicy: IfNotPresent
# upgrade strategy type (e.g. Recreate or RollingUpdate)
strategyType: Recreate
# Probes configuration
probes:
liveness:
initialDelaySeconds: 180
failureThreshold: 10
periodSeconds: 60
timeoutSeconds: 10
readiness:
initialDelaySeconds: 180
failureThreshold: 10
periodSeconds: 60
timeoutSeconds: 10
nameOverride: ""
fullnameOverride: ""
timezone: UTC
puid: 1000
pgid: 1000
strategy:
type: Recreate
service:
type: ClusterIP
port: 80
## Specify the nodePort value for the LoadBalancer and NodePort service types.
## ref: https://kubernetes.io/docs/concepts/services-networking/service/#type-nodeport
##
# nodePort:
## Provide any additional annotations which may be required. This can be used to
## set the LoadBalancer service type to internal only.
## ref: https://kubernetes.io/docs/concepts/services-networking/service/#internal-load-balancer
##
annotations: {}
labels: {}
## Use loadBalancerIP to request a specific static IP,
## otherwise leave blank
##
loadBalancerIP:
# loadBalancerSourceRanges: []
## Set the externalTrafficPolicy in the Service to either Cluster or Local
# externalTrafficPolicy: Cluster
port:
port: 80
ingress:
enabled: false
annotations: {}
# kubernetes.io/ingress.class: nginx
# kubernetes.io/tls-acme: "true"
values: {}
path: /
hosts:
- chart-example.local
tls: []
# - secretName: chart-example-tls
# hosts:
# - chart-example.local
env: {}
# TZ:
# PUID:
# PGID:
persistence:
config:
enabled: true
## Ombi configuration data Persistent Volume Storage Class
## If defined, storageClassName: <storageClass>
## If set to "-", storageClassName: "", which disables dynamic provisioning
## If undefined (the default) or set to null, no storageClassName spec is
## set, choosing the default provisioner. (gp2 on AWS, standard on
## GKE, AWS & OpenStack)
##
# storageClass: "-"
##
## If you want to reuse an existing claim, you can pass the name of the PVC using
## the existingClaim variable
# existingClaim: your-claim
accessMode: ReadWriteOnce
size: 1Gi
## If subPath is set mount a sub folder of a volume instead of the root of the volume.
## This is especially handy for volume plugins that don't natively support sub mounting (like glusterfs).
##
subPath: ""
## Do not delete the pvc upon helm uninstall
skipuninstall: false
resources: {}
# We usually recommend not to specify default resources and to leave this as a conscious
# choice for the user. This also increases chances charts run on environments with little
# resources, such as Minikube. If you do want to specify resources, uncomment the following
# lines, adjust them as necessary, and remove the curly braces after 'resources:'.
# limits:
# cpu: 100m
# memory: 128Mi
# requests:
# cpu: 100m
# memory: 128Mi
nodeSelector: {}
tolerations: []
affinity: {}
podAnnotations: {}
deploymentAnnotations: {}
enabled: false
emptyDir: false

View File

@@ -0,0 +1,24 @@
# Patterns to ignore when building packages.
# This supports shell glob matching, relative path matching, and
# negation (prefixed with !). Only one pattern per line.
.DS_Store
# Common VCS dirs
.git/
.gitignore
.bzr/
.bzrignore
.hg/
.hgignore
.svn/
# Common backup files
*.swp
*.bak
*.tmp
*~
# Various IDEs
.project
.idea/
*.tmproj
.vscode
# OWNERS file for Kubernetes
OWNERS

View File

@@ -2,7 +2,7 @@ apiVersion: v2
appVersion: 2.2.2
description: An Application dashboard and launcher
name: heimdall
version: 2.0.1
version: 3.0.0
keywords:
- heimdall
home: https://github.com/k8s-at-home/charts/tree/master/charts/heimdall
@@ -11,5 +11,9 @@ sources:
- https://hub.docker.com/r/linuxserver/heimdall/
- https://github.com/linuxserver/heimdall
maintainers:
- name: skaro13
email: simon.caron@protonmail.com
- name: billimek
email: jeff@billimek.com
dependencies:
- name: common
version: 1.7.0
repository: https://k8s-at-home.com/charts/

View File

@@ -1,4 +1,8 @@
approvers:
- billimek
- onedr0p
- bjw-s
reviewers:
- billimek
- billimek
- onedr0p
- bjw-s

View File

@@ -1,6 +1,8 @@
# Heimdall Application Dashboard
# heimdall
This is a helm chart for [heimdall](https://github.com/heimdall/heimdall) leveraging the [Linuxserver.io image](https://hub.docker.com/r/linuxserver/heimdall/)
This is a helm chart for [heimdall](https://github.com/heimdall/heimdall).
**This chart is not maintained by the upstream project and any issues with the chart should be raised [here](https://github.com/k8s-at-home/charts/issues/new/choose)**
## TL;DR;
@@ -14,7 +16,7 @@ $ helm install k8s-at-home/heimdall
To install the chart with the release name `my-release`:
```console
helm install my-release k8s-at-home/heimdall
helm install --name my-release k8s-at-home/heimdall
```
## Uninstalling the Chart
@@ -22,75 +24,46 @@ helm install my-release k8s-at-home/heimdall
To uninstall/delete the `my-release` deployment:
```console
helm delete my-release
helm delete my-release --purge
```
The command removes all the Kubernetes components associated with the chart and deletes the release.
## Configuration
The following tables lists the configurable parameters of the Sentry chart and their default values.
| Parameter | Description | Default |
|----------------------------|-------------------------------------|---------------------------------------------------------|
| `image.repository` | Image repository | `linuxserver/heimdall` |
| `image.tag` | Image tag. Possible values listed [here](https://hub.docker.com/r/linuxserver/heimdall/tags/).| `2.2.2-ls84`|
| `image.pullPolicy` | Image pull policy | `IfNotPresent` |
| `strategyType` | Specifies the strategy used to replace old Pods by new ones | `Recreate` |
| `timezone` | Timezone the heimdall instance should run as, e.g. 'America/New_York' | `UTC` |
| `puid` | process userID the heimdall instance should run as | `1001` |
| `pgid` | process groupID the heimdall instance should run as | `1001` |
| `probes.liveness.initialDelaySeconds` | Specify liveness `initialDelaySeconds` parameter for the deployment | `60` |
| `probes.liveness.failureThreshold` | Specify liveness `failureThreshold` parameter for the deployment | `5` |
| `probes.liveness.timeoutSeconds` | Specify liveness `timeoutSeconds` parameter for the deployment | `10` |
| `probes.readiness.initialDelaySeconds` | Specify readiness `initialDelaySeconds` parameter for the deployment | `60` |
| `probes.readiness.failureThreshold` | Specify readiness `failureThreshold` parameter for the deployment | `5` |
| `probes.readiness.timeoutSeconds` | Specify readiness `timeoutSeconds` parameter for the deployment | `10` |
| `Service.type` | Kubernetes service type for the heimdall GUI | `ClusterIP` |
| `Service.port` | Kubernetes port where the heimdall GUI is exposed| `80` |
| `Service.annotations` | Service annotations for the heimdall GUI | `{}` |
| `Service.labels` | Custom labels | `{}` |
| `Service.loadBalancerIP` | Loadbalance IP for the heimdall GUI | `{}` |
| `Service.loadBalancerSourceRanges` | List of IP CIDRs allowed access to load balancer (if supported) | None
| `ingress.enabled` | Enables Ingress | `false` |
| `ingress.annotations` | Ingress annotations | `{}` |
| `ingress.labels` | Custom labels | `{}`
| `ingress.path` | Ingress path | `/` |
| `ingress.hosts` | Ingress accepted hostnames | `chart-example.local` |
| `ingress.tls` | Ingress TLS configuration | `[]` |
| `persistence.config.enabled` | Use persistent volume to store configuration data | `true` |
| `persistence.config.size` | Size of persistent volume claim | `1Gi` |
| `persistence.config.existingClaim`| Use an existing PVC to persist data | `nil` |
| `persistence.config.storageClass` | Type of persistent volume claim | `-` |
| `persistence.config.accessMode` | Persistence access mode | `ReadWriteOnce` |
| `persistence.config.skipuninstall` | Do not delete the pvc upon helm uninstall | `false` |
| `persistence.extraExistingClaimMounts` | Optionally add multiple existing claims | `[]` |
| `resources` | CPU/Memory resource requests/limits | `{}` |
| `nodeSelector` | Node labels for pod assignment | `{}` |
| `tolerations` | Toleration labels for pod assignment | `[]` |
| `affinity` | Affinity settings for pod assignment | `{}` |
| `podAnnotations` | Key-value pairs to add as pod annotations | `{}` |
| `deploymentAnnotations` | Key-value pairs to add as deployment annotations | `{}` |
Read through the charts [values.yaml](https://github.com/k8s-at-home/charts/blob/master/charts/heimdall/values.yaml)
file. It has several commented out suggested values.
Additionally you can take a look at the common library [values.yaml](https://github.com/k8s-at-home/charts/blob/master/charts/common/values.yaml) for more (advanced) configuration options.
Specify each parameter using the `--set key=value[,key=value]` argument to `helm install`. For example,
```console
helm install my-release \
--set timezone="America/New York" \
helm install heimdall \
--set env.TZ="America/New_York" \
k8s-at-home/heimdall
```
Alternatively, a YAML file that specifies the values for the above parameters can be provided while installing the chart. For example,
Alternatively, a YAML file that specifies the values for the above parameters can be provided while installing the
chart. For example,
```console
helm install my-release -f values.yaml k8s-at-home/heimdall
helm install heimdall k8s-at-home/heimdall --values values.yaml
```
```yaml
image:
tag: ...
```
---
**NOTE**
If you get `Error: rendered manifests contain a resource that already exists. Unable to continue with install: existing resource conflict: ...` it may be because you uninstalled the chart with `skipuninstall` enabled, you need to manually delete the pvc or use `existingClaim`.
If you get
```console
Error: rendered manifests contain a resource that already exists. Unable to continue with install: existing resource conflict: ...`
```
it may be because you uninstalled the chart with `skipuninstall` enabled, you need to manually delete the pvc or use `existingClaim`.
---
Read through the [values.yaml](https://github.com/k8s-at-home/charts/blob/master/charts/heimdall/values.yaml) file. It has several commented out suggested values.
## Upgrading an existing Release to a new major version
A major chart version change (like 4.0.1 -> 5.0.0) indicates that there is an incompatible breaking change potentially needing manual actions.
Refer to the library values.yaml for more configuration options.

View File

@@ -0,0 +1,2 @@
ingress:
enabled: true

View File

@@ -1,19 +0,0 @@
1. Get the application URL by running these commands:
{{- if .Values.ingress.enabled }}
{{- range .Values.ingress.hosts }}
http{{ if $.Values.ingress.tls }}s{{ end }}://{{ . }}{{ $.Values.ingress.path }}
{{- end }}
{{- else if contains "NodePort" .Values.service.type }}
export NODE_PORT=$(kubectl get --namespace {{ .Release.Namespace }} -o jsonpath="{.spec.ports[0].nodePort}" services {{ include "heimdall.fullname" . }})
export NODE_IP=$(kubectl get nodes --namespace {{ .Release.Namespace }} -o jsonpath="{.items[0].status.addresses[0].address}")
echo http://$NODE_IP:$NODE_PORT
{{- else if contains "LoadBalancer" .Values.service.type }}
NOTE: It may take a few minutes for the LoadBalancer IP to be available.
You can watch the status of by running 'kubectl get svc -w {{ include "heimdall.fullname" . }}'
export SERVICE_IP=$(kubectl get svc --namespace {{ .Release.Namespace }} {{ include "heimdall.fullname" . }} -o jsonpath='{.status.loadBalancer.ingress[0].ip}')
echo http://$SERVICE_IP:{{ .Values.service.port }}
{{- else if contains "ClusterIP" .Values.service.type }}
export POD_NAME=$(kubectl get pods --namespace {{ .Release.Namespace }} -l "app.kubernetes.io/name={{ include "heimdall.name" . }},app.kubernetes.io/instance={{ .Release.Name }}" -o jsonpath="{.items[0].metadata.name}")
echo "Visit http://127.0.0.1:8080 to use your application"
kubectl port-forward $POD_NAME 8080:80
{{- end }}

View File

@@ -1,32 +0,0 @@
{{/* vim: set filetype=mustache: */}}
{{/*
Expand the name of the chart.
*/}}
{{- define "heimdall.name" -}}
{{- default .Chart.Name .Values.nameOverride | trunc 63 | trimSuffix "-" -}}
{{- end -}}
{{/*
Create a default fully qualified app name.
We truncate at 63 chars because some Kubernetes name fields are limited to this (by the DNS naming spec).
If release name contains chart name it will be used as a full name.
*/}}
{{- define "heimdall.fullname" -}}
{{- if .Values.fullnameOverride -}}
{{- .Values.fullnameOverride | trunc 63 | trimSuffix "-" -}}
{{- else -}}
{{- $name := default .Chart.Name .Values.nameOverride -}}
{{- if contains $name .Release.Name -}}
{{- .Release.Name | trunc 63 | trimSuffix "-" -}}
{{- else -}}
{{- printf "%s-%s" .Release.Name $name | trunc 63 | trimSuffix "-" -}}
{{- end -}}
{{- end -}}
{{- end -}}
{{/*
Create chart name and version as used by the chart label.
*/}}
{{- define "heimdall.chart" -}}
{{- printf "%s-%s" .Chart.Name .Chart.Version | replace "+" "_" | trunc 63 | trimSuffix "-" -}}
{{- end -}}

View File

@@ -0,0 +1 @@
{{ include "common.all" . }}

View File

@@ -1,29 +0,0 @@
{{- if and .Values.persistence.config.enabled (not .Values.persistence.config.existingClaim) }}
kind: PersistentVolumeClaim
apiVersion: v1
metadata:
name: {{ template "heimdall.fullname" . }}-config
{{- if .Values.persistence.config.skipuninstall }}
annotations:
"helm.sh/resource-policy": keep
{{- end }}
labels:
app.kubernetes.io/name: {{ include "heimdall.name" . }}
helm.sh/chart: {{ include "heimdall.chart" . }}
app.kubernetes.io/instance: {{ .Release.Name }}
app.kubernetes.io/managed-by: {{ .Release.Service }}
spec:
accessModes:
- {{ .Values.persistence.config.accessMode | quote }}
resources:
requests:
storage: {{ .Values.persistence.config.size | quote }}
{{- if .Values.persistence.config.storageClass }}
{{- if (eq "-" .Values.persistence.config.storageClass) }}
storageClassName: ""
{{- else }}
storageClassName: "{{ .Values.persistence.config.storageClass }}"
{{- end }}
{{- end }}
{{- end -}}

View File

@@ -1,98 +0,0 @@
apiVersion: apps/v1
kind: Deployment
metadata:
name: {{ include "heimdall.fullname" . }}
{{- if .Values.deploymentAnnotations }}
annotations:
{{- range $key, $value := .Values.deploymentAnnotations }}
{{ $key }}: {{ $value | quote }}
{{- end }}
{{- end }}
labels:
app.kubernetes.io/name: {{ include "heimdall.name" . }}
helm.sh/chart: {{ include "heimdall.chart" . }}
app.kubernetes.io/instance: {{ .Release.Name }}
app.kubernetes.io/managed-by: {{ .Release.Service }}
spec:
replicas: 1
revisionHistoryLimit: 3
strategy:
type: {{ .Values.strategyType }}
selector:
matchLabels:
app.kubernetes.io/name: {{ include "heimdall.name" . }}
app.kubernetes.io/instance: {{ .Release.Name }}
template:
metadata:
labels:
app.kubernetes.io/name: {{ include "heimdall.name" . }}
app.kubernetes.io/instance: {{ .Release.Name }}
{{- if .Values.podAnnotations }}
annotations:
{{- range $key, $value := .Values.podAnnotations }}
{{ $key }}: {{ $value | quote }}
{{- end }}
{{- end }}
spec:
containers:
- name: {{ .Chart.Name }}
image: "{{ .Values.image.repository }}:{{ .Values.image.tag }}"
imagePullPolicy: {{ .Values.image.pullPolicy }}
ports:
- name: http
containerPort: 80
protocol: TCP
livenessProbe:
tcpSocket:
port: http
initialDelaySeconds: {{ .Values.probes.liveness.initialDelaySeconds }}
failureThreshold: {{ .Values.probes.liveness.failureThreshold }}
timeoutSeconds: {{ .Values.probes.liveness.timeoutSeconds }}
readinessProbe:
tcpSocket:
port: http
initialDelaySeconds: {{ .Values.probes.readiness.initialDelaySeconds }}
failureThreshold: {{ .Values.probes.readiness.failureThreshold }}
timeoutSeconds: {{ .Values.probes.readiness.timeoutSeconds }}
env:
- name: TZ
value: "{{ .Values.timezone }}"
- name: PUID
value: "{{ .Values.puid }}"
- name: PGID
value: "{{ .Values.pgid }}"
volumeMounts:
- mountPath: /config
name: config
{{- range .Values.persistence.extraExistingClaimMounts }}
- name: {{ .name }}
mountPath: {{ .mountPath }}
readOnly: {{ .readOnly }}
{{- end }}
resources:
{{ toYaml .Values.resources | indent 12 }}
volumes:
- name: config
{{- if .Values.persistence.config.enabled }}
persistentVolumeClaim:
claimName: {{ if .Values.persistence.config.existingClaim }}{{ .Values.persistence.config.existingClaim }}{{- else }}{{ template "heimdall.fullname" . }}-config{{- end }}
{{- else }}
emptyDir: {}
{{- end }}
{{- range .Values.persistence.extraExistingClaimMounts }}
- name: {{ .name }}
persistentVolumeClaim:
claimName: {{ .existingClaim }}
{{- end }}
{{- with .Values.nodeSelector }}
nodeSelector:
{{ toYaml . | indent 8 }}
{{- end }}
{{- with .Values.affinity }}
affinity:
{{ toYaml . | indent 8 }}
{{- end }}
{{- with .Values.tolerations }}
tolerations:
{{ toYaml . | indent 8 }}
{{- end }}

Some files were not shown because too many files have changed in this diff Show More