Homarr¶
A Helm chart to deploy homarr for Kubernetes
This chart is not maintained by the upstream project and any issues with the chart should be raised here
Source Code¶
📄 Requirements¶
Kubernetes: >=1.22.0-0
🍱 Dependencies¶
This chart has no dependencies.
🚀 Installing the Chart¶
To install the chart with the release name homarr
OCI (Recommended)¶
Traditional¶
helm repo add oben01 https://oben01.github.io/charts/
helm repo update
helm install homarr oben01/homarr
➖ Uninstalling the Chart¶
To uninstall the homarr
deployment
The command removes all the Kubernetes components associated with the chart including persistent volumes and deletes the release.
🔧 Configuration¶
Read through the values.yaml file. It has several commented out suggested values.
Specify each parameter using the --set key=value[,key=value]
argument to helm install
.
Alternatively, a YAML file that specifies the values for the above parameters can be provided while installing the chart.
🎨 Custom configuration¶
N/A
🗃 Values¶
Key | Type | Default | Description |
---|---|---|---|
affinity | object | {} | Node affinity for pod scheduling |
autoscaling | object | {"enabled":false,"maxReplicas":100,"minReplicas":1,"targetCPUUtilizationPercentage":80} | Autoscaling configuration |
autoscaling.enabled | bool | false | Enable autoscaling |
autoscaling.maxReplicas | int | 100 | Maximum replicas |
autoscaling.minReplicas | int | 1 | Minimum replicas |
autoscaling.targetCPUUtilizationPercentage | int | 80 | Target CPU utilization for autoscaling |
env.AUTH_LDAP_ADMIN_GROUP | string | "admin" | Admin group |
env.AUTH_LDAP_BASE | string | nil | Base dn of your LDAP server |
env.AUTH_LDAP_BIND_DN | string | nil | User used for finding users and groups |
env.AUTH_LDAP_GROUP_CLASS | string | "groupOfUniqueNames" | Class used for querying groups |
env.AUTH_LDAP_GROUP_MEMBER_ATTRIBUTE | string | "member" | Attribute used for querying group member |
env.AUTH_LDAP_GROUP_MEMBER_USER_ATTRIBUTE | string | "dn" | User attribute used for comparing with group member |
env.AUTH_LDAP_OWNER_GROUP | string | "admin" | Owner group |
env.AUTH_LDAP_SEARCH_SCOPE | string | "base" | LDAP search scope between base, one or sub |
env.AUTH_LDAP_URI | string | nil | URI of your LDAP server |
env.AUTH_LDAP_USERNAME_ATTRIBUTE | string | "uid" | Attribute used for username |
env.AUTH_OIDC_ADMIN_GROUP | string | "admin" | Admin group |
env.AUTH_OIDC_CLIENT_NAME | string | "OIDC" | Display name of provider (in login screen) |
env.AUTH_OIDC_OWNER_GROUP | string | "admin" | Owner group |
env.AUTH_OIDC_SCOPE_OVERWRITE | string | "openid email profile groups" | Override the OIDC scopes |
env.AUTH_OIDC_URI | string | nil | URI of OIDC provider |
env.AUTH_PROVIDER | string | "credentials" | Enabled authentication methods. Multiple providers can be enabled with by separating them with , (ex. AUTH_PROVIDER=credentials,oidc, it is highly recommended to just enable one provider). |
env.DATABASE_URL | string | "file:/app/database/db.sqlite" | URL for the Homarr database |
env.DEFAULT_COLOR_SCHEME | string | "dark" | Colors and preferences, possible values dark / light |
env.TZ | string | "Europe/Paris" | Your local time zone |
envSecrets.AUTH_LDAP_BIND_PASSWORD | object | {"key":null,"name":null} | Password for bind user |
envSecrets.AUTH_OIDC_CLIENT_ID | object | {"key":null,"name":null} | ID of OIDC client (application) |
envSecrets.AUTH_OIDC_CLIENT_SECRET | object | {"key":null,"name":null} | Secret of OIDC client (application) |
fullnameOverride | string | "" | Overrides chart's fullname |
image.pullPolicy | string | "IfNotPresent" | Image pull policy |
image.repository | string | "ghcr.io/ajnart/homarr" | Image repository |
image.tag | string | "0.15.10" | Overrides the image tag whose default is the chart appVersion |
imagePullSecrets | list | [] | Secrets for Docker registry |
ingress | object | {"annotations":{},"enabled":true,"hosts":[{"host":"chart-example.local","paths":[{"path":"/"}]}],"ingressClassName":"","tls":[]} | Ingress configuration |
ingress.annotations | object | {} | Ingress annotations |
ingress.enabled | bool | true | Enable ingress |
ingress.hosts | list | [{"host":"chart-example.local","paths":[{"path":"/"}]}] | Ingress hosts configuration |
ingress.ingressClassName | string | "" | Ingress class name |
ingress.tls | list | [] | Ingress TLS configuration |
livenessProbe.failureThreshold | int | 5 | The number of consecutive failed liveness probe attempts before the container is considered unhealthy. |
livenessProbe.httpGet.path | string | "/" | This is the liveness check endpoint used by Kubernetes to determine if the application is still running. |
livenessProbe.httpGet.port | int | 7575 | The port on which the liveness check will be performed. This must be the same as the container port exposed by the application. |
livenessProbe.initialDelaySeconds | int | 10 | Number of seconds after the container has started before startup |
livenessProbe.periodSeconds | int | 10 | How often (in seconds) to perform the liveness probe. |
livenessProbe.timeoutSeconds | int | 3 | Number of seconds after which the probe times out |
nameOverride | string | "" | Overrides chart's name |
nodeSelector | object | {} | Node selectors for pod scheduling |
persistence | list | [{"accessMode":"ReadWriteOnce","enabled":false,"mountPath":"/app/data/configs","name":"homarr-config","size":"50Mi","storageClassName":"local-path"},{"accessMode":"ReadWriteOnce","enabled":false,"mountPath":"/app/database","name":"homarr-database","size":"50Mi","storageClassName":"local-path"},{"accessMode":"ReadWriteOnce","enabled":false,"mountPath":"/app/public/icons","name":"homarr-icons","size":"50Mi","storageClassName":"local-path"}] | Persistent storage configuration |
persistence[0].accessMode | string | "ReadWriteOnce" | homarr-config access mode |
persistence[0].enabled | bool | false | Enable homarr-config persistent storage |
persistence[0].mountPath | string | "/app/data/configs" | homarr-config mount path inside the pod |
persistence[0].size | string | "50Mi" | homarr-config storage size |
persistence[0].storageClassName | string | "local-path" | homarr-config storage class name |
persistence[1].accessMode | string | "ReadWriteOnce" | homarr-database access mode |
persistence[1].enabled | bool | false | Enable homarr-database persistent storage |
persistence[1].mountPath | string | "/app/database" | homarr-database mount path inside the pod |
persistence[1].size | string | "50Mi" | homarr-database storage size |
persistence[1].storageClassName | string | "local-path" | homarr-database storage class name |
persistence[2].accessMode | string | "ReadWriteOnce" | homarr-icons access mode |
persistence[2].enabled | bool | false | Enable homarr-icons persistent storage |
persistence[2].mountPath | string | "/app/public/icons" | homarr-icons mount path inside the pod |
persistence[2].size | string | "50Mi" | homarr-icons storage size |
persistence[2].storageClassName | string | "local-path" | homarr-icons storage class name |
podAnnotations | object | {} | Pod annotations |
podLabels | object | {} | Pod labels |
podSecurityContext | object | {} | Pod security context |
readinessProbe.failureThreshold | int | 5 | The number of consecutive failed readiness probe attempts before the container is considered "Not Ready." |
readinessProbe.httpGet.path | string | "/" | This is the readiness check endpoint used by Kubernetes to determine if the application is ready to handle traffic. |
readinessProbe.httpGet.port | int | 7575 | The port on which the readiness check will be performed. This must match the container's exposed port. |
readinessProbe.initialDelaySeconds | int | 10 | Number of seconds after the container has started before startup |
readinessProbe.periodSeconds | int | 10 | How often (in seconds) to perform the readiness probe. |
readinessProbe.timeoutSeconds | int | 3 | Number of seconds after which the probe times out |
replicaCount | int | 1 | Number of replicas |
resources | object | {} | Resource configuration |
securityContext | object | {} | Security context |
service | object | {"port":7575,"targetPort":7575,"type":"ClusterIP"} | Service configuration |
service.port | int | 7575 | Service port |
service.targetPort | int | 7575 | Service target port |
service.type | string | "ClusterIP" | Service type |
tolerations | list | [] | Node tolerations for pod scheduling |
Autogenerated from chart metadata using helm-docs