Deploying with Helm Chart (Kubernetes)

Why Helm Charts

Helm is the package manager for Kubernetes, which allows us to efficiently manage Kubernetes applications.

Helm Charts are the "packages" in the Helm world. They help us define, install, and upgrade nearly every Kubernetes application!

In this doc, we'll learn how to bootstrap a PostHog deployment on a Kubernetes cluster using Helm. We also optionally package PostgreSQL and Redis - these are required for PostHog. You can read more about this in the Packages section.


  • Kubernetes 1.4+ with Beta APIs enabled
  • Helm >= v3
  • PV provisioner support in the underlying infrastructure (with persistence storage enabled)

Step By Step Installation

  1. Ensure you meet the prerequisites above.
  2. Run the following:

NOTE: If Helm hangs while installing, try increasing the memory of your nodes. As a baseline, we suggest having at least 4gb of memory per node.

helm repo add posthog
helm repo update
helm install posthog posthog/posthog


The following table lists the configurable parameters of the PostHog Helm Chart and their default values.

Dependent charts can also have values overwritten. Preface values with PostgreSQL.

Parameter Description Default
image.repository PostHog image posthog/posthog
image.tag PostHog image tag latest
image.pullPolicy Image pull policy Always
image.imagePullSecrets Specify image pull secrets []
posthogSecret Specify SECRET_KEY. If isn't specified it will be generated automatically. nil
web.podAnnotations Web pod annotations {}
web.podLabels Web pod extra labels {}
web.replicacount Amount of web pods to run 1
web.resources.limits Web resource limits {cpu: 500m, memory: 500Mi}
web.resources.requests Web resource requests {cpu: 300m, memory: 300Mi}
web.env Additional web environment variables []
web.nodeSelector Node labels for web pod assignment {}
web.affinity Affinity settings for web pod assignment {}
web.schedulerName Name of an alternate scheduler for web pod nil
web.tolerations Toleration labels for web pod assignment []
web.livenessProbe.failureThreshold The liveness probe failure threshold 5
web.livenessProbe.initialDelaySeconds The liveness probe initial delay seconds 50
web.livenessProbe.periodSeconds The liveness probe period seconds 10
web.livenessProbe.successThreshold The liveness probe success threshold 1
web.livenessProbe.timeoutSeconds The liveness probe timeout seconds 2
web.readinessProbe.failureThreshold The readiness probe failure threshold 10
web.readinessProbe.initialDelaySeconds The readiness probe initial delay seconds 50
web.readinessProbe.periodSeconds The readiness probe period seconds 10
web.readinessProbe.successThreshold The readiness probe success threshold 1
web.readinessProbe.timeoutSeconds The readiness probe timeout seconds 2
web.priorityClassName The priorityClassName on web deployment nil
web.hpa.enabled Boolean to create a HorizontalPodAutoscaler for web deployment false
web.hpa.cputhreshold CPU threshold percent for the web HorizontalPodAutoscaler 60
web.hpa.minpods Min pods for the web HorizontalPodAutoscaler 1
web.hpa.maxpods Max pods for the web HorizontalPodAutoscaler 10
email.from_email Emails are sent are from [email protected] SMTP host for sending email smtp
email.port SMTP port 578
email.user SMTP user nil
email.password SMTP password nil
email.use_tls SMTP TLS for security false
email.use_ssl SMTP SSL for security false
email.existingSecret SMTP password from an existing secret nil
email.existingSecretKey Key to get from the email.existingSecret secret smtp-password
service.type Kubernetes service type LoadBalancer Kubernetes service name posthog
service.externalPort Kubernetes external service port 8000
service.internalPort Kubernetes internal service port 8000
service.annotations Service annotations {}
service.nodePort Kubernetes service NodePort port Randomly chosen by Kubernetes
service.loadBalancerSourceRanges Allow list for the load balancer nil
ingress.enabled Enable ingress controller resource false
ingress.annotations Ingress annotations {}
ingress.hostname URL to address your PostHog installation posthog.local
ingress.path path to address your PostHog installation /
ingress.tls Ingress TLS configuration []
postgresql.enabled Deploy postgres server (see below) true
postgresql.postgresqlDatabase Postgres database name posthog
postgresql.postgresqlUsername Postgres username postgres
postgresql.postgresqlHost External postgres host nil
postgresql.postgresqlPassword External/Internal postgres password postgres
postgresql.postgresqlPort External postgres port 5432
postgresql.existingSecret Name of existing secret to use for the PostgreSQL password nil
postgresql.existingSecretKey Key to get from the postgresql.existingSecret secret postgresql-password
redis.enabled Deploy redis server (see below) true External redis host nil
redis.password External redis password nil
redis.port External redis port 6379
redis.existingSecret Name of existing secret to use for the Redis password nil
redis.existingSecretKey Key to get from the redis.existingSecret secret redis-password
metrics.enabled Start an exporter for posthog metrics false
metrics.nodeSelector Node labels for metrics pod assignment {}
metrics.tolerations Toleration labels for metrics pod assignment []
metrics.affinity Affinity settings for metrics pod {}
metrics.schedulerName Name of an alternate scheduler for metrics pod nil
metrics.podLabels Labels for metrics pod nil
metrics.resources Metrics resource requests/limit {}
metrics.service.type Kubernetes service type for metrics service ClusterIP
metrics.service.labels Additional labels for metrics service {}
metrics.image.repository Metrics exporter image repository prom/statsd-exporter
metrics.image.tag Metrics exporter image tag v0.10.5
metrics.image.PullPolicy Metrics exporter image pull policy IfNotPresent
metrics.serviceMonitor.enabled if true, creates a Prometheus Operator ServiceMonitor (also requires metrics.enabled to be true) false
metrics.serviceMonitor.namespace Optional namespace which Prometheus is running in nil
metrics.serviceMonitor.interval How frequently to scrape metrics (use by default, falling back to Prometheus' default) nil
metrics.serviceMonitor.selector Default to kube-prometheus install (CoreOS recommended), but should be set according to Prometheus install { prometheus: kube-prometheus }
hooks.affinity Affinity settings for hooks pods {}
hooks.migrate.resources.limits Hook job resource limits {memory: 1000Mi}
hooks.migrate.resources.requests Hook job resource requests {memory: 1000Mi} name of the ServiceAccount to be used by access-controlled resources autogenerated
serviceAccount.create Configures if a ServiceAccount with this name should be created true
serviceAccount.annotations Configures annotation for the ServiceAccount {}

Dependent charts can also have values overwritten. Preface values with "postgresql."

Specify each parameter using the --set key=value[,key=value] argument to helm install. For example,

$ helm install \
  --set persistence.enabled=false, \
  my-release .

Alternatively, a YAML file that specifies the values for the above parameters can be provided while installing the chart. For example,

$ helm install -f my-values.yaml my-release .



By default, PostgreSQL is installed as part of the chart. To use an external PostgreSQL server set postgresql.enabled to false and then set postgresql.postgresHost and postgresql.postgresqlPassword.

The other options (postgresql.postgresqlDatabase, postgresql.postgresqlUsername and postgresql.postgresqlPort) may also need changes from their default values.

To avoid issues when upgrading this chart, provide postgresql.postgresqlPassword for subsequent upgrades. This is due to an issue in the PostgreSQL chart where the password will be overwritten with randomly generated passwords otherwise. See this doc for more details.


By default, Redis is installed as part of the chart. To use an external Redis server/cluster set redis.enabled to false and then set If your Redis cluster uses a password, you should define it with redis.password, otherwise just omit it. Check the table above for more configuration options.

To avoid issues when upgrading this chart, provide redis.password for subsequent upgrades. If you do not, the Redis pods will get recreated on every update, potentially incurring some downtime.


This chart provides support for the Ingress resource. If you have an available Ingress Controller such as Nginx or Traefik you maybe want to set ingress.enabled to true and choose an ingress.hostname for the URL. You should then be able to access the installation using that address.