Replies: 1 comment 4 replies
-
In default installation, prometheus is deployed with 2 replicas and alertmanager in 3 replicas. Storage is up to the user and we don't specify if this is ReadWriteOnce or ReadWriteMany.
That depends on the storage provider and it is out of scope for prometheus-operator. We define volume claim option and storage provider is doing the rest.
You should use node anti-affinity (provided by an addon in https://github.com/prometheus-operator/kube-prometheus/blob/main/jsonnet/kube-prometheus/addons/anti-affinity.libsonnet) to ensure pods are not landing on the same node and thus node cannot bring down whole monitoring.
Prometheus and alertmanager instances are managed by prometheus-operator and StatefuleSet creation is embedded in the logic of the operator itself. There is no easy way to change it.
This project doesn't provide helm chart. |
Beta Was this translation helpful? Give feedback.
-
Hi,
As I understand, currently prometheus and alertmanager instance(s) which need persistent storage are deployed as single-instance StatefulSet(s) that create ReadWriteOnce volume claims for their storage needs. The resulting PODs are inherently bound to the cluster node to which they are 1st scheduled. But this is not HA setup. In the event that this node goes down, prometheus and/or altert manager are not available any more.
If one has a volume provisioner available that provisions ReadWriteMany type of volumes, then an alternative way to deploy prometheus and altertmanager would be to create single-instance Deployment(s) which would then be HA. Would that be something this project is interested to provide as an alternative "mode" of Helm chart?
Beta Was this translation helpful? Give feedback.
All reactions