Kind: SGShardedCluster
listKind: SGShardedClusterList
plural: sgshardedclusters
singular: sgshardedcluster
shortNames sgscl
StackGres PostgreSQL sharded clusters are created using the SGShardedCluster
Custom Resource.
Example:
apiVersion: stackgres.io/v1alpha1
kind: SGShardedCluster
metadata:
name: stackgres
spec:
postgres:
version: 'latest'
coordinator:
instances: 1
pods:
persistentVolume:
size: '5Gi'
shards:
clusters: 2
instancesPerCluster: 1
pods:
persistentVolume:
size: '5Gi'
See also Sharded Cluster Creation section.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
apiVersion | ✓ | string | stackgres.io/v1alpha1 | ||
kind | ✓ | string | SGShardedCluster | ||
metadata | ✓ | ✓ | object | Refer to the Kubernetes API documentation for the fields of the metadata field. |
|
spec | ✓ | ✓ | object |
Specification of the desired behavior of a StackGres sharded cluster.
|
|
status | ✓ | object |
Current status of a StackGres sharded cluster. |
Specification of the desired behavior of a StackGres sharded cluster.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
coordinator | ✓ | ✓ | object |
The coordinator is a StackGres cluster responsible of coordinating data storage and access from the shards.
|
|
database | ✓ | ✓ | string |
The database name that will be created and used across all node and where “partitioned” (distributed) tables will live in.
|
|
postgres | ✓ | ✓ | object |
This section allows to configure Postgres features
|
|
shards | ✓ | ✓ | object |
The shards are a group of StackGres clusters where the partitioned data chunks are stored.
When referring to the cluster in the descriptions belove it apply to any shard’s StackGres cluster.
|
|
configurations | ✓ | object |
Sharded cluster custom configurations.
|
||
distributedLogs | ✓ | object |
StackGres features a functionality for all pods to send Postgres, Patroni and PgBouncer logs to a central (distributed) location, which is in turn another Postgres database. Logs can then be accessed via SQL interface or from the web UI. This section controls whether to enable this feature or not. If not enabled, logs are send to the pod's standard output.
|
||
metadata | ✓ | object |
Metadata information from any cluster created resources. |
||
nonProductionOptions | ✓ | object |
|
||
postgresServices | ✓ | object |
Kubernetes [services](https://kubernetes.io/docs/concepts/services-networking/service/) created or managed by StackGres. |
||
prometheusAutobind | ✓ | boolean |
If enabled, a ServiceMonitor is created for each Prometheus instance found in order to collect metrics.
|
||
replication | ✓ | object |
This section allows to configure the global Postgres replication mode.
The main replication group is implicit and contains the total number of instances less the sum of all instances in other replication groups. The total number of instances is always specified by |
||
type | ✓ | string |
The sharding technology that will be used for the sharded cluster.
Currently the only possible value for this field is |
The coordinator is a StackGres cluster responsible of coordinating data storage and access from the shards.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
instances | ✓ | ✓ | integer |
Number of StackGres instances for the cluster. Each instance contains one Postgres server.
Out of all of the Postgres servers, one is elected as the primary, the rest remain as read-only replicas.
Minimum: 1 Maximum: 16 |
|
pods | ✓ | ✓ | object |
Cluster pod’s configuration. |
|
configurations | ✓ | object |
Coordinator custom configurations.
|
||
managedSql | ✓ | object |
This section allows to reference SQL scripts that will be applied to the cluster live.
|
||
metadata | ✓ | object |
Metadata information from coordinator cluster created resources. |
||
replication | ✓ | object |
This section allows to configure the global Postgres replication mode.
The main replication group is implicit and contains the total number of instances less the sum of all instances in other replication groups. The total number of instances is always specified by |
||
sgInstanceProfile | ✓ | ✓ | string |
Name of the SGInstanceProfile.
A SGInstanceProfile defines CPU and memory limits. Must exist before creating a cluster. When no profile is set, a default (1 core, 2 GiB RAM) one is used. Changing this field may require a restart.
|
Cluster pod’s configuration.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
persistentVolume | ✓ | ✓ | object |
Pod’s persistent volume configuration. |
|
customContainers | ✓ | []object |
A list of custom application containers that run within the coordinator cluster’s Pods.
The name used in this section will be prefixed with the string See: https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.25/#container-v1-core
|
||
customInitContainers | ✓ | ✓ | []object |
A list of custom application init containers that run within the shards cluster’s Pods. The
custom init containers will run following the defined sequence as the end of
cluster’s Pods init containers.
The name used in this section will be prefixed with the string Changing this field may require a restart. See: https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.25/#container-v1-core
|
|
customVolumes | ✓ | ✓ | []object |
A list of custom volumes that may be used along with any container defined in
customInitContainers or customContainers sections for the coordinator.
The name used in this section will be prefixed with the string Only the following volume types are allowed: configMap, downwardAPI, emptyDir, gitRepo, glusterfs, hostPath, nfs, projected and secret Changing this field may require a restart. See: https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.25/#volume-v1-core
|
|
disableConnectionPooling | ✓ | ✓ | boolean |
If set to true , avoids creating a connection pooling (using PgBouncer) sidecar.
Changing this field may require a restart.
|
|
disableMetricsExporter | ✓ | boolean |
If set to true , avoids creating the Prometheus exporter sidecar. Recommended when there’s no intention to use Prometheus for monitoring.
|
||
disablePostgresUtil | ✓ | ✓ | boolean |
If set to true , avoids creating the postgres-util sidecar. This sidecar contains usual Postgres administration utilities that are not present in the main (patroni ) container, like psql . Only disable if you know what you are doing.
Changing this field may require a restart.
|
|
managementPolicy | ✓ | string |
managementPolicy controls how pods are created during initial scale up, when replacing pods
on nodes, or when scaling down. The default policy is OrderedReady , where pods are created
in increasing order (pod-0, then pod-1, etc) and the controller will wait until each pod is
ready before continuing. When scaling down, the pods are removed in the opposite order.
The alternative policy is Parallel which will create pods in parallel to match the desired
scale without waiting, and on scale down will delete all pods at once.
|
||
resources | ✓ | object |
Pod custom resources configuration. |
||
scheduling | ✓ | ✓ | object |
Pod custom scheduling, affinity and topology spread constratins configuration.
Changing this field may require a restart.
|
Pod’s persistent volume configuration.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
size | ✓ | ✓ | string |
Size of the PersistentVolume set for each instance of the cluster. This size is specified either in Mebibytes, Gibibytes or Tebibytes (multiples of 2^20, 2^30 or 2^40, respectively).
|
|
storageClass | ✓ | string |
Name of an existing StorageClass in the Kubernetes cluster, used to create the PersistentVolumes for the instances of the cluster.
|
Pod custom resources configuration.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
enableClusterLimitsRequirements | ✓ | ✓ | boolean |
When enabled resource limits for containers other than the patroni container wil be set just like for patroni contianer as specified in the SGInstanceProfile.
Changing this field may require a restart.
|
Pod custom scheduling, affinity and topology spread constratins configuration.
Changing this field may require a restart.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
backup | ✓ | ✓ | object |
Backup Pod custom scheduling and affinity configuration. |
|
nodeAffinity | ✓ | ✓ | object |
Node affinity is a group of node affinity scheduling rules.
See: https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.25/#nodeaffinity-v1-core
|
|
nodeSelector | ✓ | ✓ | map[string]string |
NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node’s labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/
|
|
podAffinity | ✓ | ✓ | object |
Pod affinity is a group of inter pod affinity scheduling rules.
See: https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.25/#podaffinity-v1-core
|
|
podAntiAffinity | ✓ | ✓ | object |
Pod anti affinity is a group of inter pod anti affinity scheduling rules.
See: https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.25/#podantiaffinity-v1-core
|
|
priorityClassName | ✓ | ✓ | string |
Priority indicates the importance of a Pod relative to other Pods. If a Pod cannot be scheduled, the scheduler tries to preempt (evict) lower priority Pods to make scheduling of the pending Pod possible.
|
|
tolerations | ✓ | ✓ | []object |
If specified, the pod’s tolerations.
See: https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.25/#toleration-v1-core
|
|
topologySpreadConstraints | ✓ | ✓ | []object |
TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed.
|
Backup Pod custom scheduling and affinity configuration.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
nodeAffinity | ✓ | ✓ | object |
Node affinity is a group of node affinity scheduling rules.
See: https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.25/#nodeaffinity-v1-core
|
|
nodeSelector | ✓ | ✓ | object |
Node affinity is a group of node affinity scheduling rules.
See: https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.25/#nodeaffinity-v1-core
|
|
podAffinity | ✓ | ✓ | object |
Pod affinity is a group of inter pod affinity scheduling rules.
See: https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.25/#podaffinity-v1-core
|
|
podAntiAffinity | ✓ | ✓ | object |
Pod anti affinity is a group of inter pod anti affinity scheduling rules.
See: https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.25/#podantiaffinity-v1-core
|
|
priorityClassName | ✓ | ✓ | string |
Priority indicates the importance of a Pod relative to other Pods. If a Pod cannot be scheduled, the scheduler tries to preempt (evict) lower priority Pods to make scheduling of the pending Pod possible.
|
|
tolerations | ✓ | ✓ | object |
Node affinity is a group of node affinity scheduling rules.
See: https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.25/#nodeaffinity-v1-core
|
TopologySpreadConstraint specifies how to spread matching pods among the given topology.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
maxSkew | ✓ | ✓ | ✓ | integer |
MaxSkew describes the degree to which pods may be unevenly distributed. When whenUnsatisfiable=DoNotSchedule , it is the maximum permitted difference between the number of matching pods in the target topology and the global minimum. The global minimum is the minimum number of matching pods in an eligible domain or zero if the number of eligible domains is less than MinDomains. For example, in a 3-zone cluster, MaxSkew is set to 1, and pods with the same labelSelector spread as 2/2/1: In this case, the global minimum is 1. | zone1 | zone2 | zone3 | | P P | P P | P | - if MaxSkew is 1, incoming pod can only be scheduled to zone3 to become 2/2/2; scheduling it onto zone1(zone2) would make the ActualSkew(3-1) on zone1(zone2) violate MaxSkew(1). - if MaxSkew is 2, incoming pod can be scheduled onto any zone. When whenUnsatisfiable=ScheduleAnyway , it is used to give higher precedence to topologies that satisfy it. It’s a required field. Default value is 1 and 0 is not allowed.Format: int32 |
topologyKey | ✓ | ✓ | ✓ | string |
TopologyKey is the key of node labels. Nodes that have a label with this key and identical values are considered to be in the same topology. We consider each <key, value> as a “bucket”, and try to put balanced number of pods into each bucket. We define a domain as a particular instance of a topology. Also, we define an eligible domain as a domain whose nodes meet the requirements of nodeAffinityPolicy and nodeTaintsPolicy. e.g. If TopologyKey is “kubernetes.io/hostname”, each Node is a domain of that topology. And, if TopologyKey is “topology.kubernetes.io/zone”, each zone is a domain of that topology. It’s a required field. |
whenUnsatisfiable | ✓ | ✓ | ✓ | string |
WhenUnsatisfiable indicates how to deal with a pod if it doesn’t satisfy the spread constraint. - DoNotSchedule (default) tells the scheduler not to schedule it. - ScheduleAnyway tells the scheduler to schedule the pod in any location,
but giving higher precedence to topologies that would help reduce the
skew.
A constraint is considered “Unsatisfiable” for an incoming pod if and only if every possible node assignment for that pod would violate “MaxSkew” on some topology. For example, in a 3-zone cluster, MaxSkew is set to 1, and pods with the same labelSelector spread as 3/1/1: | zone1 | zone2 | zone3 | | P P P | P | P | If WhenUnsatisfiable is set to DoNotSchedule, incoming pod can only be scheduled to zone2(zone3) to become 3/2/1(3/1/2) as ActualSkew(2-1) on zone2(zone3) satisfies MaxSkew(1). In other words, the cluster can still be imbalanced, but scheduler won’t make it more imbalanced. It’s a required field.
|
labelSelector | ✓ | ✓ | object |
A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects. |
|
matchLabelKeys | ✓ | ✓ | []string |
MatchLabelKeys is a set of pod label keys to select the pods over which spreading will be calculated. The keys are used to lookup values from the incoming pod labels, those key-value labels are ANDed with labelSelector to select the group of existing pods over which spreading will be calculated for the incoming pod. Keys that don't exist in the incoming pod labels will be ignored. A null or empty list means only match against labelSelector. |
|
minDomains | ✓ | ✓ | integer |
MinDomains indicates a minimum number of eligible domains. When the number of eligible domains with matching topology keys is less than minDomains, Pod Topology Spread treats "global minimum" as 0, and then the calculation of Skew is performed. And when the number of eligible domains with matching topology keys equals or greater than minDomains, this value has no effect on scheduling. As a result, when the number of eligible domains is less than minDomains, scheduler won't schedule more than maxSkew Pods to those domains. If value is nil, the constraint behaves as if MinDomains is equal to 1. Valid values are integers greater than 0. When value is not nil, WhenUnsatisfiable must be DoNotSchedule.
For example, in a 3-zone cluster, MaxSkew is set to 2, MinDomains is set to 5 and pods with the same labelSelector spread as 2/2/2: | zone1 | zone2 | zone3 | | P P | P P | P P | The number of domains is less than 5(MinDomains), so “global minimum” is treated as 0. In this situation, new pod with the same labelSelector cannot be scheduled, because computed skew will be 3(3 - 0) if new Pod is scheduled to any of the three zones, it will violate MaxSkew. This is a beta field and requires the MinDomainsInPodTopologySpread feature gate to be enabled (enabled by default). |
|
nodeAffinityPolicy | ✓ | ✓ | string |
NodeAffinityPolicy indicates how we will treat Pod’s nodeAffinity/nodeSelector when calculating pod topology spread skew. Options are: - Honor: only nodes matching nodeAffinity/nodeSelector are included in the calculations. - Ignore: nodeAffinity/nodeSelector are ignored. All nodes are included in the calculations.
If this value is nil, the behavior is equivalent to the Honor policy. This is a alpha-level feature enabled by the NodeInclusionPolicyInPodTopologySpread feature flag. |
|
nodeTaintsPolicy | ✓ | ✓ | string |
NodeTaintsPolicy indicates how we will treat node taints when calculating pod topology spread skew. Options are: - Honor: nodes without taints, along with tainted nodes for which the incoming pod has a toleration, are included. - Ignore: node taints are ignored. All nodes are included.
If this value is nil, the behavior is equivalent to the Ignore policy. This is a alpha-level feature enabled by the NodeInclusionPolicyInPodTopologySpread feature flag. |
A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
matchExpressions | ✓ | ✓ | []object |
matchExpressions is a list of label selector requirements. The requirements are ANDed. |
|
matchLabels | ✓ | ✓ | map[string]string |
matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is “key”, the operator is “In”, and the values array contains only “value”. The requirements are ANDed. |
A label selector requirement is a selector that contains values, a key, and an operator that relates the key and values.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
key | ✓ | ✓ | ✓ | string |
key is the label key that the selector applies to. |
operator | ✓ | ✓ | ✓ | string |
operator represents a key’s relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist. |
values | ✓ | ✓ | []string |
values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch. |
Coordinator custom configurations.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
sgPoolingConfig | ✓ | ✓ | string |
Name of the SGPoolingConfig used for this cluster. Each pod contains a sidecar with a connection pooler (currently: PgBouncer). The connection pooler is implemented as a sidecar.
If not set, a default configuration will be used. Disabling connection pooling altogether is possible if the disableConnectionPooling property of the pods object is set to true. Changing this field may require a restart.
|
|
sgPostgresConfig | ✓ | ✓ | string |
Name of the SGPostgresConfig used for the cluster. It must exist. When not set, a default Postgres config, for the major version selected, is used.
Changing this field may require a restart.
|
This section allows to reference SQL scripts that will be applied to the cluster live.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
continueOnSGScriptError | ✓ | boolean |
If true, when any entry of any SGScript fail will not prevent subsequent SGScript from being executed. By default is false . |
||
scripts | ✓ | []object |
A list of script references that will be executed in sequence.
|
A script reference. Each version of each entry of the script referenced will be executed exactly once following the sequence defined in the referenced script and skipping any script entry that have already been executed.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
id | ✓ | integer |
The id is immutable and must be unique across all the SGScript entries. It is replaced by the operator and is used to identify the SGScript entry. |
||
sgScript | ✓ | string |
A reference to an SGScript |
Metadata information from coordinator cluster created resources.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
annotations | ✓ | object |
Custom Kubernetes annotations to be passed to resources created and managed by StackGres. |
||
labels | ✓ | object |
Custom Kubernetes labels to be passed to resources created and managed by StackGres. |
Custom Kubernetes annotations to be passed to resources created and managed by StackGres.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
allResources | ✓ | map[string]string |
Annotations to attach to any resource created or managed by StackGres. |
||
clusterPods | ✓ | map[string]string |
Annotations to attach to pods created or managed by StackGres. |
||
primaryService | ✓ | map[string]string |
Custom Kubernetes annotations passed to the -primary service. |
||
replicasService | ✓ | map[string]string |
Custom Kubernetes annotations passed to the -replicas service. |
||
services | ✓ | map[string]string |
Annotations to attach to all services created or managed by StackGres. |
Custom Kubernetes labels to be passed to resources created and managed by StackGres.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
clusterPods | ✓ | map[string]string |
Labels to attach to Pods created or managed by StackGres. |
||
services | ✓ | map[string]string |
Labels to attach to Services and Endpoints created or managed by StackGres. |
This section allows to configure the global Postgres replication mode.
The main replication group is implicit and contains the total number of instances less the sum of all instances in other replication groups.
The total number of instances is always specified by .spec.instances
.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
mode | ✓ | string |
The replication mode applied to the whole cluster.
Possible values are:
async When in asynchronous mode the cluster is allowed to lose some committed transactions. When the primary server fails or becomes unavailable for any other reason a sufficiently healthy standby will automatically be promoted to primary. Any transactions that have not been replicated to that standby remain in a “forked timeline” on the primary, and are effectively unrecoverable (the data is still there, but recovering it requires a manual recovery effort by data recovery specialists). sync When in synchronous mode a standby will not be promoted unless it is certain that the standby contains all
transactions that may have returned a successful commit status to client (clients can change the behavior
per transaction using PostgreSQL’s Synchronous mode does not guarantee multi node durability of commits under all circumstances. When no suitable standby is available, primary server will still accept writes, but does not guarantee their replication. When the primary fails in this mode no standby will be promoted. When the host that used to be the primary comes back it will get promoted automatically, unless system administrator performed a manual failover. This behavior makes synchronous mode usable with 2 node clusters. When synchronous mode is used and a standby crashes, commits will block until the primary is switched to standalone mode. Manually shutting down or restarting a standby will not cause a commit service interruption. Standby will signal the primary to release itself from synchronous standby duties before PostgreSQL shutdown is initiated. strict-sync When it is absolutely necessary to guarantee that each write is stored durably on at least two nodes, use the strict
synchronous mode. This mode prevents synchronous replication to be switched off on the primary when no synchronous
standby candidates are available. As a downside, the primary will not be available for writes (unless the Postgres
transaction explicitly turns off Note: Because of the way synchronous replication is implemented in PostgreSQL it is still possible to lose transactions even when using strict synchronous mode. If the PostgreSQL backend is cancelled while waiting to acknowledge replication (as a result of packet cancellation due to client timeout or backend failure) transaction changes become visible for other backends. Such changes are not yet replicated and may be lost in case of standby promotion. sync-all The same as strict-sync-all The same as |
||
syncInstances | ✓ | integer |
Number of synchronous standby instances. Must be less than the total number of instances. It is set to 1 by default.
Only setteable if mode is sync or strict-sync .
Minimum: 1 |
This section allows to configure Postgres features
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
version | ✓ | ✓ | string |
Postgres version used on the cluster. It is either of:
|
|
extensions | ✓ | []object |
StackGres support deploy of extensions at runtime by simply adding an entry to this array. A deployed extension still
requires the creation in a database using the A cluster restart is required for:
Exmaple:
|
||
flavor | string |
Postgres flavor used on the cluster. It is either of:
* `babelfish` will use the [Babelfish for Postgres](https://babelfish-for-postgresql.github.io/babelfish-for-postgresql/).
If not specified then the vanilla Postgres will be used for the cluster. This field can only be set on creation.
|
|||
ssl | ✓ | object |
This section allows to use SSL when connecting to Postgres
|
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
name | ✓ | ✓ | string |
The name of the extension to deploy. |
|
publisher | ✓ | string |
The id of the publisher of the extension to deploy. If not specified com.ongres will be used by default. |
||
repository | ✓ | string |
The repository base URL from where to obtain the extension to deploy.
This section is filled by the operator.
|
||
version | ✓ | string |
The version of the extension to deploy. If not specified version of stable channel will be used by default. |
This section allows to use SSL when connecting to Postgres
Example:
apiVersion: stackgres.io/v1alpha1
kind: SGShardedCluster
metadata:
name: stackgres
spec:
postgres:
ssl:
enabled: true
certificateSecretKeySelector:
name: stackgres-secrets
key: cert
privateKeySecretKeySelector:
name: stackgres-secrets
key: key
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
certificateSecretKeySelector | ✓ | object |
Secret key selector for the certificate or certificate chain used for SSL connections.
|
||
enabled | ✓ | boolean |
Allow to enable SSL for connections to Postgres. By default is true .
If |
||
privateKeySecretKeySelector | ✓ | object |
Secret key selector for the private key used for SSL connections.
|
Secret key selector for the certificate or certificate chain used for SSL connections.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
key | ✓ | ✓ | string |
The key of Secret that contains the certificate or certificate chain for SSL connections
|
|
name | ✓ | ✓ | string |
The name of Secret that contains the certificate or certificate chain for SSL connections
|
Secret key selector for the private key used for SSL connections.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
key | ✓ | ✓ | string |
The key of Secret that contains the private key for SSL connections
|
|
name | ✓ | ✓ | string |
The name of Secret that contains the private key for SSL connections
|
The shards are a group of StackGres clusters where the partitioned data chunks are stored.
When referring to the cluster in the descriptions belove it apply to any shard’s StackGres cluster.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
clusters | ✓ | ✓ | integer |
Number of shard’s StackGres clusters
Minimum: 1 Maximum: 16 |
|
instancesPerCluster | ✓ | ✓ | integer |
Number of StackGres instances per shard’s StackGres cluster. Each instance contains one Postgres server.
Out of all of the Postgres servers, one is elected as the primary, the rest remain as read-only replicas.
Minimum: 1 Maximum: 16 |
|
pods | ✓ | ✓ | object |
Cluster pod’s configuration. |
|
configurations | ✓ | object |
Shards custom configurations.
|
||
managedSql | ✓ | object |
This section allows to reference SQL scripts that will be applied to the cluster live.
|
||
metadata | ✓ | object |
Metadata information from shards cluster created resources. |
||
overrides | ✓ | []object |
Any shard can be overriden by this section.
|
||
replication | ✓ | object |
This section allows to configure the global Postgres replication mode.
The main replication group is implicit and contains the total number of instances less the sum of all instances in other replication groups. The total number of instances is always specified by |
||
sgInstanceProfile | ✓ | ✓ | string |
Name of the SGInstanceProfile.
A SGInstanceProfile defines CPU and memory limits. Must exist before creating a cluster. When no profile is set, a default (1 core, 2 GiB RAM) one is used. Changing this field may require a restart.
|
Cluster pod’s configuration.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
persistentVolume | ✓ | ✓ | object |
Pod’s persistent volume configuration. |
|
customContainers | ✓ | ✓ | []object |
A list of custom application containers that run within the shards cluster’s Pods.
The name used in this section will be prefixed with the string Changing this field may require a restart. See: https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.25/#container-v1-core
|
|
customInitContainers | ✓ | ✓ | []object |
A list of custom application init containers that run within the coordinator cluster’s Pods. The
custom init containers will run following the defined sequence as the end of
cluster’s Pods init containers.
The name used in this section will be prefixed with the string Changing this field may require a restart. See: https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.25/#container-v1-core
|
|
customVolumes | ✓ | ✓ | []object |
A list of custom volumes that may be used along with any container defined in
customInitContainers or customContainers sections for the shards.
The name used in this section will be prefixed with the string Only the following volume types are allowed: configMap, downwardAPI, emptyDir, gitRepo, glusterfs, hostPath, nfs, projected and secret Changing this field may require a restart. See: https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.25/#volume-v1-core
|
|
disableConnectionPooling | ✓ | ✓ | boolean |
If set to true , avoids creating a connection pooling (using PgBouncer) sidecar.
Changing this field may require a restart.
|
|
disableMetricsExporter | ✓ | ✓ | boolean |
If set to true , avoids creating the Prometheus exporter sidecar. Recommended when there’s no intention to use Prometheus for monitoring.
Changing this field may require a restart.
|
|
disablePostgresUtil | ✓ | ✓ | boolean |
If set to true , avoids creating the postgres-util sidecar. This sidecar contains usual Postgres administration utilities that are not present in the main (patroni ) container, like psql . Only disable if you know what you are doing.
Changing this field may require a restart.
|
|
managementPolicy | ✓ | string |
managementPolicy controls how pods are created during initial scale up, when replacing pods
on nodes, or when scaling down. The default policy is OrderedReady , where pods are created
in increasing order (pod-0, then pod-1, etc) and the controller will wait until each pod is
ready before continuing. When scaling down, the pods are removed in the opposite order.
The alternative policy is Parallel which will create pods in parallel to match the desired
scale without waiting, and on scale down will delete all pods at once.
|
||
resources | ✓ | object |
Pod custom resources configuration. |
||
scheduling | ✓ | ✓ | object |
Pod custom scheduling, affinity and topology spread constratins configuration.
Changing this field may require a restart.
|
Pod’s persistent volume configuration.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
size | ✓ | ✓ | string |
Size of the PersistentVolume set for each instance of the cluster. This size is specified either in Mebibytes, Gibibytes or Tebibytes (multiples of 2^20, 2^30 or 2^40, respectively).
|
|
storageClass | ✓ | string |
Name of an existing StorageClass in the Kubernetes cluster, used to create the PersistentVolumes for the instances of the cluster.
|
Pod custom resources configuration.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
enableClusterLimitsRequirements | ✓ | ✓ | boolean |
When enabled resource limits for containers other than the patroni container wil be set just like for patroni contianer as specified in the SGInstanceProfile.
Changing this field may require a restart.
|
Pod custom scheduling, affinity and topology spread constratins configuration.
Changing this field may require a restart.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
backup | ✓ | ✓ | object |
Backup Pod custom scheduling and affinity configuration. |
|
nodeAffinity | ✓ | ✓ | object |
Node affinity is a group of node affinity scheduling rules.
See: https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.25/#nodeaffinity-v1-core
|
|
nodeSelector | ✓ | ✓ | map[string]string |
NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node’s labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/
|
|
podAffinity | ✓ | ✓ | object |
Pod affinity is a group of inter pod affinity scheduling rules.
See: https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.25/#podaffinity-v1-core
|
|
podAntiAffinity | ✓ | ✓ | object |
Pod anti affinity is a group of inter pod anti affinity scheduling rules.
See: https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.25/#podantiaffinity-v1-core
|
|
priorityClassName | ✓ | ✓ | string |
Priority indicates the importance of a Pod relative to other Pods. If a Pod cannot be scheduled, the scheduler tries to preempt (evict) lower priority Pods to make scheduling of the pending Pod possible.
|
|
tolerations | ✓ | ✓ | []object |
If specified, the pod’s tolerations.
See: https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.25/#toleration-v1-core
|
|
topologySpreadConstraints | ✓ | ✓ | []object |
TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed.
|
Backup Pod custom scheduling and affinity configuration.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
nodeAffinity | ✓ | ✓ | object |
Node affinity is a group of node affinity scheduling rules.
See: https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.25/#nodeaffinity-v1-core
|
|
nodeSelector | ✓ | ✓ | object |
Node affinity is a group of node affinity scheduling rules.
See: https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.25/#nodeaffinity-v1-core
|
|
podAffinity | ✓ | ✓ | object |
Pod affinity is a group of inter pod affinity scheduling rules.
See: https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.25/#podaffinity-v1-core
|
|
podAntiAffinity | ✓ | ✓ | object |
Pod anti affinity is a group of inter pod anti affinity scheduling rules.
See: https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.25/#podantiaffinity-v1-core
|
|
priorityClassName | ✓ | ✓ | string |
Priority indicates the importance of a Pod relative to other Pods. If a Pod cannot be scheduled, the scheduler tries to preempt (evict) lower priority Pods to make scheduling of the pending Pod possible.
|
|
tolerations | ✓ | ✓ | object |
Node affinity is a group of node affinity scheduling rules.
See: https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.25/#nodeaffinity-v1-core
|
TopologySpreadConstraint specifies how to spread matching pods among the given topology.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
maxSkew | ✓ | ✓ | ✓ | integer |
MaxSkew describes the degree to which pods may be unevenly distributed. When whenUnsatisfiable=DoNotSchedule , it is the maximum permitted difference between the number of matching pods in the target topology and the global minimum. The global minimum is the minimum number of matching pods in an eligible domain or zero if the number of eligible domains is less than MinDomains. For example, in a 3-zone cluster, MaxSkew is set to 1, and pods with the same labelSelector spread as 2/2/1: In this case, the global minimum is 1. | zone1 | zone2 | zone3 | | P P | P P | P | - if MaxSkew is 1, incoming pod can only be scheduled to zone3 to become 2/2/2; scheduling it onto zone1(zone2) would make the ActualSkew(3-1) on zone1(zone2) violate MaxSkew(1). - if MaxSkew is 2, incoming pod can be scheduled onto any zone. When whenUnsatisfiable=ScheduleAnyway , it is used to give higher precedence to topologies that satisfy it. It’s a required field. Default value is 1 and 0 is not allowed.Format: int32 |
topologyKey | ✓ | ✓ | ✓ | string |
TopologyKey is the key of node labels. Nodes that have a label with this key and identical values are considered to be in the same topology. We consider each <key, value> as a “bucket”, and try to put balanced number of pods into each bucket. We define a domain as a particular instance of a topology. Also, we define an eligible domain as a domain whose nodes meet the requirements of nodeAffinityPolicy and nodeTaintsPolicy. e.g. If TopologyKey is “kubernetes.io/hostname”, each Node is a domain of that topology. And, if TopologyKey is “topology.kubernetes.io/zone”, each zone is a domain of that topology. It’s a required field. |
whenUnsatisfiable | ✓ | ✓ | ✓ | string |
WhenUnsatisfiable indicates how to deal with a pod if it doesn’t satisfy the spread constraint. - DoNotSchedule (default) tells the scheduler not to schedule it. - ScheduleAnyway tells the scheduler to schedule the pod in any location,
but giving higher precedence to topologies that would help reduce the
skew.
A constraint is considered “Unsatisfiable” for an incoming pod if and only if every possible node assignment for that pod would violate “MaxSkew” on some topology. For example, in a 3-zone cluster, MaxSkew is set to 1, and pods with the same labelSelector spread as 3/1/1: | zone1 | zone2 | zone3 | | P P P | P | P | If WhenUnsatisfiable is set to DoNotSchedule, incoming pod can only be scheduled to zone2(zone3) to become 3/2/1(3/1/2) as ActualSkew(2-1) on zone2(zone3) satisfies MaxSkew(1). In other words, the cluster can still be imbalanced, but scheduler won’t make it more imbalanced. It’s a required field.
|
labelSelector | ✓ | ✓ | object |
A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects. |
|
matchLabelKeys | ✓ | ✓ | []string |
MatchLabelKeys is a set of pod label keys to select the pods over which spreading will be calculated. The keys are used to lookup values from the incoming pod labels, those key-value labels are ANDed with labelSelector to select the group of existing pods over which spreading will be calculated for the incoming pod. Keys that don't exist in the incoming pod labels will be ignored. A null or empty list means only match against labelSelector. |
|
minDomains | ✓ | ✓ | integer |
MinDomains indicates a minimum number of eligible domains. When the number of eligible domains with matching topology keys is less than minDomains, Pod Topology Spread treats "global minimum" as 0, and then the calculation of Skew is performed. And when the number of eligible domains with matching topology keys equals or greater than minDomains, this value has no effect on scheduling. As a result, when the number of eligible domains is less than minDomains, scheduler won't schedule more than maxSkew Pods to those domains. If value is nil, the constraint behaves as if MinDomains is equal to 1. Valid values are integers greater than 0. When value is not nil, WhenUnsatisfiable must be DoNotSchedule.
For example, in a 3-zone cluster, MaxSkew is set to 2, MinDomains is set to 5 and pods with the same labelSelector spread as 2/2/2: | zone1 | zone2 | zone3 | | P P | P P | P P | The number of domains is less than 5(MinDomains), so “global minimum” is treated as 0. In this situation, new pod with the same labelSelector cannot be scheduled, because computed skew will be 3(3 - 0) if new Pod is scheduled to any of the three zones, it will violate MaxSkew. This is a beta field and requires the MinDomainsInPodTopologySpread feature gate to be enabled (enabled by default). |
|
nodeAffinityPolicy | ✓ | ✓ | string |
NodeAffinityPolicy indicates how we will treat Pod’s nodeAffinity/nodeSelector when calculating pod topology spread skew. Options are: - Honor: only nodes matching nodeAffinity/nodeSelector are included in the calculations. - Ignore: nodeAffinity/nodeSelector are ignored. All nodes are included in the calculations.
If this value is nil, the behavior is equivalent to the Honor policy. This is a alpha-level feature enabled by the NodeInclusionPolicyInPodTopologySpread feature flag. |
|
nodeTaintsPolicy | ✓ | ✓ | string |
NodeTaintsPolicy indicates how we will treat node taints when calculating pod topology spread skew. Options are: - Honor: nodes without taints, along with tainted nodes for which the incoming pod has a toleration, are included. - Ignore: node taints are ignored. All nodes are included.
If this value is nil, the behavior is equivalent to the Ignore policy. This is a alpha-level feature enabled by the NodeInclusionPolicyInPodTopologySpread feature flag. |
A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
matchExpressions | ✓ | ✓ | []object |
matchExpressions is a list of label selector requirements. The requirements are ANDed. |
|
matchLabels | ✓ | ✓ | map[string]string |
matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is “key”, the operator is “In”, and the values array contains only “value”. The requirements are ANDed. |
A label selector requirement is a selector that contains values, a key, and an operator that relates the key and values.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
key | ✓ | ✓ | ✓ | string |
key is the label key that the selector applies to. |
operator | ✓ | ✓ | ✓ | string |
operator represents a key’s relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist. |
values | ✓ | ✓ | []string |
values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch. |
Shards custom configurations.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
sgPoolingConfig | ✓ | ✓ | string |
Name of the SGPoolingConfig used for this cluster. Each pod contains a sidecar with a connection pooler (currently: PgBouncer). The connection pooler is implemented as a sidecar.
If not set, a default configuration will be used. Disabling connection pooling altogether is possible if the disableConnectionPooling property of the pods object is set to true. Changing this field may require a restart.
|
|
sgPostgresConfig | ✓ | ✓ | string |
Name of the SGPostgresConfig used for the cluster. It must exist. When not set, a default Postgres config, for the major version selected, is used.
Changing this field may require a restart.
|
This section allows to reference SQL scripts that will be applied to the cluster live.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
continueOnSGScriptError | ✓ | boolean |
If true, when any entry of any SGScript fail will not prevent subsequent SGScript from being executed. By default is false . |
||
scripts | ✓ | []object |
A list of script references that will be executed in sequence.
|
A script reference. Each version of each entry of the script referenced will be executed exactly once following the sequence defined in the referenced script and skipping any script entry that have already been executed.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
id | ✓ | integer |
The id is immutable and must be unique across all the SGScript entries. It is replaced by the operator and is used to identify the SGScript entry. |
||
sgScript | ✓ | string |
A reference to an SGScript |
Metadata information from shards cluster created resources.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
annotations | ✓ | object |
Custom Kubernetes annotations to be passed to resources created and managed by StackGres. |
||
labels | ✓ | object |
Custom Kubernetes labels to be passed to resources created and managed by StackGres. |
Custom Kubernetes annotations to be passed to resources created and managed by StackGres.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
allResources | ✓ | map[string]string |
Annotations to attach to any resource created or managed by StackGres. |
||
clusterPods | ✓ | map[string]string |
Annotations to attach to pods created or managed by StackGres. |
||
primaryService | ✓ | map[string]string |
Custom Kubernetes annotations passed to the -primary service. |
||
replicasService | ✓ | map[string]string |
Custom Kubernetes annotations passed to the -replicas service. |
||
services | ✓ | map[string]string |
Annotations to attach to all services created or managed by StackGres. |
Custom Kubernetes labels to be passed to resources created and managed by StackGres.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
clusterPods | ✓ | map[string]string |
Labels to attach to Pods created or managed by StackGres. |
||
services | ✓ | map[string]string |
Labels to attach to Services and Endpoints created or managed by StackGres. |
Any shard can be overriden by this section.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
index | ✓ | ✓ | integer |
Identifier of the shard StackGres cluster to override (starting from 0)
Minimum: 0 Maximum: 15 |
|
configurations | ✓ | object |
Shards custom configurations.
|
||
instancesPerCluster | ✓ | integer |
Number of StackGres instances per shard’s StackGres cluster. Each instance contains one Postgres server.
Out of all of the Postgres servers, one is elected as the primary, the rest remain as read-only replicas.
Minimum: 1 Maximum: 16 |
||
managedSql | ✓ | object |
This section allows to reference SQL scripts that will be applied to the cluster live.
|
||
metadata | ✓ | object |
Metadata information from shards cluster created resources. |
||
pods | ✓ | object |
Cluster pod’s configuration. |
||
replication | ✓ | object |
This section allows to configure the global Postgres replication mode.
The main replication group is implicit and contains the total number of instances less the sum of all instances in other replication groups. The total number of instances is always specified by |
||
sgInstanceProfile | ✓ | string |
Name of the SGInstanceProfile. A SGInstanceProfile defines CPU and memory limits. Must exist before creating a cluster. When no profile is set, a default (currently: 1 core, 2 GiB RAM) one is used.
|
Shards custom configurations.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
sgPoolingConfig | ✓ | string |
Name of the SGPoolingConfig used for this cluster. Each pod contains a sidecar with a connection pooler (currently: PgBouncer). The connection pooler is implemented as a sidecar.
If not set, a default configuration will be used. Disabling connection pooling altogether is possible if the disableConnectionPooling property of the pods object is set to true.
|
||
sgPostgresConfig | ✓ | string |
Name of the SGPostgresConfig used for the cluster. It must exist. When not set, a default Postgres config, for the major version selected, is used.
|
This section allows to reference SQL scripts that will be applied to the cluster live.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
continueOnSGScriptError | ✓ | boolean |
If true, when any entry of any SGScript fail will not prevent subsequent SGScript from being executed. By default is false . |
||
scripts | ✓ | []object |
A list of script references that will be executed in sequence.
|
A script reference. Each version of each entry of the script referenced will be executed exactly once following the sequence defined in the referenced script and skipping any script entry that have already been executed.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
id | ✓ | integer |
The id is immutable and must be unique across all the SGScript entries. It is replaced by the operator and is used to identify the SGScript entry. |
||
sgScript | ✓ | string |
A reference to an SGScript |
Metadata information from shards cluster created resources.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
annotations | ✓ | object |
Custom Kubernetes annotations to be passed to resources created and managed by StackGres. |
||
labels | ✓ | object |
Custom Kubernetes labels to be passed to resources created and managed by StackGres. |
Custom Kubernetes annotations to be passed to resources created and managed by StackGres.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
allResources | ✓ | map[string]string |
Annotations to attach to any resource created or managed by StackGres. |
||
clusterPods | ✓ | map[string]string |
Annotations to attach to pods created or managed by StackGres. |
||
primaryService | ✓ | map[string]string |
Custom Kubernetes annotations passed to the -primary service. |
||
replicasService | ✓ | map[string]string |
Custom Kubernetes annotations passed to the -replicas service. |
||
services | ✓ | map[string]string |
Annotations to attach to all services created or managed by StackGres. |
Custom Kubernetes labels to be passed to resources created and managed by StackGres.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
clusterPods | ✓ | map[string]string |
Labels to attach to Pods created or managed by StackGres. |
||
services | ✓ | map[string]string |
Labels to attach to Services and Endpoints created or managed by StackGres. |
Cluster pod’s configuration.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
customContainers | ✓ | ✓ | []object |
A list of custom application containers that run within the shards cluster’s Pods.
The name used in this section will be prefixed with the string Changing this field may require a restart. See: https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.25/#container-v1-core
|
|
customInitContainers | ✓ | ✓ | []object |
A list of custom application init containers that run within the coordinator cluster’s Pods. The
custom init containers will run following the defined sequence as the end of
cluster’s Pods init containers.
The name used in this section will be prefixed with the string Changing this field may require a restart. See: https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.25/#container-v1-core
|
|
customVolumes | ✓ | ✓ | []object |
A list of custom volumes that may be used along with any container defined in
customInitContainers or customContainers sections for the shards.
The name used in this section will be prefixed with the string Only the following volume types are allowed: configMap, downwardAPI, emptyDir, gitRepo, glusterfs, hostPath, nfs, projected and secret Changing this field may require a restart. See: https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.25/#volume-v1-core
|
|
disableConnectionPooling | ✓ | ✓ | boolean |
If set to true , avoids creating a connection pooling (using PgBouncer) sidecar.
Changing this field may require a restart.
|
|
disableMetricsExporter | ✓ | boolean |
If set to true , avoids creating the Prometheus exporter sidecar. Recommended when there’s no intention to use Prometheus for monitoring. |
||
disablePostgresUtil | ✓ | ✓ | boolean |
If set to true , avoids creating the postgres-util sidecar. This sidecar contains usual Postgres administration utilities that are not present in the main (patroni ) container, like psql . Only disable if you know what you are doing.
Changing this field may require a restart.
|
|
managementPolicy | ✓ | string |
managementPolicy controls how pods are created during initial scale up, when replacing pods
on nodes, or when scaling down. The default policy is OrderedReady , where pods are created
in increasing order (pod-0, then pod-1, etc) and the controller will wait until each pod is
ready before continuing. When scaling down, the pods are removed in the opposite order.
The alternative policy is Parallel which will create pods in parallel to match the desired
scale without waiting, and on scale down will delete all pods at once.
|
||
persistentVolume | ✓ | object |
Pod’s persistent volume configuration. |
||
resources | ✓ | object |
Pod custom resources configuration. |
||
scheduling | ✓ | ✓ | object |
Pod custom scheduling, affinity and topology spread constratins configuration.
Changing this field may require a restart.
|
Pod’s persistent volume configuration.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
size | ✓ | ✓ | string |
Size of the PersistentVolume set for each instance of the cluster. This size is specified either in Mebibytes, Gibibytes or Tebibytes (multiples of 2^20, 2^30 or 2^40, respectively).
|
|
storageClass | ✓ | string |
Name of an existing StorageClass in the Kubernetes cluster, used to create the PersistentVolumes for the instances of the cluster.
|
Pod custom resources configuration.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
enableClusterLimitsRequirements | ✓ | ✓ | boolean |
When enabled resource limits for containers other than the patroni container wil be set just like for patroni contianer as specified in the SGInstanceProfile.
Changing this field may require a restart.
|
Pod custom scheduling, affinity and topology spread constratins configuration.
Changing this field may require a restart.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
backup | ✓ | ✓ | object |
Backup Pod custom scheduling and affinity configuration. |
|
nodeAffinity | ✓ | ✓ | object |
Node affinity is a group of node affinity scheduling rules.
See: https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.25/#nodeaffinity-v1-core
|
|
nodeSelector | ✓ | ✓ | map[string]string |
NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node’s labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/
|
|
podAffinity | ✓ | ✓ | object |
Pod affinity is a group of inter pod affinity scheduling rules.
See: https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.25/#podaffinity-v1-core
|
|
podAntiAffinity | ✓ | ✓ | object |
Pod anti affinity is a group of inter pod anti affinity scheduling rules.
See: https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.25/#podantiaffinity-v1-core
|
|
priorityClassName | ✓ | ✓ | string |
Priority indicates the importance of a Pod relative to other Pods. If a Pod cannot be scheduled, the scheduler tries to preempt (evict) lower priority Pods to make scheduling of the pending Pod possible.
|
|
tolerations | ✓ | ✓ | []object |
If specified, the pod’s tolerations.
See: https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.25/#toleration-v1-core
|
|
topologySpreadConstraints | ✓ | ✓ | []object |
TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed.
|
Backup Pod custom scheduling and affinity configuration.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
nodeAffinity | ✓ | ✓ | object |
Node affinity is a group of node affinity scheduling rules.
See: https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.25/#nodeaffinity-v1-core
|
|
nodeSelector | ✓ | ✓ | object |
Node affinity is a group of node affinity scheduling rules.
See: https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.25/#nodeaffinity-v1-core
|
|
podAffinity | ✓ | ✓ | object |
Pod affinity is a group of inter pod affinity scheduling rules.
See: https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.25/#podaffinity-v1-core
|
|
podAntiAffinity | ✓ | ✓ | object |
Pod anti affinity is a group of inter pod anti affinity scheduling rules.
See: https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.25/#podantiaffinity-v1-core
|
|
priorityClassName | ✓ | ✓ | string |
Priority indicates the importance of a Pod relative to other Pods. If a Pod cannot be scheduled, the scheduler tries to preempt (evict) lower priority Pods to make scheduling of the pending Pod possible.
|
|
tolerations | ✓ | ✓ | object |
Node affinity is a group of node affinity scheduling rules.
See: https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.25/#nodeaffinity-v1-core
|
TopologySpreadConstraint specifies how to spread matching pods among the given topology.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
maxSkew | ✓ | ✓ | ✓ | integer |
MaxSkew describes the degree to which pods may be unevenly distributed. When whenUnsatisfiable=DoNotSchedule , it is the maximum permitted difference between the number of matching pods in the target topology and the global minimum. The global minimum is the minimum number of matching pods in an eligible domain or zero if the number of eligible domains is less than MinDomains. For example, in a 3-zone cluster, MaxSkew is set to 1, and pods with the same labelSelector spread as 2/2/1: In this case, the global minimum is 1. | zone1 | zone2 | zone3 | | P P | P P | P | - if MaxSkew is 1, incoming pod can only be scheduled to zone3 to become 2/2/2; scheduling it onto zone1(zone2) would make the ActualSkew(3-1) on zone1(zone2) violate MaxSkew(1). - if MaxSkew is 2, incoming pod can be scheduled onto any zone. When whenUnsatisfiable=ScheduleAnyway , it is used to give higher precedence to topologies that satisfy it. It’s a required field. Default value is 1 and 0 is not allowed.Format: int32 |
topologyKey | ✓ | ✓ | ✓ | string |
TopologyKey is the key of node labels. Nodes that have a label with this key and identical values are considered to be in the same topology. We consider each <key, value> as a “bucket”, and try to put balanced number of pods into each bucket. We define a domain as a particular instance of a topology. Also, we define an eligible domain as a domain whose nodes meet the requirements of nodeAffinityPolicy and nodeTaintsPolicy. e.g. If TopologyKey is “kubernetes.io/hostname”, each Node is a domain of that topology. And, if TopologyKey is “topology.kubernetes.io/zone”, each zone is a domain of that topology. It’s a required field. |
whenUnsatisfiable | ✓ | ✓ | ✓ | string |
WhenUnsatisfiable indicates how to deal with a pod if it doesn’t satisfy the spread constraint. - DoNotSchedule (default) tells the scheduler not to schedule it. - ScheduleAnyway tells the scheduler to schedule the pod in any location,
but giving higher precedence to topologies that would help reduce the
skew.
A constraint is considered “Unsatisfiable” for an incoming pod if and only if every possible node assignment for that pod would violate “MaxSkew” on some topology. For example, in a 3-zone cluster, MaxSkew is set to 1, and pods with the same labelSelector spread as 3/1/1: | zone1 | zone2 | zone3 | | P P P | P | P | If WhenUnsatisfiable is set to DoNotSchedule, incoming pod can only be scheduled to zone2(zone3) to become 3/2/1(3/1/2) as ActualSkew(2-1) on zone2(zone3) satisfies MaxSkew(1). In other words, the cluster can still be imbalanced, but scheduler won’t make it more imbalanced. It’s a required field.
|
labelSelector | ✓ | ✓ | object |
A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects. |
|
matchLabelKeys | ✓ | ✓ | []string |
MatchLabelKeys is a set of pod label keys to select the pods over which spreading will be calculated. The keys are used to lookup values from the incoming pod labels, those key-value labels are ANDed with labelSelector to select the group of existing pods over which spreading will be calculated for the incoming pod. Keys that don't exist in the incoming pod labels will be ignored. A null or empty list means only match against labelSelector. |
|
minDomains | ✓ | ✓ | integer |
MinDomains indicates a minimum number of eligible domains. When the number of eligible domains with matching topology keys is less than minDomains, Pod Topology Spread treats "global minimum" as 0, and then the calculation of Skew is performed. And when the number of eligible domains with matching topology keys equals or greater than minDomains, this value has no effect on scheduling. As a result, when the number of eligible domains is less than minDomains, scheduler won't schedule more than maxSkew Pods to those domains. If value is nil, the constraint behaves as if MinDomains is equal to 1. Valid values are integers greater than 0. When value is not nil, WhenUnsatisfiable must be DoNotSchedule.
For example, in a 3-zone cluster, MaxSkew is set to 2, MinDomains is set to 5 and pods with the same labelSelector spread as 2/2/2: | zone1 | zone2 | zone3 | | P P | P P | P P | The number of domains is less than 5(MinDomains), so “global minimum” is treated as 0. In this situation, new pod with the same labelSelector cannot be scheduled, because computed skew will be 3(3 - 0) if new Pod is scheduled to any of the three zones, it will violate MaxSkew. This is a beta field and requires the MinDomainsInPodTopologySpread feature gate to be enabled (enabled by default). |
|
nodeAffinityPolicy | ✓ | ✓ | string |
NodeAffinityPolicy indicates how we will treat Pod’s nodeAffinity/nodeSelector when calculating pod topology spread skew. Options are: - Honor: only nodes matching nodeAffinity/nodeSelector are included in the calculations. - Ignore: nodeAffinity/nodeSelector are ignored. All nodes are included in the calculations.
If this value is nil, the behavior is equivalent to the Honor policy. This is a alpha-level feature enabled by the NodeInclusionPolicyInPodTopologySpread feature flag. |
|
nodeTaintsPolicy | ✓ | ✓ | string |
NodeTaintsPolicy indicates how we will treat node taints when calculating pod topology spread skew. Options are: - Honor: nodes without taints, along with tainted nodes for which the incoming pod has a toleration, are included. - Ignore: node taints are ignored. All nodes are included.
If this value is nil, the behavior is equivalent to the Ignore policy. This is a alpha-level feature enabled by the NodeInclusionPolicyInPodTopologySpread feature flag. |
A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
matchExpressions | ✓ | ✓ | []object |
matchExpressions is a list of label selector requirements. The requirements are ANDed. |
|
matchLabels | ✓ | ✓ | map[string]string |
matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is “key”, the operator is “In”, and the values array contains only “value”. The requirements are ANDed. |
A label selector requirement is a selector that contains values, a key, and an operator that relates the key and values.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
key | ✓ | ✓ | ✓ | string |
key is the label key that the selector applies to. |
operator | ✓ | ✓ | ✓ | string |
operator represents a key’s relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist. |
values | ✓ | ✓ | []string |
values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch. |
This section allows to configure the global Postgres replication mode.
The main replication group is implicit and contains the total number of instances less the sum of all instances in other replication groups.
The total number of instances is always specified by .spec.instances
.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
mode | ✓ | string |
The replication mode applied to the whole cluster.
Possible values are:
async When in asynchronous mode the cluster is allowed to lose some committed transactions. When the primary server fails or becomes unavailable for any other reason a sufficiently healthy standby will automatically be promoted to primary. Any transactions that have not been replicated to that standby remain in a “forked timeline” on the primary, and are effectively unrecoverable (the data is still there, but recovering it requires a manual recovery effort by data recovery specialists). sync When in synchronous mode a standby will not be promoted unless it is certain that the standby contains all
transactions that may have returned a successful commit status to client (clients can change the behavior
per transaction using PostgreSQL’s Synchronous mode does not guarantee multi node durability of commits under all circumstances. When no suitable standby is available, primary server will still accept writes, but does not guarantee their replication. When the primary fails in this mode no standby will be promoted. When the host that used to be the primary comes back it will get promoted automatically, unless system administrator performed a manual failover. This behavior makes synchronous mode usable with 2 node clusters. When synchronous mode is used and a standby crashes, commits will block until the primary is switched to standalone mode. Manually shutting down or restarting a standby will not cause a commit service interruption. Standby will signal the primary to release itself from synchronous standby duties before PostgreSQL shutdown is initiated. strict-sync When it is absolutely necessary to guarantee that each write is stored durably on at least two nodes, use the strict
synchronous mode. This mode prevents synchronous replication to be switched off on the primary when no synchronous
standby candidates are available. As a downside, the primary will not be available for writes (unless the Postgres
transaction explicitly turns off Note: Because of the way synchronous replication is implemented in PostgreSQL it is still possible to lose transactions even when using strict synchronous mode. If the PostgreSQL backend is cancelled while waiting to acknowledge replication (as a result of packet cancellation due to client timeout or backend failure) transaction changes become visible for other backends. Such changes are not yet replicated and may be lost in case of standby promotion. sync-all The same as strict-sync-all The same as |
||
syncInstances | ✓ | integer |
Number of synchronous standby instances. Must be less than the total number of instances. It is set to 1 by default.
Only setteable if mode is sync or strict-sync .
Minimum: 1 |
This section allows to configure the global Postgres replication mode.
The main replication group is implicit and contains the total number of instances less the sum of all instances in other replication groups.
The total number of instances is always specified by .spec.instances
.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
mode | ✓ | string |
The replication mode applied to the whole cluster.
Possible values are:
async When in asynchronous mode the cluster is allowed to lose some committed transactions. When the primary server fails or becomes unavailable for any other reason a sufficiently healthy standby will automatically be promoted to primary. Any transactions that have not been replicated to that standby remain in a “forked timeline” on the primary, and are effectively unrecoverable (the data is still there, but recovering it requires a manual recovery effort by data recovery specialists). sync When in synchronous mode a standby will not be promoted unless it is certain that the standby contains all
transactions that may have returned a successful commit status to client (clients can change the behavior
per transaction using PostgreSQL’s Synchronous mode does not guarantee multi node durability of commits under all circumstances. When no suitable standby is available, primary server will still accept writes, but does not guarantee their replication. When the primary fails in this mode no standby will be promoted. When the host that used to be the primary comes back it will get promoted automatically, unless system administrator performed a manual failover. This behavior makes synchronous mode usable with 2 node clusters. When synchronous mode is used and a standby crashes, commits will block until the primary is switched to standalone mode. Manually shutting down or restarting a standby will not cause a commit service interruption. Standby will signal the primary to release itself from synchronous standby duties before PostgreSQL shutdown is initiated. strict-sync When it is absolutely necessary to guarantee that each write is stored durably on at least two nodes, use the strict
synchronous mode. This mode prevents synchronous replication to be switched off on the primary when no synchronous
standby candidates are available. As a downside, the primary will not be available for writes (unless the Postgres
transaction explicitly turns off Note: Because of the way synchronous replication is implemented in PostgreSQL it is still possible to lose transactions even when using strict synchronous mode. If the PostgreSQL backend is cancelled while waiting to acknowledge replication (as a result of packet cancellation due to client timeout or backend failure) transaction changes become visible for other backends. Such changes are not yet replicated and may be lost in case of standby promotion. sync-all The same as strict-sync-all The same as |
||
syncInstances | ✓ | integer |
Number of synchronous standby instances. Must be less than the total number of instances. It is set to 1 by default.
Only setteable if mode is sync or strict-sync .
Minimum: 1 |
Sharded cluster custom configurations.
Example:
apiVersion: stackgres.io/v1alpha1
kind: SGShardedCluster
metadata:
name: stackgres
spec:
configurations:
backups:
- sgObjectStorage: 'backupconf'
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
backups | ✓ | []object |
List of backups configurations for this SGCluster
|
||
credentials | ✓ | ✓ | object |
Allow to specify custom credentials for Postgres users and Patroni REST API
Changing this field may require a restart.
|
Backup configuration for this SGCluster
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
sgObjectStorage | ✓ | ✓ | string |
Name of the SGObjectStorage to use for the cluster. It defines the location in which the the backups will be stored.
|
|
compression | ✓ | enum |
Specifies the backup compression algorithm. Possible options are: lz4, lzma, brotli. The default method is lz4 . LZ4 is the fastest method, but compression ratio is the worst. LZMA is way slower, but it compresses backups about 6 times better than LZ4. Brotli is a good trade-off between speed and compression ratio, being about 3 times better than LZ4.
Enum: lz4, lzma, brotli |
||
cronSchedule | ✓ | string |
Continuous Archiving backups are composed of periodic base backups and all the WAL segments produced in between those base backups. This parameter specifies at what time and with what frequency to start performing a new base backup.
Use cron syntax (
Also ranges of values ( If not set, full backups are performed each day at 05:00 UTC.
|
||
paths | ✓ | []string |
The paths were the backups are stored. If not set this field is filled up by the operator.
When provided will indicate were the backups and WAL files will be stored. The first path indicate the coordinator path and the other paths indicate the shards paths
|
||
performance | ✓ | object |
Configuration that affects the backup network and disk usage performance.
|
||
retention | ✓ | integer |
When an automatic retention policy is defined to delete old base backups, this parameter specifies the number of base backups to keep, in a sliding window.
Consequently, the time range covered by backups is Default is 5.
|
Configuration that affects the backup network and disk usage performance.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
downloadConcurrency | ✓ | integer |
Backup storage may use several concurrent streams to read the data. This parameter configures the number of parallel streams to use. By default, it’s set to the minimum between the number of file to read and 10.
Minimum: 1 |
||
maxDiskBandwidth | ✓ | integer |
Maximum disk read I/O when performing a backup. In bytes (per second).
|
||
maxNetworkBandwidth | ✓ | integer |
Maximum storage upload bandwidth used when storing a backup. In bytes (per second).
|
||
uploadConcurrency | ✓ | integer |
Backup storage may use several concurrent streams to store the data. This parameter configures the number of parallel streams to use. By default, it’s set to 16.
Minimum: 1 |
||
uploadDiskConcurrency | ✓ | integer |
Backup storage may use several concurrent streams to store the data. This parameter configures the number of parallel streams to use to reading from disk. By default, it’s set to 1.
Minimum: 1 |
Allow to specify custom credentials for Postgres users and Patroni REST API
Changing this field may require a restart.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
patroni | ✓ | ✓ | object |
Kubernetes SecretKeySelectors that contains the credentials for patroni REST API.
Changing this field may require a restart.
|
|
users | ✓ | ✓ | object |
Kubernetes SecretKeySelectors that contains the credentials of the users.
Changing this field may require a manual modification of the database users to reflect the new values specified. In particular you may have to create those users if username is changed or alter password if it is changed. Here are the SQL commands to perform such operation (replace
default usernames with the new ones and
Changing this field may require a restart.
|
Kubernetes SecretKeySelectors that contains the credentials for patroni REST API.
Changing this field may require a restart.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
restApiPassword | ✓ | ✓ | object |
A Kubernetes SecretKeySelector that contains the password for the patroni REST API.
|
A Kubernetes SecretKeySelector that contains the password for the patroni REST API.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
key | ✓ | ✓ | ✓ | string |
The key of the secret to select from. Must be a valid secret key. |
name | ✓ | ✓ | ✓ | string |
Name of the referent. More information. |
Kubernetes SecretKeySelectors that contains the credentials of the users.
Changing this field may require a manual modification of the database users to reflect the new values specified.
In particular you may have to create those users if username is changed or alter password if it is changed. Here are the SQL commands to perform such operation (replace
default usernames with the new ones and ***
with their respective passwords):
CREATE ROLE postgres;
ALTER ROLE postgres WITH SUPERUSER INHERIT CREATEROLE CREATEDB LOGIN REPLICATION BYPASSRLS PASSWORD '***';
CREATE ROLE replicator;
ALTER ROLE replicator WITH NOSUPERUSER INHERIT NOCREATEROLE NOCREATEDB LOGIN REPLICATION NOBYPASSRLS PASSWORD '***';
CREATE ROLE authenticator;
ALTER ROLE authenticator WITH SUPERUSER INHERIT NOCREATEROLE NOCREATEDB LOGIN NOREPLICATION NOBYPASSRLS PASSWORD '***';
Changing this field may require a restart.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
authenticator | ✓ | ✓ | object |
A Kubernetes SecretKeySelector that contains the credentials of the authenticator user used by pgbouncer to authenticate other users.
|
|
replication | ✓ | ✓ | object |
A Kubernetes SecretKeySelector that contains the credentials of the replication user used to replicate from the primary cluster and from replicas of this cluster.
|
|
superuser | ✓ | ✓ | object |
A Kubernetes SecretKeySelector that contains the credentials of the superuser (usually the postgres user).
|
A Kubernetes SecretKeySelector that contains the credentials of the authenticator user used by pgbouncer to authenticate other users.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
password | ✓ | ✓ | object |
A Kubernetes SecretKeySelector that contains the password of the user.
|
|
username | ✓ | ✓ | object |
A Kubernetes SecretKeySelector that contains the username of the user.
|
A Kubernetes SecretKeySelector that contains the password of the user.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
key | ✓ | ✓ | ✓ | string |
The key of the secret to select from. Must be a valid secret key. |
name | ✓ | ✓ | ✓ | string |
Name of the referent. More information. |
A Kubernetes SecretKeySelector that contains the username of the user.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
key | ✓ | ✓ | ✓ | string |
The key of the secret to select from. Must be a valid secret key. |
name | ✓ | ✓ | ✓ | string |
Name of the referent. More information. |
A Kubernetes SecretKeySelector that contains the credentials of the replication user used to replicate from the primary cluster and from replicas of this cluster.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
password | ✓ | ✓ | object |
A Kubernetes SecretKeySelector that contains the password of the user.
|
|
username | ✓ | ✓ | object |
A Kubernetes SecretKeySelector that contains the username of the user.
|
A Kubernetes SecretKeySelector that contains the password of the user.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
key | ✓ | ✓ | ✓ | string |
The key of the secret to select from. Must be a valid secret key. |
name | ✓ | ✓ | ✓ | string |
Name of the referent. More information. |
A Kubernetes SecretKeySelector that contains the username of the user.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
key | ✓ | ✓ | ✓ | string |
The key of the secret to select from. Must be a valid secret key. |
name | ✓ | ✓ | ✓ | string |
Name of the referent. More information. |
A Kubernetes SecretKeySelector that contains the credentials of the superuser (usually the postgres user).
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
password | ✓ | ✓ | object |
A Kubernetes SecretKeySelector that contains the password of the user.
|
|
username | ✓ | ✓ | object |
A Kubernetes SecretKeySelector that contains the username of the user.
|
A Kubernetes SecretKeySelector that contains the password of the user.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
key | ✓ | ✓ | ✓ | string |
The key of the secret to select from. Must be a valid secret key. |
name | ✓ | ✓ | ✓ | string |
Name of the referent. More information. |
A Kubernetes SecretKeySelector that contains the username of the user.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
key | ✓ | ✓ | ✓ | string |
The key of the secret to select from. Must be a valid secret key. |
name | ✓ | ✓ | ✓ | string |
Name of the referent. More information. |
StackGres features a functionality for all pods to send Postgres, Patroni and PgBouncer logs to a central (distributed) location, which is in turn another Postgres database. Logs can then be accessed via SQL interface or from the web UI. This section controls whether to enable this feature or not. If not enabled, logs are send to the pod’s standard output.
Example:
apiVersion: stackgres.io/v1alpha1
kind: SGShardedCluster
metadata:
name: stackgres
spec:
distributedLogs:
sgDistributedLogs: distributedlogs
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
retention | ✓ | string |
Define a retention window with the syntax <integer> (minutes|hours|days|months) in which log entries are kept.
Log entries will be removed when they get older more than the double of the specified retention window.
When this field is changed the retention will be applied only to log entries that are newer than the end of
the retention window previously specified. If no retention window was previously specified it is considered
to be of 7 days. This means that if previous retention window is of |
||
sgDistributedLogs | ✓ | string |
Name of the SGDistributedLogs to use for this cluster. It must exist.
|
Metadata information from any cluster created resources.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
annotations | ✓ | object |
Custom Kubernetes annotations to be passed to resources created and managed by StackGres.
|
||
labels | ✓ | object |
Custom Kubernetes [labels](https://kubernetes.io/docs/concepts/overview/working-with-objects/labels/) to be passed to resources created and managed by StackGres.
|
Custom Kubernetes annotations to be passed to resources created and managed by StackGres.
Example:
apiVersion: stackgres.io/v1alpha1
kind: SGShardedCluster
metadata:
name: stackgres
spec:
metadata:
annotations:
clusterPods:
customAnnotations: customAnnotationValue
primaryService:
customAnnotations: customAnnotationValue
replicasService:
customAnnotations: customAnnotationValue
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
allResources | ✓ | map[string]string |
Annotations to attach to any resource created or managed by StackGres. |
||
clusterPods | ✓ | map[string]string |
Annotations to attach to pods created or managed by StackGres. |
||
primaryService | ✓ | map[string]string |
Custom Kubernetes annotations passed to the -primary service. |
||
replicasService | ✓ | map[string]string |
Custom Kubernetes annotations passed to the -replicas service. |
||
services | ✓ | map[string]string |
Annotations to attach to all services created or managed by StackGres. |
Custom Kubernetes labels to be passed to resources created and managed by StackGres.
Example:
apiVersion: stackgres.io/v1alpha1
kind: SGShardedCluster
metadata:
name: stackgres
spec:
metadata:
labels:
clusterPods:
customLabel: customLabelValue
services:
customLabel: customLabelValue
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
clusterPods | ✓ | map[string]string |
Labels to attach to Pods created or managed by StackGres. |
||
services | ✓ | map[string]string |
Labels to attach to Services and Endpoints created or managed by StackGres. |
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
disableClusterPodAntiAffinity | ✓ | boolean |
It is a best practice, on non-containerized environments, when running production workloads, to run each database server on a different server (virtual or physical), i.e., not to co-locate more than one database server per host.
The same best practice applies to databases on containers. By default, StackGres will not allow to run more than one StackGres pod on a given Kubernetes node. Set this property to true to allow more than one StackGres pod per node.
|
||
disableClusterResourceRequirements | ✓ | boolean |
It is a best practice, on containerized environments, when running production workloads, to enforce container’s resources requirements.
By default, StackGres will configure resource requirements for all the containers. Set this property to true to prevent StackGres from setting container’s resources requirements (except for patroni container, see |
||
disablePatroniResourceRequirements | ✓ | boolean |
It is a best practice, on containerized environments, when running production workloads, to enforce container’s resources requirements.
The same best practice applies to databases on containers. By default, StackGres will configure resource requirements for patroni container. Set this property to true to prevent StackGres from setting patroni container’s resources requirement.
|
||
enableSetClusterCpuRequests | ✓ | boolean |
On containerized environments, when running production workloads, enforcing container’s cpu requirements request to be equals to the limit allow to achieve the highest level of performance. Doing so, reduces the chances of leaving
the workload with less cpu than it requires. It also allow to set static CPU management policy that allows to guarantee a pod the usage exclusive CPUs on the node.
By default, StackGres will configure cpu requirements to have the same limit and request for all the containers. Set this property to true to prevent StackGres from setting container’s cpu requirements request equals to the limit (except for patroni container, see |
||
enableSetClusterMemoryRequests | ✓ | boolean |
On containerized environments, when running production workloads, enforcing container’s memory requirements request to be equals to the limit allow to achieve the highest level of performance. Doing so, reduces the chances of leaving
the workload with less memory than it requires.
By default, StackGres will configure memory requirements to have the same limit and request for all the containers. Set this property to true to prevent StackGres from setting container’s memory requirements request equals to the limit (except for patroni container, see |
||
enableSetPatroniCpuRequests | ✓ | boolean |
On containerized environments, when running production workloads, enforcing container’s cpu requirements request to be equals to the limit allow to achieve the highest level of performance. Doing so, reduces the chances of leaving
the workload with less cpu than it requires. It also allow to set static CPU management policy that allows to guarantee a pod the usage exclusive CPUs on the node.
By default, StackGres will configure cpu requirements to have the same limit and request for the patroni container. Set this property to true to prevent StackGres from setting patroni container’s cpu requirements request equals to the limit
when |
||
enableSetPatroniMemoryRequests | ✓ | boolean |
On containerized environments, when running production workloads, enforcing container’s memory requirements request to be equals to the limit allow to achieve the highest level of performance. Doing so, reduces the chances of leaving
the workload with less memory than it requires.
By default, StackGres will configure memory requirements to have the same limit and request for the patroni container. Set this property to true to prevent StackGres from setting patroni container’s memory requirements request equals to the limit
when |
||
enabledFeatureGates | ✓ | []string |
A list of StackGres feature gates to enable (not suitable for a production environment).
Available feature gates are:
|
Kubernetes services created or managed by StackGres.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
coordinator | ✓ | object |
Configuration for the coordinator services
|
||
shards | ✓ | object |
Configuration for the shards services
|
Configuration for the coordinator services
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
any | ✓ | object |
Configure the coordinator service to any instance of the coordinator with the same name as the SGShardedCluster plus the -reads suffix.
It provides a stable connection (regardless of node failures) to any Postgres server of the coordinator cluster. Servers are load-balanced via this service. See also https://kubernetes.io/docs/concepts/services-networking/service/
|
||
customPorts | ✓ | ✓ | []object |
The list of custom ports that will be exposed by the coordinator services.
The names of custom ports will be prefixed with the string The names of target ports will be prefixed with the string Changing this field may require a restart. See: https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.25/#serviceport-v1-core
|
|
primary | ✓ | object |
Configure the coordinator service to the primary of the coordinator with the name as the SGShardedCluster.
It provides a stable connection (regardless of primary failures or switchovers) to the read-write Postgres server of the coordinator cluster. See also https://kubernetes.io/docs/concepts/services-networking/service/
|
Configure the coordinator service to any instance of the coordinator with the same name as the SGShardedCluster plus the -reads
suffix.
It provides a stable connection (regardless of node failures) to any Postgres server of the coordinator cluster. Servers are load-balanced via this service.
See also https://kubernetes.io/docs/concepts/services-networking/service/
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
allocateLoadBalancerNodePorts | ✓ | boolean |
allocateLoadBalancerNodePorts defines if NodePorts will be automatically allocated for services with type LoadBalancer. Default is “true”. It may be set to “false” if the cluster load-balancer does not rely on NodePorts. If the caller requests specific NodePorts (by specifying a value), those requests will be respected, regardless of this field. This field may only be set for services with type LoadBalancer and will be cleared if the type is changed to any other type. |
||
enabled | ✓ | boolean |
Specify if the service should be created or not. |
||
externalIPs | ✓ | []string |
externalIPs is a list of IP addresses for which nodes in the cluster will also accept traffic for this service. These IPs are not managed by Kubernetes. The user is responsible for ensuring that traffic arrives at a node with this IP. A common example is external load-balancers that are not part of the Kubernetes system. |
||
externalTrafficPolicy | ✓ | string |
externalTrafficPolicy describes how nodes distribute service traffic they receive on one of the Service’s “externally-facing” addresses (NodePorts, ExternalIPs, and LoadBalancer IPs). If set to “Local”, the proxy will configure the service in a way that assumes that external load balancers will take care of balancing the service traffic between nodes, and so each node will deliver traffic only to the node-local endpoints of the service, without masquerading the client source IP. (Traffic mistakenly sent to a node with no endpoints will be dropped.) The default value, “Cluster”, uses the standard behavior of routing to all endpoints evenly (possibly modified by topology and other features). Note that traffic sent to an External IP or LoadBalancer IP from within the cluster will always get “Cluster” semantics, but clients sending to a NodePort from within the cluster may need to take traffic policy into account when picking a node.
|
||
healthCheckNodePort | ✓ | integer |
healthCheckNodePort specifies the healthcheck nodePort for the service. This only applies when type is set to LoadBalancer and externalTrafficPolicy is set to Local. If a value is specified, is in-range, and is not in use, it will be used. If not specified, a value will be automatically allocated. External systems (e.g. load-balancers) can use this port to determine if a given node holds endpoints for this service or not. If this field is specified when creating a Service which does not need it, creation will fail. This field will be wiped when updating a Service to no longer need it (e.g. changing type). This field cannot be updated once set. Format: int32 |
||
internalTrafficPolicy | ✓ | string |
InternalTrafficPolicy describes how nodes distribute service traffic they receive on the ClusterIP. If set to "Local", the proxy will assume that pods only want to talk to endpoints of the service on the same node as the pod, dropping the traffic if there are no local endpoints. The default value, "Cluster", uses the standard behavior of routing to all endpoints evenly (possibly modified by topology and other features). |
||
ipFamilies | ✓ | []string |
IPFamilies is a list of IP families (e.g. IPv4, IPv6) assigned to this service. This field is usually assigned automatically based on cluster configuration and the ipFamilyPolicy field. If this field is specified manually, the requested family is available in the cluster, and ipFamilyPolicy allows it, it will be used; otherwise creation of the service will fail. This field is conditionally mutable: it allows for adding or removing a secondary IP family, but it does not allow changing the primary IP family of the Service. Valid values are "IPv4" and "IPv6". This field only applies to Services of types ClusterIP, NodePort, and LoadBalancer, and does apply to "headless" services. This field will be wiped when updating a Service to type ExternalName.
This field may hold a maximum of two entries (dual-stack families, in either order). These families must correspond to the values of the clusterIPs field, if specified. Both clusterIPs and ipFamilies are governed by the ipFamilyPolicy field. |
||
ipFamilyPolicy | ✓ | string |
IPFamilyPolicy represents the dual-stack-ness requested or required by this Service. If there is no value provided, then this field will be set to SingleStack. Services can be “SingleStack” (a single IP family), “PreferDualStack” (two IP families on dual-stack configured clusters or a single IP family on single-stack clusters), or “RequireDualStack” (two IP families on dual-stack configured clusters, otherwise fail). The ipFamilies and clusterIPs fields depend on the value of this field. This field will be wiped when updating a service to type ExternalName. |
||
loadBalancerClass | ✓ | string |
loadBalancerClass is the class of the load balancer implementation this Service belongs to. If specified, the value of this field must be a label-style identifier, with an optional prefix, e.g. “internal-vip” or “example.com/internal-vip”. Unprefixed names are reserved for end-users. This field can only be set when the Service type is ‘LoadBalancer’. If not set, the default load balancer implementation is used, today this is typically done through the cloud provider integration, but should apply for any default implementation. If set, it is assumed that a load balancer implementation is watching for Services with a matching class. Any default load balancer implementation (e.g. cloud providers) should ignore Services that set this field. This field can only be set when creating or updating a Service to type ‘LoadBalancer’. Once set, it can not be changed. This field will be wiped when a service is updated to a non ‘LoadBalancer’ type. |
||
loadBalancerIP | ✓ | string |
Only applies to Service Type: LoadBalancer. This feature depends on whether the underlying cloud-provider supports specifying the loadBalancerIP when a load balancer is created. This field will be ignored if the cloud-provider does not support the feature. Deprecated: This field was under-specified and its meaning varies across implementations, and it cannot support dual-stack. As of Kubernetes v1.24, users are encouraged to use implementation-specific annotations when available. This field may be removed in a future API version. |
||
loadBalancerSourceRanges | ✓ | []string |
If specified and supported by the platform, this will restrict traffic through the cloud-provider load-balancer will be restricted to the specified client IPs. This field will be ignored if the cloud-provider does not support the feature.” More info: https://kubernetes.io/docs/tasks/access-application-cluster/create-external-load-balancer/ |
||
sessionAffinity | ✓ | string |
Supports “ClientIP” and “None”. Used to maintain session affinity. Enable client IP based session affinity. Must be ClientIP or None. Defaults to None. More info: https://kubernetes.io/docs/concepts/services-networking/service/#virtual-ips-and-service-proxies
|
||
sessionAffinityConfig | ✓ | object |
SessionAffinityConfig represents the configurations of session affinity. |
||
type | ✓ | enum |
type determines how the Service is exposed. Defaults to ClusterIP. Valid
options are ClusterIP, NodePort, and LoadBalancer. "ClusterIP" allocates
a cluster-internal IP address for load-balancing to endpoints.
"NodePort" builds on ClusterIP and allocates a port on every node.
"LoadBalancer" builds on NodePort and creates
an external load-balancer (if supported in the current cloud).
More info: https://kubernetes.io/docs/concepts/services-networking/service/#publishing-services-service-types
Enum: ClusterIP, LoadBalancer, NodePort |
SessionAffinityConfig represents the configurations of session affinity.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
clientIP | ✓ | object |
ClientIPConfig represents the configurations of Client IP based session affinity. |
ClientIPConfig represents the configurations of Client IP based session affinity.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
timeoutSeconds | ✓ | integer |
timeoutSeconds specifies the seconds of ClientIP type session sticky time. The value must be >0 && <=86400(for 1 day) if ServiceAffinity == “ClientIP”. Default value is 10800(for 3 hours). Format: int32 |
Configure the coordinator service to the primary of the coordinator with the name as the SGShardedCluster.
It provides a stable connection (regardless of primary failures or switchovers) to the read-write Postgres server of the coordinator cluster.
See also https://kubernetes.io/docs/concepts/services-networking/service/
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
allocateLoadBalancerNodePorts | ✓ | boolean |
allocateLoadBalancerNodePorts defines if NodePorts will be automatically allocated for services with type LoadBalancer. Default is “true”. It may be set to “false” if the cluster load-balancer does not rely on NodePorts. If the caller requests specific NodePorts (by specifying a value), those requests will be respected, regardless of this field. This field may only be set for services with type LoadBalancer and will be cleared if the type is changed to any other type. |
||
enabled | ✓ | boolean |
Specify if the service should be created or not. |
||
externalIPs | ✓ | []string |
externalIPs is a list of IP addresses for which nodes in the cluster will also accept traffic for this service. These IPs are not managed by Kubernetes. The user is responsible for ensuring that traffic arrives at a node with this IP. A common example is external load-balancers that are not part of the Kubernetes system. |
||
externalTrafficPolicy | ✓ | string |
externalTrafficPolicy describes how nodes distribute service traffic they receive on one of the Service’s “externally-facing” addresses (NodePorts, ExternalIPs, and LoadBalancer IPs). If set to “Local”, the proxy will configure the service in a way that assumes that external load balancers will take care of balancing the service traffic between nodes, and so each node will deliver traffic only to the node-local endpoints of the service, without masquerading the client source IP. (Traffic mistakenly sent to a node with no endpoints will be dropped.) The default value, “Cluster”, uses the standard behavior of routing to all endpoints evenly (possibly modified by topology and other features). Note that traffic sent to an External IP or LoadBalancer IP from within the cluster will always get “Cluster” semantics, but clients sending to a NodePort from within the cluster may need to take traffic policy into account when picking a node.
|
||
healthCheckNodePort | ✓ | integer |
healthCheckNodePort specifies the healthcheck nodePort for the service. This only applies when type is set to LoadBalancer and externalTrafficPolicy is set to Local. If a value is specified, is in-range, and is not in use, it will be used. If not specified, a value will be automatically allocated. External systems (e.g. load-balancers) can use this port to determine if a given node holds endpoints for this service or not. If this field is specified when creating a Service which does not need it, creation will fail. This field will be wiped when updating a Service to no longer need it (e.g. changing type). This field cannot be updated once set. Format: int32 |
||
internalTrafficPolicy | ✓ | string |
InternalTrafficPolicy describes how nodes distribute service traffic they receive on the ClusterIP. If set to "Local", the proxy will assume that pods only want to talk to endpoints of the service on the same node as the pod, dropping the traffic if there are no local endpoints. The default value, "Cluster", uses the standard behavior of routing to all endpoints evenly (possibly modified by topology and other features). |
||
ipFamilies | ✓ | []string |
IPFamilies is a list of IP families (e.g. IPv4, IPv6) assigned to this service. This field is usually assigned automatically based on cluster configuration and the ipFamilyPolicy field. If this field is specified manually, the requested family is available in the cluster, and ipFamilyPolicy allows it, it will be used; otherwise creation of the service will fail. This field is conditionally mutable: it allows for adding or removing a secondary IP family, but it does not allow changing the primary IP family of the Service. Valid values are "IPv4" and "IPv6". This field only applies to Services of types ClusterIP, NodePort, and LoadBalancer, and does apply to "headless" services. This field will be wiped when updating a Service to type ExternalName.
This field may hold a maximum of two entries (dual-stack families, in either order). These families must correspond to the values of the clusterIPs field, if specified. Both clusterIPs and ipFamilies are governed by the ipFamilyPolicy field. |
||
ipFamilyPolicy | ✓ | string |
IPFamilyPolicy represents the dual-stack-ness requested or required by this Service. If there is no value provided, then this field will be set to SingleStack. Services can be “SingleStack” (a single IP family), “PreferDualStack” (two IP families on dual-stack configured clusters or a single IP family on single-stack clusters), or “RequireDualStack” (two IP families on dual-stack configured clusters, otherwise fail). The ipFamilies and clusterIPs fields depend on the value of this field. This field will be wiped when updating a service to type ExternalName. |
||
loadBalancerClass | ✓ | string |
loadBalancerClass is the class of the load balancer implementation this Service belongs to. If specified, the value of this field must be a label-style identifier, with an optional prefix, e.g. “internal-vip” or “example.com/internal-vip”. Unprefixed names are reserved for end-users. This field can only be set when the Service type is ‘LoadBalancer’. If not set, the default load balancer implementation is used, today this is typically done through the cloud provider integration, but should apply for any default implementation. If set, it is assumed that a load balancer implementation is watching for Services with a matching class. Any default load balancer implementation (e.g. cloud providers) should ignore Services that set this field. This field can only be set when creating or updating a Service to type ‘LoadBalancer’. Once set, it can not be changed. This field will be wiped when a service is updated to a non ‘LoadBalancer’ type. |
||
loadBalancerIP | ✓ | string |
Only applies to Service Type: LoadBalancer. This feature depends on whether the underlying cloud-provider supports specifying the loadBalancerIP when a load balancer is created. This field will be ignored if the cloud-provider does not support the feature. Deprecated: This field was under-specified and its meaning varies across implementations, and it cannot support dual-stack. As of Kubernetes v1.24, users are encouraged to use implementation-specific annotations when available. This field may be removed in a future API version. |
||
loadBalancerSourceRanges | ✓ | []string |
If specified and supported by the platform, this will restrict traffic through the cloud-provider load-balancer will be restricted to the specified client IPs. This field will be ignored if the cloud-provider does not support the feature.” More info: https://kubernetes.io/docs/tasks/access-application-cluster/create-external-load-balancer/ |
||
sessionAffinity | ✓ | string |
Supports “ClientIP” and “None”. Used to maintain session affinity. Enable client IP based session affinity. Must be ClientIP or None. Defaults to None. More info: https://kubernetes.io/docs/concepts/services-networking/service/#virtual-ips-and-service-proxies
|
||
sessionAffinityConfig | ✓ | object |
SessionAffinityConfig represents the configurations of session affinity. |
||
type | ✓ | enum |
type determines how the Service is exposed. Defaults to ClusterIP. Valid
options are ClusterIP, NodePort, and LoadBalancer. "ClusterIP" allocates
a cluster-internal IP address for load-balancing to endpoints.
"NodePort" builds on ClusterIP and allocates a port on every node.
"LoadBalancer" builds on NodePort and creates
an external load-balancer (if supported in the current cloud).
More info: https://kubernetes.io/docs/concepts/services-networking/service/#publishing-services-service-types
Enum: ClusterIP, LoadBalancer, NodePort |
SessionAffinityConfig represents the configurations of session affinity.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
clientIP | ✓ | object |
ClientIPConfig represents the configurations of Client IP based session affinity. |
ClientIPConfig represents the configurations of Client IP based session affinity.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
timeoutSeconds | ✓ | integer |
timeoutSeconds specifies the seconds of ClientIP type session sticky time. The value must be >0 && <=86400(for 1 day) if ServiceAffinity == “ClientIP”. Default value is 10800(for 3 hours). Format: int32 |
Configuration for the shards services
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
customPorts | ✓ | ✓ | []object |
The list of custom ports that will be exposed by the shards services.
The names of custom ports will be prefixed with the string The names of target ports will be prefixed with the string Changing this field may require a restart. See: https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.25/#serviceport-v1-core
|
|
primaries | ✓ | object |
Configure the shards service to any primary in the shards with the name as the SGShardedCluster plus the -shards suffix.
It provides a stable connection (regardless of primary failures or switchovers) to read-write Postgres servers of any shard cluster. Read-write servers are load-balanced via this service. See also https://kubernetes.io/docs/concepts/services-networking/service/
|
Configure the shards service to any primary in the shards with the name as the SGShardedCluster plus the -shards
suffix.
It provides a stable connection (regardless of primary failures or switchovers) to read-write Postgres servers of any shard cluster. Read-write servers are load-balanced via this service.
See also https://kubernetes.io/docs/concepts/services-networking/service/
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
allocateLoadBalancerNodePorts | ✓ | boolean |
allocateLoadBalancerNodePorts defines if NodePorts will be automatically allocated for services with type LoadBalancer. Default is “true”. It may be set to “false” if the cluster load-balancer does not rely on NodePorts. If the caller requests specific NodePorts (by specifying a value), those requests will be respected, regardless of this field. This field may only be set for services with type LoadBalancer and will be cleared if the type is changed to any other type. |
||
enabled | ✓ | boolean |
Specify if the service should be created or not. |
||
externalIPs | ✓ | []string |
externalIPs is a list of IP addresses for which nodes in the cluster will also accept traffic for this service. These IPs are not managed by Kubernetes. The user is responsible for ensuring that traffic arrives at a node with this IP. A common example is external load-balancers that are not part of the Kubernetes system. |
||
externalTrafficPolicy | ✓ | string |
externalTrafficPolicy describes how nodes distribute service traffic they receive on one of the Service’s “externally-facing” addresses (NodePorts, ExternalIPs, and LoadBalancer IPs). If set to “Local”, the proxy will configure the service in a way that assumes that external load balancers will take care of balancing the service traffic between nodes, and so each node will deliver traffic only to the node-local endpoints of the service, without masquerading the client source IP. (Traffic mistakenly sent to a node with no endpoints will be dropped.) The default value, “Cluster”, uses the standard behavior of routing to all endpoints evenly (possibly modified by topology and other features). Note that traffic sent to an External IP or LoadBalancer IP from within the cluster will always get “Cluster” semantics, but clients sending to a NodePort from within the cluster may need to take traffic policy into account when picking a node.
|
||
healthCheckNodePort | ✓ | integer |
healthCheckNodePort specifies the healthcheck nodePort for the service. This only applies when type is set to LoadBalancer and externalTrafficPolicy is set to Local. If a value is specified, is in-range, and is not in use, it will be used. If not specified, a value will be automatically allocated. External systems (e.g. load-balancers) can use this port to determine if a given node holds endpoints for this service or not. If this field is specified when creating a Service which does not need it, creation will fail. This field will be wiped when updating a Service to no longer need it (e.g. changing type). This field cannot be updated once set. Format: int32 |
||
internalTrafficPolicy | ✓ | string |
InternalTrafficPolicy describes how nodes distribute service traffic they receive on the ClusterIP. If set to "Local", the proxy will assume that pods only want to talk to endpoints of the service on the same node as the pod, dropping the traffic if there are no local endpoints. The default value, "Cluster", uses the standard behavior of routing to all endpoints evenly (possibly modified by topology and other features). |
||
ipFamilies | ✓ | []string |
IPFamilies is a list of IP families (e.g. IPv4, IPv6) assigned to this service. This field is usually assigned automatically based on cluster configuration and the ipFamilyPolicy field. If this field is specified manually, the requested family is available in the cluster, and ipFamilyPolicy allows it, it will be used; otherwise creation of the service will fail. This field is conditionally mutable: it allows for adding or removing a secondary IP family, but it does not allow changing the primary IP family of the Service. Valid values are "IPv4" and "IPv6". This field only applies to Services of types ClusterIP, NodePort, and LoadBalancer, and does apply to "headless" services. This field will be wiped when updating a Service to type ExternalName.
This field may hold a maximum of two entries (dual-stack families, in either order). These families must correspond to the values of the clusterIPs field, if specified. Both clusterIPs and ipFamilies are governed by the ipFamilyPolicy field. |
||
ipFamilyPolicy | ✓ | string |
IPFamilyPolicy represents the dual-stack-ness requested or required by this Service. If there is no value provided, then this field will be set to SingleStack. Services can be “SingleStack” (a single IP family), “PreferDualStack” (two IP families on dual-stack configured clusters or a single IP family on single-stack clusters), or “RequireDualStack” (two IP families on dual-stack configured clusters, otherwise fail). The ipFamilies and clusterIPs fields depend on the value of this field. This field will be wiped when updating a service to type ExternalName. |
||
loadBalancerClass | ✓ | string |
loadBalancerClass is the class of the load balancer implementation this Service belongs to. If specified, the value of this field must be a label-style identifier, with an optional prefix, e.g. “internal-vip” or “example.com/internal-vip”. Unprefixed names are reserved for end-users. This field can only be set when the Service type is ‘LoadBalancer’. If not set, the default load balancer implementation is used, today this is typically done through the cloud provider integration, but should apply for any default implementation. If set, it is assumed that a load balancer implementation is watching for Services with a matching class. Any default load balancer implementation (e.g. cloud providers) should ignore Services that set this field. This field can only be set when creating or updating a Service to type ‘LoadBalancer’. Once set, it can not be changed. This field will be wiped when a service is updated to a non ‘LoadBalancer’ type. |
||
loadBalancerIP | ✓ | string |
Only applies to Service Type: LoadBalancer. This feature depends on whether the underlying cloud-provider supports specifying the loadBalancerIP when a load balancer is created. This field will be ignored if the cloud-provider does not support the feature. Deprecated: This field was under-specified and its meaning varies across implementations, and it cannot support dual-stack. As of Kubernetes v1.24, users are encouraged to use implementation-specific annotations when available. This field may be removed in a future API version. |
||
loadBalancerSourceRanges | ✓ | []string |
If specified and supported by the platform, this will restrict traffic through the cloud-provider load-balancer will be restricted to the specified client IPs. This field will be ignored if the cloud-provider does not support the feature." More info: https://kubernetes.io/docs/tasks/access-application-cluster/create-external-load-balancer/ |
||
sessionAffinity | ✓ | string |
Supports “ClientIP” and “None”. Used to maintain session affinity. Enable client IP based session affinity. Must be ClientIP or None. Defaults to None. More info: https://kubernetes.io/docs/concepts/services-networking/service/#virtual-ips-and-service-proxies
|
||
sessionAffinityConfig | ✓ | object |
SessionAffinityConfig represents the configurations of session affinity. |
||
type | ✓ | enum |
type determines how the Service is exposed. Defaults to ClusterIP. Valid
options are ClusterIP, NodePort, and LoadBalancer. "ClusterIP" allocates
a cluster-internal IP address for load-balancing to endpoints.
"NodePort" builds on ClusterIP and allocates a port on every node.
"LoadBalancer" builds on NodePort and creates
an external load-balancer (if supported in the current cloud).
More info: https://kubernetes.io/docs/concepts/services-networking/service/#publishing-services-service-types
Enum: ClusterIP, LoadBalancer, NodePort |
SessionAffinityConfig represents the configurations of session affinity.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
clientIP | ✓ | object |
ClientIPConfig represents the configurations of Client IP based session affinity. |
ClientIPConfig represents the configurations of Client IP based session affinity.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
timeoutSeconds | ✓ | integer |
timeoutSeconds specifies the seconds of ClientIP type session sticky time. The value must be >0 && <=86400(for 1 day) if ServiceAffinity == “ClientIP”. Default value is 10800(for 3 hours). Format: int32 |
This section allows to configure the global Postgres replication mode.
The main replication group is implicit and contains the total number of instances less the sum of all instances in other replication groups.
The total number of instances is always specified by .spec.instances
.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
mode | ✓ | string |
The replication mode applied to the whole cluster.
Possible values are:
async When in asynchronous mode the cluster is allowed to lose some committed transactions. When the primary server fails or becomes unavailable for any other reason a sufficiently healthy standby will automatically be promoted to primary. Any transactions that have not been replicated to that standby remain in a “forked timeline” on the primary, and are effectively unrecoverable (the data is still there, but recovering it requires a manual recovery effort by data recovery specialists). sync When in synchronous mode a standby will not be promoted unless it is certain that the standby contains all
transactions that may have returned a successful commit status to client (clients can change the behavior
per transaction using PostgreSQL’s Synchronous mode does not guarantee multi node durability of commits under all circumstances. When no suitable standby is available, primary server will still accept writes, but does not guarantee their replication. When the primary fails in this mode no standby will be promoted. When the host that used to be the primary comes back it will get promoted automatically, unless system administrator performed a manual failover. This behavior makes synchronous mode usable with 2 node clusters. When synchronous mode is used and a standby crashes, commits will block until the primary is switched to standalone mode. Manually shutting down or restarting a standby will not cause a commit service interruption. Standby will signal the primary to release itself from synchronous standby duties before PostgreSQL shutdown is initiated. strict-sync When it is absolutely necessary to guarantee that each write is stored durably on at least two nodes, use the strict
synchronous mode. This mode prevents synchronous replication to be switched off on the primary when no synchronous
standby candidates are available. As a downside, the primary will not be available for writes (unless the Postgres
transaction explicitly turns off Note: Because of the way synchronous replication is implemented in PostgreSQL it is still possible to lose transactions even when using strict synchronous mode. If the PostgreSQL backend is cancelled while waiting to acknowledge replication (as a result of packet cancellation due to client timeout or backend failure) transaction changes become visible for other backends. Such changes are not yet replicated and may be lost in case of standby promotion. sync-all The same as strict-sync-all The same as |
||
syncInstances | ✓ | integer |
Number of synchronous standby instances. Must be less than the total number of instances. It is set to 1 by default.
Only setteable if mode is sync or strict-sync .
Minimum: 1 |
Current status of a StackGres sharded cluster.
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
clusterStatuses | ✓ | []object |
The list of cluster statuses. |
||
conditions | ✓ | []object |
|
||
toInstallPostgresExtensions | ✓ | []object |
The list of Postgres extensions to install |
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
name | ✓ | ✓ | string |
The name of the cluster. |
|
pendingRestart | ✓ | boolean |
Indicates if the cluster requires restart |
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
lastTransitionTime | ✓ | string |
Last time the condition transitioned from one status to another. |
||
message | ✓ | string |
A human readable message indicating details about the transition. |
||
reason | ✓ | string |
The reason for the condition’s last transition. |
||
status | ✓ | string |
Status of the condition, one of True, False, Unknown. |
||
type | ✓ | string |
Type of deployment condition. |
Property |
Required |
Updatable |
May Require Restart |
Type |
Description |
---|---|---|---|---|---|
name | ✓ | ✓ | string |
The name of the extension to install. |
|
postgresVersion | ✓ | ✓ | string |
The postgres major version of the extension to install. |
|
publisher | ✓ | ✓ | string |
The id of the publisher of the extension to install. |
|
repository | ✓ | ✓ | string |
The repository base URL from where the extension will be installed from. |
|
version | ✓ | ✓ | string |
The version of the extension to install. |
|
build | ✓ | string |
The build version of the extension to install. |
||
extraMounts | ✓ | []string |
The extra mounts of the extension to install. |