sasquatch Helm values reference¶
Helm values reference table for the sasquatch
application.
Key |
Type |
Default |
Description |
---|---|---|---|
global.baseUrl |
string |
Set by Argo CD |
Base URL for the environment |
global.host |
string |
Set by Argo CD |
Host name for ingress |
global.vaultSecretsPath |
string |
Set by Argo CD |
Base path for Vault secrets |
chronograf.enabled |
bool |
|
Whether Chronograf is enabled |
chronograf.env |
object |
See |
Additional environment variables for Chronograf |
chronograf.envFromSecret |
string |
|
Name of secret to use. The keys |
chronograf.image.repository |
string |
|
Docker image to use for Chronograf |
chronograf.image.tag |
string |
|
Docker tag to use for Chronograf |
chronograf.ingress.className |
string |
|
Ingress class to use |
chronograf.ingress.enabled |
bool |
|
Whether to enable the Chronograf ingress |
chronograf.ingress.hostname |
string |
None, must be set if the ingress is enabled |
Hostname of the ingress |
chronograf.ingress.path |
string |
|
Path for the ingress |
chronograf.ingress.tls |
bool |
|
Whether to obtain TLS certificates for the ingress hostname |
chronograf.persistence.enabled |
bool |
|
Whether to enable persistence for Chronograf data |
chronograf.persistence.size |
string |
|
Size of data store to request, if enabled |
chronograf.resources |
object |
See |
Kubernetes resource requests and limits for Chronograf |
influxdb-enterprise.enabled |
bool |
|
Whether to use influxdb-enterprise |
influxdb.config.continuous_queries.enabled |
bool |
|
Whether continuous queries are enabled |
influxdb.config.coordinator.log-queries-after |
string |
|
Maximum duration a query can run before InfluxDB logs it as a slow query |
influxdb.config.coordinator.max-concurrent-queries |
int |
|
Maximum number of running queries allowed on the instance (0 is unlimited) |
influxdb.config.coordinator.query-timeout |
string |
|
Maximum duration a query is allowed to run before it is killed |
influxdb.config.coordinator.write-timeout |
string |
|
Duration a write request waits before timeout is returned to the caller |
influxdb.config.data.cache-max-memory-size |
int |
|
Maximum size a shared cache can reach before it starts rejecting writes |
influxdb.config.data.trace-logging-enabled |
bool |
|
Whether to enable verbose logging of additional debug information within the TSM engine and WAL |
influxdb.config.data.wal-fsync-delay |
string |
|
Duration a write will wait before fsyncing. This is useful for slower disks or when WAL write contention is present. |
influxdb.config.http.auth-enabled |
bool |
|
Whether authentication is required |
influxdb.config.http.enabled |
bool |
|
Whether to enable the HTTP endpoints |
influxdb.config.http.flux-enabled |
bool |
|
Whether to enable the Flux query endpoint |
influxdb.config.http.max-row-limit |
int |
|
Maximum number of rows the system can return from a non-chunked query (0 is unlimited) |
influxdb.config.logging.level |
string |
|
Logging level |
influxdb.enabled |
bool |
|
Whether InfluxDB is enabled |
influxdb.image.tag |
string |
|
InfluxDB image tag |
influxdb.ingress.annotations |
object |
See |
Annotations to add to the ingress |
influxdb.ingress.className |
string |
|
Ingress class to use |
influxdb.ingress.enabled |
bool |
|
Whether to enable the InfluxDB ingress |
influxdb.ingress.hostname |
string |
None, must be set if the ingress is enabled |
Hostname of the ingress |
influxdb.ingress.path |
string |
|
Path for the ingress |
influxdb.ingress.tls |
bool |
|
Whether to obtain TLS certificates for the ingress hostname |
influxdb.initScripts.enabled |
bool |
|
Whether to enable the InfluxDB custom initialization script |
influxdb.persistence.enabled |
bool |
|
Whether to use persistent volume claims. By default, |
influxdb.persistence.size |
string |
1TiB for teststand deployments |
Persistent volume size |
influxdb.resources |
object |
See |
Kubernetes resource requests and limits |
influxdb.setDefaultUser.enabled |
bool |
|
Whether the default InfluxDB user is set |
influxdb.setDefaultUser.user.existingSecret |
string |
|
Use |
kafdrop.enabled |
bool |
|
Whether Kafdrop is enabled |
kafka-connect-manager |
object |
|
Overrides for kafka-connect-manager configuration |
kafka-connect-manager-enterprise.enabled |
bool |
|
Whether enterprise kafka-connect-manager is enabled |
kapacitor.enabled |
bool |
|
Whether Kapacitor is enabled |
kapacitor.envVars |
object |
See |
Additional environment variables to set |
kapacitor.existingSecret |
string |
|
Use |
kapacitor.image.repository |
string |
|
Docker image to use for Kapacitor |
kapacitor.image.tag |
string |
|
Tag to use for Kapacitor |
kapacitor.influxURL |
string |
|
InfluxDB connection URL |
kapacitor.persistence.enabled |
bool |
|
Whether to enable Kapacitor data persistence |
kapacitor.persistence.size |
string |
|
Size of storage to request if enabled |
kapacitor.resources |
object |
See |
Kubernetes resource requests and limits for Kapacitor |
rest-proxy.enabled |
bool |
|
Whether to enable the REST proxy |
squareEvents.enabled |
bool |
|
Enable the Square Events subchart with topic and user configurations |
strimzi-kafka.connect.enabled |
bool |
|
Whether Kafka Connect is enabled |
strimzi-kafka.kafka.listeners.external.enabled |
bool |
|
Whether external listener is enabled |
strimzi-kafka.kafka.listeners.plain.enabled |
bool |
|
Whether internal plaintext listener is enabled |
strimzi-kafka.kafka.listeners.tls.enabled |
bool |
|
Whether internal TLS listener is enabled |
strimzi-registry-operator.clusterName |
string |
|
Name of the Strimzi Kafka cluster |
strimzi-registry-operator.clusterNamespace |
string |
|
Namespace where the Strimzi Kafka cluster is deployed |
strimzi-registry-operator.operatorNamespace |
string |
|
Namespace where the strimzi-registry-operator is deployed |
telegraf-kafka-consumer |
object |
|
Overrides for telegraf-kafka-consumer configuration |
influxdb-enterprise.bootstrap.auth.secretName |
string |
|
Enable authentication of the data nodes using this secret, by creating a username and password for an admin account. The secret must contain keys |
influxdb-enterprise.bootstrap.ddldml.configMap |
string |
Do not run DDL or DML |
A config map containing DDL and DML that define databases, retention policies, and inject some data. The keys |
influxdb-enterprise.bootstrap.ddldml.resources |
object |
|
Kubernetes resources and limits for the bootstrap job |
influxdb-enterprise.data.affinity |
object |
See |
Affinity rules for data pods |
influxdb-enterprise.data.config.antiEntropy.enabled |
bool |
|
Enable the anti-entropy service, which copies and repairs shards |
influxdb-enterprise.data.config.cluster.log-queries-after |
string |
|
Maximum duration a query can run before InfluxDB logs it as a slow query |
influxdb-enterprise.data.config.cluster.max-concurrent-queries |
int |
|
Maximum number of running queries allowed on the instance (0 is unlimited) |
influxdb-enterprise.data.config.cluster.query-timeout |
string |
|
Maximum duration a query is allowed to run before it is killed |
influxdb-enterprise.data.config.continuousQueries.enabled |
bool |
|
Whether continuous queries are enabled |
influxdb-enterprise.data.config.data.cache-max-memory-size |
int |
|
Maximum size a shared cache can reach before it starts rejecting writes |
influxdb-enterprise.data.config.data.trace-logging-enabled |
bool |
|
Whether to enable verbose logging of additional debug information within the TSM engine and WAL |
influxdb-enterprise.data.config.data.wal-fsync-delay |
string |
|
Duration a write will wait before fsyncing. This is useful for slower disks or when WAL write contention is present. |
influxdb-enterprise.data.config.hintedHandoff.max-size |
int |
|
Maximum size of the hinted-handoff queue in bytes |
influxdb-enterprise.data.config.http.auth-enabled |
bool |
|
Whether authentication is required |
influxdb-enterprise.data.config.http.flux-enabled |
bool |
|
Whether to enable the Flux query endpoint |
influxdb-enterprise.data.config.logging.level |
string |
|
Logging level |
influxdb-enterprise.data.env |
object |
|
Additional environment variables to set in the meta container |
influxdb-enterprise.data.image.pullPolicy |
string |
|
Pull policy for data images |
influxdb-enterprise.data.image.repository |
string |
|
Docker repository for data images |
influxdb-enterprise.data.ingress.annotations |
object |
See |
Extra annotations to add to the data ingress |
influxdb-enterprise.data.ingress.className |
string |
|
Ingress class name of the data service |
influxdb-enterprise.data.ingress.enabled |
bool |
|
Whether to enable an ingress for the data service |
influxdb-enterprise.data.ingress.hostname |
string |
None, must be set if the ingress is enabled |
Hostname of the data ingress |
influxdb-enterprise.data.ingress.path |
string |
|
Path of the data service |
influxdb-enterprise.data.nodeSelector |
object |
|
Node selection rules for data pods |
influxdb-enterprise.data.persistence.accessMode |
string |
|
Access mode for the persistent volume claim |
influxdb-enterprise.data.persistence.annotations |
object |
|
Annotations to add to the persistent volume claim |
influxdb-enterprise.data.persistence.enabled |
bool |
|
Whether to persist data to a persistent volume |
influxdb-enterprise.data.persistence.existingClaim |
string |
Use a volume claim template |
Manually managed PersistentVolumeClaim to use. If defined, this PVC must be created manually before the meta service will start |
influxdb-enterprise.data.persistence.size |
string |
|
Size of persistent volume to request |
influxdb-enterprise.data.persistence.storageClass |
string |
|
Storage class of the persistent volume (set to |
influxdb-enterprise.data.podAnnotations |
object |
|
Annotations for data pods |
influxdb-enterprise.data.podDisruptionBudget.minAvailable |
int |
|
Minimum available pods to maintain |
influxdb-enterprise.data.podSecurityContext |
object |
|
Pod security context for data pods |
influxdb-enterprise.data.preruncmds |
list |
|
Commands to run in data pods before InfluxDB is started. Each list entry should have a cmd key with the command to run and an optional description key describing that command |
influxdb-enterprise.data.replicas |
int |
|
Number of data replicas to run |
influxdb-enterprise.data.resources |
object |
|
Kubernetes resources and limits for the meta container |
influxdb-enterprise.data.securityContext |
object |
|
Security context for meta pods |
influxdb-enterprise.data.service.annotations |
object |
|
Extra annotations for the data service |
influxdb-enterprise.data.service.externalIPs |
list |
Do not allocate external IPs |
External IPs for the data service |
influxdb-enterprise.data.service.externalTrafficPolicy |
string |
Do not set an external traffic policy |
External traffic policy for the data service |
influxdb-enterprise.data.service.loadBalancerIP |
string |
Do not allocate a load balancer IP |
Load balancer IP for the data service |
influxdb-enterprise.data.service.nodePort |
int |
Do not allocate a node port |
Node port for the data service |
influxdb-enterprise.data.service.type |
string |
|
Service type for the data service |
influxdb-enterprise.data.tolerations |
list |
|
Tolerations for data pods |
influxdb-enterprise.envFromSecret |
string |
No secret |
The name of a secret in the same kubernetes namespace which contain values to be added to the environment |
influxdb-enterprise.fullnameOverride |
string |
|
Override the full name for resources (includes the release name) |
influxdb-enterprise.image.addsuffix |
bool |
|
Set to true to add a suffix for the type of image to the Docker tag (for example, |
influxdb-enterprise.image.tag |
string |
|
Tagged version of the Docker image that you want to run |
influxdb-enterprise.imagePullSecrets |
list |
|
List of pull secrets needed for images. If set, each object in the list should have one attribute, name, identifying the pull secret to use |
influxdb-enterprise.license.key |
string |
|
License key. You can put your license key here for testing this chart out, but we STRONGLY recommend using a license file stored in a secret when you ship to production. |
influxdb-enterprise.license.secret.key |
string |
|
Key within that secret that contains the license |
string |
|
Name of the secret containing the license |
|
influxdb-enterprise.meta.affinity |
object |
See |
Affinity rules for meta pods |
influxdb-enterprise.meta.env |
object |
|
Additional environment variables to set in the meta container |
influxdb-enterprise.meta.image.pullPolicy |
string |
|
Pull policy for meta images |
influxdb-enterprise.meta.image.repository |
string |
|
Docker repository for meta images |
influxdb-enterprise.meta.ingress.annotations |
object |
See |
Extra annotations to add to the meta ingress |
influxdb-enterprise.meta.ingress.className |
string |
|
Ingress class name of the meta service |
influxdb-enterprise.meta.ingress.enabled |
bool |
|
Whether to enable an ingress for the meta service |
influxdb-enterprise.meta.ingress.hostname |
string |
None, must be set if the ingress is enabled |
Hostname of the meta ingress |
influxdb-enterprise.meta.ingress.path |
string |
|
Path of the meta service |
influxdb-enterprise.meta.nodeSelector |
object |
|
Node selection rules for meta pods |
influxdb-enterprise.meta.persistence.accessMode |
string |
|
Access mode for the persistent volume claim |
influxdb-enterprise.meta.persistence.annotations |
object |
|
Annotations to add to the persistent volume claim |
influxdb-enterprise.meta.persistence.enabled |
bool |
|
Whether to persist data to a persistent volume |
influxdb-enterprise.meta.persistence.existingClaim |
string |
Use a volume claim template |
Manually managed PersistentVolumeClaim to use. If defined, this PVC must be created manually before the meta service will start |
influxdb-enterprise.meta.persistence.size |
string |
|
Size of persistent volume to request |
influxdb-enterprise.meta.persistence.storageClass |
string |
|
Storage class of the persistent volume (set to |
influxdb-enterprise.meta.podAnnotations |
object |
|
Annotations for meta pods |
influxdb-enterprise.meta.podDisruptionBudget.minAvailable |
int |
|
Minimum available pods to maintain |
influxdb-enterprise.meta.podSecurityContext |
object |
|
Pod security context for meta pods |
influxdb-enterprise.meta.preruncmds |
list |
|
Commands to run in meta pods before InfluxDB is started. Each list entry should have a cmd key with the command to run and an optional description key describing that command |
influxdb-enterprise.meta.replicas |
int |
|
Number of meta pods to run |
influxdb-enterprise.meta.resources |
object |
|
Kubernetes resources and limits for the meta container |
influxdb-enterprise.meta.securityContext |
object |
|
Security context for meta pods |
influxdb-enterprise.meta.service.annotations |
object |
|
Extra annotations for the meta service |
influxdb-enterprise.meta.service.externalIPs |
list |
Do not allocate external IPs |
External IPs for the meta service |
influxdb-enterprise.meta.service.externalTrafficPolicy |
string |
Do not set an external traffic policy |
External traffic policy for the meta service |
influxdb-enterprise.meta.service.loadBalancerIP |
string |
Do not allocate a load balancer IP |
Load balancer IP for the meta service |
influxdb-enterprise.meta.service.nodePort |
int |
Do not allocate a node port |
Node port for the meta service |
influxdb-enterprise.meta.service.type |
string |
|
Service type for the meta service |
influxdb-enterprise.meta.sharedSecret.secretName |
string |
|
Shared secret used by the internal API for JWT authentication between InfluxDB nodes. Must have a key named |
influxdb-enterprise.meta.tolerations |
list |
|
Tolerations for meta pods |
influxdb-enterprise.nameOverride |
string |
|
Override the base name for resources |
influxdb-enterprise.serviceAccount.annotations |
object |
|
Annotations to add to the service account |
influxdb-enterprise.serviceAccount.create |
bool |
|
Whether to create a Kubernetes service account to run as |
string |
Name based on the chart fullname |
Name of the Kubernetes service account to run as |
|
kafdrop.affinity |
object |
|
Affinity configuration |
kafdrop.cmdArgs |
string |
See |
Command line arguments to Kafdrop |
kafdrop.existingSecret |
string |
Do not use a secret |
Existing Kubernetes secrect use to set kafdrop environment variables. Set |
kafdrop.host |
string |
|
The hostname to report for the RMI registry (used for JMX) |
kafdrop.image.pullPolicy |
string |
|
Image pull policy |
kafdrop.image.repository |
string |
|
Kafdrop Docker image repository |
kafdrop.image.tag |
string |
|
Kafdrop image version |
kafdrop.ingress.annotations |
object |
|
Additional ingress annotations |
kafdrop.ingress.enabled |
bool |
|
Whether to enable the ingress |
kafdrop.ingress.hostname |
string |
None, must be set if ingress is enabled |
Ingress hostname |
kafdrop.ingress.path |
string |
|
Ingress path |
kafdrop.jmx.port |
int |
|
Port to use for JMX. If unspecified, JMX will not be exposed. |
kafdrop.jvm.opts |
string |
|
JVM options |
kafdrop.kafka.broker |
string |
|
Bootstrap list of Kafka host/port pairs |
kafdrop.nodeSelector |
object |
|
Node selector configuration |
kafdrop.podAnnotations |
object |
|
Pod annotations |
kafdrop.replicaCount |
int |
|
Number of kafdrop pods to run in the deployment. |
kafdrop.resources |
object |
See |
Kubernetes requests and limits for Kafdrop |
kafdrop.schemaregistry |
string |
|
The endpoint of Schema Registry |
kafdrop.server.port |
int |
|
The web server port to listen on |
kafdrop.server.servlet.contextPath |
string |
|
The context path to serve requests on |
kafdrop.service.annotations |
object |
|
Additional annotations to add to the service |
kafdrop.service.port |
int |
|
Service port |
kafdrop.tolerations |
list |
|
Tolerations configuration |
kafka-connect-manager.enabled |
bool |
|
Whether to enable Kafka Connect Manager |
kafka-connect-manager.env.kafkaBrokerUrl |
string |
|
Kafka broker URL |
kafka-connect-manager.env.kafkaConnectUrl |
string |
|
Kafka connnect URL |
kafka-connect-manager.env.kafkaUsername |
string |
|
Username for SASL authentication |
kafka-connect-manager.image.pullPolicy |
string |
|
Pull policy for Kafka Connect Manager |
kafka-connect-manager.image.repository |
string |
|
Docker image to use for Kafka Connect Manager |
kafka-connect-manager.image.tag |
string |
|
Docker tag to use for Kafka Connect Manager |
kafka-connect-manager.influxdbSink.autoUpdate |
bool |
|
Whether to check for new Kafka topics |
kafka-connect-manager.influxdbSink.checkInterval |
string |
|
The interval, in milliseconds, to check for new topics and update the connector |
kafka-connect-manager.influxdbSink.connectInfluxDb |
string |
|
InfluxDB database to write to |
kafka-connect-manager.influxdbSink.connectInfluxErrorPolicy |
string |
|
Error policy, see connector documetation for details |
kafka-connect-manager.influxdbSink.connectInfluxMaxRetries |
string |
|
The maximum number of times a message is retried |
kafka-connect-manager.influxdbSink.connectInfluxRetryInterval |
string |
|
The interval, in milliseconds, between retries. Only valid when the connectInfluxErrorPolicy is set to |
kafka-connect-manager.influxdbSink.connectInfluxUrl |
string |
|
InfluxDB URL |
kafka-connect-manager.influxdbSink.connectProgressEnabled |
bool |
|
Enables the output for how many records have been processed |
kafka-connect-manager.influxdbSink.connectors |
object |
See |
Connector instances to deploy. See |
kafka-connect-manager.influxdbSink.excludedTopicsRegex |
string |
|
Regex to exclude topics from the list of selected topics from Kafka |
kafka-connect-manager.influxdbSink.tasksMax |
int |
|
Maxium number of tasks to run the connector |
kafka-connect-manager.influxdbSink.timestamp |
string |
|
Timestamp field to be used as the InfluxDB time. If not specified use |
kafka-connect-manager.jdbcSink.autoCreate |
string |
|
Whether to automatically create the destination table |
kafka-connect-manager.jdbcSink.autoEvolve |
string |
|
Whether to automatically add columns in the table schema |
kafka-connect-manager.jdbcSink.batchSize |
string |
|
Specifies how many records to attempt to batch together for insertion into the destination table |
kafka-connect-manager.jdbcSink.connectionUrl |
string |
|
Database connection URL |
kafka-connect-manager.jdbcSink.dbTimezone |
string |
|
Name of the JDBC timezone that should be used in the connector when inserting time-based values |
kafka-connect-manager.jdbcSink.enabled |
bool |
|
Whether the JDBC Sink connector is deployed |
kafka-connect-manager.jdbcSink.insertMode |
string |
|
The insertion mode to use. Supported modes are: |
kafka-connect-manager.jdbcSink.maxRetries |
string |
|
The maximum number of times to retry on errors before failing the task |
string |
|
Name of the connector to create |
|
kafka-connect-manager.jdbcSink.retryBackoffMs |
string |
|
The time in milliseconds to wait following an error before a retry attempt is made |
kafka-connect-manager.jdbcSink.tableNameFormat |
string |
|
A format string for the destination table name |
kafka-connect-manager.jdbcSink.tasksMax |
string |
|
Number of Kafka Connect tasks |
kafka-connect-manager.jdbcSink.topicRegex |
string |
|
Regex for selecting topics |
kafka-connect-manager.s3Sink.behaviorOnNullValues |
string |
|
How to handle records with a null value (for example, Kafka tombstone records). Valid options are |
kafka-connect-manager.s3Sink.checkInterval |
string |
|
The interval, in milliseconds, to check for new topics and update the connector |
kafka-connect-manager.s3Sink.enabled |
bool |
|
Whether the Amazon S3 Sink connector is deployed |
kafka-connect-manager.s3Sink.excludedTopicRegex |
string |
|
Regex to exclude topics from the list of selected topics from Kafka |
kafka-connect-manager.s3Sink.flushSize |
string |
|
Number of records written to store before invoking file commits |
kafka-connect-manager.s3Sink.locale |
string |
|
The locale to use when partitioning with TimeBasedPartitioner |
string |
|
Name of the connector to create |
|
kafka-connect-manager.s3Sink.partitionDurationMs |
string |
|
The duration of a partition in milliseconds, used by TimeBasedPartitioner. Default is 1h for an hourly based partitioner |
kafka-connect-manager.s3Sink.pathFormat |
string |
|
Pattern used to format the path in the S3 object name |
kafka-connect-manager.s3Sink.rotateIntervalMs |
string |
|
The time interval in milliseconds to invoke file commits. Set to 10 minutes by default |
kafka-connect-manager.s3Sink.s3BucketName |
string |
|
S3 bucket name. The bucket must already exist at the s3 provider |
kafka-connect-manager.s3Sink.s3PartRetries |
int |
|
Maximum number of retry attempts for failed requests. Zero means no retries. |
kafka-connect-manager.s3Sink.s3PartSize |
int |
|
The part size in S3 multi-part uploads. Valid values: [5242880,…,2147483647] |
kafka-connect-manager.s3Sink.s3Region |
string |
|
S3 region |
kafka-connect-manager.s3Sink.s3RetryBackoffMs |
int |
|
How long to wait in milliseconds before attempting the first retry of a failed S3 request |
kafka-connect-manager.s3Sink.s3SchemaCompatibility |
string |
|
S3 schema compatibility |
kafka-connect-manager.s3Sink.schemaCacheConfig |
int |
|
The size of the schema cache used in the Avro converter |
kafka-connect-manager.s3Sink.storeUrl |
string |
|
The object storage connection URL, for non-AWS s3 providers |
kafka-connect-manager.s3Sink.tasksMax |
int |
|
Number of Kafka Connect tasks |
kafka-connect-manager.s3Sink.timestampExtractor |
string |
|
The extractor determines how to obtain a timestamp from each record |
kafka-connect-manager.s3Sink.timestampField |
string |
|
The record field to be used as timestamp by the timestamp extractor. Only applies if timestampExtractor is set to RecordField. |
kafka-connect-manager.s3Sink.timezone |
string |
|
The timezone to use when partitioning with TimeBasedPartitioner |
kafka-connect-manager.s3Sink.topicsDir |
string |
|
Top level directory to store the data ingested from Kafka |
kafka-connect-manager.s3Sink.topicsRegex |
string |
|
Regex to select topics from Kafka |
kafka-connect-manager-enterprise.enabled |
bool |
|
Whether to enable Kafka Connect Manager |
kafka-connect-manager-enterprise.env.kafkaBrokerUrl |
string |
|
Kafka broker URL |
kafka-connect-manager-enterprise.env.kafkaConnectUrl |
string |
|
Kafka connnect URL |
kafka-connect-manager-enterprise.env.kafkaUsername |
string |
|
Username for SASL authentication |
kafka-connect-manager-enterprise.image.pullPolicy |
string |
|
Pull policy for Kafka Connect Manager |
kafka-connect-manager-enterprise.image.repository |
string |
|
Docker image to use for Kafka Connect Manager |
kafka-connect-manager-enterprise.image.tag |
string |
|
Docker tag to use for Kafka Connect Manager |
kafka-connect-manager-enterprise.influxdbSink.autoUpdate |
bool |
|
Whether to check for new Kafka topics |
kafka-connect-manager-enterprise.influxdbSink.checkInterval |
string |
|
The interval, in milliseconds, to check for new topics and update the connector |
kafka-connect-manager-enterprise.influxdbSink.connectInfluxDb |
string |
|
InfluxDB database to write to |
kafka-connect-manager-enterprise.influxdbSink.connectInfluxErrorPolicy |
string |
|
Error policy, see connector documetation for details |
kafka-connect-manager-enterprise.influxdbSink.connectInfluxMaxRetries |
string |
|
The maximum number of times a message is retried |
kafka-connect-manager-enterprise.influxdbSink.connectInfluxRetryInterval |
string |
|
The interval, in milliseconds, between retries. Only valid when the connectInfluxErrorPolicy is set to |
kafka-connect-manager-enterprise.influxdbSink.connectInfluxUrl |
string |
|
InfluxDB URL |
kafka-connect-manager-enterprise.influxdbSink.connectProgressEnabled |
bool |
|
Enables the output for how many records have been processed |
kafka-connect-manager-enterprise.influxdbSink.connectors |
object |
See |
Connector instances to deploy. See |
kafka-connect-manager-enterprise.influxdbSink.excludedTopicsRegex |
string |
|
Regex to exclude topics from the list of selected topics from Kafka |
kafka-connect-manager-enterprise.influxdbSink.tasksMax |
int |
|
Maxium number of tasks to run the connector |
kafka-connect-manager-enterprise.influxdbSink.timestamp |
string |
|
Timestamp field to be used as the InfluxDB time. If not specified use |
kafka-connect-manager-enterprise.jdbcSink.autoCreate |
string |
|
Whether to automatically create the destination table |
kafka-connect-manager-enterprise.jdbcSink.autoEvolve |
string |
|
Whether to automatically add columns in the table schema |
kafka-connect-manager-enterprise.jdbcSink.batchSize |
string |
|
Specifies how many records to attempt to batch together for insertion into the destination table |
kafka-connect-manager-enterprise.jdbcSink.connectionUrl |
string |
|
Database connection URL |
kafka-connect-manager-enterprise.jdbcSink.dbTimezone |
string |
|
Name of the JDBC timezone that should be used in the connector when inserting time-based values |
kafka-connect-manager-enterprise.jdbcSink.enabled |
bool |
|
Whether the JDBC Sink connector is deployed |
kafka-connect-manager-enterprise.jdbcSink.insertMode |
string |
|
The insertion mode to use. Supported modes are: |
kafka-connect-manager-enterprise.jdbcSink.maxRetries |
string |
|
The maximum number of times to retry on errors before failing the task |
string |
|
Name of the connector to create |
|
kafka-connect-manager-enterprise.jdbcSink.retryBackoffMs |
string |
|
The time in milliseconds to wait following an error before a retry attempt is made |
kafka-connect-manager-enterprise.jdbcSink.tableNameFormat |
string |
|
A format string for the destination table name |
kafka-connect-manager-enterprise.jdbcSink.tasksMax |
string |
|
Number of Kafka Connect tasks |
kafka-connect-manager-enterprise.jdbcSink.topicRegex |
string |
|
Regex for selecting topics |
kafka-connect-manager-enterprise.s3Sink.behaviorOnNullValues |
string |
|
How to handle records with a null value (for example, Kafka tombstone records). Valid options are |
kafka-connect-manager-enterprise.s3Sink.checkInterval |
string |
|
The interval, in milliseconds, to check for new topics and update the connector |
kafka-connect-manager-enterprise.s3Sink.enabled |
bool |
|
Whether the Amazon S3 Sink connector is deployed |
kafka-connect-manager-enterprise.s3Sink.excludedTopicRegex |
string |
|
Regex to exclude topics from the list of selected topics from Kafka |
kafka-connect-manager-enterprise.s3Sink.flushSize |
string |
|
Number of records written to store before invoking file commits |
kafka-connect-manager-enterprise.s3Sink.locale |
string |
|
The locale to use when partitioning with TimeBasedPartitioner |
string |
|
Name of the connector to create |
|
kafka-connect-manager-enterprise.s3Sink.partitionDurationMs |
string |
|
The duration of a partition in milliseconds, used by TimeBasedPartitioner. Default is 1h for an hourly based partitioner |
kafka-connect-manager-enterprise.s3Sink.pathFormat |
string |
|
Pattern used to format the path in the S3 object name |
kafka-connect-manager-enterprise.s3Sink.rotateIntervalMs |
string |
|
The time interval in milliseconds to invoke file commits. Set to 10 minutes by default |
kafka-connect-manager-enterprise.s3Sink.s3BucketName |
string |
|
S3 bucket name. The bucket must already exist at the s3 provider |
kafka-connect-manager-enterprise.s3Sink.s3PartRetries |
int |
|
Maximum number of retry attempts for failed requests. Zero means no retries. |
kafka-connect-manager-enterprise.s3Sink.s3PartSize |
int |
|
The part size in S3 multi-part uploads. Valid values: [5242880,…,2147483647] |
kafka-connect-manager-enterprise.s3Sink.s3Region |
string |
|
S3 region |
kafka-connect-manager-enterprise.s3Sink.s3RetryBackoffMs |
int |
|
How long to wait in milliseconds before attempting the first retry of a failed S3 request |
kafka-connect-manager-enterprise.s3Sink.s3SchemaCompatibility |
string |
|
S3 schema compatibility |
kafka-connect-manager-enterprise.s3Sink.schemaCacheConfig |
int |
|
The size of the schema cache used in the Avro converter |
kafka-connect-manager-enterprise.s3Sink.storeUrl |
string |
|
The object storage connection URL, for non-AWS s3 providers |
kafka-connect-manager-enterprise.s3Sink.tasksMax |
int |
|
Number of Kafka Connect tasks |
kafka-connect-manager-enterprise.s3Sink.timestampExtractor |
string |
|
The extractor determines how to obtain a timestamp from each record |
kafka-connect-manager-enterprise.s3Sink.timestampField |
string |
|
The record field to be used as timestamp by the timestamp extractor. Only applies if timestampExtractor is set to RecordField. |
kafka-connect-manager-enterprise.s3Sink.timezone |
string |
|
The timezone to use when partitioning with TimeBasedPartitioner |
kafka-connect-manager-enterprise.s3Sink.topicsDir |
string |
|
Top level directory to store the data ingested from Kafka |
kafka-connect-manager-enterprise.s3Sink.topicsRegex |
string |
|
Regex to select topics from Kafka |
rest-proxy.affinity |
object |
|
Affinity configuration |
rest-proxy.configurationOverrides |
object |
See |
Kafka REST configuration options |
rest-proxy.customEnv |
object |
|
Kafka REST additional env variables |
rest-proxy.heapOptions |
string |
|
Kafka REST proxy JVM Heap Option |
rest-proxy.image.pullPolicy |
string |
|
Image pull policy |
rest-proxy.image.repository |
string |
|
Kafka REST proxy image repository |
rest-proxy.image.tag |
string |
|
Kafka REST proxy image tag |
rest-proxy.ingress.annotations |
object |
See |
Additional annotations to add to the ingress |
rest-proxy.ingress.enabled |
bool |
|
Whether to enable the ingress |
rest-proxy.ingress.hostname |
string |
None, must be set if ingress is enabled |
Ingress hostname |
rest-proxy.ingress.path |
string |
`”/sasquatch-rest-proxy(/ |
$)(.*)”` |
rest-proxy.kafka.bootstrapServers |
string |
|
Kafka bootstrap servers, use the internal listerner on port 9092 with SASL connection |
string |
|
Name of the Strimzi Kafka cluster. |
|
rest-proxy.kafka.topicPrefixes |
list |
|
List of topic prefixes to use when exposing Kafka topics to the REST Proxy v2 API. |
rest-proxy.kafka.topics |
list |
|
List of Kafka topics to create via Strimzi. Alternatively topics can be created using the REST Proxy v3 API. |
rest-proxy.nodeSelector |
object |
|
Node selector configuration |
rest-proxy.podAnnotations |
object |
|
Pod annotations |
rest-proxy.replicaCount |
int |
|
Number of Kafka REST proxy pods to run in the deployment |
rest-proxy.resources |
object |
See |
Kubernetes requests and limits for the Kafka REST proxy |
rest-proxy.schemaregistry.url |
string |
|
Schema registry URL |
rest-proxy.service.port |
int |
|
Kafka REST proxy service port |
rest-proxy.tolerations |
list |
|
Tolerations configuration |
string |
|
||
strimzi-kafka.cluster.monitorLabel |
object |
|
Site wide label required for gathering Prometheus metrics if they are enabled |
string |
|
Name used for the Kafka cluster, and used by Strimzi for many annotations |
|
strimzi-kafka.connect.config.“key.converter” |
string |
|
Set the converter for the message ke |
strimzi-kafka.connect.config.“key.converter.schema.registry.url” |
string |
|
URL for the schema registry |
strimzi-kafka.connect.config.“key.converter.schemas.enable” |
bool |
|
Enable converted schemas for the message key |
strimzi-kafka.connect.config.“value.converter” |
string |
|
Converter for the message value |
strimzi-kafka.connect.config.“value.converter.schema.registry.url” |
string |
|
URL for the schema registry |
strimzi-kafka.connect.config.“value.converter.schemas.enable” |
bool |
|
Enable converted schemas for the message value |
strimzi-kafka.connect.enabled |
bool |
|
Enable Kafka Connect |
strimzi-kafka.connect.image |
string |
|
Custom strimzi-kafka image with connector plugins used by sasquatch |
strimzi-kafka.connect.replicas |
int |
|
Number of Kafka Connect replicas to run |
strimzi-kafka.kafka.affinity |
object |
See |
Affinity for Kafka pod assignment |
strimzi-kafka.kafka.config.“log.retention.bytes” |
string |
|
How much disk space Kafka will ensure is available, set to 70% of the data partition size |
strimzi-kafka.kafka.config.“log.retention.hours” |
int |
|
Number of days for a topic’s data to be retained |
strimzi-kafka.kafka.config.“message.max.bytes” |
int |
|
The largest record batch size allowed by Kafka |
strimzi-kafka.kafka.config.“offsets.retention.minutes” |
int |
|
Number of minutes for a consumer group’s offsets to be retained |
strimzi-kafka.kafka.config.“replica.fetch.max.bytes” |
int |
|
The number of bytes of messages to attempt to fetch for each partition |
strimzi-kafka.kafka.disruption_tolerance |
int |
|
Number of down brokers that the system can tolerate |
strimzi-kafka.kafka.externalListener.bootstrap.annotations |
object |
|
Annotations that will be added to the Ingress, Route, or Service resource |
strimzi-kafka.kafka.externalListener.bootstrap.host |
string |
Do not configure TLS |
Name used for TLS hostname verification |
strimzi-kafka.kafka.externalListener.bootstrap.loadBalancerIP |
string |
Do not request a load balancer IP |
Request this load balancer IP. See |
strimzi-kafka.kafka.externalListener.brokers |
list |
|
Brokers configuration. host is used in the brokers’ advertised.brokers configuration and for TLS hostname verification. The format is a list of maps. |
strimzi-kafka.kafka.externalListener.tls.certIssuerName |
string |
|
Name of a ClusterIssuer capable of provisioning a TLS certificate for the broker |
strimzi-kafka.kafka.externalListener.tls.enabled |
bool |
|
Whether TLS encryption is enabled |
strimzi-kafka.kafka.listeners.external.enabled |
bool |
|
Whether external listener is enabled |
strimzi-kafka.kafka.listeners.plain.enabled |
bool |
|
Whether internal plaintext listener is enabled |
strimzi-kafka.kafka.listeners.tls.enabled |
bool |
|
Whether internal TLS listener is enabled |
strimzi-kafka.kafka.metricsConfig.enabled |
bool |
|
Whether metric configuration is enabled |
strimzi-kafka.kafka.replicas |
int |
|
Number of Kafka broker replicas to run |
strimzi-kafka.kafka.resources |
object |
See |
Kubernetes requests and limits for the Kafka brokers |
strimzi-kafka.kafka.storage.size |
string |
|
Size of the backing storage disk for each of the Kafka brokers |
strimzi-kafka.kafka.storage.storageClassName |
string |
|
Name of a StorageClass to use when requesting persistent volumes |
strimzi-kafka.kafka.tolerations |
list |
|
Tolerations for Kafka broker pod assignment |
strimzi-kafka.kafka.version |
string |
|
Version of Kafka to deploy |
strimzi-kafka.kafkaController.enabled |
bool |
|
Enable Kafka Controller |
strimzi-kafka.kafkaController.resources |
object |
See |
Kubernetes requests and limits for the Kafka Controller |
strimzi-kafka.kafkaController.storage.size |
string |
|
Size of the backing storage disk for each of the Kafka controllers |
strimzi-kafka.kafkaController.storage.storageClassName |
string |
|
Name of a StorageClass to use when requesting persistent volumes |
strimzi-kafka.kafkaExporter.enableSaramaLogging |
bool |
|
Enable Sarama logging for pod |
strimzi-kafka.kafkaExporter.enabled |
bool |
|
Enable Kafka exporter |
strimzi-kafka.kafkaExporter.groupRegex |
string |
|
Consumer groups to monitor |
strimzi-kafka.kafkaExporter.logging |
string |
|
Logging level |
strimzi-kafka.kafkaExporter.resources |
object |
See |
Kubernetes requests and limits for the Kafka exporter |
strimzi-kafka.kafkaExporter.topicRegex |
string |
|
Kafka topics to monitor |
strimzi-kafka.kraft.enabled |
bool |
|
Enable KRaft mode for Kafka |
strimzi-kafka.mirrormaker2.enabled |
bool |
|
Enable replication in the target (passive) cluster |
strimzi-kafka.mirrormaker2.replicas |
int |
|
Number of Mirror Maker replicas to run |
strimzi-kafka.mirrormaker2.replication.policy.class |
string |
|
Replication policy. |
strimzi-kafka.mirrormaker2.replication.policy.separator |
string |
|
Convention used to rename topics when the DefaultReplicationPolicy replication policy is used. Default is “” when the IdentityReplicationPolicy replication policy is used. |
strimzi-kafka.mirrormaker2.source.bootstrapServer |
string |
None, must be set if enabled |
Source (active) cluster to replicate from |
strimzi-kafka.mirrormaker2.source.topicsPattern |
string |
|
Topic replication from the source cluster defined as a comma-separated list or regular expression pattern |
strimzi-kafka.registry.ingress.annotations |
object |
|
Annotations that will be added to the Ingress resource |
strimzi-kafka.registry.ingress.enabled |
bool |
|
Whether to enable an ingress for the Schema Registry |
strimzi-kafka.registry.ingress.hostname |
string |
None, must be set if ingress is enabled |
Hostname for the Schema Registry |
strimzi-kafka.registry.resources |
object |
See |
Kubernetes requests and limits for the Schema Registry |
strimzi-kafka.registry.schemaTopic |
string |
|
Name of the topic used by the Schema Registry |
strimzi-kafka.superusers |
list |
|
A list of usernames for users who should have global admin permissions. These users will be created, along with their credentials. |
strimzi-kafka.users.consdb.enabled |
bool |
|
Enable user consdb |
strimzi-kafka.users.kafdrop.enabled |
bool |
|
Enable user Kafdrop (deployed by parent Sasquatch chart). |
strimzi-kafka.users.kafkaConnectManager.enabled |
bool |
|
Enable user kafka-connect-manager |
strimzi-kafka.users.promptProcessing.enabled |
bool |
|
Enable user prompt-processing |
strimzi-kafka.users.replicator.enabled |
bool |
|
Enable user replicator (used by Mirror Maker 2 and required at both source and target clusters) |
strimzi-kafka.users.telegraf.enabled |
bool |
|
Enable user telegraf (deployed by parent Sasquatch chart) |
strimzi-kafka.users.tsSalKafka.enabled |
bool |
|
Enable user ts-salkafka, used at the telescope environments |
telegraf-kafka-consumer.affinity |
object |
|
Affinity for pod assignment |
telegraf-kafka-consumer.args |
list |
|
Arguments passed to the Telegraf agent containers |
telegraf-kafka-consumer.enabled |
bool |
|
Wether the Telegraf Kafka Consumer is enabled |
telegraf-kafka-consumer.env |
list |
See |
Telegraf agent enviroment variables |
telegraf-kafka-consumer.envFromSecret |
string |
|
Name of the secret with values to be added to the environment. |
telegraf-kafka-consumer.image.pullPolicy |
string |
|
Image pull policy |
telegraf-kafka-consumer.image.repo |
string |
|
Telegraf image repository |
telegraf-kafka-consumer.image.tag |
string |
|
Telegraf image tag |
telegraf-kafka-consumer.imagePullSecrets |
list |
|
Secret names to use for Docker pulls |
telegraf-kafka-consumer.influxdb.database |
string |
|
Name of the InfluxDB v1 database to write to |
telegraf-kafka-consumer.kafkaConsumers.test.enabled |
bool |
|
Enable the Telegraf Kafka consumer. |
telegraf-kafka-consumer.kafkaConsumers.test.fields |
list |
|
List of Avro fields to be recorded as InfluxDB fields. If not specified, any Avro field that is not marked as a tag will become an InfluxDB field. |
telegraf-kafka-consumer.kafkaConsumers.test.flush_interval |
string |
|
Default data flushing interval to InfluxDB. |
telegraf-kafka-consumer.kafkaConsumers.test.interval |
string |
|
Data collection interval for the Kafka consumer. |
telegraf-kafka-consumer.kafkaConsumers.test.replicaCount |
int |
|
Number of Telegraf Kafka consumer replicas. Increase this value to increase the consumer throughput. |
telegraf-kafka-consumer.kafkaConsumers.test.tags |
list |
|
List of Avro fields to be recorded as InfluxDB tags. The Avro fields specified as tags will be converted to strings before ingestion into InfluxDB. |
telegraf-kafka-consumer.kafkaConsumers.test.timestamp_field |
string |
|
Avro field to be used as the InfluxDB timestamp (optional). If unspecified or set to the empty string, Telegraf will use the time it received the measurement. |
telegraf-kafka-consumer.kafkaConsumers.test.timestamp_format |
string |
|
Timestamp format. Possible values are |
telegraf-kafka-consumer.kafkaConsumers.test.topicRegexps |
string |
|
List of regular expressions to specify the Kafka topics consumed by this agent. |
telegraf-kafka-consumer.kafkaConsumers.test.union_field_separator |
string |
|
Union field separator: if a single Avro field is flattened into more than one InfluxDB field (e.g. an array |
telegraf-kafka-consumer.kafkaConsumers.test.union_mode |
string |
|
Union mode: this can be one of |
telegraf-kafka-consumer.nodeSelector |
object |
|
Node labels for pod assignment |
telegraf-kafka-consumer.podAnnotations |
object |
|
Annotations for telegraf-kafka-consumers pods |
telegraf-kafka-consumer.podLabels |
object |
|
Labels for telegraf-kafka-consumer pods |
telegraf-kafka-consumer.resources |
object |
See |
Kubernetes resources requests and limits |
telegraf-kafka-consumer.tolerations |
list |
|
Tolerations for pod assignment |