bigquery-kafka Helm values reference#
Helm values reference table for the bigquery-kafka application.
Key |
Type |
Default |
Description |
|---|---|---|---|
config.backendApiTimeout |
string |
|
Timeout for backend API calls |
config.backendPollInterval |
string |
|
Interval at which the backend is polled for query status in Safir |
config.backendRetryCount |
int |
|
How many times to retry after a backend API network failure |
config.backendRetryDelay |
string |
|
How long to wait between retries after a backend API network failure in Safir |
config.bigqueryLocation |
string |
|
BigQuery processing location |
config.bigqueryMaxBytesBilled |
int |
100 GB |
Maximum bytes that can be billed for a single BigQuery query. Queries exceeding this will fail. Set to null for no limit. |
config.bigqueryProject |
string |
None, must be set |
GCP project ID containing the BigQuery datasets to query |
config.consumerGroupId |
string |
|
Kafka consumer group ID |
config.enabledBackend |
string |
|
Database backend to use (QSERV or BIGQUERY) |
config.gcpServiceAccount |
string |
None, must be set for BigQuery backend |
GCP service account email for Workload Identity Format: {name}@{project-id}.iam.gserviceaccount.com |
config.jobCancelTopic |
string |
|
Kafka topic for query cancellation requests |
config.jobRunBatchSize |
int |
|
Maximum batch size for query execution requests. This should generally be the same as |
config.jobRunMaxBytes |
int |
10MiB |
Maximum size of a batch read from Kafka in bytes. Wide queries can be up to 500KiB in size, so this should be at least 500KiB * 10. |
config.jobRunTopic |
string |
|
Kafka topic for query execution requests |
config.jobStatusTopic |
string |
|
Kafka topic for query status |
config.logLevel |
string |
|
Logging level |
config.logProfile |
string |
|
Logging profile ( |
config.maxWorkerJobs |
int |
|
Maximum number of arq jobs each worker can process simultaneously |
config.metrics.application |
string |
|
Name under which to log metrics. Generally there is no reason to change this. |
config.metrics.enabled |
bool |
|
Whether to enable sending metrics |
config.metrics.events.topicPrefix |
string |
|
Topic prefix for events. It may sometimes be useful to change this in development environments. |
config.metrics.schemaManager.registryUrl |
string |
Sasquatch in the local cluster |
URL of the Confluent-compatible schema registry server |
config.metrics.schemaManager.suffix |
string |
|
Suffix to add to all registered subjects. This is sometimes useful for experimentation during development. |
config.redisMaxConnections |
int |
|
Size of the Redis connection pool. This should be set to |
config.resultTimeout |
int |
3600 (1 hour) |
How long to wait for result processing (retrieval and upload) before timing out, in seconds. This doubles as the timeout forcibly terminating result worker pods. |
config.sentry.enabled |
bool |
|
Set to true to enable the Sentry integration. |
config.sentry.tracesSampleRate |
float |
|
The percentage of requests that should be traced. This should be a float between 0 and 1 |
config.slack.enabled |
bool |
|
Set to true to enable the Slack integration. If true, the slack-webhook secret must be provided. |
config.tapService |
string |
|
Name of the TAP service for which this BigQuery Kafka instance is managing queries. This must match the name of the TAP service for the corresponding query quota in the Gafaelfawr configuration. |
frontend.affinity |
object |
|
Affinity rules for the bigquery-kafka frontend pod |
frontend.debug.disablePymalloc |
bool |
|
|
frontend.debug.enabled |
bool |
|
Set to true to allow containers to run as root and to create and mount a debug PVC. Useful ro run debug containers to diagnose issues such as memory leaks. |
frontend.nodeSelector |
object |
|
Node selection rules for the bigquery-kafka frontend pod |
frontend.podAnnotations |
object |
|
Annotations for the bigquery-kafka frontend pod |
frontend.resources |
object |
See |
Resource limits and requests for the bigquery-kafka frontend pod |
frontend.tolerations |
list |
Tolerate GKE arm64 taint |
Tolerations for the bigquery-kafka frontend pod |
global.environmentName |
string |
Set by Argo CD Application |
Name of the Phalanx environment |
global.host |
string |
Set by Argo CD |
Host name for ingress |
global.repertoireUrl |
string |
Set by Argo CD |
Base URL for Repertoire discovery API |
global.vaultSecretsPath |
string |
Set by Argo CD |
Base path for Vault secrets |
image.pullPolicy |
string |
|
Pull policy for the bigquery-kafka image |
image.repository |
string |
|
Image to use in the bigquery-kafka deployment |
image.tag |
string |
The appVersion of the chart |
Tag of image to use |
ingress.annotations |
object |
|
Additional annotations for the ingress rule |
periodicMetrics.affinity |
object |
|
Affinity rules for the bigquery-kafka metrics job |
periodicMetrics.nodeSelector |
object |
|
Node selection rules for the bigquery-kafka metrics job |
periodicMetrics.podAnnotations |
object |
|
Annotations for the bigquery-kafka metrics job |
periodicMetrics.resources |
object |
See |
Resource limits and requests for the bigquery-kafka periodic metrics pods |
periodicMetrics.schedule |
string |
|
How often to run the periodic metrics job |
periodicMetrics.tolerations |
list |
Tolerate GKE arm64 taint |
Tolerations for the bigquery-kafka metrics job |
redis.config.secretKey |
string |
|
Key inside secret from which to get the Redis password (do not change) |
redis.config.secretName |
string |
|
Name of secret containing Redis password |
redis.persistence.accessMode |
string |
|
Access mode of storage to request |
redis.persistence.enabled |
bool |
|
Whether to persist Redis storage. Setting this to false will use |
redis.persistence.size |
string |
|
Amount of persistent storage to request |
redis.persistence.storageClass |
string |
|
Class of storage to request |
redis.persistence.volumeClaimName |
string |
|
Use an existing PVC, not dynamic provisioning. If this is set, the size, storageClass, and accessMode settings are ignored. |
redis.resources |
object |
See |
Resource limits and requests for the Redis pod |
redis.tolerations |
list |
Tolerate GKE arm64 taint |
Tolerations for the bigquery-kafka Redis pod |
resultWorker.affinity |
object |
|
Affinity rules for the bigquery-kafka worker pods |
resultWorker.allowRootDebug |
bool |
|
Whether to allow containers to run as root. Set to true to allow use of debug containers to diagnose issues such as memory leaks. |
resultWorker.autoscaling.enabled |
bool |
|
Enable autoscaling of bigquery-kafka result workers |
resultWorker.autoscaling.maxReplicas |
int |
|
Maximum number of bigquery-kafka worker pods. Each replica will open database connections up to the configured pool size and overflow limits, so make sure the combined connections are under the postgres connection limit. |
resultWorker.autoscaling.minReplicas |
int |
|
Minimum number of bigquery-kafka worker pods |
resultWorker.autoscaling.targetCPUUtilizationPercentage |
int |
|
Target CPU utilization of bigquery-kafka worker pods. |
resultWorker.nodeSelector |
object |
|
Node selection rules for the bigquery-kafka worker pods |
resultWorker.podAnnotations |
object |
|
Annotations for the bigquery-kafka worker pods |
resultWorker.replicaCount |
int |
|
Number of result worker pods to start if autoscaling is disabled |
resultWorker.resources |
object |
See |
Resource limits and requests for the bigquery-kafka worker pods |
resultWorker.tolerations |
list |
Tolerate GKE arm64 taint |
Tolerations for the bigquery-kafka worker pods |