From 6dbd4a3ae42984ef2f3b9c1823d5e4176cbcb3a9 Mon Sep 17 00:00:00 2001 From: "dependabot[bot]" <49699333+dependabot[bot]@users.noreply.github.com> Date: Mon, 4 Nov 2024 00:28:28 +0000 Subject: [PATCH 1/2] chore(deps): bump the main group with 2 updates Bumps the main group with 2 updates: [github.com/aiven/go-api-schemas](https://github.com/aiven/go-api-schemas) and [github.com/aiven/go-client-codegen](https://github.com/aiven/go-client-codegen). Updates `github.com/aiven/go-api-schemas` from 1.96.0 to 1.97.0 - [Release notes](https://github.com/aiven/go-api-schemas/releases) - [Commits](https://github.com/aiven/go-api-schemas/compare/v1.96.0...v1.97.0) Updates `github.com/aiven/go-client-codegen` from 0.48.0 to 0.49.0 - [Release notes](https://github.com/aiven/go-client-codegen/releases) - [Commits](https://github.com/aiven/go-client-codegen/compare/v0.48.0...v0.49.0) --- updated-dependencies: - dependency-name: github.com/aiven/go-api-schemas dependency-type: direct:production update-type: version-update:semver-minor dependency-group: main - dependency-name: github.com/aiven/go-client-codegen dependency-type: direct:production update-type: version-update:semver-minor dependency-group: main ... Signed-off-by: dependabot[bot] --- go.mod | 4 ++-- go.sum | 8 ++++---- 2 files changed, 6 insertions(+), 6 deletions(-) diff --git a/go.mod b/go.mod index 27b3bad4..3646fd47 100644 --- a/go.mod +++ b/go.mod @@ -7,8 +7,8 @@ toolchain go1.23.0 require ( github.com/ClickHouse/clickhouse-go/v2 v2.30.0 github.com/aiven/aiven-go-client/v2 v2.30.0 - github.com/aiven/go-api-schemas v1.96.0 - github.com/aiven/go-client-codegen v0.48.0 + github.com/aiven/go-api-schemas v1.97.0 + github.com/aiven/go-client-codegen v0.49.0 github.com/avast/retry-go v3.0.0+incompatible github.com/dave/jennifer v1.7.1 github.com/docker/go-units v0.5.0 diff --git a/go.sum b/go.sum index 28b06166..efe90cdc 100644 --- a/go.sum +++ b/go.sum @@ -39,10 +39,10 @@ github.com/ClickHouse/clickhouse-go/v2 v2.30.0 h1:AG4D/hW39qa58+JHQIFOSnxyL46H6h github.com/ClickHouse/clickhouse-go/v2 v2.30.0/go.mod h1:i9ZQAojcayW3RsdCb3YR+n+wC2h65eJsZCscZ1Z1wyo= github.com/aiven/aiven-go-client/v2 v2.30.0 h1:dj1nRuO1XglnwH2IwKmqEl3SGaqKebDoxnd+SRjxQRg= github.com/aiven/aiven-go-client/v2 v2.30.0/go.mod h1:Eyxa+fNgayObmUBW94uJuEkyOe1646cEpjFzhm/NETY= -github.com/aiven/go-api-schemas v1.96.0 h1:h/8ENjiyj6G2fliXGygLgWfTjzFT7wNC86WOz9IYHow= -github.com/aiven/go-api-schemas v1.96.0/go.mod h1:ATdCq7aRp+URkFI4W4mVZaNZgoN5FRicXKH3c2fpgW4= -github.com/aiven/go-client-codegen v0.48.0 h1:4etkt+/tBMHzj8JJJDO1MK6pG33+2fTjI3MeDt5wB7w= -github.com/aiven/go-client-codegen v0.48.0/go.mod h1:FfbH32Xb+Hx5zeKTIug1Y8SfMeB+AKNRzxgrzkts2oA= +github.com/aiven/go-api-schemas v1.97.0 h1:/ykiTRissDfcrgiDyNyOx8PvByeMRHZt2ikwqUVruNA= +github.com/aiven/go-api-schemas v1.97.0/go.mod h1:+E57naSaeLp5l8x7CPEBOMiuAbPNYyMiisPqx3NTqmM= +github.com/aiven/go-client-codegen v0.49.0 h1:KwpKVykr4xWcduQ4h7+ad7Y1rAYj9Z/Tq5iKhu7/guo= +github.com/aiven/go-client-codegen v0.49.0/go.mod h1:FfbH32Xb+Hx5zeKTIug1Y8SfMeB+AKNRzxgrzkts2oA= github.com/alecthomas/template v0.0.0-20160405071501-a0175ee3bccc/go.mod h1:LOuyumcjzFXgccqObfd/Ljyb9UuFJ6TxHnclSeseNhc= github.com/alecthomas/template v0.0.0-20190718012654-fb15b899a751/go.mod h1:LOuyumcjzFXgccqObfd/Ljyb9UuFJ6TxHnclSeseNhc= github.com/alecthomas/units v0.0.0-20151022065526-2efee857e7cf/go.mod h1:ybxpYRFXyAe+OPACYpWeL0wqObRcbAqCMya13uyzqw0= From 1538493b6f3fbae1fde6642506d4a62aa0393035 Mon Sep 17 00:00:00 2001 From: "dependabot[bot]" Date: Mon, 4 Nov 2024 00:29:53 +0000 Subject: [PATCH 2/2] chore(userconfigs): generate files --- CHANGELOG.md | 64 +++++++++++++++++++ .../clickhouse_kafka/clickhouse_kafka.go | 4 +- .../kafka_mirrormaker/kafka_mirrormaker.go | 2 +- .../integrationendpoints/datadog/datadog.go | 2 +- .../external_kafka/external_kafka.go | 2 +- .../external_schema_registry.go | 2 +- .../integrationendpoints/rsyslog/rsyslog.go | 2 +- .../userconfig/service/cassandra/cassandra.go | 3 +- .../userconfig/service/grafana/grafana.go | 8 +-- .../userconfig/service/kafka/kafka.go | 16 ++--- .../service/kafka_connect/kafka_connect.go | 6 +- .../userconfig/service/mysql/mysql.go | 4 +- .../service/opensearch/opensearch.go | 2 +- api/v1alpha1/userconfig/service/pg/pg.go | 12 ++-- .../userconfig/service/redis/redis.go | 2 +- .../templates/aiven.io_cassandras.yaml | 2 +- .../templates/aiven.io_grafanas.yaml | 8 +-- .../templates/aiven.io_kafkaconnects.yaml | 8 +-- .../templates/aiven.io_kafkas.yaml | 24 +++---- .../templates/aiven.io_mysqls.yaml | 4 +- .../templates/aiven.io_opensearches.yaml | 4 +- .../templates/aiven.io_postgresqls.yaml | 14 ++-- .../templates/aiven.io_redis.yaml | 8 +-- .../aiven.io_serviceintegrationendpoints.yaml | 12 ++-- .../aiven.io_serviceintegrations.yaml | 16 ++--- config/crd/bases/aiven.io_cassandras.yaml | 2 +- config/crd/bases/aiven.io_grafanas.yaml | 8 +-- config/crd/bases/aiven.io_kafkaconnects.yaml | 8 +-- config/crd/bases/aiven.io_kafkas.yaml | 24 +++---- config/crd/bases/aiven.io_mysqls.yaml | 4 +- config/crd/bases/aiven.io_opensearches.yaml | 4 +- config/crd/bases/aiven.io_postgresqls.yaml | 14 ++-- config/crd/bases/aiven.io_redis.yaml | 8 +-- .../aiven.io_serviceintegrationendpoints.yaml | 12 ++-- .../bases/aiven.io_serviceintegrations.yaml | 16 ++--- docs/docs/api-reference/cassandra.md | 2 +- docs/docs/api-reference/grafana.md | 8 +-- docs/docs/api-reference/kafka.md | 16 ++--- docs/docs/api-reference/kafkaconnect.md | 6 +- docs/docs/api-reference/mysql.md | 4 +- docs/docs/api-reference/opensearch.md | 2 +- docs/docs/api-reference/postgresql.md | 12 ++-- docs/docs/api-reference/redis.md | 2 +- docs/docs/api-reference/serviceintegration.md | 6 +- .../serviceintegrationendpoint.md | 8 +-- 45 files changed, 230 insertions(+), 167 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index de774d33..e4207afa 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -40,6 +40,70 @@ aliases alongside their associated indexes. Default is true - Add `OpenSearch` field `userConfig.s3_migration.include_aliases`, type `boolean`: Whether to restore aliases alongside their associated indexes. Default is true +- Change `Grafana` field `userConfig.alerting_nodata_or_nullvalues`: enum ~~`[alerting, keep_state, no_data, + ok]`~~ → `[alerting, keep_state, no_data, ok]` +- Change `Grafana` field `userConfig.cookie_samesite`: enum ~~`[lax, none, strict]`~~ → `[lax, none, + strict]` +- Change `Grafana` field `userConfig.smtp_server.starttls_policy`: enum ~~`[MandatoryStartTLS, NoStartTLS, + OpportunisticStartTLS]`~~ → `[MandatoryStartTLS, NoStartTLS, OpportunisticStartTLS]` +- Change `Grafana` field `userConfig.user_auto_assign_org_role`: enum ~~`[Admin, Editor, Viewer]`~~ → + `[Admin, Editor, Viewer]` +- Change `KafkaConnect` field `userConfig.kafka_connect.connector_client_config_override_policy`: enum + ~~`[All, None]`~~ → `[All, None]` +- Change `KafkaConnect` field `userConfig.kafka_connect.consumer_isolation_level`: enum ~~`[read_committed, + read_uncommitted]`~~ → `[read_committed, read_uncommitted]` +- Change `KafkaConnect` field `userConfig.kafka_connect.producer_compression_type`: enum ~~`[gzip, lz4, + none, snappy, zstd]`~~ → `[gzip, lz4, none, snappy, zstd]` +- Change `Kafka` field `userConfig.kafka.compression_type`: enum ~~`[gzip, lz4, producer, snappy, uncompressed, + zstd]`~~ → `[gzip, lz4, producer, snappy, uncompressed, zstd]` +- Change `Kafka` field `userConfig.kafka.log_cleanup_policy`: enum ~~`[compact, compact,delete, delete]`~~ + → `[compact, compact,delete, delete]` +- Change `Kafka` field `userConfig.kafka_connect_config.connector_client_config_override_policy`: enum + ~~`[All, None]`~~ → `[All, None]` +- Change `Kafka` field `userConfig.kafka_connect_config.consumer_isolation_level`: enum ~~`[read_committed, + read_uncommitted]`~~ → `[read_committed, read_uncommitted]` +- Change `Kafka` field `userConfig.kafka_connect_config.producer_compression_type`: enum ~~`[gzip, lz4, + none, snappy, zstd]`~~ → `[gzip, lz4, none, snappy, zstd]` +- Change `Kafka` field `userConfig.kafka_rest_config.name_strategy`: enum ~~`[record_name, topic_name, + topic_record_name]`~~ → `[record_name, topic_name, topic_record_name]` +- Change `Kafka` field `userConfig.kafka_rest_config.producer_acks`: enum ~~`[-1, 0, 1, all]`~~ → `[-1, + 0, 1, all]` +- Change `Kafka` field `userConfig.kafka_rest_config.producer_compression_type`: enum ~~`[gzip, lz4, + none, snappy, zstd]`~~ → `[gzip, lz4, none, snappy, zstd]` +- Change `MySQL` field `userConfig.mysql.internal_tmp_mem_storage_engine`: enum ~~`[MEMORY, TempTable]`~~ + → `[MEMORY, TempTable]` +- Change `MySQL` field `userConfig.mysql.log_output`: enum ~~`[INSIGHTS, INSIGHTS,TABLE, NONE, TABLE]`~~ + → `[INSIGHTS, INSIGHTS,TABLE, NONE, TABLE]` +- Change `OpenSearch` field `userConfig.opensearch.search_backpressure.mode`: enum ~~`[disabled, enforced, + monitor_only]`~~ → `[disabled, enforced, monitor_only]` +- Change `PostgreSQL` field `userConfig.pg.log_error_verbosity`: enum ~~`[DEFAULT, TERSE, VERBOSE]`~~ + → `[DEFAULT, TERSE, VERBOSE]` +- Change `PostgreSQL` field `userConfig.pg.log_line_prefix`: enum ~~`['%m [%p] %q[user=%u,db=%d,app=%a] + ', '%t [%p]: [%l-1] user=%u,db=%d,app=%a,client=%h ', 'pid=%p,user=%u,db=%d,app=%a,client=%h ', + 'pid=%p,user=%u,db=%d,app=%a,client=%h,txid=%x,qid=%Q ']`~~ → `['%m [%p] %q[user=%u,db=%d,app=%a] + ', '%t [%p]: [%l-1] user=%u,db=%d,app=%a,client=%h ', 'pid=%p,user=%u,db=%d,app=%a,client=%h ', + 'pid=%p,user=%u,db=%d,app=%a,client=%h,txid=%x,qid=%Q ']` +- Change `PostgreSQL` field `userConfig.pg.pg_stat_statements.track`: enum ~~`[all, none, top]`~~ → + `[all, none, top]` +- Change `PostgreSQL` field `userConfig.pg.track_functions`: enum ~~`[all, none, pl]`~~ → `[all, none, pl]` +- Change `PostgreSQL` field `userConfig.pgbouncer.autodb_pool_mode`: enum ~~`[session, statement, transaction]`~~ + → `[session, statement, transaction]` +- Change `PostgreSQL` field `userConfig.synchronous_replication`: enum ~~`[off, quorum]`~~ → `[off, + quorum]` +- Change `Redis` field `userConfig.redis_maxmemory_policy`: enum ~~`[allkeys-lfu, allkeys-lru, allkeys-random, + noeviction, volatile-lfu, volatile-lru, volatile-random, volatile-ttl]`~~ → `[allkeys-lfu, + allkeys-lru, allkeys-random, noeviction, volatile-lfu, volatile-lru, volatile-random, volatile-ttl]` +- Change `ServiceIntegrationEndpoint` field `datadog.site`: enum ~~`[ap1.datadoghq.com, datadoghq.com, + datadoghq.eu, ddog-gov.com, us3.datadoghq.com, us5.datadoghq.com]`~~ → `[ap1.datadoghq.com, datadoghq.com, + datadoghq.eu, ddog-gov.com, us3.datadoghq.com, us5.datadoghq.com]` +- Change `ServiceIntegrationEndpoint` field `externalKafka.security_protocol`: enum ~~`[PLAINTEXT, SASL_PLAINTEXT, + SASL_SSL, SSL]`~~ → `[PLAINTEXT, SASL_PLAINTEXT, SASL_SSL, SSL]` +- Change `ServiceIntegrationEndpoint` field `externalSchemaRegistry.authentication`: enum ~~`[basic, + none]`~~ → `[basic, none]` +- Change `ServiceIntegrationEndpoint` field `rsyslog.format`: enum ~~`[custom, rfc3164, rfc5424]`~~ → + `[custom, rfc3164, rfc5424]` +- Change `ServiceIntegration` field `kafkaMirrormaker.kafka_mirrormaker.producer_compression_type`: enum + ~~`[gzip, lz4, none, snappy, zstd]`~~ → `[gzip, lz4, none, snappy, zstd]` ## v0.25.0 - 2024-09-19 diff --git a/api/v1alpha1/userconfig/integration/clickhouse_kafka/clickhouse_kafka.go b/api/v1alpha1/userconfig/integration/clickhouse_kafka/clickhouse_kafka.go index fe28ad94..9bba9a38 100644 --- a/api/v1alpha1/userconfig/integration/clickhouse_kafka/clickhouse_kafka.go +++ b/api/v1alpha1/userconfig/integration/clickhouse_kafka/clickhouse_kafka.go @@ -26,7 +26,7 @@ type Topics struct { // Table to create type Tables struct { - // +kubebuilder:validation:Enum="smallest";"earliest";"beginning";"largest";"latest";"end" + // +kubebuilder:validation:Enum="beginning";"earliest";"end";"largest";"latest";"smallest" // Action to take when there is no initial offset in offset store or the desired offset is out of range AutoOffsetReset *string `groups:"create,update" json:"auto_offset_reset,omitempty"` @@ -34,7 +34,7 @@ type Tables struct { // Table columns Columns []*Columns `groups:"create,update" json:"columns"` - // +kubebuilder:validation:Enum="Avro";"CSV";"JSONAsString";"JSONCompactEachRow";"JSONCompactStringsEachRow";"JSONEachRow";"JSONStringsEachRow";"MsgPack";"TSKV";"TSV";"TabSeparated";"RawBLOB";"AvroConfluent";"Parquet" + // +kubebuilder:validation:Enum="Avro";"AvroConfluent";"CSV";"JSONAsString";"JSONCompactEachRow";"JSONCompactStringsEachRow";"JSONEachRow";"JSONStringsEachRow";"MsgPack";"Parquet";"RawBLOB";"TSKV";"TSV";"TabSeparated" // Message data format DataFormat string `groups:"create,update" json:"data_format"` diff --git a/api/v1alpha1/userconfig/integration/kafka_mirrormaker/kafka_mirrormaker.go b/api/v1alpha1/userconfig/integration/kafka_mirrormaker/kafka_mirrormaker.go index b091f00d..64e66df4 100644 --- a/api/v1alpha1/userconfig/integration/kafka_mirrormaker/kafka_mirrormaker.go +++ b/api/v1alpha1/userconfig/integration/kafka_mirrormaker/kafka_mirrormaker.go @@ -29,7 +29,7 @@ type KafkaMirrormaker struct { // The amount of bytes producer can use for buffering data before publishing to broker. ProducerBufferMemory *int `groups:"create,update" json:"producer_buffer_memory,omitempty"` - // +kubebuilder:validation:Enum="gzip";"snappy";"lz4";"zstd";"none" + // +kubebuilder:validation:Enum="gzip";"lz4";"none";"snappy";"zstd" // Specify the default compression type for producers. This configuration accepts the standard compression codecs ('gzip', 'snappy', 'lz4', 'zstd'). It additionally accepts 'none' which is the default and equivalent to no compression. ProducerCompressionType *string `groups:"create,update" json:"producer_compression_type,omitempty"` diff --git a/api/v1alpha1/userconfig/integrationendpoints/datadog/datadog.go b/api/v1alpha1/userconfig/integrationendpoints/datadog/datadog.go index 0cb36c9b..182dc5a4 100644 --- a/api/v1alpha1/userconfig/integrationendpoints/datadog/datadog.go +++ b/api/v1alpha1/userconfig/integrationendpoints/datadog/datadog.go @@ -43,7 +43,7 @@ type DatadogUserConfig struct { // Maximum number of partition contexts to send MaxPartitionContexts *int `groups:"create,update" json:"max_partition_contexts,omitempty"` - // +kubebuilder:validation:Enum="datadoghq.com";"datadoghq.eu";"us3.datadoghq.com";"us5.datadoghq.com";"ddog-gov.com";"ap1.datadoghq.com" + // +kubebuilder:validation:Enum="ap1.datadoghq.com";"datadoghq.com";"datadoghq.eu";"ddog-gov.com";"us3.datadoghq.com";"us5.datadoghq.com" // Datadog intake site. Defaults to datadoghq.com Site *string `groups:"create,update" json:"site,omitempty"` } diff --git a/api/v1alpha1/userconfig/integrationendpoints/external_kafka/external_kafka.go b/api/v1alpha1/userconfig/integrationendpoints/external_kafka/external_kafka.go index 2371d4ef..0b182f15 100644 --- a/api/v1alpha1/userconfig/integrationendpoints/external_kafka/external_kafka.go +++ b/api/v1alpha1/userconfig/integrationendpoints/external_kafka/external_kafka.go @@ -23,7 +23,7 @@ type ExternalKafkaUserConfig struct { // Username for SASL PLAIN mechanism in the Kafka server. SaslPlainUsername *string `groups:"create,update" json:"sasl_plain_username,omitempty"` - // +kubebuilder:validation:Enum="PLAINTEXT";"SSL";"SASL_PLAINTEXT";"SASL_SSL" + // +kubebuilder:validation:Enum="PLAINTEXT";"SASL_PLAINTEXT";"SASL_SSL";"SSL" // Security protocol SecurityProtocol string `groups:"create,update" json:"security_protocol"` diff --git a/api/v1alpha1/userconfig/integrationendpoints/external_schema_registry/external_schema_registry.go b/api/v1alpha1/userconfig/integrationendpoints/external_schema_registry/external_schema_registry.go index 2f835562..86551c00 100644 --- a/api/v1alpha1/userconfig/integrationendpoints/external_schema_registry/external_schema_registry.go +++ b/api/v1alpha1/userconfig/integrationendpoints/external_schema_registry/external_schema_registry.go @@ -4,7 +4,7 @@ package externalschemaregistryuserconfig type ExternalSchemaRegistryUserConfig struct { - // +kubebuilder:validation:Enum="none";"basic" + // +kubebuilder:validation:Enum="basic";"none" // Authentication method Authentication string `groups:"create,update" json:"authentication"` diff --git a/api/v1alpha1/userconfig/integrationendpoints/rsyslog/rsyslog.go b/api/v1alpha1/userconfig/integrationendpoints/rsyslog/rsyslog.go index 53fcfd29..5840e3b6 100644 --- a/api/v1alpha1/userconfig/integrationendpoints/rsyslog/rsyslog.go +++ b/api/v1alpha1/userconfig/integrationendpoints/rsyslog/rsyslog.go @@ -12,7 +12,7 @@ type RsyslogUserConfig struct { // PEM encoded client certificate Cert *string `groups:"create,update" json:"cert,omitempty"` - // +kubebuilder:validation:Enum="rfc5424";"rfc3164";"custom" + // +kubebuilder:validation:Enum="custom";"rfc3164";"rfc5424" // Message format Format string `groups:"create,update" json:"format"` diff --git a/api/v1alpha1/userconfig/service/cassandra/cassandra.go b/api/v1alpha1/userconfig/service/cassandra/cassandra.go index 217ed5f5..111bc981 100644 --- a/api/v1alpha1/userconfig/service/cassandra/cassandra.go +++ b/api/v1alpha1/userconfig/service/cassandra/cassandra.go @@ -54,8 +54,7 @@ type PublicAccess struct { } type CassandraUserConfig struct { // +kubebuilder:validation:MaxItems=1 - // +kubebuilder:deprecatedversion:warning="additional_backup_regions is deprecated" - // Deprecated. Additional Cloud Regions for Backup Replication + // Additional Cloud Regions for Backup Replication AdditionalBackupRegions []string `groups:"create,update" json:"additional_backup_regions,omitempty"` // +kubebuilder:validation:Minimum=0 diff --git a/api/v1alpha1/userconfig/service/grafana/grafana.go b/api/v1alpha1/userconfig/service/grafana/grafana.go index b7c4af3c..c3376394 100644 --- a/api/v1alpha1/userconfig/service/grafana/grafana.go +++ b/api/v1alpha1/userconfig/service/grafana/grafana.go @@ -290,7 +290,7 @@ type SmtpServer struct { // Skip verifying server certificate. Defaults to false SkipVerify *bool `groups:"create,update" json:"skip_verify,omitempty"` - // +kubebuilder:validation:Enum="OpportunisticStartTLS";"MandatoryStartTLS";"NoStartTLS" + // +kubebuilder:validation:Enum="MandatoryStartTLS";"NoStartTLS";"OpportunisticStartTLS" // Either OpportunisticStartTLS, MandatoryStartTLS or NoStartTLS. Default is OpportunisticStartTLS. StarttlsPolicy *string `groups:"create,update" json:"starttls_policy,omitempty"` @@ -316,7 +316,7 @@ type GrafanaUserConfig struct { // Max number of alert annotations that Grafana stores. 0 (default) keeps all alert annotations. AlertingMaxAnnotationsToKeep *int `groups:"create,update" json:"alerting_max_annotations_to_keep,omitempty"` - // +kubebuilder:validation:Enum="alerting";"no_data";"keep_state";"ok" + // +kubebuilder:validation:Enum="alerting";"keep_state";"no_data";"ok" // Default value for 'no data or null values' for new alerting rules AlertingNodataOrNullvalues *string `groups:"create,update" json:"alerting_nodata_or_nullvalues,omitempty"` @@ -341,7 +341,7 @@ type GrafanaUserConfig struct { // Google Auth integration AuthGoogle *AuthGoogle `groups:"create,update" json:"auth_google,omitempty"` - // +kubebuilder:validation:Enum="lax";"strict";"none" + // +kubebuilder:validation:Enum="lax";"none";"strict" // Cookie SameSite attribute: 'strict' prevents sending cookie for cross-site requests, effectively disabling direct linking from other sites to Grafana. 'lax' is the default value. CookieSamesite *string `groups:"create,update" json:"cookie_samesite,omitempty"` @@ -438,7 +438,7 @@ type GrafanaUserConfig struct { // Auto-assign new users on signup to main organization. Defaults to false UserAutoAssignOrg *bool `groups:"create,update" json:"user_auto_assign_org,omitempty"` - // +kubebuilder:validation:Enum="Viewer";"Admin";"Editor" + // +kubebuilder:validation:Enum="Admin";"Editor";"Viewer" // Set role for new signups. Defaults to Viewer UserAutoAssignOrgRole *string `groups:"create,update" json:"user_auto_assign_org_role,omitempty"` diff --git a/api/v1alpha1/userconfig/service/kafka/kafka.go b/api/v1alpha1/userconfig/service/kafka/kafka.go index 14191f9e..5a3f575b 100644 --- a/api/v1alpha1/userconfig/service/kafka/kafka.go +++ b/api/v1alpha1/userconfig/service/kafka/kafka.go @@ -25,7 +25,7 @@ type Kafka struct { // Enable auto-creation of topics. (Default: true) AutoCreateTopicsEnable *bool `groups:"create,update" json:"auto_create_topics_enable,omitempty"` - // +kubebuilder:validation:Enum="gzip";"snappy";"lz4";"zstd";"uncompressed";"producer" + // +kubebuilder:validation:Enum="gzip";"lz4";"producer";"snappy";"uncompressed";"zstd" // Specify the final compression type for a given topic. This configuration accepts the standard compression codecs ('gzip', 'snappy', 'lz4', 'zstd'). It additionally accepts 'uncompressed' which is equivalent to no compression; and 'producer' which means retain the original compression codec set by the producer.(Default: producer) CompressionType *string `groups:"create,update" json:"compression_type,omitempty"` @@ -72,7 +72,7 @@ type Kafka struct { // The minimum time a message will remain uncompacted in the log. Only applicable for logs that are being compacted. (Default: 0 ms) LogCleanerMinCompactionLagMs *int `groups:"create,update" json:"log_cleaner_min_compaction_lag_ms,omitempty"` - // +kubebuilder:validation:Enum="delete";"compact";"compact,delete" + // +kubebuilder:validation:Enum="compact";"compact,delete";"delete" // The default cleanup policy for segments beyond the retention window (Default: delete) LogCleanupPolicy *string `groups:"create,update" json:"log_cleanup_policy,omitempty"` @@ -241,7 +241,7 @@ type KafkaAuthenticationMethods struct { // Kafka Connect configuration values type KafkaConnectConfig struct { - // +kubebuilder:validation:Enum="None";"All" + // +kubebuilder:validation:Enum="All";"None" // Defines what client configurations can be overridden by the connector. Default is None ConnectorClientConfigOverridePolicy *string `groups:"create,update" json:"connector_client_config_override_policy,omitempty"` @@ -254,7 +254,7 @@ type KafkaConnectConfig struct { // Records are fetched in batches by the consumer, and if the first record batch in the first non-empty partition of the fetch is larger than this value, the record batch will still be returned to ensure that the consumer can make progress. As such, this is not a absolute maximum. ConsumerFetchMaxBytes *int `groups:"create,update" json:"consumer_fetch_max_bytes,omitempty"` - // +kubebuilder:validation:Enum="read_uncommitted";"read_committed" + // +kubebuilder:validation:Enum="read_committed";"read_uncommitted" // Transaction read isolation level. read_uncommitted is the default, but read_committed can be used if consume-exactly-once behavior is desired. ConsumerIsolationLevel *string `groups:"create,update" json:"consumer_isolation_level,omitempty"` @@ -293,7 +293,7 @@ type KafkaConnectConfig struct { // The total bytes of memory the producer can use to buffer records waiting to be sent to the broker (defaults to 33554432). ProducerBufferMemory *int `groups:"create,update" json:"producer_buffer_memory,omitempty"` - // +kubebuilder:validation:Enum="gzip";"snappy";"lz4";"zstd";"none" + // +kubebuilder:validation:Enum="gzip";"lz4";"none";"snappy";"zstd" // Specify the default compression type for producers. This configuration accepts the standard compression codecs ('gzip', 'snappy', 'lz4', 'zstd'). It additionally accepts 'none' which is the default and equivalent to no compression. ProducerCompressionType *string `groups:"create,update" json:"producer_compression_type,omitempty"` @@ -388,18 +388,18 @@ type KafkaRestConfig struct { // The maximum total time to wait for messages for a request if the maximum number of messages has not yet been reached ConsumerRequestTimeoutMs *int `groups:"create,update" json:"consumer_request_timeout_ms,omitempty"` - // +kubebuilder:validation:Enum="topic_name";"record_name";"topic_record_name" + // +kubebuilder:validation:Enum="record_name";"topic_name";"topic_record_name" // Name strategy to use when selecting subject for storing schemas NameStrategy *string `groups:"create,update" json:"name_strategy,omitempty"` // If true, validate that given schema is registered under expected subject name by the used name strategy when producing messages. NameStrategyValidation *bool `groups:"create,update" json:"name_strategy_validation,omitempty"` - // +kubebuilder:validation:Enum="all";"-1";"0";"1" + // +kubebuilder:validation:Enum="-1";"0";"1";"all" // The number of acknowledgments the producer requires the leader to have received before considering a request complete. If set to 'all' or '-1', the leader will wait for the full set of in-sync replicas to acknowledge the record. ProducerAcks *string `groups:"create,update" json:"producer_acks,omitempty"` - // +kubebuilder:validation:Enum="gzip";"snappy";"lz4";"zstd";"none" + // +kubebuilder:validation:Enum="gzip";"lz4";"none";"snappy";"zstd" // Specify the default compression type for producers. This configuration accepts the standard compression codecs ('gzip', 'snappy', 'lz4', 'zstd'). It additionally accepts 'none' which is the default and equivalent to no compression. ProducerCompressionType *string `groups:"create,update" json:"producer_compression_type,omitempty"` diff --git a/api/v1alpha1/userconfig/service/kafka_connect/kafka_connect.go b/api/v1alpha1/userconfig/service/kafka_connect/kafka_connect.go index 063dec8e..4d9411b3 100644 --- a/api/v1alpha1/userconfig/service/kafka_connect/kafka_connect.go +++ b/api/v1alpha1/userconfig/service/kafka_connect/kafka_connect.go @@ -16,7 +16,7 @@ type IpFilter struct { // Kafka Connect configuration values type KafkaConnect struct { - // +kubebuilder:validation:Enum="None";"All" + // +kubebuilder:validation:Enum="All";"None" // Defines what client configurations can be overridden by the connector. Default is None ConnectorClientConfigOverridePolicy *string `groups:"create,update" json:"connector_client_config_override_policy,omitempty"` @@ -29,7 +29,7 @@ type KafkaConnect struct { // Records are fetched in batches by the consumer, and if the first record batch in the first non-empty partition of the fetch is larger than this value, the record batch will still be returned to ensure that the consumer can make progress. As such, this is not a absolute maximum. ConsumerFetchMaxBytes *int `groups:"create,update" json:"consumer_fetch_max_bytes,omitempty"` - // +kubebuilder:validation:Enum="read_uncommitted";"read_committed" + // +kubebuilder:validation:Enum="read_committed";"read_uncommitted" // Transaction read isolation level. read_uncommitted is the default, but read_committed can be used if consume-exactly-once behavior is desired. ConsumerIsolationLevel *string `groups:"create,update" json:"consumer_isolation_level,omitempty"` @@ -68,7 +68,7 @@ type KafkaConnect struct { // The total bytes of memory the producer can use to buffer records waiting to be sent to the broker (defaults to 33554432). ProducerBufferMemory *int `groups:"create,update" json:"producer_buffer_memory,omitempty"` - // +kubebuilder:validation:Enum="gzip";"snappy";"lz4";"zstd";"none" + // +kubebuilder:validation:Enum="gzip";"lz4";"none";"snappy";"zstd" // Specify the default compression type for producers. This configuration accepts the standard compression codecs ('gzip', 'snappy', 'lz4', 'zstd'). It additionally accepts 'none' which is the default and equivalent to no compression. ProducerCompressionType *string `groups:"create,update" json:"producer_compression_type,omitempty"` diff --git a/api/v1alpha1/userconfig/service/mysql/mysql.go b/api/v1alpha1/userconfig/service/mysql/mysql.go index 54279075..afabc81f 100644 --- a/api/v1alpha1/userconfig/service/mysql/mysql.go +++ b/api/v1alpha1/userconfig/service/mysql/mysql.go @@ -136,11 +136,11 @@ type Mysql struct { // The number of seconds the server waits for activity on an interactive connection before closing it. InteractiveTimeout *int `groups:"create,update" json:"interactive_timeout,omitempty"` - // +kubebuilder:validation:Enum="TempTable";"MEMORY" + // +kubebuilder:validation:Enum="MEMORY";"TempTable" // The storage engine for in-memory internal temporary tables. InternalTmpMemStorageEngine *string `groups:"create,update" json:"internal_tmp_mem_storage_engine,omitempty"` - // +kubebuilder:validation:Enum="INSIGHTS";"NONE";"TABLE";"INSIGHTS,TABLE" + // +kubebuilder:validation:Enum="INSIGHTS";"INSIGHTS,TABLE";"NONE";"TABLE" // The slow log output destination when slow_query_log is ON. To enable MySQL AI Insights, choose INSIGHTS. To use MySQL AI Insights and the mysql.slow_log table at the same time, choose INSIGHTS,TABLE. To only use the mysql.slow_log table, choose TABLE. To silence slow logs, choose NONE. LogOutput *string `groups:"create,update" json:"log_output,omitempty"` diff --git a/api/v1alpha1/userconfig/service/opensearch/opensearch.go b/api/v1alpha1/userconfig/service/opensearch/opensearch.go index c5326079..3fb8619f 100644 --- a/api/v1alpha1/userconfig/service/opensearch/opensearch.go +++ b/api/v1alpha1/userconfig/service/opensearch/opensearch.go @@ -399,7 +399,7 @@ type SearchTask struct { // Search Backpressure Settings type SearchBackpressure struct { - // +kubebuilder:validation:Enum="monitor_only";"enforced";"disabled" + // +kubebuilder:validation:Enum="disabled";"enforced";"monitor_only" // The search backpressure mode. Valid values are monitor_only, enforced, or disabled. Default is monitor_only Mode *string `groups:"create,update" json:"mode,omitempty"` diff --git a/api/v1alpha1/userconfig/service/pg/pg.go b/api/v1alpha1/userconfig/service/pg/pg.go index a1963ebf..d28c0c0a 100644 --- a/api/v1alpha1/userconfig/service/pg/pg.go +++ b/api/v1alpha1/userconfig/service/pg/pg.go @@ -142,11 +142,11 @@ type Pg struct { // Causes each action executed by autovacuum to be logged if it ran for at least the specified number of milliseconds. Setting this to zero logs all autovacuum actions. Minus-one (the default) disables logging autovacuum actions. LogAutovacuumMinDuration *int `groups:"create,update" json:"log_autovacuum_min_duration,omitempty"` - // +kubebuilder:validation:Enum="TERSE";"DEFAULT";"VERBOSE" + // +kubebuilder:validation:Enum="DEFAULT";"TERSE";"VERBOSE" // Controls the amount of detail written in the server log for each message that is logged. LogErrorVerbosity *string `groups:"create,update" json:"log_error_verbosity,omitempty"` - // +kubebuilder:validation:Enum="'pid=%p,user=%u,db=%d,app=%a,client=%h '";"'%t [%p]: [%l-1] user=%u,db=%d,app=%a,client=%h '";"'%m [%p] %q[user=%u,db=%d,app=%a] '";"'pid=%p,user=%u,db=%d,app=%a,client=%h,txid=%x,qid=%Q '" + // +kubebuilder:validation:Enum="'%m [%p] %q[user=%u,db=%d,app=%a] '";"'%t [%p]: [%l-1] user=%u,db=%d,app=%a,client=%h '";"'pid=%p,user=%u,db=%d,app=%a,client=%h '";"'pid=%p,user=%u,db=%d,app=%a,client=%h,txid=%x,qid=%Q '" // Choose from one of the available log formats. LogLinePrefix *string `groups:"create,update" json:"log_line_prefix,omitempty"` @@ -248,7 +248,7 @@ type Pg struct { // Sets the maximum number of buckets PgStatMonitorPgsmMaxBuckets *int `groups:"create,update" json:"pg_stat_monitor.pgsm_max_buckets,omitempty"` - // +kubebuilder:validation:Enum="all";"top";"none" + // +kubebuilder:validation:Enum="all";"none";"top" // Controls which statements are counted. Specify top to track top-level statements (those issued directly by clients), all to also track nested statements (such as statements invoked within functions), or none to disable statement statistics collection. The default value is top. PgStatStatementsTrack *string `groups:"create,update" json:"pg_stat_statements.track,omitempty"` @@ -271,7 +271,7 @@ type Pg struct { // Record commit time of transactions. TrackCommitTimestamp *string `groups:"create,update" json:"track_commit_timestamp,omitempty"` - // +kubebuilder:validation:Enum="all";"pl";"none" + // +kubebuilder:validation:Enum="all";"none";"pl" // Enables tracking of function call counts and time used. TrackFunctions *string `groups:"create,update" json:"track_functions,omitempty"` @@ -388,7 +388,7 @@ type Pgbouncer struct { // Do not allow more than this many server connections per database (regardless of user). Setting it to 0 means unlimited. AutodbMaxDbConnections *int `groups:"create,update" json:"autodb_max_db_connections,omitempty"` - // +kubebuilder:validation:Enum="session";"transaction";"statement" + // +kubebuilder:validation:Enum="session";"statement";"transaction" // PGBouncer pool mode AutodbPoolMode *string `groups:"create,update" json:"autodb_pool_mode,omitempty"` @@ -584,7 +584,7 @@ type PgUserConfig struct { // Use static public IP addresses StaticIps *bool `groups:"create,update" json:"static_ips,omitempty"` - // +kubebuilder:validation:Enum="quorum";"off" + // +kubebuilder:validation:Enum="off";"quorum" // Synchronous replication type. Note that the service plan also needs to support synchronous replication. SynchronousReplication *string `groups:"create,update" json:"synchronous_replication,omitempty"` diff --git a/api/v1alpha1/userconfig/service/redis/redis.go b/api/v1alpha1/userconfig/service/redis/redis.go index 325d8d5a..d045a001 100644 --- a/api/v1alpha1/userconfig/service/redis/redis.go +++ b/api/v1alpha1/userconfig/service/redis/redis.go @@ -140,7 +140,7 @@ type RedisUserConfig struct { // Counter logarithm factor for volatile-lfu and allkeys-lfu maxmemory-policies RedisLfuLogFactor *int `groups:"create,update" json:"redis_lfu_log_factor,omitempty"` - // +kubebuilder:validation:Enum="noeviction";"allkeys-lru";"volatile-lru";"allkeys-random";"volatile-random";"volatile-ttl";"volatile-lfu";"allkeys-lfu" + // +kubebuilder:validation:Enum="allkeys-lfu";"allkeys-lru";"allkeys-random";"noeviction";"volatile-lfu";"volatile-lru";"volatile-random";"volatile-ttl" // Redis maxmemory-policy RedisMaxmemoryPolicy *string `groups:"create,update" json:"redis_maxmemory_policy,omitempty"` diff --git a/charts/aiven-operator-crds/templates/aiven.io_cassandras.yaml b/charts/aiven-operator-crds/templates/aiven.io_cassandras.yaml index 156e2acb..4f7a9d3e 100644 --- a/charts/aiven-operator-crds/templates/aiven.io_cassandras.yaml +++ b/charts/aiven-operator-crds/templates/aiven.io_cassandras.yaml @@ -225,7 +225,7 @@ spec: description: Cassandra specific user configuration options properties: additional_backup_regions: - description: Deprecated. Additional Cloud Regions for Backup Replication + description: Additional Cloud Regions for Backup Replication items: type: string maxItems: 1 diff --git a/charts/aiven-operator-crds/templates/aiven.io_grafanas.yaml b/charts/aiven-operator-crds/templates/aiven.io_grafanas.yaml index 3a3140c7..f47b2854 100644 --- a/charts/aiven-operator-crds/templates/aiven.io_grafanas.yaml +++ b/charts/aiven-operator-crds/templates/aiven.io_grafanas.yaml @@ -256,8 +256,8 @@ spec: alerting rules enum: - alerting - - no_data - keep_state + - no_data - ok type: string allow_embedding: @@ -504,8 +504,8 @@ spec: value." enum: - lax - - strict - none + - strict type: string custom_domain: description: @@ -788,9 +788,9 @@ spec: Either OpportunisticStartTLS, MandatoryStartTLS or NoStartTLS. Default is OpportunisticStartTLS. enum: - - OpportunisticStartTLS - MandatoryStartTLS - NoStartTLS + - OpportunisticStartTLS type: string username: description: Username for SMTP authentication @@ -821,9 +821,9 @@ spec: user_auto_assign_org_role: description: Set role for new signups. Defaults to Viewer enum: - - Viewer - Admin - Editor + - Viewer type: string viewers_can_edit: description: diff --git a/charts/aiven-operator-crds/templates/aiven.io_kafkaconnects.yaml b/charts/aiven-operator-crds/templates/aiven.io_kafkaconnects.yaml index 0414c140..b784c756 100644 --- a/charts/aiven-operator-crds/templates/aiven.io_kafkaconnects.yaml +++ b/charts/aiven-operator-crds/templates/aiven.io_kafkaconnects.yaml @@ -210,8 +210,8 @@ spec: Defines what client configurations can be overridden by the connector. Default is None enum: - - None - All + - None type: string consumer_auto_offset_reset: description: @@ -238,8 +238,8 @@ spec: is the default, but read_committed can be used if consume-exactly-once behavior is desired. enum: - - read_uncommitted - read_committed + - read_uncommitted type: string consumer_max_partition_fetch_bytes: description: @@ -309,10 +309,10 @@ spec: 'none' which is the default and equivalent to no compression. enum: - gzip - - snappy - lz4 - - zstd - none + - snappy + - zstd type: string producer_linger_ms: description: diff --git a/charts/aiven-operator-crds/templates/aiven.io_kafkas.yaml b/charts/aiven-operator-crds/templates/aiven.io_kafkas.yaml index 5d40d3c8..eca5e9d7 100644 --- a/charts/aiven-operator-crds/templates/aiven.io_kafkas.yaml +++ b/charts/aiven-operator-crds/templates/aiven.io_kafkas.yaml @@ -291,11 +291,11 @@ spec: codec set by the producer.(Default: producer)" enum: - gzip - - snappy - lz4 - - zstd - - uncompressed - producer + - snappy + - uncompressed + - zstd type: string connections_max_idle_ms: description: @@ -379,9 +379,9 @@ spec: "The default cleanup policy for segments beyond the retention window (Default: delete)" enum: - - delete - compact - compact,delete + - delete type: string log_flush_interval_messages: description: @@ -661,8 +661,8 @@ spec: Defines what client configurations can be overridden by the connector. Default is None enum: - - None - All + - None type: string consumer_auto_offset_reset: description: @@ -689,8 +689,8 @@ spec: is the default, but read_committed can be used if consume-exactly-once behavior is desired. enum: - - read_uncommitted - read_committed + - read_uncommitted type: string consumer_max_partition_fetch_bytes: description: @@ -760,10 +760,10 @@ spec: 'none' which is the default and equivalent to no compression. enum: - gzip - - snappy - lz4 - - zstd - none + - snappy + - zstd type: string producer_linger_ms: description: @@ -926,8 +926,8 @@ spec: Name strategy to use when selecting subject for storing schemas enum: - - topic_name - record_name + - topic_name - topic_record_name type: string name_strategy_validation: @@ -944,10 +944,10 @@ spec: for the full set of in-sync replicas to acknowledge the record. enum: - - all - "-1" - "0" - "1" + - all type: string producer_compression_type: description: @@ -957,10 +957,10 @@ spec: 'none' which is the default and equivalent to no compression. enum: - gzip - - snappy - lz4 - - zstd - none + - snappy + - zstd type: string producer_linger_ms: description: diff --git a/charts/aiven-operator-crds/templates/aiven.io_mysqls.yaml b/charts/aiven-operator-crds/templates/aiven.io_mysqls.yaml index 9b46a342..c3b65e15 100644 --- a/charts/aiven-operator-crds/templates/aiven.io_mysqls.yaml +++ b/charts/aiven-operator-crds/templates/aiven.io_mysqls.yaml @@ -495,8 +495,8 @@ spec: The storage engine for in-memory internal temporary tables. enum: - - TempTable - MEMORY + - TempTable type: string log_output: description: @@ -507,9 +507,9 @@ spec: table, choose TABLE. To silence slow logs, choose NONE. enum: - INSIGHTS + - INSIGHTS,TABLE - NONE - TABLE - - INSIGHTS,TABLE type: string long_query_time: description: diff --git a/charts/aiven-operator-crds/templates/aiven.io_opensearches.yaml b/charts/aiven-operator-crds/templates/aiven.io_opensearches.yaml index 6801b7a9..5532543e 100644 --- a/charts/aiven-operator-crds/templates/aiven.io_opensearches.yaml +++ b/charts/aiven-operator-crds/templates/aiven.io_opensearches.yaml @@ -925,9 +925,9 @@ spec: are monitor_only, enforced, or disabled. Default is monitor_only enum: - - monitor_only - - enforced - disabled + - enforced + - monitor_only type: string node_duress: description: Node duress settings diff --git a/charts/aiven-operator-crds/templates/aiven.io_postgresqls.yaml b/charts/aiven-operator-crds/templates/aiven.io_postgresqls.yaml index e9e93bca..0756a45c 100644 --- a/charts/aiven-operator-crds/templates/aiven.io_postgresqls.yaml +++ b/charts/aiven-operator-crds/templates/aiven.io_postgresqls.yaml @@ -519,16 +519,16 @@ spec: Controls the amount of detail written in the server log for each message that is logged. enum: - - TERSE - DEFAULT + - TERSE - VERBOSE type: string log_line_prefix: description: Choose from one of the available log formats. enum: - - "'pid=%p,user=%u,db=%d,app=%a,client=%h '" - - "'%t [%p]: [%l-1] user=%u,db=%d,app=%a,client=%h '" - "'%m [%p] %q[user=%u,db=%d,app=%a] '" + - "'%t [%p]: [%l-1] user=%u,db=%d,app=%a,client=%h '" + - "'pid=%p,user=%u,db=%d,app=%a,client=%h '" - "'pid=%p,user=%u,db=%d,app=%a,client=%h,txid=%x,qid=%Q '" type: string @@ -660,8 +660,8 @@ spec: statement statistics collection. The default value is top. enum: - all - - top - none + - top type: string temp_file_limit: description: @@ -694,8 +694,8 @@ spec: time used. enum: - all - - pl - none + - pl type: string track_io_timing: description: @@ -898,8 +898,8 @@ spec: description: PGBouncer pool mode enum: - session - - transaction - statement + - transaction type: string autodb_pool_size: description: @@ -1081,8 +1081,8 @@ spec: Synchronous replication type. Note that the service plan also needs to support synchronous replication. enum: - - quorum - "off" + - quorum type: string timescaledb: description: System-wide settings for the timescaledb extension diff --git a/charts/aiven-operator-crds/templates/aiven.io_redis.yaml b/charts/aiven-operator-crds/templates/aiven.io_redis.yaml index 34f54b71..0aaa5511 100644 --- a/charts/aiven-operator-crds/templates/aiven.io_redis.yaml +++ b/charts/aiven-operator-crds/templates/aiven.io_redis.yaml @@ -426,14 +426,14 @@ spec: redis_maxmemory_policy: description: Redis maxmemory-policy enum: - - noeviction + - allkeys-lfu - allkeys-lru - - volatile-lru - allkeys-random + - noeviction + - volatile-lfu + - volatile-lru - volatile-random - volatile-ttl - - volatile-lfu - - allkeys-lfu type: string redis_notify_keyspace_events: description: Set notify-keyspace-events option diff --git a/charts/aiven-operator-crds/templates/aiven.io_serviceintegrationendpoints.yaml b/charts/aiven-operator-crds/templates/aiven.io_serviceintegrationendpoints.yaml index 2893db2a..0ae9ab6c 100644 --- a/charts/aiven-operator-crds/templates/aiven.io_serviceintegrationendpoints.yaml +++ b/charts/aiven-operator-crds/templates/aiven.io_serviceintegrationendpoints.yaml @@ -124,12 +124,12 @@ spec: site: description: Datadog intake site. Defaults to datadoghq.com enum: + - ap1.datadoghq.com - datadoghq.com - datadoghq.eu + - ddog-gov.com - us3.datadoghq.com - us5.datadoghq.com - - ddog-gov.com - - ap1.datadoghq.com type: string required: - datadog_api_key @@ -327,9 +327,9 @@ spec: description: Security protocol enum: - PLAINTEXT - - SSL - SASL_PLAINTEXT - SASL_SSL + - SSL type: string ssl_ca_cert: description: PEM-encoded CA certificate @@ -444,8 +444,8 @@ spec: authentication: description: Authentication method enum: - - none - basic + - none type: string basic_auth_password: description: Basic authentication password @@ -515,9 +515,9 @@ spec: format: description: Message format enum: - - rfc5424 - - rfc3164 - custom + - rfc3164 + - rfc5424 type: string key: description: PEM encoded client key diff --git a/charts/aiven-operator-crds/templates/aiven.io_serviceintegrations.yaml b/charts/aiven-operator-crds/templates/aiven.io_serviceintegrations.yaml index 825f7270..2b87e206 100644 --- a/charts/aiven-operator-crds/templates/aiven.io_serviceintegrations.yaml +++ b/charts/aiven-operator-crds/templates/aiven.io_serviceintegrations.yaml @@ -86,12 +86,12 @@ spec: Action to take when there is no initial offset in offset store or the desired offset is out of range enum: - - smallest - - earliest - beginning + - earliest + - end - largest - latest - - end + - smallest type: string columns: description: Table columns @@ -118,6 +118,7 @@ spec: description: Message data format enum: - Avro + - AvroConfluent - CSV - JSONAsString - JSONCompactEachRow @@ -125,12 +126,11 @@ spec: - JSONEachRow - JSONStringsEachRow - MsgPack + - Parquet + - RawBLOB - TSKV - TSV - TabSeparated - - RawBLOB - - AvroConfluent - - Parquet type: string date_time_input_format: description: Method to read DateTime from text input formats @@ -572,10 +572,10 @@ spec: 'none' which is the default and equivalent to no compression. enum: - gzip - - snappy - lz4 - - zstd - none + - snappy + - zstd type: string producer_linger_ms: description: diff --git a/config/crd/bases/aiven.io_cassandras.yaml b/config/crd/bases/aiven.io_cassandras.yaml index 156e2acb..4f7a9d3e 100644 --- a/config/crd/bases/aiven.io_cassandras.yaml +++ b/config/crd/bases/aiven.io_cassandras.yaml @@ -225,7 +225,7 @@ spec: description: Cassandra specific user configuration options properties: additional_backup_regions: - description: Deprecated. Additional Cloud Regions for Backup Replication + description: Additional Cloud Regions for Backup Replication items: type: string maxItems: 1 diff --git a/config/crd/bases/aiven.io_grafanas.yaml b/config/crd/bases/aiven.io_grafanas.yaml index 3a3140c7..f47b2854 100644 --- a/config/crd/bases/aiven.io_grafanas.yaml +++ b/config/crd/bases/aiven.io_grafanas.yaml @@ -256,8 +256,8 @@ spec: alerting rules enum: - alerting - - no_data - keep_state + - no_data - ok type: string allow_embedding: @@ -504,8 +504,8 @@ spec: value." enum: - lax - - strict - none + - strict type: string custom_domain: description: @@ -788,9 +788,9 @@ spec: Either OpportunisticStartTLS, MandatoryStartTLS or NoStartTLS. Default is OpportunisticStartTLS. enum: - - OpportunisticStartTLS - MandatoryStartTLS - NoStartTLS + - OpportunisticStartTLS type: string username: description: Username for SMTP authentication @@ -821,9 +821,9 @@ spec: user_auto_assign_org_role: description: Set role for new signups. Defaults to Viewer enum: - - Viewer - Admin - Editor + - Viewer type: string viewers_can_edit: description: diff --git a/config/crd/bases/aiven.io_kafkaconnects.yaml b/config/crd/bases/aiven.io_kafkaconnects.yaml index 0414c140..b784c756 100644 --- a/config/crd/bases/aiven.io_kafkaconnects.yaml +++ b/config/crd/bases/aiven.io_kafkaconnects.yaml @@ -210,8 +210,8 @@ spec: Defines what client configurations can be overridden by the connector. Default is None enum: - - None - All + - None type: string consumer_auto_offset_reset: description: @@ -238,8 +238,8 @@ spec: is the default, but read_committed can be used if consume-exactly-once behavior is desired. enum: - - read_uncommitted - read_committed + - read_uncommitted type: string consumer_max_partition_fetch_bytes: description: @@ -309,10 +309,10 @@ spec: 'none' which is the default and equivalent to no compression. enum: - gzip - - snappy - lz4 - - zstd - none + - snappy + - zstd type: string producer_linger_ms: description: diff --git a/config/crd/bases/aiven.io_kafkas.yaml b/config/crd/bases/aiven.io_kafkas.yaml index 5d40d3c8..eca5e9d7 100644 --- a/config/crd/bases/aiven.io_kafkas.yaml +++ b/config/crd/bases/aiven.io_kafkas.yaml @@ -291,11 +291,11 @@ spec: codec set by the producer.(Default: producer)" enum: - gzip - - snappy - lz4 - - zstd - - uncompressed - producer + - snappy + - uncompressed + - zstd type: string connections_max_idle_ms: description: @@ -379,9 +379,9 @@ spec: "The default cleanup policy for segments beyond the retention window (Default: delete)" enum: - - delete - compact - compact,delete + - delete type: string log_flush_interval_messages: description: @@ -661,8 +661,8 @@ spec: Defines what client configurations can be overridden by the connector. Default is None enum: - - None - All + - None type: string consumer_auto_offset_reset: description: @@ -689,8 +689,8 @@ spec: is the default, but read_committed can be used if consume-exactly-once behavior is desired. enum: - - read_uncommitted - read_committed + - read_uncommitted type: string consumer_max_partition_fetch_bytes: description: @@ -760,10 +760,10 @@ spec: 'none' which is the default and equivalent to no compression. enum: - gzip - - snappy - lz4 - - zstd - none + - snappy + - zstd type: string producer_linger_ms: description: @@ -926,8 +926,8 @@ spec: Name strategy to use when selecting subject for storing schemas enum: - - topic_name - record_name + - topic_name - topic_record_name type: string name_strategy_validation: @@ -944,10 +944,10 @@ spec: for the full set of in-sync replicas to acknowledge the record. enum: - - all - "-1" - "0" - "1" + - all type: string producer_compression_type: description: @@ -957,10 +957,10 @@ spec: 'none' which is the default and equivalent to no compression. enum: - gzip - - snappy - lz4 - - zstd - none + - snappy + - zstd type: string producer_linger_ms: description: diff --git a/config/crd/bases/aiven.io_mysqls.yaml b/config/crd/bases/aiven.io_mysqls.yaml index 9b46a342..c3b65e15 100644 --- a/config/crd/bases/aiven.io_mysqls.yaml +++ b/config/crd/bases/aiven.io_mysqls.yaml @@ -495,8 +495,8 @@ spec: The storage engine for in-memory internal temporary tables. enum: - - TempTable - MEMORY + - TempTable type: string log_output: description: @@ -507,9 +507,9 @@ spec: table, choose TABLE. To silence slow logs, choose NONE. enum: - INSIGHTS + - INSIGHTS,TABLE - NONE - TABLE - - INSIGHTS,TABLE type: string long_query_time: description: diff --git a/config/crd/bases/aiven.io_opensearches.yaml b/config/crd/bases/aiven.io_opensearches.yaml index 6801b7a9..5532543e 100644 --- a/config/crd/bases/aiven.io_opensearches.yaml +++ b/config/crd/bases/aiven.io_opensearches.yaml @@ -925,9 +925,9 @@ spec: are monitor_only, enforced, or disabled. Default is monitor_only enum: - - monitor_only - - enforced - disabled + - enforced + - monitor_only type: string node_duress: description: Node duress settings diff --git a/config/crd/bases/aiven.io_postgresqls.yaml b/config/crd/bases/aiven.io_postgresqls.yaml index e9e93bca..0756a45c 100644 --- a/config/crd/bases/aiven.io_postgresqls.yaml +++ b/config/crd/bases/aiven.io_postgresqls.yaml @@ -519,16 +519,16 @@ spec: Controls the amount of detail written in the server log for each message that is logged. enum: - - TERSE - DEFAULT + - TERSE - VERBOSE type: string log_line_prefix: description: Choose from one of the available log formats. enum: - - "'pid=%p,user=%u,db=%d,app=%a,client=%h '" - - "'%t [%p]: [%l-1] user=%u,db=%d,app=%a,client=%h '" - "'%m [%p] %q[user=%u,db=%d,app=%a] '" + - "'%t [%p]: [%l-1] user=%u,db=%d,app=%a,client=%h '" + - "'pid=%p,user=%u,db=%d,app=%a,client=%h '" - "'pid=%p,user=%u,db=%d,app=%a,client=%h,txid=%x,qid=%Q '" type: string @@ -660,8 +660,8 @@ spec: statement statistics collection. The default value is top. enum: - all - - top - none + - top type: string temp_file_limit: description: @@ -694,8 +694,8 @@ spec: time used. enum: - all - - pl - none + - pl type: string track_io_timing: description: @@ -898,8 +898,8 @@ spec: description: PGBouncer pool mode enum: - session - - transaction - statement + - transaction type: string autodb_pool_size: description: @@ -1081,8 +1081,8 @@ spec: Synchronous replication type. Note that the service plan also needs to support synchronous replication. enum: - - quorum - "off" + - quorum type: string timescaledb: description: System-wide settings for the timescaledb extension diff --git a/config/crd/bases/aiven.io_redis.yaml b/config/crd/bases/aiven.io_redis.yaml index 34f54b71..0aaa5511 100644 --- a/config/crd/bases/aiven.io_redis.yaml +++ b/config/crd/bases/aiven.io_redis.yaml @@ -426,14 +426,14 @@ spec: redis_maxmemory_policy: description: Redis maxmemory-policy enum: - - noeviction + - allkeys-lfu - allkeys-lru - - volatile-lru - allkeys-random + - noeviction + - volatile-lfu + - volatile-lru - volatile-random - volatile-ttl - - volatile-lfu - - allkeys-lfu type: string redis_notify_keyspace_events: description: Set notify-keyspace-events option diff --git a/config/crd/bases/aiven.io_serviceintegrationendpoints.yaml b/config/crd/bases/aiven.io_serviceintegrationendpoints.yaml index 2893db2a..0ae9ab6c 100644 --- a/config/crd/bases/aiven.io_serviceintegrationendpoints.yaml +++ b/config/crd/bases/aiven.io_serviceintegrationendpoints.yaml @@ -124,12 +124,12 @@ spec: site: description: Datadog intake site. Defaults to datadoghq.com enum: + - ap1.datadoghq.com - datadoghq.com - datadoghq.eu + - ddog-gov.com - us3.datadoghq.com - us5.datadoghq.com - - ddog-gov.com - - ap1.datadoghq.com type: string required: - datadog_api_key @@ -327,9 +327,9 @@ spec: description: Security protocol enum: - PLAINTEXT - - SSL - SASL_PLAINTEXT - SASL_SSL + - SSL type: string ssl_ca_cert: description: PEM-encoded CA certificate @@ -444,8 +444,8 @@ spec: authentication: description: Authentication method enum: - - none - basic + - none type: string basic_auth_password: description: Basic authentication password @@ -515,9 +515,9 @@ spec: format: description: Message format enum: - - rfc5424 - - rfc3164 - custom + - rfc3164 + - rfc5424 type: string key: description: PEM encoded client key diff --git a/config/crd/bases/aiven.io_serviceintegrations.yaml b/config/crd/bases/aiven.io_serviceintegrations.yaml index 825f7270..2b87e206 100644 --- a/config/crd/bases/aiven.io_serviceintegrations.yaml +++ b/config/crd/bases/aiven.io_serviceintegrations.yaml @@ -86,12 +86,12 @@ spec: Action to take when there is no initial offset in offset store or the desired offset is out of range enum: - - smallest - - earliest - beginning + - earliest + - end - largest - latest - - end + - smallest type: string columns: description: Table columns @@ -118,6 +118,7 @@ spec: description: Message data format enum: - Avro + - AvroConfluent - CSV - JSONAsString - JSONCompactEachRow @@ -125,12 +126,11 @@ spec: - JSONEachRow - JSONStringsEachRow - MsgPack + - Parquet + - RawBLOB - TSKV - TSV - TabSeparated - - RawBLOB - - AvroConfluent - - Parquet type: string date_time_input_format: description: Method to read DateTime from text input formats @@ -572,10 +572,10 @@ spec: 'none' which is the default and equivalent to no compression. enum: - gzip - - snappy - lz4 - - zstd - none + - snappy + - zstd type: string producer_linger_ms: description: diff --git a/docs/docs/api-reference/cassandra.md b/docs/docs/api-reference/cassandra.md index e5380818..4916532f 100644 --- a/docs/docs/api-reference/cassandra.md +++ b/docs/docs/api-reference/cassandra.md @@ -203,7 +203,7 @@ Cassandra specific user configuration options. **Optional** -- [`additional_backup_regions`](#spec.userConfig.additional_backup_regions-property){: name='spec.userConfig.additional_backup_regions-property'} (array of strings, MaxItems: 1). Deprecated. Additional Cloud Regions for Backup Replication. +- [`additional_backup_regions`](#spec.userConfig.additional_backup_regions-property){: name='spec.userConfig.additional_backup_regions-property'} (array of strings, MaxItems: 1). Additional Cloud Regions for Backup Replication. - [`backup_hour`](#spec.userConfig.backup_hour-property){: name='spec.userConfig.backup_hour-property'} (integer, Minimum: 0, Maximum: 23). The hour of day (in UTC) when backup for the service is started. New backup is only started if previous backup has already completed. - [`backup_minute`](#spec.userConfig.backup_minute-property){: name='spec.userConfig.backup_minute-property'} (integer, Minimum: 0, Maximum: 59). The minute of an hour when backup for the service is started. New backup is only started if previous backup has already completed. - [`cassandra`](#spec.userConfig.cassandra-property){: name='spec.userConfig.cassandra-property'} (object). cassandra configuration values. See below for [nested schema](#spec.userConfig.cassandra). diff --git a/docs/docs/api-reference/grafana.md b/docs/docs/api-reference/grafana.md index f760d063..df4560fb 100644 --- a/docs/docs/api-reference/grafana.md +++ b/docs/docs/api-reference/grafana.md @@ -205,7 +205,7 @@ Cassandra specific user configuration options. - [`alerting_enabled`](#spec.userConfig.alerting_enabled-property){: name='spec.userConfig.alerting_enabled-property'} (boolean). Enable or disable Grafana legacy alerting functionality. This should not be enabled with unified_alerting_enabled. - [`alerting_error_or_timeout`](#spec.userConfig.alerting_error_or_timeout-property){: name='spec.userConfig.alerting_error_or_timeout-property'} (string, Enum: `alerting`, `keep_state`). Default error or timeout setting for new alerting rules. - [`alerting_max_annotations_to_keep`](#spec.userConfig.alerting_max_annotations_to_keep-property){: name='spec.userConfig.alerting_max_annotations_to_keep-property'} (integer, Minimum: 0, Maximum: 1000000). Max number of alert annotations that Grafana stores. 0 (default) keeps all alert annotations. -- [`alerting_nodata_or_nullvalues`](#spec.userConfig.alerting_nodata_or_nullvalues-property){: name='spec.userConfig.alerting_nodata_or_nullvalues-property'} (string, Enum: `alerting`, `no_data`, `keep_state`, `ok`). Default value for 'no data or null values' for new alerting rules. +- [`alerting_nodata_or_nullvalues`](#spec.userConfig.alerting_nodata_or_nullvalues-property){: name='spec.userConfig.alerting_nodata_or_nullvalues-property'} (string, Enum: `alerting`, `keep_state`, `no_data`, `ok`). Default value for 'no data or null values' for new alerting rules. - [`allow_embedding`](#spec.userConfig.allow_embedding-property){: name='spec.userConfig.allow_embedding-property'} (boolean). Allow embedding Grafana dashboards with iframe/frame/object/embed tags. Disabled by default to limit impact of clickjacking. - [`auth_azuread`](#spec.userConfig.auth_azuread-property){: name='spec.userConfig.auth_azuread-property'} (object). Azure AD OAuth integration. See below for [nested schema](#spec.userConfig.auth_azuread). - [`auth_basic_enabled`](#spec.userConfig.auth_basic_enabled-property){: name='spec.userConfig.auth_basic_enabled-property'} (boolean). Enable or disable basic authentication form, used by Grafana built-in login. @@ -213,7 +213,7 @@ Cassandra specific user configuration options. - [`auth_github`](#spec.userConfig.auth_github-property){: name='spec.userConfig.auth_github-property'} (object). Github Auth integration. See below for [nested schema](#spec.userConfig.auth_github). - [`auth_gitlab`](#spec.userConfig.auth_gitlab-property){: name='spec.userConfig.auth_gitlab-property'} (object). GitLab Auth integration. See below for [nested schema](#spec.userConfig.auth_gitlab). - [`auth_google`](#spec.userConfig.auth_google-property){: name='spec.userConfig.auth_google-property'} (object). Google Auth integration. See below for [nested schema](#spec.userConfig.auth_google). -- [`cookie_samesite`](#spec.userConfig.cookie_samesite-property){: name='spec.userConfig.cookie_samesite-property'} (string, Enum: `lax`, `strict`, `none`). Cookie SameSite attribute: `strict` prevents sending cookie for cross-site requests, effectively disabling direct linking from other sites to Grafana. `lax` is the default value. +- [`cookie_samesite`](#spec.userConfig.cookie_samesite-property){: name='spec.userConfig.cookie_samesite-property'} (string, Enum: `lax`, `none`, `strict`). Cookie SameSite attribute: `strict` prevents sending cookie for cross-site requests, effectively disabling direct linking from other sites to Grafana. `lax` is the default value. - [`custom_domain`](#spec.userConfig.custom_domain-property){: name='spec.userConfig.custom_domain-property'} (string, MaxLength: 255). Serve the web frontend using a custom CNAME pointing to the Aiven DNS name. - [`dashboard_previews_enabled`](#spec.userConfig.dashboard_previews_enabled-property){: name='spec.userConfig.dashboard_previews_enabled-property'} (boolean). This feature is new in Grafana 9 and is quite resource intensive. It may cause low-end plans to work more slowly while the dashboard previews are rendering. - [`dashboards_min_refresh_interval`](#spec.userConfig.dashboards_min_refresh_interval-property){: name='spec.userConfig.dashboards_min_refresh_interval-property'} (string, Pattern: `^[0-9]+(ms|s|m|h|d)$`, MaxLength: 16). Signed sequence of decimal numbers, followed by a unit suffix (ms, s, m, h, d), e.g. 30s, 1h. @@ -239,7 +239,7 @@ Cassandra specific user configuration options. - [`static_ips`](#spec.userConfig.static_ips-property){: name='spec.userConfig.static_ips-property'} (boolean). Use static public IP addresses. - [`unified_alerting_enabled`](#spec.userConfig.unified_alerting_enabled-property){: name='spec.userConfig.unified_alerting_enabled-property'} (boolean). Enable or disable Grafana unified alerting functionality. By default this is enabled and any legacy alerts will be migrated on upgrade to Grafana 9+. To stay on legacy alerting, set unified_alerting_enabled to false and alerting_enabled to true. See https://grafana.com/docs/grafana/latest/alerting/set-up/migrating-alerts/ for more details. - [`user_auto_assign_org`](#spec.userConfig.user_auto_assign_org-property){: name='spec.userConfig.user_auto_assign_org-property'} (boolean). Auto-assign new users on signup to main organization. Defaults to false. -- [`user_auto_assign_org_role`](#spec.userConfig.user_auto_assign_org_role-property){: name='spec.userConfig.user_auto_assign_org_role-property'} (string, Enum: `Viewer`, `Admin`, `Editor`). Set role for new signups. Defaults to Viewer. +- [`user_auto_assign_org_role`](#spec.userConfig.user_auto_assign_org_role-property){: name='spec.userConfig.user_auto_assign_org_role-property'} (string, Enum: `Admin`, `Editor`, `Viewer`). Set role for new signups. Defaults to Viewer. - [`viewers_can_edit`](#spec.userConfig.viewers_can_edit-property){: name='spec.userConfig.viewers_can_edit-property'} (boolean). Users with view-only permission can edit but not save dashboards. - [`wal`](#spec.userConfig.wal-property){: name='spec.userConfig.wal-property'} (boolean). Setting to enable/disable Write-Ahead Logging. The default value is false (disabled). @@ -431,5 +431,5 @@ SMTP server settings. - [`from_name`](#spec.userConfig.smtp_server.from_name-property){: name='spec.userConfig.smtp_server.from_name-property'} (string, Pattern: `^[^\x00-\x1F]+$`, MaxLength: 128). Name used in outgoing emails, defaults to Grafana. - [`password`](#spec.userConfig.smtp_server.password-property){: name='spec.userConfig.smtp_server.password-property'} (string, Pattern: `^[^\x00-\x1F]+$`, MaxLength: 255). Password for SMTP authentication. - [`skip_verify`](#spec.userConfig.smtp_server.skip_verify-property){: name='spec.userConfig.smtp_server.skip_verify-property'} (boolean). Skip verifying server certificate. Defaults to false. -- [`starttls_policy`](#spec.userConfig.smtp_server.starttls_policy-property){: name='spec.userConfig.smtp_server.starttls_policy-property'} (string, Enum: `OpportunisticStartTLS`, `MandatoryStartTLS`, `NoStartTLS`). Either OpportunisticStartTLS, MandatoryStartTLS or NoStartTLS. Default is OpportunisticStartTLS. +- [`starttls_policy`](#spec.userConfig.smtp_server.starttls_policy-property){: name='spec.userConfig.smtp_server.starttls_policy-property'} (string, Enum: `MandatoryStartTLS`, `NoStartTLS`, `OpportunisticStartTLS`). Either OpportunisticStartTLS, MandatoryStartTLS or NoStartTLS. Default is OpportunisticStartTLS. - [`username`](#spec.userConfig.smtp_server.username-property){: name='spec.userConfig.smtp_server.username-property'} (string, Pattern: `^[^\x00-\x1F]+$`, MaxLength: 255). Username for SMTP authentication. diff --git a/docs/docs/api-reference/kafka.md b/docs/docs/api-reference/kafka.md index 4f00d46b..d2fed93b 100644 --- a/docs/docs/api-reference/kafka.md +++ b/docs/docs/api-reference/kafka.md @@ -262,7 +262,7 @@ Kafka broker configuration values. **Optional** - [`auto_create_topics_enable`](#spec.userConfig.kafka.auto_create_topics_enable-property){: name='spec.userConfig.kafka.auto_create_topics_enable-property'} (boolean). Enable auto-creation of topics. (Default: true). -- [`compression_type`](#spec.userConfig.kafka.compression_type-property){: name='spec.userConfig.kafka.compression_type-property'} (string, Enum: `gzip`, `snappy`, `lz4`, `zstd`, `uncompressed`, `producer`). Specify the final compression type for a given topic. This configuration accepts the standard compression codecs (`gzip`, `snappy`, `lz4`, `zstd`). It additionally accepts `uncompressed` which is equivalent to no compression; and `producer` which means retain the original compression codec set by the producer.(Default: producer). +- [`compression_type`](#spec.userConfig.kafka.compression_type-property){: name='spec.userConfig.kafka.compression_type-property'} (string, Enum: `gzip`, `lz4`, `producer`, `snappy`, `uncompressed`, `zstd`). Specify the final compression type for a given topic. This configuration accepts the standard compression codecs (`gzip`, `snappy`, `lz4`, `zstd`). It additionally accepts `uncompressed` which is equivalent to no compression; and `producer` which means retain the original compression codec set by the producer.(Default: producer). - [`connections_max_idle_ms`](#spec.userConfig.kafka.connections_max_idle_ms-property){: name='spec.userConfig.kafka.connections_max_idle_ms-property'} (integer, Minimum: 1000, Maximum: 3600000). Idle connections timeout: the server socket processor threads close the connections that idle for longer than this. (Default: 600000 ms (10 minutes)). - [`default_replication_factor`](#spec.userConfig.kafka.default_replication_factor-property){: name='spec.userConfig.kafka.default_replication_factor-property'} (integer, Minimum: 1, Maximum: 10). Replication factor for auto-created topics (Default: 3). - [`group_initial_rebalance_delay_ms`](#spec.userConfig.kafka.group_initial_rebalance_delay_ms-property){: name='spec.userConfig.kafka.group_initial_rebalance_delay_ms-property'} (integer, Minimum: 0, Maximum: 300000). The amount of time, in milliseconds, the group coordinator will wait for more consumers to join a new group before performing the first rebalance. A longer delay means potentially fewer rebalances, but increases the time until processing begins. The default value for this is 3 seconds. During development and testing it might be desirable to set this to 0 in order to not delay test execution time. (Default: 3000 ms (3 seconds)). @@ -272,7 +272,7 @@ Kafka broker configuration values. - [`log_cleaner_max_compaction_lag_ms`](#spec.userConfig.kafka.log_cleaner_max_compaction_lag_ms-property){: name='spec.userConfig.kafka.log_cleaner_max_compaction_lag_ms-property'} (integer, Minimum: 30000). The maximum amount of time message will remain uncompacted. Only applicable for logs that are being compacted. (Default: 9223372036854775807 ms (Long.MAX_VALUE)). - [`log_cleaner_min_cleanable_ratio`](#spec.userConfig.kafka.log_cleaner_min_cleanable_ratio-property){: name='spec.userConfig.kafka.log_cleaner_min_cleanable_ratio-property'} (number, Minimum: 0.2, Maximum: 0.9). Controls log compactor frequency. Larger value means more frequent compactions but also more space wasted for logs. Consider setting log.cleaner.max.compaction.lag.ms to enforce compactions sooner, instead of setting a very high value for this option. (Default: 0.5). - [`log_cleaner_min_compaction_lag_ms`](#spec.userConfig.kafka.log_cleaner_min_compaction_lag_ms-property){: name='spec.userConfig.kafka.log_cleaner_min_compaction_lag_ms-property'} (integer, Minimum: 0). The minimum time a message will remain uncompacted in the log. Only applicable for logs that are being compacted. (Default: 0 ms). -- [`log_cleanup_policy`](#spec.userConfig.kafka.log_cleanup_policy-property){: name='spec.userConfig.kafka.log_cleanup_policy-property'} (string, Enum: `delete`, `compact`, `compact,delete`). The default cleanup policy for segments beyond the retention window (Default: delete). +- [`log_cleanup_policy`](#spec.userConfig.kafka.log_cleanup_policy-property){: name='spec.userConfig.kafka.log_cleanup_policy-property'} (string, Enum: `compact`, `compact,delete`, `delete`). The default cleanup policy for segments beyond the retention window (Default: delete). - [`log_flush_interval_messages`](#spec.userConfig.kafka.log_flush_interval_messages-property){: name='spec.userConfig.kafka.log_flush_interval_messages-property'} (integer, Minimum: 1). The number of messages accumulated on a log partition before messages are flushed to disk (Default: 9223372036854775807 (Long.MAX_VALUE)). - [`log_flush_interval_ms`](#spec.userConfig.kafka.log_flush_interval_ms-property){: name='spec.userConfig.kafka.log_flush_interval_ms-property'} (integer, Minimum: 0). The maximum time in ms that a message in any topic is kept in memory (page-cache) before flushed to disk. If not set, the value in log.flush.scheduler.interval.ms is used (Default: null). - [`log_index_interval_bytes`](#spec.userConfig.kafka.log_index_interval_bytes-property){: name='spec.userConfig.kafka.log_index_interval_bytes-property'} (integer, Minimum: 0, Maximum: 104857600). The interval with which Kafka adds an entry to the offset index (Default: 4096 bytes (4 kibibytes)). @@ -327,10 +327,10 @@ Kafka Connect configuration values. **Optional** -- [`connector_client_config_override_policy`](#spec.userConfig.kafka_connect_config.connector_client_config_override_policy-property){: name='spec.userConfig.kafka_connect_config.connector_client_config_override_policy-property'} (string, Enum: `None`, `All`). Defines what client configurations can be overridden by the connector. Default is None. +- [`connector_client_config_override_policy`](#spec.userConfig.kafka_connect_config.connector_client_config_override_policy-property){: name='spec.userConfig.kafka_connect_config.connector_client_config_override_policy-property'} (string, Enum: `All`, `None`). Defines what client configurations can be overridden by the connector. Default is None. - [`consumer_auto_offset_reset`](#spec.userConfig.kafka_connect_config.consumer_auto_offset_reset-property){: name='spec.userConfig.kafka_connect_config.consumer_auto_offset_reset-property'} (string, Enum: `earliest`, `latest`). What to do when there is no initial offset in Kafka or if the current offset does not exist any more on the server. Default is earliest. - [`consumer_fetch_max_bytes`](#spec.userConfig.kafka_connect_config.consumer_fetch_max_bytes-property){: name='spec.userConfig.kafka_connect_config.consumer_fetch_max_bytes-property'} (integer, Minimum: 1048576, Maximum: 104857600). Records are fetched in batches by the consumer, and if the first record batch in the first non-empty partition of the fetch is larger than this value, the record batch will still be returned to ensure that the consumer can make progress. As such, this is not a absolute maximum. -- [`consumer_isolation_level`](#spec.userConfig.kafka_connect_config.consumer_isolation_level-property){: name='spec.userConfig.kafka_connect_config.consumer_isolation_level-property'} (string, Enum: `read_uncommitted`, `read_committed`). Transaction read isolation level. read_uncommitted is the default, but read_committed can be used if consume-exactly-once behavior is desired. +- [`consumer_isolation_level`](#spec.userConfig.kafka_connect_config.consumer_isolation_level-property){: name='spec.userConfig.kafka_connect_config.consumer_isolation_level-property'} (string, Enum: `read_committed`, `read_uncommitted`). Transaction read isolation level. read_uncommitted is the default, but read_committed can be used if consume-exactly-once behavior is desired. - [`consumer_max_partition_fetch_bytes`](#spec.userConfig.kafka_connect_config.consumer_max_partition_fetch_bytes-property){: name='spec.userConfig.kafka_connect_config.consumer_max_partition_fetch_bytes-property'} (integer, Minimum: 1048576, Maximum: 104857600). Records are fetched in batches by the consumer.If the first record batch in the first non-empty partition of the fetch is larger than this limit, the batch will still be returned to ensure that the consumer can make progress. - [`consumer_max_poll_interval_ms`](#spec.userConfig.kafka_connect_config.consumer_max_poll_interval_ms-property){: name='spec.userConfig.kafka_connect_config.consumer_max_poll_interval_ms-property'} (integer, Minimum: 1, Maximum: 2147483647). The maximum delay in milliseconds between invocations of poll() when using consumer group management (defaults to 300000). - [`consumer_max_poll_records`](#spec.userConfig.kafka_connect_config.consumer_max_poll_records-property){: name='spec.userConfig.kafka_connect_config.consumer_max_poll_records-property'} (integer, Minimum: 1, Maximum: 10000). The maximum number of records returned in a single call to poll() (defaults to 500). @@ -338,7 +338,7 @@ Kafka Connect configuration values. - [`offset_flush_timeout_ms`](#spec.userConfig.kafka_connect_config.offset_flush_timeout_ms-property){: name='spec.userConfig.kafka_connect_config.offset_flush_timeout_ms-property'} (integer, Minimum: 1, Maximum: 2147483647). Maximum number of milliseconds to wait for records to flush and partition offset data to be committed to offset storage before cancelling the process and restoring the offset data to be committed in a future attempt (defaults to 5000). - [`producer_batch_size`](#spec.userConfig.kafka_connect_config.producer_batch_size-property){: name='spec.userConfig.kafka_connect_config.producer_batch_size-property'} (integer, Minimum: 0, Maximum: 5242880). This setting gives the upper bound of the batch size to be sent. If there are fewer than this many bytes accumulated for this partition, the producer will `linger` for the linger.ms time waiting for more records to show up. A batch size of zero will disable batching entirely (defaults to 16384). - [`producer_buffer_memory`](#spec.userConfig.kafka_connect_config.producer_buffer_memory-property){: name='spec.userConfig.kafka_connect_config.producer_buffer_memory-property'} (integer, Minimum: 5242880, Maximum: 134217728). The total bytes of memory the producer can use to buffer records waiting to be sent to the broker (defaults to 33554432). -- [`producer_compression_type`](#spec.userConfig.kafka_connect_config.producer_compression_type-property){: name='spec.userConfig.kafka_connect_config.producer_compression_type-property'} (string, Enum: `gzip`, `snappy`, `lz4`, `zstd`, `none`). Specify the default compression type for producers. This configuration accepts the standard compression codecs (`gzip`, `snappy`, `lz4`, `zstd`). It additionally accepts `none` which is the default and equivalent to no compression. +- [`producer_compression_type`](#spec.userConfig.kafka_connect_config.producer_compression_type-property){: name='spec.userConfig.kafka_connect_config.producer_compression_type-property'} (string, Enum: `gzip`, `lz4`, `none`, `snappy`, `zstd`). Specify the default compression type for producers. This configuration accepts the standard compression codecs (`gzip`, `snappy`, `lz4`, `zstd`). It additionally accepts `none` which is the default and equivalent to no compression. - [`producer_linger_ms`](#spec.userConfig.kafka_connect_config.producer_linger_ms-property){: name='spec.userConfig.kafka_connect_config.producer_linger_ms-property'} (integer, Minimum: 0, Maximum: 5000). This setting gives the upper bound on the delay for batching: once there is batch.size worth of records for a partition it will be sent immediately regardless of this setting, however if there are fewer than this many bytes accumulated for this partition the producer will `linger` for the specified time waiting for more records to show up. Defaults to 0. - [`producer_max_request_size`](#spec.userConfig.kafka_connect_config.producer_max_request_size-property){: name='spec.userConfig.kafka_connect_config.producer_max_request_size-property'} (integer, Minimum: 131072, Maximum: 67108864). This setting will limit the number of record batches the producer will send in a single request to avoid sending huge requests. - [`scheduled_rebalance_max_delay_ms`](#spec.userConfig.kafka_connect_config.scheduled_rebalance_max_delay_ms-property){: name='spec.userConfig.kafka_connect_config.scheduled_rebalance_max_delay_ms-property'} (integer, Minimum: 0, Maximum: 600000). The maximum delay that is scheduled in order to wait for the return of one or more departed workers before rebalancing and reassigning their connectors and tasks to the group. During this period the connectors and tasks of the departed workers remain unassigned. Defaults to 5 minutes. @@ -403,10 +403,10 @@ Kafka REST configuration. - [`consumer_enable_auto_commit`](#spec.userConfig.kafka_rest_config.consumer_enable_auto_commit-property){: name='spec.userConfig.kafka_rest_config.consumer_enable_auto_commit-property'} (boolean). If true the consumer's offset will be periodically committed to Kafka in the background. - [`consumer_request_max_bytes`](#spec.userConfig.kafka_rest_config.consumer_request_max_bytes-property){: name='spec.userConfig.kafka_rest_config.consumer_request_max_bytes-property'} (integer, Minimum: 0, Maximum: 671088640). Maximum number of bytes in unencoded message keys and values by a single request. - [`consumer_request_timeout_ms`](#spec.userConfig.kafka_rest_config.consumer_request_timeout_ms-property){: name='spec.userConfig.kafka_rest_config.consumer_request_timeout_ms-property'} (integer, Enum: `1000`, `15000`, `30000`, Minimum: 1000, Maximum: 30000). The maximum total time to wait for messages for a request if the maximum number of messages has not yet been reached. -- [`name_strategy`](#spec.userConfig.kafka_rest_config.name_strategy-property){: name='spec.userConfig.kafka_rest_config.name_strategy-property'} (string, Enum: `topic_name`, `record_name`, `topic_record_name`). Name strategy to use when selecting subject for storing schemas. +- [`name_strategy`](#spec.userConfig.kafka_rest_config.name_strategy-property){: name='spec.userConfig.kafka_rest_config.name_strategy-property'} (string, Enum: `record_name`, `topic_name`, `topic_record_name`). Name strategy to use when selecting subject for storing schemas. - [`name_strategy_validation`](#spec.userConfig.kafka_rest_config.name_strategy_validation-property){: name='spec.userConfig.kafka_rest_config.name_strategy_validation-property'} (boolean). If true, validate that given schema is registered under expected subject name by the used name strategy when producing messages. -- [`producer_acks`](#spec.userConfig.kafka_rest_config.producer_acks-property){: name='spec.userConfig.kafka_rest_config.producer_acks-property'} (string, Enum: `all`, `-1`, `0`, `1`). The number of acknowledgments the producer requires the leader to have received before considering a request complete. If set to `all` or `-1`, the leader will wait for the full set of in-sync replicas to acknowledge the record. -- [`producer_compression_type`](#spec.userConfig.kafka_rest_config.producer_compression_type-property){: name='spec.userConfig.kafka_rest_config.producer_compression_type-property'} (string, Enum: `gzip`, `snappy`, `lz4`, `zstd`, `none`). Specify the default compression type for producers. This configuration accepts the standard compression codecs (`gzip`, `snappy`, `lz4`, `zstd`). It additionally accepts `none` which is the default and equivalent to no compression. +- [`producer_acks`](#spec.userConfig.kafka_rest_config.producer_acks-property){: name='spec.userConfig.kafka_rest_config.producer_acks-property'} (string, Enum: `-1`, `0`, `1`, `all`). The number of acknowledgments the producer requires the leader to have received before considering a request complete. If set to `all` or `-1`, the leader will wait for the full set of in-sync replicas to acknowledge the record. +- [`producer_compression_type`](#spec.userConfig.kafka_rest_config.producer_compression_type-property){: name='spec.userConfig.kafka_rest_config.producer_compression_type-property'} (string, Enum: `gzip`, `lz4`, `none`, `snappy`, `zstd`). Specify the default compression type for producers. This configuration accepts the standard compression codecs (`gzip`, `snappy`, `lz4`, `zstd`). It additionally accepts `none` which is the default and equivalent to no compression. - [`producer_linger_ms`](#spec.userConfig.kafka_rest_config.producer_linger_ms-property){: name='spec.userConfig.kafka_rest_config.producer_linger_ms-property'} (integer, Minimum: 0, Maximum: 5000). Wait for up to the given delay to allow batching records together. - [`producer_max_request_size`](#spec.userConfig.kafka_rest_config.producer_max_request_size-property){: name='spec.userConfig.kafka_rest_config.producer_max_request_size-property'} (integer, Minimum: 0, Maximum: 2147483647). The maximum size of a request in bytes. Note that Kafka broker can also cap the record batch size. - [`simpleconsumer_pool_size_max`](#spec.userConfig.kafka_rest_config.simpleconsumer_pool_size_max-property){: name='spec.userConfig.kafka_rest_config.simpleconsumer_pool_size_max-property'} (integer, Minimum: 10, Maximum: 250). Maximum number of SimpleConsumers that can be instantiated per broker. diff --git a/docs/docs/api-reference/kafkaconnect.md b/docs/docs/api-reference/kafkaconnect.md index c50a6b5f..d0d501cc 100644 --- a/docs/docs/api-reference/kafkaconnect.md +++ b/docs/docs/api-reference/kafkaconnect.md @@ -169,10 +169,10 @@ Kafka Connect configuration values. **Optional** -- [`connector_client_config_override_policy`](#spec.userConfig.kafka_connect.connector_client_config_override_policy-property){: name='spec.userConfig.kafka_connect.connector_client_config_override_policy-property'} (string, Enum: `None`, `All`). Defines what client configurations can be overridden by the connector. Default is None. +- [`connector_client_config_override_policy`](#spec.userConfig.kafka_connect.connector_client_config_override_policy-property){: name='spec.userConfig.kafka_connect.connector_client_config_override_policy-property'} (string, Enum: `All`, `None`). Defines what client configurations can be overridden by the connector. Default is None. - [`consumer_auto_offset_reset`](#spec.userConfig.kafka_connect.consumer_auto_offset_reset-property){: name='spec.userConfig.kafka_connect.consumer_auto_offset_reset-property'} (string, Enum: `earliest`, `latest`). What to do when there is no initial offset in Kafka or if the current offset does not exist any more on the server. Default is earliest. - [`consumer_fetch_max_bytes`](#spec.userConfig.kafka_connect.consumer_fetch_max_bytes-property){: name='spec.userConfig.kafka_connect.consumer_fetch_max_bytes-property'} (integer, Minimum: 1048576, Maximum: 104857600). Records are fetched in batches by the consumer, and if the first record batch in the first non-empty partition of the fetch is larger than this value, the record batch will still be returned to ensure that the consumer can make progress. As such, this is not a absolute maximum. -- [`consumer_isolation_level`](#spec.userConfig.kafka_connect.consumer_isolation_level-property){: name='spec.userConfig.kafka_connect.consumer_isolation_level-property'} (string, Enum: `read_uncommitted`, `read_committed`). Transaction read isolation level. read_uncommitted is the default, but read_committed can be used if consume-exactly-once behavior is desired. +- [`consumer_isolation_level`](#spec.userConfig.kafka_connect.consumer_isolation_level-property){: name='spec.userConfig.kafka_connect.consumer_isolation_level-property'} (string, Enum: `read_committed`, `read_uncommitted`). Transaction read isolation level. read_uncommitted is the default, but read_committed can be used if consume-exactly-once behavior is desired. - [`consumer_max_partition_fetch_bytes`](#spec.userConfig.kafka_connect.consumer_max_partition_fetch_bytes-property){: name='spec.userConfig.kafka_connect.consumer_max_partition_fetch_bytes-property'} (integer, Minimum: 1048576, Maximum: 104857600). Records are fetched in batches by the consumer.If the first record batch in the first non-empty partition of the fetch is larger than this limit, the batch will still be returned to ensure that the consumer can make progress. - [`consumer_max_poll_interval_ms`](#spec.userConfig.kafka_connect.consumer_max_poll_interval_ms-property){: name='spec.userConfig.kafka_connect.consumer_max_poll_interval_ms-property'} (integer, Minimum: 1, Maximum: 2147483647). The maximum delay in milliseconds between invocations of poll() when using consumer group management (defaults to 300000). - [`consumer_max_poll_records`](#spec.userConfig.kafka_connect.consumer_max_poll_records-property){: name='spec.userConfig.kafka_connect.consumer_max_poll_records-property'} (integer, Minimum: 1, Maximum: 10000). The maximum number of records returned in a single call to poll() (defaults to 500). @@ -180,7 +180,7 @@ Kafka Connect configuration values. - [`offset_flush_timeout_ms`](#spec.userConfig.kafka_connect.offset_flush_timeout_ms-property){: name='spec.userConfig.kafka_connect.offset_flush_timeout_ms-property'} (integer, Minimum: 1, Maximum: 2147483647). Maximum number of milliseconds to wait for records to flush and partition offset data to be committed to offset storage before cancelling the process and restoring the offset data to be committed in a future attempt (defaults to 5000). - [`producer_batch_size`](#spec.userConfig.kafka_connect.producer_batch_size-property){: name='spec.userConfig.kafka_connect.producer_batch_size-property'} (integer, Minimum: 0, Maximum: 5242880). This setting gives the upper bound of the batch size to be sent. If there are fewer than this many bytes accumulated for this partition, the producer will `linger` for the linger.ms time waiting for more records to show up. A batch size of zero will disable batching entirely (defaults to 16384). - [`producer_buffer_memory`](#spec.userConfig.kafka_connect.producer_buffer_memory-property){: name='spec.userConfig.kafka_connect.producer_buffer_memory-property'} (integer, Minimum: 5242880, Maximum: 134217728). The total bytes of memory the producer can use to buffer records waiting to be sent to the broker (defaults to 33554432). -- [`producer_compression_type`](#spec.userConfig.kafka_connect.producer_compression_type-property){: name='spec.userConfig.kafka_connect.producer_compression_type-property'} (string, Enum: `gzip`, `snappy`, `lz4`, `zstd`, `none`). Specify the default compression type for producers. This configuration accepts the standard compression codecs (`gzip`, `snappy`, `lz4`, `zstd`). It additionally accepts `none` which is the default and equivalent to no compression. +- [`producer_compression_type`](#spec.userConfig.kafka_connect.producer_compression_type-property){: name='spec.userConfig.kafka_connect.producer_compression_type-property'} (string, Enum: `gzip`, `lz4`, `none`, `snappy`, `zstd`). Specify the default compression type for producers. This configuration accepts the standard compression codecs (`gzip`, `snappy`, `lz4`, `zstd`). It additionally accepts `none` which is the default and equivalent to no compression. - [`producer_linger_ms`](#spec.userConfig.kafka_connect.producer_linger_ms-property){: name='spec.userConfig.kafka_connect.producer_linger_ms-property'} (integer, Minimum: 0, Maximum: 5000). This setting gives the upper bound on the delay for batching: once there is batch.size worth of records for a partition it will be sent immediately regardless of this setting, however if there are fewer than this many bytes accumulated for this partition the producer will `linger` for the specified time waiting for more records to show up. Defaults to 0. - [`producer_max_request_size`](#spec.userConfig.kafka_connect.producer_max_request_size-property){: name='spec.userConfig.kafka_connect.producer_max_request_size-property'} (integer, Minimum: 131072, Maximum: 67108864). This setting will limit the number of record batches the producer will send in a single request to avoid sending huge requests. - [`scheduled_rebalance_max_delay_ms`](#spec.userConfig.kafka_connect.scheduled_rebalance_max_delay_ms-property){: name='spec.userConfig.kafka_connect.scheduled_rebalance_max_delay_ms-property'} (integer, Minimum: 0, Maximum: 600000). The maximum delay that is scheduled in order to wait for the return of one or more departed workers before rebalancing and reassigning their connectors and tasks to the group. During this period the connectors and tasks of the departed workers remain unassigned. Defaults to 5 minutes. diff --git a/docs/docs/api-reference/mysql.md b/docs/docs/api-reference/mysql.md index 89d912e5..123af6eb 100644 --- a/docs/docs/api-reference/mysql.md +++ b/docs/docs/api-reference/mysql.md @@ -283,8 +283,8 @@ mysql.conf configuration values. - [`innodb_thread_concurrency`](#spec.userConfig.mysql.innodb_thread_concurrency-property){: name='spec.userConfig.mysql.innodb_thread_concurrency-property'} (integer, Minimum: 0, Maximum: 1000). Defines the maximum number of threads permitted inside of InnoDB. Default is 0 (infinite concurrency - no limit). - [`innodb_write_io_threads`](#spec.userConfig.mysql.innodb_write_io_threads-property){: name='spec.userConfig.mysql.innodb_write_io_threads-property'} (integer, Minimum: 1, Maximum: 64). The number of I/O threads for write operations in InnoDB. Default is 4. Changing this parameter will lead to a restart of the MySQL service. - [`interactive_timeout`](#spec.userConfig.mysql.interactive_timeout-property){: name='spec.userConfig.mysql.interactive_timeout-property'} (integer, Minimum: 30, Maximum: 604800). The number of seconds the server waits for activity on an interactive connection before closing it. -- [`internal_tmp_mem_storage_engine`](#spec.userConfig.mysql.internal_tmp_mem_storage_engine-property){: name='spec.userConfig.mysql.internal_tmp_mem_storage_engine-property'} (string, Enum: `TempTable`, `MEMORY`). The storage engine for in-memory internal temporary tables. -- [`log_output`](#spec.userConfig.mysql.log_output-property){: name='spec.userConfig.mysql.log_output-property'} (string, Enum: `INSIGHTS`, `NONE`, `TABLE`, `INSIGHTS,TABLE`). The slow log output destination when slow_query_log is ON. To enable MySQL AI Insights, choose INSIGHTS. To use MySQL AI Insights and the mysql.slow_log table at the same time, choose INSIGHTS,TABLE. To only use the mysql.slow_log table, choose TABLE. To silence slow logs, choose NONE. +- [`internal_tmp_mem_storage_engine`](#spec.userConfig.mysql.internal_tmp_mem_storage_engine-property){: name='spec.userConfig.mysql.internal_tmp_mem_storage_engine-property'} (string, Enum: `MEMORY`, `TempTable`). The storage engine for in-memory internal temporary tables. +- [`log_output`](#spec.userConfig.mysql.log_output-property){: name='spec.userConfig.mysql.log_output-property'} (string, Enum: `INSIGHTS`, `INSIGHTS,TABLE`, `NONE`, `TABLE`). The slow log output destination when slow_query_log is ON. To enable MySQL AI Insights, choose INSIGHTS. To use MySQL AI Insights and the mysql.slow_log table at the same time, choose INSIGHTS,TABLE. To only use the mysql.slow_log table, choose TABLE. To silence slow logs, choose NONE. - [`long_query_time`](#spec.userConfig.mysql.long_query_time-property){: name='spec.userConfig.mysql.long_query_time-property'} (number, Minimum: 0, Maximum: 3600). The slow_query_logs work as SQL statements that take more than long_query_time seconds to execute. - [`max_allowed_packet`](#spec.userConfig.mysql.max_allowed_packet-property){: name='spec.userConfig.mysql.max_allowed_packet-property'} (integer, Minimum: 102400, Maximum: 1073741824). Size of the largest message in bytes that can be received by the server. Default is 67108864 (64M). - [`max_heap_table_size`](#spec.userConfig.mysql.max_heap_table_size-property){: name='spec.userConfig.mysql.max_heap_table_size-property'} (integer, Minimum: 1048576, Maximum: 1073741824). Limits the size of internal in-memory tables. Also set tmp_table_size. Default is 16777216 (16M). diff --git a/docs/docs/api-reference/opensearch.md b/docs/docs/api-reference/opensearch.md index 117c5754..06d3d159 100644 --- a/docs/docs/api-reference/opensearch.md +++ b/docs/docs/api-reference/opensearch.md @@ -445,7 +445,7 @@ Search Backpressure Settings. **Optional** -- [`mode`](#spec.userConfig.opensearch.search_backpressure.mode-property){: name='spec.userConfig.opensearch.search_backpressure.mode-property'} (string, Enum: `monitor_only`, `enforced`, `disabled`). The search backpressure mode. Valid values are monitor_only, enforced, or disabled. Default is monitor_only. +- [`mode`](#spec.userConfig.opensearch.search_backpressure.mode-property){: name='spec.userConfig.opensearch.search_backpressure.mode-property'} (string, Enum: `disabled`, `enforced`, `monitor_only`). The search backpressure mode. Valid values are monitor_only, enforced, or disabled. Default is monitor_only. - [`node_duress`](#spec.userConfig.opensearch.search_backpressure.node_duress-property){: name='spec.userConfig.opensearch.search_backpressure.node_duress-property'} (object). Node duress settings. See below for [nested schema](#spec.userConfig.opensearch.search_backpressure.node_duress). - [`search_shard_task`](#spec.userConfig.opensearch.search_backpressure.search_shard_task-property){: name='spec.userConfig.opensearch.search_backpressure.search_shard_task-property'} (object). Search shard settings. See below for [nested schema](#spec.userConfig.opensearch.search_backpressure.search_shard_task). - [`search_task`](#spec.userConfig.opensearch.search_backpressure.search_task-property){: name='spec.userConfig.opensearch.search_backpressure.search_task-property'} (object). Search task settings. See below for [nested schema](#spec.userConfig.opensearch.search_backpressure.search_task). diff --git a/docs/docs/api-reference/postgresql.md b/docs/docs/api-reference/postgresql.md index bf9556bf..7d637c19 100644 --- a/docs/docs/api-reference/postgresql.md +++ b/docs/docs/api-reference/postgresql.md @@ -224,7 +224,7 @@ PostgreSQL specific user configuration options. - [`service_to_fork_from`](#spec.userConfig.service_to_fork_from-property){: name='spec.userConfig.service_to_fork_from-property'} (string, Immutable, Pattern: `^[a-z][-a-z0-9]{0,63}$|^$`, MaxLength: 64). Name of another service to fork from. This has effect only when a new service is being created. - [`shared_buffers_percentage`](#spec.userConfig.shared_buffers_percentage-property){: name='spec.userConfig.shared_buffers_percentage-property'} (number, Minimum: 20, Maximum: 60). Percentage of total RAM that the database server uses for shared memory buffers. Valid range is 20-60 (float), which corresponds to 20% - 60%. This setting adjusts the shared_buffers configuration value. - [`static_ips`](#spec.userConfig.static_ips-property){: name='spec.userConfig.static_ips-property'} (boolean). Use static public IP addresses. -- [`synchronous_replication`](#spec.userConfig.synchronous_replication-property){: name='spec.userConfig.synchronous_replication-property'} (string, Enum: `quorum`, `off`). Synchronous replication type. Note that the service plan also needs to support synchronous replication. +- [`synchronous_replication`](#spec.userConfig.synchronous_replication-property){: name='spec.userConfig.synchronous_replication-property'} (string, Enum: `off`, `quorum`). Synchronous replication type. Note that the service plan also needs to support synchronous replication. - [`timescaledb`](#spec.userConfig.timescaledb-property){: name='spec.userConfig.timescaledb-property'} (object). System-wide settings for the timescaledb extension. See below for [nested schema](#spec.userConfig.timescaledb). - [`variant`](#spec.userConfig.variant-property){: name='spec.userConfig.variant-property'} (string, Enum: `aiven`, `timescale`). Variant of the PostgreSQL service, may affect the features that are exposed by default. - [`work_mem`](#spec.userConfig.work_mem-property){: name='spec.userConfig.work_mem-property'} (integer, Minimum: 1, Maximum: 1024). Sets the maximum amount of memory to be used by a query operation (such as a sort or hash table) before writing to temporary disk files, in MB. Default is 1MB + 0.075% of total RAM (up to 32MB). @@ -290,8 +290,8 @@ postgresql.conf configuration values. - [`idle_in_transaction_session_timeout`](#spec.userConfig.pg.idle_in_transaction_session_timeout-property){: name='spec.userConfig.pg.idle_in_transaction_session_timeout-property'} (integer, Minimum: 0, Maximum: 604800000). Time out sessions with open transactions after this number of milliseconds. - [`jit`](#spec.userConfig.pg.jit-property){: name='spec.userConfig.pg.jit-property'} (boolean). Controls system-wide use of Just-in-Time Compilation (JIT). - [`log_autovacuum_min_duration`](#spec.userConfig.pg.log_autovacuum_min_duration-property){: name='spec.userConfig.pg.log_autovacuum_min_duration-property'} (integer, Minimum: -1, Maximum: 2147483647). Causes each action executed by autovacuum to be logged if it ran for at least the specified number of milliseconds. Setting this to zero logs all autovacuum actions. Minus-one (the default) disables logging autovacuum actions. -- [`log_error_verbosity`](#spec.userConfig.pg.log_error_verbosity-property){: name='spec.userConfig.pg.log_error_verbosity-property'} (string, Enum: `TERSE`, `DEFAULT`, `VERBOSE`). Controls the amount of detail written in the server log for each message that is logged. -- [`log_line_prefix`](#spec.userConfig.pg.log_line_prefix-property){: name='spec.userConfig.pg.log_line_prefix-property'} (string, Enum: `'pid=%p,user=%u,db=%d,app=%a,client=%h '`, `'%t [%p]: [%l-1] user=%u,db=%d,app=%a,client=%h '`, `'%m [%p] %q[user=%u,db=%d,app=%a] '`, `'pid=%p,user=%u,db=%d,app=%a,client=%h,txid=%x,qid=%Q '`). Choose from one of the available log formats. +- [`log_error_verbosity`](#spec.userConfig.pg.log_error_verbosity-property){: name='spec.userConfig.pg.log_error_verbosity-property'} (string, Enum: `DEFAULT`, `TERSE`, `VERBOSE`). Controls the amount of detail written in the server log for each message that is logged. +- [`log_line_prefix`](#spec.userConfig.pg.log_line_prefix-property){: name='spec.userConfig.pg.log_line_prefix-property'} (string, Enum: `'%m [%p] %q[user=%u,db=%d,app=%a] '`, `'%t [%p]: [%l-1] user=%u,db=%d,app=%a,client=%h '`, `'pid=%p,user=%u,db=%d,app=%a,client=%h '`, `'pid=%p,user=%u,db=%d,app=%a,client=%h,txid=%x,qid=%Q '`). Choose from one of the available log formats. - [`log_min_duration_statement`](#spec.userConfig.pg.log_min_duration_statement-property){: name='spec.userConfig.pg.log_min_duration_statement-property'} (integer, Minimum: -1, Maximum: 86400000). Log statements that take more than this number of milliseconds to run, -1 disables. - [`log_temp_files`](#spec.userConfig.pg.log_temp_files-property){: name='spec.userConfig.pg.log_temp_files-property'} (integer, Minimum: -1, Maximum: 2147483647). Log statements for each temporary file created larger than this number of kilobytes, -1 disables. - [`max_files_per_process`](#spec.userConfig.pg.max_files_per_process-property){: name='spec.userConfig.pg.max_files_per_process-property'} (integer, Minimum: 1000, Maximum: 4096). PostgreSQL maximum number of files that can be open per process. @@ -312,12 +312,12 @@ postgresql.conf configuration values. - [`pg_partman_bgw.role`](#spec.userConfig.pg.pg_partman_bgw.role-property){: name='spec.userConfig.pg.pg_partman_bgw.role-property'} (string, Pattern: `^[_A-Za-z0-9][-._A-Za-z0-9]{0,63}$`, MaxLength: 64). Controls which role to use for pg_partman's scheduled background tasks. - [`pg_stat_monitor.pgsm_enable_query_plan`](#spec.userConfig.pg.pg_stat_monitor.pgsm_enable_query_plan-property){: name='spec.userConfig.pg.pg_stat_monitor.pgsm_enable_query_plan-property'} (boolean). Enables or disables query plan monitoring. - [`pg_stat_monitor.pgsm_max_buckets`](#spec.userConfig.pg.pg_stat_monitor.pgsm_max_buckets-property){: name='spec.userConfig.pg.pg_stat_monitor.pgsm_max_buckets-property'} (integer, Minimum: 1, Maximum: 10). Sets the maximum number of buckets. -- [`pg_stat_statements.track`](#spec.userConfig.pg.pg_stat_statements.track-property){: name='spec.userConfig.pg.pg_stat_statements.track-property'} (string, Enum: `all`, `top`, `none`). Controls which statements are counted. Specify top to track top-level statements (those issued directly by clients), all to also track nested statements (such as statements invoked within functions), or none to disable statement statistics collection. The default value is top. +- [`pg_stat_statements.track`](#spec.userConfig.pg.pg_stat_statements.track-property){: name='spec.userConfig.pg.pg_stat_statements.track-property'} (string, Enum: `all`, `none`, `top`). Controls which statements are counted. Specify top to track top-level statements (those issued directly by clients), all to also track nested statements (such as statements invoked within functions), or none to disable statement statistics collection. The default value is top. - [`temp_file_limit`](#spec.userConfig.pg.temp_file_limit-property){: name='spec.userConfig.pg.temp_file_limit-property'} (integer, Minimum: -1, Maximum: 2147483647). PostgreSQL temporary file limit in KiB, -1 for unlimited. - [`timezone`](#spec.userConfig.pg.timezone-property){: name='spec.userConfig.pg.timezone-property'} (string, Pattern: `^[\w/]*$`, MaxLength: 64). PostgreSQL service timezone. - [`track_activity_query_size`](#spec.userConfig.pg.track_activity_query_size-property){: name='spec.userConfig.pg.track_activity_query_size-property'} (integer, Minimum: 1024, Maximum: 10240). Specifies the number of bytes reserved to track the currently executing command for each active session. - [`track_commit_timestamp`](#spec.userConfig.pg.track_commit_timestamp-property){: name='spec.userConfig.pg.track_commit_timestamp-property'} (string, Enum: `off`, `on`). Record commit time of transactions. -- [`track_functions`](#spec.userConfig.pg.track_functions-property){: name='spec.userConfig.pg.track_functions-property'} (string, Enum: `all`, `pl`, `none`). Enables tracking of function call counts and time used. +- [`track_functions`](#spec.userConfig.pg.track_functions-property){: name='spec.userConfig.pg.track_functions-property'} (string, Enum: `all`, `none`, `pl`). Enables tracking of function call counts and time used. - [`track_io_timing`](#spec.userConfig.pg.track_io_timing-property){: name='spec.userConfig.pg.track_io_timing-property'} (string, Enum: `off`, `on`). Enables timing of database I/O calls. This parameter is off by default, because it will repeatedly query the operating system for the current time, which may cause significant overhead on some platforms. - [`wal_sender_timeout`](#spec.userConfig.pg.wal_sender_timeout-property){: name='spec.userConfig.pg.wal_sender_timeout-property'} (integer). Terminate replication connections that are inactive for longer than this amount of time, in milliseconds. Setting this value to zero disables the timeout. - [`wal_writer_delay`](#spec.userConfig.pg.wal_writer_delay-property){: name='spec.userConfig.pg.wal_writer_delay-property'} (integer, Minimum: 10, Maximum: 200). WAL flush interval in milliseconds. Note that setting this value to lower than the default 200ms may negatively impact performance. @@ -369,7 +369,7 @@ PGBouncer connection pooling settings. - [`autodb_idle_timeout`](#spec.userConfig.pgbouncer.autodb_idle_timeout-property){: name='spec.userConfig.pgbouncer.autodb_idle_timeout-property'} (integer, Minimum: 0, Maximum: 86400). If the automatically created database pools have been unused this many seconds, they are freed. If 0 then timeout is disabled. [seconds]. - [`autodb_max_db_connections`](#spec.userConfig.pgbouncer.autodb_max_db_connections-property){: name='spec.userConfig.pgbouncer.autodb_max_db_connections-property'} (integer, Minimum: 0, Maximum: 2147483647). Do not allow more than this many server connections per database (regardless of user). Setting it to 0 means unlimited. -- [`autodb_pool_mode`](#spec.userConfig.pgbouncer.autodb_pool_mode-property){: name='spec.userConfig.pgbouncer.autodb_pool_mode-property'} (string, Enum: `session`, `transaction`, `statement`). PGBouncer pool mode. +- [`autodb_pool_mode`](#spec.userConfig.pgbouncer.autodb_pool_mode-property){: name='spec.userConfig.pgbouncer.autodb_pool_mode-property'} (string, Enum: `session`, `statement`, `transaction`). PGBouncer pool mode. - [`autodb_pool_size`](#spec.userConfig.pgbouncer.autodb_pool_size-property){: name='spec.userConfig.pgbouncer.autodb_pool_size-property'} (integer, Minimum: 0, Maximum: 10000). If non-zero then create automatically a pool of that size per user when a pool doesn't exist. - [`ignore_startup_parameters`](#spec.userConfig.pgbouncer.ignore_startup_parameters-property){: name='spec.userConfig.pgbouncer.ignore_startup_parameters-property'} (array of strings, MaxItems: 32). List of parameters to ignore when given in startup packet. - [`max_prepared_statements`](#spec.userConfig.pgbouncer.max_prepared_statements-property){: name='spec.userConfig.pgbouncer.max_prepared_statements-property'} (integer, Minimum: 0, Maximum: 3000). PgBouncer tracks protocol-level named prepared statements related commands sent by the client in transaction and statement pooling modes when max_prepared_statements is set to a non-zero value. Setting it to 0 disables prepared statements. max_prepared_statements defaults to 100, and its maximum is 3000. diff --git a/docs/docs/api-reference/redis.md b/docs/docs/api-reference/redis.md index 4090ef4d..9d65cf7e 100644 --- a/docs/docs/api-reference/redis.md +++ b/docs/docs/api-reference/redis.md @@ -208,7 +208,7 @@ Redis specific user configuration options. - [`redis_io_threads`](#spec.userConfig.redis_io_threads-property){: name='spec.userConfig.redis_io_threads-property'} (integer, Minimum: 1, Maximum: 32). Set Redis IO thread count. Changing this will cause a restart of the Redis service. - [`redis_lfu_decay_time`](#spec.userConfig.redis_lfu_decay_time-property){: name='spec.userConfig.redis_lfu_decay_time-property'} (integer, Minimum: 1, Maximum: 120). LFU maxmemory-policy counter decay time in minutes. - [`redis_lfu_log_factor`](#spec.userConfig.redis_lfu_log_factor-property){: name='spec.userConfig.redis_lfu_log_factor-property'} (integer, Minimum: 0, Maximum: 100). Counter logarithm factor for volatile-lfu and allkeys-lfu maxmemory-policies. -- [`redis_maxmemory_policy`](#spec.userConfig.redis_maxmemory_policy-property){: name='spec.userConfig.redis_maxmemory_policy-property'} (string, Enum: `noeviction`, `allkeys-lru`, `volatile-lru`, `allkeys-random`, `volatile-random`, `volatile-ttl`, `volatile-lfu`, `allkeys-lfu`). Redis maxmemory-policy. +- [`redis_maxmemory_policy`](#spec.userConfig.redis_maxmemory_policy-property){: name='spec.userConfig.redis_maxmemory_policy-property'} (string, Enum: `allkeys-lfu`, `allkeys-lru`, `allkeys-random`, `noeviction`, `volatile-lfu`, `volatile-lru`, `volatile-random`, `volatile-ttl`). Redis maxmemory-policy. - [`redis_notify_keyspace_events`](#spec.userConfig.redis_notify_keyspace_events-property){: name='spec.userConfig.redis_notify_keyspace_events-property'} (string, Pattern: `^[KEg\$lshzxentdmA]*$`, MaxLength: 32). Set notify-keyspace-events option. - [`redis_number_of_databases`](#spec.userConfig.redis_number_of_databases-property){: name='spec.userConfig.redis_number_of_databases-property'} (integer, Minimum: 1, Maximum: 128). Set number of Redis databases. Changing this will cause a restart of the Redis service. - [`redis_persistence`](#spec.userConfig.redis_persistence-property){: name='spec.userConfig.redis_persistence-property'} (string, Enum: `off`, `rdb`). When persistence is `rdb`, Redis does RDB dumps each 10 minutes if any key is changed. Also RDB dumps are done according to the backup schedule for backup purposes. When persistence is `off`, no RDB dumps or backups are done, so data can be lost at any moment if the service is restarted for any reason, or if the service is powered off. Also, the service can't be forked. diff --git a/docs/docs/api-reference/serviceintegration.md b/docs/docs/api-reference/serviceintegration.md index 1a0ed1cf..a2725be4 100644 --- a/docs/docs/api-reference/serviceintegration.md +++ b/docs/docs/api-reference/serviceintegration.md @@ -300,14 +300,14 @@ Table to create. **Required** - [`columns`](#spec.clickhouseKafka.tables.columns-property){: name='spec.clickhouseKafka.tables.columns-property'} (array of objects, MaxItems: 100). Table columns. See below for [nested schema](#spec.clickhouseKafka.tables.columns). -- [`data_format`](#spec.clickhouseKafka.tables.data_format-property){: name='spec.clickhouseKafka.tables.data_format-property'} (string, Enum: `Avro`, `CSV`, `JSONAsString`, `JSONCompactEachRow`, `JSONCompactStringsEachRow`, `JSONEachRow`, `JSONStringsEachRow`, `MsgPack`, `TSKV`, `TSV`, `TabSeparated`, `RawBLOB`, `AvroConfluent`, `Parquet`). Message data format. +- [`data_format`](#spec.clickhouseKafka.tables.data_format-property){: name='spec.clickhouseKafka.tables.data_format-property'} (string, Enum: `Avro`, `AvroConfluent`, `CSV`, `JSONAsString`, `JSONCompactEachRow`, `JSONCompactStringsEachRow`, `JSONEachRow`, `JSONStringsEachRow`, `MsgPack`, `Parquet`, `RawBLOB`, `TSKV`, `TSV`, `TabSeparated`). Message data format. - [`group_name`](#spec.clickhouseKafka.tables.group_name-property){: name='spec.clickhouseKafka.tables.group_name-property'} (string, MinLength: 1, MaxLength: 249). Kafka consumers group. - [`name`](#spec.clickhouseKafka.tables.name-property){: name='spec.clickhouseKafka.tables.name-property'} (string, MinLength: 1, MaxLength: 40). Name of the table. - [`topics`](#spec.clickhouseKafka.tables.topics-property){: name='spec.clickhouseKafka.tables.topics-property'} (array of objects, MaxItems: 100). Kafka topics. See below for [nested schema](#spec.clickhouseKafka.tables.topics). **Optional** -- [`auto_offset_reset`](#spec.clickhouseKafka.tables.auto_offset_reset-property){: name='spec.clickhouseKafka.tables.auto_offset_reset-property'} (string, Enum: `smallest`, `earliest`, `beginning`, `largest`, `latest`, `end`). Action to take when there is no initial offset in offset store or the desired offset is out of range. +- [`auto_offset_reset`](#spec.clickhouseKafka.tables.auto_offset_reset-property){: name='spec.clickhouseKafka.tables.auto_offset_reset-property'} (string, Enum: `beginning`, `earliest`, `end`, `largest`, `latest`, `smallest`). Action to take when there is no initial offset in offset store or the desired offset is out of range. - [`date_time_input_format`](#spec.clickhouseKafka.tables.date_time_input_format-property){: name='spec.clickhouseKafka.tables.date_time_input_format-property'} (string, Enum: `basic`, `best_effort`, `best_effort_us`). Method to read DateTime from text input formats. - [`handle_error_mode`](#spec.clickhouseKafka.tables.handle_error_mode-property){: name='spec.clickhouseKafka.tables.handle_error_mode-property'} (string, Enum: `default`, `stream`). How to handle errors for Kafka engine. - [`max_block_size`](#spec.clickhouseKafka.tables.max_block_size-property){: name='spec.clickhouseKafka.tables.max_block_size-property'} (integer, Minimum: 0, Maximum: 1000000000). Number of row collected by poll(s) for flushing data from Kafka. @@ -512,7 +512,7 @@ Kafka MirrorMaker configuration values. - [`consumer_max_poll_records`](#spec.kafkaMirrormaker.kafka_mirrormaker.consumer_max_poll_records-property){: name='spec.kafkaMirrormaker.kafka_mirrormaker.consumer_max_poll_records-property'} (integer, Minimum: 100, Maximum: 20000). Set consumer max.poll.records. The default is 500. - [`producer_batch_size`](#spec.kafkaMirrormaker.kafka_mirrormaker.producer_batch_size-property){: name='spec.kafkaMirrormaker.kafka_mirrormaker.producer_batch_size-property'} (integer, Minimum: 0, Maximum: 5242880). The batch size in bytes producer will attempt to collect before publishing to broker. - [`producer_buffer_memory`](#spec.kafkaMirrormaker.kafka_mirrormaker.producer_buffer_memory-property){: name='spec.kafkaMirrormaker.kafka_mirrormaker.producer_buffer_memory-property'} (integer, Minimum: 5242880, Maximum: 134217728). The amount of bytes producer can use for buffering data before publishing to broker. -- [`producer_compression_type`](#spec.kafkaMirrormaker.kafka_mirrormaker.producer_compression_type-property){: name='spec.kafkaMirrormaker.kafka_mirrormaker.producer_compression_type-property'} (string, Enum: `gzip`, `snappy`, `lz4`, `zstd`, `none`). Specify the default compression type for producers. This configuration accepts the standard compression codecs (`gzip`, `snappy`, `lz4`, `zstd`). It additionally accepts `none` which is the default and equivalent to no compression. +- [`producer_compression_type`](#spec.kafkaMirrormaker.kafka_mirrormaker.producer_compression_type-property){: name='spec.kafkaMirrormaker.kafka_mirrormaker.producer_compression_type-property'} (string, Enum: `gzip`, `lz4`, `none`, `snappy`, `zstd`). Specify the default compression type for producers. This configuration accepts the standard compression codecs (`gzip`, `snappy`, `lz4`, `zstd`). It additionally accepts `none` which is the default and equivalent to no compression. - [`producer_linger_ms`](#spec.kafkaMirrormaker.kafka_mirrormaker.producer_linger_ms-property){: name='spec.kafkaMirrormaker.kafka_mirrormaker.producer_linger_ms-property'} (integer, Minimum: 0, Maximum: 5000). The linger time (ms) for waiting new data to arrive for publishing. - [`producer_max_request_size`](#spec.kafkaMirrormaker.kafka_mirrormaker.producer_max_request_size-property){: name='spec.kafkaMirrormaker.kafka_mirrormaker.producer_max_request_size-property'} (integer, Minimum: 0, Maximum: 268435456). The maximum request size in bytes. diff --git a/docs/docs/api-reference/serviceintegrationendpoint.md b/docs/docs/api-reference/serviceintegrationendpoint.md index ce48b8a7..672f0242 100644 --- a/docs/docs/api-reference/serviceintegrationendpoint.md +++ b/docs/docs/api-reference/serviceintegrationendpoint.md @@ -138,7 +138,7 @@ Datadog configuration values. - [`kafka_consumer_check_instances`](#spec.datadog.kafka_consumer_check_instances-property){: name='spec.datadog.kafka_consumer_check_instances-property'} (integer, Minimum: 1, Maximum: 100). Number of separate instances to fetch kafka consumer statistics with. - [`kafka_consumer_stats_timeout`](#spec.datadog.kafka_consumer_stats_timeout-property){: name='spec.datadog.kafka_consumer_stats_timeout-property'} (integer, Minimum: 2, Maximum: 300). Number of seconds that datadog will wait to get consumer statistics from brokers. - [`max_partition_contexts`](#spec.datadog.max_partition_contexts-property){: name='spec.datadog.max_partition_contexts-property'} (integer, Minimum: 200, Maximum: 200000). Maximum number of partition contexts to send. -- [`site`](#spec.datadog.site-property){: name='spec.datadog.site-property'} (string, Enum: `datadoghq.com`, `datadoghq.eu`, `us3.datadoghq.com`, `us5.datadoghq.com`, `ddog-gov.com`, `ap1.datadoghq.com`). Datadog intake site. Defaults to datadoghq.com. +- [`site`](#spec.datadog.site-property){: name='spec.datadog.site-property'} (string, Enum: `ap1.datadoghq.com`, `datadoghq.com`, `datadoghq.eu`, `ddog-gov.com`, `us3.datadoghq.com`, `us5.datadoghq.com`). Datadog intake site. Defaults to datadoghq.com. ### datadog_tags {: #spec.datadog.datadog_tags } @@ -232,7 +232,7 @@ ExternalKafka configuration values. **Required** - [`bootstrap_servers`](#spec.externalKafka.bootstrap_servers-property){: name='spec.externalKafka.bootstrap_servers-property'} (string, MinLength: 3, MaxLength: 256). Bootstrap servers. -- [`security_protocol`](#spec.externalKafka.security_protocol-property){: name='spec.externalKafka.security_protocol-property'} (string, Enum: `PLAINTEXT`, `SSL`, `SASL_PLAINTEXT`, `SASL_SSL`). Security protocol. +- [`security_protocol`](#spec.externalKafka.security_protocol-property){: name='spec.externalKafka.security_protocol-property'} (string, Enum: `PLAINTEXT`, `SASL_PLAINTEXT`, `SASL_SSL`, `SSL`). Security protocol. **Optional** @@ -290,7 +290,7 @@ ExternalSchemaRegistry configuration values. **Required** -- [`authentication`](#spec.externalSchemaRegistry.authentication-property){: name='spec.externalSchemaRegistry.authentication-property'} (string, Enum: `none`, `basic`). Authentication method. +- [`authentication`](#spec.externalSchemaRegistry.authentication-property){: name='spec.externalSchemaRegistry.authentication-property'} (string, Enum: `basic`, `none`). Authentication method. - [`url`](#spec.externalSchemaRegistry.url-property){: name='spec.externalSchemaRegistry.url-property'} (string, MaxLength: 2048). Schema Registry URL. **Optional** @@ -328,7 +328,7 @@ Rsyslog configuration values. **Required** -- [`format`](#spec.rsyslog.format-property){: name='spec.rsyslog.format-property'} (string, Enum: `rfc5424`, `rfc3164`, `custom`). Message format. +- [`format`](#spec.rsyslog.format-property){: name='spec.rsyslog.format-property'} (string, Enum: `custom`, `rfc3164`, `rfc5424`). Message format. - [`port`](#spec.rsyslog.port-property){: name='spec.rsyslog.port-property'} (integer, Minimum: 1, Maximum: 65535). Rsyslog server port. - [`server`](#spec.rsyslog.server-property){: name='spec.rsyslog.server-property'} (string, MinLength: 4, MaxLength: 255). Rsyslog server IP address or hostname. - [`tls`](#spec.rsyslog.tls-property){: name='spec.rsyslog.tls-property'} (boolean). Require TLS.