GetShard
Returns the specified shard.
- TypeScript
- Python
import { cloudApi, serviceClients, Session } from "@yandex-cloud/nodejs-sdk";
const GetClusterShardRequest =
cloudApi.mdb.clickhouse_cluster_service.GetClusterShardRequest;
(async () => {
const authToken = process.env["YC_OAUTH_TOKEN"];
const session = new Session({ oauthToken: authToken });
const client = session.client(serviceClients.ClusterServiceClient);
const result = await client.getShard(
GetClusterShardRequest.fromPartial({
clusterId: "clusterId",
shardName: "shardName",
})
);
console.log(result);
})();
import os
import grpc
import yandexcloud
from yandex.cloud.dataproc.v1.cluster_service_pb2_grpc import ClusterServiceStub
from yandex.cloud.mdb.clickhouse.v1.cluster_service_pb2 import GetClusterShardRequest
token = os.getenv("YC_OAUTH_TOKEN")
sdk = yandexcloud.SDK(token=token)
service = sdk.client(ClusterServiceStub)
response = service.GetShard(GetClusterShardRequest(cluster_id="clusterId", shard_name="shardName"))
print(response)
GetClusterShardRequest
clusterId
: string
ID of the cluster that the shard belongs to. To get the cluster ID, use a ClusterService.List request. To get the name of the database, use a ClusterService.List request.
shardName
: string
Name of the shard to request information about. To get the name of a shard, use a ClusterService.ListShards request.
Shard
name
: string
Name of the shard.
clusterId
: string
ID of the cluster that the shard belongs to.
config
: ShardConfig
Configuration of the shard.
ShardConfig
Clickhouse
config
: config.ClickhouseConfigSet
ClickHouse settings for a shard.
resources
: Resources
Computational resources for a shard.
weight
: google.protobuf.Int64Value
Relative weight of a shard considered when writing data to the cluster. For details, see ClickHouse documentation.
clickhouse
: Clickhouse
ClickHouse configuration for a shard.
ClickhouseConfigSet
effectiveConfig
: ClickhouseConfig
Effective settings for a ClickHouse cluster (a combination of settings defined in user_config and default_config).
userConfig
: ClickhouseConfig
User-defined settings for a ClickHouse cluster.
defaultConfig
: ClickhouseConfig
Default configuration for a ClickHouse cluster.
Resources
resourcePresetId
: string
ID of the preset for computational resources available to a host (CPU, memory etc.). All available presets are listed in the documentation
diskSize
: int64
Volume of the storage available to a host, in bytes.
diskTypeId
: string
Type of the storage environment for the host. Possible values:
- network-hdd - network HDD drive,
- network-ssd - network SSD drive,
- local-ssd - local SSD storage.
Clickhouse
config
: config.ClickhouseConfigSet
ClickHouse settings for a shard.
resources
: Resources
Computational resources for a shard.
weight
: google.protobuf.Int64Value
Relative weight of a shard considered when writing data to the cluster. For details, see ClickHouse documentation.
ClickhouseConfig
ClickHouse configuration options. Detailed description for each set of options is available in ClickHouse documentation.
Any options not listed here are not supported.
LogLevel
LOG_LEVEL_UNSPECIFIED
TRACE
DEBUG
INFORMATION
WARNING
ERROR
MergeTree
Options specific to the MergeTree table engine.
replicatedDeduplicationWindow
: google.protobuf.Int64Value
Number of blocks of hashes to keep in ZooKeeper.
replicatedDeduplicationWindowSeconds
: google.protobuf.Int64Value
Period of time to keep blocks of hashes for.
partsToDelayInsert
: google.protobuf.Int64Value
If table contains at least that many active parts in single partition, artificially slow down insert into table.
partsToThrowInsert
: google.protobuf.Int64Value
If more than this number active parts in single partition, throw 'Too many parts ...' exception.
inactivePartsToDelayInsert
: google.protobuf.Int64Value
inactivePartsToThrowInsert
: google.protobuf.Int64Value
maxReplicatedMergesInQueue
: google.protobuf.Int64Value
How many tasks of merging and mutating parts are allowed simultaneously in ReplicatedMergeTree queue.
numberOfFreeEntriesInPoolToLowerMaxSizeOfMerge
: google.protobuf.Int64Value
If there is less than specified number of free entries in background pool (or replicated queue), start to lower maximum size of merge to process.
maxBytesToMergeAtMinSpaceInPool
: google.protobuf.Int64Value
Maximum in total size of parts to merge, when there are minimum free threads in background pool (or entries in replication queue).
maxBytesToMergeAtMaxSpaceInPool
: google.protobuf.Int64Value
minBytesForWidePart
: google.protobuf.Int64Value
Minimum number of bytes in a data part that can be stored in Wide format.
More info see in ClickHouse documentation.
minRowsForWidePart
: google.protobuf.Int64Value
Minimum number of rows in a data part that can be stored in Wide format.
More info see in ClickHouse documentation.
ttlOnlyDropParts
: google.protobuf.BoolValue
Enables or disables complete dropping of data parts where all rows are expired in MergeTree tables.
More info see in ClickHouse documentation.
allowRemoteFsZeroCopyReplication
: google.protobuf.BoolValue
mergeWithTtlTimeout
: google.protobuf.Int64Value
mergeWithRecompressionTtlTimeout
: google.protobuf.Int64Value
maxPartsInTotal
: google.protobuf.Int64Value
maxNumberOfMergesWithTtlInPool
: google.protobuf.Int64Value
cleanupDelayPeriod
: google.protobuf.Int64Value
numberOfFreeEntriesInPoolToExecuteMutation
: google.protobuf.Int64Value
maxAvgPartSizeForTooManyParts
: google.protobuf.Int64Value
The 'too many parts' check according to 'parts_to_delay_insert' and 'parts_to_throw_insert' will be active only if the average part size (in the relevant partition) is not larger than the specified threshold. If it is larger than the specified threshold, the INSERTs will be neither delayed or rejected. This allows to have hundreds of terabytes in a single table on a single server if the parts are successfully merged to larger parts. This does not affect the thresholds on inactive parts or total parts. Default: 1 GiB Min version: 22.10 See in-depth description in ClickHouse GitHub
minAgeToForceMergeSeconds
: google.protobuf.Int64Value
Merge parts if every part in the range is older than the value of min_age_to_force_merge_seconds. Default: 0 - disabled Min_version: 22.10 See in-depth description in ClickHouse documentation
minAgeToForceMergeOnPartitionOnly
: google.protobuf.BoolValue
Whether min_age_to_force_merge_seconds should be applied only on the entire partition and not on subset. Default: false Min_version: 22.11 See in-depth description in ClickHouse documentation
mergeSelectingSleepMs
: google.protobuf.Int64Value
Sleep time for merge selecting when no part is selected. A lower setting triggers selecting tasks in background_schedule_pool frequently, which results in a large number of requests to ClickHouse Keeper in large-scale clusters. Default: 5000 Min_version: 21.10 See in-depth description in ClickHouse documentation
mergeMaxBlockSize
: google.protobuf.Int64Value
The number of rows that are read from the merged parts into memory. Default: 8192 See in-depth description in ClickHouse documentation
checkSampleColumnIsCorrect
: google.protobuf.BoolValue
Enables the check at table creation, that the data type of a column for sampling or sampling expression is correct. The data type must be one of unsigned integer types: UInt8, UInt16, UInt32, UInt64. Default: true See in-depth description in ClickHouse documentation
maxMergeSelectingSleepMs
: google.protobuf.Int64Value
Maximum sleep time for merge selecting, a lower setting will trigger selecting tasks in background_schedule_pool frequently which result in large amount of requests to zookeeper in large-scale clusters. Default: 60000 Min_version: 23.6 See in-depth description in ClickHouse GitHub
maxCleanupDelayPeriod
: google.protobuf.Int64Value
Maximum period to clean old queue logs, blocks hashes and parts. Default: 300 Min_version: 23.6 See in-depth description in ClickHouse GitHub
Kafka
SecurityProtocol
SECURITY_PROTOCOL_UNSPECIFIED
SECURITY_PROTOCOL_PLAINTEXT
SECURITY_PROTOCOL_SSL
SECURITY_PROTOCOL_SASL_PLAINTEXT
SECURITY_PROTOCOL_SASL_SSL
SaslMechanism
SASL_MECHANISM_UNSPECIFIED
SASL_MECHANISM_GSSAPI
SASL_MECHANISM_PLAIN
SASL_MECHANISM_SCRAM_SHA_256
SASL_MECHANISM_SCRAM_SHA_512
Debug
DEBUG_UNSPECIFIED
DEBUG_GENERIC
DEBUG_BROKER
DEBUG_TOPIC
DEBUG_METADATA
DEBUG_FEATURE
DEBUG_QUEUE
DEBUG_MSG
DEBUG_PROTOCOL
DEBUG_CGRP
DEBUG_SECURITY
DEBUG_FETCH
DEBUG_INTERCEPTOR
DEBUG_PLUGIN
DEBUG_CONSUMER
DEBUG_ADMIN
DEBUG_EOS
DEBUG_MOCK
DEBUG_ASSIGNOR
DEBUG_CONF
DEBUG_TELEMETRY
DEBUG_ALL
AutoOffsetReset
AUTO_OFFSET_RESET_UNSPECIFIED
AUTO_OFFSET_RESET_SMALLEST
AUTO_OFFSET_RESET_EARLIEST
AUTO_OFFSET_RESET_BEGINNING
AUTO_OFFSET_RESET_LARGEST
AUTO_OFFSET_RESET_LATEST
AUTO_OFFSET_RESET_END
AUTO_OFFSET_RESET_ERROR
securityProtocol
: SecurityProtocol
saslMechanism
: SaslMechanism
saslUsername
: string
saslPassword
: string
enableSslCertificateVerification
: google.protobuf.BoolValue
maxPollIntervalMs
: google.protobuf.Int64Value
sessionTimeoutMs
: google.protobuf.Int64Value
debug
: Debug
autoOffsetReset
: AutoOffsetReset
KafkaTopic
name
: string
settings
: Kafka
Rabbitmq
username
: string
RabbitMQ username
password
: string
RabbitMQ password
vhost
: string
RabbitMQ virtual host
Compression
Method
METHOD_UNSPECIFIED
LZ4
ZSTD
method
: Method
Compression method to use for the specified combination of min_part_size and min_part_size_ratio.
minPartSize
: int64
Minimum size of a part of a table.
minPartSizeRatio
: double
Minimum ratio of a part relative to the size of all the data in the table.
level
: google.protobuf.Int64Value
ExternalDictionary
HttpSource
####### Header {#httpsource_header}
name
: string
value
: string
url
: string
URL of the source dictionary available over HTTP.
format
: string
The data format. Valid values are all formats supported by ClickHouse SQL dialect.
headers
: Header
HTTP headers.
MysqlSource
####### Replica {#mysqlsource_replica}
host
: string
MySQL host of the replica.
priority
: int64
The priority of the replica that ClickHouse takes into account when connecting. Replica with the highest priority should have this field set to the lowest number.
port
: int64
Port to use when connecting to the replica. If a port is not specified for a replica, ClickHouse uses the port specified for the source.
user
: string
Name of the MySQL database user.
password
: string
Password of the MySQL database user.
db
: string
Name of the MySQL database to connect to.
table
: string
Name of the database table to use as a ClickHouse dictionary.
port
: int64
Default port to use when connecting to a replica of the dictionary source.
user
: string
Name of the default user for replicas of the dictionary source.
password
: string
Password of the default user for replicas of the dictionary source.
replicas
: Replica
List of MySQL replicas of the database used as dictionary source.
where
: string
Selection criteria for the data in the specified MySQL table.
invalidateQuery
: string
Query for checking the dictionary status, to pull only updated data. For more details, see ClickHouse documentation on dictionaries.
closeConnection
: google.protobuf.BoolValue
Should the connection be closed after each request.
shareConnection
: google.protobuf.BoolValue
Should a connection be shared for some requests.
ClickhouseSource
db
: string
Name of the ClickHouse database.
table
: string
Name of the table in the specified database to be used as the dictionary source.
host
: string
ClickHouse host of the specified database.
port
: int64
Port to use when connecting to the host.
user
: string
Name of the ClickHouse database user.
password
: string
Password of the ClickHouse database user.
where
: string
Selection criteria for the data in the specified ClickHouse table.
secure
: google.protobuf.BoolValue
Use ssl for connection.
MongodbSource
db
: string
Name of the MongoDB database.
collection
: string
Name of the collection in the specified database to be used as the dictionary source.
host
: string
MongoDB host of the specified database.
port
: int64
Port to use when connecting to the host.
user
: string
Name of the MongoDB database user.
password
: string
Password of the MongoDB database user.
options
: string
PostgresqlSource
####### SslMode {#postgresqlsource_sslmode}
SSL_MODE_UNSPECIFIED
DISABLE
Only try a non-SSL connection.
ALLOW
First try a non-SSL connection; if that fails, try an SSL connection.
PREFER
First try an SSL connection; if that fails, try a non-SSL connection.
VERIFY_CA
Only try an SSL connection, and verify that the server certificate is issued by a trusted certificate authority (CA).
VERIFY_FULL
Only try an SSL connection, verify that the server certificate is issued by a trusted CA and that the requested server host name matches that in the certificate.
db
: string
Name of the PostrgreSQL database.
table
: string
Name of the table in the specified database to be used as the dictionary source.
hosts
: string
Name of the PostrgreSQL host
port
: int64
Port to use when connecting to the host.
user
: string
Name of the PostrgreSQL database user.
password
: string
Password of the PostrgreSQL database user.
invalidateQuery
: string
Query for checking the dictionary status, to pull only updated data. For more details, see ClickHouse documentation on dictionaries.
sslMode
: SslMode
Mode of SSL TCP/IP connection to the PostgreSQL host. For more details, see PostgreSQL documentation.
Structure
####### Attribute {#structure_attribute}
name
: string
Name of the column.
type
: string
Type of the column.
nullValue
: string
Default value for an element without data (for example, an empty string).
expression
: string
Expression, describing the attribute, if applicable.
hierarchical
: bool
Indication of hierarchy support.
Default value: false
.
injective
: bool
Indication of injective mapping "id -> attribute".
Default value: false
.
####### Id {#structure_id}
Numeric key.
name
: string
Name of the numeric key.
####### Key {#structure_key}
Complex key.
attributes
: Attribute
Attributes of a complex key.
id
: Id
Single numeric key column for the dictionary.
key
: Key
Composite key for the dictionary, containing of one or more key columns. For details, see ClickHouse documentation.
rangeMin
: Attribute
Field holding the beginning of the range for dictionaries with RANGE_HASHED
layout.
For details, see ClickHouse documentation.
rangeMax
: Attribute
Field holding the end of the range for dictionaries with RANGE_HASHED
layout.
For details, see ClickHouse documentation.
attributes
: Attribute
Description of the fields available for database queries. For details, see ClickHouse documentation.
Layout
Layout determining how to store the dictionary in memory.
####### Type {#layout_type}
TYPE_UNSPECIFIED
FLAT
The entire dictionary is stored in memory in the form of flat arrays. Available for all dictionary sources.
HASHED
The entire dictionary is stored in memory in the form of a hash table. Available for all dictionary sources.
COMPLEX_KEY_HASHED
Similar to HASHED, to be used with composite keys. Available for all dictionary sources.
RANGE_HASHED
The entire dictionary is stored in memory in the form of a hash table, with an ordered array of ranges and their corresponding values. Available for all dictionary sources.
CACHE
The dictionary is stored in a cache with a set number of cells. Available for MySQL, ClickHouse and HTTP dictionary sources.
COMPLEX_KEY_CACHE
Similar to CACHE, to be used with composite keys. Available for MySQL, ClickHouse and HTTP dictionary sources.
type
: Type
Layout type for an external dictionary.
sizeInCells
: int64
Number of cells in the cache. Rounded up to a power of two. Applicable only for CACHE and COMPLEX_KEY_CACHE layout types.
maxArraySize
: int64
Maximum dictionary key size. Applicable only for FLAT layout type.
Range
min
: int64
Minimum dictionary lifetime.
max
: int64
Maximum dictionary lifetime.
name
: string
Name of the external dictionary.
structure
: Structure
Set of attributes for the external dictionary. For in-depth description, see ClickHouse documentation.
layout
: Layout
Layout for storing the dictionary in memory. For in-depth description, see ClickHouse documentation.
One of lifetime
Setting for the period of time between dictionary updates. For details, see ClickHouse documentation.
fixedLifetime
: int64Fixed interval between dictionary updates.
lifetimeRange
: RangeRange of intervals between dictionary updates for ClickHouse to choose from.
One of source
Description of the source for the external dictionary.
httpSource
: HttpSourceHTTP source for the dictionary.
mysqlSource
: MysqlSourceMySQL source for the dictionary.
clickhouseSource
: ClickhouseSourceClickHouse source for the dictionary.
mongodbSource
: MongodbSourceMongoDB source for the dictionary.
postgresqlSource
: PostgresqlSourcePostgreSQL source for the dictionary.
GraphiteRollup
Rollup settings for the GraphiteMergeTree table engine.
Pattern
####### Retention {#pattern_retention}
age
: int64
Minimum age of the data in seconds.
precision
: int64
Precision of determining the age of the data, in seconds.
regexp
: string
Pattern for metric names.
function
: string
Name of the aggregating function to apply to data of the age specified in retention.
retention
: Retention
Age of data to use for thinning.
name
: string
Name for the specified combination of settings for Graphite rollup.
patterns
: Pattern
Pattern to use for the rollup.
pathColumnName
: string
The name of the column storing the metric name (Graphite sensor). Default: Path See in-depth description in ClickHouse documentation
timeColumnName
: string
The name of the column storing the time of measuring the metric. Default: Time See in-depth description in ClickHouse documentation
valueColumnName
: string
The name of the column storing the value of the metric at the time set in time_column_name. Default: Value See in-depth description in ClickHouse documentation
versionColumnName
: string
The name of the column storing the version of the metric. Default: Timestamp See in-depth description in ClickHouse documentation
QueryMaskingRule
name
: string
Name for the rule.
regexp
: string
RE2 compatible regular expression. Required.
replace
: string
Substitution string for sensitive data. Default: six asterisks
QueryCache
maxSizeInBytes
: google.protobuf.Int64Value
The maximum cache size in bytes. Default: 1073741824 (1 GiB)
maxEntries
: google.protobuf.Int64Value
The maximum number of SELECT query results stored in the cache. Default: 1024
maxEntrySizeInBytes
: google.protobuf.Int64Value
The maximum size in bytes SELECT query results may have to be saved in the cache. Dafault: 1048576 (1 MiB)
maxEntrySizeInRows
: google.protobuf.Int64Value
The maximum number of rows SELECT query results may have to be saved in the cache. Default: 30000000 (30 mil)
logLevel
: LogLevel
Logging level for the ClickHouse cluster. Possible values: TRACE, DEBUG, INFORMATION, WARNING, ERROR.
mergeTree
: MergeTree
Settings for the MergeTree engine. See description in ClickHouse documentation.
compression
: Compression
Compression settings for the ClickHouse cluster. See in-depth description in ClickHouse documentation.
dictionaries
: ExternalDictionary
Configuration of external dictionaries to be used by the ClickHouse cluster. See in-depth description in ClickHouse documentation.
graphiteRollup
: GraphiteRollup
Settings for thinning Graphite data. See in-depth description in ClickHouse documentation.
kafka
: Kafka
kafkaTopics
: KafkaTopic
rabbitmq
: Rabbitmq
maxConnections
: google.protobuf.Int64Value
Maximum number of inbound connections.
maxConcurrentQueries
: google.protobuf.Int64Value
Maximum number of simultaneously processed requests.
keepAliveTimeout
: google.protobuf.Int64Value
Number of milliseconds that ClickHouse waits for incoming requests before closing the connection.
uncompressedCacheSize
: google.protobuf.Int64Value
Cache size (in bytes) for uncompressed data used by MergeTree tables.
markCacheSize
: google.protobuf.Int64Value
Approximate size (in bytes) of the cache of "marks" used by MergeTree tables.
maxTableSizeToDrop
: google.protobuf.Int64Value
Maximum size of the table that can be deleted using a DROP query.
maxPartitionSizeToDrop
: google.protobuf.Int64Value
Maximum size of the partition that can be deleted using a DROP query.
builtinDictionariesReloadInterval
: google.protobuf.Int64Value
The setting is deprecated and has no effect.
timezone
: string
The server's time zone to be used in DateTime fields conversions. Specified as an IANA identifier.
geobaseEnabled
: google.protobuf.BoolValue
Enable or disable geobase.
geobaseUri
: string
Address of the archive with the user geobase in Object Storage.
queryLogRetentionSize
: google.protobuf.Int64Value
The maximum size that query_log can grow to before old data will be removed. If set to 0, automatic removal of query_log data based on size is disabled.
queryLogRetentionTime
: google.protobuf.Int64Value
The maximum time that query_log records will be retained before removal. If set to 0, automatic removal of query_log data based on time is disabled.
queryThreadLogEnabled
: google.protobuf.BoolValue
Whether query_thread_log system table is enabled.
queryThreadLogRetentionSize
: google.protobuf.Int64Value
The maximum size that query_thread_log can grow to before old data will be removed. If set to 0, automatic removal of query_thread_log data based on size is disabled.
queryThreadLogRetentionTime
: google.protobuf.Int64Value
The maximum time that query_thread_log records will be retained before removal. If set to 0, automatic removal of query_thread_log data based on time is disabled.
partLogRetentionSize
: google.protobuf.Int64Value
The maximum size that part_log can grow to before old data will be removed. If set to 0, automatic removal of part_log data based on size is disabled.
partLogRetentionTime
: google.protobuf.Int64Value
The maximum time that part_log records will be retained before removal. If set to 0, automatic removal of part_log data based on time is disabled.
metricLogEnabled
: google.protobuf.BoolValue
Whether metric_log system table is enabled.
metricLogRetentionSize
: google.protobuf.Int64Value
The maximum size that metric_log can grow to before old data will be removed. If set to 0, automatic removal of metric_log data based on size is disabled.
metricLogRetentionTime
: google.protobuf.Int64Value
The maximum time that metric_log records will be retained before removal. If set to 0, automatic removal of metric_log data based on time is disabled.
traceLogEnabled
: google.protobuf.BoolValue
Whether trace_log system table is enabled.
traceLogRetentionSize
: google.protobuf.Int64Value
The maximum size that trace_log can grow to before old data will be removed. If set to 0, automatic removal of trace_log data based on size is disabled.
traceLogRetentionTime
: google.protobuf.Int64Value
The maximum time that trace_log records will be retained before removal. If set to 0, automatic removal of trace_log data based on time is disabled.
textLogEnabled
: google.protobuf.BoolValue
Whether text_log system table is enabled.
textLogRetentionSize
: google.protobuf.Int64Value
The maximum size that text_log can grow to before old data will be removed. If set to 0, automatic removal of text_log data based on size is disabled.
textLogRetentionTime
: google.protobuf.Int64Value
The maximum time that text_log records will be retained before removal. If set to 0, automatic removal of text_log data based on time is disabled.
textLogLevel
: LogLevel
Logging level for text_log system table. Possible values: TRACE, DEBUG, INFORMATION, WARNING, ERROR.
opentelemetrySpanLogEnabled
: google.protobuf.BoolValue
Enable or disable opentelemetry_span_log system table. Default value: false.
opentelemetrySpanLogRetentionSize
: google.protobuf.Int64Value
The maximum size that opentelemetry_span_log can grow to before old data will be removed. If set to 0 (default), automatic removal of opentelemetry_span_log data based on size is disabled.
opentelemetrySpanLogRetentionTime
: google.protobuf.Int64Value
The maximum time that opentelemetry_span_log records will be retained before removal. If set to 0, automatic removal of opentelemetry_span_log data based on time is disabled.
queryViewsLogEnabled
: google.protobuf.BoolValue
Enable or disable query_views_log system table. Default value: false.
queryViewsLogRetentionSize
: google.protobuf.Int64Value
The maximum size that query_views_log can grow to before old data will be removed. If set to 0 (default), automatic removal of query_views_log data based on size is disabled.
queryViewsLogRetentionTime
: google.protobuf.Int64Value
The maximum time that query_views_log records will be retained before removal. If set to 0, automatic removal of query_views_log data based on time is disabled.
asynchronousMetricLogEnabled
: google.protobuf.BoolValue
Enable or disable asynchronous_metric_log system table. Default value: false.
asynchronousMetricLogRetentionSize
: google.protobuf.Int64Value
The maximum size that asynchronous_metric_log can grow to before old data will be removed. If set to 0 (default), automatic removal of asynchronous_metric_log data based on size is disabled.
asynchronousMetricLogRetentionTime
: google.protobuf.Int64Value
The maximum time that asynchronous_metric_log records will be retained before removal. If set to 0, automatic removal of asynchronous_metric_log data based on time is disabled.
sessionLogEnabled
: google.protobuf.BoolValue
Enable or disable session_log system table. Default value: false.
sessionLogRetentionSize
: google.protobuf.Int64Value
The maximum size that session_log can grow to before old data will be removed. If set to 0 (default), automatic removal of session_log data based on size is disabled.
sessionLogRetentionTime
: google.protobuf.Int64Value
The maximum time that session_log records will be retained before removal. If set to 0, automatic removal of session_log data based on time is disabled.
zookeeperLogEnabled
: google.protobuf.BoolValue
Enable or disable zookeeper_log system table. Default value: false.
zookeeperLogRetentionSize
: google.protobuf.Int64Value
The maximum size that zookeeper_log can grow to before old data will be removed. If set to 0 (default), automatic removal of zookeeper_log data based on size is disabled.
zookeeperLogRetentionTime
: google.protobuf.Int64Value
The maximum time that zookeeper_log records will be retained before removal. If set to 0, automatic removal of zookeeper_log data based on time is disabled.
asynchronousInsertLogEnabled
: google.protobuf.BoolValue
Enable or disable asynchronous_insert_log system table. Default value: false. Minimal required ClickHouse version: 22.10.
asynchronousInsertLogRetentionSize
: google.protobuf.Int64Value
The maximum size that asynchronous_insert_log can grow to before old data will be removed. If set to 0 (default), automatic removal of asynchronous_insert_log data based on size is disabled.
asynchronousInsertLogRetentionTime
: google.protobuf.Int64Value
The maximum time that asynchronous_insert_log records will be retained before removal. If set to 0, automatic removal of asynchronous_insert_log data based on time is disabled.
backgroundPoolSize
: google.protobuf.Int64Value
backgroundMergesMutationsConcurrencyRatio
: google.protobuf.Int64Value
Sets a ratio between the number of threads and the number of background merges and mutations that can be executed concurrently. For example, if the ratio equals to 2 and background_pool_size is set to 16 then ClickHouse can execute 32 background merges concurrently. This is possible, because background operations could be suspended and postponed. This is needed to give small merges more execution priority. You can only increase this ratio at runtime. To lower it you have to restart the server. The same as for background_pool_size setting background_merges_mutations_concurrency_ratio could be applied from the default profile for backward compatibility. Default: 2 See in-depth description in ClickHouse documentation
backgroundSchedulePoolSize
: google.protobuf.Int64Value
backgroundFetchesPoolSize
: google.protobuf.Int64Value
Sets the number of threads performing background fetches for tables with ReplicatedMergeTree engines. Default value: 8.
More info see in ClickHouse documentation.
backgroundMovePoolSize
: google.protobuf.Int64Value
backgroundDistributedSchedulePoolSize
: google.protobuf.Int64Value
backgroundBufferFlushSchedulePoolSize
: google.protobuf.Int64Value
backgroundMessageBrokerSchedulePoolSize
: google.protobuf.Int64Value
backgroundCommonPoolSize
: google.protobuf.Int64Value
The maximum number of threads that will be used for performing a variety of operations (mostly garbage collection) for *MergeTree-engine tables in a background. Default: 8 See in-depth description in ClickHouse documentation
defaultDatabase
: google.protobuf.StringValue
The default database.
To get a list of cluster databases, see Yandex Managed ClickHouse documentation.
totalMemoryProfilerStep
: google.protobuf.Int64Value
Sets the memory size (in bytes) for a stack trace at every peak allocation step. Default value: 4194304.
More info see in ClickHouse documentation.
totalMemoryTrackerSampleProbability
: google.protobuf.DoubleValue
queryMaskingRules
: QueryMaskingRule
Regexp-based rules, which will be applied to queries as well as all log messages before storing them in server logs, system.query_log, system.text_log, system.processes tables, and in logs sent to the client. That allows preventing sensitive data leakage from SQL queries (like names, emails, personal identifiers or credit card numbers) to logs. Change of these settings is applied with ClickHouse restart See in-depth description in ClickHouse documentation
dictionariesLazyLoad
: google.protobuf.BoolValue
Lazy loading of dictionaries. Default: true See in-depth description in ClickHouse documentation
queryCache
: QueryCache
Query cache configuration. Min version: 23.5 See in-depth description in ClickHouse documentation
Kafka
SecurityProtocol
SECURITY_PROTOCOL_UNSPECIFIED
SECURITY_PROTOCOL_PLAINTEXT
SECURITY_PROTOCOL_SSL
SECURITY_PROTOCOL_SASL_PLAINTEXT
SECURITY_PROTOCOL_SASL_SSL
SaslMechanism
SASL_MECHANISM_UNSPECIFIED
SASL_MECHANISM_GSSAPI
SASL_MECHANISM_PLAIN
SASL_MECHANISM_SCRAM_SHA_256
SASL_MECHANISM_SCRAM_SHA_512
Debug
DEBUG_UNSPECIFIED
DEBUG_GENERIC
DEBUG_BROKER
DEBUG_TOPIC
DEBUG_METADATA
DEBUG_FEATURE
DEBUG_QUEUE
DEBUG_MSG
DEBUG_PROTOCOL
DEBUG_CGRP
DEBUG_SECURITY
DEBUG_FETCH
DEBUG_INTERCEPTOR
DEBUG_PLUGIN
DEBUG_CONSUMER
DEBUG_ADMIN
DEBUG_EOS
DEBUG_MOCK
DEBUG_ASSIGNOR
DEBUG_CONF
DEBUG_TELEMETRY
DEBUG_ALL
AutoOffsetReset
AUTO_OFFSET_RESET_UNSPECIFIED
AUTO_OFFSET_RESET_SMALLEST
AUTO_OFFSET_RESET_EARLIEST
AUTO_OFFSET_RESET_BEGINNING
AUTO_OFFSET_RESET_LARGEST
AUTO_OFFSET_RESET_LATEST
AUTO_OFFSET_RESET_END
AUTO_OFFSET_RESET_ERROR
securityProtocol
: SecurityProtocol
saslMechanism
: SaslMechanism
saslUsername
: string
saslPassword
: string
enableSslCertificateVerification
: google.protobuf.BoolValue
maxPollIntervalMs
: google.protobuf.Int64Value
sessionTimeoutMs
: google.protobuf.Int64Value
debug
: Debug
autoOffsetReset
: AutoOffsetReset
Header
name
: string
value
: string
Replica
host
: string
MySQL host of the replica.
priority
: int64
The priority of the replica that ClickHouse takes into account when connecting. Replica with the highest priority should have this field set to the lowest number.
port
: int64
Port to use when connecting to the replica. If a port is not specified for a replica, ClickHouse uses the port specified for the source.
user
: string
Name of the MySQL database user.
password
: string
Password of the MySQL database user.
Attribute
name
: string
Name of the column.
type
: string
Type of the column.
nullValue
: string
Default value for an element without data (for example, an empty string).
expression
: string
Expression, describing the attribute, if applicable.
hierarchical
: bool
Indication of hierarchy support.
Default value: false
.
injective
: bool
Indication of injective mapping "id -> attribute".
Default value: false
.
Id
Numeric key.
name
: string
Name of the numeric key.
Key
Complex key.
attributes
: Attribute
Attributes of a complex key.
Structure
Attribute
name
: string
Name of the column.
type
: string
Type of the column.
nullValue
: string
Default value for an element without data (for example, an empty string).
expression
: string
Expression, describing the attribute, if applicable.
hierarchical
: bool
Indication of hierarchy support.
Default value: false
.
injective
: bool
Indication of injective mapping "id -> attribute".
Default value: false
.
Id
Numeric key.
name
: string
Name of the numeric key.
Key
Complex key.
attributes
: Attribute
Attributes of a complex key.
id
: Id
Single numeric key column for the dictionary.
key
: Key
Composite key for the dictionary, containing of one or more key columns. For details, see ClickHouse documentation.
rangeMin
: Attribute
Field holding the beginning of the range for dictionaries with RANGE_HASHED
layout.
For details, see ClickHouse documentation.
rangeMax
: Attribute
Field holding the end of the range for dictionaries with RANGE_HASHED
layout.
For details, see ClickHouse documentation.
attributes
: Attribute
Description of the fields available for database queries. For details, see ClickHouse documentation.
Layout
Layout determining how to store the dictionary in memory.
Type
TYPE_UNSPECIFIED
FLAT
The entire dictionary is stored in memory in the form of flat arrays. Available for all dictionary sources.
HASHED
The entire dictionary is stored in memory in the form of a hash table. Available for all dictionary sources.
COMPLEX_KEY_HASHED
Similar to HASHED, to be used with composite keys. Available for all dictionary sources.
RANGE_HASHED
The entire dictionary is stored in memory in the form of a hash table, with an ordered array of ranges and their corresponding values. Available for all dictionary sources.
CACHE
The dictionary is stored in a cache with a set number of cells. Available for MySQL, ClickHouse and HTTP dictionary sources.
COMPLEX_KEY_CACHE
Similar to CACHE, to be used with composite keys. Available for MySQL, ClickHouse and HTTP dictionary sources.
type
: Type
Layout type for an external dictionary.
sizeInCells
: int64
Number of cells in the cache. Rounded up to a power of two. Applicable only for CACHE and COMPLEX_KEY_CACHE layout types.
maxArraySize
: int64
Maximum dictionary key size. Applicable only for FLAT layout type.
Range
min
: int64
Minimum dictionary lifetime.
max
: int64
Maximum dictionary lifetime.
HttpSource
Header
name
: string
value
: string
url
: string
URL of the source dictionary available over HTTP.
format
: string
The data format. Valid values are all formats supported by ClickHouse SQL dialect.
headers
: Header
HTTP headers.
MysqlSource
Replica
host
: string
MySQL host of the replica.
priority
: int64
The priority of the replica that ClickHouse takes into account when connecting. Replica with the highest priority should have this field set to the lowest number.
port
: int64
Port to use when connecting to the replica. If a port is not specified for a replica, ClickHouse uses the port specified for the source.
user
: string
Name of the MySQL database user.
password
: string
Password of the MySQL database user.
db
: string
Name of the MySQL database to connect to.
table
: string
Name of the database table to use as a ClickHouse dictionary.
port
: int64
Default port to use when connecting to a replica of the dictionary source.
user
: string
Name of the default user for replicas of the dictionary source.
password
: string
Password of the default user for replicas of the dictionary source.
replicas
: Replica
List of MySQL replicas of the database used as dictionary source.
where
: string
Selection criteria for the data in the specified MySQL table.
invalidateQuery
: string
Query for checking the dictionary status, to pull only updated data. For more details, see ClickHouse documentation on dictionaries.
closeConnection
: google.protobuf.BoolValue
Should the connection be closed after each request.
shareConnection
: google.protobuf.BoolValue
Should a connection be shared for some requests.
ClickhouseSource
db
: string
Name of the ClickHouse database.
table
: string
Name of the table in the specified database to be used as the dictionary source.
host
: string
ClickHouse host of the specified database.
port
: int64
Port to use when connecting to the host.
user
: string
Name of the ClickHouse database user.
password
: string
Password of the ClickHouse database user.
where
: string
Selection criteria for the data in the specified ClickHouse table.
secure
: google.protobuf.BoolValue
Use ssl for connection.
MongodbSource
db
: string
Name of the MongoDB database.
collection
: string
Name of the collection in the specified database to be used as the dictionary source.
host
: string
MongoDB host of the specified database.
port
: int64
Port to use when connecting to the host.
user
: string
Name of the MongoDB database user.
password
: string
Password of the MongoDB database user.
options
: string
PostgresqlSource
SslMode
SSL_MODE_UNSPECIFIED
DISABLE
Only try a non-SSL connection.
ALLOW
First try a non-SSL connection; if that fails, try an SSL connection.
PREFER
First try an SSL connection; if that fails, try a non-SSL connection.
VERIFY_CA
Only try an SSL connection, and verify that the server certificate is issued by a trusted certificate authority (CA).
VERIFY_FULL
Only try an SSL connection, verify that the server certificate is issued by a trusted CA and that the requested server host name matches that in the certificate.
db
: string
Name of the PostrgreSQL database.
table
: string
Name of the table in the specified database to be used as the dictionary source.
hosts
: string
Name of the PostrgreSQL host
port
: int64
Port to use when connecting to the host.
user
: string
Name of the PostrgreSQL database user.
password
: string
Password of the PostrgreSQL database user.
invalidateQuery
: string
Query for checking the dictionary status, to pull only updated data. For more details, see ClickHouse documentation on dictionaries.
sslMode
: SslMode
Mode of SSL TCP/IP connection to the PostgreSQL host. For more details, see PostgreSQL documentation.
Retention
age
: int64
Minimum age of the data in seconds.
precision
: int64
Precision of determining the age of the data, in seconds.
Pattern
Retention
age
: int64
Minimum age of the data in seconds.
precision
: int64
Precision of determining the age of the data, in seconds.
regexp
: string
Pattern for metric names.
function
: string
Name of the aggregating function to apply to data of the age specified in retention.
retention
: Retention
Age of data to use for thinning.
MergeTree
Options specific to the MergeTree table engine.
replicatedDeduplicationWindow
: google.protobuf.Int64Value
Number of blocks of hashes to keep in ZooKeeper.
replicatedDeduplicationWindowSeconds
: google.protobuf.Int64Value
Period of time to keep blocks of hashes for.
partsToDelayInsert
: google.protobuf.Int64Value
If table contains at least that many active parts in single partition, artificially slow down insert into table.
partsToThrowInsert
: google.protobuf.Int64Value
If more than this number active parts in single partition, throw 'Too many parts ...' exception.
inactivePartsToDelayInsert
: google.protobuf.Int64Value
inactivePartsToThrowInsert
: google.protobuf.Int64Value
maxReplicatedMergesInQueue
: google.protobuf.Int64Value
How many tasks of merging and mutating parts are allowed simultaneously in ReplicatedMergeTree queue.
numberOfFreeEntriesInPoolToLowerMaxSizeOfMerge
: google.protobuf.Int64Value
If there is less than specified number of free entries in background pool (or replicated queue), start to lower maximum size of merge to process.
maxBytesToMergeAtMinSpaceInPool
: google.protobuf.Int64Value
Maximum in total size of parts to merge, when there are minimum free threads in background pool (or entries in replication queue).
maxBytesToMergeAtMaxSpaceInPool
: google.protobuf.Int64Value
minBytesForWidePart
: google.protobuf.Int64Value
Minimum number of bytes in a data part that can be stored in Wide format.
More info see in ClickHouse documentation.
minRowsForWidePart
: google.protobuf.Int64Value
Minimum number of rows in a data part that can be stored in Wide format.
More info see in ClickHouse documentation.
ttlOnlyDropParts
: google.protobuf.BoolValue
Enables or disables complete dropping of data parts where all rows are expired in MergeTree tables.
More info see in ClickHouse documentation.
allowRemoteFsZeroCopyReplication
: google.protobuf.BoolValue
mergeWithTtlTimeout
: google.protobuf.Int64Value
mergeWithRecompressionTtlTimeout
: google.protobuf.Int64Value
maxPartsInTotal
: google.protobuf.Int64Value
maxNumberOfMergesWithTtlInPool
: google.protobuf.Int64Value
cleanupDelayPeriod
: google.protobuf.Int64Value
numberOfFreeEntriesInPoolToExecuteMutation
: google.protobuf.Int64Value
maxAvgPartSizeForTooManyParts
: google.protobuf.Int64Value
The 'too many parts' check according to 'parts_to_delay_insert' and 'parts_to_throw_insert' will be active only if the average part size (in the relevant partition) is not larger than the specified threshold. If it is larger than the specified threshold, the INSERTs will be neither delayed or rejected. This allows to have hundreds of terabytes in a single table on a single server if the parts are successfully merged to larger parts. This does not affect the thresholds on inactive parts or total parts. Default: 1 GiB Min version: 22.10 See in-depth description in ClickHouse GitHub
minAgeToForceMergeSeconds
: google.protobuf.Int64Value
Merge parts if every part in the range is older than the value of min_age_to_force_merge_seconds. Default: 0 - disabled Min_version: 22.10 See in-depth description in ClickHouse documentation
minAgeToForceMergeOnPartitionOnly
: google.protobuf.BoolValue
Whether min_age_to_force_merge_seconds should be applied only on the entire partition and not on subset. Default: false Min_version: 22.11 See in-depth description in ClickHouse documentation
mergeSelectingSleepMs
: google.protobuf.Int64Value
Sleep time for merge selecting when no part is selected. A lower setting triggers selecting tasks in background_schedule_pool frequently, which results in a large number of requests to ClickHouse Keeper in large-scale clusters. Default: 5000 Min_version: 21.10 See in-depth description in ClickHouse documentation
mergeMaxBlockSize
: google.protobuf.Int64Value
The number of rows that are read from the merged parts into memory. Default: 8192 See in-depth description in ClickHouse documentation
checkSampleColumnIsCorrect
: google.protobuf.BoolValue
Enables the check at table creation, that the data type of a column for sampling or sampling expression is correct. The data type must be one of unsigned integer types: UInt8, UInt16, UInt32, UInt64. Default: true See in-depth description in ClickHouse documentation
maxMergeSelectingSleepMs
: google.protobuf.Int64Value
Maximum sleep time for merge selecting, a lower setting will trigger selecting tasks in background_schedule_pool frequently which result in large amount of requests to zookeeper in large-scale clusters. Default: 60000 Min_version: 23.6 See in-depth description in ClickHouse GitHub
maxCleanupDelayPeriod
: google.protobuf.Int64Value
Maximum period to clean old queue logs, blocks hashes and parts. Default: 300 Min_version: 23.6 See in-depth description in ClickHouse GitHub
Compression
Method
METHOD_UNSPECIFIED
LZ4
ZSTD
method
: Method
Compression method to use for the specified combination of min_part_size and min_part_size_ratio.
minPartSize
: int64
Minimum size of a part of a table.
minPartSizeRatio
: double
Minimum ratio of a part relative to the size of all the data in the table.
level
: google.protobuf.Int64Value
ExternalDictionary
HttpSource
Header
name
: string
value
: string
url
: string
URL of the source dictionary available over HTTP.
format
: string
The data format. Valid values are all formats supported by ClickHouse SQL dialect.
headers
: Header
HTTP headers.
MysqlSource
Replica
host
: string
MySQL host of the replica.
priority
: int64
The priority of the replica that ClickHouse takes into account when connecting. Replica with the highest priority should have this field set to the lowest number.
port
: int64
Port to use when connecting to the replica. If a port is not specified for a replica, ClickHouse uses the port specified for the source.
user
: string
Name of the MySQL database user.
password
: string
Password of the MySQL database user.
db
: string
Name of the MySQL database to connect to.
table
: string
Name of the database table to use as a ClickHouse dictionary.
port
: int64
Default port to use when connecting to a replica of the dictionary source.
user
: string
Name of the default user for replicas of the dictionary source.
password
: string
Password of the default user for replicas of the dictionary source.
replicas
: Replica
List of MySQL replicas of the database used as dictionary source.
where
: string
Selection criteria for the data in the specified MySQL table.
invalidateQuery
: string
Query for checking the dictionary status, to pull only updated data. For more details, see ClickHouse documentation on dictionaries.
closeConnection
: google.protobuf.BoolValue
Should the connection be closed after each request.
shareConnection
: google.protobuf.BoolValue
Should a connection be shared for some requests.
ClickhouseSource
db
: string
Name of the ClickHouse database.
table
: string
Name of the table in the specified database to be used as the dictionary source.
host
: string
ClickHouse host of the specified database.
port
: int64
Port to use when connecting to the host.
user
: string
Name of the ClickHouse database user.
password
: string
Password of the ClickHouse database user.
where
: string
Selection criteria for the data in the specified ClickHouse table.
secure
: google.protobuf.BoolValue
Use ssl for connection.
MongodbSource
db
: string
Name of the MongoDB database.
collection
: string
Name of the collection in the specified database to be used as the dictionary source.
host
: string
MongoDB host of the specified database.
port
: int64
Port to use when connecting to the host.
user
: string
Name of the MongoDB database user.
password
: string
Password of the MongoDB database user.
options
: string
PostgresqlSource
SslMode
SSL_MODE_UNSPECIFIED
DISABLE
Only try a non-SSL connection.
ALLOW
First try a non-SSL connection; if that fails, try an SSL connection.
PREFER
First try an SSL connection; if that fails, try a non-SSL connection.
VERIFY_CA
Only try an SSL connection, and verify that the server certificate is issued by a trusted certificate authority (CA).
VERIFY_FULL
Only try an SSL connection, verify that the server certificate is issued by a trusted CA and that the requested server host name matches that in the certificate.
db
: string
Name of the PostrgreSQL database.
table
: string
Name of the table in the specified database to be used as the dictionary source.
hosts
: string
Name of the PostrgreSQL host
port
: int64
Port to use when connecting to the host.
user
: string
Name of the PostrgreSQL database user.
password
: string
Password of the PostrgreSQL database user.
invalidateQuery
: string
Query for checking the dictionary status, to pull only updated data. For more details, see ClickHouse documentation on dictionaries.
sslMode
: SslMode
Mode of SSL TCP/IP connection to the PostgreSQL host. For more details, see PostgreSQL documentation.
Structure
Attribute
name
: string
Name of the column.
type
: string
Type of the column.
nullValue
: string
Default value for an element without data (for example, an empty string).
expression
: string
Expression, describing the attribute, if applicable.
hierarchical
: bool
Indication of hierarchy support.
Default value: false
.
injective
: bool
Indication of injective mapping "id -> attribute".
Default value: false
.
Id
Numeric key.
name
: string
Name of the numeric key.
Key
Complex key.
attributes
: Attribute
Attributes of a complex key.
id
: Id
Single numeric key column for the dictionary.
key
: Key
Composite key for the dictionary, containing of one or more key columns. For details, see ClickHouse documentation.
rangeMin
: Attribute
Field holding the beginning of the range for dictionaries with RANGE_HASHED
layout.
For details, see ClickHouse documentation.
rangeMax
: Attribute
Field holding the end of the range for dictionaries with RANGE_HASHED
layout.
For details, see ClickHouse documentation.
attributes
: Attribute
Description of the fields available for database queries. For details, see ClickHouse documentation.
Layout
Layout determining how to store the dictionary in memory.
Type
TYPE_UNSPECIFIED
FLAT
The entire dictionary is stored in memory in the form of flat arrays. Available for all dictionary sources.
HASHED
The entire dictionary is stored in memory in the form of a hash table. Available for all dictionary sources.
COMPLEX_KEY_HASHED
Similar to HASHED, to be used with composite keys. Available for all dictionary sources.
RANGE_HASHED
The entire dictionary is stored in memory in the form of a hash table, with an ordered array of ranges and their corresponding values. Available for all dictionary sources.
CACHE
The dictionary is stored in a cache with a set number of cells. Available for MySQL, ClickHouse and HTTP dictionary sources.
COMPLEX_KEY_CACHE
Similar to CACHE, to be used with composite keys. Available for MySQL, ClickHouse and HTTP dictionary sources.
type
: Type
Layout type for an external dictionary.
sizeInCells
: int64
Number of cells in the cache. Rounded up to a power of two. Applicable only for CACHE and COMPLEX_KEY_CACHE layout types.
maxArraySize
: int64
Maximum dictionary key size. Applicable only for FLAT layout type.
Range
min
: int64
Minimum dictionary lifetime.
max
: int64
Maximum dictionary lifetime.
name
: string
Name of the external dictionary.
structure
: Structure
Set of attributes for the external dictionary. For in-depth description, see ClickHouse documentation.
layout
: Layout
Layout for storing the dictionary in memory. For in-depth description, see ClickHouse documentation.
fixedLifetime
: int64Fixed interval between dictionary updates.
lifetimeRange
: RangeRange of intervals between dictionary updates for ClickHouse to choose from.
httpSource
: HttpSourceHTTP source for the dictionary.
mysqlSource
: MysqlSourceMySQL source for the dictionary.
clickhouseSource
: ClickhouseSourceClickHouse source for the dictionary.
mongodbSource
: MongodbSourceMongoDB source for the dictionary.
postgresqlSource
: PostgresqlSourcePostgreSQL source for the dictionary.
GraphiteRollup
Rollup settings for the GraphiteMergeTree table engine.
Pattern
Retention
age
: int64
Minimum age of the data in seconds.
precision
: int64
Precision of determining the age of the data, in seconds.
regexp
: string
Pattern for metric names.
function
: string
Name of the aggregating function to apply to data of the age specified in retention.
retention
: Retention
Age of data to use for thinning.
name
: string
Name for the specified combination of settings for Graphite rollup.
patterns
: Pattern
Pattern to use for the rollup.
pathColumnName
: string
The name of the column storing the metric name (Graphite sensor). Default: Path See in-depth description in ClickHouse documentation
timeColumnName
: string
The name of the column storing the time of measuring the metric. Default: Time See in-depth description in ClickHouse documentation
valueColumnName
: string
The name of the column storing the value of the metric at the time set in time_column_name. Default: Value See in-depth description in ClickHouse documentation
versionColumnName
: string
The name of the column storing the version of the metric. Default: Timestamp See in-depth description in ClickHouse documentation
KafkaTopic
name
: string
settings
: Kafka
Rabbitmq
username
: string
RabbitMQ username
password
: string
RabbitMQ password
vhost
: string
RabbitMQ virtual host
QueryMaskingRule
name
: string
Name for the rule.
regexp
: string
RE2 compatible regular expression. Required.
replace
: string
Substitution string for sensitive data. Default: six asterisks
QueryCache
maxSizeInBytes
: google.protobuf.Int64Value
The maximum cache size in bytes. Default: 1073741824 (1 GiB)
maxEntries
: google.protobuf.Int64Value
The maximum number of SELECT query results stored in the cache. Default: 1024
maxEntrySizeInBytes
: google.protobuf.Int64Value
The maximum size in bytes SELECT query results may have to be saved in the cache. Dafault: 1048576 (1 MiB)
maxEntrySizeInRows
: google.protobuf.Int64Value
The maximum number of rows SELECT query results may have to be saved in the cache. Default: 30000000 (30 mil)