Create
Creates a ClickHouse cluster in the specified folder.
- TypeScript
- Python
import {
cloudApi,
decodeMessage,
serviceClients,
Session,
waitForOperation,
} from "@yandex-cloud/nodejs-sdk";
const Cluster = cloudApi.dataproc.cluster.Cluster;
const CreateClusterRequest =
cloudApi.dataproc.cluster_service.CreateClusterRequest;
const HadoopConfig_Service = cloudApi.dataproc.cluster.HadoopConfig_Service;
const Role = cloudApi.dataproc.subcluster.Role;
(async () => {
const authToken = process.env["YC_OAUTH_TOKEN"];
const session = new Session({ oauthToken: authToken });
const client = session.client(serviceClients.ClusterServiceClient);
const operation = await client.create(
CreateClusterRequest.fromPartial({
folderId: "folderId",
// name: "name",
// description: "description",
// labels: {"key": "labels"},
configSpec: {
// versionId: "versionId",
// hadoop: {
// services: [HadoopConfig_Service.HDFS],
// properties: {"key": "properties"},
// sshPublicKeys: ["sshPublicKeys"],
// initializationActions: [{
// uri: "uri",
// args: ["args"],
// timeout: 0
// }]
// },
// subclustersSpec: [{
// name: "name",
// role: Role.MASTERNODE,
// resources: {
// resourcePresetId: "resourcePresetId",
// diskTypeId: "diskTypeId",
// diskSize: 0
// },
// subnetId: "subnetId",
// hostsCount: 0,
// assignPublicIp: true,
// autoscalingConfig: {
// maxHostsCount: 0,
// preemptible: true,
// measurementDuration: {
// seconds: 0,
// nanos: 0
// },
// warmupDuration: {
// seconds: 0,
// nanos: 0
// },
// stabilizationDuration: {
// seconds: 0,
// nanos: 0
// },
// cpuUtilizationTarget: 0,
// decommissionTimeout: 0
// }
// }]
},
zoneId: "zoneId",
serviceAccountId: "serviceAccountId",
// bucket: "bucket",
// uiProxy: true,
// securityGroupIds: ["securityGroupIds"],
// hostGroupIds: ["hostGroupIds"],
// deletionProtection: true,
// logGroupId: "logGroupId"
})
);
const finishedOp = await waitForOperation(operation, session);
if (finishedOp.response) {
const result = decodeMessage<typeof Cluster>(finishedOp.response);
console.log(result);
}
})();
import os
import grpc
import yandexcloud
from yandex.cloud.dataproc.v1.subcluster_pb2 import AutoscalingConfig
from yandex.cloud.dataproc.v1.cluster_pb2 import Cluster
from yandex.cloud.dataproc.v1.cluster_service_pb2_grpc import ClusterServiceStub
from yandex.cloud.dataproc.v1.cluster_service_pb2 import CreateClusterConfigSpec
from yandex.cloud.dataproc.v1.cluster_service_pb2 import CreateClusterMetadata
from yandex.cloud.dataproc.v1.cluster_service_pb2 import CreateClusterRequest
from yandex.cloud.dataproc.v1.cluster_service_pb2 import CreateSubclusterConfigSpec
from yandex.cloud.dataproc.v1.cluster_pb2 import HadoopConfig
from yandex.cloud.dataproc.v1.cluster_pb2 import InitializationAction
from yandex.cloud.dataproc.v1.common_pb2 import Resources
from yandex.cloud.dataproc.v1.subcluster_pb2 import Role
token = os.getenv("YC_OAUTH_TOKEN")
sdk = yandexcloud.SDK(token=token)
service = sdk.client(ClusterServiceStub)
operation = service.Create(
CreateClusterRequest(
folder_id="folderId",
# name = "name",
# description = "description",
# labels = {"key": "labels"},
config_spec=CreateClusterConfigSpec(
# version_id = "versionId",
# hadoop = HadoopConfig(
# services = [HadoopConfig.Service.HDFS],
# properties = {"key": "properties"},
# ssh_public_keys = ["sshPublicKeys"],
# initialization_actions = [InitializationAction(
# uri = "uri",
# args = ["args"],
# timeout = 0
# )]
# ),
# subclusters_spec = [CreateSubclusterConfigSpec(
# name = "name",
# role = Role.MASTERNODE,
# resources = Resources(
# resource_preset_id = "resourcePresetId",
# disk_type_id = "diskTypeId",
# disk_size = 0
# ),
# subnet_id = "subnetId",
# hosts_count = 0,
# assign_public_ip = true,
# autoscaling_config = AutoscalingConfig(
# max_hosts_count = 0,
# preemptible = true,
# measurement_duration = Duration(
# seconds = 0,
# nanos = 0
# ),
# warmup_duration = Duration(
# seconds = 0,
# nanos = 0
# ),
# stabilization_duration = Duration(
# seconds = 0,
# nanos = 0
# ),
# cpu_utilization_target = 0,
# decommission_timeout = 0
# )
# )]
),
zone_id="zoneId",
service_account_id="serviceAccountId",
# bucket = "bucket",
# ui_proxy = true,
# security_group_ids = ["securityGroupIds"],
# host_group_ids = ["hostGroupIds"],
# deletion_protection = true,
# log_group_id = "logGroupId"
)
)
operation_result = sdk.wait_operation_and_get_result(
operation,
response_type=Cluster,
meta_type=CreateClusterMetadata,
)
print(operation_result)
CreateClusterRequest
folderId
: string
ID of the folder to create the ClickHouse cluster in.
name
: string
Name of the ClickHouse cluster. The name must be unique within the folder.
description
: string
Description of the ClickHouse cluster.
labels
: string
Custom labels for the ClickHouse cluster as key:value
pairs. Maximum 64 per resource.
For example, "project": "mvp" or "source": "dictionary".
environment
: Cluster.Environment
Deployment environment of the ClickHouse cluster.
configSpec
: ConfigSpec
Configuration and resources for hosts that should be created for the ClickHouse cluster.
databaseSpecs
: DatabaseSpec
Descriptions of databases to be created in the ClickHouse cluster.
userSpecs
: UserSpec
Descriptions of database users to be created in the ClickHouse cluster.
hostSpecs
: HostSpec
Individual configurations for hosts that should be created for the ClickHouse cluster.
networkId
: string
ID of the network to create the cluster in.
shardName
: string
Name of the first shard in cluster. If not set, defaults to the value 'shard1'.
serviceAccountId
: string
ID of the service account used for access to Object Storage.
securityGroupIds
: string
User security groups
deletionProtection
: bool
Deletion Protection inhibits deletion of the cluster
maintenanceWindow
: MaintenanceWindow
Window of maintenance operations.
ConfigSpec
Clickhouse
config
: config.ClickhouseConfig
Configuration for a ClickHouse server.
resources
: Resources
Resources allocated to ClickHouse hosts.
Zookeeper
resources
: Resources
Resources allocated to ZooKeeper hosts. If not set, minimal available resources will be used. All available resource presets can be retrieved with a ResourcePresetService.List request.
version
: string
Version of the ClickHouse server software.
clickhouse
: Clickhouse
Configuration and resources for a ClickHouse server.
zookeeper
: Zookeeper
Configuration and resources for a ZooKeeper server.
backupWindowStart
: google.type.TimeOfDay
Time to start the daily backup, in the UTC timezone.
access
: Access
Access policy for external services.
If you want a specific service to access the ClickHouse cluster, then set the necessary values in this policy.
cloudStorage
: CloudStorage
sqlDatabaseManagement
: google.protobuf.BoolValue
Whether database management through SQL commands is enabled.
sqlUserManagement
: google.protobuf.BoolValue
Whether user management through SQL commands is enabled.
adminPassword
: string
Password for user 'admin' that has SQL user management access.
embeddedKeeper
: google.protobuf.BoolValue
Whether cluster should use embedded Keeper instead of Zookeeper
backupRetainPeriodDays
: google.protobuf.Int64Value
Retain period of automatically created backup in days
DatabaseSpec
name
: string
Name of the ClickHouse database. 1-63 characters long.
UserSpec
name
: string
Name of the ClickHouse user.
password
: string
Password of the ClickHouse user.
permissions
: Permission
Set of permissions to grant to the user. If not set, it's granted permissions to access all databases.
settings
: UserSettings
quotas
: UserQuota
Set of quotas assigned to the user.
HostSpec
zoneId
: string
ID of the availability zone where the host resides. To get a list of available zones, use the yandex.cloud.compute.v1.ZoneService.List request.
type
: Host.Type
Type of the host to be deployed.
subnetId
: string
ID of the subnet that the host should belong to. This subnet should be a part of the network that the cluster belongs to. The ID of the network is set in the Cluster.network_id field.
assignPublicIp
: bool
Whether the host should get a public IP address on creation.
After a host has been created, this setting cannot be changed. To remove an assigned public IP, or to assign a public IP to a host without one, recreate the host with assign_public_ip set as needed.
Possible values:
- false - don't assign a public IP to the host.
- true - the host should have a public IP address.
shardName
: string
Name of the shard that the host is assigned to.
MaintenanceWindow
A maintenance window settings.
One of policy
The maintenance policy in effect.
anytime
: AnytimeMaintenanceWindowMaintenance operation can be scheduled anytime.
weeklyMaintenanceWindow
: WeeklyMaintenanceWindowMaintenance operation can be scheduled on a weekly basis.
ClickhouseConfig
ClickHouse configuration options. Detailed description for each set of options is available in ClickHouse documentation.
Any options not listed here are not supported.
LogLevel
LOG_LEVEL_UNSPECIFIED
TRACE
DEBUG
INFORMATION
WARNING
ERROR
MergeTree
Options specific to the MergeTree table engine.
replicatedDeduplicationWindow
: google.protobuf.Int64Value
Number of blocks of hashes to keep in ZooKeeper.
replicatedDeduplicationWindowSeconds
: google.protobuf.Int64Value
Period of time to keep blocks of hashes for.
partsToDelayInsert
: google.protobuf.Int64Value
If table contains at least that many active parts in single partition, artificially slow down insert into table.
partsToThrowInsert
: google.protobuf.Int64Value
If more than this number active parts in single partition, throw 'Too many parts ...' exception.
inactivePartsToDelayInsert
: google.protobuf.Int64Value
inactivePartsToThrowInsert
: google.protobuf.Int64Value
maxReplicatedMergesInQueue
: google.protobuf.Int64Value
How many tasks of merging and mutating parts are allowed simultaneously in ReplicatedMergeTree queue.
numberOfFreeEntriesInPoolToLowerMaxSizeOfMerge
: google.protobuf.Int64Value
If there is less than specified number of free entries in background pool (or replicated queue), start to lower maximum size of merge to process.
maxBytesToMergeAtMinSpaceInPool
: google.protobuf.Int64Value
Maximum in total size of parts to merge, when there are minimum free threads in background pool (or entries in replication queue).
maxBytesToMergeAtMaxSpaceInPool
: google.protobuf.Int64Value
minBytesForWidePart
: google.protobuf.Int64Value
Minimum number of bytes in a data part that can be stored in Wide format.
More info see in ClickHouse documentation.
minRowsForWidePart
: google.protobuf.Int64Value
Minimum number of rows in a data part that can be stored in Wide format.
More info see in ClickHouse documentation.
ttlOnlyDropParts
: google.protobuf.BoolValue
Enables or disables complete dropping of data parts where all rows are expired in MergeTree tables.
More info see in ClickHouse documentation.
allowRemoteFsZeroCopyReplication
: google.protobuf.BoolValue
mergeWithTtlTimeout
: google.protobuf.Int64Value
mergeWithRecompressionTtlTimeout
: google.protobuf.Int64Value
maxPartsInTotal
: google.protobuf.Int64Value
maxNumberOfMergesWithTtlInPool
: google.protobuf.Int64Value
cleanupDelayPeriod
: google.protobuf.Int64Value
numberOfFreeEntriesInPoolToExecuteMutation
: google.protobuf.Int64Value
maxAvgPartSizeForTooManyParts
: google.protobuf.Int64Value
The 'too many parts' check according to 'parts_to_delay_insert' and 'parts_to_throw_insert' will be active only if the average part size (in the relevant partition) is not larger than the specified threshold. If it is larger than the specified threshold, the INSERTs will be neither delayed or rejected. This allows to have hundreds of terabytes in a single table on a single server if the parts are successfully merged to larger parts. This does not affect the thresholds on inactive parts or total parts. Default: 1 GiB Min version: 22.10 See in-depth description in ClickHouse GitHub
minAgeToForceMergeSeconds
: google.protobuf.Int64Value
Merge parts if every part in the range is older than the value of min_age_to_force_merge_seconds. Default: 0 - disabled Min_version: 22.10 See in-depth description in ClickHouse documentation
minAgeToForceMergeOnPartitionOnly
: google.protobuf.BoolValue
Whether min_age_to_force_merge_seconds should be applied only on the entire partition and not on subset. Default: false Min_version: 22.11 See in-depth description in ClickHouse documentation
mergeSelectingSleepMs
: google.protobuf.Int64Value
Sleep time for merge selecting when no part is selected. A lower setting triggers selecting tasks in background_schedule_pool frequently, which results in a large number of requests to ClickHouse Keeper in large-scale clusters. Default: 5000 Min_version: 21.10 See in-depth description in ClickHouse documentation
mergeMaxBlockSize
: google.protobuf.Int64Value
The number of rows that are read from the merged parts into memory. Default: 8192 See in-depth description in ClickHouse documentation
checkSampleColumnIsCorrect
: google.protobuf.BoolValue
Enables the check at table creation, that the data type of a column for sampling or sampling expression is correct. The data type must be one of unsigned integer types: UInt8, UInt16, UInt32, UInt64. Default: true See in-depth description in ClickHouse documentation
maxMergeSelectingSleepMs
: google.protobuf.Int64Value
Maximum sleep time for merge selecting, a lower setting will trigger selecting tasks in background_schedule_pool frequently which result in large amount of requests to zookeeper in large-scale clusters. Default: 60000 Min_version: 23.6 See in-depth description in ClickHouse GitHub
maxCleanupDelayPeriod
: google.protobuf.Int64Value
Maximum period to clean old queue logs, blocks hashes and parts. Default: 300 Min_version: 23.6 See in-depth description in ClickHouse GitHub
Kafka
SecurityProtocol
SECURITY_PROTOCOL_UNSPECIFIED
SECURITY_PROTOCOL_PLAINTEXT
SECURITY_PROTOCOL_SSL
SECURITY_PROTOCOL_SASL_PLAINTEXT
SECURITY_PROTOCOL_SASL_SSL
SaslMechanism
SASL_MECHANISM_UNSPECIFIED
SASL_MECHANISM_GSSAPI
SASL_MECHANISM_PLAIN
SASL_MECHANISM_SCRAM_SHA_256
SASL_MECHANISM_SCRAM_SHA_512
Debug
DEBUG_UNSPECIFIED
DEBUG_GENERIC
DEBUG_BROKER
DEBUG_TOPIC
DEBUG_METADATA
DEBUG_FEATURE
DEBUG_QUEUE
DEBUG_MSG
DEBUG_PROTOCOL
DEBUG_CGRP
DEBUG_SECURITY
DEBUG_FETCH
DEBUG_INTERCEPTOR
DEBUG_PLUGIN
DEBUG_CONSUMER
DEBUG_ADMIN
DEBUG_EOS
DEBUG_MOCK
DEBUG_ASSIGNOR
DEBUG_CONF
DEBUG_TELEMETRY
DEBUG_ALL
AutoOffsetReset
AUTO_OFFSET_RESET_UNSPECIFIED
AUTO_OFFSET_RESET_SMALLEST
AUTO_OFFSET_RESET_EARLIEST
AUTO_OFFSET_RESET_BEGINNING
AUTO_OFFSET_RESET_LARGEST
AUTO_OFFSET_RESET_LATEST
AUTO_OFFSET_RESET_END
AUTO_OFFSET_RESET_ERROR
securityProtocol
: SecurityProtocol
saslMechanism
: SaslMechanism
saslUsername
: string
saslPassword
: string
enableSslCertificateVerification
: google.protobuf.BoolValue
maxPollIntervalMs
: google.protobuf.Int64Value
sessionTimeoutMs
: google.protobuf.Int64Value
debug
: Debug
autoOffsetReset
: AutoOffsetReset
KafkaTopic
name
: string
settings
: Kafka
Rabbitmq
username
: string
RabbitMQ username
password
: string
RabbitMQ password
vhost
: string
RabbitMQ virtual host
Compression
Method
METHOD_UNSPECIFIED
LZ4
ZSTD
method
: Method
Compression method to use for the specified combination of min_part_size and min_part_size_ratio.
minPartSize
: int64
Minimum size of a part of a table.
minPartSizeRatio
: double
Minimum ratio of a part relative to the size of all the data in the table.
level
: google.protobuf.Int64Value
ExternalDictionary
HttpSource
####### Header {#httpsource_header}
name
: string
value
: string
url
: string
URL of the source dictionary available over HTTP.
format
: string
The data format. Valid values are all formats supported by ClickHouse SQL dialect.
headers
: Header
HTTP headers.
MysqlSource
####### Replica {#mysqlsource_replica}
host
: string
MySQL host of the replica.
priority
: int64
The priority of the replica that ClickHouse takes into account when connecting. Replica with the highest priority should have this field set to the lowest number.
port
: int64
Port to use when connecting to the replica. If a port is not specified for a replica, ClickHouse uses the port specified for the source.
user
: string
Name of the MySQL database user.
password
: string
Password of the MySQL database user.
db
: string
Name of the MySQL database to connect to.
table
: string
Name of the database table to use as a ClickHouse dictionary.
port
: int64
Default port to use when connecting to a replica of the dictionary source.
user
: string
Name of the default user for replicas of the dictionary source.
password
: string
Password of the default user for replicas of the dictionary source.
replicas
: Replica
List of MySQL replicas of the database used as dictionary source.
where
: string
Selection criteria for the data in the specified MySQL table.
invalidateQuery
: string
Query for checking the dictionary status, to pull only updated data. For more details, see ClickHouse documentation on dictionaries.
closeConnection
: google.protobuf.BoolValue
Should the connection be closed after each request.
shareConnection
: google.protobuf.BoolValue
Should a connection be shared for some requests.
ClickhouseSource
db
: string
Name of the ClickHouse database.
table
: string
Name of the table in the specified database to be used as the dictionary source.
host
: string
ClickHouse host of the specified database.
port
: int64
Port to use when connecting to the host.
user
: string
Name of the ClickHouse database user.
password
: string
Password of the ClickHouse database user.
where
: string
Selection criteria for the data in the specified ClickHouse table.
secure
: google.protobuf.BoolValue
Use ssl for connection.
MongodbSource
db
: string
Name of the MongoDB database.
collection
: string
Name of the collection in the specified database to be used as the dictionary source.
host
: string
MongoDB host of the specified database.
port
: int64
Port to use when connecting to the host.
user
: string
Name of the MongoDB database user.
password
: string
Password of the MongoDB database user.
options
: string
PostgresqlSource
####### SslMode {#postgresqlsource_sslmode}
SSL_MODE_UNSPECIFIED
DISABLE
Only try a non-SSL connection.
ALLOW
First try a non-SSL connection; if that fails, try an SSL connection.
PREFER
First try an SSL connection; if that fails, try a non-SSL connection.
VERIFY_CA
Only try an SSL connection, and verify that the server certificate is issued by a trusted certificate authority (CA).
VERIFY_FULL
Only try an SSL connection, verify that the server certificate is issued by a trusted CA and that the requested server host name matches that in the certificate.
db
: string
Name of the PostrgreSQL database.
table
: string
Name of the table in the specified database to be used as the dictionary source.
hosts
: string
Name of the PostrgreSQL host
port
: int64
Port to use when connecting to the host.
user
: string
Name of the PostrgreSQL database user.
password
: string
Password of the PostrgreSQL database user.
invalidateQuery
: string
Query for checking the dictionary status, to pull only updated data. For more details, see ClickHouse documentation on dictionaries.
sslMode
: SslMode
Mode of SSL TCP/IP connection to the PostgreSQL host. For more details, see PostgreSQL documentation.
Structure
####### Attribute {#structure_attribute}
name
: string
Name of the column.
type
: string
Type of the column.
nullValue
: string
Default value for an element without data (for example, an empty string).
expression
: string
Expression, describing the attribute, if applicable.
hierarchical
: bool
Indication of hierarchy support.
Default value: false
.
injective
: bool
Indication of injective mapping "id -> attribute".
Default value: false
.
####### Id {#structure_id}
Numeric key.
name
: string
Name of the numeric key.
####### Key {#structure_key}
Complex key.
attributes
: Attribute
Attributes of a complex key.
id
: Id
Single numeric key column for the dictionary.
key
: Key
Composite key for the dictionary, containing of one or more key columns. For details, see ClickHouse documentation.
rangeMin
: Attribute
Field holding the beginning of the range for dictionaries with RANGE_HASHED
layout.
For details, see ClickHouse documentation.
rangeMax
: Attribute
Field holding the end of the range for dictionaries with RANGE_HASHED
layout.
For details, see ClickHouse documentation.
attributes
: Attribute
Description of the fields available for database queries. For details, see ClickHouse documentation.
Layout
Layout determining how to store the dictionary in memory.
####### Type {#layout_type}
TYPE_UNSPECIFIED
FLAT
The entire dictionary is stored in memory in the form of flat arrays. Available for all dictionary sources.
HASHED
The entire dictionary is stored in memory in the form of a hash table. Available for all dictionary sources.
COMPLEX_KEY_HASHED
Similar to HASHED, to be used with composite keys. Available for all dictionary sources.
RANGE_HASHED
The entire dictionary is stored in memory in the form of a hash table, with an ordered array of ranges and their corresponding values. Available for all dictionary sources.
CACHE
The dictionary is stored in a cache with a set number of cells. Available for MySQL, ClickHouse and HTTP dictionary sources.
COMPLEX_KEY_CACHE
Similar to CACHE, to be used with composite keys. Available for MySQL, ClickHouse and HTTP dictionary sources.
type
: Type
Layout type for an external dictionary.
sizeInCells
: int64
Number of cells in the cache. Rounded up to a power of two. Applicable only for CACHE and COMPLEX_KEY_CACHE layout types.
maxArraySize
: int64
Maximum dictionary key size. Applicable only for FLAT layout type.
Range
min
: int64
Minimum dictionary lifetime.
max
: int64
Maximum dictionary lifetime.
name
: string
Name of the external dictionary.
structure
: Structure
Set of attributes for the external dictionary. For in-depth description, see ClickHouse documentation.
layout
: Layout
Layout for storing the dictionary in memory. For in-depth description, see ClickHouse documentation.
One of lifetime
Setting for the period of time between dictionary updates. For details, see ClickHouse documentation.
fixedLifetime
: int64Fixed interval between dictionary updates.
lifetimeRange
: RangeRange of intervals between dictionary updates for ClickHouse to choose from.
One of source
Description of the source for the external dictionary.
httpSource
: HttpSourceHTTP source for the dictionary.
mysqlSource
: MysqlSourceMySQL source for the dictionary.
clickhouseSource
: ClickhouseSourceClickHouse source for the dictionary.
mongodbSource
: MongodbSourceMongoDB source for the dictionary.
postgresqlSource
: PostgresqlSourcePostgreSQL source for the dictionary.
GraphiteRollup
Rollup settings for the GraphiteMergeTree table engine.
Pattern
####### Retention {#pattern_retention}
age
: int64
Minimum age of the data in seconds.
precision
: int64
Precision of determining the age of the data, in seconds.
regexp
: string
Pattern for metric names.
function
: string
Name of the aggregating function to apply to data of the age specified in retention.
retention
: Retention
Age of data to use for thinning.
name
: string
Name for the specified combination of settings for Graphite rollup.
patterns
: Pattern
Pattern to use for the rollup.
pathColumnName
: string
The name of the column storing the metric name (Graphite sensor). Default: Path See in-depth description in ClickHouse documentation
timeColumnName
: string
The name of the column storing the time of measuring the metric. Default: Time See in-depth description in ClickHouse documentation
valueColumnName
: string
The name of the column storing the value of the metric at the time set in time_column_name. Default: Value See in-depth description in ClickHouse documentation
versionColumnName
: string
The name of the column storing the version of the metric. Default: Timestamp See in-depth description in ClickHouse documentation
QueryMaskingRule
name
: string
Name for the rule.
regexp
: string
RE2 compatible regular expression. Required.
replace
: string
Substitution string for sensitive data. Default: six asterisks
QueryCache
maxSizeInBytes
: google.protobuf.Int64Value
The maximum cache size in bytes. Default: 1073741824 (1 GiB)
maxEntries
: google.protobuf.Int64Value
The maximum number of SELECT query results stored in the cache. Default: 1024
maxEntrySizeInBytes
: google.protobuf.Int64Value
The maximum size in bytes SELECT query results may have to be saved in the cache. Dafault: 1048576 (1 MiB)
maxEntrySizeInRows
: google.protobuf.Int64Value
The maximum number of rows SELECT query results may have to be saved in the cache. Default: 30000000 (30 mil)
logLevel
: LogLevel
Logging level for the ClickHouse cluster. Possible values: TRACE, DEBUG, INFORMATION, WARNING, ERROR.
mergeTree
: MergeTree
Settings for the MergeTree engine. See description in ClickHouse documentation.
compression
: Compression
Compression settings for the ClickHouse cluster. See in-depth description in ClickHouse documentation.
dictionaries
: ExternalDictionary
Configuration of external dictionaries to be used by the ClickHouse cluster. See in-depth description in ClickHouse documentation.
graphiteRollup
: GraphiteRollup
Settings for thinning Graphite data. See in-depth description in ClickHouse documentation.
kafka
: Kafka
kafkaTopics
: KafkaTopic
rabbitmq
: Rabbitmq
maxConnections
: google.protobuf.Int64Value
Maximum number of inbound connections.
maxConcurrentQueries
: google.protobuf.Int64Value
Maximum number of simultaneously processed requests.
keepAliveTimeout
: google.protobuf.Int64Value
Number of milliseconds that ClickHouse waits for incoming requests before closing the connection.
uncompressedCacheSize
: google.protobuf.Int64Value
Cache size (in bytes) for uncompressed data used by MergeTree tables.
markCacheSize
: google.protobuf.Int64Value
Approximate size (in bytes) of the cache of "marks" used by MergeTree tables.
maxTableSizeToDrop
: google.protobuf.Int64Value
Maximum size of the table that can be deleted using a DROP query.
maxPartitionSizeToDrop
: google.protobuf.Int64Value
Maximum size of the partition that can be deleted using a DROP query.
builtinDictionariesReloadInterval
: google.protobuf.Int64Value
The setting is deprecated and has no effect.
timezone
: string
The server's time zone to be used in DateTime fields conversions. Specified as an IANA identifier.
geobaseEnabled
: google.protobuf.BoolValue
Enable or disable geobase.
geobaseUri
: string
Address of the archive with the user geobase in Object Storage.
queryLogRetentionSize
: google.protobuf.Int64Value
The maximum size that query_log can grow to before old data will be removed. If set to 0, automatic removal of query_log data based on size is disabled.
queryLogRetentionTime
: google.protobuf.Int64Value
The maximum time that query_log records will be retained before removal. If set to 0, automatic removal of query_log data based on time is disabled.
queryThreadLogEnabled
: google.protobuf.BoolValue
Whether query_thread_log system table is enabled.
queryThreadLogRetentionSize
: google.protobuf.Int64Value
The maximum size that query_thread_log can grow to before old data will be removed. If set to 0, automatic removal of query_thread_log data based on size is disabled.
queryThreadLogRetentionTime
: google.protobuf.Int64Value
The maximum time that query_thread_log records will be retained before removal. If set to 0, automatic removal of query_thread_log data based on time is disabled.
partLogRetentionSize
: google.protobuf.Int64Value
The maximum size that part_log can grow to before old data will be removed. If set to 0, automatic removal of part_log data based on size is disabled.
partLogRetentionTime
: google.protobuf.Int64Value
The maximum time that part_log records will be retained before removal. If set to 0, automatic removal of part_log data based on time is disabled.
metricLogEnabled
: google.protobuf.BoolValue
Whether metric_log system table is enabled.
metricLogRetentionSize
: google.protobuf.Int64Value
The maximum size that metric_log can grow to before old data will be removed. If set to 0, automatic removal of metric_log data based on size is disabled.
metricLogRetentionTime
: google.protobuf.Int64Value
The maximum time that metric_log records will be retained before removal. If set to 0, automatic removal of metric_log data based on time is disabled.
traceLogEnabled
: google.protobuf.BoolValue
Whether trace_log system table is enabled.
traceLogRetentionSize
: google.protobuf.Int64Value
The maximum size that trace_log can grow to before old data will be removed. If set to 0, automatic removal of trace_log data based on size is disabled.
traceLogRetentionTime
: google.protobuf.Int64Value
The maximum time that trace_log records will be retained before removal. If set to 0, automatic removal of trace_log data based on time is disabled.
textLogEnabled
: google.protobuf.BoolValue
Whether text_log system table is enabled.
textLogRetentionSize
: google.protobuf.Int64Value
The maximum size that text_log can grow to before old data will be removed. If set to 0, automatic removal of text_log data based on size is disabled.
textLogRetentionTime
: google.protobuf.Int64Value
The maximum time that text_log records will be retained before removal. If set to 0, automatic removal of text_log data based on time is disabled.
textLogLevel
: LogLevel
Logging level for text_log system table. Possible values: TRACE, DEBUG, INFORMATION, WARNING, ERROR.
opentelemetrySpanLogEnabled
: google.protobuf.BoolValue
Enable or disable opentelemetry_span_log system table. Default value: false.
opentelemetrySpanLogRetentionSize
: google.protobuf.Int64Value
The maximum size that opentelemetry_span_log can grow to before old data will be removed. If set to 0 (default), automatic removal of opentelemetry_span_log data based on size is disabled.
opentelemetrySpanLogRetentionTime
: google.protobuf.Int64Value
The maximum time that opentelemetry_span_log records will be retained before removal. If set to 0, automatic removal of opentelemetry_span_log data based on time is disabled.
queryViewsLogEnabled
: google.protobuf.BoolValue
Enable or disable query_views_log system table. Default value: false.
queryViewsLogRetentionSize
: google.protobuf.Int64Value
The maximum size that query_views_log can grow to before old data will be removed. If set to 0 (default), automatic removal of query_views_log data based on size is disabled.
queryViewsLogRetentionTime
: google.protobuf.Int64Value
The maximum time that query_views_log records will be retained before removal. If set to 0, automatic removal of query_views_log data based on time is disabled.
asynchronousMetricLogEnabled
: google.protobuf.BoolValue
Enable or disable asynchronous_metric_log system table. Default value: false.
asynchronousMetricLogRetentionSize
: google.protobuf.Int64Value
The maximum size that asynchronous_metric_log can grow to before old data will be removed. If set to 0 (default), automatic removal of asynchronous_metric_log data based on size is disabled.
asynchronousMetricLogRetentionTime
: google.protobuf.Int64Value
The maximum time that asynchronous_metric_log records will be retained before removal. If set to 0, automatic removal of asynchronous_metric_log data based on time is disabled.
sessionLogEnabled
: google.protobuf.BoolValue
Enable or disable session_log system table. Default value: false.
sessionLogRetentionSize
: google.protobuf.Int64Value
The maximum size that session_log can grow to before old data will be removed. If set to 0 (default), automatic removal of session_log data based on size is disabled.
sessionLogRetentionTime
: google.protobuf.Int64Value
The maximum time that session_log records will be retained before removal. If set to 0, automatic removal of session_log data based on time is disabled.
zookeeperLogEnabled
: google.protobuf.BoolValue
Enable or disable zookeeper_log system table. Default value: false.
zookeeperLogRetentionSize
: google.protobuf.Int64Value
The maximum size that zookeeper_log can grow to before old data will be removed. If set to 0 (default), automatic removal of zookeeper_log data based on size is disabled.
zookeeperLogRetentionTime
: google.protobuf.Int64Value
The maximum time that zookeeper_log records will be retained before removal. If set to 0, automatic removal of zookeeper_log data based on time is disabled.
asynchronousInsertLogEnabled
: google.protobuf.BoolValue
Enable or disable asynchronous_insert_log system table. Default value: false. Minimal required ClickHouse version: 22.10.
asynchronousInsertLogRetentionSize
: google.protobuf.Int64Value
The maximum size that asynchronous_insert_log can grow to before old data will be removed. If set to 0 (default), automatic removal of asynchronous_insert_log data based on size is disabled.
asynchronousInsertLogRetentionTime
: google.protobuf.Int64Value
The maximum time that asynchronous_insert_log records will be retained before removal. If set to 0, automatic removal of asynchronous_insert_log data based on time is disabled.
backgroundPoolSize
: google.protobuf.Int64Value
backgroundMergesMutationsConcurrencyRatio
: google.protobuf.Int64Value
Sets a ratio between the number of threads and the number of background merges and mutations that can be executed concurrently. For example, if the ratio equals to 2 and background_pool_size is set to 16 then ClickHouse can execute 32 background merges concurrently. This is possible, because background operations could be suspended and postponed. This is needed to give small merges more execution priority. You can only increase this ratio at runtime. To lower it you have to restart the server. The same as for background_pool_size setting background_merges_mutations_concurrency_ratio could be applied from the default profile for backward compatibility. Default: 2 See in-depth description in ClickHouse documentation
backgroundSchedulePoolSize
: google.protobuf.Int64Value
backgroundFetchesPoolSize
: google.protobuf.Int64Value
Sets the number of threads performing background fetches for tables with ReplicatedMergeTree engines. Default value: 8.
More info see in ClickHouse documentation.
backgroundMovePoolSize
: google.protobuf.Int64Value
backgroundDistributedSchedulePoolSize
: google.protobuf.Int64Value
backgroundBufferFlushSchedulePoolSize
: google.protobuf.Int64Value
backgroundMessageBrokerSchedulePoolSize
: google.protobuf.Int64Value
backgroundCommonPoolSize
: google.protobuf.Int64Value
The maximum number of threads that will be used for performing a variety of operations (mostly garbage collection) for *MergeTree-engine tables in a background. Default: 8 See in-depth description in ClickHouse documentation
defaultDatabase
: google.protobuf.StringValue
The default database.
To get a list of cluster databases, see Yandex Managed ClickHouse documentation.
totalMemoryProfilerStep
: google.protobuf.Int64Value
Sets the memory size (in bytes) for a stack trace at every peak allocation step. Default value: 4194304.
More info see in ClickHouse documentation.
totalMemoryTrackerSampleProbability
: google.protobuf.DoubleValue
queryMaskingRules
: QueryMaskingRule
Regexp-based rules, which will be applied to queries as well as all log messages before storing them in server logs, system.query_log, system.text_log, system.processes tables, and in logs sent to the client. That allows preventing sensitive data leakage from SQL queries (like names, emails, personal identifiers or credit card numbers) to logs. Change of these settings is applied with ClickHouse restart See in-depth description in ClickHouse documentation
dictionariesLazyLoad
: google.protobuf.BoolValue
Lazy loading of dictionaries. Default: true See in-depth description in ClickHouse documentation
queryCache
: QueryCache
Query cache configuration. Min version: 23.5 See in-depth description in ClickHouse documentation
Resources
resourcePresetId
: string
ID of the preset for computational resources available to a host (CPU, memory etc.). All available presets are listed in the documentation
diskSize
: int64
Volume of the storage available to a host, in bytes.
diskTypeId
: string
Type of the storage environment for the host. Possible values:
- network-hdd - network HDD drive,
- network-ssd - network SSD drive,
- local-ssd - local SSD storage.
Clickhouse
config
: config.ClickhouseConfig
Configuration for a ClickHouse server.
resources
: Resources
Resources allocated to ClickHouse hosts.
Zookeeper
resources
: Resources
Resources allocated to ZooKeeper hosts. If not set, minimal available resources will be used. All available resource presets can be retrieved with a ResourcePresetService.List request.
Access
dataLens
: bool
Allow to export data from the cluster to DataLens.
webSql
: bool
Allow SQL queries to the cluster databases from the management console.
See SQL queries in the management console for more details.
metrika
: bool
Allow to import data from Yandex Metrica and AppMetrica to the cluster.
See AppMetrica documentation for more details.
serverless
: bool
Allow access to cluster for Serverless.
dataTransfer
: bool
Allow access for DataTransfer
yandexQuery
: bool
Allow access for Query
CloudStorage
enabled
: bool
Whether to use Object Storage for storing ClickHouse data.
moveFactor
: google.protobuf.DoubleValue
dataCacheEnabled
: google.protobuf.BoolValue
dataCacheMaxSize
: google.protobuf.Int64Value
preferNotToMerge
: google.protobuf.BoolValue
Permission
databaseName
: string
Name of the database that the permission grants access to.
UserSettings
ClickHouse user settings. Supported settings are a limited subset of all settings described in ClickHouse documentation.
OverflowMode
OVERFLOW_MODE_UNSPECIFIED
OVERFLOW_MODE_THROW
OVERFLOW_MODE_BREAK
GroupByOverflowMode
GROUP_BY_OVERFLOW_MODE_UNSPECIFIED
GROUP_BY_OVERFLOW_MODE_THROW
GROUP_BY_OVERFLOW_MODE_BREAK
GROUP_BY_OVERFLOW_MODE_ANY
DistributedProductMode
DISTRIBUTED_PRODUCT_MODE_UNSPECIFIED
DISTRIBUTED_PRODUCT_MODE_DENY
Default value. Prohibits using these types of subqueries (returns the "Double-distributed in/JOIN subqueries is denied" exception).
DISTRIBUTED_PRODUCT_MODE_LOCAL
Replaces the database and table in the subquery with local ones for the destination server (shard), leaving the normal IN/JOIN.
DISTRIBUTED_PRODUCT_MODE_GLOBAL
Replaces the IN/JOIN query with GLOBAL IN/GLOBAL JOIN.
DISTRIBUTED_PRODUCT_MODE_ALLOW
Allows the use of these types of subqueries.
QuotaMode
QUOTA_MODE_UNSPECIFIED
QUOTA_MODE_DEFAULT
QUOTA_MODE_KEYED
QUOTA_MODE_KEYED_BY_IP
CountDistinctImplementation
COUNT_DISTINCT_IMPLEMENTATION_UNSPECIFIED
COUNT_DISTINCT_IMPLEMENTATION_UNIQ
COUNT_DISTINCT_IMPLEMENTATION_UNIQ_COMBINED
COUNT_DISTINCT_IMPLEMENTATION_UNIQ_COMBINED_64
COUNT_DISTINCT_IMPLEMENTATION_UNIQ_HLL_12
COUNT_DISTINCT_IMPLEMENTATION_UNIQ_EXACT
JoinAlgorithm
JOIN_ALGORITHM_UNSPECIFIED
JOIN_ALGORITHM_HASH
JOIN_ALGORITHM_PARALLEL_HASH
JOIN_ALGORITHM_PARTIAL_MERGE
JOIN_ALGORITHM_DIRECT
JOIN_ALGORITHM_AUTO
JOIN_ALGORITHM_FULL_SORTING_MERGE
JOIN_ALGORITHM_PREFER_PARTIAL_MERGE
FormatRegexpEscapingRule
FORMAT_REGEXP_ESCAPING_RULE_UNSPECIFIED
FORMAT_REGEXP_ESCAPING_RULE_ESCAPED
FORMAT_REGEXP_ESCAPING_RULE_QUOTED
FORMAT_REGEXP_ESCAPING_RULE_CSV
FORMAT_REGEXP_ESCAPING_RULE_JSON
FORMAT_REGEXP_ESCAPING_RULE_XML
FORMAT_REGEXP_ESCAPING_RULE_RAW
DateTimeInputFormat
DATE_TIME_INPUT_FORMAT_UNSPECIFIED
DATE_TIME_INPUT_FORMAT_BEST_EFFORT
DATE_TIME_INPUT_FORMAT_BASIC
DATE_TIME_INPUT_FORMAT_BEST_EFFORT_US
DateTimeOutputFormat
DATE_TIME_OUTPUT_FORMAT_UNSPECIFIED
DATE_TIME_OUTPUT_FORMAT_SIMPLE
DATE_TIME_OUTPUT_FORMAT_ISO
DATE_TIME_OUTPUT_FORMAT_UNIX_TIMESTAMP
LocalFilesystemReadMethod
LOCAL_FILESYSTEM_READ_METHOD_UNSPECIFIED
LOCAL_FILESYSTEM_READ_METHOD_READ
LOCAL_FILESYSTEM_READ_METHOD_PREAD_THREADPOOL
LOCAL_FILESYSTEM_READ_METHOD_PREAD
LOCAL_FILESYSTEM_READ_METHOD_NMAP
RemoteFilesystemReadMethod
REMOTE_FILESYSTEM_READ_METHOD_UNSPECIFIED
REMOTE_FILESYSTEM_READ_METHOD_READ
REMOTE_FILESYSTEM_READ_METHOD_THREADPOOL
LoadBalancing
LOAD_BALANCING_UNSPECIFIED
LOAD_BALANCING_RANDOM
LOAD_BALANCING_NEAREST_HOSTNAME
LOAD_BALANCING_IN_ORDER
LOAD_BALANCING_FIRST_OR_RANDOM
LOAD_BALANCING_ROUND_ROBIN
readonly
: google.protobuf.Int64Value
Restricts permissions for non-DDL queries. To restrict permissions for DDL queries, use allow_ddl instead.
- 0 (default)-no restrictions.
- 1-only read data queries are allowed.
- 2-read data and change settings queries are allowed.
See in-depth description in ClickHouse documentation.
allowDdl
: google.protobuf.BoolValue
Determines whether DDL queries are allowed (e.g., CREATE, ALTER, RENAME, etc).
Default value: true.
See in-depth description in ClickHouse documentation.
allowIntrospectionFunctions
: google.protobuf.BoolValue
Enables introspections functions for query profiling.
More info see in ClickHouse documentation.
connectTimeout
: google.protobuf.Int64Value
Connection timeout in milliseconds.
Value must be greater than 0 (default: 10000, 10 seconds).
connectTimeoutWithFailover
: google.protobuf.Int64Value
The timeout in milliseconds for connecting to a remote server for a Distributed table engine. Applies only if the cluster uses sharding and replication. If unsuccessful, several attempts are made to connect to various replicas.
Default value: 50.
More info see in ClickHouse documentation.
receiveTimeout
: google.protobuf.Int64Value
Receive timeout in milliseconds.
Value must be greater than 0 (default: 300000, 300 seconds or 5 minutes).
sendTimeout
: google.protobuf.Int64Value
Send timeout in milliseconds.
Value must be greater than 0 (default: 300000, 300 seconds or 5 minutes).
timeoutBeforeCheckingExecutionSpeed
: google.protobuf.Int64Value
Timeout (in seconds) between checks of execution speed. It is checked that execution speed is not less that specified in min_execution_speed parameter.
Default value: 10.
insertQuorum
: google.protobuf.Int64Value
Enables or disables write quorum for ClickHouse cluster. If the value is less than 2, then write quorum is disabled, otherwise it is enabled.
When used, write quorum guarantees that ClickHouse has written data to the quorum of insert_quorum replicas with no errors until the insert_quorum_timeout expires. All replicas in the quorum are in the consistent state, meaning that they contain linearized data from the previous INSERT queries. Employ write quorum, if you need the guarantees that the written data would not be lost in case of one or more replicas failure.
You can use select_sequential_consistency setting to read the data written with write quorum.
See in-depth description in ClickHouse documentation.
insertQuorumTimeout
: google.protobuf.Int64Value
Quorum write timeout in milliseconds.
If the write quorum is enabled in the cluster, this timeout expires and some data is not written to the insert_quorum replicas, then ClickHouse will abort the execution of INSERT query and return an error. In this case, the client must send the query again to write the data block into the same or another replica.
Minimum value: 1000, 1 second (default: 60000, 1 minute).
insertQuorumParallel
: google.protobuf.BoolValue
See in-depth description in ClickHouse documentation.
insertNullAsDefault
: google.protobuf.BoolValue
Enables the insertion of default values instead of NULL into columns with not nullable data type.
Default value: true.
More info see in ClickHouse documentation.
selectSequentialConsistency
: google.protobuf.BoolValue
Determines the behavior of SELECT queries from the replicated table: if enabled, ClickHouse will terminate a query with error message in case the replica does not have a chunk written with the quorum and will not read the parts that have not yet been written with the quorum.
Default value: false (sequential consistency is disabled).
deduplicateBlocksInDependentMaterializedViews
: google.protobuf.BoolValue
See in-depth description in ClickHouse documentation.
replicationAlterPartitionsSync
: google.protobuf.Int64Value
Wait mode for asynchronous actions in ALTER queries on replicated tables:
- 0-do not wait for replicas.
- 1-only wait for own execution (default).
- 2-wait for all replicas.
See in-depth description in ClickHouse documentation.
maxReplicaDelayForDistributedQueries
: google.protobuf.Int64Value
Max replica delay in milliseconds. If a replica lags more than the set value, this replica is not used and becomes a stale one.
Minimum value: 1000, 1 second (default: 300000, 300 seconds or 5 minutes).
See in-depth description in ClickHouse documentation.
fallbackToStaleReplicasForDistributedQueries
: google.protobuf.BoolValue
Enables or disables query forcing to a stale replica in case the actual data is unavailable. If enabled, ClickHouse will choose the most up-to-date replica and force the query to use the data in this replica. This setting can be used when doing SELECT query from a distributed table that points to replicated tables.
Default value: true (query forcing is enabled).
See in-depth description in ClickHouse documentation.
distributedProductMode
: DistributedProductMode
Determine the behavior of distributed subqueries.
See in-depth description in ClickHouse documentation.
distributedAggregationMemoryEfficient
: google.protobuf.BoolValue
Enables of disables memory saving mode when doing distributed aggregation.
When ClickHouse works with a distributed query, external aggregation is done on remote servers. Enable this setting to achieve a smaller memory footprint on the server that sourced such a distributed query.
Default value: false (memory saving mode is disabled).
See in-depth description in ClickHouse documentation.
distributedDdlTaskTimeout
: google.protobuf.Int64Value
Timeout for DDL queries, in milliseconds.
skipUnavailableShards
: google.protobuf.BoolValue
Enables or disables silent skipping of unavailable shards.
A shard is considered unavailable if all its replicas are also unavailable.
Default value: false (silent skipping is disabled).
See in-depth description in ClickHouse documentation.
compileExpressions
: google.protobuf.BoolValue
Enables or disables expression compilation. If you execute a lot of queries that contain identical expressions, then enable this setting. As a result, such queries may be executed faster due to use of compiled expressions.
Use this setting in combination with min_count_to_compile_expression setting.
Default value: false (expression compilation is disabled).
minCountToCompileExpression
: google.protobuf.Int64Value
How many identical expressions ClickHouse has to encounter before they are compiled.
Minimum value: 0 (default: 3).
For the 0 value compilation is synchronous: a query waits for expression compilation process to complete prior to continuing execution. It is recommended to set this value only for testing purposes.
For all other values, compilation is asynchronous: the compilation process executes in a separate thread. When a compiled expression is ready, it will be used by ClickHouse for eligible queries, including the ones that are currently running.
maxBlockSize
: google.protobuf.Int64Value
The maximum block size for reading.
Data in ClickHouse is organized and processed by blocks (block is a set of columns' parts). The internal processing cycles for a single block are efficient enough, but there are noticeable expenditures on each block.
This setting is a recommendation for size of block (in a count of rows) that should be loaded from tables.
Value must be greater than 0 (default: 65536).
See in-depth description in ClickHouse documentation.
minInsertBlockSizeRows
: google.protobuf.Int64Value
Limits the minimum number of rows in a block to be inserted in a table by INSERT query. Blocks that are smaller than the specified value, will be squashed together into the bigger blocks.
Minimal value: 0, block squashing is disabled (default: 1048576).
minInsertBlockSizeBytes
: google.protobuf.Int64Value
Limits the minimum number of bytes in a block to be inserted in a table by INSERT query. Blocks that are smaller than the specified value, will be squashed together into the bigger blocks.
Minimal value: 0, block squashing is disabled (default: 268435456, 256 MB).
maxInsertBlockSize
: google.protobuf.Int64Value
Allows to form blocks of the specified size (in bytes) when inserting data in a table. This setting has effect only if server is creating such blocks by itself.
Value must be greater than 0 (default: 1048576).
See in-depth description in ClickHouse documentation.
minBytesToUseDirectIo
: google.protobuf.Int64Value
Limits the minimum number of bytes to enable unbuffered direct reads from disk (Direct I/O).
By default, ClickHouse does not read data directly from disk, but relies on the filesystem and its cache instead. Such reading strategy is effective when the data volume is small. If the amount of the data to read is huge, it is more effective to read directly from the disk, bypassing the filesystem cache.
If the total amount of the data to read is greater than the value of this setting, then ClickHouse will fetch this data directly from the disk.
Minimal value and default value: 0, Direct I/O is disabled.
useUncompressedCache
: google.protobuf.BoolValue
Determines whether to use the cache of uncompressed blocks, or not. Using this cache can significantly reduce latency and increase the throughput when a huge amount of small queries is to be processed. Enable this setting for the users who instantiates small queries frequently.
This setting has effect only for tables of the MergeTree family.
Default value: false (uncompressed cache is disabled).
See in-depth description in ClickHouse documentation.
mergeTreeMaxRowsToUseCache
: google.protobuf.Int64Value
Limits the maximum size in rows of the request that can use the cache of uncompressed data. The cache is not used for requests larger than the specified value.
Use this setting in combination with use_uncompressed_cache setting.
Value must be greater than 0 (default: 128x8192).
mergeTreeMaxBytesToUseCache
: google.protobuf.Int64Value
Limits the maximum size in bytes of the request that can use the cache of uncompressed data. The cache is not used for requests larger than the specified value.
Use this setting in combination with use_uncompressed_cache setting.
Value must be greater than 0 (default: 192x10x1024x1024).
mergeTreeMinRowsForConcurrentRead
: google.protobuf.Int64Value
Limits the minimum number of rows to be read from a file to enable concurrent read. If the number of rows to be read exceeds this value, then ClickHouse will try to use a few threads to read from a file concurrently.
This setting has effect only for tables of the MergeTree family.
Value must be greater than 0 (default: 20x8192).
mergeTreeMinBytesForConcurrentRead
: google.protobuf.Int64Value
Limits the number of bytes to be read from a file to enable concurrent read. If the number of bytes to be read exceeds this value, then ClickHouse will try to use a few threads to read from a file concurrently.
This setting has effect only for tables of the MergeTree family.
Value must be greater than 0 (default: 24x10x1024x1024).
maxBytesBeforeExternalGroupBy
: google.protobuf.Int64Value
Sets the threshold of RAM consumption (in bytes) after that the temporary data, collected during the GROUP BY operation, should be flushed to disk to limit the RAM comsumption.
By default, aggregation is done by employing hash table that resides in RAM. A query can result in aggregation of huge data volumes that can lead to memory exhaustion and abortion of the query (see the max_memory_usage setting). For such queries, you can use this setting to force ClickHouse to do flushing and complete aggregation successfully.
Minimal value and default value: 0, GROUP BY in the external memory is disabled.
When using aggregation in external memory, it is recommended to set the value of this setting twice as low as the max_memory_usage setting value (by default, the maximum memory usage is limited to ten gigabytes).
See in-depth description in ClickHouse documentation.
See also: the distributed_aggregation_memory_efficient setting.
maxBytesBeforeExternalSort
: google.protobuf.Int64Value
This setting is equivalent of the max_bytes_before_external_group_by setting, except for it is for sort operation (ORDER BY), not aggregation.
groupByTwoLevelThreshold
: google.protobuf.Int64Value
Sets the threshold of the number of keys, after that the two-level aggregation should be used.
Minimal value: 0, threshold is not set (default: 10000).
groupByTwoLevelThresholdBytes
: google.protobuf.Int64Value
Sets the threshold of the number of bytes, after that the two-level aggregation should be used.
Minimal value: 0, threshold is not set (default: 100000000).
priority
: google.protobuf.Int64Value
Sets the priority of a query.
- 0-priority is not used.
- 1-the highest priority.
- and so on. The higher the number, the lower a query's priority.
This setting should be set up for each query individually.
If ClickHouse is working with the high-priority queries, and a low-priority query enters, then the low-priority query is paused until higher-priority queries are completed.
Minimal value and default value: 0, priority is not used.
maxThreads
: google.protobuf.Int64Value
Limits the maximum number of threads to process the request (setting does not take threads that read data from remote servers into account).
This setting applies to threads that perform the same stages of the query processing pipeline in parallel.
Minimal value and default value: 0 (the thread number is calculated automatically based on the number of physical CPU cores, no HyperThreading cores are taken into account).
See in-depth description in ClickHouse documentation.
maxMemoryUsage
: google.protobuf.Int64Value
Limits the maximum memory usage (in bytes) for processing of a single user's query on a single server. This setting does not take server's free RAM amount or total RAM amount into account.
This limitation is enforced for any user's single query on a single server.
Minimal value: 0, no limitation is set. Value that is set in the ClickHouse default config file: 10737418240 (10 GB).
If you use max_bytes_before_external_group_by or max_bytes_before_external_sort setting, then it is recommended to set their values twice as low as max_memory_usage setting value.
See in-depth description in ClickHouse documentation.
maxMemoryUsageForUser
: google.protobuf.Int64Value
Limits the maximum memory usage (in bytes) for processing of user's queries on a single server. This setting does not take server's free RAM amount or total RAM amount into account.
This limitation is enforced for all queries that belong to one user and run simultaneously on a single server.
Minimal value and default value: 0, no limitation is set.
maxNetworkBandwidth
: google.protobuf.Int64Value
The maximum speed of data exchange over the network in bytes per second for a query.
Minimal value and default value: 0, no limitation is set.
maxNetworkBandwidthForUser
: google.protobuf.Int64Value
The maximum speed of data exchange over the network in bytes per second for all concurrently running user queries.
Minimal value and default value: 0, no limitation is set.
maxPartitionsPerInsertBlock
: google.protobuf.Int64Value
See in-depth description in ClickHouse documentation.
maxConcurrentQueriesForUser
: google.protobuf.Int64Value
The maximum number of concurrent requests per user. Default value: 0 (no limit).
forceIndexByDate
: google.protobuf.BoolValue
If enabled, query is not executed if the ClickHouse can't use index by date. This setting has effect only for tables of the MergeTree family.
Default value: false (setting is disabled, query executes even if ClickHouse can't use index by date).
See in-depth description in ClickHouse documentation.
forcePrimaryKey
: google.protobuf.BoolValue
If enabled, query is not executed if the ClickHouse can't use index by primary key. This setting has effect only for tables of the MergeTree family.
Default value: false (setting is disabled, query executes even if ClickHouse can't use index by primary key).
See in-depth description in ClickHouse documentation.
maxRowsToRead
: google.protobuf.Int64Value
Limits the maximum number of rows that can be read from a table when running a query.
Minimal value and default value: 0, no limitation is set.
See in-depth description in ClickHouse documentation.
maxBytesToRead
: google.protobuf.Int64Value
Limits the maximum number of bytes (uncompressed data) that can be read from a table when running a query.
Minimal value and default value: 0, no limitation is set.
readOverflowMode
: OverflowMode
Determines the behavior on exceeding limits while reading the data.
- throw-abort query execution, return an error.
- break-stop query execution, return partial result.
maxRowsToGroupBy
: google.protobuf.Int64Value
Limits the maximum number of unique keys received from aggregation function. This setting helps to reduce RAM consumption while doing aggregation.
Minimal value and default value: 0, no limitation is set.
groupByOverflowMode
: GroupByOverflowMode
Determines the behavior on exceeding limits while doing aggregation.
- throw-abort query execution, return an error.
- break-stop query execution, return partial result.
- any-perform approximate GROUP BY operation by continuing aggregation for the keys that got into the set, but don't add new keys to the set.
maxRowsToSort
: google.protobuf.Int64Value
Limits the maximum number of rows that can be read from a table for sorting. This setting helps to reduce RAM consumption.
Minimal value and default value: 0, no limitation is set.
maxBytesToSort
: google.protobuf.Int64Value
Limits the maximum number of bytes (uncompressed data) that can be read from a table for sorting. This setting helps to reduce RAM consumption.
Minimal value and default value: 0, no limitation is set.
sortOverflowMode
: OverflowMode
Determines the behavior on exceeding limits while sorting.
- throw-abort query execution, return an error.
- break-stop query execution, return partial result.
maxResultRows
: google.protobuf.Int64Value
Limits the number of rows in the result. This limitation is also checked for subqueries and parts of distributed queries that run on remote servers.
Minimal value and default value: 0, no limitation is set.
maxResultBytes
: google.protobuf.Int64Value
Limits the number of bytes in the result. This limitation is also checked for subqueries and parts of distributed queries that run on remote servers.
Minimal value and default value: 0, no limitation is set.
resultOverflowMode
: OverflowMode
Determines the behavior on exceeding limits while forming result.
- throw-abort query execution, return an error.
- break-stop query execution, return partial result.
maxRowsInDistinct
: google.protobuf.Int64Value
Limits the maximum number of different rows when using DISTINCT.
Minimal value and default value: 0, no limitation is set.
maxBytesInDistinct
: google.protobuf.Int64Value
Limits the maximum size of a hash table in bytes (uncompressed data) when using DISTINCT.
distinctOverflowMode
: OverflowMode
Determines the behavior on exceeding limits while doing DISCTINCT.
- throw-abort query execution, return an error.
- break-stop query execution, return partial result.
maxRowsToTransfer
: google.protobuf.Int64Value
Limits the maximum number of rows that can be passed to a remote server or saved in a temporary table when using GLOBAL IN.
Minimal value and default value: 0, no limitation is set.
maxBytesToTransfer
: google.protobuf.Int64Value
Limits the maximum number of bytes (uncompressed data) that can be passed to a remote server or saved in a temporary table when using GLOBAL IN.
Minimal value and default value: 0, no limitation is set.
transferOverflowMode
: OverflowMode
Determines the behavior on exceeding limits while doing transfers.
- throw-abort query execution, return an error.
- break-stop query execution, return partial result.
maxExecutionTime
: google.protobuf.Int64Value
Limits the maximum query execution time in milliseconds. At this moment, this limitation is not checked when passing one of the sorting stages, as well as merging and finalizing aggregation funictions.
Minimal value and default value: 0, no limitation is set.
timeoutOverflowMode
: OverflowMode
Determines the behavior on exceeding limits of execution time.
- throw-abort query execution, return an error.
- break-stop query execution, return partial result.
maxRowsInSet
: google.protobuf.Int64Value
Limit on the number of rows in the set resulting from the execution of the IN section.
maxBytesInSet
: google.protobuf.Int64Value
Limit on the number of bytes in the set resulting from the execution of the IN section.
setOverflowMode
: OverflowMode
Determine the behavior on exceeding max_rows_in_set or max_bytes_in_set limit. Possible values: OVERFLOW_MODE_THROW, OVERFLOW_MODE_BREAK.
maxRowsInJoin
: google.protobuf.Int64Value
Limit on maximum size of the hash table for JOIN, in rows.
maxBytesInJoin
: google.protobuf.Int64Value
Limit on maximum size of the hash table for JOIN, in bytes.
joinOverflowMode
: OverflowMode
Determine the behavior on exceeding max_rows_in_join or max_bytes_in_join limit. Possible values: OVERFLOW_MODE_THROW, OVERFLOW_MODE_BREAK.
joinAlgorithm
: JoinAlgorithm
See in-depth description in ClickHouse documentation.
anyJoinDistinctRightTableKeys
: google.protobuf.BoolValue
See in-depth description in ClickHouse documentation.
maxColumnsToRead
: google.protobuf.Int64Value
Limits the maximum number of columns that can be read from a table in a single query. If the query requires to read more columns to complete, then it will be aborted.
Minimal value and default value: 0, no limitation is set.
maxTemporaryColumns
: google.protobuf.Int64Value
Limits the maximum number of temporary columns that must be kept in RAM at the same time when running a query, including constant columns.
Minimal value and default value: 0, no limitation is set.
maxTemporaryNonConstColumns
: google.protobuf.Int64Value
Limits the maximum number of temporary columns that must be kept in RAM at the same time when running a query, excluding constant columns.
Minimal value and default value: 0, no limitation is set.
maxQuerySize
: google.protobuf.Int64Value
Limits the size of the part of a query that can be transferred to RAM for parsing with the SQL parser, in bytes.
Value must be greater than 0 (default: 262144).
See in-depth description in ClickHouse documentation.
maxAstDepth
: google.protobuf.Int64Value
Limits the maximum depth of query syntax tree.
Executing a big and complex query may result in building a syntax tree of enormous depth. By using this setting, you can prohibit execution of over-sized or non-optimized queries for huge tables.
For example, the SELECT * query may result in more complex and deeper syntax tree, compared to the SELECT ... WHERE ... query, containing constraints and conditions, in the most cases. A user can be forced to construct more optimized queries, if this setting is used.
Value must be greater than 0 (default: 1000). If a too small value is set, it may render ClickHouse unable to execute even simple queries.
See in-depth description in ClickHouse documentation.
maxAstElements
: google.protobuf.Int64Value
Limits the maximum size of query syntax tree in number of nodes.
Executing a big and complex query may result in building a syntax tree of enormous size. By using this setting, you can prohibit execution of over-sized or non-optimized queries for huge tables.
Value must be greater than 0 (default: 50000). If a too small value is set, it may render ClickHouse unable to execute even simple queries.
See in-depth description in ClickHouse documentation.
maxExpandedAstElements
: google.protobuf.Int64Value
Limits the maximum size of query syntax tree in number of nodes after expansion of aliases and the asterisk values.
Executing a big and complex query may result in building a syntax tree of enormous size. By using this setting, you can prohibit execution of over-sized or non-optimized queries for huge tables.
Value must be greater than 0 (default: 500000). If a too small value is set, it may render ClickHouse unable to execute even simple queries.
minExecutionSpeed
: google.protobuf.Int64Value
Minimal execution speed in rows per second.
minExecutionSpeedBytes
: google.protobuf.Int64Value
Minimal execution speed in bytes per second.
countDistinctImplementation
: CountDistinctImplementation
Aggregate function to use for implementation of count(DISTINCT ...).
inputFormatValuesInterpretExpressions
: google.protobuf.BoolValue
Enables or disables SQL parser if the fast stream parser cannot parse the data.
Enable this setting, if the data that you want to insert into a table contains SQL expressions.
For example, the stream parser is unable to parse a value that contains now() expression; therefore an INSERT query for this value will fail and no data will be inserted into a table. With enabled SQL parser, this expression is parsed correctly: the now() expression will be parsed as SQL function, interpreted, and the current date and time will be inserted into the table as a result.
This setting has effect only if you use Values format when inserting data.
Default value: true (SQL parser is enabled).
See in-depth description in ClickHouse documentation.
inputFormatDefaultsForOmittedFields
: google.protobuf.BoolValue
Enables or disables replacing omitted input values with default values of the respective columns when performing INSERT queries.
Default value: true (replacing is enabled).
inputFormatNullAsDefault
: google.protobuf.BoolValue
See in-depth description in ClickHouse documentation.
dateTimeInputFormat
: DateTimeInputFormat
See in-depth description in ClickHouse documentation.
inputFormatWithNamesUseHeader
: google.protobuf.BoolValue
See in-depth description in ClickHouse documentation.
outputFormatJsonQuote_64bitIntegers
: google.protobuf.BoolValue
Enables quoting of 64-bit integers in JSON output format.
If this setting is enabled, then 64-bit integers (UInt64 and Int64) will be quoted when written to JSON output in order to maintain compatibility with the most of the JavaScript engines. Otherwise, such integers will not be quoted.
Default value: false (quoting 64-bit integers is disabled).
outputFormatJsonQuoteDenormals
: google.protobuf.BoolValue
Enables special floating-point values (+nan, -nan, +inf and -inf) in JSON output format.
Default value: false (special values do not present in output).
dateTimeOutputFormat
: DateTimeOutputFormat
See in-depth description in ClickHouse documentation.
lowCardinalityAllowInNativeFormat
: google.protobuf.BoolValue
Determines whether to use LowCardinality type in Native format.
- true (default)-yes, use.
- false-convert LowCardinality columns to regular columns when doing SELECT, and convert regular columns to LowCardinality when doing INSERT.
LowCardinality columns (aka sparse columns) store data in more effective way, compared to regular columns, by using hash tables. If data to insert suits this storage format, ClickHouse will place them into LowCardinality column.
If you use a third-party ClickHouse client that can't work with LowCardinality columns, then this client will not be able to correctly interpret the result of the query that asks for data stored in LowCardinality column. Disable this setting to convert LowCardinality column to regular column when creating the result, so such clients will be able to process the result.
Official ClickHouse client works with LowCardinality columns out-of-the-box.
Default value: true (LowCardinality columns are used in Native format).
allowSuspiciousLowCardinalityTypes
: google.protobuf.BoolValue
Allows specifying LowCardinality modifier for types of small fixed size (8 or less) in CREATE TABLE statements. Enabling this may increase merge times and memory consumption.
More info see in ClickHouse documentation.
emptyResultForAggregationByEmptySet
: google.protobuf.BoolValue
Enables returning of empty result when aggregating without keys (with GROUP BY operation absent) on empty set (e.g., SELECT count(*) FROM table WHERE 0).
- true-ClickHouse will return an empty result for such queries.
- false (default)-ClickHouse will return a single-line result consisting of NULL values for aggregation functions, in accordance with SQL standard.
httpConnectionTimeout
: google.protobuf.Int64Value
HTTP connection timeout, in milliseconds.
Value must be greater than 0 (default: 1000, 1 second).
httpReceiveTimeout
: google.protobuf.Int64Value
HTTP receive timeout, in milliseconds.
Value must be greater than 0 (default: 1800000, 1800 seconds, 30 minutes).
httpSendTimeout
: google.protobuf.Int64Value
HTTP send timeout, in milliseconds.
Value must be greater than 0 (default: 1800000, 1800 seconds, 30 minutes).
enableHttpCompression
: google.protobuf.BoolValue
Enables or disables data compression in HTTP responses.
By default, ClickHouse stores data compressed. When executing a query, its result is uncompressed. Use this setting to command ClickHouse to compress the result when sending it via HTTP.
Enable this setting and add the Accept-Encoding: <compression method> HTTP header in a HTTP request to force compression of HTTP response from ClickHouse.
ClickHouse support the following compression methods: gzip, br and deflate.
Default value: false (compression is disabled).
See in-depth description in ClickHouse documentation.
sendProgressInHttpHeaders
: google.protobuf.BoolValue
Enables progress notifications using X-ClickHouse-Progress HTTP header.
Default value: false (notifications disabled).
httpHeadersProgressInterval
: google.protobuf.Int64Value
Minimum interval between progress notifications with X-ClickHouse-Progress HTTP header, in milliseconds.
Value must be greater than 0 (default: 100).
addHttpCorsHeader
: google.protobuf.BoolValue
Adds CORS header in HTTP responses.
Default value: false (header is not added).
cancelHttpReadonlyQueriesOnClientClose
: google.protobuf.BoolValue
Cancels HTTP read-only queries (e.g. SELECT) when a client closes the connection without waiting for the response.
Default value: false.
More info see in ClickHouse documentation.
maxHttpGetRedirects
: google.protobuf.Int64Value
Limits the maximum number of HTTP GET redirect hops for URL-engine tables.
If the parameter is set to 0 (default), no hops is allowed.
More info see in ClickHouse documentation.
joinedSubqueryRequiresAlias
: google.protobuf.BoolValue
joinUseNulls
: google.protobuf.BoolValue
transformNullIn
: google.protobuf.BoolValue
quotaMode
: QuotaMode
Quota accounting mode. Possible values: QUOTA_MODE_DEFAULT, QUOTA_MODE_KEYED and QUOTA_MODE_KEYED_BY_IP.
flattenNested
: google.protobuf.BoolValue
Sets the data format of a nested columns.
More info see in ClickHouse documentation.
formatRegexp
: string
Regular expression (for Regexp format)
formatRegexpEscapingRule
: FormatRegexpEscapingRule
See in-depth description in ClickHouse documentation.
formatRegexpSkipUnmatched
: google.protobuf.BoolValue
See in-depth description in ClickHouse documentation.
asyncInsert
: google.protobuf.BoolValue
Enables asynchronous inserts.
Disabled by default.
More info see in ClickHouse documentation.
asyncInsertThreads
: google.protobuf.Int64Value
The maximum number of threads for background data parsing and insertion.
If the parameter is set to 0, asynchronous insertions are disabled. Default value: 16.
More info see in ClickHouse documentation.
waitForAsyncInsert
: google.protobuf.BoolValue
Enables waiting for processing of asynchronous insertion. If enabled, server returns OK only after the data is inserted.
More info see in ClickHouse documentation.
waitForAsyncInsertTimeout
: google.protobuf.Int64Value
The timeout (in seconds) for waiting for processing of asynchronous insertion.
Default value: 120.
More info see in ClickHouse documentation.
asyncInsertMaxDataSize
: google.protobuf.Int64Value
The maximum size of the unparsed data in bytes collected per query before being inserted.
If the parameter is set to 0, asynchronous insertions are disabled. Default value: 100000.
More info see in ClickHouse documentation.
asyncInsertBusyTimeout
: google.protobuf.Int64Value
The maximum timeout in milliseconds since the first INSERT query before inserting collected data.
If the parameter is set to 0, the timeout is disabled. Default value: 200.
More info see in ClickHouse documentation.
asyncInsertStaleTimeout
: google.protobuf.Int64Value
The maximum timeout in milliseconds since the last INSERT query before dumping collected data. If enabled, the settings prolongs the async_insert_busy_timeout with every INSERT query as long as async_insert_max_data_size is not exceeded.
More info see in ClickHouse documentation.
memoryProfilerStep
: google.protobuf.Int64Value
Memory profiler step (in bytes).
If the next query step requires more memory than this parameter specifies, the memory profiler collects the allocating stack trace. Values lower than a few megabytes slow down query processing.
Default value: 4194304 (4 MB). Zero means disabled memory profiler.
memoryProfilerSampleProbability
: google.protobuf.DoubleValue
Collect random allocations and deallocations and write them into system.trace_log with 'MemorySample' trace_type. The probability is for every alloc/free regardless to the size of the allocation.
Possible values: from 0 to 1. Default: 0.
maxFinalThreads
: google.protobuf.Int64Value
Sets the maximum number of parallel threads for the SELECT query data read phase with the FINAL modifier. See in-depth description in ClickHouse documentation.
inputFormatParallelParsing
: google.protobuf.BoolValue
Enables or disables order-preserving parallel parsing of data formats. Supported only for TSV, TKSV, CSV and JSONEachRow formats. See in-depth description in ClickHouse documentation
inputFormatImportNestedJson
: google.protobuf.BoolValue
Enables or disables the insertion of JSON data with nested objects. See in-depth description in ClickHouse documentation
localFilesystemReadMethod
: LocalFilesystemReadMethod
Method of reading data from local filesystem, one of: read, pread, mmap, io_uring, pread_threadpool. The 'io_uring' method is experimental and does not work for Log, TinyLog, StripeLog, File, Set and Join, and other tables with append-able files in presence of concurrent reads and writes.
maxReadBufferSize
: google.protobuf.Int64Value
The maximum size of the buffer to read from the filesystem. See in-depth description in ClickHouse documentation
insertKeeperMaxRetries
: google.protobuf.Int64Value
The setting sets the maximum number of retries for ClickHouse Keeper (or ZooKeeper) requests during insert into replicated MergeTree. Only Keeper requests which failed due to network error, Keeper session timeout, or request timeout are considered for retries. Default: 20 from 23.2, 0(disabled) before Min_version: 22.11 See in-depth description in ClickHouse documentation
maxTemporaryDataOnDiskSizeForUser
: google.protobuf.Int64Value
The maximum amount of data consumed by temporary files on disk in bytes for all concurrently running user queries. Zero means unlimited. Default: 0 - unlimited Min_version: 22.10 See in-depth description in ClickHouse documentation
maxTemporaryDataOnDiskSizeForQuery
: google.protobuf.Int64Value
The maximum amount of data consumed by temporary files on disk in bytes for all concurrently running queries. Zero means unlimited. Default: 0 - unlimited Min_version: 22.10 See in-depth description in ClickHouse documentation
maxParserDepth
: google.protobuf.Int64Value
Limits maximum recursion depth in the recursive descent parser. Allows controlling the stack size. Default: 1000 Special: 0 - unlimited See in-depth description in ClickHouse documentation
remoteFilesystemReadMethod
: RemoteFilesystemReadMethod
Method of reading data from remote filesystem, one of: read, threadpool. Default: read Min_version: 21.11 See in-depth description in ClickHouse GitHub
memoryOvercommitRatioDenominator
: google.protobuf.Int64Value
It represents soft memory limit in case when hard limit is reached on user level. This value is used to compute overcommit ratio for the query. Zero means skip the query. Default: 1GiB Min_version: 22.5 See in-depth description in ClickHouse documentation
memoryOvercommitRatioDenominatorForUser
: google.protobuf.Int64Value
It represents soft memory limit in case when hard limit is reached on global level. This value is used to compute overcommit ratio for the query. Zero means skip the query. Default: 1GiB Min_version: 22.5 See in-depth description in ClickHouse documentation
memoryUsageOvercommitMaxWaitMicroseconds
: google.protobuf.Int64Value
Maximum time thread will wait for memory to be freed in the case of memory overcommit on a user level. If the timeout is reached and memory is not freed, an exception is thrown. Default: 5000000 Min_version: 22.5 See in-depth description in ClickHouse documentation
logQueryThreads
: google.protobuf.BoolValue
Setting up query threads logging. Query threads log into the system.query_thread_log table. This setting has effect only when log_queries is true. Queries threads run by ClickHouse with this setup are logged according to the rules in the query_thread_log server configuration parameter. Default: true See in-depth description in ClickHouse documentation
maxInsertThreads
: google.protobuf.Int64Value
The maximum number of threads to execute the INSERT SELECT query. Default: 0 See in-depth description in ClickHouse documentation
useHedgedRequests
: google.protobuf.BoolValue
Enables hedged requests logic for remote queries. It allows to establish many connections with different replicas for query. New connection is enabled in case existent connection(s) with replica(s) were not established within hedged_connection_timeout or no data was received within receive_data_timeout. Query uses the first connection which send non empty progress packet (or data packet, if allow_changing_replica_until_first_data_packet); other connections are cancelled. Queries with max_parallel_replicas > 1 are supported. Default: true See in-depth description in ClickHouse documentation
idleConnectionTimeout
: google.protobuf.Int64Value
Timeout to close idle TCP connections after specified number of milliseconds. Default: 360000 (3600 seconds) See in-depth description in ClickHouse documentation
hedgedConnectionTimeoutMs
: google.protobuf.Int64Value
Connection timeout for establishing connection with replica for Hedged requests. Default: 50 See in-depth description in ClickHouse GitHub
loadBalancing
: LoadBalancing
Specifies the algorithm of replicas selection that is used for distributed query processing, one of: random, nearest_hostname, in_order, first_or_random, round_robin. Default: random See in-depth description in ClickHouse documentation
preferLocalhostReplica
: google.protobuf.BoolValue
Enables/disables preferable using the localhost replica when processing distributed queries. Default: true See in-depth description in ClickHouse documentation
compile
: google.protobuf.BoolValue
The setting is deprecated and has no effect.
minCountToCompile
: google.protobuf.Int64Value
The setting is deprecated and has no effect.
UserQuota
ClickHouse quota representation. Each quota associated with an user and limits it resource usage for an interval. See in-depth description ClickHouse documentation.
intervalDuration
: google.protobuf.Int64Value
Duration of interval for quota in milliseconds. Minimal value is 1 second.
queries
: google.protobuf.Int64Value
The total number of queries. 0 - unlimited.
errors
: google.protobuf.Int64Value
The number of queries that threw exception. 0 - unlimited.
resultRows
: google.protobuf.Int64Value
The total number of rows given as the result.. 0 - unlimited.
readRows
: google.protobuf.Int64Value
The total number of source rows read from tables for running the query, on all remote servers. 0 - unlimited.
executionTime
: google.protobuf.Int64Value
The total query execution time, in milliseconds (wall time). 0 - unlimited.
AnytimeMaintenanceWindow
WeeklyMaintenanceWindow
Weelky maintenance window settings.
WeekDay
WEEK_DAY_UNSPECIFIED
MON
TUE
WED
THU
FRI
SAT
SUN
day
: WeekDay
Day of the week (in DDD
format).
hour
: int64
Hour of the day in UTC (in HH
format).
Kafka
SecurityProtocol
SECURITY_PROTOCOL_UNSPECIFIED
SECURITY_PROTOCOL_PLAINTEXT
SECURITY_PROTOCOL_SSL
SECURITY_PROTOCOL_SASL_PLAINTEXT
SECURITY_PROTOCOL_SASL_SSL
SaslMechanism
SASL_MECHANISM_UNSPECIFIED
SASL_MECHANISM_GSSAPI
SASL_MECHANISM_PLAIN
SASL_MECHANISM_SCRAM_SHA_256
SASL_MECHANISM_SCRAM_SHA_512
Debug
DEBUG_UNSPECIFIED
DEBUG_GENERIC
DEBUG_BROKER
DEBUG_TOPIC
DEBUG_METADATA
DEBUG_FEATURE
DEBUG_QUEUE
DEBUG_MSG
DEBUG_PROTOCOL
DEBUG_CGRP
DEBUG_SECURITY
DEBUG_FETCH
DEBUG_INTERCEPTOR
DEBUG_PLUGIN
DEBUG_CONSUMER
DEBUG_ADMIN
DEBUG_EOS
DEBUG_MOCK
DEBUG_ASSIGNOR
DEBUG_CONF
DEBUG_TELEMETRY
DEBUG_ALL
AutoOffsetReset
AUTO_OFFSET_RESET_UNSPECIFIED
AUTO_OFFSET_RESET_SMALLEST
AUTO_OFFSET_RESET_EARLIEST
AUTO_OFFSET_RESET_BEGINNING
AUTO_OFFSET_RESET_LARGEST
AUTO_OFFSET_RESET_LATEST
AUTO_OFFSET_RESET_END
AUTO_OFFSET_RESET_ERROR
securityProtocol
: SecurityProtocol
saslMechanism
: SaslMechanism
saslUsername
: string
saslPassword
: string
enableSslCertificateVerification
: google.protobuf.BoolValue
maxPollIntervalMs
: google.protobuf.Int64Value
sessionTimeoutMs
: google.protobuf.Int64Value
debug
: Debug
autoOffsetReset
: AutoOffsetReset
Header
name
: string
value
: string
Replica
host
: string
MySQL host of the replica.
priority
: int64
The priority of the replica that ClickHouse takes into account when connecting. Replica with the highest priority should have this field set to the lowest number.
port
: int64
Port to use when connecting to the replica. If a port is not specified for a replica, ClickHouse uses the port specified for the source.
user
: string
Name of the MySQL database user.
password
: string
Password of the MySQL database user.
Attribute
name
: string
Name of the column.
type
: string
Type of the column.
nullValue
: string
Default value for an element without data (for example, an empty string).
expression
: string
Expression, describing the attribute, if applicable.
hierarchical
: bool
Indication of hierarchy support.
Default value: false
.
injective
: bool
Indication of injective mapping "id -> attribute".
Default value: false
.
Id
Numeric key.
name
: string
Name of the numeric key.
Key
Complex key.
attributes
: Attribute
Attributes of a complex key.
Structure
Attribute
name
: string
Name of the column.
type
: string
Type of the column.
nullValue
: string
Default value for an element without data (for example, an empty string).
expression
: string
Expression, describing the attribute, if applicable.
hierarchical
: bool
Indication of hierarchy support.
Default value: false
.
injective
: bool
Indication of injective mapping "id -> attribute".
Default value: false
.
Id
Numeric key.
name
: string
Name of the numeric key.
Key
Complex key.
attributes
: Attribute
Attributes of a complex key.
id
: Id
Single numeric key column for the dictionary.
key
: Key
Composite key for the dictionary, containing of one or more key columns. For details, see ClickHouse documentation.
rangeMin
: Attribute
Field holding the beginning of the range for dictionaries with RANGE_HASHED
layout.
For details, see ClickHouse documentation.
rangeMax
: Attribute
Field holding the end of the range for dictionaries with RANGE_HASHED
layout.
For details, see ClickHouse documentation.
attributes
: Attribute
Description of the fields available for database queries. For details, see ClickHouse documentation.
Layout
Layout determining how to store the dictionary in memory.
Type
TYPE_UNSPECIFIED
FLAT
The entire dictionary is stored in memory in the form of flat arrays. Available for all dictionary sources.
HASHED
The entire dictionary is stored in memory in the form of a hash table. Available for all dictionary sources.
COMPLEX_KEY_HASHED
Similar to HASHED, to be used with composite keys. Available for all dictionary sources.
RANGE_HASHED
The entire dictionary is stored in memory in the form of a hash table, with an ordered array of ranges and their corresponding values. Available for all dictionary sources.
CACHE
The dictionary is stored in a cache with a set number of cells. Available for MySQL, ClickHouse and HTTP dictionary sources.
COMPLEX_KEY_CACHE
Similar to CACHE, to be used with composite keys. Available for MySQL, ClickHouse and HTTP dictionary sources.
type
: Type
Layout type for an external dictionary.
sizeInCells
: int64
Number of cells in the cache. Rounded up to a power of two. Applicable only for CACHE and COMPLEX_KEY_CACHE layout types.
maxArraySize
: int64
Maximum dictionary key size. Applicable only for FLAT layout type.
Range
min
: int64
Minimum dictionary lifetime.
max
: int64
Maximum dictionary lifetime.
HttpSource
Header
name
: string
value
: string
url
: string
URL of the source dictionary available over HTTP.
format
: string
The data format. Valid values are all formats supported by ClickHouse SQL dialect.
headers
: Header
HTTP headers.
MysqlSource
Replica
host
: string
MySQL host of the replica.
priority
: int64
The priority of the replica that ClickHouse takes into account when connecting. Replica with the highest priority should have this field set to the lowest number.
port
: int64
Port to use when connecting to the replica. If a port is not specified for a replica, ClickHouse uses the port specified for the source.
user
: string
Name of the MySQL database user.
password
: string
Password of the MySQL database user.
db
: string
Name of the MySQL database to connect to.
table
: string
Name of the database table to use as a ClickHouse dictionary.
port
: int64
Default port to use when connecting to a replica of the dictionary source.
user
: string
Name of the default user for replicas of the dictionary source.
password
: string
Password of the default user for replicas of the dictionary source.
replicas
: Replica
List of MySQL replicas of the database used as dictionary source.
where
: string
Selection criteria for the data in the specified MySQL table.
invalidateQuery
: string
Query for checking the dictionary status, to pull only updated data. For more details, see ClickHouse documentation on dictionaries.
closeConnection
: google.protobuf.BoolValue
Should the connection be closed after each request.
shareConnection
: google.protobuf.BoolValue
Should a connection be shared for some requests.
ClickhouseSource
db
: string
Name of the ClickHouse database.
table
: string
Name of the table in the specified database to be used as the dictionary source.
host
: string
ClickHouse host of the specified database.
port
: int64
Port to use when connecting to the host.
user
: string
Name of the ClickHouse database user.
password
: string
Password of the ClickHouse database user.
where
: string
Selection criteria for the data in the specified ClickHouse table.
secure
: google.protobuf.BoolValue
Use ssl for connection.
MongodbSource
db
: string
Name of the MongoDB database.
collection
: string
Name of the collection in the specified database to be used as the dictionary source.
host
: string
MongoDB host of the specified database.
port
: int64
Port to use when connecting to the host.
user
: string
Name of the MongoDB database user.
password
: string
Password of the MongoDB database user.
options
: string
PostgresqlSource
SslMode
SSL_MODE_UNSPECIFIED
DISABLE
Only try a non-SSL connection.
ALLOW
First try a non-SSL connection; if that fails, try an SSL connection.
PREFER
First try an SSL connection; if that fails, try a non-SSL connection.
VERIFY_CA
Only try an SSL connection, and verify that the server certificate is issued by a trusted certificate authority (CA).
VERIFY_FULL
Only try an SSL connection, verify that the server certificate is issued by a trusted CA and that the requested server host name matches that in the certificate.
db
: string
Name of the PostrgreSQL database.
table
: string
Name of the table in the specified database to be used as the dictionary source.
hosts
: string
Name of the PostrgreSQL host
port
: int64
Port to use when connecting to the host.
user
: string
Name of the PostrgreSQL database user.
password
: string
Password of the PostrgreSQL database user.
invalidateQuery
: string
Query for checking the dictionary status, to pull only updated data. For more details, see ClickHouse documentation on dictionaries.
sslMode
: SslMode
Mode of SSL TCP/IP connection to the PostgreSQL host. For more details, see PostgreSQL documentation.
Retention
age
: int64
Minimum age of the data in seconds.
precision
: int64
Precision of determining the age of the data, in seconds.
Pattern
Retention
age
: int64
Minimum age of the data in seconds.
precision
: int64
Precision of determining the age of the data, in seconds.
regexp
: string
Pattern for metric names.
function
: string
Name of the aggregating function to apply to data of the age specified in retention.
retention
: Retention
Age of data to use for thinning.
MergeTree
Options specific to the MergeTree table engine.
replicatedDeduplicationWindow
: google.protobuf.Int64Value
Number of blocks of hashes to keep in ZooKeeper.
replicatedDeduplicationWindowSeconds
: google.protobuf.Int64Value
Period of time to keep blocks of hashes for.
partsToDelayInsert
: google.protobuf.Int64Value
If table contains at least that many active parts in single partition, artificially slow down insert into table.
partsToThrowInsert
: google.protobuf.Int64Value
If more than this number active parts in single partition, throw 'Too many parts ...' exception.
inactivePartsToDelayInsert
: google.protobuf.Int64Value
inactivePartsToThrowInsert
: google.protobuf.Int64Value
maxReplicatedMergesInQueue
: google.protobuf.Int64Value
How many tasks of merging and mutating parts are allowed simultaneously in ReplicatedMergeTree queue.
numberOfFreeEntriesInPoolToLowerMaxSizeOfMerge
: google.protobuf.Int64Value
If there is less than specified number of free entries in background pool (or replicated queue), start to lower maximum size of merge to process.
maxBytesToMergeAtMinSpaceInPool
: google.protobuf.Int64Value
Maximum in total size of parts to merge, when there are minimum free threads in background pool (or entries in replication queue).
maxBytesToMergeAtMaxSpaceInPool
: google.protobuf.Int64Value
minBytesForWidePart
: google.protobuf.Int64Value
Minimum number of bytes in a data part that can be stored in Wide format.
More info see in ClickHouse documentation.
minRowsForWidePart
: google.protobuf.Int64Value
Minimum number of rows in a data part that can be stored in Wide format.
More info see in ClickHouse documentation.
ttlOnlyDropParts
: google.protobuf.BoolValue
Enables or disables complete dropping of data parts where all rows are expired in MergeTree tables.
More info see in ClickHouse documentation.
allowRemoteFsZeroCopyReplication
: google.protobuf.BoolValue
mergeWithTtlTimeout
: google.protobuf.Int64Value
mergeWithRecompressionTtlTimeout
: google.protobuf.Int64Value
maxPartsInTotal
: google.protobuf.Int64Value
maxNumberOfMergesWithTtlInPool
: google.protobuf.Int64Value
cleanupDelayPeriod
: google.protobuf.Int64Value
numberOfFreeEntriesInPoolToExecuteMutation
: google.protobuf.Int64Value
maxAvgPartSizeForTooManyParts
: google.protobuf.Int64Value
The 'too many parts' check according to 'parts_to_delay_insert' and 'parts_to_throw_insert' will be active only if the average part size (in the relevant partition) is not larger than the specified threshold. If it is larger than the specified threshold, the INSERTs will be neither delayed or rejected. This allows to have hundreds of terabytes in a single table on a single server if the parts are successfully merged to larger parts. This does not affect the thresholds on inactive parts or total parts. Default: 1 GiB Min version: 22.10 See in-depth description in ClickHouse GitHub
minAgeToForceMergeSeconds
: google.protobuf.Int64Value
Merge parts if every part in the range is older than the value of min_age_to_force_merge_seconds. Default: 0 - disabled Min_version: 22.10 See in-depth description in ClickHouse documentation
minAgeToForceMergeOnPartitionOnly
: google.protobuf.BoolValue
Whether min_age_to_force_merge_seconds should be applied only on the entire partition and not on subset. Default: false Min_version: 22.11 See in-depth description in ClickHouse documentation
mergeSelectingSleepMs
: google.protobuf.Int64Value
Sleep time for merge selecting when no part is selected. A lower setting triggers selecting tasks in background_schedule_pool frequently, which results in a large number of requests to ClickHouse Keeper in large-scale clusters. Default: 5000 Min_version: 21.10 See in-depth description in ClickHouse documentation
mergeMaxBlockSize
: google.protobuf.Int64Value
The number of rows that are read from the merged parts into memory. Default: 8192 See in-depth description in ClickHouse documentation
checkSampleColumnIsCorrect
: google.protobuf.BoolValue
Enables the check at table creation, that the data type of a column for sampling or sampling expression is correct. The data type must be one of unsigned integer types: UInt8, UInt16, UInt32, UInt64. Default: true See in-depth description in ClickHouse documentation
maxMergeSelectingSleepMs
: google.protobuf.Int64Value
Maximum sleep time for merge selecting, a lower setting will trigger selecting tasks in background_schedule_pool frequently which result in large amount of requests to zookeeper in large-scale clusters. Default: 60000 Min_version: 23.6 See in-depth description in ClickHouse GitHub
maxCleanupDelayPeriod
: google.protobuf.Int64Value
Maximum period to clean old queue logs, blocks hashes and parts. Default: 300 Min_version: 23.6 See in-depth description in ClickHouse GitHub
Compression
Method
METHOD_UNSPECIFIED
LZ4
ZSTD
method
: Method
Compression method to use for the specified combination of min_part_size and min_part_size_ratio.
minPartSize
: int64
Minimum size of a part of a table.
minPartSizeRatio
: double
Minimum ratio of a part relative to the size of all the data in the table.
level
: google.protobuf.Int64Value
ExternalDictionary
HttpSource
Header
name
: string
value
: string
url
: string
URL of the source dictionary available over HTTP.
format
: string
The data format. Valid values are all formats supported by ClickHouse SQL dialect.
headers
: Header
HTTP headers.
MysqlSource
Replica
host
: string
MySQL host of the replica.
priority
: int64
The priority of the replica that ClickHouse takes into account when connecting. Replica with the highest priority should have this field set to the lowest number.
port
: int64
Port to use when connecting to the replica. If a port is not specified for a replica, ClickHouse uses the port specified for the source.
user
: string
Name of the MySQL database user.
password
: string
Password of the MySQL database user.
db
: string
Name of the MySQL database to connect to.
table
: string
Name of the database table to use as a ClickHouse dictionary.
port
: int64
Default port to use when connecting to a replica of the dictionary source.
user
: string
Name of the default user for replicas of the dictionary source.
password
: string
Password of the default user for replicas of the dictionary source.
replicas
: Replica
List of MySQL replicas of the database used as dictionary source.
where
: string
Selection criteria for the data in the specified MySQL table.
invalidateQuery
: string
Query for checking the dictionary status, to pull only updated data. For more details, see ClickHouse documentation on dictionaries.
closeConnection
: google.protobuf.BoolValue
Should the connection be closed after each request.
shareConnection
: google.protobuf.BoolValue
Should a connection be shared for some requests.
ClickhouseSource
db
: string
Name of the ClickHouse database.
table
: string
Name of the table in the specified database to be used as the dictionary source.
host
: string
ClickHouse host of the specified database.
port
: int64
Port to use when connecting to the host.
user
: string
Name of the ClickHouse database user.
password
: string
Password of the ClickHouse database user.
where
: string
Selection criteria for the data in the specified ClickHouse table.
secure
: google.protobuf.BoolValue
Use ssl for connection.
MongodbSource
db
: string
Name of the MongoDB database.
collection
: string
Name of the collection in the specified database to be used as the dictionary source.
host
: string
MongoDB host of the specified database.
port
: int64
Port to use when connecting to the host.
user
: string
Name of the MongoDB database user.
password
: string
Password of the MongoDB database user.
options
: string
PostgresqlSource
SslMode
SSL_MODE_UNSPECIFIED
DISABLE
Only try a non-SSL connection.
ALLOW
First try a non-SSL connection; if that fails, try an SSL connection.
PREFER
First try an SSL connection; if that fails, try a non-SSL connection.
VERIFY_CA
Only try an SSL connection, and verify that the server certificate is issued by a trusted certificate authority (CA).
VERIFY_FULL
Only try an SSL connection, verify that the server certificate is issued by a trusted CA and that the requested server host name matches that in the certificate.
db
: string
Name of the PostrgreSQL database.
table
: string
Name of the table in the specified database to be used as the dictionary source.
hosts
: string
Name of the PostrgreSQL host
port
: int64
Port to use when connecting to the host.
user
: string
Name of the PostrgreSQL database user.
password
: string
Password of the PostrgreSQL database user.
invalidateQuery
: string
Query for checking the dictionary status, to pull only updated data. For more details, see ClickHouse documentation on dictionaries.
sslMode
: SslMode
Mode of SSL TCP/IP connection to the PostgreSQL host. For more details, see PostgreSQL documentation.
Structure
Attribute
name
: string
Name of the column.
type
: string
Type of the column.
nullValue
: string
Default value for an element without data (for example, an empty string).
expression
: string
Expression, describing the attribute, if applicable.
hierarchical
: bool
Indication of hierarchy support.
Default value: false
.
injective
: bool
Indication of injective mapping "id -> attribute".
Default value: false
.
Id
Numeric key.
name
: string
Name of the numeric key.
Key
Complex key.
attributes
: Attribute
Attributes of a complex key.
id
: Id
Single numeric key column for the dictionary.
key
: Key
Composite key for the dictionary, containing of one or more key columns. For details, see ClickHouse documentation.
rangeMin
: Attribute
Field holding the beginning of the range for dictionaries with RANGE_HASHED
layout.
For details, see ClickHouse documentation.
rangeMax
: Attribute
Field holding the end of the range for dictionaries with RANGE_HASHED
layout.
For details, see ClickHouse documentation.
attributes
: Attribute
Description of the fields available for database queries. For details, see ClickHouse documentation.
Layout
Layout determining how to store the dictionary in memory.
Type
TYPE_UNSPECIFIED
FLAT
The entire dictionary is stored in memory in the form of flat arrays. Available for all dictionary sources.
HASHED
The entire dictionary is stored in memory in the form of a hash table. Available for all dictionary sources.
COMPLEX_KEY_HASHED
Similar to HASHED, to be used with composite keys. Available for all dictionary sources.
RANGE_HASHED
The entire dictionary is stored in memory in the form of a hash table, with an ordered array of ranges and their corresponding values. Available for all dictionary sources.
CACHE
The dictionary is stored in a cache with a set number of cells. Available for MySQL, ClickHouse and HTTP dictionary sources.
COMPLEX_KEY_CACHE
Similar to CACHE, to be used with composite keys. Available for MySQL, ClickHouse and HTTP dictionary sources.
type
: Type
Layout type for an external dictionary.
sizeInCells
: int64
Number of cells in the cache. Rounded up to a power of two. Applicable only for CACHE and COMPLEX_KEY_CACHE layout types.
maxArraySize
: int64
Maximum dictionary key size. Applicable only for FLAT layout type.
Range
min
: int64
Minimum dictionary lifetime.
max
: int64
Maximum dictionary lifetime.
name
: string
Name of the external dictionary.
structure
: Structure
Set of attributes for the external dictionary. For in-depth description, see ClickHouse documentation.
layout
: Layout
Layout for storing the dictionary in memory. For in-depth description, see ClickHouse documentation.
fixedLifetime
: int64Fixed interval between dictionary updates.
lifetimeRange
: RangeRange of intervals between dictionary updates for ClickHouse to choose from.
httpSource
: HttpSourceHTTP source for the dictionary.
mysqlSource
: MysqlSourceMySQL source for the dictionary.
clickhouseSource
: ClickhouseSourceClickHouse source for the dictionary.
mongodbSource
: MongodbSourceMongoDB source for the dictionary.
postgresqlSource
: PostgresqlSourcePostgreSQL source for the dictionary.
GraphiteRollup
Rollup settings for the GraphiteMergeTree table engine.
Pattern
Retention
age
: int64
Minimum age of the data in seconds.
precision
: int64
Precision of determining the age of the data, in seconds.
regexp
: string
Pattern for metric names.
function
: string
Name of the aggregating function to apply to data of the age specified in retention.
retention
: Retention
Age of data to use for thinning.
name
: string
Name for the specified combination of settings for Graphite rollup.
patterns
: Pattern
Pattern to use for the rollup.
pathColumnName
: string
The name of the column storing the metric name (Graphite sensor). Default: Path See in-depth description in ClickHouse documentation
timeColumnName
: string
The name of the column storing the time of measuring the metric. Default: Time See in-depth description in ClickHouse documentation
valueColumnName
: string
The name of the column storing the value of the metric at the time set in time_column_name. Default: Value See in-depth description in ClickHouse documentation
versionColumnName
: string
The name of the column storing the version of the metric. Default: Timestamp See in-depth description in ClickHouse documentation
KafkaTopic
name
: string
settings
: Kafka
Rabbitmq
username
: string
RabbitMQ username
password
: string
RabbitMQ password
vhost
: string
RabbitMQ virtual host
QueryMaskingRule
name
: string
Name for the rule.
regexp
: string
RE2 compatible regular expression. Required.
replace
: string
Substitution string for sensitive data. Default: six asterisks
QueryCache
maxSizeInBytes
: google.protobuf.Int64Value
The maximum cache size in bytes. Default: 1073741824 (1 GiB)
maxEntries
: google.protobuf.Int64Value
The maximum number of SELECT query results stored in the cache. Default: 1024
maxEntrySizeInBytes
: google.protobuf.Int64Value
The maximum size in bytes SELECT query results may have to be saved in the cache. Dafault: 1048576 (1 MiB)
maxEntrySizeInRows
: google.protobuf.Int64Value
The maximum number of rows SELECT query results may have to be saved in the cache. Default: 30000000 (30 mil)
Operation
An Operation resource. For more information, see Operation.
id
: string
ID of the operation.
description
: string
Description of the operation. 0-256 characters long.
createdAt
: google.protobuf.Timestamp
Creation timestamp.
createdBy
: string
ID of the user or service account who initiated the operation.
modifiedAt
: google.protobuf.Timestamp
The time when the Operation resource was last modified.
done
: bool
If the value is false
, it means the operation is still in progress.
If true
, the operation is completed, and either error
or response
is available.
metadata
: google.protobuf.Any
Service-specific metadata associated with the operation. It typically contains the ID of the target resource that the operation is performed on. Any method that returns a long-running operation should document the metadata type, if any.
One of result
The operation result.
If done == false
and there was no failure detected, neither error
nor response
is set.
If done == false
and there was a failure detected, error
is set.
If done == true
, exactly one of error
or response
is set.
error
: google.rpc.StatusThe error result of the operation in case of failure or cancellation.
response
: google.protobuf.AnyThe normal response of the operation in case of success.
If the original method returns no data on success, such as Delete, the response is google.protobuf.Empty. If the original method is the standard Create/Update, the response should be the target resource of the operation. Any method that returns a long-running operation should document the response type, if any.