Yandex Cloud
  • Сервисы
  • Решения
  • Почему Yandex Cloud
  • Сообщество
  • Тарифы
  • Документация
  • Связаться с нами
Подключиться
Language / Region
© 2022 ООО «Яндекс.Облако»
Yandex Managed Service for ClickHouse
  • Начало работы
  • Пошаговые инструкции
    • Все инструкции
    • Информация об имеющихся кластерах
    • Создание кластера
    • Подключение к базе данных
    • Остановка и запуск кластера
    • SQL-запросы в консоли управления
    • Изменение настроек кластера и базы данных
    • Настройка доступа к ObjectStorage
    • Подключение внешних словарей
    • Подключение собственной геобазы
    • Управление моделями машинного обучения
    • Управление схемами формата данных
    • Подключение к DataLens
    • Изменение версии ClickHouse
    • Управление хостами ZooKeeper
    • Управление хостами ClickHouse
    • Управление базами данных
    • Управление пользователями БД
    • Управление резервными копиями
    • Управление шардами
    • Управление группами шардов
    • Просмотр логов кластера
    • Удаление кластера
    • Мониторинг состояния кластера и хостов
  • Практические руководства
    • Добавление данных в БД
    • Шардирование таблиц
    • Использование гибридного хранилища
    • Получение данных из Managed Service for Apache Kafka®
    • Получение данных из RabbitMQ
    • Обмен данными с Data Proc
    • Настройка Yandex Cloud DNS для доступа к кластеру из других облачных сетей
    • Анализ логов Object Storage при помощи DataLens
    • Настройка кластера под Graphite
  • Концепции
    • Взаимосвязь ресурсов сервиса
    • Классы хостов
      • Действующие классы хостов
      • Архив
        • До 1 июня 2020 года
      • Использование устаревших классов хостов
    • Сеть в Managed Service for ClickHouse
    • Квоты и лимиты
    • Типы хранилища
    • Резервные копии
    • Репликация
    • Словари
    • Шардирование
    • Техническое обслуживание
    • Поддерживаемые клиенты
    • Управление памятью в Managed Service for ClickHouse
    • Политика работы с версиями ClickHouse
    • Настройки ClickHouse
  • Управление доступом
  • Правила тарификации
    • Действующие правила
    • Архив
      • До 1 января 2019 года
      • С 1 января до 1 марта 2019 года
      • С 1 марта 2019 года до 1 февраля 2020 года
  • Справочник API
    • Аутентификация в API
    • gRPC (англ.)
      • Overview
      • BackupService
      • ClusterService
      • DatabaseService
      • FormatSchemaService
      • MlModelService
      • ResourcePresetService
      • UserService
      • VersionsService
      • OperationService
    • REST (англ.)
      • Overview
      • Backup
        • Overview
        • get
        • list
      • Cluster
        • Overview
        • addHosts
        • addShard
        • addZookeeper
        • backup
        • create
        • createExternalDictionary
        • createShardGroup
        • delete
        • deleteExternalDictionary
        • deleteHosts
        • deleteShard
        • deleteShardGroup
        • get
        • getShard
        • getShardGroup
        • list
        • listBackups
        • listHosts
        • listLogs
        • listOperations
        • listShardGroups
        • listShards
        • move
        • rescheduleMaintenance
        • restore
        • start
        • stop
        • streamLogs
        • update
        • updateHosts
        • updateShard
        • updateShardGroup
      • Database
        • Overview
        • create
        • delete
        • get
        • list
      • FormatSchema
        • Overview
        • create
        • delete
        • get
        • list
        • update
      • MlModel
        • Overview
        • create
        • delete
        • get
        • list
        • update
      • ResourcePreset
        • Overview
        • get
        • list
      • User
        • Overview
        • create
        • delete
        • get
        • grantPermission
        • list
        • revokePermission
        • update
      • Versions
        • Overview
        • list
      • Operation
        • Overview
        • get
  • История изменений
  • Вопросы и ответы
    • Общие вопросы
    • Вопросы о ClickHouse
    • Подключение
    • Изменение кластера
    • Настройки параметров кластера
    • Перемещение и восстановление кластера
    • Мониторинг и логи
    • Все вопросы на одной странице
  1. Справочник API
  2. gRPC (англ.)
  3. UserService

UserService

Статья создана
Yandex Cloud
  • Calls UserService
  • Get
    • GetUserRequest
    • User
    • Permission
    • UserSettings
    • UserQuota
  • List
    • ListUsersRequest
    • ListUsersResponse
    • User
    • Permission
    • UserSettings
    • UserQuota
  • Create
    • CreateUserRequest
    • UserSpec
    • Permission
    • UserSettings
    • UserQuota
    • Operation
    • CreateUserMetadata
    • User
    • Permission
    • UserSettings
    • UserQuota
  • Update
    • UpdateUserRequest
    • Permission
    • UserSettings
    • UserQuota
    • Operation
    • UpdateUserMetadata
    • User
    • Permission
    • UserSettings
    • UserQuota
  • Delete
    • DeleteUserRequest
    • Operation
    • DeleteUserMetadata
  • GrantPermission
    • GrantUserPermissionRequest
    • Permission
    • Operation
    • GrantUserPermissionMetadata
    • User
    • Permission
    • UserSettings
    • UserQuota
  • RevokePermission
    • RevokeUserPermissionRequest
    • Operation
    • RevokeUserPermissionMetadata
    • User
    • Permission
    • UserSettings
    • UserQuota

A set of methods for managing ClickHouse User resources.
NOTE: these methods are available only if user management through SQL is disabled.

Call Description
Get Returns the specified ClickHouse User resource.
List Retrieves the list of ClickHouse User resources in the specified cluster.
Create Creates a ClickHouse user in the specified cluster.
Update Updates the specified ClickHouse user.
Delete Deletes the specified ClickHouse user.
GrantPermission Grants a permission to the specified ClickHouse user.
RevokePermission Revokes a permission from the specified ClickHouse user.

Calls UserService

Get

Returns the specified ClickHouse User resource.
To get the list of available ClickHouse User resources, make a List request.

rpc Get (GetUserRequest) returns (User)

GetUserRequest

Field Description
cluster_id string
Required. ID of the ClickHouse cluster the user belongs to. To get the cluster ID, use a ClusterService.List request. The maximum string length in characters is 50.
user_name string
Required. Name of the ClickHouse User resource to return. To get the name of the user, use a UserService.List request. The maximum string length in characters is 63. Value must match the regular expression [a-zA-Z0-9_]*.

User

Field Description
name string
Name of the ClickHouse user.
cluster_id string
ID of the ClickHouse cluster the user belongs to.
permissions[] Permission
Set of permissions granted to the user.
settings UserSettings
quotas[] UserQuota
Set of quotas assigned to the user.

Permission

Field Description
database_name string
Name of the database that the permission grants access to.

UserSettings

Field Description
readonly google.protobuf.Int64Value
Restricts permissions for non-DDL queries. To restrict permissions for DDL queries, use allow_ddl instead.
  • 0 (default)-no restrictions.
  • 1-only read data queries are allowed.
  • 2-read data and change settings queries are allowed.

See in-depth description in ClickHouse documentation. Acceptable values are 0 to 2, inclusive.
allow_ddl google.protobuf.BoolValue
Determines whether DDL queries are allowed (e.g., CREATE, ALTER, RENAME, etc).
Default value: true.
See in-depth description in ClickHouse documentation.
insert_quorum google.protobuf.Int64Value
Enables or disables write quorum for ClickHouse cluster. If the value is less than 2, then write quorum is disabled, otherwise it is enabled.
When used, write quorum guarantees that ClickHouse has written data to the quorum of insert_quorum replicas with no errors until the insert_quorum_timeout expires. All replicas in the quorum are in the consistent state, meaning that they contain linearized data from the previous INSERT queries. Employ write quorum, if you need the guarantees that the written data would not be lost in case of one or more replicas failure.
You can use select_sequential_consistency setting to read the data written with write quorum.
See in-depth description in ClickHouse documentation. The minimum value is 0.
connect_timeout google.protobuf.Int64Value
Connection timeout in milliseconds.
Value must be greater than 0 (default: 10000, 10 seconds). Value must be greater than 0.
receive_timeout google.protobuf.Int64Value
Receive timeout in milliseconds.
Value must be greater than 0 (default: 300000, 300 seconds or 5 minutes). Value must be greater than 0.
send_timeout google.protobuf.Int64Value
Send timeout in milliseconds.
Value must be greater than 0 (default: 300000, 300 seconds or 5 minutes). Value must be greater than 0.
insert_quorum_timeout google.protobuf.Int64Value
Quorum write timeout in milliseconds.
If the write quorum is enabled in the cluster, this timeout expires and some data is not written to the insert_quorum replicas, then ClickHouse will abort the execution of INSERT query and return an error. In this case, the client must send the query again to write the data block into the same or another replica.
Minimum value: 1000, 1 second (default: 60000, 1 minute). The minimum value is 1000.
select_sequential_consistency google.protobuf.BoolValue
Determines the behavior of SELECT queries from the replicated table: if enabled, ClickHouse will terminate a query with error message in case the replica does not have a chunk written with the quorum and will not read the parts that have not yet been written with the quorum.
Default value: false (sequential consistency is disabled).
max_replica_delay_for_distributed_queries google.protobuf.Int64Value
Max replica delay in milliseconds. If a replica lags more than the set value, this replica is not used and becomes a stale one.
Minimum value: 1000, 1 second (default: 300000, 300 seconds or 5 minutes).
See in-depth description in ClickHouse documentation. The minimum value is 1000.
fallback_to_stale_replicas_for_distributed_queries google.protobuf.BoolValue
Enables or disables query forcing to a stale replica in case the actual data is unavailable. If enabled, ClickHouse will choose the most up-to-date replica and force the query to use the data in this replica. This setting can be used when doing SELECT query from a distributed table that points to replicated tables.
Default value: true (query forcing is enabled).
See in-depth description in ClickHouse documentation.
replication_alter_partitions_sync google.protobuf.Int64Value
Wait mode for asynchronous actions in ALTER queries on replicated tables:
  • 0-do not wait for replicas.
  • 1-only wait for own execution (default).
  • 2-wait for all replicas.

See in-depth description in ClickHouse documentation. Acceptable values are 0 to 2, inclusive.
distributed_product_mode enum DistributedProductMode
Determine the behavior of distributed subqueries.
See in-depth description in ClickHouse documentation.
  • DISTRIBUTED_PRODUCT_MODE_DENY: Default value. Prohibits using these types of subqueries (returns the "Double-distributed in/JOIN subqueries is denied" exception).
  • DISTRIBUTED_PRODUCT_MODE_LOCAL: Replaces the database and table in the subquery with local ones for the destination server (shard), leaving the normal IN/JOIN.
  • DISTRIBUTED_PRODUCT_MODE_GLOBAL: Replaces the IN/JOIN query with GLOBAL IN/GLOBAL JOIN.
  • DISTRIBUTED_PRODUCT_MODE_ALLOW: Allows the use of these types of subqueries.
distributed_aggregation_memory_efficient google.protobuf.BoolValue
Enables of disables memory saving mode when doing distributed aggregation.
When ClickHouse works with a distributed query, external aggregation is done on remote servers. Enable this setting to achieve a smaller memory footprint on the server that sourced such a distributed query.
Default value: false (memory saving mode is disabled).
See in-depth description in ClickHouse documentation.
distributed_ddl_task_timeout google.protobuf.Int64Value
Timeout for DDL queries, in milliseconds.
skip_unavailable_shards google.protobuf.BoolValue
Enables or disables silent skipping of unavailable shards.
A shard is considered unavailable if all its replicas are also unavailable.
Default value: false (silent skipping is disabled).
See in-depth description in ClickHouse documentation.
compile google.protobuf.BoolValue
Enables or disables query compilation. If you execute a lot of structurally identical queries, then enable this setting. As a result, such queries may be executed faster due to use of queries' compiled parts.
Use this setting in combination with min_count_to_compile setting.
Default value: false (compilation is disabled).
See in-depth description in ClickHouse documentation.
min_count_to_compile google.protobuf.Int64Value
How many structurally identical queries ClickHouse has to encounter before they are compiled.
Minimum value: 0 (default: 3).
For the 0 value compilation is synchronous: a query waits for compilation process to complete prior to continuing execution. It is recommended to set this value only for testing purposes.
For all other values, compilation is asynchronous: the compilation process executes in a separate thread. When a compiled part of query is ready, it will be used by ClickHouse for eligible queries, including the ones that are currently running.
See in-depth description in ClickHouse documentation. The minimum value is 0.
compile_expressions google.protobuf.BoolValue
Enables or disables expression compilation. If you execute a lot of queries that contain identical expressions, then enable this setting. As a result, such queries may be executed faster due to use of compiled expressions.
Use this setting in combination with min_count_to_compile_expression setting.
Default value: false (expression compilation is disabled).
min_count_to_compile_expression google.protobuf.Int64Value
How many identical expressions ClickHouse has to encounter before they are compiled.
Minimum value: 0 (default: 3).
For the 0 value compilation is synchronous: a query waits for expression compilation process to complete prior to continuing execution. It is recommended to set this value only for testing purposes.
For all other values, compilation is asynchronous: the compilation process executes in a separate thread. When a compiled expression is ready, it will be used by ClickHouse for eligible queries, including the ones that are currently running. The minimum value is 0.
max_block_size google.protobuf.Int64Value
The maximum block size for reading.
Data in ClickHouse is organized and processed by blocks (block is a set of columns' parts). The internal processing cycles for a single block are efficient enough, but there are noticeable expenditures on each block.
This setting is a recommendation for size of block (in a count of rows) that should be loaded from tables.
Value must be greater than 0 (default: 65536).
See in-depth description in ClickHouse documentation. Value must be greater than 0.
min_insert_block_size_rows google.protobuf.Int64Value
Limits the minimum number of rows in a block to be inserted in a table by INSERT query. Blocks that are smaller than the specified value, will be squashed together into the bigger blocks.
Minimal value: 0, block squashing is disabled (default: 1048576). The minimum value is 0.
min_insert_block_size_bytes google.protobuf.Int64Value
Limits the minimum number of bytes in a block to be inserted in a table by INSERT query. Blocks that are smaller than the specified value, will be squashed together into the bigger blocks.
Minimal value: 0, block squashing is disabled (default: 268435456, 256 MB). The minimum value is 0.
max_insert_block_size google.protobuf.Int64Value
Allows to form blocks of the specified size (in bytes) when inserting data in a table. This setting has effect only if server is creating such blocks by itself.
Value must be greater than 0 (default: 1048576).
See in-depth description in ClickHouse documentation. Value must be greater than 0.
min_bytes_to_use_direct_io google.protobuf.Int64Value
Limits the minimum number of bytes to enable unbuffered direct reads from disk (Direct I/O).
By default, ClickHouse does not read data directly from disk, but relies on the filesystem and its cache instead. Such reading strategy is effective when the data volume is small. If the amount of the data to read is huge, it is more effective to read directly from the disk, bypassing the filesystem cache.
If the total amount of the data to read is greater than the value of this setting, then ClickHouse will fetch this data directly from the disk.
Minimal value and default value: 0, Direct I/O is disabled. The minimum value is 0.
use_uncompressed_cache google.protobuf.BoolValue
Determines whether to use the cache of uncompressed blocks, or not. Using this cache can significantly reduce latency and increase the throughput when a huge amount of small queries is to be processed. Enable this setting for the users who instantiates small queries frequently.
This setting has effect only for tables of the MergeTree family.
Default value: false (uncompressed cache is disabled).
See in-depth description in ClickHouse documentation.
merge_tree_max_rows_to_use_cache google.protobuf.Int64Value
Limits the maximum size in rows of the request that can use the cache of uncompressed data. The cache is not used for requests larger than the specified value.
Use this setting in combination with use_uncompressed_cache setting.
Value must be greater than 0 (default: 128x8192). Value must be greater than 0.
merge_tree_max_bytes_to_use_cache google.protobuf.Int64Value
Limits the maximum size in bytes of the request that can use the cache of uncompressed data. The cache is not used for requests larger than the specified value.
Use this setting in combination with use_uncompressed_cache setting.
Value must be greater than 0 (default: 192x10x1024x1024). Value must be greater than 0.
merge_tree_min_rows_for_concurrent_read google.protobuf.Int64Value
Limits the minimum number of rows to be read from a file to enable concurrent read. If the number of rows to be read exceeds this value, then ClickHouse will try to use a few threads to read from a file concurrently.
This setting has effect only for tables of the MergeTree family.
Value must be greater than 0 (default: 20x8192). Value must be greater than 0.
merge_tree_min_bytes_for_concurrent_read google.protobuf.Int64Value
Limits the number of bytes to be read from a file to enable concurrent read. If the number of bytes to be read exceeds this value, then ClickHouse will try to use a few threads to read from a file concurrently.
This setting has effect only for tables of the MergeTree family.
Value must be greater than 0 (default: 24x10x1024x1024). Value must be greater than 0.
max_bytes_before_external_group_by google.protobuf.Int64Value
Sets the threshold of RAM consumption (in bytes) after that the temporary data, collected during the GROUP BY operation, should be flushed to disk to limit the RAM comsumption.
By default, aggregation is done by employing hash table that resides in RAM. A query can result in aggregation of huge data volumes that can lead to memory exhaustion and abortion of the query (see the max_memory_usage setting). For such queries, you can use this setting to force ClickHouse to do flushing and complete aggregation successfully.
Minimal value and default value: 0, GROUP BY in the external memory is disabled.
When using aggregation in external memory, it is recommended to set the value of this setting twice as low as the max_memory_usage setting value (by default, the maximum memory usage is limited to ten gigabytes).
See in-depth description in ClickHouse documentation.
See also: the distributed_aggregation_memory_efficient setting.
max_bytes_before_external_sort google.protobuf.Int64Value
This setting is equivalent of the max_bytes_before_external_group_by setting, except for it is for sort operation (ORDER BY), not aggregation.
group_by_two_level_threshold google.protobuf.Int64Value
Sets the threshold of the number of keys, after that the two-level aggregation should be used.
Minimal value: 0, threshold is not set (default: 10000).
group_by_two_level_threshold_bytes google.protobuf.Int64Value
Sets the threshold of the number of bytes, after that the two-level aggregation should be used.
Minimal value: 0, threshold is not set (default: 100000000).
priority google.protobuf.Int64Value
Sets the priority of a query.
  • 0-priority is not used.
  • 1-the highest priority.
  • and so on. The higher the number, the lower a query's priority.

This setting should be set up for each query individually.
If ClickHouse is working with the high-priority queries, and a low-priority query enters, then the low-priority query is paused until higher-priority queries are completed.
Minimal value and default value: 0, priority is not used. The minimum value is 0.
max_threads google.protobuf.Int64Value
Limits the maximum number of threads to process the request (setting does not take threads that read data from remote servers into account).
This setting applies to threads that perform the same stages of the query processing pipeline in parallel.
Minimal value and default value: 0 (the thread number is calculated automatically based on the number of physical CPU cores, no HyperThreading cores are taken into account).
See in-depth description in ClickHouse documentation. Value must be greater than 0.
max_memory_usage google.protobuf.Int64Value
Limits the maximum memory usage (in bytes) for processing of a single user's query on a single server. This setting does not take server's free RAM amount or total RAM amount into account.
This limitation is enforced for any user's single query on a single server.
Minimal value: 0, no limitation is set. Value that is set in the ClickHouse default config file: 10737418240 (10 GB).
If you use max_bytes_before_external_group_by or max_bytes_before_external_sort setting, then it is recommended to set their values twice as low as max_memory_usage setting value.
See in-depth description in ClickHouse documentation. The minimum value is 0.
max_memory_usage_for_user google.protobuf.Int64Value
Limits the maximum memory usage (in bytes) for processing of user's queries on a single server. This setting does not take server's free RAM amount or total RAM amount into account.
This limitation is enforced for all queries that belong to one user and run simultaneously on a single server.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_network_bandwidth google.protobuf.Int64Value
The maximum speed of data exchange over the network in bytes per second for a query.
Minimal value and default value: 0, no limitation is set.
max_network_bandwidth_for_user google.protobuf.Int64Value
The maximum speed of data exchange over the network in bytes per second for all concurrently running user queries.
Minimal value and default value: 0, no limitation is set.
force_index_by_date google.protobuf.BoolValue
If enabled, query is not executed if the ClickHouse can't use index by date. This setting has effect only for tables of the MergeTree family.
Default value: false (setting is disabled, query executes even if ClickHouse can't use index by date).
See in-depth description in ClickHouse documentation.
force_primary_key google.protobuf.BoolValue
If enabled, query is not executed if the ClickHouse can't use index by primary key. This setting has effect only for tables of the MergeTree family.
Default value: false (setting is disabled, query executes even if ClickHouse can't use index by primary key).
See in-depth description in ClickHouse documentation.
max_rows_to_read google.protobuf.Int64Value
Limits the maximum number of rows that can be read from a table when running a query.
Minimal value and default value: 0, no limitation is set.
See in-depth description in ClickHouse documentation. The minimum value is 0.
max_bytes_to_read google.protobuf.Int64Value
Limits the maximum number of bytes (uncompressed data) that can be read from a table when running a query.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
read_overflow_mode enum OverflowMode
Determines the behavior on exceeding limits while reading the data.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
max_rows_to_group_by google.protobuf.Int64Value
Limits the maximum number of unique keys received from aggregation function. This setting helps to reduce RAM consumption while doing aggregation.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
group_by_overflow_mode enum GroupByOverflowMode
Determines the behavior on exceeding limits while doing aggregation.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
  • any-perform approximate GROUP BY operation by continuing aggregation for the keys that got into the set, but don't add new keys to the set.
max_rows_to_sort google.protobuf.Int64Value
Limits the maximum number of rows that can be read from a table for sorting. This setting helps to reduce RAM consumption.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_bytes_to_sort google.protobuf.Int64Value
Limits the maximum number of bytes (uncompressed data) that can be read from a table for sorting. This setting helps to reduce RAM consumption.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
sort_overflow_mode enum OverflowMode
Determines the behavior on exceeding limits while sorting.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
max_result_rows google.protobuf.Int64Value
Limits the number of rows in the result. This limitation is also checked for subqueries and parts of distributed queries that run on remote servers.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_result_bytes google.protobuf.Int64Value
Limits the number of bytes in the result. This limitation is also checked for subqueries and parts of distributed queries that run on remote servers.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
result_overflow_mode enum OverflowMode
Determines the behavior on exceeding limits while forming result.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
max_rows_in_distinct google.protobuf.Int64Value
Limits the maximum number of different rows when using DISTINCT.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_bytes_in_distinct google.protobuf.Int64Value
Limits the maximum size of a hash table in bytes (uncompressed data) when using DISTINCT. The minimum value is 0.
distinct_overflow_mode enum OverflowMode
Determines the behavior on exceeding limits while doing DISCTINCT.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
max_rows_to_transfer google.protobuf.Int64Value
Limits the maximum number of rows that can be passed to a remote server or saved in a temporary table when using GLOBAL IN.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_bytes_to_transfer google.protobuf.Int64Value
Limits the maximum number of bytes (uncompressed data) that can be passed to a remote server or saved in a temporary table when using GLOBAL IN.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
transfer_overflow_mode enum OverflowMode
Determines the behavior on exceeding limits while doing transfers.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
max_execution_time google.protobuf.Int64Value
Limits the maximum query execution time in milliseconds. At this moment, this limitation is not checked when passing one of the sorting stages, as well as merging and finalizing aggregation funictions.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
timeout_overflow_mode enum OverflowMode
Determines the behavior on exceeding limits of execution time.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
max_rows_in_set google.protobuf.Int64Value
Limit on the number of rows in the set resulting from the execution of the IN section. The minimum value is 0.
max_bytes_in_set google.protobuf.Int64Value
Limit on the number of bytes in the set resulting from the execution of the IN section. The minimum value is 0.
set_overflow_mode enum OverflowMode
Determine the behavior on exceeding max_rows_in_set or max_bytes_in_set limit. Possible values: OVERFLOW_MODE_THROW, OVERFLOW_MODE_BREAK.
max_rows_in_join google.protobuf.Int64Value
Limit on maximum size of the hash table for JOIN, in rows. The minimum value is 0.
max_bytes_in_join google.protobuf.Int64Value
Limit on maximum size of the hash table for JOIN, in bytes. The minimum value is 0.
join_overflow_mode enum OverflowMode
Determine the behavior on exceeding max_rows_in_join or max_bytes_in_join limit. Possible values: OVERFLOW_MODE_THROW, OVERFLOW_MODE_BREAK.
max_columns_to_read google.protobuf.Int64Value
Limits the maximum number of columns that can be read from a table in a single query. If the query requires to read more columns to complete, then it will be aborted.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_temporary_columns google.protobuf.Int64Value
Limits the maximum number of temporary columns that must be kept in RAM at the same time when running a query, including constant columns.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_temporary_non_const_columns google.protobuf.Int64Value
Limits the maximum number of temporary columns that must be kept in RAM at the same time when running a query, excluding constant columns.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_query_size google.protobuf.Int64Value
Limits the size of the part of a query that can be transferred to RAM for parsing with the SQL parser, in bytes.
Value must be greater than 0 (default: 262144).
See in-depth description in ClickHouse documentation. Value must be greater than 0.
max_ast_depth google.protobuf.Int64Value
Limits the maximum depth of query syntax tree.
Executing a big and complex query may result in building a syntax tree of enormous depth. By using this setting, you can prohibit execution of over-sized or non-optimized queries for huge tables.
For example, the **SELECT *** query may result in more complex and deeper syntax tree, compared to the SELECT ... WHERE ... query, containing constraints and conditions, in the most cases. A user can be forced to construct more optimized queries, if this setting is used.
Value must be greater than 0 (default: 1000). If a too small value is set, it may render ClickHouse unable to execute even simple queries.
See in-depth description in ClickHouse documentation. Value must be greater than 0.
max_ast_elements google.protobuf.Int64Value
Limits the maximum size of query syntax tree in number of nodes.
Executing a big and complex query may result in building a syntax tree of enormous size. By using this setting, you can prohibit execution of over-sized or non-optimized queries for huge tables.
Value must be greater than 0 (default: 50000). If a too small value is set, it may render ClickHouse unable to execute even simple queries.
See in-depth description in ClickHouse documentation. Value must be greater than 0.
max_expanded_ast_elements google.protobuf.Int64Value
Limits the maximum size of query syntax tree in number of nodes after expansion of aliases and the asterisk values.
Executing a big and complex query may result in building a syntax tree of enormous size. By using this setting, you can prohibit execution of over-sized or non-optimized queries for huge tables.
Value must be greater than 0 (default: 500000). If a too small value is set, it may render ClickHouse unable to execute even simple queries. Value must be greater than 0.
min_execution_speed google.protobuf.Int64Value
Minimal execution speed in rows per second. The minimum value is 0.
min_execution_speed_bytes google.protobuf.Int64Value
Minimal execution speed in bytes per second. The minimum value is 0.
count_distinct_implementation enum CountDistinctImplementation
Aggregate function to use for implementation of count(DISTINCT ...).
input_format_values_interpret_expressions google.protobuf.BoolValue
Enables or disables SQL parser if the fast stream parser cannot parse the data.
Enable this setting, if the data that you want to insert into a table contains SQL expressions.
For example, the stream parser is unable to parse a value that contains now() expression; therefore an INSERT query for this value will fail and no data will be inserted into a table. With enabled SQL parser, this expression is parsed correctly: the now() expression will be parsed as SQL function, interpreted, and the current date and time will be inserted into the table as a result.
This setting has effect only if you use Values format when inserting data.
Default value: true (SQL parser is enabled).
See in-depth description in ClickHouse documentation.
input_format_defaults_for_omitted_fields google.protobuf.BoolValue
Enables or disables replacing omitted input values with default values of the respective columns when performing INSERT queries.
Default value: true (replacing is enabled).
output_format_json_quote_64bit_integers google.protobuf.BoolValue
Enables quoting of 64-bit integers in JSON output format.
If this setting is enabled, then 64-bit integers (UInt64 and Int64) will be quoted when written to JSON output in order to maintain compatibility with the most of the JavaScript engines. Otherwise, such integers will not be quoted.
Default value: false (quoting 64-bit integers is disabled).
output_format_json_quote_denormals google.protobuf.BoolValue
Enables special floating-point values (+nan, -nan, +inf and -inf) in JSON output format.
Default value: false (special values do not present in output).
low_cardinality_allow_in_native_format google.protobuf.BoolValue
Determines whether to use LowCardinality type in Native format.
  • true (default)-yes, use.
  • false-convert LowCardinality columns to regular columns when doing SELECT, and convert regular columns to LowCardinality when doing INSERT.

LowCardinality columns (aka sparse columns) store data in more effective way, compared to regular columns, by using hash tables. If data to insert suits this storage format, ClickHouse will place them into LowCardinality column.
If you use a third-party ClickHouse client that can't work with LowCardinality columns, then this client will not be able to correctly interpret the result of the query that asks for data stored in LowCardinality column. Disable this setting to convert LowCardinality column to regular column when creating the result, so such clients will be able to process the result.
Official ClickHouse client works with LowCardinality columns out-of-the-box.
Default value: true (LowCardinality columns are used in Native format).
empty_result_for_aggregation_by_empty_set google.protobuf.BoolValue
Enables returning of empty result when aggregating without keys (with GROUP BY operation absent) on empty set (e.g., SELECT count(*) FROM table WHERE 0).
  • true-ClickHouse will return an empty result for such queries.
  • false (default)-ClickHouse will return a single-line result consisting of NULL values for aggregation functions, in accordance with SQL standard.
joined_subquery_requires_alias google.protobuf.BoolValue
join_use_nulls google.protobuf.BoolValue
transform_null_in google.protobuf.BoolValue
http_connection_timeout google.protobuf.Int64Value
HTTP connection timeout, in milliseconds.
Value must be greater than 0 (default: 1000, 1 second).
http_receive_timeout google.protobuf.Int64Value
HTTP receive timeout, in milliseconds.
Value must be greater than 0 (default: 1800000, 1800 seconds, 30 minutes).
http_send_timeout google.protobuf.Int64Value
HTTP send timeout, in milliseconds.
Value must be greater than 0 (default: 1800000, 1800 seconds, 30 minutes).
enable_http_compression google.protobuf.BoolValue
Enables or disables data compression in HTTP responses.
By default, ClickHouse stores data compressed. When executing a query, its result is uncompressed. Use this setting to command ClickHouse to compress the result when sending it via HTTP.
Enable this setting and add the Accept-Encoding: HTTP header in a HTTP request to force compression of HTTP response from ClickHouse.
ClickHouse support the following compression methods: gzip, br and deflate.
Default value: false (compression is disabled).
See in-depth description in ClickHouse documentation.
send_progress_in_http_headers google.protobuf.BoolValue
Enables progress notifications using X-ClickHouse-Progress HTTP header.
Default value: false (notifications disabled).
http_headers_progress_interval google.protobuf.Int64Value
Minimum interval between progress notifications with X-ClickHouse-Progress HTTP header, in milliseconds.
Value must be greater than 0 (default: 100).
add_http_cors_header google.protobuf.BoolValue
Adds CORS header in HTTP responses.
Default value: false (header is not added).
quota_mode enum QuotaMode
Quota accounting mode. Possible values: QUOTA_MODE_DEFAULT, QUOTA_MODE_KEYED and QUOTA_MODE_KEYED_BY_IP.

UserQuota

Field Description
interval_duration google.protobuf.Int64Value
Duration of interval for quota in milliseconds. Minimal value is 1 second. The minimum value is 1000.
queries google.protobuf.Int64Value
The total number of queries. 0 - unlimited. The minimum value is 0.
errors google.protobuf.Int64Value
The number of queries that threw exception. 0 - unlimited. The minimum value is 0.
result_rows google.protobuf.Int64Value
The total number of rows given as the result.. 0 - unlimited. The minimum value is 0.
read_rows google.protobuf.Int64Value
The total number of source rows read from tables for running the query, on all remote servers. 0 - unlimited. The minimum value is 0.
execution_time google.protobuf.Int64Value
The total query execution time, in milliseconds (wall time). 0 - unlimited. The minimum value is 0.

List

Retrieves the list of ClickHouse User resources in the specified cluster.

rpc List (ListUsersRequest) returns (ListUsersResponse)

ListUsersRequest

Field Description
cluster_id string
Required. ID of the cluster to list ClickHouse users in. To get the cluster ID, use a ClusterService.List request. The maximum string length in characters is 50.
page_size int64
The maximum number of results per page to return. If the number of available results is larger than page_size, the service returns a ListUsersResponse.next_page_token that can be used to get the next page of results in subsequent list requests. The maximum value is 1000.
page_token string
Page token. To get the next page of results, set page_token to the ListUsersResponse.next_page_token returned by a previous list request. The maximum string length in characters is 100.

ListUsersResponse

Field Description
users[] User
List of ClickHouse User resources.
next_page_token string
This token allows you to get the next page of results for list requests. If the number of results is larger than ListUsersRequest.page_size, use the next_page_token as the value for the ListUsersRequest.page_token parameter in the next list request. Each subsequent list request will have its own next_page_token to continue paging through the results.

User

Field Description
name string
Name of the ClickHouse user.
cluster_id string
ID of the ClickHouse cluster the user belongs to.
permissions[] Permission
Set of permissions granted to the user.
settings UserSettings
quotas[] UserQuota
Set of quotas assigned to the user.

Permission

Field Description
database_name string
Name of the database that the permission grants access to.

UserSettings

Field Description
readonly google.protobuf.Int64Value
Restricts permissions for non-DDL queries. To restrict permissions for DDL queries, use allow_ddl instead.
  • 0 (default)-no restrictions.
  • 1-only read data queries are allowed.
  • 2-read data and change settings queries are allowed.

See in-depth description in ClickHouse documentation. Acceptable values are 0 to 2, inclusive.
allow_ddl google.protobuf.BoolValue
Determines whether DDL queries are allowed (e.g., CREATE, ALTER, RENAME, etc).
Default value: true.
See in-depth description in ClickHouse documentation.
insert_quorum google.protobuf.Int64Value
Enables or disables write quorum for ClickHouse cluster. If the value is less than 2, then write quorum is disabled, otherwise it is enabled.
When used, write quorum guarantees that ClickHouse has written data to the quorum of insert_quorum replicas with no errors until the insert_quorum_timeout expires. All replicas in the quorum are in the consistent state, meaning that they contain linearized data from the previous INSERT queries. Employ write quorum, if you need the guarantees that the written data would not be lost in case of one or more replicas failure.
You can use select_sequential_consistency setting to read the data written with write quorum.
See in-depth description in ClickHouse documentation. The minimum value is 0.
connect_timeout google.protobuf.Int64Value
Connection timeout in milliseconds.
Value must be greater than 0 (default: 10000, 10 seconds). Value must be greater than 0.
receive_timeout google.protobuf.Int64Value
Receive timeout in milliseconds.
Value must be greater than 0 (default: 300000, 300 seconds or 5 minutes). Value must be greater than 0.
send_timeout google.protobuf.Int64Value
Send timeout in milliseconds.
Value must be greater than 0 (default: 300000, 300 seconds or 5 minutes). Value must be greater than 0.
insert_quorum_timeout google.protobuf.Int64Value
Quorum write timeout in milliseconds.
If the write quorum is enabled in the cluster, this timeout expires and some data is not written to the insert_quorum replicas, then ClickHouse will abort the execution of INSERT query and return an error. In this case, the client must send the query again to write the data block into the same or another replica.
Minimum value: 1000, 1 second (default: 60000, 1 minute). The minimum value is 1000.
select_sequential_consistency google.protobuf.BoolValue
Determines the behavior of SELECT queries from the replicated table: if enabled, ClickHouse will terminate a query with error message in case the replica does not have a chunk written with the quorum and will not read the parts that have not yet been written with the quorum.
Default value: false (sequential consistency is disabled).
max_replica_delay_for_distributed_queries google.protobuf.Int64Value
Max replica delay in milliseconds. If a replica lags more than the set value, this replica is not used and becomes a stale one.
Minimum value: 1000, 1 second (default: 300000, 300 seconds or 5 minutes).
See in-depth description in ClickHouse documentation. The minimum value is 1000.
fallback_to_stale_replicas_for_distributed_queries google.protobuf.BoolValue
Enables or disables query forcing to a stale replica in case the actual data is unavailable. If enabled, ClickHouse will choose the most up-to-date replica and force the query to use the data in this replica. This setting can be used when doing SELECT query from a distributed table that points to replicated tables.
Default value: true (query forcing is enabled).
See in-depth description in ClickHouse documentation.
replication_alter_partitions_sync google.protobuf.Int64Value
Wait mode for asynchronous actions in ALTER queries on replicated tables:
  • 0-do not wait for replicas.
  • 1-only wait for own execution (default).
  • 2-wait for all replicas.

See in-depth description in ClickHouse documentation. Acceptable values are 0 to 2, inclusive.
distributed_product_mode enum DistributedProductMode
Determine the behavior of distributed subqueries.
See in-depth description in ClickHouse documentation.
  • DISTRIBUTED_PRODUCT_MODE_DENY: Default value. Prohibits using these types of subqueries (returns the "Double-distributed in/JOIN subqueries is denied" exception).
  • DISTRIBUTED_PRODUCT_MODE_LOCAL: Replaces the database and table in the subquery with local ones for the destination server (shard), leaving the normal IN/JOIN.
  • DISTRIBUTED_PRODUCT_MODE_GLOBAL: Replaces the IN/JOIN query with GLOBAL IN/GLOBAL JOIN.
  • DISTRIBUTED_PRODUCT_MODE_ALLOW: Allows the use of these types of subqueries.
distributed_aggregation_memory_efficient google.protobuf.BoolValue
Enables of disables memory saving mode when doing distributed aggregation.
When ClickHouse works with a distributed query, external aggregation is done on remote servers. Enable this setting to achieve a smaller memory footprint on the server that sourced such a distributed query.
Default value: false (memory saving mode is disabled).
See in-depth description in ClickHouse documentation.
distributed_ddl_task_timeout google.protobuf.Int64Value
Timeout for DDL queries, in milliseconds.
skip_unavailable_shards google.protobuf.BoolValue
Enables or disables silent skipping of unavailable shards.
A shard is considered unavailable if all its replicas are also unavailable.
Default value: false (silent skipping is disabled).
See in-depth description in ClickHouse documentation.
compile google.protobuf.BoolValue
Enables or disables query compilation. If you execute a lot of structurally identical queries, then enable this setting. As a result, such queries may be executed faster due to use of queries' compiled parts.
Use this setting in combination with min_count_to_compile setting.
Default value: false (compilation is disabled).
See in-depth description in ClickHouse documentation.
min_count_to_compile google.protobuf.Int64Value
How many structurally identical queries ClickHouse has to encounter before they are compiled.
Minimum value: 0 (default: 3).
For the 0 value compilation is synchronous: a query waits for compilation process to complete prior to continuing execution. It is recommended to set this value only for testing purposes.
For all other values, compilation is asynchronous: the compilation process executes in a separate thread. When a compiled part of query is ready, it will be used by ClickHouse for eligible queries, including the ones that are currently running.
See in-depth description in ClickHouse documentation. The minimum value is 0.
compile_expressions google.protobuf.BoolValue
Enables or disables expression compilation. If you execute a lot of queries that contain identical expressions, then enable this setting. As a result, such queries may be executed faster due to use of compiled expressions.
Use this setting in combination with min_count_to_compile_expression setting.
Default value: false (expression compilation is disabled).
min_count_to_compile_expression google.protobuf.Int64Value
How many identical expressions ClickHouse has to encounter before they are compiled.
Minimum value: 0 (default: 3).
For the 0 value compilation is synchronous: a query waits for expression compilation process to complete prior to continuing execution. It is recommended to set this value only for testing purposes.
For all other values, compilation is asynchronous: the compilation process executes in a separate thread. When a compiled expression is ready, it will be used by ClickHouse for eligible queries, including the ones that are currently running. The minimum value is 0.
max_block_size google.protobuf.Int64Value
The maximum block size for reading.
Data in ClickHouse is organized and processed by blocks (block is a set of columns' parts). The internal processing cycles for a single block are efficient enough, but there are noticeable expenditures on each block.
This setting is a recommendation for size of block (in a count of rows) that should be loaded from tables.
Value must be greater than 0 (default: 65536).
See in-depth description in ClickHouse documentation. Value must be greater than 0.
min_insert_block_size_rows google.protobuf.Int64Value
Limits the minimum number of rows in a block to be inserted in a table by INSERT query. Blocks that are smaller than the specified value, will be squashed together into the bigger blocks.
Minimal value: 0, block squashing is disabled (default: 1048576). The minimum value is 0.
min_insert_block_size_bytes google.protobuf.Int64Value
Limits the minimum number of bytes in a block to be inserted in a table by INSERT query. Blocks that are smaller than the specified value, will be squashed together into the bigger blocks.
Minimal value: 0, block squashing is disabled (default: 268435456, 256 MB). The minimum value is 0.
max_insert_block_size google.protobuf.Int64Value
Allows to form blocks of the specified size (in bytes) when inserting data in a table. This setting has effect only if server is creating such blocks by itself.
Value must be greater than 0 (default: 1048576).
See in-depth description in ClickHouse documentation. Value must be greater than 0.
min_bytes_to_use_direct_io google.protobuf.Int64Value
Limits the minimum number of bytes to enable unbuffered direct reads from disk (Direct I/O).
By default, ClickHouse does not read data directly from disk, but relies on the filesystem and its cache instead. Such reading strategy is effective when the data volume is small. If the amount of the data to read is huge, it is more effective to read directly from the disk, bypassing the filesystem cache.
If the total amount of the data to read is greater than the value of this setting, then ClickHouse will fetch this data directly from the disk.
Minimal value and default value: 0, Direct I/O is disabled. The minimum value is 0.
use_uncompressed_cache google.protobuf.BoolValue
Determines whether to use the cache of uncompressed blocks, or not. Using this cache can significantly reduce latency and increase the throughput when a huge amount of small queries is to be processed. Enable this setting for the users who instantiates small queries frequently.
This setting has effect only for tables of the MergeTree family.
Default value: false (uncompressed cache is disabled).
See in-depth description in ClickHouse documentation.
merge_tree_max_rows_to_use_cache google.protobuf.Int64Value
Limits the maximum size in rows of the request that can use the cache of uncompressed data. The cache is not used for requests larger than the specified value.
Use this setting in combination with use_uncompressed_cache setting.
Value must be greater than 0 (default: 128x8192). Value must be greater than 0.
merge_tree_max_bytes_to_use_cache google.protobuf.Int64Value
Limits the maximum size in bytes of the request that can use the cache of uncompressed data. The cache is not used for requests larger than the specified value.
Use this setting in combination with use_uncompressed_cache setting.
Value must be greater than 0 (default: 192x10x1024x1024). Value must be greater than 0.
merge_tree_min_rows_for_concurrent_read google.protobuf.Int64Value
Limits the minimum number of rows to be read from a file to enable concurrent read. If the number of rows to be read exceeds this value, then ClickHouse will try to use a few threads to read from a file concurrently.
This setting has effect only for tables of the MergeTree family.
Value must be greater than 0 (default: 20x8192). Value must be greater than 0.
merge_tree_min_bytes_for_concurrent_read google.protobuf.Int64Value
Limits the number of bytes to be read from a file to enable concurrent read. If the number of bytes to be read exceeds this value, then ClickHouse will try to use a few threads to read from a file concurrently.
This setting has effect only for tables of the MergeTree family.
Value must be greater than 0 (default: 24x10x1024x1024). Value must be greater than 0.
max_bytes_before_external_group_by google.protobuf.Int64Value
Sets the threshold of RAM consumption (in bytes) after that the temporary data, collected during the GROUP BY operation, should be flushed to disk to limit the RAM comsumption.
By default, aggregation is done by employing hash table that resides in RAM. A query can result in aggregation of huge data volumes that can lead to memory exhaustion and abortion of the query (see the max_memory_usage setting). For such queries, you can use this setting to force ClickHouse to do flushing and complete aggregation successfully.
Minimal value and default value: 0, GROUP BY in the external memory is disabled.
When using aggregation in external memory, it is recommended to set the value of this setting twice as low as the max_memory_usage setting value (by default, the maximum memory usage is limited to ten gigabytes).
See in-depth description in ClickHouse documentation.
See also: the distributed_aggregation_memory_efficient setting.
max_bytes_before_external_sort google.protobuf.Int64Value
This setting is equivalent of the max_bytes_before_external_group_by setting, except for it is for sort operation (ORDER BY), not aggregation.
group_by_two_level_threshold google.protobuf.Int64Value
Sets the threshold of the number of keys, after that the two-level aggregation should be used.
Minimal value: 0, threshold is not set (default: 10000).
group_by_two_level_threshold_bytes google.protobuf.Int64Value
Sets the threshold of the number of bytes, after that the two-level aggregation should be used.
Minimal value: 0, threshold is not set (default: 100000000).
priority google.protobuf.Int64Value
Sets the priority of a query.
  • 0-priority is not used.
  • 1-the highest priority.
  • and so on. The higher the number, the lower a query's priority.

This setting should be set up for each query individually.
If ClickHouse is working with the high-priority queries, and a low-priority query enters, then the low-priority query is paused until higher-priority queries are completed.
Minimal value and default value: 0, priority is not used. The minimum value is 0.
max_threads google.protobuf.Int64Value
Limits the maximum number of threads to process the request (setting does not take threads that read data from remote servers into account).
This setting applies to threads that perform the same stages of the query processing pipeline in parallel.
Minimal value and default value: 0 (the thread number is calculated automatically based on the number of physical CPU cores, no HyperThreading cores are taken into account).
See in-depth description in ClickHouse documentation. Value must be greater than 0.
max_memory_usage google.protobuf.Int64Value
Limits the maximum memory usage (in bytes) for processing of a single user's query on a single server. This setting does not take server's free RAM amount or total RAM amount into account.
This limitation is enforced for any user's single query on a single server.
Minimal value: 0, no limitation is set. Value that is set in the ClickHouse default config file: 10737418240 (10 GB).
If you use max_bytes_before_external_group_by or max_bytes_before_external_sort setting, then it is recommended to set their values twice as low as max_memory_usage setting value.
See in-depth description in ClickHouse documentation. The minimum value is 0.
max_memory_usage_for_user google.protobuf.Int64Value
Limits the maximum memory usage (in bytes) for processing of user's queries on a single server. This setting does not take server's free RAM amount or total RAM amount into account.
This limitation is enforced for all queries that belong to one user and run simultaneously on a single server.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_network_bandwidth google.protobuf.Int64Value
The maximum speed of data exchange over the network in bytes per second for a query.
Minimal value and default value: 0, no limitation is set.
max_network_bandwidth_for_user google.protobuf.Int64Value
The maximum speed of data exchange over the network in bytes per second for all concurrently running user queries.
Minimal value and default value: 0, no limitation is set.
force_index_by_date google.protobuf.BoolValue
If enabled, query is not executed if the ClickHouse can't use index by date. This setting has effect only for tables of the MergeTree family.
Default value: false (setting is disabled, query executes even if ClickHouse can't use index by date).
See in-depth description in ClickHouse documentation.
force_primary_key google.protobuf.BoolValue
If enabled, query is not executed if the ClickHouse can't use index by primary key. This setting has effect only for tables of the MergeTree family.
Default value: false (setting is disabled, query executes even if ClickHouse can't use index by primary key).
See in-depth description in ClickHouse documentation.
max_rows_to_read google.protobuf.Int64Value
Limits the maximum number of rows that can be read from a table when running a query.
Minimal value and default value: 0, no limitation is set.
See in-depth description in ClickHouse documentation. The minimum value is 0.
max_bytes_to_read google.protobuf.Int64Value
Limits the maximum number of bytes (uncompressed data) that can be read from a table when running a query.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
read_overflow_mode enum OverflowMode
Determines the behavior on exceeding limits while reading the data.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
max_rows_to_group_by google.protobuf.Int64Value
Limits the maximum number of unique keys received from aggregation function. This setting helps to reduce RAM consumption while doing aggregation.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
group_by_overflow_mode enum GroupByOverflowMode
Determines the behavior on exceeding limits while doing aggregation.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
  • any-perform approximate GROUP BY operation by continuing aggregation for the keys that got into the set, but don't add new keys to the set.
max_rows_to_sort google.protobuf.Int64Value
Limits the maximum number of rows that can be read from a table for sorting. This setting helps to reduce RAM consumption.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_bytes_to_sort google.protobuf.Int64Value
Limits the maximum number of bytes (uncompressed data) that can be read from a table for sorting. This setting helps to reduce RAM consumption.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
sort_overflow_mode enum OverflowMode
Determines the behavior on exceeding limits while sorting.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
max_result_rows google.protobuf.Int64Value
Limits the number of rows in the result. This limitation is also checked for subqueries and parts of distributed queries that run on remote servers.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_result_bytes google.protobuf.Int64Value
Limits the number of bytes in the result. This limitation is also checked for subqueries and parts of distributed queries that run on remote servers.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
result_overflow_mode enum OverflowMode
Determines the behavior on exceeding limits while forming result.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
max_rows_in_distinct google.protobuf.Int64Value
Limits the maximum number of different rows when using DISTINCT.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_bytes_in_distinct google.protobuf.Int64Value
Limits the maximum size of a hash table in bytes (uncompressed data) when using DISTINCT. The minimum value is 0.
distinct_overflow_mode enum OverflowMode
Determines the behavior on exceeding limits while doing DISCTINCT.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
max_rows_to_transfer google.protobuf.Int64Value
Limits the maximum number of rows that can be passed to a remote server or saved in a temporary table when using GLOBAL IN.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_bytes_to_transfer google.protobuf.Int64Value
Limits the maximum number of bytes (uncompressed data) that can be passed to a remote server or saved in a temporary table when using GLOBAL IN.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
transfer_overflow_mode enum OverflowMode
Determines the behavior on exceeding limits while doing transfers.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
max_execution_time google.protobuf.Int64Value
Limits the maximum query execution time in milliseconds. At this moment, this limitation is not checked when passing one of the sorting stages, as well as merging and finalizing aggregation funictions.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
timeout_overflow_mode enum OverflowMode
Determines the behavior on exceeding limits of execution time.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
max_rows_in_set google.protobuf.Int64Value
Limit on the number of rows in the set resulting from the execution of the IN section. The minimum value is 0.
max_bytes_in_set google.protobuf.Int64Value
Limit on the number of bytes in the set resulting from the execution of the IN section. The minimum value is 0.
set_overflow_mode enum OverflowMode
Determine the behavior on exceeding max_rows_in_set or max_bytes_in_set limit. Possible values: OVERFLOW_MODE_THROW, OVERFLOW_MODE_BREAK.
max_rows_in_join google.protobuf.Int64Value
Limit on maximum size of the hash table for JOIN, in rows. The minimum value is 0.
max_bytes_in_join google.protobuf.Int64Value
Limit on maximum size of the hash table for JOIN, in bytes. The minimum value is 0.
join_overflow_mode enum OverflowMode
Determine the behavior on exceeding max_rows_in_join or max_bytes_in_join limit. Possible values: OVERFLOW_MODE_THROW, OVERFLOW_MODE_BREAK.
max_columns_to_read google.protobuf.Int64Value
Limits the maximum number of columns that can be read from a table in a single query. If the query requires to read more columns to complete, then it will be aborted.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_temporary_columns google.protobuf.Int64Value
Limits the maximum number of temporary columns that must be kept in RAM at the same time when running a query, including constant columns.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_temporary_non_const_columns google.protobuf.Int64Value
Limits the maximum number of temporary columns that must be kept in RAM at the same time when running a query, excluding constant columns.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_query_size google.protobuf.Int64Value
Limits the size of the part of a query that can be transferred to RAM for parsing with the SQL parser, in bytes.
Value must be greater than 0 (default: 262144).
See in-depth description in ClickHouse documentation. Value must be greater than 0.
max_ast_depth google.protobuf.Int64Value
Limits the maximum depth of query syntax tree.
Executing a big and complex query may result in building a syntax tree of enormous depth. By using this setting, you can prohibit execution of over-sized or non-optimized queries for huge tables.
For example, the **SELECT *** query may result in more complex and deeper syntax tree, compared to the SELECT ... WHERE ... query, containing constraints and conditions, in the most cases. A user can be forced to construct more optimized queries, if this setting is used.
Value must be greater than 0 (default: 1000). If a too small value is set, it may render ClickHouse unable to execute even simple queries.
See in-depth description in ClickHouse documentation. Value must be greater than 0.
max_ast_elements google.protobuf.Int64Value
Limits the maximum size of query syntax tree in number of nodes.
Executing a big and complex query may result in building a syntax tree of enormous size. By using this setting, you can prohibit execution of over-sized or non-optimized queries for huge tables.
Value must be greater than 0 (default: 50000). If a too small value is set, it may render ClickHouse unable to execute even simple queries.
See in-depth description in ClickHouse documentation. Value must be greater than 0.
max_expanded_ast_elements google.protobuf.Int64Value
Limits the maximum size of query syntax tree in number of nodes after expansion of aliases and the asterisk values.
Executing a big and complex query may result in building a syntax tree of enormous size. By using this setting, you can prohibit execution of over-sized or non-optimized queries for huge tables.
Value must be greater than 0 (default: 500000). If a too small value is set, it may render ClickHouse unable to execute even simple queries. Value must be greater than 0.
min_execution_speed google.protobuf.Int64Value
Minimal execution speed in rows per second. The minimum value is 0.
min_execution_speed_bytes google.protobuf.Int64Value
Minimal execution speed in bytes per second. The minimum value is 0.
count_distinct_implementation enum CountDistinctImplementation
Aggregate function to use for implementation of count(DISTINCT ...).
input_format_values_interpret_expressions google.protobuf.BoolValue
Enables or disables SQL parser if the fast stream parser cannot parse the data.
Enable this setting, if the data that you want to insert into a table contains SQL expressions.
For example, the stream parser is unable to parse a value that contains now() expression; therefore an INSERT query for this value will fail and no data will be inserted into a table. With enabled SQL parser, this expression is parsed correctly: the now() expression will be parsed as SQL function, interpreted, and the current date and time will be inserted into the table as a result.
This setting has effect only if you use Values format when inserting data.
Default value: true (SQL parser is enabled).
See in-depth description in ClickHouse documentation.
input_format_defaults_for_omitted_fields google.protobuf.BoolValue
Enables or disables replacing omitted input values with default values of the respective columns when performing INSERT queries.
Default value: true (replacing is enabled).
output_format_json_quote_64bit_integers google.protobuf.BoolValue
Enables quoting of 64-bit integers in JSON output format.
If this setting is enabled, then 64-bit integers (UInt64 and Int64) will be quoted when written to JSON output in order to maintain compatibility with the most of the JavaScript engines. Otherwise, such integers will not be quoted.
Default value: false (quoting 64-bit integers is disabled).
output_format_json_quote_denormals google.protobuf.BoolValue
Enables special floating-point values (+nan, -nan, +inf and -inf) in JSON output format.
Default value: false (special values do not present in output).
low_cardinality_allow_in_native_format google.protobuf.BoolValue
Determines whether to use LowCardinality type in Native format.
  • true (default)-yes, use.
  • false-convert LowCardinality columns to regular columns when doing SELECT, and convert regular columns to LowCardinality when doing INSERT.

LowCardinality columns (aka sparse columns) store data in more effective way, compared to regular columns, by using hash tables. If data to insert suits this storage format, ClickHouse will place them into LowCardinality column.
If you use a third-party ClickHouse client that can't work with LowCardinality columns, then this client will not be able to correctly interpret the result of the query that asks for data stored in LowCardinality column. Disable this setting to convert LowCardinality column to regular column when creating the result, so such clients will be able to process the result.
Official ClickHouse client works with LowCardinality columns out-of-the-box.
Default value: true (LowCardinality columns are used in Native format).
empty_result_for_aggregation_by_empty_set google.protobuf.BoolValue
Enables returning of empty result when aggregating without keys (with GROUP BY operation absent) on empty set (e.g., SELECT count(*) FROM table WHERE 0).
  • true-ClickHouse will return an empty result for such queries.
  • false (default)-ClickHouse will return a single-line result consisting of NULL values for aggregation functions, in accordance with SQL standard.
joined_subquery_requires_alias google.protobuf.BoolValue
join_use_nulls google.protobuf.BoolValue
transform_null_in google.protobuf.BoolValue
http_connection_timeout google.protobuf.Int64Value
HTTP connection timeout, in milliseconds.
Value must be greater than 0 (default: 1000, 1 second).
http_receive_timeout google.protobuf.Int64Value
HTTP receive timeout, in milliseconds.
Value must be greater than 0 (default: 1800000, 1800 seconds, 30 minutes).
http_send_timeout google.protobuf.Int64Value
HTTP send timeout, in milliseconds.
Value must be greater than 0 (default: 1800000, 1800 seconds, 30 minutes).
enable_http_compression google.protobuf.BoolValue
Enables or disables data compression in HTTP responses.
By default, ClickHouse stores data compressed. When executing a query, its result is uncompressed. Use this setting to command ClickHouse to compress the result when sending it via HTTP.
Enable this setting and add the Accept-Encoding: HTTP header in a HTTP request to force compression of HTTP response from ClickHouse.
ClickHouse support the following compression methods: gzip, br and deflate.
Default value: false (compression is disabled).
See in-depth description in ClickHouse documentation.
send_progress_in_http_headers google.protobuf.BoolValue
Enables progress notifications using X-ClickHouse-Progress HTTP header.
Default value: false (notifications disabled).
http_headers_progress_interval google.protobuf.Int64Value
Minimum interval between progress notifications with X-ClickHouse-Progress HTTP header, in milliseconds.
Value must be greater than 0 (default: 100).
add_http_cors_header google.protobuf.BoolValue
Adds CORS header in HTTP responses.
Default value: false (header is not added).
quota_mode enum QuotaMode
Quota accounting mode. Possible values: QUOTA_MODE_DEFAULT, QUOTA_MODE_KEYED and QUOTA_MODE_KEYED_BY_IP.

UserQuota

Field Description
interval_duration google.protobuf.Int64Value
Duration of interval for quota in milliseconds. Minimal value is 1 second. The minimum value is 1000.
queries google.protobuf.Int64Value
The total number of queries. 0 - unlimited. The minimum value is 0.
errors google.protobuf.Int64Value
The number of queries that threw exception. 0 - unlimited. The minimum value is 0.
result_rows google.protobuf.Int64Value
The total number of rows given as the result.. 0 - unlimited. The minimum value is 0.
read_rows google.protobuf.Int64Value
The total number of source rows read from tables for running the query, on all remote servers. 0 - unlimited. The minimum value is 0.
execution_time google.protobuf.Int64Value
The total query execution time, in milliseconds (wall time). 0 - unlimited. The minimum value is 0.

Create

Creates a ClickHouse user in the specified cluster.

rpc Create (CreateUserRequest) returns (operation.Operation)

Metadata and response of Operation:

    Operation.metadata:CreateUserMetadata

    Operation.response:User

CreateUserRequest

Field Description
cluster_id string
Required. ID of the ClickHouse cluster to create a user in. To get the cluster ID, use a ClusterService.List request. The maximum string length in characters is 50.
user_spec UserSpec
Required. Properties of the user to be created.

UserSpec

Field Description
name string
Required. Name of the ClickHouse user. The maximum string length in characters is 63. Value must match the regular expression [a-zA-Z0-9_]*.
password string
Required. Password of the ClickHouse user. The string length in characters must be 8-128.
permissions[] Permission
Set of permissions to grant to the user. If not set, it's granted permissions to access all databases.
settings UserSettings
quotas[] UserQuota
Set of quotas assigned to the user.

Permission

Field Description
database_name string
Name of the database that the permission grants access to.

UserSettings

Field Description
readonly google.protobuf.Int64Value
Restricts permissions for non-DDL queries. To restrict permissions for DDL queries, use allow_ddl instead.
  • 0 (default)-no restrictions.
  • 1-only read data queries are allowed.
  • 2-read data and change settings queries are allowed.

See in-depth description in ClickHouse documentation. Acceptable values are 0 to 2, inclusive.
allow_ddl google.protobuf.BoolValue
Determines whether DDL queries are allowed (e.g., CREATE, ALTER, RENAME, etc).
Default value: true.
See in-depth description in ClickHouse documentation.
insert_quorum google.protobuf.Int64Value
Enables or disables write quorum for ClickHouse cluster. If the value is less than 2, then write quorum is disabled, otherwise it is enabled.
When used, write quorum guarantees that ClickHouse has written data to the quorum of insert_quorum replicas with no errors until the insert_quorum_timeout expires. All replicas in the quorum are in the consistent state, meaning that they contain linearized data from the previous INSERT queries. Employ write quorum, if you need the guarantees that the written data would not be lost in case of one or more replicas failure.
You can use select_sequential_consistency setting to read the data written with write quorum.
See in-depth description in ClickHouse documentation. The minimum value is 0.
connect_timeout google.protobuf.Int64Value
Connection timeout in milliseconds.
Value must be greater than 0 (default: 10000, 10 seconds). Value must be greater than 0.
receive_timeout google.protobuf.Int64Value
Receive timeout in milliseconds.
Value must be greater than 0 (default: 300000, 300 seconds or 5 minutes). Value must be greater than 0.
send_timeout google.protobuf.Int64Value
Send timeout in milliseconds.
Value must be greater than 0 (default: 300000, 300 seconds or 5 minutes). Value must be greater than 0.
insert_quorum_timeout google.protobuf.Int64Value
Quorum write timeout in milliseconds.
If the write quorum is enabled in the cluster, this timeout expires and some data is not written to the insert_quorum replicas, then ClickHouse will abort the execution of INSERT query and return an error. In this case, the client must send the query again to write the data block into the same or another replica.
Minimum value: 1000, 1 second (default: 60000, 1 minute). The minimum value is 1000.
select_sequential_consistency google.protobuf.BoolValue
Determines the behavior of SELECT queries from the replicated table: if enabled, ClickHouse will terminate a query with error message in case the replica does not have a chunk written with the quorum and will not read the parts that have not yet been written with the quorum.
Default value: false (sequential consistency is disabled).
max_replica_delay_for_distributed_queries google.protobuf.Int64Value
Max replica delay in milliseconds. If a replica lags more than the set value, this replica is not used and becomes a stale one.
Minimum value: 1000, 1 second (default: 300000, 300 seconds or 5 minutes).
See in-depth description in ClickHouse documentation. The minimum value is 1000.
fallback_to_stale_replicas_for_distributed_queries google.protobuf.BoolValue
Enables or disables query forcing to a stale replica in case the actual data is unavailable. If enabled, ClickHouse will choose the most up-to-date replica and force the query to use the data in this replica. This setting can be used when doing SELECT query from a distributed table that points to replicated tables.
Default value: true (query forcing is enabled).
See in-depth description in ClickHouse documentation.
replication_alter_partitions_sync google.protobuf.Int64Value
Wait mode for asynchronous actions in ALTER queries on replicated tables:
  • 0-do not wait for replicas.
  • 1-only wait for own execution (default).
  • 2-wait for all replicas.

See in-depth description in ClickHouse documentation. Acceptable values are 0 to 2, inclusive.
distributed_product_mode enum DistributedProductMode
Determine the behavior of distributed subqueries.
See in-depth description in ClickHouse documentation.
  • DISTRIBUTED_PRODUCT_MODE_DENY: Default value. Prohibits using these types of subqueries (returns the "Double-distributed in/JOIN subqueries is denied" exception).
  • DISTRIBUTED_PRODUCT_MODE_LOCAL: Replaces the database and table in the subquery with local ones for the destination server (shard), leaving the normal IN/JOIN.
  • DISTRIBUTED_PRODUCT_MODE_GLOBAL: Replaces the IN/JOIN query with GLOBAL IN/GLOBAL JOIN.
  • DISTRIBUTED_PRODUCT_MODE_ALLOW: Allows the use of these types of subqueries.
distributed_aggregation_memory_efficient google.protobuf.BoolValue
Enables of disables memory saving mode when doing distributed aggregation.
When ClickHouse works with a distributed query, external aggregation is done on remote servers. Enable this setting to achieve a smaller memory footprint on the server that sourced such a distributed query.
Default value: false (memory saving mode is disabled).
See in-depth description in ClickHouse documentation.
distributed_ddl_task_timeout google.protobuf.Int64Value
Timeout for DDL queries, in milliseconds.
skip_unavailable_shards google.protobuf.BoolValue
Enables or disables silent skipping of unavailable shards.
A shard is considered unavailable if all its replicas are also unavailable.
Default value: false (silent skipping is disabled).
See in-depth description in ClickHouse documentation.
compile google.protobuf.BoolValue
Enables or disables query compilation. If you execute a lot of structurally identical queries, then enable this setting. As a result, such queries may be executed faster due to use of queries' compiled parts.
Use this setting in combination with min_count_to_compile setting.
Default value: false (compilation is disabled).
See in-depth description in ClickHouse documentation.
min_count_to_compile google.protobuf.Int64Value
How many structurally identical queries ClickHouse has to encounter before they are compiled.
Minimum value: 0 (default: 3).
For the 0 value compilation is synchronous: a query waits for compilation process to complete prior to continuing execution. It is recommended to set this value only for testing purposes.
For all other values, compilation is asynchronous: the compilation process executes in a separate thread. When a compiled part of query is ready, it will be used by ClickHouse for eligible queries, including the ones that are currently running.
See in-depth description in ClickHouse documentation. The minimum value is 0.
compile_expressions google.protobuf.BoolValue
Enables or disables expression compilation. If you execute a lot of queries that contain identical expressions, then enable this setting. As a result, such queries may be executed faster due to use of compiled expressions.
Use this setting in combination with min_count_to_compile_expression setting.
Default value: false (expression compilation is disabled).
min_count_to_compile_expression google.protobuf.Int64Value
How many identical expressions ClickHouse has to encounter before they are compiled.
Minimum value: 0 (default: 3).
For the 0 value compilation is synchronous: a query waits for expression compilation process to complete prior to continuing execution. It is recommended to set this value only for testing purposes.
For all other values, compilation is asynchronous: the compilation process executes in a separate thread. When a compiled expression is ready, it will be used by ClickHouse for eligible queries, including the ones that are currently running. The minimum value is 0.
max_block_size google.protobuf.Int64Value
The maximum block size for reading.
Data in ClickHouse is organized and processed by blocks (block is a set of columns' parts). The internal processing cycles for a single block are efficient enough, but there are noticeable expenditures on each block.
This setting is a recommendation for size of block (in a count of rows) that should be loaded from tables.
Value must be greater than 0 (default: 65536).
See in-depth description in ClickHouse documentation. Value must be greater than 0.
min_insert_block_size_rows google.protobuf.Int64Value
Limits the minimum number of rows in a block to be inserted in a table by INSERT query. Blocks that are smaller than the specified value, will be squashed together into the bigger blocks.
Minimal value: 0, block squashing is disabled (default: 1048576). The minimum value is 0.
min_insert_block_size_bytes google.protobuf.Int64Value
Limits the minimum number of bytes in a block to be inserted in a table by INSERT query. Blocks that are smaller than the specified value, will be squashed together into the bigger blocks.
Minimal value: 0, block squashing is disabled (default: 268435456, 256 MB). The minimum value is 0.
max_insert_block_size google.protobuf.Int64Value
Allows to form blocks of the specified size (in bytes) when inserting data in a table. This setting has effect only if server is creating such blocks by itself.
Value must be greater than 0 (default: 1048576).
See in-depth description in ClickHouse documentation. Value must be greater than 0.
min_bytes_to_use_direct_io google.protobuf.Int64Value
Limits the minimum number of bytes to enable unbuffered direct reads from disk (Direct I/O).
By default, ClickHouse does not read data directly from disk, but relies on the filesystem and its cache instead. Such reading strategy is effective when the data volume is small. If the amount of the data to read is huge, it is more effective to read directly from the disk, bypassing the filesystem cache.
If the total amount of the data to read is greater than the value of this setting, then ClickHouse will fetch this data directly from the disk.
Minimal value and default value: 0, Direct I/O is disabled. The minimum value is 0.
use_uncompressed_cache google.protobuf.BoolValue
Determines whether to use the cache of uncompressed blocks, or not. Using this cache can significantly reduce latency and increase the throughput when a huge amount of small queries is to be processed. Enable this setting for the users who instantiates small queries frequently.
This setting has effect only for tables of the MergeTree family.
Default value: false (uncompressed cache is disabled).
See in-depth description in ClickHouse documentation.
merge_tree_max_rows_to_use_cache google.protobuf.Int64Value
Limits the maximum size in rows of the request that can use the cache of uncompressed data. The cache is not used for requests larger than the specified value.
Use this setting in combination with use_uncompressed_cache setting.
Value must be greater than 0 (default: 128x8192). Value must be greater than 0.
merge_tree_max_bytes_to_use_cache google.protobuf.Int64Value
Limits the maximum size in bytes of the request that can use the cache of uncompressed data. The cache is not used for requests larger than the specified value.
Use this setting in combination with use_uncompressed_cache setting.
Value must be greater than 0 (default: 192x10x1024x1024). Value must be greater than 0.
merge_tree_min_rows_for_concurrent_read google.protobuf.Int64Value
Limits the minimum number of rows to be read from a file to enable concurrent read. If the number of rows to be read exceeds this value, then ClickHouse will try to use a few threads to read from a file concurrently.
This setting has effect only for tables of the MergeTree family.
Value must be greater than 0 (default: 20x8192). Value must be greater than 0.
merge_tree_min_bytes_for_concurrent_read google.protobuf.Int64Value
Limits the number of bytes to be read from a file to enable concurrent read. If the number of bytes to be read exceeds this value, then ClickHouse will try to use a few threads to read from a file concurrently.
This setting has effect only for tables of the MergeTree family.
Value must be greater than 0 (default: 24x10x1024x1024). Value must be greater than 0.
max_bytes_before_external_group_by google.protobuf.Int64Value
Sets the threshold of RAM consumption (in bytes) after that the temporary data, collected during the GROUP BY operation, should be flushed to disk to limit the RAM comsumption.
By default, aggregation is done by employing hash table that resides in RAM. A query can result in aggregation of huge data volumes that can lead to memory exhaustion and abortion of the query (see the max_memory_usage setting). For such queries, you can use this setting to force ClickHouse to do flushing and complete aggregation successfully.
Minimal value and default value: 0, GROUP BY in the external memory is disabled.
When using aggregation in external memory, it is recommended to set the value of this setting twice as low as the max_memory_usage setting value (by default, the maximum memory usage is limited to ten gigabytes).
See in-depth description in ClickHouse documentation.
See also: the distributed_aggregation_memory_efficient setting.
max_bytes_before_external_sort google.protobuf.Int64Value
This setting is equivalent of the max_bytes_before_external_group_by setting, except for it is for sort operation (ORDER BY), not aggregation.
group_by_two_level_threshold google.protobuf.Int64Value
Sets the threshold of the number of keys, after that the two-level aggregation should be used.
Minimal value: 0, threshold is not set (default: 10000).
group_by_two_level_threshold_bytes google.protobuf.Int64Value
Sets the threshold of the number of bytes, after that the two-level aggregation should be used.
Minimal value: 0, threshold is not set (default: 100000000).
priority google.protobuf.Int64Value
Sets the priority of a query.
  • 0-priority is not used.
  • 1-the highest priority.
  • and so on. The higher the number, the lower a query's priority.

This setting should be set up for each query individually.
If ClickHouse is working with the high-priority queries, and a low-priority query enters, then the low-priority query is paused until higher-priority queries are completed.
Minimal value and default value: 0, priority is not used. The minimum value is 0.
max_threads google.protobuf.Int64Value
Limits the maximum number of threads to process the request (setting does not take threads that read data from remote servers into account).
This setting applies to threads that perform the same stages of the query processing pipeline in parallel.
Minimal value and default value: 0 (the thread number is calculated automatically based on the number of physical CPU cores, no HyperThreading cores are taken into account).
See in-depth description in ClickHouse documentation. Value must be greater than 0.
max_memory_usage google.protobuf.Int64Value
Limits the maximum memory usage (in bytes) for processing of a single user's query on a single server. This setting does not take server's free RAM amount or total RAM amount into account.
This limitation is enforced for any user's single query on a single server.
Minimal value: 0, no limitation is set. Value that is set in the ClickHouse default config file: 10737418240 (10 GB).
If you use max_bytes_before_external_group_by or max_bytes_before_external_sort setting, then it is recommended to set their values twice as low as max_memory_usage setting value.
See in-depth description in ClickHouse documentation. The minimum value is 0.
max_memory_usage_for_user google.protobuf.Int64Value
Limits the maximum memory usage (in bytes) for processing of user's queries on a single server. This setting does not take server's free RAM amount or total RAM amount into account.
This limitation is enforced for all queries that belong to one user and run simultaneously on a single server.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_network_bandwidth google.protobuf.Int64Value
The maximum speed of data exchange over the network in bytes per second for a query.
Minimal value and default value: 0, no limitation is set.
max_network_bandwidth_for_user google.protobuf.Int64Value
The maximum speed of data exchange over the network in bytes per second for all concurrently running user queries.
Minimal value and default value: 0, no limitation is set.
force_index_by_date google.protobuf.BoolValue
If enabled, query is not executed if the ClickHouse can't use index by date. This setting has effect only for tables of the MergeTree family.
Default value: false (setting is disabled, query executes even if ClickHouse can't use index by date).
See in-depth description in ClickHouse documentation.
force_primary_key google.protobuf.BoolValue
If enabled, query is not executed if the ClickHouse can't use index by primary key. This setting has effect only for tables of the MergeTree family.
Default value: false (setting is disabled, query executes even if ClickHouse can't use index by primary key).
See in-depth description in ClickHouse documentation.
max_rows_to_read google.protobuf.Int64Value
Limits the maximum number of rows that can be read from a table when running a query.
Minimal value and default value: 0, no limitation is set.
See in-depth description in ClickHouse documentation. The minimum value is 0.
max_bytes_to_read google.protobuf.Int64Value
Limits the maximum number of bytes (uncompressed data) that can be read from a table when running a query.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
read_overflow_mode enum OverflowMode
Determines the behavior on exceeding limits while reading the data.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
max_rows_to_group_by google.protobuf.Int64Value
Limits the maximum number of unique keys received from aggregation function. This setting helps to reduce RAM consumption while doing aggregation.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
group_by_overflow_mode enum GroupByOverflowMode
Determines the behavior on exceeding limits while doing aggregation.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
  • any-perform approximate GROUP BY operation by continuing aggregation for the keys that got into the set, but don't add new keys to the set.
max_rows_to_sort google.protobuf.Int64Value
Limits the maximum number of rows that can be read from a table for sorting. This setting helps to reduce RAM consumption.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_bytes_to_sort google.protobuf.Int64Value
Limits the maximum number of bytes (uncompressed data) that can be read from a table for sorting. This setting helps to reduce RAM consumption.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
sort_overflow_mode enum OverflowMode
Determines the behavior on exceeding limits while sorting.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
max_result_rows google.protobuf.Int64Value
Limits the number of rows in the result. This limitation is also checked for subqueries and parts of distributed queries that run on remote servers.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_result_bytes google.protobuf.Int64Value
Limits the number of bytes in the result. This limitation is also checked for subqueries and parts of distributed queries that run on remote servers.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
result_overflow_mode enum OverflowMode
Determines the behavior on exceeding limits while forming result.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
max_rows_in_distinct google.protobuf.Int64Value
Limits the maximum number of different rows when using DISTINCT.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_bytes_in_distinct google.protobuf.Int64Value
Limits the maximum size of a hash table in bytes (uncompressed data) when using DISTINCT. The minimum value is 0.
distinct_overflow_mode enum OverflowMode
Determines the behavior on exceeding limits while doing DISCTINCT.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
max_rows_to_transfer google.protobuf.Int64Value
Limits the maximum number of rows that can be passed to a remote server or saved in a temporary table when using GLOBAL IN.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_bytes_to_transfer google.protobuf.Int64Value
Limits the maximum number of bytes (uncompressed data) that can be passed to a remote server or saved in a temporary table when using GLOBAL IN.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
transfer_overflow_mode enum OverflowMode
Determines the behavior on exceeding limits while doing transfers.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
max_execution_time google.protobuf.Int64Value
Limits the maximum query execution time in milliseconds. At this moment, this limitation is not checked when passing one of the sorting stages, as well as merging and finalizing aggregation funictions.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
timeout_overflow_mode enum OverflowMode
Determines the behavior on exceeding limits of execution time.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
max_rows_in_set google.protobuf.Int64Value
Limit on the number of rows in the set resulting from the execution of the IN section. The minimum value is 0.
max_bytes_in_set google.protobuf.Int64Value
Limit on the number of bytes in the set resulting from the execution of the IN section. The minimum value is 0.
set_overflow_mode enum OverflowMode
Determine the behavior on exceeding max_rows_in_set or max_bytes_in_set limit. Possible values: OVERFLOW_MODE_THROW, OVERFLOW_MODE_BREAK.
max_rows_in_join google.protobuf.Int64Value
Limit on maximum size of the hash table for JOIN, in rows. The minimum value is 0.
max_bytes_in_join google.protobuf.Int64Value
Limit on maximum size of the hash table for JOIN, in bytes. The minimum value is 0.
join_overflow_mode enum OverflowMode
Determine the behavior on exceeding max_rows_in_join or max_bytes_in_join limit. Possible values: OVERFLOW_MODE_THROW, OVERFLOW_MODE_BREAK.
max_columns_to_read google.protobuf.Int64Value
Limits the maximum number of columns that can be read from a table in a single query. If the query requires to read more columns to complete, then it will be aborted.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_temporary_columns google.protobuf.Int64Value
Limits the maximum number of temporary columns that must be kept in RAM at the same time when running a query, including constant columns.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_temporary_non_const_columns google.protobuf.Int64Value
Limits the maximum number of temporary columns that must be kept in RAM at the same time when running a query, excluding constant columns.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_query_size google.protobuf.Int64Value
Limits the size of the part of a query that can be transferred to RAM for parsing with the SQL parser, in bytes.
Value must be greater than 0 (default: 262144).
See in-depth description in ClickHouse documentation. Value must be greater than 0.
max_ast_depth google.protobuf.Int64Value
Limits the maximum depth of query syntax tree.
Executing a big and complex query may result in building a syntax tree of enormous depth. By using this setting, you can prohibit execution of over-sized or non-optimized queries for huge tables.
For example, the **SELECT *** query may result in more complex and deeper syntax tree, compared to the SELECT ... WHERE ... query, containing constraints and conditions, in the most cases. A user can be forced to construct more optimized queries, if this setting is used.
Value must be greater than 0 (default: 1000). If a too small value is set, it may render ClickHouse unable to execute even simple queries.
See in-depth description in ClickHouse documentation. Value must be greater than 0.
max_ast_elements google.protobuf.Int64Value
Limits the maximum size of query syntax tree in number of nodes.
Executing a big and complex query may result in building a syntax tree of enormous size. By using this setting, you can prohibit execution of over-sized or non-optimized queries for huge tables.
Value must be greater than 0 (default: 50000). If a too small value is set, it may render ClickHouse unable to execute even simple queries.
See in-depth description in ClickHouse documentation. Value must be greater than 0.
max_expanded_ast_elements google.protobuf.Int64Value
Limits the maximum size of query syntax tree in number of nodes after expansion of aliases and the asterisk values.
Executing a big and complex query may result in building a syntax tree of enormous size. By using this setting, you can prohibit execution of over-sized or non-optimized queries for huge tables.
Value must be greater than 0 (default: 500000). If a too small value is set, it may render ClickHouse unable to execute even simple queries. Value must be greater than 0.
min_execution_speed google.protobuf.Int64Value
Minimal execution speed in rows per second. The minimum value is 0.
min_execution_speed_bytes google.protobuf.Int64Value
Minimal execution speed in bytes per second. The minimum value is 0.
count_distinct_implementation enum CountDistinctImplementation
Aggregate function to use for implementation of count(DISTINCT ...).
input_format_values_interpret_expressions google.protobuf.BoolValue
Enables or disables SQL parser if the fast stream parser cannot parse the data.
Enable this setting, if the data that you want to insert into a table contains SQL expressions.
For example, the stream parser is unable to parse a value that contains now() expression; therefore an INSERT query for this value will fail and no data will be inserted into a table. With enabled SQL parser, this expression is parsed correctly: the now() expression will be parsed as SQL function, interpreted, and the current date and time will be inserted into the table as a result.
This setting has effect only if you use Values format when inserting data.
Default value: true (SQL parser is enabled).
See in-depth description in ClickHouse documentation.
input_format_defaults_for_omitted_fields google.protobuf.BoolValue
Enables or disables replacing omitted input values with default values of the respective columns when performing INSERT queries.
Default value: true (replacing is enabled).
output_format_json_quote_64bit_integers google.protobuf.BoolValue
Enables quoting of 64-bit integers in JSON output format.
If this setting is enabled, then 64-bit integers (UInt64 and Int64) will be quoted when written to JSON output in order to maintain compatibility with the most of the JavaScript engines. Otherwise, such integers will not be quoted.
Default value: false (quoting 64-bit integers is disabled).
output_format_json_quote_denormals google.protobuf.BoolValue
Enables special floating-point values (+nan, -nan, +inf and -inf) in JSON output format.
Default value: false (special values do not present in output).
low_cardinality_allow_in_native_format google.protobuf.BoolValue
Determines whether to use LowCardinality type in Native format.
  • true (default)-yes, use.
  • false-convert LowCardinality columns to regular columns when doing SELECT, and convert regular columns to LowCardinality when doing INSERT.

LowCardinality columns (aka sparse columns) store data in more effective way, compared to regular columns, by using hash tables. If data to insert suits this storage format, ClickHouse will place them into LowCardinality column.
If you use a third-party ClickHouse client that can't work with LowCardinality columns, then this client will not be able to correctly interpret the result of the query that asks for data stored in LowCardinality column. Disable this setting to convert LowCardinality column to regular column when creating the result, so such clients will be able to process the result.
Official ClickHouse client works with LowCardinality columns out-of-the-box.
Default value: true (LowCardinality columns are used in Native format).
empty_result_for_aggregation_by_empty_set google.protobuf.BoolValue
Enables returning of empty result when aggregating without keys (with GROUP BY operation absent) on empty set (e.g., SELECT count(*) FROM table WHERE 0).
  • true-ClickHouse will return an empty result for such queries.
  • false (default)-ClickHouse will return a single-line result consisting of NULL values for aggregation functions, in accordance with SQL standard.
joined_subquery_requires_alias google.protobuf.BoolValue
join_use_nulls google.protobuf.BoolValue
transform_null_in google.protobuf.BoolValue
http_connection_timeout google.protobuf.Int64Value
HTTP connection timeout, in milliseconds.
Value must be greater than 0 (default: 1000, 1 second).
http_receive_timeout google.protobuf.Int64Value
HTTP receive timeout, in milliseconds.
Value must be greater than 0 (default: 1800000, 1800 seconds, 30 minutes).
http_send_timeout google.protobuf.Int64Value
HTTP send timeout, in milliseconds.
Value must be greater than 0 (default: 1800000, 1800 seconds, 30 minutes).
enable_http_compression google.protobuf.BoolValue
Enables or disables data compression in HTTP responses.
By default, ClickHouse stores data compressed. When executing a query, its result is uncompressed. Use this setting to command ClickHouse to compress the result when sending it via HTTP.
Enable this setting and add the Accept-Encoding: HTTP header in a HTTP request to force compression of HTTP response from ClickHouse.
ClickHouse support the following compression methods: gzip, br and deflate.
Default value: false (compression is disabled).
See in-depth description in ClickHouse documentation.
send_progress_in_http_headers google.protobuf.BoolValue
Enables progress notifications using X-ClickHouse-Progress HTTP header.
Default value: false (notifications disabled).
http_headers_progress_interval google.protobuf.Int64Value
Minimum interval between progress notifications with X-ClickHouse-Progress HTTP header, in milliseconds.
Value must be greater than 0 (default: 100).
add_http_cors_header google.protobuf.BoolValue
Adds CORS header in HTTP responses.
Default value: false (header is not added).
quota_mode enum QuotaMode
Quota accounting mode. Possible values: QUOTA_MODE_DEFAULT, QUOTA_MODE_KEYED and QUOTA_MODE_KEYED_BY_IP.

UserQuota

Field Description
interval_duration google.protobuf.Int64Value
Duration of interval for quota in milliseconds. Minimal value is 1 second. The minimum value is 1000.
queries google.protobuf.Int64Value
The total number of queries. 0 - unlimited. The minimum value is 0.
errors google.protobuf.Int64Value
The number of queries that threw exception. 0 - unlimited. The minimum value is 0.
result_rows google.protobuf.Int64Value
The total number of rows given as the result.. 0 - unlimited. The minimum value is 0.
read_rows google.protobuf.Int64Value
The total number of source rows read from tables for running the query, on all remote servers. 0 - unlimited. The minimum value is 0.
execution_time google.protobuf.Int64Value
The total query execution time, in milliseconds (wall time). 0 - unlimited. The minimum value is 0.

Operation

Field Description
id string
ID of the operation.
description string
Description of the operation. 0-256 characters long.
created_at google.protobuf.Timestamp
Creation timestamp.
created_by string
ID of the user or service account who initiated the operation.
modified_at google.protobuf.Timestamp
The time when the Operation resource was last modified.
done bool
If the value is false, it means the operation is still in progress. If true, the operation is completed, and either error or response is available.
metadata google.protobuf.Any<CreateUserMetadata>
Service-specific metadata associated with the operation. It typically contains the ID of the target resource that the operation is performed on. Any method that returns a long-running operation should document the metadata type, if any.
result oneof: error or response
The operation result. If done == false and there was no failure detected, neither error nor response is set. If done == false and there was a failure detected, error is set. If done == true, exactly one of error or response is set.
  error google.rpc.Status
The error result of the operation in case of failure or cancellation.
  response google.protobuf.Any<User>
if operation finished successfully.

CreateUserMetadata

Field Description
cluster_id string
ID of the ClickHouse cluster the user is being created in.
user_name string
Name of the user that is being created.

User

Field Description
name string
Name of the ClickHouse user.
cluster_id string
ID of the ClickHouse cluster the user belongs to.
permissions[] Permission
Set of permissions granted to the user.
settings UserSettings
quotas[] UserQuota
Set of quotas assigned to the user.

Permission

Field Description
database_name string
Name of the database that the permission grants access to.

UserSettings

Field Description
readonly google.protobuf.Int64Value
Restricts permissions for non-DDL queries. To restrict permissions for DDL queries, use allow_ddl instead.
  • 0 (default)-no restrictions.
  • 1-only read data queries are allowed.
  • 2-read data and change settings queries are allowed.

See in-depth description in ClickHouse documentation. Acceptable values are 0 to 2, inclusive.
allow_ddl google.protobuf.BoolValue
Determines whether DDL queries are allowed (e.g., CREATE, ALTER, RENAME, etc).
Default value: true.
See in-depth description in ClickHouse documentation.
insert_quorum google.protobuf.Int64Value
Enables or disables write quorum for ClickHouse cluster. If the value is less than 2, then write quorum is disabled, otherwise it is enabled.
When used, write quorum guarantees that ClickHouse has written data to the quorum of insert_quorum replicas with no errors until the insert_quorum_timeout expires. All replicas in the quorum are in the consistent state, meaning that they contain linearized data from the previous INSERT queries. Employ write quorum, if you need the guarantees that the written data would not be lost in case of one or more replicas failure.
You can use select_sequential_consistency setting to read the data written with write quorum.
See in-depth description in ClickHouse documentation. The minimum value is 0.
connect_timeout google.protobuf.Int64Value
Connection timeout in milliseconds.
Value must be greater than 0 (default: 10000, 10 seconds). Value must be greater than 0.
receive_timeout google.protobuf.Int64Value
Receive timeout in milliseconds.
Value must be greater than 0 (default: 300000, 300 seconds or 5 minutes). Value must be greater than 0.
send_timeout google.protobuf.Int64Value
Send timeout in milliseconds.
Value must be greater than 0 (default: 300000, 300 seconds or 5 minutes). Value must be greater than 0.
insert_quorum_timeout google.protobuf.Int64Value
Quorum write timeout in milliseconds.
If the write quorum is enabled in the cluster, this timeout expires and some data is not written to the insert_quorum replicas, then ClickHouse will abort the execution of INSERT query and return an error. In this case, the client must send the query again to write the data block into the same or another replica.
Minimum value: 1000, 1 second (default: 60000, 1 minute). The minimum value is 1000.
select_sequential_consistency google.protobuf.BoolValue
Determines the behavior of SELECT queries from the replicated table: if enabled, ClickHouse will terminate a query with error message in case the replica does not have a chunk written with the quorum and will not read the parts that have not yet been written with the quorum.
Default value: false (sequential consistency is disabled).
max_replica_delay_for_distributed_queries google.protobuf.Int64Value
Max replica delay in milliseconds. If a replica lags more than the set value, this replica is not used and becomes a stale one.
Minimum value: 1000, 1 second (default: 300000, 300 seconds or 5 minutes).
See in-depth description in ClickHouse documentation. The minimum value is 1000.
fallback_to_stale_replicas_for_distributed_queries google.protobuf.BoolValue
Enables or disables query forcing to a stale replica in case the actual data is unavailable. If enabled, ClickHouse will choose the most up-to-date replica and force the query to use the data in this replica. This setting can be used when doing SELECT query from a distributed table that points to replicated tables.
Default value: true (query forcing is enabled).
See in-depth description in ClickHouse documentation.
replication_alter_partitions_sync google.protobuf.Int64Value
Wait mode for asynchronous actions in ALTER queries on replicated tables:
  • 0-do not wait for replicas.
  • 1-only wait for own execution (default).
  • 2-wait for all replicas.

See in-depth description in ClickHouse documentation. Acceptable values are 0 to 2, inclusive.
distributed_product_mode enum DistributedProductMode
Determine the behavior of distributed subqueries.
See in-depth description in ClickHouse documentation.
  • DISTRIBUTED_PRODUCT_MODE_DENY: Default value. Prohibits using these types of subqueries (returns the "Double-distributed in/JOIN subqueries is denied" exception).
  • DISTRIBUTED_PRODUCT_MODE_LOCAL: Replaces the database and table in the subquery with local ones for the destination server (shard), leaving the normal IN/JOIN.
  • DISTRIBUTED_PRODUCT_MODE_GLOBAL: Replaces the IN/JOIN query with GLOBAL IN/GLOBAL JOIN.
  • DISTRIBUTED_PRODUCT_MODE_ALLOW: Allows the use of these types of subqueries.
distributed_aggregation_memory_efficient google.protobuf.BoolValue
Enables of disables memory saving mode when doing distributed aggregation.
When ClickHouse works with a distributed query, external aggregation is done on remote servers. Enable this setting to achieve a smaller memory footprint on the server that sourced such a distributed query.
Default value: false (memory saving mode is disabled).
See in-depth description in ClickHouse documentation.
distributed_ddl_task_timeout google.protobuf.Int64Value
Timeout for DDL queries, in milliseconds.
skip_unavailable_shards google.protobuf.BoolValue
Enables or disables silent skipping of unavailable shards.
A shard is considered unavailable if all its replicas are also unavailable.
Default value: false (silent skipping is disabled).
See in-depth description in ClickHouse documentation.
compile google.protobuf.BoolValue
Enables or disables query compilation. If you execute a lot of structurally identical queries, then enable this setting. As a result, such queries may be executed faster due to use of queries' compiled parts.
Use this setting in combination with min_count_to_compile setting.
Default value: false (compilation is disabled).
See in-depth description in ClickHouse documentation.
min_count_to_compile google.protobuf.Int64Value
How many structurally identical queries ClickHouse has to encounter before they are compiled.
Minimum value: 0 (default: 3).
For the 0 value compilation is synchronous: a query waits for compilation process to complete prior to continuing execution. It is recommended to set this value only for testing purposes.
For all other values, compilation is asynchronous: the compilation process executes in a separate thread. When a compiled part of query is ready, it will be used by ClickHouse for eligible queries, including the ones that are currently running.
See in-depth description in ClickHouse documentation. The minimum value is 0.
compile_expressions google.protobuf.BoolValue
Enables or disables expression compilation. If you execute a lot of queries that contain identical expressions, then enable this setting. As a result, such queries may be executed faster due to use of compiled expressions.
Use this setting in combination with min_count_to_compile_expression setting.
Default value: false (expression compilation is disabled).
min_count_to_compile_expression google.protobuf.Int64Value
How many identical expressions ClickHouse has to encounter before they are compiled.
Minimum value: 0 (default: 3).
For the 0 value compilation is synchronous: a query waits for expression compilation process to complete prior to continuing execution. It is recommended to set this value only for testing purposes.
For all other values, compilation is asynchronous: the compilation process executes in a separate thread. When a compiled expression is ready, it will be used by ClickHouse for eligible queries, including the ones that are currently running. The minimum value is 0.
max_block_size google.protobuf.Int64Value
The maximum block size for reading.
Data in ClickHouse is organized and processed by blocks (block is a set of columns' parts). The internal processing cycles for a single block are efficient enough, but there are noticeable expenditures on each block.
This setting is a recommendation for size of block (in a count of rows) that should be loaded from tables.
Value must be greater than 0 (default: 65536).
See in-depth description in ClickHouse documentation. Value must be greater than 0.
min_insert_block_size_rows google.protobuf.Int64Value
Limits the minimum number of rows in a block to be inserted in a table by INSERT query. Blocks that are smaller than the specified value, will be squashed together into the bigger blocks.
Minimal value: 0, block squashing is disabled (default: 1048576). The minimum value is 0.
min_insert_block_size_bytes google.protobuf.Int64Value
Limits the minimum number of bytes in a block to be inserted in a table by INSERT query. Blocks that are smaller than the specified value, will be squashed together into the bigger blocks.
Minimal value: 0, block squashing is disabled (default: 268435456, 256 MB). The minimum value is 0.
max_insert_block_size google.protobuf.Int64Value
Allows to form blocks of the specified size (in bytes) when inserting data in a table. This setting has effect only if server is creating such blocks by itself.
Value must be greater than 0 (default: 1048576).
See in-depth description in ClickHouse documentation. Value must be greater than 0.
min_bytes_to_use_direct_io google.protobuf.Int64Value
Limits the minimum number of bytes to enable unbuffered direct reads from disk (Direct I/O).
By default, ClickHouse does not read data directly from disk, but relies on the filesystem and its cache instead. Such reading strategy is effective when the data volume is small. If the amount of the data to read is huge, it is more effective to read directly from the disk, bypassing the filesystem cache.
If the total amount of the data to read is greater than the value of this setting, then ClickHouse will fetch this data directly from the disk.
Minimal value and default value: 0, Direct I/O is disabled. The minimum value is 0.
use_uncompressed_cache google.protobuf.BoolValue
Determines whether to use the cache of uncompressed blocks, or not. Using this cache can significantly reduce latency and increase the throughput when a huge amount of small queries is to be processed. Enable this setting for the users who instantiates small queries frequently.
This setting has effect only for tables of the MergeTree family.
Default value: false (uncompressed cache is disabled).
See in-depth description in ClickHouse documentation.
merge_tree_max_rows_to_use_cache google.protobuf.Int64Value
Limits the maximum size in rows of the request that can use the cache of uncompressed data. The cache is not used for requests larger than the specified value.
Use this setting in combination with use_uncompressed_cache setting.
Value must be greater than 0 (default: 128x8192). Value must be greater than 0.
merge_tree_max_bytes_to_use_cache google.protobuf.Int64Value
Limits the maximum size in bytes of the request that can use the cache of uncompressed data. The cache is not used for requests larger than the specified value.
Use this setting in combination with use_uncompressed_cache setting.
Value must be greater than 0 (default: 192x10x1024x1024). Value must be greater than 0.
merge_tree_min_rows_for_concurrent_read google.protobuf.Int64Value
Limits the minimum number of rows to be read from a file to enable concurrent read. If the number of rows to be read exceeds this value, then ClickHouse will try to use a few threads to read from a file concurrently.
This setting has effect only for tables of the MergeTree family.
Value must be greater than 0 (default: 20x8192). Value must be greater than 0.
merge_tree_min_bytes_for_concurrent_read google.protobuf.Int64Value
Limits the number of bytes to be read from a file to enable concurrent read. If the number of bytes to be read exceeds this value, then ClickHouse will try to use a few threads to read from a file concurrently.
This setting has effect only for tables of the MergeTree family.
Value must be greater than 0 (default: 24x10x1024x1024). Value must be greater than 0.
max_bytes_before_external_group_by google.protobuf.Int64Value
Sets the threshold of RAM consumption (in bytes) after that the temporary data, collected during the GROUP BY operation, should be flushed to disk to limit the RAM comsumption.
By default, aggregation is done by employing hash table that resides in RAM. A query can result in aggregation of huge data volumes that can lead to memory exhaustion and abortion of the query (see the max_memory_usage setting). For such queries, you can use this setting to force ClickHouse to do flushing and complete aggregation successfully.
Minimal value and default value: 0, GROUP BY in the external memory is disabled.
When using aggregation in external memory, it is recommended to set the value of this setting twice as low as the max_memory_usage setting value (by default, the maximum memory usage is limited to ten gigabytes).
See in-depth description in ClickHouse documentation.
See also: the distributed_aggregation_memory_efficient setting.
max_bytes_before_external_sort google.protobuf.Int64Value
This setting is equivalent of the max_bytes_before_external_group_by setting, except for it is for sort operation (ORDER BY), not aggregation.
group_by_two_level_threshold google.protobuf.Int64Value
Sets the threshold of the number of keys, after that the two-level aggregation should be used.
Minimal value: 0, threshold is not set (default: 10000).
group_by_two_level_threshold_bytes google.protobuf.Int64Value
Sets the threshold of the number of bytes, after that the two-level aggregation should be used.
Minimal value: 0, threshold is not set (default: 100000000).
priority google.protobuf.Int64Value
Sets the priority of a query.
  • 0-priority is not used.
  • 1-the highest priority.
  • and so on. The higher the number, the lower a query's priority.

This setting should be set up for each query individually.
If ClickHouse is working with the high-priority queries, and a low-priority query enters, then the low-priority query is paused until higher-priority queries are completed.
Minimal value and default value: 0, priority is not used. The minimum value is 0.
max_threads google.protobuf.Int64Value
Limits the maximum number of threads to process the request (setting does not take threads that read data from remote servers into account).
This setting applies to threads that perform the same stages of the query processing pipeline in parallel.
Minimal value and default value: 0 (the thread number is calculated automatically based on the number of physical CPU cores, no HyperThreading cores are taken into account).
See in-depth description in ClickHouse documentation. Value must be greater than 0.
max_memory_usage google.protobuf.Int64Value
Limits the maximum memory usage (in bytes) for processing of a single user's query on a single server. This setting does not take server's free RAM amount or total RAM amount into account.
This limitation is enforced for any user's single query on a single server.
Minimal value: 0, no limitation is set. Value that is set in the ClickHouse default config file: 10737418240 (10 GB).
If you use max_bytes_before_external_group_by or max_bytes_before_external_sort setting, then it is recommended to set their values twice as low as max_memory_usage setting value.
See in-depth description in ClickHouse documentation. The minimum value is 0.
max_memory_usage_for_user google.protobuf.Int64Value
Limits the maximum memory usage (in bytes) for processing of user's queries on a single server. This setting does not take server's free RAM amount or total RAM amount into account.
This limitation is enforced for all queries that belong to one user and run simultaneously on a single server.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_network_bandwidth google.protobuf.Int64Value
The maximum speed of data exchange over the network in bytes per second for a query.
Minimal value and default value: 0, no limitation is set.
max_network_bandwidth_for_user google.protobuf.Int64Value
The maximum speed of data exchange over the network in bytes per second for all concurrently running user queries.
Minimal value and default value: 0, no limitation is set.
force_index_by_date google.protobuf.BoolValue
If enabled, query is not executed if the ClickHouse can't use index by date. This setting has effect only for tables of the MergeTree family.
Default value: false (setting is disabled, query executes even if ClickHouse can't use index by date).
See in-depth description in ClickHouse documentation.
force_primary_key google.protobuf.BoolValue
If enabled, query is not executed if the ClickHouse can't use index by primary key. This setting has effect only for tables of the MergeTree family.
Default value: false (setting is disabled, query executes even if ClickHouse can't use index by primary key).
See in-depth description in ClickHouse documentation.
max_rows_to_read google.protobuf.Int64Value
Limits the maximum number of rows that can be read from a table when running a query.
Minimal value and default value: 0, no limitation is set.
See in-depth description in ClickHouse documentation. The minimum value is 0.
max_bytes_to_read google.protobuf.Int64Value
Limits the maximum number of bytes (uncompressed data) that can be read from a table when running a query.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
read_overflow_mode enum OverflowMode
Determines the behavior on exceeding limits while reading the data.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
max_rows_to_group_by google.protobuf.Int64Value
Limits the maximum number of unique keys received from aggregation function. This setting helps to reduce RAM consumption while doing aggregation.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
group_by_overflow_mode enum GroupByOverflowMode
Determines the behavior on exceeding limits while doing aggregation.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
  • any-perform approximate GROUP BY operation by continuing aggregation for the keys that got into the set, but don't add new keys to the set.
max_rows_to_sort google.protobuf.Int64Value
Limits the maximum number of rows that can be read from a table for sorting. This setting helps to reduce RAM consumption.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_bytes_to_sort google.protobuf.Int64Value
Limits the maximum number of bytes (uncompressed data) that can be read from a table for sorting. This setting helps to reduce RAM consumption.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
sort_overflow_mode enum OverflowMode
Determines the behavior on exceeding limits while sorting.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
max_result_rows google.protobuf.Int64Value
Limits the number of rows in the result. This limitation is also checked for subqueries and parts of distributed queries that run on remote servers.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_result_bytes google.protobuf.Int64Value
Limits the number of bytes in the result. This limitation is also checked for subqueries and parts of distributed queries that run on remote servers.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
result_overflow_mode enum OverflowMode
Determines the behavior on exceeding limits while forming result.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
max_rows_in_distinct google.protobuf.Int64Value
Limits the maximum number of different rows when using DISTINCT.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_bytes_in_distinct google.protobuf.Int64Value
Limits the maximum size of a hash table in bytes (uncompressed data) when using DISTINCT. The minimum value is 0.
distinct_overflow_mode enum OverflowMode
Determines the behavior on exceeding limits while doing DISCTINCT.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
max_rows_to_transfer google.protobuf.Int64Value
Limits the maximum number of rows that can be passed to a remote server or saved in a temporary table when using GLOBAL IN.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_bytes_to_transfer google.protobuf.Int64Value
Limits the maximum number of bytes (uncompressed data) that can be passed to a remote server or saved in a temporary table when using GLOBAL IN.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
transfer_overflow_mode enum OverflowMode
Determines the behavior on exceeding limits while doing transfers.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
max_execution_time google.protobuf.Int64Value
Limits the maximum query execution time in milliseconds. At this moment, this limitation is not checked when passing one of the sorting stages, as well as merging and finalizing aggregation funictions.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
timeout_overflow_mode enum OverflowMode
Determines the behavior on exceeding limits of execution time.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
max_rows_in_set google.protobuf.Int64Value
Limit on the number of rows in the set resulting from the execution of the IN section. The minimum value is 0.
max_bytes_in_set google.protobuf.Int64Value
Limit on the number of bytes in the set resulting from the execution of the IN section. The minimum value is 0.
set_overflow_mode enum OverflowMode
Determine the behavior on exceeding max_rows_in_set or max_bytes_in_set limit. Possible values: OVERFLOW_MODE_THROW, OVERFLOW_MODE_BREAK.
max_rows_in_join google.protobuf.Int64Value
Limit on maximum size of the hash table for JOIN, in rows. The minimum value is 0.
max_bytes_in_join google.protobuf.Int64Value
Limit on maximum size of the hash table for JOIN, in bytes. The minimum value is 0.
join_overflow_mode enum OverflowMode
Determine the behavior on exceeding max_rows_in_join or max_bytes_in_join limit. Possible values: OVERFLOW_MODE_THROW, OVERFLOW_MODE_BREAK.
max_columns_to_read google.protobuf.Int64Value
Limits the maximum number of columns that can be read from a table in a single query. If the query requires to read more columns to complete, then it will be aborted.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_temporary_columns google.protobuf.Int64Value
Limits the maximum number of temporary columns that must be kept in RAM at the same time when running a query, including constant columns.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_temporary_non_const_columns google.protobuf.Int64Value
Limits the maximum number of temporary columns that must be kept in RAM at the same time when running a query, excluding constant columns.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_query_size google.protobuf.Int64Value
Limits the size of the part of a query that can be transferred to RAM for parsing with the SQL parser, in bytes.
Value must be greater than 0 (default: 262144).
See in-depth description in ClickHouse documentation. Value must be greater than 0.
max_ast_depth google.protobuf.Int64Value
Limits the maximum depth of query syntax tree.
Executing a big and complex query may result in building a syntax tree of enormous depth. By using this setting, you can prohibit execution of over-sized or non-optimized queries for huge tables.
For example, the **SELECT *** query may result in more complex and deeper syntax tree, compared to the SELECT ... WHERE ... query, containing constraints and conditions, in the most cases. A user can be forced to construct more optimized queries, if this setting is used.
Value must be greater than 0 (default: 1000). If a too small value is set, it may render ClickHouse unable to execute even simple queries.
See in-depth description in ClickHouse documentation. Value must be greater than 0.
max_ast_elements google.protobuf.Int64Value
Limits the maximum size of query syntax tree in number of nodes.
Executing a big and complex query may result in building a syntax tree of enormous size. By using this setting, you can prohibit execution of over-sized or non-optimized queries for huge tables.
Value must be greater than 0 (default: 50000). If a too small value is set, it may render ClickHouse unable to execute even simple queries.
See in-depth description in ClickHouse documentation. Value must be greater than 0.
max_expanded_ast_elements google.protobuf.Int64Value
Limits the maximum size of query syntax tree in number of nodes after expansion of aliases and the asterisk values.
Executing a big and complex query may result in building a syntax tree of enormous size. By using this setting, you can prohibit execution of over-sized or non-optimized queries for huge tables.
Value must be greater than 0 (default: 500000). If a too small value is set, it may render ClickHouse unable to execute even simple queries. Value must be greater than 0.
min_execution_speed google.protobuf.Int64Value
Minimal execution speed in rows per second. The minimum value is 0.
min_execution_speed_bytes google.protobuf.Int64Value
Minimal execution speed in bytes per second. The minimum value is 0.
count_distinct_implementation enum CountDistinctImplementation
Aggregate function to use for implementation of count(DISTINCT ...).
input_format_values_interpret_expressions google.protobuf.BoolValue
Enables or disables SQL parser if the fast stream parser cannot parse the data.
Enable this setting, if the data that you want to insert into a table contains SQL expressions.
For example, the stream parser is unable to parse a value that contains now() expression; therefore an INSERT query for this value will fail and no data will be inserted into a table. With enabled SQL parser, this expression is parsed correctly: the now() expression will be parsed as SQL function, interpreted, and the current date and time will be inserted into the table as a result.
This setting has effect only if you use Values format when inserting data.
Default value: true (SQL parser is enabled).
See in-depth description in ClickHouse documentation.
input_format_defaults_for_omitted_fields google.protobuf.BoolValue
Enables or disables replacing omitted input values with default values of the respective columns when performing INSERT queries.
Default value: true (replacing is enabled).
output_format_json_quote_64bit_integers google.protobuf.BoolValue
Enables quoting of 64-bit integers in JSON output format.
If this setting is enabled, then 64-bit integers (UInt64 and Int64) will be quoted when written to JSON output in order to maintain compatibility with the most of the JavaScript engines. Otherwise, such integers will not be quoted.
Default value: false (quoting 64-bit integers is disabled).
output_format_json_quote_denormals google.protobuf.BoolValue
Enables special floating-point values (+nan, -nan, +inf and -inf) in JSON output format.
Default value: false (special values do not present in output).
low_cardinality_allow_in_native_format google.protobuf.BoolValue
Determines whether to use LowCardinality type in Native format.
  • true (default)-yes, use.
  • false-convert LowCardinality columns to regular columns when doing SELECT, and convert regular columns to LowCardinality when doing INSERT.

LowCardinality columns (aka sparse columns) store data in more effective way, compared to regular columns, by using hash tables. If data to insert suits this storage format, ClickHouse will place them into LowCardinality column.
If you use a third-party ClickHouse client that can't work with LowCardinality columns, then this client will not be able to correctly interpret the result of the query that asks for data stored in LowCardinality column. Disable this setting to convert LowCardinality column to regular column when creating the result, so such clients will be able to process the result.
Official ClickHouse client works with LowCardinality columns out-of-the-box.
Default value: true (LowCardinality columns are used in Native format).
empty_result_for_aggregation_by_empty_set google.protobuf.BoolValue
Enables returning of empty result when aggregating without keys (with GROUP BY operation absent) on empty set (e.g., SELECT count(*) FROM table WHERE 0).
  • true-ClickHouse will return an empty result for such queries.
  • false (default)-ClickHouse will return a single-line result consisting of NULL values for aggregation functions, in accordance with SQL standard.
joined_subquery_requires_alias google.protobuf.BoolValue
join_use_nulls google.protobuf.BoolValue
transform_null_in google.protobuf.BoolValue
http_connection_timeout google.protobuf.Int64Value
HTTP connection timeout, in milliseconds.
Value must be greater than 0 (default: 1000, 1 second).
http_receive_timeout google.protobuf.Int64Value
HTTP receive timeout, in milliseconds.
Value must be greater than 0 (default: 1800000, 1800 seconds, 30 minutes).
http_send_timeout google.protobuf.Int64Value
HTTP send timeout, in milliseconds.
Value must be greater than 0 (default: 1800000, 1800 seconds, 30 minutes).
enable_http_compression google.protobuf.BoolValue
Enables or disables data compression in HTTP responses.
By default, ClickHouse stores data compressed. When executing a query, its result is uncompressed. Use this setting to command ClickHouse to compress the result when sending it via HTTP.
Enable this setting and add the Accept-Encoding: HTTP header in a HTTP request to force compression of HTTP response from ClickHouse.
ClickHouse support the following compression methods: gzip, br and deflate.
Default value: false (compression is disabled).
See in-depth description in ClickHouse documentation.
send_progress_in_http_headers google.protobuf.BoolValue
Enables progress notifications using X-ClickHouse-Progress HTTP header.
Default value: false (notifications disabled).
http_headers_progress_interval google.protobuf.Int64Value
Minimum interval between progress notifications with X-ClickHouse-Progress HTTP header, in milliseconds.
Value must be greater than 0 (default: 100).
add_http_cors_header google.protobuf.BoolValue
Adds CORS header in HTTP responses.
Default value: false (header is not added).
quota_mode enum QuotaMode
Quota accounting mode. Possible values: QUOTA_MODE_DEFAULT, QUOTA_MODE_KEYED and QUOTA_MODE_KEYED_BY_IP.

UserQuota

Field Description
interval_duration google.protobuf.Int64Value
Duration of interval for quota in milliseconds. Minimal value is 1 second. The minimum value is 1000.
queries google.protobuf.Int64Value
The total number of queries. 0 - unlimited. The minimum value is 0.
errors google.protobuf.Int64Value
The number of queries that threw exception. 0 - unlimited. The minimum value is 0.
result_rows google.protobuf.Int64Value
The total number of rows given as the result.. 0 - unlimited. The minimum value is 0.
read_rows google.protobuf.Int64Value
The total number of source rows read from tables for running the query, on all remote servers. 0 - unlimited. The minimum value is 0.
execution_time google.protobuf.Int64Value
The total query execution time, in milliseconds (wall time). 0 - unlimited. The minimum value is 0.

Update

Updates the specified ClickHouse user.

rpc Update (UpdateUserRequest) returns (operation.Operation)

Metadata and response of Operation:

    Operation.metadata:UpdateUserMetadata

    Operation.response:User

UpdateUserRequest

Field Description
cluster_id string
Required. ID of the ClickHouse cluster the user belongs to. To get the cluster ID, use a ClusterService.List request. The maximum string length in characters is 50.
user_name string
Required. Name of the user to be updated. To get the name of the user, use a UserService.List request. The maximum string length in characters is 63. Value must match the regular expression [a-zA-Z0-9_]*.
update_mask google.protobuf.FieldMask
Field mask that specifies which attributes of the ClickHouse user should be updated.
password string
New password for the user. The string length in characters must be 8-128.
permissions[] Permission
New set of permissions for the user.
settings UserSettings
quotas[] UserQuota

Permission

Field Description
database_name string
Name of the database that the permission grants access to.

UserSettings

Field Description
readonly google.protobuf.Int64Value
Restricts permissions for non-DDL queries. To restrict permissions for DDL queries, use allow_ddl instead.
  • 0 (default)-no restrictions.
  • 1-only read data queries are allowed.
  • 2-read data and change settings queries are allowed.

See in-depth description in ClickHouse documentation. Acceptable values are 0 to 2, inclusive.
allow_ddl google.protobuf.BoolValue
Determines whether DDL queries are allowed (e.g., CREATE, ALTER, RENAME, etc).
Default value: true.
See in-depth description in ClickHouse documentation.
insert_quorum google.protobuf.Int64Value
Enables or disables write quorum for ClickHouse cluster. If the value is less than 2, then write quorum is disabled, otherwise it is enabled.
When used, write quorum guarantees that ClickHouse has written data to the quorum of insert_quorum replicas with no errors until the insert_quorum_timeout expires. All replicas in the quorum are in the consistent state, meaning that they contain linearized data from the previous INSERT queries. Employ write quorum, if you need the guarantees that the written data would not be lost in case of one or more replicas failure.
You can use select_sequential_consistency setting to read the data written with write quorum.
See in-depth description in ClickHouse documentation. The minimum value is 0.
connect_timeout google.protobuf.Int64Value
Connection timeout in milliseconds.
Value must be greater than 0 (default: 10000, 10 seconds). Value must be greater than 0.
receive_timeout google.protobuf.Int64Value
Receive timeout in milliseconds.
Value must be greater than 0 (default: 300000, 300 seconds or 5 minutes). Value must be greater than 0.
send_timeout google.protobuf.Int64Value
Send timeout in milliseconds.
Value must be greater than 0 (default: 300000, 300 seconds or 5 minutes). Value must be greater than 0.
insert_quorum_timeout google.protobuf.Int64Value
Quorum write timeout in milliseconds.
If the write quorum is enabled in the cluster, this timeout expires and some data is not written to the insert_quorum replicas, then ClickHouse will abort the execution of INSERT query and return an error. In this case, the client must send the query again to write the data block into the same or another replica.
Minimum value: 1000, 1 second (default: 60000, 1 minute). The minimum value is 1000.
select_sequential_consistency google.protobuf.BoolValue
Determines the behavior of SELECT queries from the replicated table: if enabled, ClickHouse will terminate a query with error message in case the replica does not have a chunk written with the quorum and will not read the parts that have not yet been written with the quorum.
Default value: false (sequential consistency is disabled).
max_replica_delay_for_distributed_queries google.protobuf.Int64Value
Max replica delay in milliseconds. If a replica lags more than the set value, this replica is not used and becomes a stale one.
Minimum value: 1000, 1 second (default: 300000, 300 seconds or 5 minutes).
See in-depth description in ClickHouse documentation. The minimum value is 1000.
fallback_to_stale_replicas_for_distributed_queries google.protobuf.BoolValue
Enables or disables query forcing to a stale replica in case the actual data is unavailable. If enabled, ClickHouse will choose the most up-to-date replica and force the query to use the data in this replica. This setting can be used when doing SELECT query from a distributed table that points to replicated tables.
Default value: true (query forcing is enabled).
See in-depth description in ClickHouse documentation.
replication_alter_partitions_sync google.protobuf.Int64Value
Wait mode for asynchronous actions in ALTER queries on replicated tables:
  • 0-do not wait for replicas.
  • 1-only wait for own execution (default).
  • 2-wait for all replicas.

See in-depth description in ClickHouse documentation. Acceptable values are 0 to 2, inclusive.
distributed_product_mode enum DistributedProductMode
Determine the behavior of distributed subqueries.
See in-depth description in ClickHouse documentation.
  • DISTRIBUTED_PRODUCT_MODE_DENY: Default value. Prohibits using these types of subqueries (returns the "Double-distributed in/JOIN subqueries is denied" exception).
  • DISTRIBUTED_PRODUCT_MODE_LOCAL: Replaces the database and table in the subquery with local ones for the destination server (shard), leaving the normal IN/JOIN.
  • DISTRIBUTED_PRODUCT_MODE_GLOBAL: Replaces the IN/JOIN query with GLOBAL IN/GLOBAL JOIN.
  • DISTRIBUTED_PRODUCT_MODE_ALLOW: Allows the use of these types of subqueries.
distributed_aggregation_memory_efficient google.protobuf.BoolValue
Enables of disables memory saving mode when doing distributed aggregation.
When ClickHouse works with a distributed query, external aggregation is done on remote servers. Enable this setting to achieve a smaller memory footprint on the server that sourced such a distributed query.
Default value: false (memory saving mode is disabled).
See in-depth description in ClickHouse documentation.
distributed_ddl_task_timeout google.protobuf.Int64Value
Timeout for DDL queries, in milliseconds.
skip_unavailable_shards google.protobuf.BoolValue
Enables or disables silent skipping of unavailable shards.
A shard is considered unavailable if all its replicas are also unavailable.
Default value: false (silent skipping is disabled).
See in-depth description in ClickHouse documentation.
compile google.protobuf.BoolValue
Enables or disables query compilation. If you execute a lot of structurally identical queries, then enable this setting. As a result, such queries may be executed faster due to use of queries' compiled parts.
Use this setting in combination with min_count_to_compile setting.
Default value: false (compilation is disabled).
See in-depth description in ClickHouse documentation.
min_count_to_compile google.protobuf.Int64Value
How many structurally identical queries ClickHouse has to encounter before they are compiled.
Minimum value: 0 (default: 3).
For the 0 value compilation is synchronous: a query waits for compilation process to complete prior to continuing execution. It is recommended to set this value only for testing purposes.
For all other values, compilation is asynchronous: the compilation process executes in a separate thread. When a compiled part of query is ready, it will be used by ClickHouse for eligible queries, including the ones that are currently running.
See in-depth description in ClickHouse documentation. The minimum value is 0.
compile_expressions google.protobuf.BoolValue
Enables or disables expression compilation. If you execute a lot of queries that contain identical expressions, then enable this setting. As a result, such queries may be executed faster due to use of compiled expressions.
Use this setting in combination with min_count_to_compile_expression setting.
Default value: false (expression compilation is disabled).
min_count_to_compile_expression google.protobuf.Int64Value
How many identical expressions ClickHouse has to encounter before they are compiled.
Minimum value: 0 (default: 3).
For the 0 value compilation is synchronous: a query waits for expression compilation process to complete prior to continuing execution. It is recommended to set this value only for testing purposes.
For all other values, compilation is asynchronous: the compilation process executes in a separate thread. When a compiled expression is ready, it will be used by ClickHouse for eligible queries, including the ones that are currently running. The minimum value is 0.
max_block_size google.protobuf.Int64Value
The maximum block size for reading.
Data in ClickHouse is organized and processed by blocks (block is a set of columns' parts). The internal processing cycles for a single block are efficient enough, but there are noticeable expenditures on each block.
This setting is a recommendation for size of block (in a count of rows) that should be loaded from tables.
Value must be greater than 0 (default: 65536).
See in-depth description in ClickHouse documentation. Value must be greater than 0.
min_insert_block_size_rows google.protobuf.Int64Value
Limits the minimum number of rows in a block to be inserted in a table by INSERT query. Blocks that are smaller than the specified value, will be squashed together into the bigger blocks.
Minimal value: 0, block squashing is disabled (default: 1048576). The minimum value is 0.
min_insert_block_size_bytes google.protobuf.Int64Value
Limits the minimum number of bytes in a block to be inserted in a table by INSERT query. Blocks that are smaller than the specified value, will be squashed together into the bigger blocks.
Minimal value: 0, block squashing is disabled (default: 268435456, 256 MB). The minimum value is 0.
max_insert_block_size google.protobuf.Int64Value
Allows to form blocks of the specified size (in bytes) when inserting data in a table. This setting has effect only if server is creating such blocks by itself.
Value must be greater than 0 (default: 1048576).
See in-depth description in ClickHouse documentation. Value must be greater than 0.
min_bytes_to_use_direct_io google.protobuf.Int64Value
Limits the minimum number of bytes to enable unbuffered direct reads from disk (Direct I/O).
By default, ClickHouse does not read data directly from disk, but relies on the filesystem and its cache instead. Such reading strategy is effective when the data volume is small. If the amount of the data to read is huge, it is more effective to read directly from the disk, bypassing the filesystem cache.
If the total amount of the data to read is greater than the value of this setting, then ClickHouse will fetch this data directly from the disk.
Minimal value and default value: 0, Direct I/O is disabled. The minimum value is 0.
use_uncompressed_cache google.protobuf.BoolValue
Determines whether to use the cache of uncompressed blocks, or not. Using this cache can significantly reduce latency and increase the throughput when a huge amount of small queries is to be processed. Enable this setting for the users who instantiates small queries frequently.
This setting has effect only for tables of the MergeTree family.
Default value: false (uncompressed cache is disabled).
See in-depth description in ClickHouse documentation.
merge_tree_max_rows_to_use_cache google.protobuf.Int64Value
Limits the maximum size in rows of the request that can use the cache of uncompressed data. The cache is not used for requests larger than the specified value.
Use this setting in combination with use_uncompressed_cache setting.
Value must be greater than 0 (default: 128x8192). Value must be greater than 0.
merge_tree_max_bytes_to_use_cache google.protobuf.Int64Value
Limits the maximum size in bytes of the request that can use the cache of uncompressed data. The cache is not used for requests larger than the specified value.
Use this setting in combination with use_uncompressed_cache setting.
Value must be greater than 0 (default: 192x10x1024x1024). Value must be greater than 0.
merge_tree_min_rows_for_concurrent_read google.protobuf.Int64Value
Limits the minimum number of rows to be read from a file to enable concurrent read. If the number of rows to be read exceeds this value, then ClickHouse will try to use a few threads to read from a file concurrently.
This setting has effect only for tables of the MergeTree family.
Value must be greater than 0 (default: 20x8192). Value must be greater than 0.
merge_tree_min_bytes_for_concurrent_read google.protobuf.Int64Value
Limits the number of bytes to be read from a file to enable concurrent read. If the number of bytes to be read exceeds this value, then ClickHouse will try to use a few threads to read from a file concurrently.
This setting has effect only for tables of the MergeTree family.
Value must be greater than 0 (default: 24x10x1024x1024). Value must be greater than 0.
max_bytes_before_external_group_by google.protobuf.Int64Value
Sets the threshold of RAM consumption (in bytes) after that the temporary data, collected during the GROUP BY operation, should be flushed to disk to limit the RAM comsumption.
By default, aggregation is done by employing hash table that resides in RAM. A query can result in aggregation of huge data volumes that can lead to memory exhaustion and abortion of the query (see the max_memory_usage setting). For such queries, you can use this setting to force ClickHouse to do flushing and complete aggregation successfully.
Minimal value and default value: 0, GROUP BY in the external memory is disabled.
When using aggregation in external memory, it is recommended to set the value of this setting twice as low as the max_memory_usage setting value (by default, the maximum memory usage is limited to ten gigabytes).
See in-depth description in ClickHouse documentation.
See also: the distributed_aggregation_memory_efficient setting.
max_bytes_before_external_sort google.protobuf.Int64Value
This setting is equivalent of the max_bytes_before_external_group_by setting, except for it is for sort operation (ORDER BY), not aggregation.
group_by_two_level_threshold google.protobuf.Int64Value
Sets the threshold of the number of keys, after that the two-level aggregation should be used.
Minimal value: 0, threshold is not set (default: 10000).
group_by_two_level_threshold_bytes google.protobuf.Int64Value
Sets the threshold of the number of bytes, after that the two-level aggregation should be used.
Minimal value: 0, threshold is not set (default: 100000000).
priority google.protobuf.Int64Value
Sets the priority of a query.
  • 0-priority is not used.
  • 1-the highest priority.
  • and so on. The higher the number, the lower a query's priority.

This setting should be set up for each query individually.
If ClickHouse is working with the high-priority queries, and a low-priority query enters, then the low-priority query is paused until higher-priority queries are completed.
Minimal value and default value: 0, priority is not used. The minimum value is 0.
max_threads google.protobuf.Int64Value
Limits the maximum number of threads to process the request (setting does not take threads that read data from remote servers into account).
This setting applies to threads that perform the same stages of the query processing pipeline in parallel.
Minimal value and default value: 0 (the thread number is calculated automatically based on the number of physical CPU cores, no HyperThreading cores are taken into account).
See in-depth description in ClickHouse documentation. Value must be greater than 0.
max_memory_usage google.protobuf.Int64Value
Limits the maximum memory usage (in bytes) for processing of a single user's query on a single server. This setting does not take server's free RAM amount or total RAM amount into account.
This limitation is enforced for any user's single query on a single server.
Minimal value: 0, no limitation is set. Value that is set in the ClickHouse default config file: 10737418240 (10 GB).
If you use max_bytes_before_external_group_by or max_bytes_before_external_sort setting, then it is recommended to set their values twice as low as max_memory_usage setting value.
See in-depth description in ClickHouse documentation. The minimum value is 0.
max_memory_usage_for_user google.protobuf.Int64Value
Limits the maximum memory usage (in bytes) for processing of user's queries on a single server. This setting does not take server's free RAM amount or total RAM amount into account.
This limitation is enforced for all queries that belong to one user and run simultaneously on a single server.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_network_bandwidth google.protobuf.Int64Value
The maximum speed of data exchange over the network in bytes per second for a query.
Minimal value and default value: 0, no limitation is set.
max_network_bandwidth_for_user google.protobuf.Int64Value
The maximum speed of data exchange over the network in bytes per second for all concurrently running user queries.
Minimal value and default value: 0, no limitation is set.
force_index_by_date google.protobuf.BoolValue
If enabled, query is not executed if the ClickHouse can't use index by date. This setting has effect only for tables of the MergeTree family.
Default value: false (setting is disabled, query executes even if ClickHouse can't use index by date).
See in-depth description in ClickHouse documentation.
force_primary_key google.protobuf.BoolValue
If enabled, query is not executed if the ClickHouse can't use index by primary key. This setting has effect only for tables of the MergeTree family.
Default value: false (setting is disabled, query executes even if ClickHouse can't use index by primary key).
See in-depth description in ClickHouse documentation.
max_rows_to_read google.protobuf.Int64Value
Limits the maximum number of rows that can be read from a table when running a query.
Minimal value and default value: 0, no limitation is set.
See in-depth description in ClickHouse documentation. The minimum value is 0.
max_bytes_to_read google.protobuf.Int64Value
Limits the maximum number of bytes (uncompressed data) that can be read from a table when running a query.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
read_overflow_mode enum OverflowMode
Determines the behavior on exceeding limits while reading the data.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
max_rows_to_group_by google.protobuf.Int64Value
Limits the maximum number of unique keys received from aggregation function. This setting helps to reduce RAM consumption while doing aggregation.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
group_by_overflow_mode enum GroupByOverflowMode
Determines the behavior on exceeding limits while doing aggregation.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
  • any-perform approximate GROUP BY operation by continuing aggregation for the keys that got into the set, but don't add new keys to the set.
max_rows_to_sort google.protobuf.Int64Value
Limits the maximum number of rows that can be read from a table for sorting. This setting helps to reduce RAM consumption.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_bytes_to_sort google.protobuf.Int64Value
Limits the maximum number of bytes (uncompressed data) that can be read from a table for sorting. This setting helps to reduce RAM consumption.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
sort_overflow_mode enum OverflowMode
Determines the behavior on exceeding limits while sorting.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
max_result_rows google.protobuf.Int64Value
Limits the number of rows in the result. This limitation is also checked for subqueries and parts of distributed queries that run on remote servers.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_result_bytes google.protobuf.Int64Value
Limits the number of bytes in the result. This limitation is also checked for subqueries and parts of distributed queries that run on remote servers.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
result_overflow_mode enum OverflowMode
Determines the behavior on exceeding limits while forming result.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
max_rows_in_distinct google.protobuf.Int64Value
Limits the maximum number of different rows when using DISTINCT.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_bytes_in_distinct google.protobuf.Int64Value
Limits the maximum size of a hash table in bytes (uncompressed data) when using DISTINCT. The minimum value is 0.
distinct_overflow_mode enum OverflowMode
Determines the behavior on exceeding limits while doing DISCTINCT.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
max_rows_to_transfer google.protobuf.Int64Value
Limits the maximum number of rows that can be passed to a remote server or saved in a temporary table when using GLOBAL IN.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_bytes_to_transfer google.protobuf.Int64Value
Limits the maximum number of bytes (uncompressed data) that can be passed to a remote server or saved in a temporary table when using GLOBAL IN.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
transfer_overflow_mode enum OverflowMode
Determines the behavior on exceeding limits while doing transfers.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
max_execution_time google.protobuf.Int64Value
Limits the maximum query execution time in milliseconds. At this moment, this limitation is not checked when passing one of the sorting stages, as well as merging and finalizing aggregation funictions.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
timeout_overflow_mode enum OverflowMode
Determines the behavior on exceeding limits of execution time.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
max_rows_in_set google.protobuf.Int64Value
Limit on the number of rows in the set resulting from the execution of the IN section. The minimum value is 0.
max_bytes_in_set google.protobuf.Int64Value
Limit on the number of bytes in the set resulting from the execution of the IN section. The minimum value is 0.
set_overflow_mode enum OverflowMode
Determine the behavior on exceeding max_rows_in_set or max_bytes_in_set limit. Possible values: OVERFLOW_MODE_THROW, OVERFLOW_MODE_BREAK.
max_rows_in_join google.protobuf.Int64Value
Limit on maximum size of the hash table for JOIN, in rows. The minimum value is 0.
max_bytes_in_join google.protobuf.Int64Value
Limit on maximum size of the hash table for JOIN, in bytes. The minimum value is 0.
join_overflow_mode enum OverflowMode
Determine the behavior on exceeding max_rows_in_join or max_bytes_in_join limit. Possible values: OVERFLOW_MODE_THROW, OVERFLOW_MODE_BREAK.
max_columns_to_read google.protobuf.Int64Value
Limits the maximum number of columns that can be read from a table in a single query. If the query requires to read more columns to complete, then it will be aborted.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_temporary_columns google.protobuf.Int64Value
Limits the maximum number of temporary columns that must be kept in RAM at the same time when running a query, including constant columns.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_temporary_non_const_columns google.protobuf.Int64Value
Limits the maximum number of temporary columns that must be kept in RAM at the same time when running a query, excluding constant columns.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_query_size google.protobuf.Int64Value
Limits the size of the part of a query that can be transferred to RAM for parsing with the SQL parser, in bytes.
Value must be greater than 0 (default: 262144).
See in-depth description in ClickHouse documentation. Value must be greater than 0.
max_ast_depth google.protobuf.Int64Value
Limits the maximum depth of query syntax tree.
Executing a big and complex query may result in building a syntax tree of enormous depth. By using this setting, you can prohibit execution of over-sized or non-optimized queries for huge tables.
For example, the **SELECT *** query may result in more complex and deeper syntax tree, compared to the SELECT ... WHERE ... query, containing constraints and conditions, in the most cases. A user can be forced to construct more optimized queries, if this setting is used.
Value must be greater than 0 (default: 1000). If a too small value is set, it may render ClickHouse unable to execute even simple queries.
See in-depth description in ClickHouse documentation. Value must be greater than 0.
max_ast_elements google.protobuf.Int64Value
Limits the maximum size of query syntax tree in number of nodes.
Executing a big and complex query may result in building a syntax tree of enormous size. By using this setting, you can prohibit execution of over-sized or non-optimized queries for huge tables.
Value must be greater than 0 (default: 50000). If a too small value is set, it may render ClickHouse unable to execute even simple queries.
See in-depth description in ClickHouse documentation. Value must be greater than 0.
max_expanded_ast_elements google.protobuf.Int64Value
Limits the maximum size of query syntax tree in number of nodes after expansion of aliases and the asterisk values.
Executing a big and complex query may result in building a syntax tree of enormous size. By using this setting, you can prohibit execution of over-sized or non-optimized queries for huge tables.
Value must be greater than 0 (default: 500000). If a too small value is set, it may render ClickHouse unable to execute even simple queries. Value must be greater than 0.
min_execution_speed google.protobuf.Int64Value
Minimal execution speed in rows per second. The minimum value is 0.
min_execution_speed_bytes google.protobuf.Int64Value
Minimal execution speed in bytes per second. The minimum value is 0.
count_distinct_implementation enum CountDistinctImplementation
Aggregate function to use for implementation of count(DISTINCT ...).
input_format_values_interpret_expressions google.protobuf.BoolValue
Enables or disables SQL parser if the fast stream parser cannot parse the data.
Enable this setting, if the data that you want to insert into a table contains SQL expressions.
For example, the stream parser is unable to parse a value that contains now() expression; therefore an INSERT query for this value will fail and no data will be inserted into a table. With enabled SQL parser, this expression is parsed correctly: the now() expression will be parsed as SQL function, interpreted, and the current date and time will be inserted into the table as a result.
This setting has effect only if you use Values format when inserting data.
Default value: true (SQL parser is enabled).
See in-depth description in ClickHouse documentation.
input_format_defaults_for_omitted_fields google.protobuf.BoolValue
Enables or disables replacing omitted input values with default values of the respective columns when performing INSERT queries.
Default value: true (replacing is enabled).
output_format_json_quote_64bit_integers google.protobuf.BoolValue
Enables quoting of 64-bit integers in JSON output format.
If this setting is enabled, then 64-bit integers (UInt64 and Int64) will be quoted when written to JSON output in order to maintain compatibility with the most of the JavaScript engines. Otherwise, such integers will not be quoted.
Default value: false (quoting 64-bit integers is disabled).
output_format_json_quote_denormals google.protobuf.BoolValue
Enables special floating-point values (+nan, -nan, +inf and -inf) in JSON output format.
Default value: false (special values do not present in output).
low_cardinality_allow_in_native_format google.protobuf.BoolValue
Determines whether to use LowCardinality type in Native format.
  • true (default)-yes, use.
  • false-convert LowCardinality columns to regular columns when doing SELECT, and convert regular columns to LowCardinality when doing INSERT.

LowCardinality columns (aka sparse columns) store data in more effective way, compared to regular columns, by using hash tables. If data to insert suits this storage format, ClickHouse will place them into LowCardinality column.
If you use a third-party ClickHouse client that can't work with LowCardinality columns, then this client will not be able to correctly interpret the result of the query that asks for data stored in LowCardinality column. Disable this setting to convert LowCardinality column to regular column when creating the result, so such clients will be able to process the result.
Official ClickHouse client works with LowCardinality columns out-of-the-box.
Default value: true (LowCardinality columns are used in Native format).
empty_result_for_aggregation_by_empty_set google.protobuf.BoolValue
Enables returning of empty result when aggregating without keys (with GROUP BY operation absent) on empty set (e.g., SELECT count(*) FROM table WHERE 0).
  • true-ClickHouse will return an empty result for such queries.
  • false (default)-ClickHouse will return a single-line result consisting of NULL values for aggregation functions, in accordance with SQL standard.
joined_subquery_requires_alias google.protobuf.BoolValue
join_use_nulls google.protobuf.BoolValue
transform_null_in google.protobuf.BoolValue
http_connection_timeout google.protobuf.Int64Value
HTTP connection timeout, in milliseconds.
Value must be greater than 0 (default: 1000, 1 second).
http_receive_timeout google.protobuf.Int64Value
HTTP receive timeout, in milliseconds.
Value must be greater than 0 (default: 1800000, 1800 seconds, 30 minutes).
http_send_timeout google.protobuf.Int64Value
HTTP send timeout, in milliseconds.
Value must be greater than 0 (default: 1800000, 1800 seconds, 30 minutes).
enable_http_compression google.protobuf.BoolValue
Enables or disables data compression in HTTP responses.
By default, ClickHouse stores data compressed. When executing a query, its result is uncompressed. Use this setting to command ClickHouse to compress the result when sending it via HTTP.
Enable this setting and add the Accept-Encoding: HTTP header in a HTTP request to force compression of HTTP response from ClickHouse.
ClickHouse support the following compression methods: gzip, br and deflate.
Default value: false (compression is disabled).
See in-depth description in ClickHouse documentation.
send_progress_in_http_headers google.protobuf.BoolValue
Enables progress notifications using X-ClickHouse-Progress HTTP header.
Default value: false (notifications disabled).
http_headers_progress_interval google.protobuf.Int64Value
Minimum interval between progress notifications with X-ClickHouse-Progress HTTP header, in milliseconds.
Value must be greater than 0 (default: 100).
add_http_cors_header google.protobuf.BoolValue
Adds CORS header in HTTP responses.
Default value: false (header is not added).
quota_mode enum QuotaMode
Quota accounting mode. Possible values: QUOTA_MODE_DEFAULT, QUOTA_MODE_KEYED and QUOTA_MODE_KEYED_BY_IP.

UserQuota

Field Description
interval_duration google.protobuf.Int64Value
Duration of interval for quota in milliseconds. Minimal value is 1 second. The minimum value is 1000.
queries google.protobuf.Int64Value
The total number of queries. 0 - unlimited. The minimum value is 0.
errors google.protobuf.Int64Value
The number of queries that threw exception. 0 - unlimited. The minimum value is 0.
result_rows google.protobuf.Int64Value
The total number of rows given as the result.. 0 - unlimited. The minimum value is 0.
read_rows google.protobuf.Int64Value
The total number of source rows read from tables for running the query, on all remote servers. 0 - unlimited. The minimum value is 0.
execution_time google.protobuf.Int64Value
The total query execution time, in milliseconds (wall time). 0 - unlimited. The minimum value is 0.

Operation

Field Description
id string
ID of the operation.
description string
Description of the operation. 0-256 characters long.
created_at google.protobuf.Timestamp
Creation timestamp.
created_by string
ID of the user or service account who initiated the operation.
modified_at google.protobuf.Timestamp
The time when the Operation resource was last modified.
done bool
If the value is false, it means the operation is still in progress. If true, the operation is completed, and either error or response is available.
metadata google.protobuf.Any<UpdateUserMetadata>
Service-specific metadata associated with the operation. It typically contains the ID of the target resource that the operation is performed on. Any method that returns a long-running operation should document the metadata type, if any.
result oneof: error or response
The operation result. If done == false and there was no failure detected, neither error nor response is set. If done == false and there was a failure detected, error is set. If done == true, exactly one of error or response is set.
  error google.rpc.Status
The error result of the operation in case of failure or cancellation.
  response google.protobuf.Any<User>
if operation finished successfully.

UpdateUserMetadata

Field Description
cluster_id string
ID of the ClickHouse cluster the user belongs to.
user_name string
Name of the user that is being updated.

User

Field Description
name string
Name of the ClickHouse user.
cluster_id string
ID of the ClickHouse cluster the user belongs to.
permissions[] Permission
Set of permissions granted to the user.
settings UserSettings
quotas[] UserQuota
Set of quotas assigned to the user.

Permission

Field Description
database_name string
Name of the database that the permission grants access to.

UserSettings

Field Description
readonly google.protobuf.Int64Value
Restricts permissions for non-DDL queries. To restrict permissions for DDL queries, use allow_ddl instead.
  • 0 (default)-no restrictions.
  • 1-only read data queries are allowed.
  • 2-read data and change settings queries are allowed.

See in-depth description in ClickHouse documentation. Acceptable values are 0 to 2, inclusive.
allow_ddl google.protobuf.BoolValue
Determines whether DDL queries are allowed (e.g., CREATE, ALTER, RENAME, etc).
Default value: true.
See in-depth description in ClickHouse documentation.
insert_quorum google.protobuf.Int64Value
Enables or disables write quorum for ClickHouse cluster. If the value is less than 2, then write quorum is disabled, otherwise it is enabled.
When used, write quorum guarantees that ClickHouse has written data to the quorum of insert_quorum replicas with no errors until the insert_quorum_timeout expires. All replicas in the quorum are in the consistent state, meaning that they contain linearized data from the previous INSERT queries. Employ write quorum, if you need the guarantees that the written data would not be lost in case of one or more replicas failure.
You can use select_sequential_consistency setting to read the data written with write quorum.
See in-depth description in ClickHouse documentation. The minimum value is 0.
connect_timeout google.protobuf.Int64Value
Connection timeout in milliseconds.
Value must be greater than 0 (default: 10000, 10 seconds). Value must be greater than 0.
receive_timeout google.protobuf.Int64Value
Receive timeout in milliseconds.
Value must be greater than 0 (default: 300000, 300 seconds or 5 minutes). Value must be greater than 0.
send_timeout google.protobuf.Int64Value
Send timeout in milliseconds.
Value must be greater than 0 (default: 300000, 300 seconds or 5 minutes). Value must be greater than 0.
insert_quorum_timeout google.protobuf.Int64Value
Quorum write timeout in milliseconds.
If the write quorum is enabled in the cluster, this timeout expires and some data is not written to the insert_quorum replicas, then ClickHouse will abort the execution of INSERT query and return an error. In this case, the client must send the query again to write the data block into the same or another replica.
Minimum value: 1000, 1 second (default: 60000, 1 minute). The minimum value is 1000.
select_sequential_consistency google.protobuf.BoolValue
Determines the behavior of SELECT queries from the replicated table: if enabled, ClickHouse will terminate a query with error message in case the replica does not have a chunk written with the quorum and will not read the parts that have not yet been written with the quorum.
Default value: false (sequential consistency is disabled).
max_replica_delay_for_distributed_queries google.protobuf.Int64Value
Max replica delay in milliseconds. If a replica lags more than the set value, this replica is not used and becomes a stale one.
Minimum value: 1000, 1 second (default: 300000, 300 seconds or 5 minutes).
See in-depth description in ClickHouse documentation. The minimum value is 1000.
fallback_to_stale_replicas_for_distributed_queries google.protobuf.BoolValue
Enables or disables query forcing to a stale replica in case the actual data is unavailable. If enabled, ClickHouse will choose the most up-to-date replica and force the query to use the data in this replica. This setting can be used when doing SELECT query from a distributed table that points to replicated tables.
Default value: true (query forcing is enabled).
See in-depth description in ClickHouse documentation.
replication_alter_partitions_sync google.protobuf.Int64Value
Wait mode for asynchronous actions in ALTER queries on replicated tables:
  • 0-do not wait for replicas.
  • 1-only wait for own execution (default).
  • 2-wait for all replicas.

See in-depth description in ClickHouse documentation. Acceptable values are 0 to 2, inclusive.
distributed_product_mode enum DistributedProductMode
Determine the behavior of distributed subqueries.
See in-depth description in ClickHouse documentation.
  • DISTRIBUTED_PRODUCT_MODE_DENY: Default value. Prohibits using these types of subqueries (returns the "Double-distributed in/JOIN subqueries is denied" exception).
  • DISTRIBUTED_PRODUCT_MODE_LOCAL: Replaces the database and table in the subquery with local ones for the destination server (shard), leaving the normal IN/JOIN.
  • DISTRIBUTED_PRODUCT_MODE_GLOBAL: Replaces the IN/JOIN query with GLOBAL IN/GLOBAL JOIN.
  • DISTRIBUTED_PRODUCT_MODE_ALLOW: Allows the use of these types of subqueries.
distributed_aggregation_memory_efficient google.protobuf.BoolValue
Enables of disables memory saving mode when doing distributed aggregation.
When ClickHouse works with a distributed query, external aggregation is done on remote servers. Enable this setting to achieve a smaller memory footprint on the server that sourced such a distributed query.
Default value: false (memory saving mode is disabled).
See in-depth description in ClickHouse documentation.
distributed_ddl_task_timeout google.protobuf.Int64Value
Timeout for DDL queries, in milliseconds.
skip_unavailable_shards google.protobuf.BoolValue
Enables or disables silent skipping of unavailable shards.
A shard is considered unavailable if all its replicas are also unavailable.
Default value: false (silent skipping is disabled).
See in-depth description in ClickHouse documentation.
compile google.protobuf.BoolValue
Enables or disables query compilation. If you execute a lot of structurally identical queries, then enable this setting. As a result, such queries may be executed faster due to use of queries' compiled parts.
Use this setting in combination with min_count_to_compile setting.
Default value: false (compilation is disabled).
See in-depth description in ClickHouse documentation.
min_count_to_compile google.protobuf.Int64Value
How many structurally identical queries ClickHouse has to encounter before they are compiled.
Minimum value: 0 (default: 3).
For the 0 value compilation is synchronous: a query waits for compilation process to complete prior to continuing execution. It is recommended to set this value only for testing purposes.
For all other values, compilation is asynchronous: the compilation process executes in a separate thread. When a compiled part of query is ready, it will be used by ClickHouse for eligible queries, including the ones that are currently running.
See in-depth description in ClickHouse documentation. The minimum value is 0.
compile_expressions google.protobuf.BoolValue
Enables or disables expression compilation. If you execute a lot of queries that contain identical expressions, then enable this setting. As a result, such queries may be executed faster due to use of compiled expressions.
Use this setting in combination with min_count_to_compile_expression setting.
Default value: false (expression compilation is disabled).
min_count_to_compile_expression google.protobuf.Int64Value
How many identical expressions ClickHouse has to encounter before they are compiled.
Minimum value: 0 (default: 3).
For the 0 value compilation is synchronous: a query waits for expression compilation process to complete prior to continuing execution. It is recommended to set this value only for testing purposes.
For all other values, compilation is asynchronous: the compilation process executes in a separate thread. When a compiled expression is ready, it will be used by ClickHouse for eligible queries, including the ones that are currently running. The minimum value is 0.
max_block_size google.protobuf.Int64Value
The maximum block size for reading.
Data in ClickHouse is organized and processed by blocks (block is a set of columns' parts). The internal processing cycles for a single block are efficient enough, but there are noticeable expenditures on each block.
This setting is a recommendation for size of block (in a count of rows) that should be loaded from tables.
Value must be greater than 0 (default: 65536).
See in-depth description in ClickHouse documentation. Value must be greater than 0.
min_insert_block_size_rows google.protobuf.Int64Value
Limits the minimum number of rows in a block to be inserted in a table by INSERT query. Blocks that are smaller than the specified value, will be squashed together into the bigger blocks.
Minimal value: 0, block squashing is disabled (default: 1048576). The minimum value is 0.
min_insert_block_size_bytes google.protobuf.Int64Value
Limits the minimum number of bytes in a block to be inserted in a table by INSERT query. Blocks that are smaller than the specified value, will be squashed together into the bigger blocks.
Minimal value: 0, block squashing is disabled (default: 268435456, 256 MB). The minimum value is 0.
max_insert_block_size google.protobuf.Int64Value
Allows to form blocks of the specified size (in bytes) when inserting data in a table. This setting has effect only if server is creating such blocks by itself.
Value must be greater than 0 (default: 1048576).
See in-depth description in ClickHouse documentation. Value must be greater than 0.
min_bytes_to_use_direct_io google.protobuf.Int64Value
Limits the minimum number of bytes to enable unbuffered direct reads from disk (Direct I/O).
By default, ClickHouse does not read data directly from disk, but relies on the filesystem and its cache instead. Such reading strategy is effective when the data volume is small. If the amount of the data to read is huge, it is more effective to read directly from the disk, bypassing the filesystem cache.
If the total amount of the data to read is greater than the value of this setting, then ClickHouse will fetch this data directly from the disk.
Minimal value and default value: 0, Direct I/O is disabled. The minimum value is 0.
use_uncompressed_cache google.protobuf.BoolValue
Determines whether to use the cache of uncompressed blocks, or not. Using this cache can significantly reduce latency and increase the throughput when a huge amount of small queries is to be processed. Enable this setting for the users who instantiates small queries frequently.
This setting has effect only for tables of the MergeTree family.
Default value: false (uncompressed cache is disabled).
See in-depth description in ClickHouse documentation.
merge_tree_max_rows_to_use_cache google.protobuf.Int64Value
Limits the maximum size in rows of the request that can use the cache of uncompressed data. The cache is not used for requests larger than the specified value.
Use this setting in combination with use_uncompressed_cache setting.
Value must be greater than 0 (default: 128x8192). Value must be greater than 0.
merge_tree_max_bytes_to_use_cache google.protobuf.Int64Value
Limits the maximum size in bytes of the request that can use the cache of uncompressed data. The cache is not used for requests larger than the specified value.
Use this setting in combination with use_uncompressed_cache setting.
Value must be greater than 0 (default: 192x10x1024x1024). Value must be greater than 0.
merge_tree_min_rows_for_concurrent_read google.protobuf.Int64Value
Limits the minimum number of rows to be read from a file to enable concurrent read. If the number of rows to be read exceeds this value, then ClickHouse will try to use a few threads to read from a file concurrently.
This setting has effect only for tables of the MergeTree family.
Value must be greater than 0 (default: 20x8192). Value must be greater than 0.
merge_tree_min_bytes_for_concurrent_read google.protobuf.Int64Value
Limits the number of bytes to be read from a file to enable concurrent read. If the number of bytes to be read exceeds this value, then ClickHouse will try to use a few threads to read from a file concurrently.
This setting has effect only for tables of the MergeTree family.
Value must be greater than 0 (default: 24x10x1024x1024). Value must be greater than 0.
max_bytes_before_external_group_by google.protobuf.Int64Value
Sets the threshold of RAM consumption (in bytes) after that the temporary data, collected during the GROUP BY operation, should be flushed to disk to limit the RAM comsumption.
By default, aggregation is done by employing hash table that resides in RAM. A query can result in aggregation of huge data volumes that can lead to memory exhaustion and abortion of the query (see the max_memory_usage setting). For such queries, you can use this setting to force ClickHouse to do flushing and complete aggregation successfully.
Minimal value and default value: 0, GROUP BY in the external memory is disabled.
When using aggregation in external memory, it is recommended to set the value of this setting twice as low as the max_memory_usage setting value (by default, the maximum memory usage is limited to ten gigabytes).
See in-depth description in ClickHouse documentation.
See also: the distributed_aggregation_memory_efficient setting.
max_bytes_before_external_sort google.protobuf.Int64Value
This setting is equivalent of the max_bytes_before_external_group_by setting, except for it is for sort operation (ORDER BY), not aggregation.
group_by_two_level_threshold google.protobuf.Int64Value
Sets the threshold of the number of keys, after that the two-level aggregation should be used.
Minimal value: 0, threshold is not set (default: 10000).
group_by_two_level_threshold_bytes google.protobuf.Int64Value
Sets the threshold of the number of bytes, after that the two-level aggregation should be used.
Minimal value: 0, threshold is not set (default: 100000000).
priority google.protobuf.Int64Value
Sets the priority of a query.
  • 0-priority is not used.
  • 1-the highest priority.
  • and so on. The higher the number, the lower a query's priority.

This setting should be set up for each query individually.
If ClickHouse is working with the high-priority queries, and a low-priority query enters, then the low-priority query is paused until higher-priority queries are completed.
Minimal value and default value: 0, priority is not used. The minimum value is 0.
max_threads google.protobuf.Int64Value
Limits the maximum number of threads to process the request (setting does not take threads that read data from remote servers into account).
This setting applies to threads that perform the same stages of the query processing pipeline in parallel.
Minimal value and default value: 0 (the thread number is calculated automatically based on the number of physical CPU cores, no HyperThreading cores are taken into account).
See in-depth description in ClickHouse documentation. Value must be greater than 0.
max_memory_usage google.protobuf.Int64Value
Limits the maximum memory usage (in bytes) for processing of a single user's query on a single server. This setting does not take server's free RAM amount or total RAM amount into account.
This limitation is enforced for any user's single query on a single server.
Minimal value: 0, no limitation is set. Value that is set in the ClickHouse default config file: 10737418240 (10 GB).
If you use max_bytes_before_external_group_by or max_bytes_before_external_sort setting, then it is recommended to set their values twice as low as max_memory_usage setting value.
See in-depth description in ClickHouse documentation. The minimum value is 0.
max_memory_usage_for_user google.protobuf.Int64Value
Limits the maximum memory usage (in bytes) for processing of user's queries on a single server. This setting does not take server's free RAM amount or total RAM amount into account.
This limitation is enforced for all queries that belong to one user and run simultaneously on a single server.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_network_bandwidth google.protobuf.Int64Value
The maximum speed of data exchange over the network in bytes per second for a query.
Minimal value and default value: 0, no limitation is set.
max_network_bandwidth_for_user google.protobuf.Int64Value
The maximum speed of data exchange over the network in bytes per second for all concurrently running user queries.
Minimal value and default value: 0, no limitation is set.
force_index_by_date google.protobuf.BoolValue
If enabled, query is not executed if the ClickHouse can't use index by date. This setting has effect only for tables of the MergeTree family.
Default value: false (setting is disabled, query executes even if ClickHouse can't use index by date).
See in-depth description in ClickHouse documentation.
force_primary_key google.protobuf.BoolValue
If enabled, query is not executed if the ClickHouse can't use index by primary key. This setting has effect only for tables of the MergeTree family.
Default value: false (setting is disabled, query executes even if ClickHouse can't use index by primary key).
See in-depth description in ClickHouse documentation.
max_rows_to_read google.protobuf.Int64Value
Limits the maximum number of rows that can be read from a table when running a query.
Minimal value and default value: 0, no limitation is set.
See in-depth description in ClickHouse documentation. The minimum value is 0.
max_bytes_to_read google.protobuf.Int64Value
Limits the maximum number of bytes (uncompressed data) that can be read from a table when running a query.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
read_overflow_mode enum OverflowMode
Determines the behavior on exceeding limits while reading the data.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
max_rows_to_group_by google.protobuf.Int64Value
Limits the maximum number of unique keys received from aggregation function. This setting helps to reduce RAM consumption while doing aggregation.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
group_by_overflow_mode enum GroupByOverflowMode
Determines the behavior on exceeding limits while doing aggregation.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
  • any-perform approximate GROUP BY operation by continuing aggregation for the keys that got into the set, but don't add new keys to the set.
max_rows_to_sort google.protobuf.Int64Value
Limits the maximum number of rows that can be read from a table for sorting. This setting helps to reduce RAM consumption.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_bytes_to_sort google.protobuf.Int64Value
Limits the maximum number of bytes (uncompressed data) that can be read from a table for sorting. This setting helps to reduce RAM consumption.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
sort_overflow_mode enum OverflowMode
Determines the behavior on exceeding limits while sorting.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
max_result_rows google.protobuf.Int64Value
Limits the number of rows in the result. This limitation is also checked for subqueries and parts of distributed queries that run on remote servers.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_result_bytes google.protobuf.Int64Value
Limits the number of bytes in the result. This limitation is also checked for subqueries and parts of distributed queries that run on remote servers.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
result_overflow_mode enum OverflowMode
Determines the behavior on exceeding limits while forming result.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
max_rows_in_distinct google.protobuf.Int64Value
Limits the maximum number of different rows when using DISTINCT.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_bytes_in_distinct google.protobuf.Int64Value
Limits the maximum size of a hash table in bytes (uncompressed data) when using DISTINCT. The minimum value is 0.
distinct_overflow_mode enum OverflowMode
Determines the behavior on exceeding limits while doing DISCTINCT.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
max_rows_to_transfer google.protobuf.Int64Value
Limits the maximum number of rows that can be passed to a remote server or saved in a temporary table when using GLOBAL IN.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_bytes_to_transfer google.protobuf.Int64Value
Limits the maximum number of bytes (uncompressed data) that can be passed to a remote server or saved in a temporary table when using GLOBAL IN.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
transfer_overflow_mode enum OverflowMode
Determines the behavior on exceeding limits while doing transfers.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
max_execution_time google.protobuf.Int64Value
Limits the maximum query execution time in milliseconds. At this moment, this limitation is not checked when passing one of the sorting stages, as well as merging and finalizing aggregation funictions.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
timeout_overflow_mode enum OverflowMode
Determines the behavior on exceeding limits of execution time.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
max_rows_in_set google.protobuf.Int64Value
Limit on the number of rows in the set resulting from the execution of the IN section. The minimum value is 0.
max_bytes_in_set google.protobuf.Int64Value
Limit on the number of bytes in the set resulting from the execution of the IN section. The minimum value is 0.
set_overflow_mode enum OverflowMode
Determine the behavior on exceeding max_rows_in_set or max_bytes_in_set limit. Possible values: OVERFLOW_MODE_THROW, OVERFLOW_MODE_BREAK.
max_rows_in_join google.protobuf.Int64Value
Limit on maximum size of the hash table for JOIN, in rows. The minimum value is 0.
max_bytes_in_join google.protobuf.Int64Value
Limit on maximum size of the hash table for JOIN, in bytes. The minimum value is 0.
join_overflow_mode enum OverflowMode
Determine the behavior on exceeding max_rows_in_join or max_bytes_in_join limit. Possible values: OVERFLOW_MODE_THROW, OVERFLOW_MODE_BREAK.
max_columns_to_read google.protobuf.Int64Value
Limits the maximum number of columns that can be read from a table in a single query. If the query requires to read more columns to complete, then it will be aborted.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_temporary_columns google.protobuf.Int64Value
Limits the maximum number of temporary columns that must be kept in RAM at the same time when running a query, including constant columns.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_temporary_non_const_columns google.protobuf.Int64Value
Limits the maximum number of temporary columns that must be kept in RAM at the same time when running a query, excluding constant columns.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_query_size google.protobuf.Int64Value
Limits the size of the part of a query that can be transferred to RAM for parsing with the SQL parser, in bytes.
Value must be greater than 0 (default: 262144).
See in-depth description in ClickHouse documentation. Value must be greater than 0.
max_ast_depth google.protobuf.Int64Value
Limits the maximum depth of query syntax tree.
Executing a big and complex query may result in building a syntax tree of enormous depth. By using this setting, you can prohibit execution of over-sized or non-optimized queries for huge tables.
For example, the **SELECT *** query may result in more complex and deeper syntax tree, compared to the SELECT ... WHERE ... query, containing constraints and conditions, in the most cases. A user can be forced to construct more optimized queries, if this setting is used.
Value must be greater than 0 (default: 1000). If a too small value is set, it may render ClickHouse unable to execute even simple queries.
See in-depth description in ClickHouse documentation. Value must be greater than 0.
max_ast_elements google.protobuf.Int64Value
Limits the maximum size of query syntax tree in number of nodes.
Executing a big and complex query may result in building a syntax tree of enormous size. By using this setting, you can prohibit execution of over-sized or non-optimized queries for huge tables.
Value must be greater than 0 (default: 50000). If a too small value is set, it may render ClickHouse unable to execute even simple queries.
See in-depth description in ClickHouse documentation. Value must be greater than 0.
max_expanded_ast_elements google.protobuf.Int64Value
Limits the maximum size of query syntax tree in number of nodes after expansion of aliases and the asterisk values.
Executing a big and complex query may result in building a syntax tree of enormous size. By using this setting, you can prohibit execution of over-sized or non-optimized queries for huge tables.
Value must be greater than 0 (default: 500000). If a too small value is set, it may render ClickHouse unable to execute even simple queries. Value must be greater than 0.
min_execution_speed google.protobuf.Int64Value
Minimal execution speed in rows per second. The minimum value is 0.
min_execution_speed_bytes google.protobuf.Int64Value
Minimal execution speed in bytes per second. The minimum value is 0.
count_distinct_implementation enum CountDistinctImplementation
Aggregate function to use for implementation of count(DISTINCT ...).
input_format_values_interpret_expressions google.protobuf.BoolValue
Enables or disables SQL parser if the fast stream parser cannot parse the data.
Enable this setting, if the data that you want to insert into a table contains SQL expressions.
For example, the stream parser is unable to parse a value that contains now() expression; therefore an INSERT query for this value will fail and no data will be inserted into a table. With enabled SQL parser, this expression is parsed correctly: the now() expression will be parsed as SQL function, interpreted, and the current date and time will be inserted into the table as a result.
This setting has effect only if you use Values format when inserting data.
Default value: true (SQL parser is enabled).
See in-depth description in ClickHouse documentation.
input_format_defaults_for_omitted_fields google.protobuf.BoolValue
Enables or disables replacing omitted input values with default values of the respective columns when performing INSERT queries.
Default value: true (replacing is enabled).
output_format_json_quote_64bit_integers google.protobuf.BoolValue
Enables quoting of 64-bit integers in JSON output format.
If this setting is enabled, then 64-bit integers (UInt64 and Int64) will be quoted when written to JSON output in order to maintain compatibility with the most of the JavaScript engines. Otherwise, such integers will not be quoted.
Default value: false (quoting 64-bit integers is disabled).
output_format_json_quote_denormals google.protobuf.BoolValue
Enables special floating-point values (+nan, -nan, +inf and -inf) in JSON output format.
Default value: false (special values do not present in output).
low_cardinality_allow_in_native_format google.protobuf.BoolValue
Determines whether to use LowCardinality type in Native format.
  • true (default)-yes, use.
  • false-convert LowCardinality columns to regular columns when doing SELECT, and convert regular columns to LowCardinality when doing INSERT.

LowCardinality columns (aka sparse columns) store data in more effective way, compared to regular columns, by using hash tables. If data to insert suits this storage format, ClickHouse will place them into LowCardinality column.
If you use a third-party ClickHouse client that can't work with LowCardinality columns, then this client will not be able to correctly interpret the result of the query that asks for data stored in LowCardinality column. Disable this setting to convert LowCardinality column to regular column when creating the result, so such clients will be able to process the result.
Official ClickHouse client works with LowCardinality columns out-of-the-box.
Default value: true (LowCardinality columns are used in Native format).
empty_result_for_aggregation_by_empty_set google.protobuf.BoolValue
Enables returning of empty result when aggregating without keys (with GROUP BY operation absent) on empty set (e.g., SELECT count(*) FROM table WHERE 0).
  • true-ClickHouse will return an empty result for such queries.
  • false (default)-ClickHouse will return a single-line result consisting of NULL values for aggregation functions, in accordance with SQL standard.
joined_subquery_requires_alias google.protobuf.BoolValue
join_use_nulls google.protobuf.BoolValue
transform_null_in google.protobuf.BoolValue
http_connection_timeout google.protobuf.Int64Value
HTTP connection timeout, in milliseconds.
Value must be greater than 0 (default: 1000, 1 second).
http_receive_timeout google.protobuf.Int64Value
HTTP receive timeout, in milliseconds.
Value must be greater than 0 (default: 1800000, 1800 seconds, 30 minutes).
http_send_timeout google.protobuf.Int64Value
HTTP send timeout, in milliseconds.
Value must be greater than 0 (default: 1800000, 1800 seconds, 30 minutes).
enable_http_compression google.protobuf.BoolValue
Enables or disables data compression in HTTP responses.
By default, ClickHouse stores data compressed. When executing a query, its result is uncompressed. Use this setting to command ClickHouse to compress the result when sending it via HTTP.
Enable this setting and add the Accept-Encoding: HTTP header in a HTTP request to force compression of HTTP response from ClickHouse.
ClickHouse support the following compression methods: gzip, br and deflate.
Default value: false (compression is disabled).
See in-depth description in ClickHouse documentation.
send_progress_in_http_headers google.protobuf.BoolValue
Enables progress notifications using X-ClickHouse-Progress HTTP header.
Default value: false (notifications disabled).
http_headers_progress_interval google.protobuf.Int64Value
Minimum interval between progress notifications with X-ClickHouse-Progress HTTP header, in milliseconds.
Value must be greater than 0 (default: 100).
add_http_cors_header google.protobuf.BoolValue
Adds CORS header in HTTP responses.
Default value: false (header is not added).
quota_mode enum QuotaMode
Quota accounting mode. Possible values: QUOTA_MODE_DEFAULT, QUOTA_MODE_KEYED and QUOTA_MODE_KEYED_BY_IP.

UserQuota

Field Description
interval_duration google.protobuf.Int64Value
Duration of interval for quota in milliseconds. Minimal value is 1 second. The minimum value is 1000.
queries google.protobuf.Int64Value
The total number of queries. 0 - unlimited. The minimum value is 0.
errors google.protobuf.Int64Value
The number of queries that threw exception. 0 - unlimited. The minimum value is 0.
result_rows google.protobuf.Int64Value
The total number of rows given as the result.. 0 - unlimited. The minimum value is 0.
read_rows google.protobuf.Int64Value
The total number of source rows read from tables for running the query, on all remote servers. 0 - unlimited. The minimum value is 0.
execution_time google.protobuf.Int64Value
The total query execution time, in milliseconds (wall time). 0 - unlimited. The minimum value is 0.

Delete

Deletes the specified ClickHouse user.

rpc Delete (DeleteUserRequest) returns (operation.Operation)

Metadata and response of Operation:

    Operation.metadata:DeleteUserMetadata

    Operation.response:google.protobuf.Empty

DeleteUserRequest

Field Description
cluster_id string
Required. ID of the ClickHouse cluster the user belongs to. To get the cluster ID, use a ClusterService.List request. The maximum string length in characters is 50.
user_name string
Required. Name of the user to delete. To get the name of the user, use a UserService.List request. The maximum string length in characters is 63. Value must match the regular expression [a-zA-Z0-9_]*.

Operation

Field Description
id string
ID of the operation.
description string
Description of the operation. 0-256 characters long.
created_at google.protobuf.Timestamp
Creation timestamp.
created_by string
ID of the user or service account who initiated the operation.
modified_at google.protobuf.Timestamp
The time when the Operation resource was last modified.
done bool
If the value is false, it means the operation is still in progress. If true, the operation is completed, and either error or response is available.
metadata google.protobuf.Any<DeleteUserMetadata>
Service-specific metadata associated with the operation. It typically contains the ID of the target resource that the operation is performed on. Any method that returns a long-running operation should document the metadata type, if any.
result oneof: error or response
The operation result. If done == false and there was no failure detected, neither error nor response is set. If done == false and there was a failure detected, error is set. If done == true, exactly one of error or response is set.
  error google.rpc.Status
The error result of the operation in case of failure or cancellation.
  response google.protobuf.Any<google.protobuf.Empty>
if operation finished successfully.

DeleteUserMetadata

Field Description
cluster_id string
ID of the ClickHouse cluster the user belongs to.
user_name string
Name of the user that is being deleted.

GrantPermission

Grants a permission to the specified ClickHouse user.

rpc GrantPermission (GrantUserPermissionRequest) returns (operation.Operation)

Metadata and response of Operation:

    Operation.metadata:GrantUserPermissionMetadata

    Operation.response:User

GrantUserPermissionRequest

Field Description
cluster_id string
Required. ID of the ClickHouse cluster the user belongs to. To get the cluster ID, use a ClusterService.List request. The maximum string length in characters is 50.
user_name string
Required. Name of the user to grant the permission to. To get the name of the user, use a UserService.List request. The maximum string length in characters is 63. Value must match the regular expression [a-zA-Z0-9_]*.
permission Permission
Permission that should be granted to the specified user.

Permission

Field Description
database_name string
Name of the database that the permission grants access to.

Operation

Field Description
id string
ID of the operation.
description string
Description of the operation. 0-256 characters long.
created_at google.protobuf.Timestamp
Creation timestamp.
created_by string
ID of the user or service account who initiated the operation.
modified_at google.protobuf.Timestamp
The time when the Operation resource was last modified.
done bool
If the value is false, it means the operation is still in progress. If true, the operation is completed, and either error or response is available.
metadata google.protobuf.Any<GrantUserPermissionMetadata>
Service-specific metadata associated with the operation. It typically contains the ID of the target resource that the operation is performed on. Any method that returns a long-running operation should document the metadata type, if any.
result oneof: error or response
The operation result. If done == false and there was no failure detected, neither error nor response is set. If done == false and there was a failure detected, error is set. If done == true, exactly one of error or response is set.
  error google.rpc.Status
The error result of the operation in case of failure or cancellation.
  response google.protobuf.Any<User>
if operation finished successfully.

GrantUserPermissionMetadata

Field Description
cluster_id string
ID of the ClickHouse cluster the user belongs to.
user_name string
Name of the user that is being granted a permission.

User

Field Description
name string
Name of the ClickHouse user.
cluster_id string
ID of the ClickHouse cluster the user belongs to.
permissions[] Permission
Set of permissions granted to the user.
settings UserSettings
quotas[] UserQuota
Set of quotas assigned to the user.

Permission

Field Description
database_name string
Name of the database that the permission grants access to.

UserSettings

Field Description
readonly google.protobuf.Int64Value
Restricts permissions for non-DDL queries. To restrict permissions for DDL queries, use allow_ddl instead.
  • 0 (default)-no restrictions.
  • 1-only read data queries are allowed.
  • 2-read data and change settings queries are allowed.

See in-depth description in ClickHouse documentation. Acceptable values are 0 to 2, inclusive.
allow_ddl google.protobuf.BoolValue
Determines whether DDL queries are allowed (e.g., CREATE, ALTER, RENAME, etc).
Default value: true.
See in-depth description in ClickHouse documentation.
insert_quorum google.protobuf.Int64Value
Enables or disables write quorum for ClickHouse cluster. If the value is less than 2, then write quorum is disabled, otherwise it is enabled.
When used, write quorum guarantees that ClickHouse has written data to the quorum of insert_quorum replicas with no errors until the insert_quorum_timeout expires. All replicas in the quorum are in the consistent state, meaning that they contain linearized data from the previous INSERT queries. Employ write quorum, if you need the guarantees that the written data would not be lost in case of one or more replicas failure.
You can use select_sequential_consistency setting to read the data written with write quorum.
See in-depth description in ClickHouse documentation. The minimum value is 0.
connect_timeout google.protobuf.Int64Value
Connection timeout in milliseconds.
Value must be greater than 0 (default: 10000, 10 seconds). Value must be greater than 0.
receive_timeout google.protobuf.Int64Value
Receive timeout in milliseconds.
Value must be greater than 0 (default: 300000, 300 seconds or 5 minutes). Value must be greater than 0.
send_timeout google.protobuf.Int64Value
Send timeout in milliseconds.
Value must be greater than 0 (default: 300000, 300 seconds or 5 minutes). Value must be greater than 0.
insert_quorum_timeout google.protobuf.Int64Value
Quorum write timeout in milliseconds.
If the write quorum is enabled in the cluster, this timeout expires and some data is not written to the insert_quorum replicas, then ClickHouse will abort the execution of INSERT query and return an error. In this case, the client must send the query again to write the data block into the same or another replica.
Minimum value: 1000, 1 second (default: 60000, 1 minute). The minimum value is 1000.
select_sequential_consistency google.protobuf.BoolValue
Determines the behavior of SELECT queries from the replicated table: if enabled, ClickHouse will terminate a query with error message in case the replica does not have a chunk written with the quorum and will not read the parts that have not yet been written with the quorum.
Default value: false (sequential consistency is disabled).
max_replica_delay_for_distributed_queries google.protobuf.Int64Value
Max replica delay in milliseconds. If a replica lags more than the set value, this replica is not used and becomes a stale one.
Minimum value: 1000, 1 second (default: 300000, 300 seconds or 5 minutes).
See in-depth description in ClickHouse documentation. The minimum value is 1000.
fallback_to_stale_replicas_for_distributed_queries google.protobuf.BoolValue
Enables or disables query forcing to a stale replica in case the actual data is unavailable. If enabled, ClickHouse will choose the most up-to-date replica and force the query to use the data in this replica. This setting can be used when doing SELECT query from a distributed table that points to replicated tables.
Default value: true (query forcing is enabled).
See in-depth description in ClickHouse documentation.
replication_alter_partitions_sync google.protobuf.Int64Value
Wait mode for asynchronous actions in ALTER queries on replicated tables:
  • 0-do not wait for replicas.
  • 1-only wait for own execution (default).
  • 2-wait for all replicas.

See in-depth description in ClickHouse documentation. Acceptable values are 0 to 2, inclusive.
distributed_product_mode enum DistributedProductMode
Determine the behavior of distributed subqueries.
See in-depth description in ClickHouse documentation.
  • DISTRIBUTED_PRODUCT_MODE_DENY: Default value. Prohibits using these types of subqueries (returns the "Double-distributed in/JOIN subqueries is denied" exception).
  • DISTRIBUTED_PRODUCT_MODE_LOCAL: Replaces the database and table in the subquery with local ones for the destination server (shard), leaving the normal IN/JOIN.
  • DISTRIBUTED_PRODUCT_MODE_GLOBAL: Replaces the IN/JOIN query with GLOBAL IN/GLOBAL JOIN.
  • DISTRIBUTED_PRODUCT_MODE_ALLOW: Allows the use of these types of subqueries.
distributed_aggregation_memory_efficient google.protobuf.BoolValue
Enables of disables memory saving mode when doing distributed aggregation.
When ClickHouse works with a distributed query, external aggregation is done on remote servers. Enable this setting to achieve a smaller memory footprint on the server that sourced such a distributed query.
Default value: false (memory saving mode is disabled).
See in-depth description in ClickHouse documentation.
distributed_ddl_task_timeout google.protobuf.Int64Value
Timeout for DDL queries, in milliseconds.
skip_unavailable_shards google.protobuf.BoolValue
Enables or disables silent skipping of unavailable shards.
A shard is considered unavailable if all its replicas are also unavailable.
Default value: false (silent skipping is disabled).
See in-depth description in ClickHouse documentation.
compile google.protobuf.BoolValue
Enables or disables query compilation. If you execute a lot of structurally identical queries, then enable this setting. As a result, such queries may be executed faster due to use of queries' compiled parts.
Use this setting in combination with min_count_to_compile setting.
Default value: false (compilation is disabled).
See in-depth description in ClickHouse documentation.
min_count_to_compile google.protobuf.Int64Value
How many structurally identical queries ClickHouse has to encounter before they are compiled.
Minimum value: 0 (default: 3).
For the 0 value compilation is synchronous: a query waits for compilation process to complete prior to continuing execution. It is recommended to set this value only for testing purposes.
For all other values, compilation is asynchronous: the compilation process executes in a separate thread. When a compiled part of query is ready, it will be used by ClickHouse for eligible queries, including the ones that are currently running.
See in-depth description in ClickHouse documentation. The minimum value is 0.
compile_expressions google.protobuf.BoolValue
Enables or disables expression compilation. If you execute a lot of queries that contain identical expressions, then enable this setting. As a result, such queries may be executed faster due to use of compiled expressions.
Use this setting in combination with min_count_to_compile_expression setting.
Default value: false (expression compilation is disabled).
min_count_to_compile_expression google.protobuf.Int64Value
How many identical expressions ClickHouse has to encounter before they are compiled.
Minimum value: 0 (default: 3).
For the 0 value compilation is synchronous: a query waits for expression compilation process to complete prior to continuing execution. It is recommended to set this value only for testing purposes.
For all other values, compilation is asynchronous: the compilation process executes in a separate thread. When a compiled expression is ready, it will be used by ClickHouse for eligible queries, including the ones that are currently running. The minimum value is 0.
max_block_size google.protobuf.Int64Value
The maximum block size for reading.
Data in ClickHouse is organized and processed by blocks (block is a set of columns' parts). The internal processing cycles for a single block are efficient enough, but there are noticeable expenditures on each block.
This setting is a recommendation for size of block (in a count of rows) that should be loaded from tables.
Value must be greater than 0 (default: 65536).
See in-depth description in ClickHouse documentation. Value must be greater than 0.
min_insert_block_size_rows google.protobuf.Int64Value
Limits the minimum number of rows in a block to be inserted in a table by INSERT query. Blocks that are smaller than the specified value, will be squashed together into the bigger blocks.
Minimal value: 0, block squashing is disabled (default: 1048576). The minimum value is 0.
min_insert_block_size_bytes google.protobuf.Int64Value
Limits the minimum number of bytes in a block to be inserted in a table by INSERT query. Blocks that are smaller than the specified value, will be squashed together into the bigger blocks.
Minimal value: 0, block squashing is disabled (default: 268435456, 256 MB). The minimum value is 0.
max_insert_block_size google.protobuf.Int64Value
Allows to form blocks of the specified size (in bytes) when inserting data in a table. This setting has effect only if server is creating such blocks by itself.
Value must be greater than 0 (default: 1048576).
See in-depth description in ClickHouse documentation. Value must be greater than 0.
min_bytes_to_use_direct_io google.protobuf.Int64Value
Limits the minimum number of bytes to enable unbuffered direct reads from disk (Direct I/O).
By default, ClickHouse does not read data directly from disk, but relies on the filesystem and its cache instead. Such reading strategy is effective when the data volume is small. If the amount of the data to read is huge, it is more effective to read directly from the disk, bypassing the filesystem cache.
If the total amount of the data to read is greater than the value of this setting, then ClickHouse will fetch this data directly from the disk.
Minimal value and default value: 0, Direct I/O is disabled. The minimum value is 0.
use_uncompressed_cache google.protobuf.BoolValue
Determines whether to use the cache of uncompressed blocks, or not. Using this cache can significantly reduce latency and increase the throughput when a huge amount of small queries is to be processed. Enable this setting for the users who instantiates small queries frequently.
This setting has effect only for tables of the MergeTree family.
Default value: false (uncompressed cache is disabled).
See in-depth description in ClickHouse documentation.
merge_tree_max_rows_to_use_cache google.protobuf.Int64Value
Limits the maximum size in rows of the request that can use the cache of uncompressed data. The cache is not used for requests larger than the specified value.
Use this setting in combination with use_uncompressed_cache setting.
Value must be greater than 0 (default: 128x8192). Value must be greater than 0.
merge_tree_max_bytes_to_use_cache google.protobuf.Int64Value
Limits the maximum size in bytes of the request that can use the cache of uncompressed data. The cache is not used for requests larger than the specified value.
Use this setting in combination with use_uncompressed_cache setting.
Value must be greater than 0 (default: 192x10x1024x1024). Value must be greater than 0.
merge_tree_min_rows_for_concurrent_read google.protobuf.Int64Value
Limits the minimum number of rows to be read from a file to enable concurrent read. If the number of rows to be read exceeds this value, then ClickHouse will try to use a few threads to read from a file concurrently.
This setting has effect only for tables of the MergeTree family.
Value must be greater than 0 (default: 20x8192). Value must be greater than 0.
merge_tree_min_bytes_for_concurrent_read google.protobuf.Int64Value
Limits the number of bytes to be read from a file to enable concurrent read. If the number of bytes to be read exceeds this value, then ClickHouse will try to use a few threads to read from a file concurrently.
This setting has effect only for tables of the MergeTree family.
Value must be greater than 0 (default: 24x10x1024x1024). Value must be greater than 0.
max_bytes_before_external_group_by google.protobuf.Int64Value
Sets the threshold of RAM consumption (in bytes) after that the temporary data, collected during the GROUP BY operation, should be flushed to disk to limit the RAM comsumption.
By default, aggregation is done by employing hash table that resides in RAM. A query can result in aggregation of huge data volumes that can lead to memory exhaustion and abortion of the query (see the max_memory_usage setting). For such queries, you can use this setting to force ClickHouse to do flushing and complete aggregation successfully.
Minimal value and default value: 0, GROUP BY in the external memory is disabled.
When using aggregation in external memory, it is recommended to set the value of this setting twice as low as the max_memory_usage setting value (by default, the maximum memory usage is limited to ten gigabytes).
See in-depth description in ClickHouse documentation.
See also: the distributed_aggregation_memory_efficient setting.
max_bytes_before_external_sort google.protobuf.Int64Value
This setting is equivalent of the max_bytes_before_external_group_by setting, except for it is for sort operation (ORDER BY), not aggregation.
group_by_two_level_threshold google.protobuf.Int64Value
Sets the threshold of the number of keys, after that the two-level aggregation should be used.
Minimal value: 0, threshold is not set (default: 10000).
group_by_two_level_threshold_bytes google.protobuf.Int64Value
Sets the threshold of the number of bytes, after that the two-level aggregation should be used.
Minimal value: 0, threshold is not set (default: 100000000).
priority google.protobuf.Int64Value
Sets the priority of a query.
  • 0-priority is not used.
  • 1-the highest priority.
  • and so on. The higher the number, the lower a query's priority.

This setting should be set up for each query individually.
If ClickHouse is working with the high-priority queries, and a low-priority query enters, then the low-priority query is paused until higher-priority queries are completed.
Minimal value and default value: 0, priority is not used. The minimum value is 0.
max_threads google.protobuf.Int64Value
Limits the maximum number of threads to process the request (setting does not take threads that read data from remote servers into account).
This setting applies to threads that perform the same stages of the query processing pipeline in parallel.
Minimal value and default value: 0 (the thread number is calculated automatically based on the number of physical CPU cores, no HyperThreading cores are taken into account).
See in-depth description in ClickHouse documentation. Value must be greater than 0.
max_memory_usage google.protobuf.Int64Value
Limits the maximum memory usage (in bytes) for processing of a single user's query on a single server. This setting does not take server's free RAM amount or total RAM amount into account.
This limitation is enforced for any user's single query on a single server.
Minimal value: 0, no limitation is set. Value that is set in the ClickHouse default config file: 10737418240 (10 GB).
If you use max_bytes_before_external_group_by or max_bytes_before_external_sort setting, then it is recommended to set their values twice as low as max_memory_usage setting value.
See in-depth description in ClickHouse documentation. The minimum value is 0.
max_memory_usage_for_user google.protobuf.Int64Value
Limits the maximum memory usage (in bytes) for processing of user's queries on a single server. This setting does not take server's free RAM amount or total RAM amount into account.
This limitation is enforced for all queries that belong to one user and run simultaneously on a single server.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_network_bandwidth google.protobuf.Int64Value
The maximum speed of data exchange over the network in bytes per second for a query.
Minimal value and default value: 0, no limitation is set.
max_network_bandwidth_for_user google.protobuf.Int64Value
The maximum speed of data exchange over the network in bytes per second for all concurrently running user queries.
Minimal value and default value: 0, no limitation is set.
force_index_by_date google.protobuf.BoolValue
If enabled, query is not executed if the ClickHouse can't use index by date. This setting has effect only for tables of the MergeTree family.
Default value: false (setting is disabled, query executes even if ClickHouse can't use index by date).
See in-depth description in ClickHouse documentation.
force_primary_key google.protobuf.BoolValue
If enabled, query is not executed if the ClickHouse can't use index by primary key. This setting has effect only for tables of the MergeTree family.
Default value: false (setting is disabled, query executes even if ClickHouse can't use index by primary key).
See in-depth description in ClickHouse documentation.
max_rows_to_read google.protobuf.Int64Value
Limits the maximum number of rows that can be read from a table when running a query.
Minimal value and default value: 0, no limitation is set.
See in-depth description in ClickHouse documentation. The minimum value is 0.
max_bytes_to_read google.protobuf.Int64Value
Limits the maximum number of bytes (uncompressed data) that can be read from a table when running a query.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
read_overflow_mode enum OverflowMode
Determines the behavior on exceeding limits while reading the data.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
max_rows_to_group_by google.protobuf.Int64Value
Limits the maximum number of unique keys received from aggregation function. This setting helps to reduce RAM consumption while doing aggregation.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
group_by_overflow_mode enum GroupByOverflowMode
Determines the behavior on exceeding limits while doing aggregation.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
  • any-perform approximate GROUP BY operation by continuing aggregation for the keys that got into the set, but don't add new keys to the set.
max_rows_to_sort google.protobuf.Int64Value
Limits the maximum number of rows that can be read from a table for sorting. This setting helps to reduce RAM consumption.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_bytes_to_sort google.protobuf.Int64Value
Limits the maximum number of bytes (uncompressed data) that can be read from a table for sorting. This setting helps to reduce RAM consumption.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
sort_overflow_mode enum OverflowMode
Determines the behavior on exceeding limits while sorting.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
max_result_rows google.protobuf.Int64Value
Limits the number of rows in the result. This limitation is also checked for subqueries and parts of distributed queries that run on remote servers.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_result_bytes google.protobuf.Int64Value
Limits the number of bytes in the result. This limitation is also checked for subqueries and parts of distributed queries that run on remote servers.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
result_overflow_mode enum OverflowMode
Determines the behavior on exceeding limits while forming result.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
max_rows_in_distinct google.protobuf.Int64Value
Limits the maximum number of different rows when using DISTINCT.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_bytes_in_distinct google.protobuf.Int64Value
Limits the maximum size of a hash table in bytes (uncompressed data) when using DISTINCT. The minimum value is 0.
distinct_overflow_mode enum OverflowMode
Determines the behavior on exceeding limits while doing DISCTINCT.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
max_rows_to_transfer google.protobuf.Int64Value
Limits the maximum number of rows that can be passed to a remote server or saved in a temporary table when using GLOBAL IN.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_bytes_to_transfer google.protobuf.Int64Value
Limits the maximum number of bytes (uncompressed data) that can be passed to a remote server or saved in a temporary table when using GLOBAL IN.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
transfer_overflow_mode enum OverflowMode
Determines the behavior on exceeding limits while doing transfers.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
max_execution_time google.protobuf.Int64Value
Limits the maximum query execution time in milliseconds. At this moment, this limitation is not checked when passing one of the sorting stages, as well as merging and finalizing aggregation funictions.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
timeout_overflow_mode enum OverflowMode
Determines the behavior on exceeding limits of execution time.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
max_rows_in_set google.protobuf.Int64Value
Limit on the number of rows in the set resulting from the execution of the IN section. The minimum value is 0.
max_bytes_in_set google.protobuf.Int64Value
Limit on the number of bytes in the set resulting from the execution of the IN section. The minimum value is 0.
set_overflow_mode enum OverflowMode
Determine the behavior on exceeding max_rows_in_set or max_bytes_in_set limit. Possible values: OVERFLOW_MODE_THROW, OVERFLOW_MODE_BREAK.
max_rows_in_join google.protobuf.Int64Value
Limit on maximum size of the hash table for JOIN, in rows. The minimum value is 0.
max_bytes_in_join google.protobuf.Int64Value
Limit on maximum size of the hash table for JOIN, in bytes. The minimum value is 0.
join_overflow_mode enum OverflowMode
Determine the behavior on exceeding max_rows_in_join or max_bytes_in_join limit. Possible values: OVERFLOW_MODE_THROW, OVERFLOW_MODE_BREAK.
max_columns_to_read google.protobuf.Int64Value
Limits the maximum number of columns that can be read from a table in a single query. If the query requires to read more columns to complete, then it will be aborted.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_temporary_columns google.protobuf.Int64Value
Limits the maximum number of temporary columns that must be kept in RAM at the same time when running a query, including constant columns.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_temporary_non_const_columns google.protobuf.Int64Value
Limits the maximum number of temporary columns that must be kept in RAM at the same time when running a query, excluding constant columns.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_query_size google.protobuf.Int64Value
Limits the size of the part of a query that can be transferred to RAM for parsing with the SQL parser, in bytes.
Value must be greater than 0 (default: 262144).
See in-depth description in ClickHouse documentation. Value must be greater than 0.
max_ast_depth google.protobuf.Int64Value
Limits the maximum depth of query syntax tree.
Executing a big and complex query may result in building a syntax tree of enormous depth. By using this setting, you can prohibit execution of over-sized or non-optimized queries for huge tables.
For example, the **SELECT *** query may result in more complex and deeper syntax tree, compared to the SELECT ... WHERE ... query, containing constraints and conditions, in the most cases. A user can be forced to construct more optimized queries, if this setting is used.
Value must be greater than 0 (default: 1000). If a too small value is set, it may render ClickHouse unable to execute even simple queries.
See in-depth description in ClickHouse documentation. Value must be greater than 0.
max_ast_elements google.protobuf.Int64Value
Limits the maximum size of query syntax tree in number of nodes.
Executing a big and complex query may result in building a syntax tree of enormous size. By using this setting, you can prohibit execution of over-sized or non-optimized queries for huge tables.
Value must be greater than 0 (default: 50000). If a too small value is set, it may render ClickHouse unable to execute even simple queries.
See in-depth description in ClickHouse documentation. Value must be greater than 0.
max_expanded_ast_elements google.protobuf.Int64Value
Limits the maximum size of query syntax tree in number of nodes after expansion of aliases and the asterisk values.
Executing a big and complex query may result in building a syntax tree of enormous size. By using this setting, you can prohibit execution of over-sized or non-optimized queries for huge tables.
Value must be greater than 0 (default: 500000). If a too small value is set, it may render ClickHouse unable to execute even simple queries. Value must be greater than 0.
min_execution_speed google.protobuf.Int64Value
Minimal execution speed in rows per second. The minimum value is 0.
min_execution_speed_bytes google.protobuf.Int64Value
Minimal execution speed in bytes per second. The minimum value is 0.
count_distinct_implementation enum CountDistinctImplementation
Aggregate function to use for implementation of count(DISTINCT ...).
input_format_values_interpret_expressions google.protobuf.BoolValue
Enables or disables SQL parser if the fast stream parser cannot parse the data.
Enable this setting, if the data that you want to insert into a table contains SQL expressions.
For example, the stream parser is unable to parse a value that contains now() expression; therefore an INSERT query for this value will fail and no data will be inserted into a table. With enabled SQL parser, this expression is parsed correctly: the now() expression will be parsed as SQL function, interpreted, and the current date and time will be inserted into the table as a result.
This setting has effect only if you use Values format when inserting data.
Default value: true (SQL parser is enabled).
See in-depth description in ClickHouse documentation.
input_format_defaults_for_omitted_fields google.protobuf.BoolValue
Enables or disables replacing omitted input values with default values of the respective columns when performing INSERT queries.
Default value: true (replacing is enabled).
output_format_json_quote_64bit_integers google.protobuf.BoolValue
Enables quoting of 64-bit integers in JSON output format.
If this setting is enabled, then 64-bit integers (UInt64 and Int64) will be quoted when written to JSON output in order to maintain compatibility with the most of the JavaScript engines. Otherwise, such integers will not be quoted.
Default value: false (quoting 64-bit integers is disabled).
output_format_json_quote_denormals google.protobuf.BoolValue
Enables special floating-point values (+nan, -nan, +inf and -inf) in JSON output format.
Default value: false (special values do not present in output).
low_cardinality_allow_in_native_format google.protobuf.BoolValue
Determines whether to use LowCardinality type in Native format.
  • true (default)-yes, use.
  • false-convert LowCardinality columns to regular columns when doing SELECT, and convert regular columns to LowCardinality when doing INSERT.

LowCardinality columns (aka sparse columns) store data in more effective way, compared to regular columns, by using hash tables. If data to insert suits this storage format, ClickHouse will place them into LowCardinality column.
If you use a third-party ClickHouse client that can't work with LowCardinality columns, then this client will not be able to correctly interpret the result of the query that asks for data stored in LowCardinality column. Disable this setting to convert LowCardinality column to regular column when creating the result, so such clients will be able to process the result.
Official ClickHouse client works with LowCardinality columns out-of-the-box.
Default value: true (LowCardinality columns are used in Native format).
empty_result_for_aggregation_by_empty_set google.protobuf.BoolValue
Enables returning of empty result when aggregating without keys (with GROUP BY operation absent) on empty set (e.g., SELECT count(*) FROM table WHERE 0).
  • true-ClickHouse will return an empty result for such queries.
  • false (default)-ClickHouse will return a single-line result consisting of NULL values for aggregation functions, in accordance with SQL standard.
joined_subquery_requires_alias google.protobuf.BoolValue
join_use_nulls google.protobuf.BoolValue
transform_null_in google.protobuf.BoolValue
http_connection_timeout google.protobuf.Int64Value
HTTP connection timeout, in milliseconds.
Value must be greater than 0 (default: 1000, 1 second).
http_receive_timeout google.protobuf.Int64Value
HTTP receive timeout, in milliseconds.
Value must be greater than 0 (default: 1800000, 1800 seconds, 30 minutes).
http_send_timeout google.protobuf.Int64Value
HTTP send timeout, in milliseconds.
Value must be greater than 0 (default: 1800000, 1800 seconds, 30 minutes).
enable_http_compression google.protobuf.BoolValue
Enables or disables data compression in HTTP responses.
By default, ClickHouse stores data compressed. When executing a query, its result is uncompressed. Use this setting to command ClickHouse to compress the result when sending it via HTTP.
Enable this setting and add the Accept-Encoding: HTTP header in a HTTP request to force compression of HTTP response from ClickHouse.
ClickHouse support the following compression methods: gzip, br and deflate.
Default value: false (compression is disabled).
See in-depth description in ClickHouse documentation.
send_progress_in_http_headers google.protobuf.BoolValue
Enables progress notifications using X-ClickHouse-Progress HTTP header.
Default value: false (notifications disabled).
http_headers_progress_interval google.protobuf.Int64Value
Minimum interval between progress notifications with X-ClickHouse-Progress HTTP header, in milliseconds.
Value must be greater than 0 (default: 100).
add_http_cors_header google.protobuf.BoolValue
Adds CORS header in HTTP responses.
Default value: false (header is not added).
quota_mode enum QuotaMode
Quota accounting mode. Possible values: QUOTA_MODE_DEFAULT, QUOTA_MODE_KEYED and QUOTA_MODE_KEYED_BY_IP.

UserQuota

Field Description
interval_duration google.protobuf.Int64Value
Duration of interval for quota in milliseconds. Minimal value is 1 second. The minimum value is 1000.
queries google.protobuf.Int64Value
The total number of queries. 0 - unlimited. The minimum value is 0.
errors google.protobuf.Int64Value
The number of queries that threw exception. 0 - unlimited. The minimum value is 0.
result_rows google.protobuf.Int64Value
The total number of rows given as the result.. 0 - unlimited. The minimum value is 0.
read_rows google.protobuf.Int64Value
The total number of source rows read from tables for running the query, on all remote servers. 0 - unlimited. The minimum value is 0.
execution_time google.protobuf.Int64Value
The total query execution time, in milliseconds (wall time). 0 - unlimited. The minimum value is 0.

RevokePermission

Revokes a permission from the specified ClickHouse user.

rpc RevokePermission (RevokeUserPermissionRequest) returns (operation.Operation)

Metadata and response of Operation:

    Operation.metadata:RevokeUserPermissionMetadata

    Operation.response:User

RevokeUserPermissionRequest

Field Description
cluster_id string
Required. ID of the ClickHouse cluster the user belongs to. To get the cluster ID, use a ClusterService.List request. The maximum string length in characters is 50.
user_name string
Required. Name of the user to revoke a permission from. To get the name of the user, use a UserService.List request. The maximum string length in characters is 63. Value must match the regular expression [a-zA-Z0-9_]*.
database_name string
Name of the database that the user should lose access to. The maximum string length in characters is 63. Value must match the regular expression [a-zA-Z0-9_-]*.

Operation

Field Description
id string
ID of the operation.
description string
Description of the operation. 0-256 characters long.
created_at google.protobuf.Timestamp
Creation timestamp.
created_by string
ID of the user or service account who initiated the operation.
modified_at google.protobuf.Timestamp
The time when the Operation resource was last modified.
done bool
If the value is false, it means the operation is still in progress. If true, the operation is completed, and either error or response is available.
metadata google.protobuf.Any<RevokeUserPermissionMetadata>
Service-specific metadata associated with the operation. It typically contains the ID of the target resource that the operation is performed on. Any method that returns a long-running operation should document the metadata type, if any.
result oneof: error or response
The operation result. If done == false and there was no failure detected, neither error nor response is set. If done == false and there was a failure detected, error is set. If done == true, exactly one of error or response is set.
  error google.rpc.Status
The error result of the operation in case of failure or cancellation.
  response google.protobuf.Any<User>
if operation finished successfully.

RevokeUserPermissionMetadata

Field Description
cluster_id string
ID of the ClickHouse cluster the user belongs to.
user_name string
Name of the user whose permission is being revoked.

User

Field Description
name string
Name of the ClickHouse user.
cluster_id string
ID of the ClickHouse cluster the user belongs to.
permissions[] Permission
Set of permissions granted to the user.
settings UserSettings
quotas[] UserQuota
Set of quotas assigned to the user.

Permission

Field Description
database_name string
Name of the database that the permission grants access to.

UserSettings

Field Description
readonly google.protobuf.Int64Value
Restricts permissions for non-DDL queries. To restrict permissions for DDL queries, use allow_ddl instead.
  • 0 (default)-no restrictions.
  • 1-only read data queries are allowed.
  • 2-read data and change settings queries are allowed.

See in-depth description in ClickHouse documentation. Acceptable values are 0 to 2, inclusive.
allow_ddl google.protobuf.BoolValue
Determines whether DDL queries are allowed (e.g., CREATE, ALTER, RENAME, etc).
Default value: true.
See in-depth description in ClickHouse documentation.
insert_quorum google.protobuf.Int64Value
Enables or disables write quorum for ClickHouse cluster. If the value is less than 2, then write quorum is disabled, otherwise it is enabled.
When used, write quorum guarantees that ClickHouse has written data to the quorum of insert_quorum replicas with no errors until the insert_quorum_timeout expires. All replicas in the quorum are in the consistent state, meaning that they contain linearized data from the previous INSERT queries. Employ write quorum, if you need the guarantees that the written data would not be lost in case of one or more replicas failure.
You can use select_sequential_consistency setting to read the data written with write quorum.
See in-depth description in ClickHouse documentation. The minimum value is 0.
connect_timeout google.protobuf.Int64Value
Connection timeout in milliseconds.
Value must be greater than 0 (default: 10000, 10 seconds). Value must be greater than 0.
receive_timeout google.protobuf.Int64Value
Receive timeout in milliseconds.
Value must be greater than 0 (default: 300000, 300 seconds or 5 minutes). Value must be greater than 0.
send_timeout google.protobuf.Int64Value
Send timeout in milliseconds.
Value must be greater than 0 (default: 300000, 300 seconds or 5 minutes). Value must be greater than 0.
insert_quorum_timeout google.protobuf.Int64Value
Quorum write timeout in milliseconds.
If the write quorum is enabled in the cluster, this timeout expires and some data is not written to the insert_quorum replicas, then ClickHouse will abort the execution of INSERT query and return an error. In this case, the client must send the query again to write the data block into the same or another replica.
Minimum value: 1000, 1 second (default: 60000, 1 minute). The minimum value is 1000.
select_sequential_consistency google.protobuf.BoolValue
Determines the behavior of SELECT queries from the replicated table: if enabled, ClickHouse will terminate a query with error message in case the replica does not have a chunk written with the quorum and will not read the parts that have not yet been written with the quorum.
Default value: false (sequential consistency is disabled).
max_replica_delay_for_distributed_queries google.protobuf.Int64Value
Max replica delay in milliseconds. If a replica lags more than the set value, this replica is not used and becomes a stale one.
Minimum value: 1000, 1 second (default: 300000, 300 seconds or 5 minutes).
See in-depth description in ClickHouse documentation. The minimum value is 1000.
fallback_to_stale_replicas_for_distributed_queries google.protobuf.BoolValue
Enables or disables query forcing to a stale replica in case the actual data is unavailable. If enabled, ClickHouse will choose the most up-to-date replica and force the query to use the data in this replica. This setting can be used when doing SELECT query from a distributed table that points to replicated tables.
Default value: true (query forcing is enabled).
See in-depth description in ClickHouse documentation.
replication_alter_partitions_sync google.protobuf.Int64Value
Wait mode for asynchronous actions in ALTER queries on replicated tables:
  • 0-do not wait for replicas.
  • 1-only wait for own execution (default).
  • 2-wait for all replicas.

See in-depth description in ClickHouse documentation. Acceptable values are 0 to 2, inclusive.
distributed_product_mode enum DistributedProductMode
Determine the behavior of distributed subqueries.
See in-depth description in ClickHouse documentation.
  • DISTRIBUTED_PRODUCT_MODE_DENY: Default value. Prohibits using these types of subqueries (returns the "Double-distributed in/JOIN subqueries is denied" exception).
  • DISTRIBUTED_PRODUCT_MODE_LOCAL: Replaces the database and table in the subquery with local ones for the destination server (shard), leaving the normal IN/JOIN.
  • DISTRIBUTED_PRODUCT_MODE_GLOBAL: Replaces the IN/JOIN query with GLOBAL IN/GLOBAL JOIN.
  • DISTRIBUTED_PRODUCT_MODE_ALLOW: Allows the use of these types of subqueries.
distributed_aggregation_memory_efficient google.protobuf.BoolValue
Enables of disables memory saving mode when doing distributed aggregation.
When ClickHouse works with a distributed query, external aggregation is done on remote servers. Enable this setting to achieve a smaller memory footprint on the server that sourced such a distributed query.
Default value: false (memory saving mode is disabled).
See in-depth description in ClickHouse documentation.
distributed_ddl_task_timeout google.protobuf.Int64Value
Timeout for DDL queries, in milliseconds.
skip_unavailable_shards google.protobuf.BoolValue
Enables or disables silent skipping of unavailable shards.
A shard is considered unavailable if all its replicas are also unavailable.
Default value: false (silent skipping is disabled).
See in-depth description in ClickHouse documentation.
compile google.protobuf.BoolValue
Enables or disables query compilation. If you execute a lot of structurally identical queries, then enable this setting. As a result, such queries may be executed faster due to use of queries' compiled parts.
Use this setting in combination with min_count_to_compile setting.
Default value: false (compilation is disabled).
See in-depth description in ClickHouse documentation.
min_count_to_compile google.protobuf.Int64Value
How many structurally identical queries ClickHouse has to encounter before they are compiled.
Minimum value: 0 (default: 3).
For the 0 value compilation is synchronous: a query waits for compilation process to complete prior to continuing execution. It is recommended to set this value only for testing purposes.
For all other values, compilation is asynchronous: the compilation process executes in a separate thread. When a compiled part of query is ready, it will be used by ClickHouse for eligible queries, including the ones that are currently running.
See in-depth description in ClickHouse documentation. The minimum value is 0.
compile_expressions google.protobuf.BoolValue
Enables or disables expression compilation. If you execute a lot of queries that contain identical expressions, then enable this setting. As a result, such queries may be executed faster due to use of compiled expressions.
Use this setting in combination with min_count_to_compile_expression setting.
Default value: false (expression compilation is disabled).
min_count_to_compile_expression google.protobuf.Int64Value
How many identical expressions ClickHouse has to encounter before they are compiled.
Minimum value: 0 (default: 3).
For the 0 value compilation is synchronous: a query waits for expression compilation process to complete prior to continuing execution. It is recommended to set this value only for testing purposes.
For all other values, compilation is asynchronous: the compilation process executes in a separate thread. When a compiled expression is ready, it will be used by ClickHouse for eligible queries, including the ones that are currently running. The minimum value is 0.
max_block_size google.protobuf.Int64Value
The maximum block size for reading.
Data in ClickHouse is organized and processed by blocks (block is a set of columns' parts). The internal processing cycles for a single block are efficient enough, but there are noticeable expenditures on each block.
This setting is a recommendation for size of block (in a count of rows) that should be loaded from tables.
Value must be greater than 0 (default: 65536).
See in-depth description in ClickHouse documentation. Value must be greater than 0.
min_insert_block_size_rows google.protobuf.Int64Value
Limits the minimum number of rows in a block to be inserted in a table by INSERT query. Blocks that are smaller than the specified value, will be squashed together into the bigger blocks.
Minimal value: 0, block squashing is disabled (default: 1048576). The minimum value is 0.
min_insert_block_size_bytes google.protobuf.Int64Value
Limits the minimum number of bytes in a block to be inserted in a table by INSERT query. Blocks that are smaller than the specified value, will be squashed together into the bigger blocks.
Minimal value: 0, block squashing is disabled (default: 268435456, 256 MB). The minimum value is 0.
max_insert_block_size google.protobuf.Int64Value
Allows to form blocks of the specified size (in bytes) when inserting data in a table. This setting has effect only if server is creating such blocks by itself.
Value must be greater than 0 (default: 1048576).
See in-depth description in ClickHouse documentation. Value must be greater than 0.
min_bytes_to_use_direct_io google.protobuf.Int64Value
Limits the minimum number of bytes to enable unbuffered direct reads from disk (Direct I/O).
By default, ClickHouse does not read data directly from disk, but relies on the filesystem and its cache instead. Such reading strategy is effective when the data volume is small. If the amount of the data to read is huge, it is more effective to read directly from the disk, bypassing the filesystem cache.
If the total amount of the data to read is greater than the value of this setting, then ClickHouse will fetch this data directly from the disk.
Minimal value and default value: 0, Direct I/O is disabled. The minimum value is 0.
use_uncompressed_cache google.protobuf.BoolValue
Determines whether to use the cache of uncompressed blocks, or not. Using this cache can significantly reduce latency and increase the throughput when a huge amount of small queries is to be processed. Enable this setting for the users who instantiates small queries frequently.
This setting has effect only for tables of the MergeTree family.
Default value: false (uncompressed cache is disabled).
See in-depth description in ClickHouse documentation.
merge_tree_max_rows_to_use_cache google.protobuf.Int64Value
Limits the maximum size in rows of the request that can use the cache of uncompressed data. The cache is not used for requests larger than the specified value.
Use this setting in combination with use_uncompressed_cache setting.
Value must be greater than 0 (default: 128x8192). Value must be greater than 0.
merge_tree_max_bytes_to_use_cache google.protobuf.Int64Value
Limits the maximum size in bytes of the request that can use the cache of uncompressed data. The cache is not used for requests larger than the specified value.
Use this setting in combination with use_uncompressed_cache setting.
Value must be greater than 0 (default: 192x10x1024x1024). Value must be greater than 0.
merge_tree_min_rows_for_concurrent_read google.protobuf.Int64Value
Limits the minimum number of rows to be read from a file to enable concurrent read. If the number of rows to be read exceeds this value, then ClickHouse will try to use a few threads to read from a file concurrently.
This setting has effect only for tables of the MergeTree family.
Value must be greater than 0 (default: 20x8192). Value must be greater than 0.
merge_tree_min_bytes_for_concurrent_read google.protobuf.Int64Value
Limits the number of bytes to be read from a file to enable concurrent read. If the number of bytes to be read exceeds this value, then ClickHouse will try to use a few threads to read from a file concurrently.
This setting has effect only for tables of the MergeTree family.
Value must be greater than 0 (default: 24x10x1024x1024). Value must be greater than 0.
max_bytes_before_external_group_by google.protobuf.Int64Value
Sets the threshold of RAM consumption (in bytes) after that the temporary data, collected during the GROUP BY operation, should be flushed to disk to limit the RAM comsumption.
By default, aggregation is done by employing hash table that resides in RAM. A query can result in aggregation of huge data volumes that can lead to memory exhaustion and abortion of the query (see the max_memory_usage setting). For such queries, you can use this setting to force ClickHouse to do flushing and complete aggregation successfully.
Minimal value and default value: 0, GROUP BY in the external memory is disabled.
When using aggregation in external memory, it is recommended to set the value of this setting twice as low as the max_memory_usage setting value (by default, the maximum memory usage is limited to ten gigabytes).
See in-depth description in ClickHouse documentation.
See also: the distributed_aggregation_memory_efficient setting.
max_bytes_before_external_sort google.protobuf.Int64Value
This setting is equivalent of the max_bytes_before_external_group_by setting, except for it is for sort operation (ORDER BY), not aggregation.
group_by_two_level_threshold google.protobuf.Int64Value
Sets the threshold of the number of keys, after that the two-level aggregation should be used.
Minimal value: 0, threshold is not set (default: 10000).
group_by_two_level_threshold_bytes google.protobuf.Int64Value
Sets the threshold of the number of bytes, after that the two-level aggregation should be used.
Minimal value: 0, threshold is not set (default: 100000000).
priority google.protobuf.Int64Value
Sets the priority of a query.
  • 0-priority is not used.
  • 1-the highest priority.
  • and so on. The higher the number, the lower a query's priority.

This setting should be set up for each query individually.
If ClickHouse is working with the high-priority queries, and a low-priority query enters, then the low-priority query is paused until higher-priority queries are completed.
Minimal value and default value: 0, priority is not used. The minimum value is 0.
max_threads google.protobuf.Int64Value
Limits the maximum number of threads to process the request (setting does not take threads that read data from remote servers into account).
This setting applies to threads that perform the same stages of the query processing pipeline in parallel.
Minimal value and default value: 0 (the thread number is calculated automatically based on the number of physical CPU cores, no HyperThreading cores are taken into account).
See in-depth description in ClickHouse documentation. Value must be greater than 0.
max_memory_usage google.protobuf.Int64Value
Limits the maximum memory usage (in bytes) for processing of a single user's query on a single server. This setting does not take server's free RAM amount or total RAM amount into account.
This limitation is enforced for any user's single query on a single server.
Minimal value: 0, no limitation is set. Value that is set in the ClickHouse default config file: 10737418240 (10 GB).
If you use max_bytes_before_external_group_by or max_bytes_before_external_sort setting, then it is recommended to set their values twice as low as max_memory_usage setting value.
See in-depth description in ClickHouse documentation. The minimum value is 0.
max_memory_usage_for_user google.protobuf.Int64Value
Limits the maximum memory usage (in bytes) for processing of user's queries on a single server. This setting does not take server's free RAM amount or total RAM amount into account.
This limitation is enforced for all queries that belong to one user and run simultaneously on a single server.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_network_bandwidth google.protobuf.Int64Value
The maximum speed of data exchange over the network in bytes per second for a query.
Minimal value and default value: 0, no limitation is set.
max_network_bandwidth_for_user google.protobuf.Int64Value
The maximum speed of data exchange over the network in bytes per second for all concurrently running user queries.
Minimal value and default value: 0, no limitation is set.
force_index_by_date google.protobuf.BoolValue
If enabled, query is not executed if the ClickHouse can't use index by date. This setting has effect only for tables of the MergeTree family.
Default value: false (setting is disabled, query executes even if ClickHouse can't use index by date).
See in-depth description in ClickHouse documentation.
force_primary_key google.protobuf.BoolValue
If enabled, query is not executed if the ClickHouse can't use index by primary key. This setting has effect only for tables of the MergeTree family.
Default value: false (setting is disabled, query executes even if ClickHouse can't use index by primary key).
See in-depth description in ClickHouse documentation.
max_rows_to_read google.protobuf.Int64Value
Limits the maximum number of rows that can be read from a table when running a query.
Minimal value and default value: 0, no limitation is set.
See in-depth description in ClickHouse documentation. The minimum value is 0.
max_bytes_to_read google.protobuf.Int64Value
Limits the maximum number of bytes (uncompressed data) that can be read from a table when running a query.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
read_overflow_mode enum OverflowMode
Determines the behavior on exceeding limits while reading the data.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
max_rows_to_group_by google.protobuf.Int64Value
Limits the maximum number of unique keys received from aggregation function. This setting helps to reduce RAM consumption while doing aggregation.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
group_by_overflow_mode enum GroupByOverflowMode
Determines the behavior on exceeding limits while doing aggregation.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
  • any-perform approximate GROUP BY operation by continuing aggregation for the keys that got into the set, but don't add new keys to the set.
max_rows_to_sort google.protobuf.Int64Value
Limits the maximum number of rows that can be read from a table for sorting. This setting helps to reduce RAM consumption.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_bytes_to_sort google.protobuf.Int64Value
Limits the maximum number of bytes (uncompressed data) that can be read from a table for sorting. This setting helps to reduce RAM consumption.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
sort_overflow_mode enum OverflowMode
Determines the behavior on exceeding limits while sorting.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
max_result_rows google.protobuf.Int64Value
Limits the number of rows in the result. This limitation is also checked for subqueries and parts of distributed queries that run on remote servers.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_result_bytes google.protobuf.Int64Value
Limits the number of bytes in the result. This limitation is also checked for subqueries and parts of distributed queries that run on remote servers.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
result_overflow_mode enum OverflowMode
Determines the behavior on exceeding limits while forming result.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
max_rows_in_distinct google.protobuf.Int64Value
Limits the maximum number of different rows when using DISTINCT.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_bytes_in_distinct google.protobuf.Int64Value
Limits the maximum size of a hash table in bytes (uncompressed data) when using DISTINCT. The minimum value is 0.
distinct_overflow_mode enum OverflowMode
Determines the behavior on exceeding limits while doing DISCTINCT.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
max_rows_to_transfer google.protobuf.Int64Value
Limits the maximum number of rows that can be passed to a remote server or saved in a temporary table when using GLOBAL IN.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_bytes_to_transfer google.protobuf.Int64Value
Limits the maximum number of bytes (uncompressed data) that can be passed to a remote server or saved in a temporary table when using GLOBAL IN.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
transfer_overflow_mode enum OverflowMode
Determines the behavior on exceeding limits while doing transfers.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
max_execution_time google.protobuf.Int64Value
Limits the maximum query execution time in milliseconds. At this moment, this limitation is not checked when passing one of the sorting stages, as well as merging and finalizing aggregation funictions.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
timeout_overflow_mode enum OverflowMode
Determines the behavior on exceeding limits of execution time.
  • throw-abort query execution, return an error.
  • break-stop query execution, return partial result.
max_rows_in_set google.protobuf.Int64Value
Limit on the number of rows in the set resulting from the execution of the IN section. The minimum value is 0.
max_bytes_in_set google.protobuf.Int64Value
Limit on the number of bytes in the set resulting from the execution of the IN section. The minimum value is 0.
set_overflow_mode enum OverflowMode
Determine the behavior on exceeding max_rows_in_set or max_bytes_in_set limit. Possible values: OVERFLOW_MODE_THROW, OVERFLOW_MODE_BREAK.
max_rows_in_join google.protobuf.Int64Value
Limit on maximum size of the hash table for JOIN, in rows. The minimum value is 0.
max_bytes_in_join google.protobuf.Int64Value
Limit on maximum size of the hash table for JOIN, in bytes. The minimum value is 0.
join_overflow_mode enum OverflowMode
Determine the behavior on exceeding max_rows_in_join or max_bytes_in_join limit. Possible values: OVERFLOW_MODE_THROW, OVERFLOW_MODE_BREAK.
max_columns_to_read google.protobuf.Int64Value
Limits the maximum number of columns that can be read from a table in a single query. If the query requires to read more columns to complete, then it will be aborted.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_temporary_columns google.protobuf.Int64Value
Limits the maximum number of temporary columns that must be kept in RAM at the same time when running a query, including constant columns.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_temporary_non_const_columns google.protobuf.Int64Value
Limits the maximum number of temporary columns that must be kept in RAM at the same time when running a query, excluding constant columns.
Minimal value and default value: 0, no limitation is set. The minimum value is 0.
max_query_size google.protobuf.Int64Value
Limits the size of the part of a query that can be transferred to RAM for parsing with the SQL parser, in bytes.
Value must be greater than 0 (default: 262144).
See in-depth description in ClickHouse documentation. Value must be greater than 0.
max_ast_depth google.protobuf.Int64Value
Limits the maximum depth of query syntax tree.
Executing a big and complex query may result in building a syntax tree of enormous depth. By using this setting, you can prohibit execution of over-sized or non-optimized queries for huge tables.
For example, the **SELECT *** query may result in more complex and deeper syntax tree, compared to the SELECT ... WHERE ... query, containing constraints and conditions, in the most cases. A user can be forced to construct more optimized queries, if this setting is used.
Value must be greater than 0 (default: 1000). If a too small value is set, it may render ClickHouse unable to execute even simple queries.
See in-depth description in ClickHouse documentation. Value must be greater than 0.
max_ast_elements google.protobuf.Int64Value
Limits the maximum size of query syntax tree in number of nodes.
Executing a big and complex query may result in building a syntax tree of enormous size. By using this setting, you can prohibit execution of over-sized or non-optimized queries for huge tables.
Value must be greater than 0 (default: 50000). If a too small value is set, it may render ClickHouse unable to execute even simple queries.
See in-depth description in ClickHouse documentation. Value must be greater than 0.
max_expanded_ast_elements google.protobuf.Int64Value
Limits the maximum size of query syntax tree in number of nodes after expansion of aliases and the asterisk values.
Executing a big and complex query may result in building a syntax tree of enormous size. By using this setting, you can prohibit execution of over-sized or non-optimized queries for huge tables.
Value must be greater than 0 (default: 500000). If a too small value is set, it may render ClickHouse unable to execute even simple queries. Value must be greater than 0.
min_execution_speed google.protobuf.Int64Value
Minimal execution speed in rows per second. The minimum value is 0.
min_execution_speed_bytes google.protobuf.Int64Value
Minimal execution speed in bytes per second. The minimum value is 0.
count_distinct_implementation enum CountDistinctImplementation
Aggregate function to use for implementation of count(DISTINCT ...).
input_format_values_interpret_expressions google.protobuf.BoolValue
Enables or disables SQL parser if the fast stream parser cannot parse the data.
Enable this setting, if the data that you want to insert into a table contains SQL expressions.
For example, the stream parser is unable to parse a value that contains now() expression; therefore an INSERT query for this value will fail and no data will be inserted into a table. With enabled SQL parser, this expression is parsed correctly: the now() expression will be parsed as SQL function, interpreted, and the current date and time will be inserted into the table as a result.
This setting has effect only if you use Values format when inserting data.
Default value: true (SQL parser is enabled).
See in-depth description in ClickHouse documentation.
input_format_defaults_for_omitted_fields google.protobuf.BoolValue
Enables or disables replacing omitted input values with default values of the respective columns when performing INSERT queries.
Default value: true (replacing is enabled).
output_format_json_quote_64bit_integers google.protobuf.BoolValue
Enables quoting of 64-bit integers in JSON output format.
If this setting is enabled, then 64-bit integers (UInt64 and Int64) will be quoted when written to JSON output in order to maintain compatibility with the most of the JavaScript engines. Otherwise, such integers will not be quoted.
Default value: false (quoting 64-bit integers is disabled).
output_format_json_quote_denormals google.protobuf.BoolValue
Enables special floating-point values (+nan, -nan, +inf and -inf) in JSON output format.
Default value: false (special values do not present in output).
low_cardinality_allow_in_native_format google.protobuf.BoolValue
Determines whether to use LowCardinality type in Native format.
  • true (default)-yes, use.
  • false-convert LowCardinality columns to regular columns when doing SELECT, and convert regular columns to LowCardinality when doing INSERT.

LowCardinality columns (aka sparse columns) store data in more effective way, compared to regular columns, by using hash tables. If data to insert suits this storage format, ClickHouse will place them into LowCardinality column.
If you use a third-party ClickHouse client that can't work with LowCardinality columns, then this client will not be able to correctly interpret the result of the query that asks for data stored in LowCardinality column. Disable this setting to convert LowCardinality column to regular column when creating the result, so such clients will be able to process the result.
Official ClickHouse client works with LowCardinality columns out-of-the-box.
Default value: true (LowCardinality columns are used in Native format).
empty_result_for_aggregation_by_empty_set google.protobuf.BoolValue
Enables returning of empty result when aggregating without keys (with GROUP BY operation absent) on empty set (e.g., SELECT count(*) FROM table WHERE 0).
  • true-ClickHouse will return an empty result for such queries.
  • false (default)-ClickHouse will return a single-line result consisting of NULL values for aggregation functions, in accordance with SQL standard.
joined_subquery_requires_alias google.protobuf.BoolValue
join_use_nulls google.protobuf.BoolValue
transform_null_in google.protobuf.BoolValue
http_connection_timeout google.protobuf.Int64Value
HTTP connection timeout, in milliseconds.
Value must be greater than 0 (default: 1000, 1 second).
http_receive_timeout google.protobuf.Int64Value
HTTP receive timeout, in milliseconds.
Value must be greater than 0 (default: 1800000, 1800 seconds, 30 minutes).
http_send_timeout google.protobuf.Int64Value
HTTP send timeout, in milliseconds.
Value must be greater than 0 (default: 1800000, 1800 seconds, 30 minutes).
enable_http_compression google.protobuf.BoolValue
Enables or disables data compression in HTTP responses.
By default, ClickHouse stores data compressed. When executing a query, its result is uncompressed. Use this setting to command ClickHouse to compress the result when sending it via HTTP.
Enable this setting and add the Accept-Encoding: HTTP header in a HTTP request to force compression of HTTP response from ClickHouse.
ClickHouse support the following compression methods: gzip, br and deflate.
Default value: false (compression is disabled).
See in-depth description in ClickHouse documentation.
send_progress_in_http_headers google.protobuf.BoolValue
Enables progress notifications using X-ClickHouse-Progress HTTP header.
Default value: false (notifications disabled).
http_headers_progress_interval google.protobuf.Int64Value
Minimum interval between progress notifications with X-ClickHouse-Progress HTTP header, in milliseconds.
Value must be greater than 0 (default: 100).
add_http_cors_header google.protobuf.BoolValue
Adds CORS header in HTTP responses.
Default value: false (header is not added).
quota_mode enum QuotaMode
Quota accounting mode. Possible values: QUOTA_MODE_DEFAULT, QUOTA_MODE_KEYED and QUOTA_MODE_KEYED_BY_IP.

UserQuota

Field Description
interval_duration google.protobuf.Int64Value
Duration of interval for quota in milliseconds. Minimal value is 1 second. The minimum value is 1000.
queries google.protobuf.Int64Value
The total number of queries. 0 - unlimited. The minimum value is 0.
errors google.protobuf.Int64Value
The number of queries that threw exception. 0 - unlimited. The minimum value is 0.
result_rows google.protobuf.Int64Value
The total number of rows given as the result.. 0 - unlimited. The minimum value is 0.
read_rows google.protobuf.Int64Value
The total number of source rows read from tables for running the query, on all remote servers. 0 - unlimited. The minimum value is 0.
execution_time google.protobuf.Int64Value
The total query execution time, in milliseconds (wall time). 0 - unlimited. The minimum value is 0.

Была ли статья полезна?

Language / Region
© 2022 ООО «Яндекс.Облако»
В этой статье:
  • Calls UserService
  • Get
  • GetUserRequest
  • User
  • Permission
  • UserSettings
  • UserQuota
  • List
  • ListUsersRequest
  • ListUsersResponse
  • User
  • Permission
  • UserSettings
  • UserQuota
  • Create
  • CreateUserRequest
  • UserSpec
  • Permission
  • UserSettings
  • UserQuota
  • Operation
  • CreateUserMetadata
  • User
  • Permission
  • UserSettings
  • UserQuota
  • Update
  • UpdateUserRequest
  • Permission
  • UserSettings
  • UserQuota
  • Operation
  • UpdateUserMetadata
  • User
  • Permission
  • UserSettings
  • UserQuota
  • Delete
  • DeleteUserRequest
  • Operation
  • DeleteUserMetadata
  • GrantPermission
  • GrantUserPermissionRequest
  • Permission
  • Operation
  • GrantUserPermissionMetadata
  • User
  • Permission
  • UserSettings
  • UserQuota
  • RevokePermission
  • RevokeUserPermissionRequest
  • Operation
  • RevokeUserPermissionMetadata
  • User
  • Permission
  • UserSettings
  • UserQuota