+ ln -s /usr/share/clickhouse-test/ci/stress.py /usr/bin/stress + ln -s /usr/share/clickhouse-test/clickhouse-test /usr/bin/clickhouse-test + source /attach_gdb.lib ++ source /utils.lib + source /stress_tests.lib ++ sysctl kernel.core_pattern=core.%e.%p-%P kernel.core_pattern = core.%e.%p-%P ++ OK='\tOK\t\N\t' ++ FAIL='\tFAIL\t\N\t' ++ FAILURE_CONTEXT_LINES=100 ++ FAILURE_CONTEXT_MAX_LINE_WIDTH=300 + install_packages package_folder + dpkg -i package_folder/clickhouse-common-static_24.3.12.76.altinitystable+msan_amd64.deb Selecting previously unselected package clickhouse-common-static. (Reading database ... 49227 files and directories currently installed.) Preparing to unpack .../clickhouse-common-static_24.3.12.76.altinitystable+msan_amd64.deb ... Unpacking clickhouse-common-static (24.3.12.76.altinitystable+msan) ... Setting up clickhouse-common-static (24.3.12.76.altinitystable+msan) ... + dpkg -i package_folder/clickhouse-common-static-dbg_24.3.12.76.altinitystable+msan_amd64.deb Selecting previously unselected package clickhouse-common-static-dbg. (Reading database ... 49256 files and directories currently installed.) Preparing to unpack .../clickhouse-common-static-dbg_24.3.12.76.altinitystable+msan_amd64.deb ... Unpacking clickhouse-common-static-dbg (24.3.12.76.altinitystable+msan) ... Setting up clickhouse-common-static-dbg (24.3.12.76.altinitystable+msan) ... + dpkg -i package_folder/clickhouse-server_24.3.12.76.altinitystable+msan_amd64.deb Selecting previously unselected package clickhouse-server. (Reading database ... 49265 files and directories currently installed.) Preparing to unpack .../clickhouse-server_24.3.12.76.altinitystable+msan_amd64.deb ... Unpacking clickhouse-server (24.3.12.76.altinitystable+msan) ... Setting up clickhouse-server (24.3.12.76.altinitystable+msan) ... ClickHouse binary is already located at /usr/bin/clickhouse Symlink /usr/bin/clickhouse-server already exists but it points to /clickhouse. Will replace the old symlink to /usr/bin/clickhouse. Creating symlink /usr/bin/clickhouse-server to /usr/bin/clickhouse. Creating symlink /usr/bin/clickhouse-client to /usr/bin/clickhouse. Creating symlink /usr/bin/clickhouse-local to /usr/bin/clickhouse. Creating symlink /usr/bin/clickhouse-benchmark to /usr/bin/clickhouse. Creating symlink /usr/bin/clickhouse-obfuscator to /usr/bin/clickhouse. Creating symlink /usr/bin/clickhouse-git-import to /usr/bin/clickhouse. Creating symlink /usr/bin/clickhouse-compressor to /usr/bin/clickhouse. Creating symlink /usr/bin/clickhouse-format to /usr/bin/clickhouse. Symlink /usr/bin/clickhouse-extract-from-config already exists but it points to /clickhouse. Will replace the old symlink to /usr/bin/clickhouse. Creating symlink /usr/bin/clickhouse-extract-from-config to /usr/bin/clickhouse. Symlink /usr/bin/clickhouse-keeper already exists but it points to /clickhouse. Will replace the old symlink to /usr/bin/clickhouse. Creating symlink /usr/bin/clickhouse-keeper to /usr/bin/clickhouse. Symlink /usr/bin/clickhouse-keeper-converter already exists but it points to /clickhouse. Will replace the old symlink to /usr/bin/clickhouse. Creating symlink /usr/bin/clickhouse-keeper-converter to /usr/bin/clickhouse. Creating symlink /usr/bin/clickhouse-disks to /usr/bin/clickhouse. Creating symlink /usr/bin/ch to /usr/bin/clickhouse. Creating symlink /usr/bin/chl to /usr/bin/clickhouse. Creating symlink /usr/bin/chc to /usr/bin/clickhouse. Creating clickhouse group if it does not exist. groupadd -r clickhouse Creating clickhouse user if it does not exist. useradd -r --shell /bin/false --home-dir /nonexistent -g clickhouse clickhouse Will set ulimits for clickhouse user in /etc/security/limits.d/clickhouse.conf. Creating config directory /etc/clickhouse-server/config.d that is used for tweaks of main server configuration. Creating config directory /etc/clickhouse-server/users.d that is used for tweaks of users configuration. Config file /etc/clickhouse-server/config.xml already exists, will keep it and extract path info from it. /etc/clickhouse-server/config.xml has /var/lib/clickhouse/ as data path. /etc/clickhouse-server/config.xml has /var/log/clickhouse-server/ as log path. Users config file /etc/clickhouse-server/users.xml already exists, will keep it and extract users info from it. Log directory /var/log/clickhouse-server/ already exists. Creating data directory /var/lib/clickhouse/. Creating pid directory /var/run/clickhouse-server. chown -R clickhouse:clickhouse '/var/log/clickhouse-server/' chown -R clickhouse:clickhouse '/var/run/clickhouse-server' chown clickhouse:clickhouse '/var/lib/clickhouse/' groupadd -r clickhouse-bridge useradd -r --shell /bin/false --home-dir /nonexistent -g clickhouse-bridge clickhouse-bridge chown -R clickhouse-bridge:clickhouse-bridge '/usr/bin/clickhouse-odbc-bridge' chown -R clickhouse-bridge:clickhouse-bridge '/usr/bin/clickhouse-library-bridge' Password for the default user is an empty string. See /etc/clickhouse-server/users.xml and /etc/clickhouse-server/users.d to change it. Setting capabilities for clickhouse binary. This is optional. chown -R clickhouse:clickhouse '/etc/clickhouse-server' ClickHouse has been successfully installed. Start clickhouse-server with: sudo clickhouse start Start clickhouse-client with: clickhouse-client + dpkg -i package_folder/clickhouse-client_24.3.12.76.altinitystable+msan_amd64.deb Selecting previously unselected package clickhouse-client. (Reading database ... 49282 files and directories currently installed.) Preparing to unpack .../clickhouse-client_24.3.12.76.altinitystable+msan_amd64.deb ... Unpacking clickhouse-client (24.3.12.76.altinitystable+msan) ... Setting up clickhouse-client (24.3.12.76.altinitystable+msan) ... + export THREAD_FUZZER_CPU_TIME_PERIOD_US=1000 + THREAD_FUZZER_CPU_TIME_PERIOD_US=1000 + export THREAD_FUZZER_SLEEP_PROBABILITY=0.1 + THREAD_FUZZER_SLEEP_PROBABILITY=0.1 + export THREAD_FUZZER_SLEEP_TIME_US_MAX=100000 + THREAD_FUZZER_SLEEP_TIME_US_MAX=100000 + export THREAD_FUZZER_pthread_mutex_lock_BEFORE_MIGRATE_PROBABILITY=1 + THREAD_FUZZER_pthread_mutex_lock_BEFORE_MIGRATE_PROBABILITY=1 + export THREAD_FUZZER_pthread_mutex_lock_AFTER_MIGRATE_PROBABILITY=1 + THREAD_FUZZER_pthread_mutex_lock_AFTER_MIGRATE_PROBABILITY=1 + export THREAD_FUZZER_pthread_mutex_unlock_BEFORE_MIGRATE_PROBABILITY=1 + THREAD_FUZZER_pthread_mutex_unlock_BEFORE_MIGRATE_PROBABILITY=1 + export THREAD_FUZZER_pthread_mutex_unlock_AFTER_MIGRATE_PROBABILITY=1 + THREAD_FUZZER_pthread_mutex_unlock_AFTER_MIGRATE_PROBABILITY=1 + export THREAD_FUZZER_pthread_mutex_lock_BEFORE_SLEEP_PROBABILITY=0.001 + THREAD_FUZZER_pthread_mutex_lock_BEFORE_SLEEP_PROBABILITY=0.001 + export THREAD_FUZZER_pthread_mutex_lock_AFTER_SLEEP_PROBABILITY=0.001 + THREAD_FUZZER_pthread_mutex_lock_AFTER_SLEEP_PROBABILITY=0.001 + export THREAD_FUZZER_pthread_mutex_unlock_BEFORE_SLEEP_PROBABILITY=0.001 + THREAD_FUZZER_pthread_mutex_unlock_BEFORE_SLEEP_PROBABILITY=0.001 + export THREAD_FUZZER_pthread_mutex_unlock_AFTER_SLEEP_PROBABILITY=0.001 + THREAD_FUZZER_pthread_mutex_unlock_AFTER_SLEEP_PROBABILITY=0.001 + export THREAD_FUZZER_pthread_mutex_lock_BEFORE_SLEEP_TIME_US_MAX=10000 + THREAD_FUZZER_pthread_mutex_lock_BEFORE_SLEEP_TIME_US_MAX=10000 + export THREAD_FUZZER_pthread_mutex_lock_AFTER_SLEEP_TIME_US_MAX=10000 + THREAD_FUZZER_pthread_mutex_lock_AFTER_SLEEP_TIME_US_MAX=10000 + export THREAD_FUZZER_pthread_mutex_unlock_BEFORE_SLEEP_TIME_US_MAX=10000 + THREAD_FUZZER_pthread_mutex_unlock_BEFORE_SLEEP_TIME_US_MAX=10000 + export THREAD_FUZZER_pthread_mutex_unlock_AFTER_SLEEP_TIME_US_MAX=10000 + THREAD_FUZZER_pthread_mutex_unlock_AFTER_SLEEP_TIME_US_MAX=10000 + export THREAD_FUZZER_EXPLICIT_SLEEP_PROBABILITY=0.01 + THREAD_FUZZER_EXPLICIT_SLEEP_PROBABILITY=0.01 + export THREAD_FUZZER_EXPLICIT_MEMORY_EXCEPTION_PROBABILITY=0.01 + THREAD_FUZZER_EXPLICIT_MEMORY_EXCEPTION_PROBABILITY=0.01 + export ZOOKEEPER_FAULT_INJECTION=1 + ZOOKEEPER_FAULT_INJECTION=1 + configure + export USE_DATABASE_ORDINARY=1 + USE_DATABASE_ORDINARY=1 + export EXPORT_S3_STORAGE_POLICIES=1 + EXPORT_S3_STORAGE_POLICIES=1 + /usr/share/clickhouse-test/config/install.sh + DEST_SERVER_PATH=/etc/clickhouse-server + DEST_CLIENT_PATH=/etc/clickhouse-client +++ dirname /usr/share/clickhouse-test/config/install.sh ++ cd /usr/share/clickhouse-test/config ++ pwd -P + SRC_PATH=/usr/share/clickhouse-test/config + echo 'Going to install test configs from /usr/share/clickhouse-test/config into /etc/clickhouse-server' + mkdir -p /etc/clickhouse-server/config.d/ Going to install test configs from /usr/share/clickhouse-test/config into /etc/clickhouse-server + mkdir -p /etc/clickhouse-server/users.d/ + mkdir -p /etc/clickhouse-client + ln -sf /usr/share/clickhouse-test/config/config.d/zookeeper_write.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/max_num_to_warn.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/listen.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/text_log.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/blob_storage_log.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/custom_settings_prefixes.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/enable_access_control_improvements.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/macros.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/secure_ports.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/clusters.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/graphite.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/graphite_alternative.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/database_atomic.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/max_concurrent_queries.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/merge_tree_settings.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/backoff_failed_mutation.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/merge_tree_old_dirs_cleanup.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/test_cluster_with_incorrect_pw.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/keeper_port.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/logging_no_rotate.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/merge_tree.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/lost_forever_check.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/tcp_with_proxy.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/prometheus.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/top_level_domains_lists.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/top_level_domains_path.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/transactions.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/encryption.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/CORS.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/zookeeper_log.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/logger_trace.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/named_collection.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/ssl_certs.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/filesystem_cache_log.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/session_log.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/system_unfreeze.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/enable_zero_copy_replication.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/nlp.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/forbidden_headers.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/enable_keeper_map.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/custom_disks_base_path.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/display_name.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/reverse_dns_query_function.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/compressed_marks_and_index.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/disable_s3_env_credentials.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/enable_wait_for_shutdown_replicated_tables.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/backups.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/filesystem_caches_path.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/validate_tcp_client_information.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/zero_copy_destructive_operations.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/block_number.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/handlers.yaml /etc/clickhouse-server/config.d/ + '[' /etc/clickhouse-server = /etc/clickhouse-server ']' + ln -sf /usr/share/clickhouse-test/config/config.d/legacy_geobase.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/users.d/log_queries.xml /etc/clickhouse-server/users.d/ + ln -sf /usr/share/clickhouse-test/config/users.d/readonly.xml /etc/clickhouse-server/users.d/ + ln -sf /usr/share/clickhouse-test/config/users.d/access_management.xml /etc/clickhouse-server/users.d/ + ln -sf /usr/share/clickhouse-test/config/users.d/database_atomic_drop_detach_sync.xml /etc/clickhouse-server/users.d/ + ln -sf /usr/share/clickhouse-test/config/users.d/opentelemetry.xml /etc/clickhouse-server/users.d/ + ln -sf /usr/share/clickhouse-test/config/users.d/remote_queries.xml /etc/clickhouse-server/users.d/ + ln -sf /usr/share/clickhouse-test/config/users.d/session_log_test.xml /etc/clickhouse-server/users.d/ + ln -sf /usr/share/clickhouse-test/config/users.d/memory_profiler.xml /etc/clickhouse-server/users.d/ + ln -sf /usr/share/clickhouse-test/config/users.d/no_fsync_metadata.xml /etc/clickhouse-server/users.d/ + ln -sf /usr/share/clickhouse-test/config/users.d/filelog.xml /etc/clickhouse-server/users.d/ + ln -sf /usr/share/clickhouse-test/config/users.d/enable_blobs_check.xml /etc/clickhouse-server/users.d/ + ln -sf /usr/share/clickhouse-test/config/users.d/marks.xml /etc/clickhouse-server/users.d/ + ln -sf /usr/share/clickhouse-test/config/users.d/insert_keeper_retries.xml /etc/clickhouse-server/users.d/ + ln -sf /usr/share/clickhouse-test/config/users.d/prefetch_settings.xml /etc/clickhouse-server/users.d/ + ln -sf /usr/share/clickhouse-test/config/users.d/nonconst_timezone.xml /etc/clickhouse-server/users.d/ + ln -sf /usr/share/clickhouse-test/config/users.d/allow_introspection_functions.yaml /etc/clickhouse-server/users.d/ + ln -sf /usr/share/clickhouse-test/config/users.d/replicated_ddl_entry.xml /etc/clickhouse-server/users.d/ + [[ -n '' ]] + ln -sf /usr/share/clickhouse-test/config/users.d/timeouts.xml /etc/clickhouse-server/users.d/ + ln -sf /usr/share/clickhouse-test/config/ints_dictionary.xml /etc/clickhouse-server/ + ln -sf /usr/share/clickhouse-test/config/strings_dictionary.xml /etc/clickhouse-server/ + ln -sf /usr/share/clickhouse-test/config/decimals_dictionary.xml /etc/clickhouse-server/ + ln -sf /usr/share/clickhouse-test/config/executable_dictionary.xml /etc/clickhouse-server/ + ln -sf /usr/share/clickhouse-test/config/executable_pool_dictionary.xml /etc/clickhouse-server/ + ln -sf /usr/share/clickhouse-test/config/test_function.xml /etc/clickhouse-server/ + ln -sf /usr/share/clickhouse-test/config/top_level_domains /etc/clickhouse-server/ + ln -sf /usr/share/clickhouse-test/config/regions_hierarchy.txt /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/regions_names_en.txt /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/ext-en.txt /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/ext-ru.txt /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/lem-en.bin /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/server.key /etc/clickhouse-server/ + ln -sf /usr/share/clickhouse-test/config/server.crt /etc/clickhouse-server/ + ln -sf /usr/share/clickhouse-test/config/dhparam.pem /etc/clickhouse-server/ + ln -sf --backup=simple --suffix=_original.xml /usr/share/clickhouse-test/config/config.d/query_masking_rules.xml /etc/clickhouse-server/config.d/ + [[ -n 1 ]] + [[ 1 -eq 1 ]] + rm -f /etc/clickhouse-server/config.d/zookeeper.xml + ln -sf /usr/share/clickhouse-test/config/config.d/zookeeper_fault_injection.xml /etc/clickhouse-server/config.d/ + [[ -n '' ]] + rm -f /etc/clickhouse-server/config.d/cannot_allocate_thread_injection.xml + value=0 + sed --follow-symlinks -i 's|[01]|0|' /etc/clickhouse-server/config.d/keeper_port.xml + value=24709120 + sed --follow-symlinks -i 's|[[:digit:]]\+|24709120|' /etc/clickhouse-server/config.d/keeper_port.xml + value=55672832 + sed --follow-symlinks -i 's|[[:digit:]]\+|55672832|' /etc/clickhouse-server/config.d/keeper_port.xml + [[ -n '' ]] + [[ -n 1 ]] + [[ 1 -eq 1 ]] + ln -sf /usr/share/clickhouse-test/config/users.d/database_ordinary.xml /etc/clickhouse-server/users.d/ + [[ -n '' ]] + ARM=aarch64 ++ uname -m x86_64 Adding azure configuration + OS=x86_64 + [[ -n 1 ]] + echo x86_64 + [[ '' -eq 1 ]] + [[ x86_64 == \a\a\r\c\h\6\4 ]] + echo 'Adding azure configuration' + ln -sf /usr/share/clickhouse-test/config/config.d/azure_storage_conf.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/storage_conf.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/storage_conf_02944.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/storage_conf_02963.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/storage_conf_02961.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/users.d/s3_cache.xml /etc/clickhouse-server/users.d/ + ln -sf /usr/share/clickhouse-test/config/users.d/s3_cache_new.xml /etc/clickhouse-server/users.d/ + [[ -n '' ]] + ln -sf /usr/share/clickhouse-test/config/client_config.xml /etc/clickhouse-client/config.xml + sudo cat /etc/clickhouse-server/config.d/keeper_port.xml + sed 's|100000|10000|' + sudo mv /etc/clickhouse-server/config.d/keeper_port.xml.tmp /etc/clickhouse-server/config.d/keeper_port.xml + [[ -n 1 ]] + [[ 1 -eq 1 ]] + randomize_config_boolean_value filtered_list keeper_port + value=0 + sudo cat /etc/clickhouse-server/config.d/keeper_port.xml + sed 's|[01]|0|' + sudo mv /etc/clickhouse-server/config.d/keeper_port.xml.tmp /etc/clickhouse-server/config.d/keeper_port.xml + randomize_config_boolean_value multi_read keeper_port + value=1 + sudo cat /etc/clickhouse-server/config.d/keeper_port.xml + sed 's|[01]|1|' + sudo mv /etc/clickhouse-server/config.d/keeper_port.xml.tmp /etc/clickhouse-server/config.d/keeper_port.xml + randomize_config_boolean_value check_not_exists keeper_port + value=1 + sed 's|[01]|1|' + sudo cat /etc/clickhouse-server/config.d/keeper_port.xml + sudo mv /etc/clickhouse-server/config.d/keeper_port.xml.tmp /etc/clickhouse-server/config.d/keeper_port.xml + randomize_config_boolean_value create_if_not_exists keeper_port + value=1 + sudo cat /etc/clickhouse-server/config.d/keeper_port.xml + sed 's|[01]|1|' + sudo mv /etc/clickhouse-server/config.d/keeper_port.xml.tmp /etc/clickhouse-server/config.d/keeper_port.xml + sudo chown clickhouse /etc/clickhouse-server/config.d/keeper_port.xml + sudo chgrp clickhouse /etc/clickhouse-server/config.d/keeper_port.xml + [[ -n 1 ]] + [[ 1 -eq 1 ]] + randomize_config_boolean_value use_compression zookeeper_fault_injection + value=1 + sudo cat /etc/clickhouse-server/config.d/zookeeper_fault_injection.xml + sed 's|[01]|1|' + sudo mv /etc/clickhouse-server/config.d/zookeeper_fault_injection.xml.tmp /etc/clickhouse-server/config.d/zookeeper_fault_injection.xml + randomize_config_boolean_value allow_experimental_block_number_column block_number + value=0 + sudo cat /etc/clickhouse-server/config.d/block_number.xml + sed 's|[01]|0|' + sudo mv /etc/clickhouse-server/config.d/block_number.xml.tmp /etc/clickhouse-server/config.d/block_number.xml + echo 'ASAN_OPTIONS='\''malloc_context_size=10 verbosity=1 allocator_release_to_os_interval_ms=10000'\''' + export 'ASAN_OPTIONS=malloc_context_size=10 allocator_release_to_os_interval_ms=10000' + ASAN_OPTIONS='malloc_context_size=10 allocator_release_to_os_interval_ms=10000' + sudo chown root: /var/lib/clickhouse + echo '1' + local total_mem ++ awk '/MemTotal/ { print $(NF-1) }' /proc/meminfo + total_mem=32086428 + total_mem=32856502272 + max_server_memory_usage_to_ram_ratio=0.5 Setting max_server_memory_usage_to_ram_ratio to 0.5 + echo 'Setting max_server_memory_usage_to_ram_ratio to 0.5' + cat + local max_users_mem + max_users_mem=9856950681 + echo 'Setting max_memory_usage_for_user=9856950681 and max_memory_usage for queries to 10G' + cat Setting max_memory_usage_for_user=9856950681 and max_memory_usage for queries to 10G + cat + ./setup_minio.sh stateless + azurite-blob --blobHost 0.0.0.0 --blobPort 10000 --debug /azurite_log + export MINIO_ROOT_USER=clickhouse + MINIO_ROOT_USER=clickhouse + export MINIO_ROOT_PASSWORD=clickhouse + MINIO_ROOT_PASSWORD=clickhouse + main stateless + local query_dir ++ check_arg stateless ++ local query_dir ++ '[' '!' 1 -eq 1 ']' ++ case "$1" in ++ query_dir=0_stateless ++ echo 0_stateless + query_dir=0_stateless + '[' '!' -f ./minio ']' + start_minio + mkdir -p ./minio_data + ./minio --version Azurite Blob service is starting on 0.0.0.0:10000 Azurite Blob service successfully listens on http://0.0.0.0:10000 minio version RELEASE.2022-01-03T18-22-58Z + wait_for_it + local counter=0 + local max_counter=60 + local url=http://localhost:11111 + params=('--silent' '--verbose') + local params + ./minio server --address :11111 ./minio_data + curl --silent --verbose http://localhost:11111 + grep AccessDenied trying to connect to minio + [[ 0 == \6\0 ]] + echo 'trying to connect to minio' + sleep 1 + counter=1 + grep AccessDenied + curl --silent --verbose http://localhost:11111 + [[ 1 == \6\0 ]] + echo 'trying to connect to minio' + sleep 1 trying to connect to minio + counter=2 + curl --silent --verbose http://localhost:11111 + grep AccessDenied + [[ 2 == \6\0 ]] + echo 'trying to connect to minio' + sleep 1 trying to connect to minio + counter=3 + curl --silent --verbose http://localhost:11111 + grep AccessDenied trying to connect to minio + [[ 3 == \6\0 ]] + echo 'trying to connect to minio' + sleep 1 + counter=4 + curl --silent --verbose http://localhost:11111 + grep AccessDenied trying to connect to minio + [[ 4 == \6\0 ]] + echo 'trying to connect to minio' + sleep 1 + counter=5 + curl --silent --verbose http://localhost:11111 + grep AccessDenied + [[ 5 == \6\0 ]] + echo 'trying to connect to minio' + sleep 1 trying to connect to minio + counter=6 + curl --silent --verbose http://localhost:11111 + grep AccessDenied + [[ 6 == \6\0 ]] + echo 'trying to connect to minio' + sleep 1 trying to connect to minio + counter=7 + curl --silent --verbose http://localhost:11111 + grep AccessDenied + [[ 7 == \6\0 ]] + echo 'trying to connect to minio' + sleep 1 trying to connect to minio + counter=8 + curl --silent --verbose http://localhost:11111 + grep AccessDenied trying to connect to minio + [[ 8 == \6\0 ]] + echo 'trying to connect to minio' + sleep 1 + counter=9 + curl --silent --verbose http://localhost:11111 + grep AccessDenied + [[ 9 == \6\0 ]] + echo 'trying to connect to minio' + sleep 1 trying to connect to minio + counter=10 + curl --silent --verbose http://localhost:11111 + grep AccessDenied + [[ 10 == \6\0 ]] + echo 'trying to connect to minio' + sleep 1 trying to connect to minio API: http://172.17.0.2:11111 http://127.0.0.1:11111 Console: http://172.17.0.2:45545 http://127.0.0.1:45545 Documentation: https://docs.min.io WARNING: Console endpoint is listening on a dynamic port (45545), please use --console-address ":PORT" to choose a static port. + counter=11 + curl --silent --verbose http://localhost:11111 + grep AccessDenied AccessDeniedAccess Denied./18119AF0415413B1c978493f-9314-4977-ae9d-01cf681f6eb5 + lsof -i :11111 COMMAND PID USER FD TYPE DEVICE SIZE/OFF NODE NAME minio 291 root 10u IPv6 42225 0t0 TCP *:11111 (LISTEN) + sleep 5 + setup_minio stateless + local test_type=stateless + ./mc alias set clickminio http://localhost:11111 clickhouse clickhouse Added `clickminio` successfully. + ./mc admin user add clickminio test testtest Added user `test` successfully. + ./mc admin policy set clickminio readwrite user=test Policy `readwrite` is set on user `test` + ./mc mb clickminio/test Bucket created successfully `clickminio/test`. + '[' stateless = stateless ']' + ./mc policy set public clickminio/test Access permission for `clickminio/test` is set to `public` + upload_data 0_stateless /usr/share/clickhouse-test + local query_dir=0_stateless + local test_path=/usr/share/clickhouse-test + local data_path=/usr/share/clickhouse-test/queries/0_stateless/data_minio ++ ls /usr/share/clickhouse-test/queries/0_stateless/data_minio + for file in $(ls "${data_path}") + echo 02366_data.jsonl + ./mc cp /usr/share/clickhouse-test/queries/0_stateless/data_minio/02366_data.jsonl clickminio/test/02366_data.jsonl 02366_data.jsonl `/usr/share/clickhouse-test/queries/0_stateless/data_minio/02366_data.jsonl` -> `clickminio/test/02366_data.jsonl` Total: 0 B, Transferred: 0 B, Speed: 0 B/s + for file in $(ls "${data_path}") + echo 02731.arrow + ./mc cp /usr/share/clickhouse-test/queries/0_stateless/data_minio/02731.arrow clickminio/test/02731.arrow 02731.arrow `/usr/share/clickhouse-test/queries/0_stateless/data_minio/02731.arrow` -> `clickminio/test/02731.arrow` Total: 0 B, Transferred: 3.82 MiB, Speed: 80.88 MiB/s + for file in $(ls "${data_path}") + echo 02731.parquet + ./mc cp /usr/share/clickhouse-test/queries/0_stateless/data_minio/02731.parquet clickminio/test/02731.parquet 02731.parquet `/usr/share/clickhouse-test/queries/0_stateless/data_minio/02731.parquet` -> `clickminio/test/02731.parquet` Total: 0 B, Transferred: 1.57 MiB, Speed: 65.21 MiB/s 02876.parquet + for file in $(ls "${data_path}") + echo 02876.parquet + ./mc cp /usr/share/clickhouse-test/queries/0_stateless/data_minio/02876.parquet clickminio/test/02876.parquet `/usr/share/clickhouse-test/queries/0_stateless/data_minio/02876.parquet` -> `clickminio/test/02876.parquet` Total: 0 B, Transferred: 293 B, Speed: 29.44 KiB/s a.tsv + for file in $(ls "${data_path}") + echo a.tsv + ./mc cp /usr/share/clickhouse-test/queries/0_stateless/data_minio/a.tsv clickminio/test/a.tsv `/usr/share/clickhouse-test/queries/0_stateless/data_minio/a.tsv` -> `clickminio/test/a.tsv` Total: 0 B, Transferred: 24 B, Speed: 2.63 KiB/s b.tsv + for file in $(ls "${data_path}") + echo b.tsv + ./mc cp /usr/share/clickhouse-test/queries/0_stateless/data_minio/b.tsv clickminio/test/b.tsv `/usr/share/clickhouse-test/queries/0_stateless/data_minio/b.tsv` -> `clickminio/test/b.tsv` Total: 0 B, Transferred: 33 B, Speed: 2.06 KiB/s + for file in $(ls "${data_path}") + echo c.tsv + ./mc cp /usr/share/clickhouse-test/queries/0_stateless/data_minio/c.tsv clickminio/test/c.tsv c.tsv `/usr/share/clickhouse-test/queries/0_stateless/data_minio/c.tsv` -> `clickminio/test/c.tsv` Total: 0 B, Transferred: 33 B, Speed: 4.25 KiB/s + for file in $(ls "${data_path}") + echo tsv_with_header.tsv + ./mc cp /usr/share/clickhouse-test/queries/0_stateless/data_minio/tsv_with_header.tsv clickminio/test/tsv_with_header.tsv tsv_with_header.tsv `/usr/share/clickhouse-test/queries/0_stateless/data_minio/tsv_with_header.tsv` -> `clickminio/test/tsv_with_header.tsv` Total: 0 B, Transferred: 44 B, Speed: 5.31 KiB/s + setup_aws_credentials + local minio_root_user=clickhouse + local minio_root_password=clickhouse + mkdir -p /root/.aws + cat + config_logs_export_cluster /etc/clickhouse-server/config.d/system_logs_export.yaml + set +x File /tmp/export-logs-config.sh does not exist, do not setup + start_server + counter=0 + max_attempt=120 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 0 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=1 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 1 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=2 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 2 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=3 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 3 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=4 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 4 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=5 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 5 -gt 120 ']' + clickhouse start --user root 127.0.0.1 - - [16/Dec/2024:08:28:19 +0000] "PUT /devstoreaccount1/cont?restype=container HTTP/1.1" 201 - 127.0.0.1 - - [16/Dec/2024:08:28:19 +0000] "PUT /devstoreaccount1/cont/qrxrldwkjvdicymikkiicluqzqnxblaw?blockid=cqkqjinybvxvkmjaqxuudslqbpztxyfovgdzusbllwprqmzygbavmflysjswmqdj&comp=block HTTP/1.1" 201 - + sleep 0.5 127.0.0.1 - - [16/Dec/2024:08:28:19 +0000] "PUT /devstoreaccount1/cont/qrxrldwkjvdicymikkiicluqzqnxblaw?comp=blocklist HTTP/1.1" 201 - 127.0.0.1 - - [16/Dec/2024:08:28:19 +0000] "GET /devstoreaccount1/cont/qrxrldwkjvdicymikkiicluqzqnxblaw HTTP/1.1" 206 4 127.0.0.1 - - [16/Dec/2024:08:28:19 +0000] "GET /devstoreaccount1/cont/qrxrldwkjvdicymikkiicluqzqnxblaw HTTP/1.1" 206 2 127.0.0.1 - - [16/Dec/2024:08:28:19 +0000] "DELETE /devstoreaccount1/cont/qrxrldwkjvdicymikkiicluqzqnxblaw HTTP/1.1" 202 - + counter=6 + clickhouse-client --query 'SELECT 1' 1 + attach_gdb_to_clickhouse ++ kill -l SIGRTMIN + RTMIN=34 + echo ' set follow-fork-mode parent handle SIGHUP nostop noprint pass handle SIGINT nostop noprint pass handle SIGQUIT nostop noprint pass handle SIGPIPE nostop noprint pass handle SIGTERM nostop noprint pass handle SIGUSR1 nostop noprint pass handle SIGUSR2 nostop noprint pass handle SIG34 nostop noprint pass info signals continue backtrace full thread apply all backtrace full info registers disassemble /s up disassemble /s up disassemble /s p "done" detach quit ' + sleep 5 + ts '%Y-%m-%d %H:%M:%S' ++ cat /var/run/clickhouse-server/clickhouse-server.pid + gdb -batch -command script.gdb -p 577 + run_with_retry 60 clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' + [[ hxB =~ e ]] + set_e=false + set +e + local total_retries=60 + shift + local retry=0 + '[' 0 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Connected to clickhouse-server after attaching gdb + false + return + setup_logs_replication + set +x File /tmp/export-logs-config.sh does not exist, do not setup + clickhouse-client --query 'CREATE DATABASE datasets' + clickhouse-client --multiquery + clickhouse-client --query 'SHOW TABLES FROM datasets' hits_v1 visits_v1 + clickhouse-client --query 'CREATE DATABASE IF NOT EXISTS test' + stop_server + local max_tries=90 + local check_hang=true + local pid ++ cat /var/run/clickhouse-server/clickhouse-server.pid + pid=577 + clickhouse stop --max-tries 90 --do-not-kill script.gdb:13: Error in sourced command file: No stack. /var/run/clickhouse-server/clickhouse-server.pid file exists and contains pid = 577. The process with pid = 577 is running. Sent terminate signal to process with pid 577. Waiting for server to stop /var/run/clickhouse-server/clickhouse-server.pid file exists and contains pid = 577. The process with pid = 577 is running. Waiting for server to stop /var/run/clickhouse-server/clickhouse-server.pid file exists and contains pid = 577. The process with pid = 577 is running. Waiting for server to stop /var/run/clickhouse-server/clickhouse-server.pid file exists and contains pid = 577. The process with pid = 577 is running. Waiting for server to stop /var/run/clickhouse-server/clickhouse-server.pid file exists and contains pid = 577. The process with pid = 577 is running. Waiting for server to stop /var/run/clickhouse-server/clickhouse-server.pid file exists and contains pid = 577. The process with pid = 577 is running. Waiting for server to stop /var/run/clickhouse-server/clickhouse-server.pid file exists and contains pid = 577. The process with pid = 577 is running. Waiting for server to stop Now there is no clickhouse-server process. Server stopped + return + mv /var/log/clickhouse-server/clickhouse-server.log /var/log/clickhouse-server/clickhouse-server.initial.log + cache_policy= + '[' 1 -eq 1 ']' + cache_policy=SLRU + echo 'Using cache policy: SLRU' + '[' SLRU = SLRU ']' Using cache policy: SLRU + sudo cat /etc/clickhouse-server/config.d/storage_conf.xml + sed 's|LRU|SLRU|' + mv /etc/clickhouse-server/config.d/storage_conf.xml.tmp /etc/clickhouse-server/config.d/storage_conf.xml + start_server + counter=0 + max_attempt=120 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 0 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=1 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 1 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=2 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 2 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=3 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 3 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=4 + clickhouse-client --query 'SELECT 1' 127.0.0.1 - - [16/Dec/2024:08:29:28 +0000] "PUT /devstoreaccount1/cont?restype=container HTTP/1.1" 409 - 127.0.0.1 - - [16/Dec/2024:08:29:28 +0000] "PUT /devstoreaccount1/cont/geieuxexxdhbruvtikzvakhigmcohjki?blockid=eiskqqbzezntbvrlphnkmylrqmsbzimozppuwnqjmdhjwbxhrdrabxusjegaklug&comp=block HTTP/1.1" 201 - 127.0.0.1 - - [16/Dec/2024:08:29:28 +0000] "PUT /devstoreaccount1/cont/geieuxexxdhbruvtikzvakhigmcohjki?comp=blocklist HTTP/1.1" 201 - 127.0.0.1 - - [16/Dec/2024:08:29:28 +0000] "GET /devstoreaccount1/cont/geieuxexxdhbruvtikzvakhigmcohjki HTTP/1.1" 206 4 127.0.0.1 - - [16/Dec/2024:08:29:28 +0000] "GET /devstoreaccount1/cont/geieuxexxdhbruvtikzvakhigmcohjki HTTP/1.1" 206 2 127.0.0.1 - - [16/Dec/2024:08:29:28 +0000] "DELETE /devstoreaccount1/cont/geieuxexxdhbruvtikzvakhigmcohjki HTTP/1.1" 202 - Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 4 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=5 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 5 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=6 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 6 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=7 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 7 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=8 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 8 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=9 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 9 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=10 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 10 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=11 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 11 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=12 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 12 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=13 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 13 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=14 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 14 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=15 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 15 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=16 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 16 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=17 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 17 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=18 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 18 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=19 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 19 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=20 + clickhouse-client --query 'SELECT 1' 1 + attach_gdb_to_clickhouse ++ kill -l SIGRTMIN + RTMIN=34 + echo ' set follow-fork-mode parent handle SIGHUP nostop noprint pass handle SIGINT nostop noprint pass handle SIGQUIT nostop noprint pass handle SIGPIPE nostop noprint pass handle SIGTERM nostop noprint pass handle SIGUSR1 nostop noprint pass handle SIGUSR2 nostop noprint pass handle SIG34 nostop noprint pass info signals continue backtrace full thread apply all backtrace full info registers disassemble /s up disassemble /s up disassemble /s p "done" detach quit ' + sleep 5 + ts '%Y-%m-%d %H:%M:%S' ++ cat /var/run/clickhouse-server/clickhouse-server.pid + gdb -batch -command script.gdb -p 1413 + run_with_retry 60 clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' + [[ hxB =~ e ]] + set_e=false + set +e + local total_retries=60 + shift + local retry=0 + '[' 0 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Connected to clickhouse-server after attaching gdb + false + return + clickhouse-client --query 'SHOW TABLES FROM datasets' hits_v1 visits_v1 + clickhouse-client --query 'SHOW TABLES FROM test' + clickhouse-client --query 'CREATE TABLE test.hits_s3 (WatchID UInt64, JavaEnable UInt8, Title String, GoodEvent Int16, EventTime DateTime, EventDate Date, CounterID UInt32, ClientIP UInt32, ClientIP6 FixedString(16), RegionID UInt32, UserID UInt64, CounterClass Int8, OS UInt8, UserAgent UInt8, URL String, Referer String, URLDomain String, RefererDomain String, Refresh UInt8, IsRobot UInt8, RefererCategories Array(UInt16), URLCategories Array(UInt16), URLRegions Array(UInt32), RefererRegions Array(UInt32), ResolutionWidth UInt16, ResolutionHeight UInt16, ResolutionDepth UInt8, FlashMajor UInt8, FlashMinor UInt8, FlashMinor2 String, NetMajor UInt8, NetMinor UInt8, UserAgentMajor UInt16, UserAgentMinor FixedString(2), CookieEnable UInt8, JavascriptEnable UInt8, IsMobile UInt8, MobilePhone UInt8, MobilePhoneModel String, Params String, IPNetworkID UInt32, TraficSourceID Int8, SearchEngineID UInt16, SearchPhrase String, AdvEngineID UInt8, IsArtifical UInt8, WindowClientWidth UInt16, WindowClientHeight UInt16, ClientTimeZone Int16, ClientEventTime DateTime, SilverlightVersion1 UInt8, SilverlightVersion2 UInt8, SilverlightVersion3 UInt32, SilverlightVersion4 UInt16, PageCharset String, CodeVersion UInt32, IsLink UInt8, IsDownload UInt8, IsNotBounce UInt8, FUniqID UInt64, HID UInt32, IsOldCounter UInt8, IsEvent UInt8, IsParameter UInt8, DontCountHits UInt8, WithHash UInt8, HitColor FixedString(1), UTCEventTime DateTime, Age UInt8, Sex UInt8, Income UInt8, Interests UInt16, Robotness UInt8, GeneralInterests Array(UInt16), RemoteIP UInt32, RemoteIP6 FixedString(16), WindowName Int32, OpenerName Int32, HistoryLength Int16, BrowserLanguage FixedString(2), BrowserCountry FixedString(2), SocialNetwork String, SocialAction String, HTTPError UInt16, SendTiming Int32, DNSTiming Int32, ConnectTiming Int32, ResponseStartTiming Int32, ResponseEndTiming Int32, FetchTiming Int32, RedirectTiming Int32, DOMInteractiveTiming Int32, DOMContentLoadedTiming Int32, DOMCompleteTiming Int32, LoadEventStartTiming Int32, LoadEventEndTiming Int32, NSToDOMContentLoadedTiming Int32, FirstPaintTiming Int32, RedirectCount Int8, SocialSourceNetworkID UInt8, SocialSourcePage String, ParamPrice Int64, ParamOrderID String, ParamCurrency FixedString(3), ParamCurrencyID UInt16, GoalsReached Array(UInt32), OpenstatServiceName String, OpenstatCampaignID String, OpenstatAdID String, OpenstatSourceID String, UTMSource String, UTMMedium String, UTMCampaign String, UTMContent String, UTMTerm String, FromTag String, HasGCLID UInt8, RefererHash UInt64, URLHash UInt64, CLID UInt32, YCLID UInt64, ShareService String, ShareURL String, ShareTitle String, ParsedParams Nested(Key1 String, Key2 String, Key3 String, Key4 String, Key5 String, ValueDouble Float64), IslandID FixedString(16), RequestNum UInt32, RequestTry UInt8) ENGINE = MergeTree() PARTITION BY toYYYYMM(EventDate) ORDER BY (CounterID, EventDate, intHash32(UserID)) SAMPLE BY intHash32(UserID) SETTINGS index_granularity = 8192, storage_policy='\''s3_cache'\''' + clickhouse-client --query 'CREATE TABLE test.hits (WatchID UInt64, JavaEnable UInt8, Title String, GoodEvent Int16, EventTime DateTime, EventDate Date, CounterID UInt32, ClientIP UInt32, ClientIP6 FixedString(16), RegionID UInt32, UserID UInt64, CounterClass Int8, OS UInt8, UserAgent UInt8, URL String, Referer String, URLDomain String, RefererDomain String, Refresh UInt8, IsRobot UInt8, RefererCategories Array(UInt16), URLCategories Array(UInt16), URLRegions Array(UInt32), RefererRegions Array(UInt32), ResolutionWidth UInt16, ResolutionHeight UInt16, ResolutionDepth UInt8, FlashMajor UInt8, FlashMinor UInt8, FlashMinor2 String, NetMajor UInt8, NetMinor UInt8, UserAgentMajor UInt16, UserAgentMinor FixedString(2), CookieEnable UInt8, JavascriptEnable UInt8, IsMobile UInt8, MobilePhone UInt8, MobilePhoneModel String, Params String, IPNetworkID UInt32, TraficSourceID Int8, SearchEngineID UInt16, SearchPhrase String, AdvEngineID UInt8, IsArtifical UInt8, WindowClientWidth UInt16, WindowClientHeight UInt16, ClientTimeZone Int16, ClientEventTime DateTime, SilverlightVersion1 UInt8, SilverlightVersion2 UInt8, SilverlightVersion3 UInt32, SilverlightVersion4 UInt16, PageCharset String, CodeVersion UInt32, IsLink UInt8, IsDownload UInt8, IsNotBounce UInt8, FUniqID UInt64, HID UInt32, IsOldCounter UInt8, IsEvent UInt8, IsParameter UInt8, DontCountHits UInt8, WithHash UInt8, HitColor FixedString(1), UTCEventTime DateTime, Age UInt8, Sex UInt8, Income UInt8, Interests UInt16, Robotness UInt8, GeneralInterests Array(UInt16), RemoteIP UInt32, RemoteIP6 FixedString(16), WindowName Int32, OpenerName Int32, HistoryLength Int16, BrowserLanguage FixedString(2), BrowserCountry FixedString(2), SocialNetwork String, SocialAction String, HTTPError UInt16, SendTiming Int32, DNSTiming Int32, ConnectTiming Int32, ResponseStartTiming Int32, ResponseEndTiming Int32, FetchTiming Int32, RedirectTiming Int32, DOMInteractiveTiming Int32, DOMContentLoadedTiming Int32, DOMCompleteTiming Int32, LoadEventStartTiming Int32, LoadEventEndTiming Int32, NSToDOMContentLoadedTiming Int32, FirstPaintTiming Int32, RedirectCount Int8, SocialSourceNetworkID UInt8, SocialSourcePage String, ParamPrice Int64, ParamOrderID String, ParamCurrency FixedString(3), ParamCurrencyID UInt16, GoalsReached Array(UInt32), OpenstatServiceName String, OpenstatCampaignID String, OpenstatAdID String, OpenstatSourceID String, UTMSource String, UTMMedium String, UTMCampaign String, UTMContent String, UTMTerm String, FromTag String, HasGCLID UInt8, RefererHash UInt64, URLHash UInt64, CLID UInt32, YCLID UInt64, ShareService String, ShareURL String, ShareTitle String, ParsedParams Nested(Key1 String, Key2 String, Key3 String, Key4 String, Key5 String, ValueDouble Float64), IslandID FixedString(16), RequestNum UInt32, RequestTry UInt8) ENGINE = MergeTree() PARTITION BY toYYYYMM(EventDate) ORDER BY (CounterID, EventDate, intHash32(UserID)) SAMPLE BY intHash32(UserID) SETTINGS index_granularity = 8192, storage_policy='\''s3_cache'\''' + clickhouse-client --query 'CREATE TABLE test.visits (CounterID UInt32, StartDate Date, Sign Int8, IsNew UInt8, VisitID UInt64, UserID UInt64, StartTime DateTime, Duration UInt32, UTCStartTime DateTime, PageViews Int32, Hits Int32, IsBounce UInt8, Referer String, StartURL String, RefererDomain String, StartURLDomain String, EndURL String, LinkURL String, IsDownload UInt8, TraficSourceID Int8, SearchEngineID UInt16, SearchPhrase String, AdvEngineID UInt8, PlaceID Int32, RefererCategories Array(UInt16), URLCategories Array(UInt16), URLRegions Array(UInt32), RefererRegions Array(UInt32), IsYandex UInt8, GoalReachesDepth Int32, GoalReachesURL Int32, GoalReachesAny Int32, SocialSourceNetworkID UInt8, SocialSourcePage String, MobilePhoneModel String, ClientEventTime DateTime, RegionID UInt32, ClientIP UInt32, ClientIP6 FixedString(16), RemoteIP UInt32, RemoteIP6 FixedString(16), IPNetworkID UInt32, SilverlightVersion3 UInt32, CodeVersion UInt32, ResolutionWidth UInt16, ResolutionHeight UInt16, UserAgentMajor UInt16, UserAgentMinor UInt16, WindowClientWidth UInt16, WindowClientHeight UInt16, SilverlightVersion2 UInt8, SilverlightVersion4 UInt16, FlashVersion3 UInt16, FlashVersion4 UInt16, ClientTimeZone Int16, OS UInt8, UserAgent UInt8, ResolutionDepth UInt8, FlashMajor UInt8, FlashMinor UInt8, NetMajor UInt8, NetMinor UInt8, MobilePhone UInt8, SilverlightVersion1 UInt8, Age UInt8, Sex UInt8, Income UInt8, JavaEnable UInt8, CookieEnable UInt8, JavascriptEnable UInt8, IsMobile UInt8, BrowserLanguage UInt16, BrowserCountry UInt16, Interests UInt16, Robotness UInt8, GeneralInterests Array(UInt16), Params Array(String), Goals Nested(ID UInt32, Serial UInt32, EventTime DateTime, Price Int64, OrderID String, CurrencyID UInt32), WatchIDs Array(UInt64), ParamSumPrice Int64, ParamCurrency FixedString(3), ParamCurrencyID UInt16, ClickLogID UInt64, ClickEventID Int32, ClickGoodEvent Int32, ClickEventTime DateTime, ClickPriorityID Int32, ClickPhraseID Int32, ClickPageID Int32, ClickPlaceID Int32, ClickTypeID Int32, ClickResourceID Int32, ClickCost UInt32, ClickClientIP UInt32, ClickDomainID UInt32, ClickURL String, ClickAttempt UInt8, ClickOrderID UInt32, ClickBannerID UInt32, ClickMarketCategoryID UInt32, ClickMarketPP UInt32, ClickMarketCategoryName String, ClickMarketPPName String, ClickAWAPSCampaignName String, ClickPageName String, ClickTargetType UInt16, ClickTargetPhraseID UInt64, ClickContextType UInt8, ClickSelectType Int8, ClickOptions String, ClickGroupBannerID Int32, OpenstatServiceName String, OpenstatCampaignID String, OpenstatAdID String, OpenstatSourceID String, UTMSource String, UTMMedium String, UTMCampaign String, UTMContent String, UTMTerm String, FromTag String, HasGCLID UInt8, FirstVisit DateTime, PredLastVisit Date, LastVisit Date, TotalVisits UInt32, TraficSource Nested(ID Int8, SearchEngineID UInt16, AdvEngineID UInt8, PlaceID UInt16, SocialSourceNetworkID UInt8, Domain String, SearchPhrase String, SocialSourcePage String), Attendance FixedString(16), CLID UInt32, YCLID UInt64, NormalizedRefererHash UInt64, SearchPhraseHash UInt64, RefererDomainHash UInt64, NormalizedStartURLHash UInt64, StartURLDomainHash UInt64, NormalizedEndURLHash UInt64, TopLevelDomain UInt64, URLScheme UInt64, OpenstatServiceNameHash UInt64, OpenstatCampaignIDHash UInt64, OpenstatAdIDHash UInt64, OpenstatSourceIDHash UInt64, UTMSourceHash UInt64, UTMMediumHash UInt64, UTMCampaignHash UInt64, UTMContentHash UInt64, UTMTermHash UInt64, FromHash UInt64, WebVisorEnabled UInt8, WebVisorActivity UInt32, ParsedParams Nested(Key1 String, Key2 String, Key3 String, Key4 String, Key5 String, ValueDouble Float64), Market Nested(Type UInt8, GoalID UInt32, OrderID String, OrderPrice Int64, PP UInt32, DirectPlaceID UInt32, DirectOrderID UInt32, DirectBannerID UInt32, GoodID String, GoodName String, GoodQuantity Int32, GoodPrice Int64), IslandID FixedString(16)) ENGINE = CollapsingMergeTree(Sign) PARTITION BY toYYYYMM(StartDate) ORDER BY (CounterID, StartDate, intHash32(UserID), VisitID) SAMPLE BY intHash32(UserID) SETTINGS index_granularity = 8192, storage_policy='\''s3_cache'\''' + clickhouse-client --query 'INSERT INTO test.hits_s3 SELECT * FROM datasets.hits_v1 SETTINGS enable_filesystem_cache_on_write_operations=0' + clickhouse-client --query 'INSERT INTO test.hits SELECT * FROM datasets.hits_v1 SETTINGS enable_filesystem_cache_on_write_operations=0' + clickhouse-client --query 'INSERT INTO test.visits SELECT * FROM datasets.visits_v1 SETTINGS enable_filesystem_cache_on_write_operations=0' + clickhouse-client --query 'DROP TABLE datasets.visits_v1 SYNC' + clickhouse-client --query 'DROP TABLE datasets.hits_v1 SYNC' + clickhouse-client --query 'SHOW TABLES FROM test' hits hits_s3 visits + clickhouse-client --query 'SYSTEM STOP THREAD FUZZER' + stop_server + local max_tries=90 + local check_hang=true + local pid ++ cat /var/run/clickhouse-server/clickhouse-server.pid + pid=1413 + clickhouse stop --max-tries 90 --do-not-kill /var/run/clickhouse-server/clickhouse-server.pid file exists and contains pid = 1413. The process with pid = 1413 is running. Sent terminate signal to process with pid 1413. Waiting for server to stop /var/run/clickhouse-server/clickhouse-server.pid file exists and contains pid = 1413. The process with pid = 1413 is running. Waiting for server to stop /var/run/clickhouse-server/clickhouse-server.pid file exists and contains pid = 1413. The process with pid = 1413 is running. Waiting for server to stop /var/run/clickhouse-server/clickhouse-server.pid file exists and contains pid = 1413. The process with pid = 1413 is running. Waiting for server to stop /var/run/clickhouse-server/clickhouse-server.pid file exists and contains pid = 1413. The process with pid = 1413 is running. Waiting for server to stop /var/run/clickhouse-server/clickhouse-server.pid file exists and contains pid = 1413. The process with pid = 1413 is running. Waiting for server to stop Now there is no clickhouse-server process. Server stopped + return + export USE_S3_STORAGE_FOR_MERGE_TREE=1 + USE_S3_STORAGE_FOR_MERGE_TREE=1 + export RANDOMIZE_OBJECT_KEY_TYPE=1 + RANDOMIZE_OBJECT_KEY_TYPE=1 + export ZOOKEEPER_FAULT_INJECTION=1 + ZOOKEEPER_FAULT_INJECTION=1 + export THREAD_POOL_FAULT_INJECTION=1 + THREAD_POOL_FAULT_INJECTION=1 + configure + export USE_DATABASE_ORDINARY=1 + USE_DATABASE_ORDINARY=1 + export EXPORT_S3_STORAGE_POLICIES=1 + EXPORT_S3_STORAGE_POLICIES=1 + /usr/share/clickhouse-test/config/install.sh + DEST_SERVER_PATH=/etc/clickhouse-server + DEST_CLIENT_PATH=/etc/clickhouse-client +++ dirname /usr/share/clickhouse-test/config/install.sh ++ cd /usr/share/clickhouse-test/config ++ pwd -P Going to install test configs from /usr/share/clickhouse-test/config into /etc/clickhouse-server + SRC_PATH=/usr/share/clickhouse-test/config + echo 'Going to install test configs from /usr/share/clickhouse-test/config into /etc/clickhouse-server' + mkdir -p /etc/clickhouse-server/config.d/ + mkdir -p /etc/clickhouse-server/users.d/ + mkdir -p /etc/clickhouse-client + ln -sf /usr/share/clickhouse-test/config/config.d/zookeeper_write.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/max_num_to_warn.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/listen.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/text_log.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/blob_storage_log.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/custom_settings_prefixes.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/enable_access_control_improvements.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/macros.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/secure_ports.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/clusters.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/graphite.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/graphite_alternative.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/database_atomic.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/max_concurrent_queries.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/merge_tree_settings.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/backoff_failed_mutation.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/merge_tree_old_dirs_cleanup.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/test_cluster_with_incorrect_pw.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/keeper_port.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/logging_no_rotate.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/merge_tree.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/lost_forever_check.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/tcp_with_proxy.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/prometheus.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/top_level_domains_lists.xml /etc/clickhouse-server/config.d/ script.gdb:13: Error in sourced command file: No stack. + ln -sf /usr/share/clickhouse-test/config/config.d/top_level_domains_path.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/transactions.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/encryption.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/CORS.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/zookeeper_log.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/logger_trace.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/named_collection.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/ssl_certs.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/filesystem_cache_log.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/session_log.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/system_unfreeze.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/enable_zero_copy_replication.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/nlp.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/forbidden_headers.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/enable_keeper_map.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/custom_disks_base_path.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/display_name.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/reverse_dns_query_function.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/compressed_marks_and_index.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/disable_s3_env_credentials.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/enable_wait_for_shutdown_replicated_tables.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/backups.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/filesystem_caches_path.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/validate_tcp_client_information.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/zero_copy_destructive_operations.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/block_number.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/handlers.yaml /etc/clickhouse-server/config.d/ + '[' /etc/clickhouse-server = /etc/clickhouse-server ']' + ln -sf /usr/share/clickhouse-test/config/config.d/legacy_geobase.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/users.d/log_queries.xml /etc/clickhouse-server/users.d/ + ln -sf /usr/share/clickhouse-test/config/users.d/readonly.xml /etc/clickhouse-server/users.d/ + ln -sf /usr/share/clickhouse-test/config/users.d/access_management.xml /etc/clickhouse-server/users.d/ + ln -sf /usr/share/clickhouse-test/config/users.d/database_atomic_drop_detach_sync.xml /etc/clickhouse-server/users.d/ + ln -sf /usr/share/clickhouse-test/config/users.d/opentelemetry.xml /etc/clickhouse-server/users.d/ + ln -sf /usr/share/clickhouse-test/config/users.d/remote_queries.xml /etc/clickhouse-server/users.d/ + ln -sf /usr/share/clickhouse-test/config/users.d/session_log_test.xml /etc/clickhouse-server/users.d/ + ln -sf /usr/share/clickhouse-test/config/users.d/memory_profiler.xml /etc/clickhouse-server/users.d/ + ln -sf /usr/share/clickhouse-test/config/users.d/no_fsync_metadata.xml /etc/clickhouse-server/users.d/ + ln -sf /usr/share/clickhouse-test/config/users.d/filelog.xml /etc/clickhouse-server/users.d/ + ln -sf /usr/share/clickhouse-test/config/users.d/enable_blobs_check.xml /etc/clickhouse-server/users.d/ + ln -sf /usr/share/clickhouse-test/config/users.d/marks.xml /etc/clickhouse-server/users.d/ + ln -sf /usr/share/clickhouse-test/config/users.d/insert_keeper_retries.xml /etc/clickhouse-server/users.d/ + ln -sf /usr/share/clickhouse-test/config/users.d/prefetch_settings.xml /etc/clickhouse-server/users.d/ + ln -sf /usr/share/clickhouse-test/config/users.d/nonconst_timezone.xml /etc/clickhouse-server/users.d/ + ln -sf /usr/share/clickhouse-test/config/users.d/allow_introspection_functions.yaml /etc/clickhouse-server/users.d/ + ln -sf /usr/share/clickhouse-test/config/users.d/replicated_ddl_entry.xml /etc/clickhouse-server/users.d/ + [[ -n '' ]] + ln -sf /usr/share/clickhouse-test/config/users.d/timeouts.xml /etc/clickhouse-server/users.d/ + ln -sf /usr/share/clickhouse-test/config/ints_dictionary.xml /etc/clickhouse-server/ + ln -sf /usr/share/clickhouse-test/config/strings_dictionary.xml /etc/clickhouse-server/ + ln -sf /usr/share/clickhouse-test/config/decimals_dictionary.xml /etc/clickhouse-server/ + ln -sf /usr/share/clickhouse-test/config/executable_dictionary.xml /etc/clickhouse-server/ + ln -sf /usr/share/clickhouse-test/config/executable_pool_dictionary.xml /etc/clickhouse-server/ + ln -sf /usr/share/clickhouse-test/config/test_function.xml /etc/clickhouse-server/ + ln -sf /usr/share/clickhouse-test/config/top_level_domains /etc/clickhouse-server/ + ln -sf /usr/share/clickhouse-test/config/regions_hierarchy.txt /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/regions_names_en.txt /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/ext-en.txt /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/ext-ru.txt /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/lem-en.bin /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/server.key /etc/clickhouse-server/ + ln -sf /usr/share/clickhouse-test/config/server.crt /etc/clickhouse-server/ + ln -sf /usr/share/clickhouse-test/config/dhparam.pem /etc/clickhouse-server/ + ln -sf --backup=simple --suffix=_original.xml /usr/share/clickhouse-test/config/config.d/query_masking_rules.xml /etc/clickhouse-server/config.d/ + [[ -n 1 ]] + [[ 1 -eq 1 ]] + rm -f /etc/clickhouse-server/config.d/zookeeper.xml + ln -sf /usr/share/clickhouse-test/config/config.d/zookeeper_fault_injection.xml /etc/clickhouse-server/config.d/ + [[ -n 1 ]] + [[ 1 -eq 1 ]] + ln -sf /usr/share/clickhouse-test/config/config.d/cannot_allocate_thread_injection.xml /etc/clickhouse-server/config.d/ + value=0 + sed --follow-symlinks -i 's|[01]|0|' /etc/clickhouse-server/config.d/keeper_port.xml + value=61054976 + sed --follow-symlinks -i 's|[[:digit:]]\+|61054976|' /etc/clickhouse-server/config.d/keeper_port.xml + value=16164864 + sed --follow-symlinks -i 's|[[:digit:]]\+|16164864|' /etc/clickhouse-server/config.d/keeper_port.xml + [[ -n '' ]] + [[ -n 1 ]] + [[ 1 -eq 1 ]] + ln -sf /usr/share/clickhouse-test/config/users.d/database_ordinary.xml /etc/clickhouse-server/users.d/ + [[ -n 1 ]] + [[ 1 -eq 1 ]] + object_key_types_options=("generate-suffix" "generate-full-key" "generate-template-key") + object_key_type=generate-suffix + [[ -n 1 ]] + [[ 1 -eq 1 ]] + object_key_type=generate-full-key + case $object_key_type in + ln -sf /usr/share/clickhouse-test/config/config.d/storage_metadata_with_full_object_key.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/s3_storage_policy_by_default.xml /etc/clickhouse-server/config.d/ + ARM=aarch64 ++ uname -m + OS=x86_64 x86_64 Adding azure configuration + [[ -n 1 ]] + echo x86_64 + [[ '' -eq 1 ]] + [[ x86_64 == \a\a\r\c\h\6\4 ]] + echo 'Adding azure configuration' + ln -sf /usr/share/clickhouse-test/config/config.d/azure_storage_conf.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/storage_conf.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/storage_conf_02944.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/storage_conf_02963.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/config.d/storage_conf_02961.xml /etc/clickhouse-server/config.d/ + ln -sf /usr/share/clickhouse-test/config/users.d/s3_cache.xml /etc/clickhouse-server/users.d/ + ln -sf /usr/share/clickhouse-test/config/users.d/s3_cache_new.xml /etc/clickhouse-server/users.d/ + [[ -n '' ]] + ln -sf /usr/share/clickhouse-test/config/client_config.xml /etc/clickhouse-client/config.xml + sudo cat /etc/clickhouse-server/config.d/keeper_port.xml + sed 's|100000|10000|' + sudo mv /etc/clickhouse-server/config.d/keeper_port.xml.tmp /etc/clickhouse-server/config.d/keeper_port.xml + [[ -n 1 ]] + [[ 1 -eq 1 ]] + randomize_config_boolean_value filtered_list keeper_port + value=0 + sudo cat /etc/clickhouse-server/config.d/keeper_port.xml + sed 's|[01]|0|' + sudo mv /etc/clickhouse-server/config.d/keeper_port.xml.tmp /etc/clickhouse-server/config.d/keeper_port.xml + randomize_config_boolean_value multi_read keeper_port + value=0 + sudo cat /etc/clickhouse-server/config.d/keeper_port.xml + sed 's|[01]|0|' + sudo mv /etc/clickhouse-server/config.d/keeper_port.xml.tmp /etc/clickhouse-server/config.d/keeper_port.xml + randomize_config_boolean_value check_not_exists keeper_port + value=1 + sed 's|[01]|1|' + sudo cat /etc/clickhouse-server/config.d/keeper_port.xml + sudo mv /etc/clickhouse-server/config.d/keeper_port.xml.tmp /etc/clickhouse-server/config.d/keeper_port.xml + randomize_config_boolean_value create_if_not_exists keeper_port + value=1 + sudo cat /etc/clickhouse-server/config.d/keeper_port.xml + sed 's|[01]|1|' + sudo mv /etc/clickhouse-server/config.d/keeper_port.xml.tmp /etc/clickhouse-server/config.d/keeper_port.xml + sudo chown clickhouse /etc/clickhouse-server/config.d/keeper_port.xml + sudo chgrp clickhouse /etc/clickhouse-server/config.d/keeper_port.xml + [[ -n 1 ]] + [[ 1 -eq 1 ]] + randomize_config_boolean_value use_compression zookeeper_fault_injection + value=1 + sudo cat /etc/clickhouse-server/config.d/zookeeper_fault_injection.xml + sed 's|[01]|1|' + sudo mv /etc/clickhouse-server/config.d/zookeeper_fault_injection.xml.tmp /etc/clickhouse-server/config.d/zookeeper_fault_injection.xml + randomize_config_boolean_value allow_experimental_block_number_column block_number + value=1 + sudo cat /etc/clickhouse-server/config.d/block_number.xml + sed 's|[01]|1|' + sudo mv /etc/clickhouse-server/config.d/block_number.xml.tmp /etc/clickhouse-server/config.d/block_number.xml + echo 'ASAN_OPTIONS='\''malloc_context_size=10 verbosity=1 allocator_release_to_os_interval_ms=10000'\''' + export 'ASAN_OPTIONS=malloc_context_size=10 allocator_release_to_os_interval_ms=10000' + ASAN_OPTIONS='malloc_context_size=10 allocator_release_to_os_interval_ms=10000' + sudo chown root: /var/lib/clickhouse + echo '1' + local total_mem ++ awk '/MemTotal/ { print $(NF-1) }' /proc/meminfo + total_mem=32086428 + total_mem=32856502272 + max_server_memory_usage_to_ram_ratio=0.5 + echo 'Setting max_server_memory_usage_to_ram_ratio to 0.5' + cat Setting max_server_memory_usage_to_ram_ratio to 0.5 Setting max_memory_usage_for_user=9856950681 and max_memory_usage for queries to 10G + local max_users_mem + max_users_mem=9856950681 + echo 'Setting max_memory_usage_for_user=9856950681 and max_memory_usage for queries to 10G' + cat + cat + sudo cat /etc/clickhouse-server/config.d/s3_storage_policy_by_default.xml + sed 's|
s3
|
s3
default|' + mv /etc/clickhouse-server/config.d/s3_storage_policy_by_default.xml.tmp /etc/clickhouse-server/config.d/s3_storage_policy_by_default.xml + sudo chown clickhouse /etc/clickhouse-server/config.d/s3_storage_policy_by_default.xml + sudo chgrp clickhouse /etc/clickhouse-server/config.d/s3_storage_policy_by_default.xml + sudo cat /etc/clickhouse-server/config.d/logger_trace.xml + sed 's|trace|test|' + mv /etc/clickhouse-server/config.d/logger_trace.xml.tmp /etc/clickhouse-server/config.d/logger_trace.xml + '[' SLRU = SLRU ']' + sudo cat /etc/clickhouse-server/config.d/storage_conf.xml + sed 's|LRU|SLRU|' + mv /etc/clickhouse-server/config.d/storage_conf.xml.tmp /etc/clickhouse-server/config.d/storage_conf.xml ++ date +%-d + '[' 0 -eq 1 ']' + start_server + counter=0 + max_attempt=120 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 0 -gt 120 ']' + clickhouse start --user root + sleep 0.5 127.0.0.1 - - [16/Dec/2024:08:55:28 +0000] "PUT /devstoreaccount1/cont?restype=container HTTP/1.1" 409 - 127.0.0.1 - - [16/Dec/2024:08:55:28 +0000] "PUT /devstoreaccount1/cont/iphcsbsfmiywhjazhdsickinbrcmpgmw?blockid=lieeoyqhklqiylnglyouidhyobephgvkrendcuhsqnbqpbnltyhknurmtmdbnukn&comp=block HTTP/1.1" 201 - 127.0.0.1 - - [16/Dec/2024:08:55:28 +0000] "PUT /devstoreaccount1/cont/iphcsbsfmiywhjazhdsickinbrcmpgmw?comp=blocklist HTTP/1.1" 201 - 127.0.0.1 - - [16/Dec/2024:08:55:28 +0000] "GET /devstoreaccount1/cont/iphcsbsfmiywhjazhdsickinbrcmpgmw HTTP/1.1" 206 4 127.0.0.1 - - [16/Dec/2024:08:55:28 +0000] "GET /devstoreaccount1/cont/iphcsbsfmiywhjazhdsickinbrcmpgmw HTTP/1.1" 206 2 127.0.0.1 - - [16/Dec/2024:08:55:28 +0000] "DELETE /devstoreaccount1/cont/iphcsbsfmiywhjazhdsickinbrcmpgmw HTTP/1.1" 202 - + counter=1 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 1 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=2 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 2 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=3 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 3 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=4 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 4 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=5 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 5 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=6 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 6 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=7 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 7 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=8 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 8 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=9 + clickhouse-client --query 'SELECT 1' 1 + attach_gdb_to_clickhouse ++ kill -l SIGRTMIN + RTMIN=34 + echo ' set follow-fork-mode parent handle SIGHUP nostop noprint pass handle SIGINT nostop noprint pass handle SIGQUIT nostop noprint pass handle SIGPIPE nostop noprint pass handle SIGTERM nostop noprint pass handle SIGUSR1 nostop noprint pass handle SIGUSR2 nostop noprint pass handle SIG34 nostop noprint pass info signals continue backtrace full thread apply all backtrace full info registers disassemble /s up disassemble /s up disassemble /s p "done" detach quit ' + sleep 5 + ts '%Y-%m-%d %H:%M:%S' ++ cat /var/run/clickhouse-server/clickhouse-server.pid + gdb -batch -command script.gdb -p 2778 + run_with_retry 60 clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' + [[ hxB =~ e ]] + set_e=false + set +e + local total_retries=60 + shift + local retry=0 + '[' 0 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Connected to clickhouse-server after attaching gdb + false + return + stress --hung-check --drop-databases --output-folder test_output --skip-func-tests '' --global-time-limit 1200 2024-12-16 09:56:23,644 Run func tests '/usr/bin/clickhouse-test --global_time_limit=1200 ' 2024-12-16 09:56:24,146 Run func tests '/usr/bin/clickhouse-test --order=random --database=test_1 --client-option join_use_nulls=1 join_algorithm='parallel_hash' memory_tracker_fault_probability=0.001 merge_tree_read_split_ranges_into_intersecting_and_non_intersecting_injection_probability=0.05 group_by_use_nulls=1 --global_time_limit=1200 ' 2024-12-16 09:56:24,647 Run func tests '/usr/bin/clickhouse-test --order=random --db-engine="Replicated('/test/db/test_2', 's1', 'r1')" --client-option allow_experimental_database_replicated=1 enable_deflate_qpl_codec=1 enable_zstd_qat_codec=1 use_query_cache=1 http_make_head_request=1 --global_time_limit=1200 ' 2024-12-16 09:56:25,148 Run func tests '/usr/bin/clickhouse-test --order=random --database=test_3 --client-option join_algorithm='partial_merge' group_by_use_nulls=1 --global_time_limit=1200 ' 2024-12-16 09:56:25,650 Run func tests '/usr/bin/clickhouse-test --order=random --client-option join_use_nulls=1 --global_time_limit=1200 ' 2024-12-16 09:56:26,151 Run func tests '/usr/bin/clickhouse-test --order=random --db-engine="Replicated('/test/db/test_5', 's1', 'r1')" --database=test_5 --client-option allow_experimental_database_replicated=1 enable_deflate_qpl_codec=1 enable_zstd_qat_codec=1 join_algorithm='full_sorting_merge' use_query_cache=1 group_by_use_nulls=1 http_make_head_request=1 --global_time_limit=1200 ' 2024-12-16 09:56:26,652 Run func tests '/usr/bin/clickhouse-test --order=random --client-option use_query_cache=1 memory_tracker_fault_probability=0.001 merge_tree_read_split_ranges_into_intersecting_and_non_intersecting_injection_probability=0.05 --global_time_limit=1200 ' 2024-12-16 09:56:27,154 Run func tests '/usr/bin/clickhouse-test --order=random --database=test_7 --client-option join_use_nulls=1 join_algorithm='grace_hash' group_by_use_nulls=1 http_make_head_request=1 --global_time_limit=1200 ' 2024-12-16 09:56:27,655 Run func tests '/usr/bin/clickhouse-test --order=random --db-engine="Replicated('/test/db/test_8', 's1', 'r1')" --client-option allow_experimental_database_replicated=1 enable_deflate_qpl_codec=1 enable_zstd_qat_codec=1 --global_time_limit=1200 ' 2024-12-16 09:56:28,156 Run func tests '/usr/bin/clickhouse-test --order=random --database=test_9 --client-option join_algorithm='auto' max_rows_in_join=1000 group_by_use_nulls=1 --global_time_limit=1200 ' 2024-12-16 09:56:28,657 Run func tests '/usr/bin/clickhouse-test --order=random --client-option join_use_nulls=1 use_query_cache=1 --global_time_limit=1200 ' 2024-12-16 09:56:29,158 Run func tests '/usr/bin/clickhouse-test --order=random --db-engine="Replicated('/test/db/test_11', 's1', 'r1')" --database=test_11 --client-option allow_experimental_database_replicated=1 enable_deflate_qpl_codec=1 enable_zstd_qat_codec=1 join_algorithm='parallel_hash' memory_tracker_fault_probability=0.001 merge_tree_read_split_ranges_into_intersecting_and_non_intersecting_injection_probability=0.05 group_by_use_nulls=1 http_make_head_request=1 --global_time_limit=1200 ' 2024-12-16 09:56:29,660 Run func tests '/usr/bin/clickhouse-test --order=random --client-option implicit_transaction=1 throw_on_unsupported_query_inside_transaction=0 --global_time_limit=1200 ' 2024-12-16 09:56:30,161 Run func tests '/usr/bin/clickhouse-test --order=random --database=test_13 --client-option join_use_nulls=1 join_algorithm='partial_merge' use_query_cache=1 group_by_use_nulls=1 --global_time_limit=1200 ' 2024-12-16 09:56:30,662 Run func tests '/usr/bin/clickhouse-test --order=random --db-engine="Replicated('/test/db/test_14', 's1', 'r1')" --client-option allow_experimental_database_replicated=1 enable_deflate_qpl_codec=1 enable_zstd_qat_codec=1 use_query_cache=1 --global_time_limit=1200 ' 2024-12-16 09:56:31,163 Run func tests '/usr/bin/clickhouse-test --order=random --database=test_15 --client-option join_algorithm='full_sorting_merge' group_by_use_nulls=1 optimize_trivial_approximate_count_query=1 http_make_head_request=1 --global_time_limit=1200 ' 2024-12-16 09:56:31,664 Will wait functests to finish 2024-12-16 09:56:31,665 Finished 6 from 16 processes 2024-12-16 09:56:36,670 Finished 10 from 16 processes 2024-12-16 09:56:41,674 Finished 10 from 16 processes 2024-12-16 09:56:46,675 Finished 10 from 16 processes 2024-12-16 09:56:51,679 Finished 10 from 16 processes 2024-12-16 09:56:56,685 Finished 10 from 16 processes 2024-12-16 09:57:01,690 Finished 10 from 16 processes 2024-12-16 09:57:06,693 Finished 10 from 16 processes 2024-12-16 09:57:11,698 Finished 10 from 16 processes 2024-12-16 09:57:16,703 Finished 10 from 16 processes 2024-12-16 09:57:21,707 Finished 10 from 16 processes 2024-12-16 09:57:26,711 Finished 10 from 16 processes 2024-12-16 09:57:31,713 Finished 10 from 16 processes 2024-12-16 09:57:36,715 Finished 10 from 16 processes 2024-12-16 09:57:41,721 Finished 10 from 16 processes 2024-12-16 09:57:46,726 Finished 10 from 16 processes 2024-12-16 09:57:51,727 Finished 10 from 16 processes 2024-12-16 09:57:56,731 Finished 10 from 16 processes 2024-12-16 09:58:01,733 Finished 10 from 16 processes 2024-12-16 09:58:06,733 Finished 10 from 16 processes 2024-12-16 09:58:11,734 Finished 10 from 16 processes 2024-12-16 09:58:16,735 Finished 10 from 16 processes 2024-12-16 09:58:21,741 Finished 10 from 16 processes 2024-12-16 09:58:26,743 Finished 10 from 16 processes 2024-12-16 09:58:31,747 Finished 10 from 16 processes 2024-12-16 09:58:36,753 Finished 10 from 16 processes 2024-12-16 09:58:41,756 Finished 10 from 16 processes 2024-12-16 09:58:46,759 Finished 10 from 16 processes 2024-12-16 09:58:51,765 Finished 10 from 16 processes 2024-12-16 09:58:56,770 Finished 10 from 16 processes 2024-12-16 09:59:01,771 Finished 10 from 16 processes 2024-12-16 09:59:06,775 Finished 10 from 16 processes 2024-12-16 09:59:11,779 Finished 10 from 16 processes 2024-12-16 09:59:16,783 Finished 10 from 16 processes 2024-12-16 09:59:21,787 Finished 10 from 16 processes 2024-12-16 09:59:26,792 Finished 10 from 16 processes 2024-12-16 09:59:31,795 Finished 10 from 16 processes 2024-12-16 09:59:36,801 Finished 10 from 16 processes 2024-12-16 09:59:41,806 Finished 10 from 16 processes 2024-12-16 09:59:46,811 Finished 10 from 16 processes 2024-12-16 09:59:51,815 Finished 10 from 16 processes 2024-12-16 09:59:56,819 Finished 10 from 16 processes 2024-12-16 10:00:01,823 Finished 10 from 16 processes 2024-12-16 10:00:06,827 Finished 10 from 16 processes 2024-12-16 10:00:11,831 Finished 10 from 16 processes 2024-12-16 10:00:16,835 Finished 10 from 16 processes 2024-12-16 10:00:21,836 Finished 10 from 16 processes 2024-12-16 10:00:26,839 Finished 10 from 16 processes 2024-12-16 10:00:31,843 Finished 10 from 16 processes 2024-12-16 10:00:36,847 Finished 10 from 16 processes 2024-12-16 10:00:41,851 Finished 10 from 16 processes 2024-12-16 10:00:46,853 Finished 10 from 16 processes 2024-12-16 10:00:51,855 Finished 10 from 16 processes 2024-12-16 10:00:56,859 Finished 10 from 16 processes 2024-12-16 10:01:01,863 Finished 10 from 16 processes 2024-12-16 10:01:06,867 Finished 10 from 16 processes 2024-12-16 10:01:11,871 Finished 10 from 16 processes 2024-12-16 10:01:16,875 Finished 10 from 16 processes 2024-12-16 10:01:21,879 Finished 10 from 16 processes 2024-12-16 10:01:26,881 Finished 10 from 16 processes 2024-12-16 10:01:31,886 Finished 10 from 16 processes 2024-12-16 10:01:36,891 Finished 10 from 16 processes 2024-12-16 10:01:41,893 Finished 10 from 16 processes 2024-12-16 10:01:46,898 Finished 10 from 16 processes 2024-12-16 10:01:51,899 Finished 10 from 16 processes 2024-12-16 10:01:56,900 Finished 10 from 16 processes API: PutObjectPart(bucket=test, object=s3/nyz/wwhxoufenstyrqsvhwaqqfyiopzal) Time: 09:01:59 UTC 12/16/2024 DeploymentID: c978493f-9314-4977-ae9d-01cf681f6eb5 RequestID: 18119CC93EF98D5B RemoteHost: ::1 Host: localhost:11111 UserAgent: aws-sdk-cpp/1.11.234 ua/2.0 md/aws-crt#0.24.11-dev+cddd1639 os/Linux/5.15.0-125-generic md/arch#x86_64 lang/c++#C++23 md/Clang#17.0.6 cfg/retry-mode#custom api/S3 Error: open /minio_data/.minio.sys/multipart/c2aa4bafc6769590ccefa38e3db3ea5f9b5acf937e22d638d180d318655b2b4b/7b0a65b2-9eaf-4d4d-a8d9-ed2d132933e5/00002.d9f15be772929726021274af7b4f6b80.16777216: no such file or directory (*fs.PathError) uploadID=7b0a65b2-9eaf-4d4d-a8d9-ed2d132933e5, partPath=/minio_data/.minio.sys/multipart/c2aa4bafc6769590ccefa38e3db3ea5f9b5acf937e22d638d180d318655b2b4b/7b0a65b2-9eaf-4d4d-a8d9-ed2d132933e5/00002.d9f15be772929726021274af7b4f6b80.16777216, filepath=/minio_data/.minio.sys/tmp/aec027a9-59b4-4a69-9e58-05cc47c90dc9/7b0a65b2-9eaf-4d4d-a8d9-ed2d132933e5.b2125ea7-ba44-4405-a89d-a69eabbc6986 1: cmd/fs-v1-multipart.go:121:cmd.(*FSObjects).backgroundAppend() 2024-12-16 10:02:01,901 Finished 10 from 16 processes 2024-12-16 10:02:06,903 Finished 10 from 16 processes 2024-12-16 10:02:11,907 Finished 10 from 16 processes 2024-12-16 10:02:16,913 Finished 10 from 16 processes 2024-12-16 10:02:21,915 Finished 10 from 16 processes 2024-12-16 10:02:26,920 Finished 10 from 16 processes 2024-12-16 10:02:31,923 Finished 10 from 16 processes 2024-12-16 10:02:36,927 Finished 10 from 16 processes 2024-12-16 10:02:41,932 Finished 10 from 16 processes 2024-12-16 10:02:46,935 Finished 10 from 16 processes 2024-12-16 10:02:51,941 Finished 10 from 16 processes 2024-12-16 10:02:56,943 Finished 10 from 16 processes 2024-12-16 10:03:01,946 Finished 10 from 16 processes 2024-12-16 10:03:06,951 Finished 10 from 16 processes 2024-12-16 10:03:11,957 Finished 10 from 16 processes 2024-12-16 10:03:16,959 Finished 10 from 16 processes 2024-12-16 10:03:21,963 Finished 10 from 16 processes 2024-12-16 10:03:26,967 Finished 10 from 16 processes 2024-12-16 10:03:31,972 Finished 10 from 16 processes 2024-12-16 10:03:36,977 Finished 10 from 16 processes 2024-12-16 10:03:41,979 Finished 10 from 16 processes 2024-12-16 10:03:46,983 Finished 10 from 16 processes 2024-12-16 10:03:51,987 Finished 10 from 16 processes 2024-12-16 10:03:56,992 Finished 11 from 16 processes 2024-12-16 10:04:01,995 Finished 11 from 16 processes 2024-12-16 10:04:06,999 Finished 11 from 16 processes 2024-12-16 10:04:12,003 Finished 11 from 16 processes 2024-12-16 10:04:17,009 Finished 11 from 16 processes 2024-12-16 10:04:22,011 Finished 11 from 16 processes 2024-12-16 10:04:27,017 Finished 11 from 16 processes 2024-12-16 10:04:32,022 Finished 11 from 16 processes 2024-12-16 10:04:37,023 Finished 12 from 16 processes 2024-12-16 10:04:42,027 Finished 12 from 16 processes 2024-12-16 10:04:47,031 Finished 12 from 16 processes 2024-12-16 10:04:52,037 Finished 12 from 16 processes 2024-12-16 10:04:57,039 Finished 12 from 16 processes 2024-12-16 10:05:02,043 Finished 12 from 16 processes 2024-12-16 10:05:07,047 Finished 12 from 16 processes 2024-12-16 10:05:12,053 Finished 12 from 16 processes 2024-12-16 10:05:17,055 Finished 12 from 16 processes 2024-12-16 10:05:22,059 Finished 12 from 16 processes 2024-12-16 10:05:27,063 Finished 12 from 16 processes 2024-12-16 10:05:32,066 Finished 12 from 16 processes 2024-12-16 10:05:37,070 Finished 12 from 16 processes 2024-12-16 10:05:42,071 Finished 12 from 16 processes 2024-12-16 10:05:47,077 Finished 12 from 16 processes 2024-12-16 10:05:52,079 Finished 12 from 16 processes 2024-12-16 10:05:57,085 Finished 12 from 16 processes 2024-12-16 10:06:02,090 Finished 12 from 16 processes 2024-12-16 10:06:07,091 Finished 12 from 16 processes 2024-12-16 10:06:12,093 Finished 12 from 16 processes 2024-12-16 10:06:17,098 Finished 12 from 16 processes 2024-12-16 10:06:22,099 Finished 12 from 16 processes 2024-12-16 10:06:27,103 Finished 12 from 16 processes 2024-12-16 10:06:32,107 Finished 12 from 16 processes 2024-12-16 10:06:37,113 Finished 12 from 16 processes 2024-12-16 10:06:42,118 Finished 12 from 16 processes 2024-12-16 10:06:47,119 Finished 12 from 16 processes 2024-12-16 10:06:52,123 Finished 12 from 16 processes 2024-12-16 10:06:57,127 Finished 12 from 16 processes 2024-12-16 10:07:02,133 Finished 12 from 16 processes 2024-12-16 10:07:07,138 Finished 12 from 16 processes 2024-12-16 10:07:12,139 Finished 12 from 16 processes 2024-12-16 10:07:17,143 Finished 12 from 16 processes 2024-12-16 10:07:22,147 Finished 12 from 16 processes 2024-12-16 10:07:27,151 Finished 12 from 16 processes 2024-12-16 10:07:32,156 Finished 12 from 16 processes 2024-12-16 10:07:37,159 Finished 12 from 16 processes 2024-12-16 10:07:42,163 Finished 12 from 16 processes 2024-12-16 10:07:47,169 Finished 12 from 16 processes 2024-12-16 10:07:52,170 Finished 12 from 16 processes 2024-12-16 10:07:57,172 Finished 12 from 16 processes 2024-12-16 10:08:02,177 Finished 12 from 16 processes 2024-12-16 10:08:07,178 Finished 12 from 16 processes 2024-12-16 10:08:12,179 Finished 12 from 16 processes 2024-12-16 10:08:17,185 Finished 12 from 16 processes 2024-12-16 10:08:22,187 Finished 12 from 16 processes 2024-12-16 10:08:27,191 Finished 12 from 16 processes 2024-12-16 10:08:32,197 Finished 12 from 16 processes 2024-12-16 10:08:37,201 Finished 12 from 16 processes 2024-12-16 10:08:42,206 Finished 12 from 16 processes 2024-12-16 10:08:47,207 Finished 12 from 16 processes 2024-12-16 10:08:52,211 Finished 12 from 16 processes 2024-12-16 10:08:57,217 Finished 12 from 16 processes 2024-12-16 10:09:02,222 Finished 12 from 16 processes 2024-12-16 10:09:07,223 Finished 12 from 16 processes 2024-12-16 10:09:12,227 Finished 12 from 16 processes 2024-12-16 10:09:17,229 Finished 12 from 16 processes 2024-12-16 10:09:22,233 Finished 12 from 16 processes 2024-12-16 10:09:27,235 Finished 12 from 16 processes 2024-12-16 10:09:32,236 Finished 12 from 16 processes 2024-12-16 10:09:37,239 Finished 12 from 16 processes 2024-12-16 10:09:42,243 Finished 12 from 16 processes 2024-12-16 10:09:47,247 Finished 12 from 16 processes 2024-12-16 10:09:52,251 Finished 12 from 16 processes 2024-12-16 10:09:57,256 Finished 12 from 16 processes 2024-12-16 10:10:02,259 Finished 12 from 16 processes 2024-12-16 10:10:07,263 Finished 12 from 16 processes 2024-12-16 10:10:12,267 Finished 12 from 16 processes 2024-12-16 10:10:17,271 Finished 12 from 16 processes 2024-12-16 10:10:22,275 Finished 12 from 16 processes 2024-12-16 10:10:27,279 Finished 12 from 16 processes 2024-12-16 10:10:32,283 Finished 12 from 16 processes 2024-12-16 10:10:37,289 Finished 12 from 16 processes 2024-12-16 10:10:42,291 Finished 12 from 16 processes 2024-12-16 10:10:47,295 Finished 12 from 16 processes 2024-12-16 10:10:52,299 Finished 12 from 16 processes 2024-12-16 10:10:57,303 Finished 12 from 16 processes 2024-12-16 10:11:02,307 Finished 12 from 16 processes 2024-12-16 10:11:07,313 Finished 12 from 16 processes 2024-12-16 10:11:12,315 Finished 12 from 16 processes 2024-12-16 10:11:17,319 Finished 12 from 16 processes 2024-12-16 10:11:22,325 Finished 12 from 16 processes 2024-12-16 10:11:27,327 Finished 12 from 16 processes 2024-12-16 10:11:32,333 Finished 12 from 16 processes 2024-12-16 10:11:37,335 Finished 12 from 16 processes 2024-12-16 10:11:42,339 Finished 12 from 16 processes 2024-12-16 10:11:47,345 Finished 12 from 16 processes 2024-12-16 10:11:52,346 Finished 12 from 16 processes 2024-12-16 10:11:57,348 Finished 12 from 16 processes 2024-12-16 10:12:02,352 Finished 12 from 16 processes 2024-12-16 10:12:07,355 Finished 12 from 16 processes 2024-12-16 10:12:12,359 Finished 12 from 16 processes 2024-12-16 10:12:17,365 Finished 12 from 16 processes 2024-12-16 10:12:22,370 Finished 12 from 16 processes 2024-12-16 10:12:27,374 Finished 12 from 16 processes 2024-12-16 10:12:32,375 Finished 12 from 16 processes 2024-12-16 10:12:37,379 Finished 12 from 16 processes 2024-12-16 10:12:42,383 Finished 12 from 16 processes 2024-12-16 10:12:47,389 Finished 12 from 16 processes 2024-12-16 10:12:52,391 Finished 12 from 16 processes 2024-12-16 10:12:57,397 Finished 12 from 16 processes 2024-12-16 10:13:02,399 Finished 12 from 16 processes 2024-12-16 10:13:07,403 Finished 12 from 16 processes 2024-12-16 10:13:12,406 Finished 12 from 16 processes 2024-12-16 10:13:17,411 Finished 12 from 16 processes 2024-12-16 10:13:22,411 Finished 12 from 16 processes 2024-12-16 10:13:27,415 Finished 12 from 16 processes 2024-12-16 10:13:32,419 Finished 12 from 16 processes 2024-12-16 10:13:37,423 Finished 12 from 16 processes 2024-12-16 10:13:42,429 Finished 12 from 16 processes 2024-12-16 10:13:47,434 Finished 12 from 16 processes 2024-12-16 10:13:52,435 Finished 12 from 16 processes 2024-12-16 10:13:57,439 Finished 12 from 16 processes 2024-12-16 10:14:02,443 Finished 12 from 16 processes 2024-12-16 10:14:07,448 Finished 12 from 16 processes 2024-12-16 10:14:12,451 Finished 12 from 16 processes 2024-12-16 10:14:17,455 Finished 12 from 16 processes 2024-12-16 10:14:22,461 Finished 12 from 16 processes 2024-12-16 10:14:27,463 Finished 12 from 16 processes 2024-12-16 10:14:32,469 Finished 12 from 16 processes 2024-12-16 10:14:37,471 Finished 12 from 16 processes 2024-12-16 10:14:42,474 Finished 12 from 16 processes 2024-12-16 10:14:47,479 Finished 12 from 16 processes 2024-12-16 10:14:52,483 Finished 12 from 16 processes 2024-12-16 10:14:57,487 Finished 12 from 16 processes 2024-12-16 10:15:02,491 Finished 12 from 16 processes 2024-12-16 10:15:07,495 Finished 12 from 16 processes 2024-12-16 10:15:12,501 Finished 12 from 16 processes 2024-12-16 10:15:17,503 Finished 12 from 16 processes 2024-12-16 10:15:22,507 Finished 12 from 16 processes 2024-12-16 10:15:27,511 Finished 12 from 16 processes 2024-12-16 10:15:32,516 Finished 12 from 16 processes 2024-12-16 10:15:37,519 Finished 12 from 16 processes 2024-12-16 10:15:42,525 Finished 12 from 16 processes API: PutObjectPart(bucket=test, object=s3/kwq/ndnoodgkffmnhrprhvuuivsmsvkab) Time: 09:15:43 UTC 12/16/2024 DeploymentID: c978493f-9314-4977-ae9d-01cf681f6eb5 RequestID: 18119D88FBCA3FB1 RemoteHost: 127.0.0.1 Host: localhost:11111 UserAgent: aws-sdk-cpp/1.11.234 ua/2.0 md/aws-crt#0.24.11-dev+cddd1639 os/Linux/5.15.0-125-generic md/arch#x86_64 lang/c++#C++23 md/Clang#17.0.6 cfg/retry-mode#custom api/S3 Error: open /minio_data/.minio.sys/multipart/61084e9fa33e69b1e4b685e8b85d1d5ed761fc7156a3624616e426700bee72c6/e725e2e7-3c21-4461-b14f-2235f579a345/00002.6a18e0dd5d08bd9ee533a0f0ada111ca.10513603: no such file or directory (*fs.PathError) partPath=/minio_data/.minio.sys/multipart/61084e9fa33e69b1e4b685e8b85d1d5ed761fc7156a3624616e426700bee72c6/e725e2e7-3c21-4461-b14f-2235f579a345/00002.6a18e0dd5d08bd9ee533a0f0ada111ca.10513603, filepath=/minio_data/.minio.sys/tmp/aec027a9-59b4-4a69-9e58-05cc47c90dc9/e725e2e7-3c21-4461-b14f-2235f579a345.32fe14ff-7be7-46d6-b2cf-f84790da9a81, uploadID=e725e2e7-3c21-4461-b14f-2235f579a345 1: cmd/fs-v1-multipart.go:121:cmd.(*FSObjects).backgroundAppend() 2024-12-16 10:15:47,527 Finished 12 from 16 processes 2024-12-16 10:15:52,531 Finished 12 from 16 processes 2024-12-16 10:15:57,534 Finished 12 from 16 processes 2024-12-16 10:16:02,535 Finished 12 from 16 processes 2024-12-16 10:16:07,539 Finished 12 from 16 processes 2024-12-16 10:16:12,543 Finished 12 from 16 processes 2024-12-16 10:16:17,548 Finished 12 from 16 processes 2024-12-16 10:16:22,551 Finished 12 from 16 processes 2024-12-16 10:16:27,555 Finished 12 from 16 processes 2024-12-16 10:16:32,560 Finished 12 from 16 processes 2024-12-16 10:16:37,564 Finished 12 from 16 processes 2024-12-16 10:16:42,569 Finished 12 from 16 processes 2024-12-16 10:16:47,572 Finished 12 from 16 processes 2024-12-16 10:16:52,577 Finished 12 from 16 processes 2024-12-16 10:16:57,579 Finished 13 from 16 processes 2024-12-16 10:17:02,585 Finished 13 from 16 processes 2024-12-16 10:17:07,590 Finished 13 from 16 processes 2024-12-16 10:17:12,591 Finished 13 from 16 processes 2024-12-16 10:17:17,593 Finished 13 from 16 processes 2024-12-16 10:17:22,595 Finished 13 from 16 processes 2024-12-16 10:17:27,599 Finished 13 from 16 processes 2024-12-16 10:17:32,603 Finished 13 from 16 processes 2024-12-16 10:17:37,607 Finished 14 from 16 processes 2024-12-16 10:17:42,611 Finished 14 from 16 processes API: PutObjectPart(bucket=test, object=s3/xib/xamiwcyybhvekapnuikcvjfgstfpv) Time: 09:17:46 UTC 12/16/2024 DeploymentID: c978493f-9314-4977-ae9d-01cf681f6eb5 RequestID: 18119DA5C039FB5C RemoteHost: ::1 Host: localhost:11111 UserAgent: aws-sdk-cpp/1.11.234 ua/2.0 md/aws-crt#0.24.11-dev+cddd1639 os/Linux/5.15.0-125-generic md/arch#x86_64 lang/c++#C++23 md/Clang#17.0.6 cfg/retry-mode#custom api/S3 Error: open /minio_data/.minio.sys/multipart/f7fc31108a20400ff673ba3464cba4bd28e73330ef1754b590fe0ccc5cd61c40/02902513-edb5-4d5b-b82c-8f948692aabe/00002.98e45c79f70897931f79611e0ab1fe23.16032449: no such file or directory (*fs.PathError) uploadID=02902513-edb5-4d5b-b82c-8f948692aabe, partPath=/minio_data/.minio.sys/multipart/f7fc31108a20400ff673ba3464cba4bd28e73330ef1754b590fe0ccc5cd61c40/02902513-edb5-4d5b-b82c-8f948692aabe/00002.98e45c79f70897931f79611e0ab1fe23.16032449, filepath=/minio_data/.minio.sys/tmp/aec027a9-59b4-4a69-9e58-05cc47c90dc9/02902513-edb5-4d5b-b82c-8f948692aabe.865e021a-9234-41c1-b316-22540f54fce4 1: cmd/fs-v1-multipart.go:121:cmd.(*FSObjects).backgroundAppend() 2024-12-16 10:17:47,614 Finished 14 from 16 processes 2024-12-16 10:17:52,617 Finished 14 from 16 processes 2024-12-16 10:17:57,619 Finished 14 from 16 processes 2024-12-16 10:18:02,625 Finished 14 from 16 processes 2024-12-16 10:18:07,627 Finished 15 from 16 processes 2024-12-16 10:18:12,633 Finished 15 from 16 processes 2024-12-16 10:18:17,638 Finished 15 from 16 processes 2024-12-16 10:18:22,639 Finished 15 from 16 processes 2024-12-16 10:18:27,643 Finished 15 from 16 processes 2024-12-16 10:18:32,647 Finished 15 from 16 processes 2024-12-16 10:18:37,651 Finished 15 from 16 processes 2024-12-16 10:18:42,655 Finished 15 from 16 processes 2024-12-16 10:18:47,659 Finished 15 from 16 processes 2024-12-16 10:18:52,663 Finished 15 from 16 processes 2024-12-16 10:18:57,667 Finished 15 from 16 processes 2024-12-16 10:19:02,671 Finished 15 from 16 processes 2024-12-16 10:19:07,675 Finished 15 from 16 processes 2024-12-16 10:19:12,679 Finished 15 from 16 processes 2024-12-16 10:19:17,685 Finished 15 from 16 processes 2024-12-16 10:19:22,690 Finished 15 from 16 processes 2024-12-16 10:19:27,695 Finished 15 from 16 processes 2024-12-16 10:19:32,699 Finished 15 from 16 processes 2024-12-16 10:19:37,703 Finished 15 from 16 processes 2024-12-16 10:19:42,707 Finished 15 from 16 processes 2024-12-16 10:19:47,713 Finished 15 from 16 processes 2024-12-16 10:19:52,713 Finished 15 from 16 processes 2024-12-16 10:19:57,715 Finished 15 from 16 processes 2024-12-16 10:20:02,716 Finished 15 from 16 processes 2024-12-16 10:20:07,720 Finished 15 from 16 processes 2024-12-16 10:20:12,725 Finished 15 from 16 processes 2024-12-16 10:20:17,730 Finished 15 from 16 processes 2024-12-16 10:20:22,734 Finished 15 from 16 processes 2024-12-16 10:20:27,737 Finished 15 from 16 processes API: PutObjectPart(bucket=test, object=s3/iwo/ulpcttpeqreyucbyemjfnqwkruark) Time: 09:20:30 UTC 12/16/2024 DeploymentID: c978493f-9314-4977-ae9d-01cf681f6eb5 RequestID: 18119DCBBA1D46DE RemoteHost: 127.0.0.1 Host: localhost:11111 UserAgent: aws-sdk-cpp/1.11.234 ua/2.0 md/aws-crt#0.24.11-dev+cddd1639 os/Linux/5.15.0-125-generic md/arch#x86_64 lang/c++#C++23 md/Clang#17.0.6 cfg/retry-mode#custom api/S3 Error: open /minio_data/.minio.sys/multipart/19db0fcdb162e2e3aaaa1cbfb9a9f68e9d27fc7e1bd20e2f4ee08cde68384d76/374a8652-26ff-4e52-9683-b379a179eadd/00002.d9f15be772929726021274af7b4f6b80.16777216: no such file or directory (*fs.PathError) uploadID=374a8652-26ff-4e52-9683-b379a179eadd, partPath=/minio_data/.minio.sys/multipart/19db0fcdb162e2e3aaaa1cbfb9a9f68e9d27fc7e1bd20e2f4ee08cde68384d76/374a8652-26ff-4e52-9683-b379a179eadd/00002.d9f15be772929726021274af7b4f6b80.16777216, filepath=/minio_data/.minio.sys/tmp/aec027a9-59b4-4a69-9e58-05cc47c90dc9/374a8652-26ff-4e52-9683-b379a179eadd.0cebc8d6-24fc-4069-bf9a-16ec62e5281d 1: cmd/fs-v1-multipart.go:121:cmd.(*FSObjects).backgroundAppend() 2024-12-16 10:20:32,741 Finished 15 from 16 processes 2024-12-16 10:20:37,743 Finished 15 from 16 processes 2024-12-16 10:20:42,747 Finished 15 from 16 processes 2024-12-16 10:20:47,751 Finished 15 from 16 processes 2024-12-16 10:20:52,755 Finished 15 from 16 processes 2024-12-16 10:20:57,756 Finished 15 from 16 processes 2024-12-16 10:21:02,759 Finished 15 from 16 processes 2024-12-16 10:21:07,763 Finished 15 from 16 processes 2024-12-16 10:21:12,769 Finished 15 from 16 processes 2024-12-16 10:21:17,770 Finished 15 from 16 processes 2024-12-16 10:21:22,771 Finished 15 from 16 processes 2024-12-16 10:21:27,775 Finished 15 from 16 processes 2024-12-16 10:21:32,781 Finished 15 from 16 processes 2024-12-16 10:21:37,786 Finished 15 from 16 processes 2024-12-16 10:21:42,787 Finished 15 from 16 processes 2024-12-16 10:21:47,790 Finished 15 from 16 processes 2024-12-16 10:21:52,792 Finished 15 from 16 processes 2024-12-16 10:21:57,798 Finished 15 from 16 processes 2024-12-16 10:22:02,803 Finished 15 from 16 processes 2024-12-16 10:22:07,807 Finished 15 from 16 processes 2024-12-16 10:22:12,813 Finished 15 from 16 processes 2024-12-16 10:22:17,815 Finished 15 from 16 processes 2024-12-16 10:22:22,821 Finished 15 from 16 processes 2024-12-16 10:22:27,823 Finished 15 from 16 processes 2024-12-16 10:22:32,829 Finished 15 from 16 processes 2024-12-16 10:22:37,829 Finished 15 from 16 processes 2024-12-16 10:22:42,831 Finished 15 from 16 processes 2024-12-16 10:22:47,837 Finished 15 from 16 processes 2024-12-16 10:22:52,841 Finished 15 from 16 processes 2024-12-16 10:22:57,846 Finished 15 from 16 processes 2024-12-16 10:23:02,851 Finished 15 from 16 processes 2024-12-16 10:23:07,857 Finished 15 from 16 processes 2024-12-16 10:23:12,862 Finished 15 from 16 processes 2024-12-16 10:23:17,867 Finished 15 from 16 processes 2024-12-16 10:23:22,872 Finished 15 from 16 processes 2024-12-16 10:23:27,875 Finished 15 from 16 processes 2024-12-16 10:23:32,879 Finished 15 from 16 processes 2024-12-16 10:23:37,883 Finished 15 from 16 processes 2024-12-16 10:23:42,887 Finished 15 from 16 processes 2024-12-16 10:23:47,893 Finished 15 from 16 processes 2024-12-16 10:23:52,898 Finished 15 from 16 processes 2024-12-16 10:23:57,899 Finished 15 from 16 processes 2024-12-16 10:24:02,903 Finished 15 from 16 processes 2024-12-16 10:24:07,907 Finished 15 from 16 processes API: PutObjectPart(bucket=test, object=s3/bon/cxkotwriywcavhdnsflurkakmaung) Time: 09:24:12 UTC 12/16/2024 DeploymentID: c978493f-9314-4977-ae9d-01cf681f6eb5 RequestID: 18119DFF8E176AC6 RemoteHost: 127.0.0.1 Host: localhost:11111 UserAgent: aws-sdk-cpp/1.11.234 ua/2.0 md/aws-crt#0.24.11-dev+cddd1639 os/Linux/5.15.0-125-generic md/arch#x86_64 lang/c++#C++23 md/Clang#17.0.6 cfg/retry-mode#custom api/S3 Error: open /minio_data/.minio.sys/multipart/5ced7cb5268c40791320a895822dbb9f8e5a5358c7211e04e9c716b62c5f4b4d/14459e94-516f-49cc-a301-4a1b08bcca54/00002.d9f15be772929726021274af7b4f6b80.16777216: no such file or directory (*fs.PathError) uploadID=14459e94-516f-49cc-a301-4a1b08bcca54, partPath=/minio_data/.minio.sys/multipart/5ced7cb5268c40791320a895822dbb9f8e5a5358c7211e04e9c716b62c5f4b4d/14459e94-516f-49cc-a301-4a1b08bcca54/00002.d9f15be772929726021274af7b4f6b80.16777216, filepath=/minio_data/.minio.sys/tmp/aec027a9-59b4-4a69-9e58-05cc47c90dc9/14459e94-516f-49cc-a301-4a1b08bcca54.0f107174-e09a-4f04-948c-3b60203c0a5c 1: cmd/fs-v1-multipart.go:121:cmd.(*FSObjects).backgroundAppend() 2024-12-16 10:24:12,911 Finished 15 from 16 processes 2024-12-16 10:24:17,917 Finished 15 from 16 processes 2024-12-16 10:24:22,919 Finished 15 from 16 processes 2024-12-16 10:24:27,923 Finished 15 from 16 processes 2024-12-16 10:24:32,925 Finished 15 from 16 processes 2024-12-16 10:24:37,927 Finished 15 from 16 processes 2024-12-16 10:24:42,933 Finished 15 from 16 processes 2024-12-16 10:24:47,935 Finished 15 from 16 processes 2024-12-16 10:24:52,939 Finished 15 from 16 processes 2024-12-16 10:24:57,943 Finished 15 from 16 processes 2024-12-16 10:25:02,945 All processes finished 2024-12-16 10:25:02,945 Compressing stress logs 2024-12-16 10:25:02,974 Logs compressed 2024-12-16 10:25:02,974 Will terminate gdb (if any) 2024-12-16 10:25:02,974 Running command: kill -TERM $(pidof gdb) 2024-12-16 10:25:02,978 Running command: timeout 50s tail --pid=$(pidof gdb) -f /dev/null || kill -9 $(pidof gdb) ||: Quit 2024-12-16 10:25:03,995 Running command: kill -CONT $(cat /var/run/clickhouse-server/clickhouse-server.pid) && clickhouse client -q 'SELECT 1 FORMAT Null' 2024-12-16 10:25:04,360 Running command: clickhouse client -q "SYSTEM STOP THREAD FUZZER" --max_untracked_memory=1Gi --memory_profiler_step=1Gi --max_memory_usage_for_user=0 --max_memory_usage_in_client=1000000000 2024-12-16 10:25:04,625 Running command: clickhouse client -q "SYSTEM START MERGES" --max_untracked_memory=1Gi --memory_profiler_step=1Gi --max_memory_usage_for_user=0 --max_memory_usage_in_client=1000000000 2024-12-16 10:25:04,890 Running command: clickhouse client -q "SYSTEM START DISTRIBUTED SENDS" --max_untracked_memory=1Gi --memory_profiler_step=1Gi --max_memory_usage_for_user=0 --max_memory_usage_in_client=1000000000 2024-12-16 10:25:05,154 Running command: clickhouse client -q "SYSTEM START TTL MERGES" --max_untracked_memory=1Gi --memory_profiler_step=1Gi --max_memory_usage_for_user=0 --max_memory_usage_in_client=1000000000 2024-12-16 10:25:05,419 Running command: clickhouse client -q "SYSTEM START MOVES" --max_untracked_memory=1Gi --memory_profiler_step=1Gi --max_memory_usage_for_user=0 --max_memory_usage_in_client=1000000000 2024-12-16 10:25:05,684 Running command: clickhouse client -q "SYSTEM START FETCHES" --max_untracked_memory=1Gi --memory_profiler_step=1Gi --max_memory_usage_for_user=0 --max_memory_usage_in_client=1000000000 2024-12-16 10:25:05,948 Running command: clickhouse client -q "SYSTEM START REPLICATED SENDS" --max_untracked_memory=1Gi --memory_profiler_step=1Gi --max_memory_usage_for_user=0 --max_memory_usage_in_client=1000000000 2024-12-16 10:25:06,213 Running command: clickhouse client -q "SYSTEM START REPLICATION QUEUES" --max_untracked_memory=1Gi --memory_profiler_step=1Gi --max_memory_usage_for_user=0 --max_memory_usage_in_client=1000000000 2024-12-16 10:25:06,478 Running command: clickhouse client -q "SYSTEM DROP MARK CACHE" --max_untracked_memory=1Gi --memory_profiler_step=1Gi --max_memory_usage_for_user=0 --max_memory_usage_in_client=1000000000 2024-12-16 10:25:06,742 Running command: clickhouse client -q "KILL QUERY WHERE upper(query) LIKE 'WATCH %'" --max_untracked_memory=1Gi --memory_profiler_step=1Gi --max_memory_usage_for_user=0 --max_memory_usage_in_client=1000000000 2024-12-16 10:25:07,007 Running command: clickhouse client -q "KILL QUERY WHERE query LIKE 'insert into tableB select %'" --max_untracked_memory=1Gi --memory_profiler_step=1Gi --max_memory_usage_for_user=0 --max_memory_usage_in_client=1000000000 2024-12-16 10:25:07,322 Running command: clickhouse client -q "KILL QUERY WHERE query LIKE 'SELECT URL, uniq(SearchPhrase) AS u FROM test.hits GROUP BY URL ORDER BY u %'" --max_untracked_memory=1Gi --memory_profiler_step=1Gi --max_memory_usage_for_user=0 --max_memory_usage_in_client=1000000000 2024-12-16 10:25:07,586 Running command: clickhouse client -q "KILL QUERY WHERE query LIKE 'SELECT (SELECT number FROM system.numbers WHERE number = 1000000000000)%'" --max_untracked_memory=1Gi --memory_profiler_step=1Gi --max_memory_usage_for_user=0 --max_memory_usage_in_client=1000000000 2024-12-16 10:30:09,906 Checking if some queries hung Using queries from '/usr/share/clickhouse-test/queries' directory Connecting to ClickHouse server... OK Connected to server 24.3.12.76.altinitystable @ cddd1639b2e412ae7156bb59ac7d06b73e68c286 HEAD Running 1 stateless tests (MainProcess). 00001_select_1: [ OK ] 1 tests passed. 0 tests skipped. 0.64 s elapsed (MainProcess). Won't run stateful tests because test data wasn't loaded. Checking the hung queries: done No queries hung. All tests have finished. Top patterns of log messages: count count_% size size_% uniq_loggers uniq_threads levels background_% message_format_string 1. 19422 0.078 4.74 MiB 0.058 1 63 ['Information'] 0.087 WriteBufferFromS3 is not finalized in destructor. The file might not be written to S3. {}. 2. 19044 0.076 2.07 MiB 0.026 1 19 ['Debug'] 0 Reading {} marks from part {}, total {} rows starting from the beginning of the part, column {} 3. 12738 0.051 759.32 KiB 0.009 2 105 ['Trace'] 0 Access granted: {}{} 4. 10460 0.042 2.51 MiB 0.031 1 84 ['Debug'] 0 (from {}{}{}){}{} {} (stage: {}) 5. 10366 0.041 1.55 MiB 0.019 2 83 ['Trace'] 1 {} Creating query context from {} context, user_id: {}, parent context user: {} 6. 9209 0.037 259.87 KiB 0.003 1 83 ['Debug'] 0 Processed in {} sec. 7. 7755 0.031 734.13 KiB 0.009 1 26 ['Debug'] 0 Reading {} marks from part {}, total {} rows starting from the beginning of the part 8. 5636 0.023 335.74 KiB 0.004 1 57 ['Debug'] 0.916 Key {} is removed from metadata 9. 4942 0.02 10.71 MiB 0.132 1 72 ['Trace'] 0 Query {} to stage {}{} 10. 4924 0.02 283.71 KiB 0.003 1 236 ['Debug'] 0.039 Submitting key {} for removal 11. 4911 0.02 10.82 MiB 0.133 1 72 ['Trace'] 0 Query {} from stage {} to stage {}{} 12. 4820 0.019 380.83 KiB 0.005 1 70 ['Debug'] 0 Read {} rows, {} in {} sec., {} rows/sec., {}/sec. 13. 4558 0.018 316.31 KiB 0.004 165 581 ['Trace'] 0.892 Trying to reserve {} using storage policy from min volume index {} 14. 4124 0.016 10.34 MiB 0.127 3 512 ['Error'] 1 Table is in readonly mode (replica path: {}) 15. 4031 0.016 180.21 KiB 0.002 1 1 ['Trace'] 1 Processing requests batch, size: {}, bytes: {} 16. 3884 0.016 187.69 KiB 0.002 1 122 ['Trace'] 0.703 filled checksums {} 17. 3641 0.015 1.47 MiB 0.018 1 502 ['Debug'] 0.987 Objects with paths [{}] were removed from S3 18. 3622 0.014 434.01 KiB 0.005 148 123 ['Trace'] 0.711 Renaming temporary part {} to {} with tid {}. 19. 3418 0.014 180.85 KiB 0.002 1 69 ['Debug'] 0.357 Peak memory usage{}: {}. 20. 3255 0.013 266.11 KiB 0.003 3 90 ['Debug'] 0.998 {} Authenticating user '{}' from {} 21. 3232 0.013 362.97 KiB 0.004 3 90 ['Debug'] 0.998 {} Authenticated with global context as user {} 22. 3226 0.013 283.54 KiB 0.003 3 90 ['Debug'] 0.998 {} Logout, user_id: {} 23. 3223 0.013 352.52 KiB 0.004 2 83 ['Debug'] 1 {} Creating session context with user_id: {} 24. 2887 0.012 3.15 MiB 0.039 1 177 ['Debug'] 0.987 metadata and objects were removed for [{}], only metadata were removed for [{}]. 25. 2792 0.011 161.86 KiB 0.002 15 15 ['Trace'] 1 Flushing system log, {} entries to flush up to offset {} 26. 2790 0.011 100.36 KiB 0.001 15 15 ['Trace'] 1 Flushed system log up to offset {} 27. 2585 0.01 227.48 KiB 0.003 14 32 ['Trace'] 0 Reading {} ranges in{}order from part {}, approx. {} rows starting from {} 28. 2522 0.01 82.48 KiB 0.001 1 83 ['Trace'] 1 TCP Request. Address: {} 29. 2521 0.01 232.32 KiB 0.003 1 83 ['Debug'] 1 Connected {} version {}.{}.{}, revision: {}{}{}. 30. 2457 0.01 13.37 MiB 0.165 28 91 ['Error'] 0.396 Cannot schedule a task: {} (threads={}, jobs={}) 31. 2338 0.009 61.65 KiB 0.001 1 83 ['Debug'] 1 Done processing connection. 32. 1663 0.007 203.23 KiB 0.002 1 1 ['Trace'] 1 MemoryTracking: was {}, peak {}, free memory in arenas {}, will set to {} (RSS), difference: {} 33. 1632 0.007 145.82 KiB 0.002 1 253 ['Trace'] 0 Aggregated. {} to {} rows (from {}) in {} sec. ({:.3f} rows/sec., {}/sec.) 34. 1612 0.006 50.07 KiB 0.001 1 261 ['Trace'] 0 Aggregation method: {} 35. 1584 0.006 261.06 KiB 0.003 1 448 ['Debug'] 0.638 Recursively remove path {}: metadata and objects were removed for [{}], only metadata were removed for [{}]. 36. 1374 0.005 203.95 KiB 0.002 1 39 ['Debug'] 0 Merged partially aggregated blocks for bucket #{}. Got {} rows, {} from {} source rows in {} sec. ({:.3f} rows/sec., {}/sec.) 37. 1373 0.005 67.76 KiB 0.001 1 38 ['Trace'] 0 Merging partially aggregated blocks (bucket = {}). 38. 1304 0.005 69.98 KiB 0.001 39 483 ['Debug'] 0.985 Selected {} parts from {} to {} 39. 1291 0.005 106.55 KiB 0.001 1 28 ['Debug'] 0 Merging {} parts: from {} to {} into {} with storage {} 40. 1290 0.005 43.27 KiB 0.001 1 28 ['Debug'] 0 Selected MergeAlgorithm: {} 41. 1207 0.005 208.02 KiB 0.002 1 18 ['Trace'] 0 {}Keys: {}, datatype: {}, kind: {}, strictness: {}, right header: {} 42. 1206 0.005 85.82 KiB 0.001 1 21 ['Trace'] 0.002 {}Join data is being destroyed, {} bytes and {} rows in hash table 43. 1176 0.005 158.48 KiB 0.002 1 123 ['Information'] 0 Sorting and writing part of data into temporary file {} 44. 1084 0.004 150.11 KiB 0.002 1 34 ['Trace'] 0.015 PREWHERE condition was split into {} steps: {} 45. 1059 0.004 81.70 KiB 0.001 1 43 ['Trace'] 0 Query span trace_id for opentelemetry log: {} 46. 938 0.004 163.58 KiB 0.002 1 118 ['Information'] 0 Done writing part of data into temporary file {}, compressed {}, uncompressed {} 47. 928 0.004 9.97 KiB 0 1 222 ['Trace'] 0 Aggregating 48. 883 0.004 109.68 KiB 0.001 1 412 ['Information'] 1 Have {} tables in drop queue ({} of them are in use), will try drop {} 49. 880 0.004 65.31 KiB 0.001 1 30 ['Debug'] 0 Waiting for table {} to be finally dropped 50. 880 0.004 27.68 KiB 0 19 18 ['Debug'] 0 Requested flush up to offset {} 51. 879 0.004 168.05 KiB 0.002 1 412 ['Information'] 1 Removing metadata {} of dropped table {} 52. 824 0.003 14.48 KiB 0 784 39 ['Debug'] 0.008 Loading data parts 53. 802 0.003 2.46 MiB 0.03 2 1 ['Error'] 0 Table {}.{} does not exist. Maybe you meant {}.{}? 54. 778 0.003 17.47 KiB 0 772 38 ['Debug'] 0.009 There are no data parts 55. 774 0.003 16.95 KiB 0 25 404 ['Trace'] 1 Execution took {} ms. 56. 772 0.003 414.16 KiB 0.005 2 36 ['Trace'] 1 Request URI: {} 57. 772 0.003 137.17 KiB 0.002 2 36 ['Trace'] 1 HTTP Request for {}. Method: {}, Address: {}, User-Agent: {}{}, Content Type: {}, Transfer Encoding: {}, X-Forwarded-For: {} 58. 770 0.003 100.64 KiB 0.001 1 27 ['Debug'] 0 Merge sorted {} rows, containing {} columns ({} merged, {} gathered) in {} sec., {} rows/sec., {}/sec. 59. 744 0.003 100.27 KiB 0.001 1 46 ['Debug'] 0 Writing part of aggregation data into temporary file {} 60. 744 0.003 33.32 KiB 0 1 46 ['Debug'] 0 Max size of temporary block: {} rows, {}. 61. 743 0.003 184.12 KiB 0.002 1 46 ['Debug'] 0 Written part in {:.3f} sec., {} rows, {} uncompressed, {} compressed, {:.3f} uncompressed bytes per row, {:.3f} compressed bytes per row, compression rate: {:.3f} ({:.3f} rows/sec., {}/sec. uncompressed, {}/sec. compressed) 62. 736 0.003 22.60 KiB 0 93 52 ['Debug'] 0.003 Key condition: {} 63. 725 0.003 41.78 KiB 0.001 722 384 ['Trace'] 0.999 dropAllData: removing data parts (count {}) from filesystem. 64. 725 0.003 21.95 KiB 0 722 384 ['Trace'] 0.999 dropAllData: waiting for locks. 65. 722 0.003 46.54 KiB 0.001 722 384 ['Information'] 0.999 dropAllData: removing table directory recursive to cleanup garbage 66. 722 0.003 54.29 KiB 0.001 722 384 ['Information'] 0.999 dropAllData: remove format_version.txt, detached, moving and write ahead logs 67. 722 0.003 30.32 KiB 0 722 384 ['Information'] 0.999 dropAllData: clearing temporary directories 68. 722 0.003 12.69 KiB 0 722 384 ['Trace'] 0.999 dropAllData: done. 69. 722 0.003 34.55 KiB 0 722 384 ['Trace'] 0.999 dropAllData: removing all data parts from memory. 70. 689 0.003 14.13 KiB 0 2 36 ['Debug'] 0.065 Done processing query 71. 675 0.003 79.01 KiB 0.001 92 51 ['Debug'] 0.001 Selected {}/{} parts by partition key, {} parts by primary key, {}/{} marks by primary key, {} marks to read from {} ranges 72. 633 0.003 41.96 KiB 0.001 32 27 ['Trace'] 0 Merged {} parts: [{}, {}] -> {} 73. 628 0.003 49.83 KiB 0.001 1 30 ['Debug'] 0 Prioritize load job '{}': {} -> {} 74. 628 0.003 43.71 KiB 0.001 1 155 ['Debug'] 1 Finish load job '{}' with status {} 75. 628 0.003 41.25 KiB 0 1 155 ['Debug'] 1 Execute load job '{}' in {} 76. 628 0.003 43.09 KiB 0.001 1 30 ['Debug'] 0 Schedule load job '{}' into {} 77. 628 0.003 19.63 KiB 0 1 185 ['Debug'] 0.546 Spawn loader worker #{} in {} 78. 628 0.003 12.88 KiB 0 1 211 ['Debug'] 1 Stop worker in {} 79. 609 0.002 31.52 KiB 0 84 50 ['Trace'] 0.002 Spreading mark ranges among streams (default reading) 80. 608 0.002 23.88 KiB 0 84 50 ['Debug'] 0.002 Reading approx. {} rows with {} streams 81. 570 0.002 5.01 KiB 0 2 30 ['Trace'] 0 No tables 82. 566 0.002 18.79 KiB 0 1 184 ['Debug'] 0.5 Change current priority: {} -> {} 83. 554 0.002 71.49 KiB 0.001 94 78 ['Debug'] 0.977 Removing {} parts from filesystem (serially): Parts: [{}] 84. 532 0.002 43.42 KiB 0.001 1 189 ['Trace'] 0 An entry for key={} found in cache: sum_of_sizes={}, median_size={} 85. 493 0.002 58.04 KiB 0.001 27 29 ['Trace'] 0.974 Found {} old parts to remove. Parts: [{}] 86. 490 0.002 58.26 KiB 0.001 27 27 ['Debug'] 0.973 Removing {} parts from memory: Parts: [{}] 87. 488 0.002 15.39 KiB 0 1 50 ['Debug'] 0.002 min_marks_for_concurrent_read={} 88. 475 0.002 13.79 KiB 0 1 10 ['Trace'] 0 {} -> {} 89. 472 0.002 18.59 KiB 0 34 47 ['Debug'] 0.004 MinMax index condition: {} 90. 437 0.002 25.18 KiB 0 1 8 ['Trace'] 0.979 Loading config file '{}'. 91. 433 0.002 13.04 KiB 0 22 186 ['Trace'] 0 Found {} range in {} steps 92. 433 0.002 12.59 KiB 0 22 186 ['Trace'] 0 Found (LEFT) boundary mark: {} 93. 433 0.002 13.05 KiB 0 22 186 ['Trace'] 0 Found (RIGHT) boundary mark: {} 94. 433 0.002 29.73 KiB 0 22 186 ['Trace'] 0 Running binary search on index range for part {} ({} marks) 95. 418 0.002 132.15 KiB 0.002 309 113 ['Information'] 0.024 Setting download as failed: {} 96. 400 0.002 14.06 KiB 0 1 155 ['Trace'] 0 Converting aggregated data to blocks 97. 400 0.002 17.19 KiB 0 1 153 ['Trace'] 0 Reading blocks of partially aggregated data. 98. 400 0.002 42.14 KiB 0.001 1 155 ['Debug'] 0 Converted aggregated data to blocks. {} rows, {} in {} sec. ({:.3f} rows/sec., {}/sec.) 99. 396 0.002 17.79 KiB 0 1 152 ['Trace'] 0 Merged partially aggregated single-level data. 100. 396 0.002 18.18 KiB 0 1 152 ['Trace'] 0 Merging partially aggregated single-level data. Top messages without format string (fmt::runtime): count pattern runtime_message line 1. 18 CodeDBExceptionReceivedfromDBExc Code: 439. DB::Exception: Received from 127.0.0.1:9000. DB::Exception: Cannot schedule a task: fault injected (threads=50, jobs=60): While executing MergeTreeSelect(pool: ReadPoolInOrder, algorithm: InOrder). Stack trace: 0. ./contrib/llvm-project/libcxx/ ('/executeQuery.cpp',218) 2. 12 CodeDBExceptionSyntaxerrorfailed Code: 62. DB::Exception: Syntax error: failed at position 32 ('DROP'): DROP COLUMN c. Expected one of: ALTER command, token, OpeningRoundBracket: In scope SELECT formatQuery('ALTER TABLE a (DROP COLUMN b), DROP COLUMN c'). (SYNTAX_ERROR) (version 24.3.12.7 ('/executeQuery.cpp',218) 3. 12 autoDBStorageReplicatedMergeTree auto DB::StorageReplicatedMergeTree::processQueueEntry(ReplicatedMergeTreeQueue::SelectedEntryPtr)::(anonymous class)::operator()(LogEntryPtr &) const: Code: 999. Coordination::Exception: Session expired. (KEEPER_EXCEPTION), Stack trace (when copying this ('/Exception.cpp',222) 4. 5 CodeCoordinationExceptionSession Code: 999. Coordination::Exception: Session expired. (KEEPER_EXCEPTION) (version 24.3.12.76.altinitystable) (from [::1]:54982) (comment: 00992_system_parts_race_condition_zookeeper_long.sh) (in query: KILL MUTATION WHERE database=currentDatabase() AND tabl ('/executeQuery.cpp',218) 5. 4 CodeDBExceptionEmptyqueryInscope Code: 62. DB::Exception: Empty query: In scope SELECT formatQuery(''). (SYNTAX_ERROR) (version 24.3.12.76.altinitystable) (from [::1]:53776) (comment: 02882_formatQuery.sql) (in query: SELECT formatQuery('');), Stack trace (when copying this message, alway ('/executeQuery.cpp',218) 6. 4 voidDBStorageReplicatedMergeTree void DB::StorageReplicatedMergeTree::mergeSelectingTask(): Code: 999. Coordination::Exception: Session expired. (KEEPER_EXCEPTION), Stack trace (when copying this message, always include the lines below): 0. ./contrib/llvm-project/libcxx/include/exception ('/Exception.cpp',222) 7. 2 CodeDBExceptionReceivedfromlocal Code: 439. DB::Exception: Received from localhost:9000. DB::Exception: Cannot schedule a task: fault injected (threads=28, jobs=28). Stack trace: 0. ./contrib/llvm-project/libcxx/include/exception:141: Poco::Exception::Exception(String const&, int) @ 0x00 ('/executeQuery.cpp',218) 8. 2 voidDBReplicatedMergeTreeCleanup void DB::ReplicatedMergeTreeCleanupThread::run(): Code: 999. Coordination::Exception: Session expired. (KEEPER_EXCEPTION), Stack trace (when copying this message, always include the lines below): 0. ./contrib/llvm-project/libcxx/include/exception:141: Poc ('/Exception.cpp',222) 9. 1 CodeDBExceptionExceptionUnsuppor Code: 1000. DB::Exception: Exception: Unsupported language for region name. Supported languages are: 'ru', 'en', 'ua', 'uk', 'by', 'kz', 'tr', 'de', 'uz', 'lv', 'lt', 'et', 'pt', 'he', 'vi', 'es'. (POCO_EXCEPTION), Stack trace (when copying this message, a ('/TCPHandler.cpp',701) 10. 1 Cannotdroptabletestkeepermapback Cannot drop table test_1_02911_keeper_map.`02911_backup_restore_keeper_map3` (695bff83-23ec-4e48-a0a9-cb1941cd6ded). Will retry later.: Code: 999. Coordination::Exception: Session expired (fault injected on send). (KEEPER_EXCEPTION), Stack trace (when copy ('/Exception.cpp',222) 11. 1 PocoExceptionCodeecodeExceptionU Poco::Exception. Code: 1000, e.code() = 0, Exception: Unsupported language for region name. Supported languages are: 'ru', 'en', 'ua', 'uk', 'by', 'kz', 'tr', 'de', 'uz', 'lv', 'lt', 'et', 'pt', 'he', 'vi', 'es'. (version 24.3.12.76.altinitystable) (from [ ('/executeQuery.cpp',218) Top messages not matching their format strings: message_format_string count() any_message 1. {} is in use (by merge/mutation/INSERT) (consider increasing temporary_directories_lifetime setting) 176 /var/lib/clickhouse/disks/s3_disk/store/b60/b608a48e-3043-4573-bd9c-1ee892c1d911/tmp_merge_201403_1_6_1/ is in use (by merge/mutation/INSERT) (consider increasing temporary_directories_lifetime setting) (skipped 1 similar messages) 2. Query {} to stage {}{} 40 Query SELECT IPv6NumToString(bitAnd(materialize(_CAST(' \r����:�ڊ.ps4', 'FixedString(16)')), _CAST('������\0\0\0\0\0\0\0\0\0\0', 'FixedString(16)'))) AS `IPv6NumToString(bitAnd(materialize(IPv6StringToNum('2001:0db8:85a3:8d3a:b2da:8a2e:0370:7334')), IPv6StringToNum('ffff:ffff:ffff:0000:0000:0000:0000:0000')))` FROM system.numbers AS __table1 LIMIT _CAST(10, 'UInt64') to stage Complete 3. Query {} from stage {} to stage {}{} 40 Query SELECT IPv6NumToString(bitAnd(materialize(_CAST(' \r����:�ڊ.ps4', 'FixedString(16)')), _CAST('������\0\0\0\0\0\0\0\0\0\0', 'FixedString(16)'))) AS `IPv6NumToString(bitAnd(materialize(IPv6StringToNum('2001:0db8:85a3:8d3a:b2da:8a2e:0370:7334')), IPv6StringToNum('ffff:ffff:ffff:0000:0000:0000:0000:0000')))` FROM system.numbers AS __table1 LIMIT _CAST(10, 'UInt64') from stage FetchColumns to stage Complete Top short messages: c message_format_string substr(any(message), 1, 120) min_length_without_exception_boilerplate 1. 3 Froze {} parts Froze 0 parts -13 2. 3 Creating {}: {} Creating table test_1_02911_keeper_map.`02911_backup_restore_keeper_map2`: CREATE TABLE IF NOT EXISTS test_1_02911_keepe 290 3. 1 {} Server has databases (for example `ordinary_test_11`) with Ordinary engine, which was deprecated. To convert this databa 396 Top messages by level: (0.016449480868105,'Table is in readonly mode (replica path: {})') Error (0.0001475826363045308,'Client has gone away.') Warning (0.07746891790017829,'WriteBufferFromS3 is not finalized in destructor. The file might not be written to S3. {}.') Information (0.07596118177793201,'Reading {} marks from part {}, total {} rows starting from the beginning of the part, column {}') Debug (0.05082825768726043,'Access granted: {}{}') Trace 2024-12-16 10:30:29,925 Stress test finished + echo -e 'Test script exit code\tOK\t\N\t' + rg -Fa 'No queries hung' /test_output/test_results.tsv + grep -Fa OK No queries hung OK \N + stop_server + local max_tries=90 + local check_hang=true + local pid ++ cat /var/run/clickhouse-server/clickhouse-server.pid + pid=2778 + clickhouse stop --max-tries 90 --do-not-kill /var/run/clickhouse-server/clickhouse-server.pid file exists and contains pid = 2778. The process with pid = 2778 is running. Sent terminate signal to process with pid 2778. Waiting for server to stop /var/run/clickhouse-server/clickhouse-server.pid file exists and contains pid = 2778. The process with pid = 2778 is running. Waiting for server to stop /var/run/clickhouse-server/clickhouse-server.pid file exists and contains pid = 2778. The process with pid = 2778 is running. Waiting for server to stop /var/run/clickhouse-server/clickhouse-server.pid file exists and contains pid = 2778. The process with pid = 2778 is running. Waiting for server to stop /var/run/clickhouse-server/clickhouse-server.pid file exists and contains pid = 2778. The process with pid = 2778 is running. Waiting for server to stop /var/run/clickhouse-server/clickhouse-server.pid file exists and contains pid = 2778. The process with pid = 2778 is running. Waiting for server to stop /var/run/clickhouse-server/clickhouse-server.pid file exists and contains pid = 2778. The process with pid = 2778 is running. Waiting for server to stop /var/run/clickhouse-server/clickhouse-server.pid file exists and contains pid = 2778. The process with pid = 2778 is running. Waiting for server to stop /var/run/clickhouse-server/clickhouse-server.pid file exists and contains pid = 2778. The process with pid = 2778 does not exist. Server stopped + return + mv /var/log/clickhouse-server/clickhouse-server.log /var/log/clickhouse-server/clickhouse-server.stress.log + unset THREAD_FUZZER_CPU_TIME_PERIOD_US THREAD_FUZZER_EXPLICIT_MEMORY_EXCEPTION_PROBABILITY THREAD_FUZZER_EXPLICIT_SLEEP_PROBABILITY THREAD_FUZZER_SLEEP_PROBABILITY THREAD_FUZZER_SLEEP_TIME_US_MAX THREAD_FUZZER_pthread_mutex_lock_AFTER_MIGRATE_PROBABILITY THREAD_FUZZER_pthread_mutex_lock_AFTER_SLEEP_PROBABILITY THREAD_FUZZER_pthread_mutex_lock_AFTER_SLEEP_TIME_US_MAX THREAD_FUZZER_pthread_mutex_lock_BEFORE_MIGRATE_PROBABILITY THREAD_FUZZER_pthread_mutex_lock_BEFORE_SLEEP_PROBABILITY THREAD_FUZZER_pthread_mutex_lock_BEFORE_SLEEP_TIME_US_MAX THREAD_FUZZER_pthread_mutex_unlock_AFTER_MIGRATE_PROBABILITY THREAD_FUZZER_pthread_mutex_unlock_AFTER_SLEEP_PROBABILITY THREAD_FUZZER_pthread_mutex_unlock_AFTER_SLEEP_TIME_US_MAX THREAD_FUZZER_pthread_mutex_unlock_BEFORE_MIGRATE_PROBABILITY THREAD_FUZZER_pthread_mutex_unlock_BEFORE_SLEEP_PROBABILITY THREAD_FUZZER_pthread_mutex_unlock_BEFORE_SLEEP_TIME_US_MAX THREAD_POOL_FAULT_INJECTION + start_server + counter=0 + max_attempt=120 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 0 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=1 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 1 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=2 + clickhouse-client --query 'SELECT 1' 127.0.0.1 - - [16/Dec/2024:09:30:38 +0000] "PUT /devstoreaccount1/cont?restype=container HTTP/1.1" 409 - 127.0.0.1 - - [16/Dec/2024:09:30:38 +0000] "PUT /devstoreaccount1/cont/hqkodwapfssttldwmvgbonhttrmvyxhl?blockid=ucfgmzipicxtmqhsosuonmwwdhdgccovuldwlqvwltcjeyuqejimgycnuvkfqira&comp=block HTTP/1.1" 201 - Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) 127.0.0.1 - - [16/Dec/2024:09:30:38 +0000] "PUT /devstoreaccount1/cont/hqkodwapfssttldwmvgbonhttrmvyxhl?comp=blocklist HTTP/1.1" 201 - 127.0.0.1 - - [16/Dec/2024:09:30:38 +0000] "GET /devstoreaccount1/cont/hqkodwapfssttldwmvgbonhttrmvyxhl HTTP/1.1" 206 4 127.0.0.1 - - [16/Dec/2024:09:30:38 +0000] "GET /devstoreaccount1/cont/hqkodwapfssttldwmvgbonhttrmvyxhl HTTP/1.1" 206 2 + '[' 2 -gt 120 ']' + clickhouse start --user root 127.0.0.1 - - [16/Dec/2024:09:30:38 +0000] "DELETE /devstoreaccount1/cont/hqkodwapfssttldwmvgbonhttrmvyxhl HTTP/1.1" 202 - + sleep 0.5 + counter=3 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 3 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=4 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 4 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=5 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 5 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=6 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 6 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=7 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 7 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=8 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 8 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=9 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 9 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=10 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 10 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=11 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 11 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=12 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 12 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=13 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 13 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=14 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 14 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=15 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 15 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=16 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 16 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=17 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 17 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=18 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 18 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=19 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 19 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=20 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 20 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=21 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 21 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=22 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 22 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=23 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 23 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=24 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 24 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=25 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 25 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=26 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 26 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=27 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 27 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=28 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 28 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=29 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 29 -gt 120 ']' + clickhouse start --user root + sleep 0.5 127.0.0.1 - - [16/Dec/2024:09:30:59 +0000] "PUT /devstoreaccount1/cont?restype=container HTTP/1.1" 409 - 127.0.0.1 - - [16/Dec/2024:09:30:59 +0000] "PUT /devstoreaccount1/cont/bcwwdagvfwcnmdqodcfipnhfwoxcjxgc?blockid=qniuyggtooyonsjkjrrbrwunnrrgtzckhrydzidkfxunpmpfsbsqaxvcmkcssonz&comp=block HTTP/1.1" 201 - 127.0.0.1 - - [16/Dec/2024:09:30:59 +0000] "PUT /devstoreaccount1/cont/bcwwdagvfwcnmdqodcfipnhfwoxcjxgc?comp=blocklist HTTP/1.1" 201 - 127.0.0.1 - - [16/Dec/2024:09:30:59 +0000] "GET /devstoreaccount1/cont/bcwwdagvfwcnmdqodcfipnhfwoxcjxgc HTTP/1.1" 206 4 127.0.0.1 - - [16/Dec/2024:09:30:59 +0000] "GET /devstoreaccount1/cont/bcwwdagvfwcnmdqodcfipnhfwoxcjxgc HTTP/1.1" 206 2 127.0.0.1 - - [16/Dec/2024:09:30:59 +0000] "DELETE /devstoreaccount1/cont/bcwwdagvfwcnmdqodcfipnhfwoxcjxgc HTTP/1.1" 202 - + counter=30 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 30 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=31 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 31 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=32 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 32 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=33 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 33 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=34 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 34 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=35 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 35 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=36 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 36 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=37 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 37 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=38 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 38 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=39 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 39 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=40 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 40 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=41 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 41 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=42 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 42 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=43 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 43 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=44 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 44 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=45 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 45 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=46 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 46 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=47 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 47 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=48 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 48 -gt 120 ']' + clickhouse start --user root + sleep 0.5 127.0.0.1 - - [16/Dec/2024:09:31:13 +0000] "PUT /devstoreaccount1/cont?restype=container HTTP/1.1" 409 - 127.0.0.1 - - [16/Dec/2024:09:31:13 +0000] "PUT /devstoreaccount1/cont/fmzlprxvjruuxjyrfmhksuxsywzvsidh?blockid=lyryjneldfruksrxhigripbnnvnzuciokqriurkxqidqaydmjfvsfsyobaczodlq&comp=block HTTP/1.1" 201 - 127.0.0.1 - - [16/Dec/2024:09:31:13 +0000] "PUT /devstoreaccount1/cont/fmzlprxvjruuxjyrfmhksuxsywzvsidh?comp=blocklist HTTP/1.1" 201 - 127.0.0.1 - - [16/Dec/2024:09:31:13 +0000] "GET /devstoreaccount1/cont/fmzlprxvjruuxjyrfmhksuxsywzvsidh HTTP/1.1" 206 4 127.0.0.1 - - [16/Dec/2024:09:31:13 +0000] "GET /devstoreaccount1/cont/fmzlprxvjruuxjyrfmhksuxsywzvsidh HTTP/1.1" 206 2 127.0.0.1 - - [16/Dec/2024:09:31:13 +0000] "DELETE /devstoreaccount1/cont/fmzlprxvjruuxjyrfmhksuxsywzvsidh HTTP/1.1" 202 - + counter=49 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 49 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=50 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 50 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=51 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 51 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=52 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 52 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=53 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 53 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=54 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 54 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=55 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 55 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=56 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 56 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=57 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 57 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=58 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 58 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=59 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 59 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=60 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 60 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=61 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 61 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=62 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 62 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=63 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 63 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=64 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 64 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=65 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 65 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=66 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 66 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=67 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 67 -gt 120 ']' + clickhouse start --user root + sleep 0.5 127.0.0.1 - - [16/Dec/2024:09:31:28 +0000] "PUT /devstoreaccount1/cont?restype=container HTTP/1.1" 409 - 127.0.0.1 - - [16/Dec/2024:09:31:28 +0000] "PUT /devstoreaccount1/cont/pllzgftilcvcnjgkgfhyvjufqykvtsty?blockid=nmyabmybnflqiinzclaqxfbtcffwuhryshwynxgbylfztlxiddvfksbjxhzqejxr&comp=block HTTP/1.1" 201 - 127.0.0.1 - - [16/Dec/2024:09:31:28 +0000] "PUT /devstoreaccount1/cont/pllzgftilcvcnjgkgfhyvjufqykvtsty?comp=blocklist HTTP/1.1" 201 - 127.0.0.1 - - [16/Dec/2024:09:31:28 +0000] "GET /devstoreaccount1/cont/pllzgftilcvcnjgkgfhyvjufqykvtsty HTTP/1.1" 206 4 127.0.0.1 - - [16/Dec/2024:09:31:28 +0000] "GET /devstoreaccount1/cont/pllzgftilcvcnjgkgfhyvjufqykvtsty HTTP/1.1" 206 2 127.0.0.1 - - [16/Dec/2024:09:31:28 +0000] "DELETE /devstoreaccount1/cont/pllzgftilcvcnjgkgfhyvjufqykvtsty HTTP/1.1" 202 - + counter=68 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 68 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=69 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 69 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=70 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 70 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=71 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 71 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=72 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 72 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=73 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 73 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=74 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 74 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=75 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 75 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=76 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 76 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=77 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 77 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=78 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 78 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=79 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 79 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=80 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 80 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=81 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 81 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=82 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 82 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=83 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 83 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=84 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 84 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=85 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 85 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=86 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 86 -gt 120 ']' + clickhouse start --user root + sleep 0.5 127.0.0.1 - - [16/Dec/2024:09:31:43 +0000] "PUT /devstoreaccount1/cont?restype=container HTTP/1.1" 409 - 127.0.0.1 - - [16/Dec/2024:09:31:43 +0000] "PUT /devstoreaccount1/cont/xwnljnbwtnuzxxyosmipafjmbddfhxdn?blockid=kbnccbtdddntbdhkzsicqfpcdcnwtcfegqqgptxysnfxtjaidccqrnmzapvaggpu&comp=block HTTP/1.1" 201 - 127.0.0.1 - - [16/Dec/2024:09:31:43 +0000] "PUT /devstoreaccount1/cont/xwnljnbwtnuzxxyosmipafjmbddfhxdn?comp=blocklist HTTP/1.1" 201 - 127.0.0.1 - - [16/Dec/2024:09:31:43 +0000] "GET /devstoreaccount1/cont/xwnljnbwtnuzxxyosmipafjmbddfhxdn HTTP/1.1" 206 4 127.0.0.1 - - [16/Dec/2024:09:31:43 +0000] "GET /devstoreaccount1/cont/xwnljnbwtnuzxxyosmipafjmbddfhxdn HTTP/1.1" 206 2 127.0.0.1 - - [16/Dec/2024:09:31:43 +0000] "DELETE /devstoreaccount1/cont/xwnljnbwtnuzxxyosmipafjmbddfhxdn HTTP/1.1" 202 - + counter=87 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 87 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=88 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 88 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=89 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 89 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=90 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 90 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=91 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 91 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=92 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 92 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=93 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 93 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=94 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 94 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=95 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 95 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=96 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 96 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=97 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 97 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=98 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 98 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=99 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 99 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=100 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 100 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=101 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 101 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=102 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 102 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=103 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 103 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=104 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 104 -gt 120 ']' + clickhouse start --user root + sleep 0.5 127.0.0.1 - - [16/Dec/2024:09:31:56 +0000] "PUT /devstoreaccount1/cont?restype=container HTTP/1.1" 409 - 127.0.0.1 - - [16/Dec/2024:09:31:56 +0000] "PUT /devstoreaccount1/cont/hldrpwvuzkrxtcaafylrbcpywfrvygfx?blockid=yohbfondlksmodvwldscnirflhppekdemfiuxrvsjbrlktylesxjfgberdonigjs&comp=block HTTP/1.1" 201 - 127.0.0.1 - - [16/Dec/2024:09:31:56 +0000] "PUT /devstoreaccount1/cont/hldrpwvuzkrxtcaafylrbcpywfrvygfx?comp=blocklist HTTP/1.1" 201 - 127.0.0.1 - - [16/Dec/2024:09:31:56 +0000] "GET /devstoreaccount1/cont/hldrpwvuzkrxtcaafylrbcpywfrvygfx HTTP/1.1" 206 4 127.0.0.1 - - [16/Dec/2024:09:31:56 +0000] "GET /devstoreaccount1/cont/hldrpwvuzkrxtcaafylrbcpywfrvygfx HTTP/1.1" 206 2 127.0.0.1 - - [16/Dec/2024:09:31:56 +0000] "DELETE /devstoreaccount1/cont/hldrpwvuzkrxtcaafylrbcpywfrvygfx HTTP/1.1" 202 - + counter=105 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 105 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=106 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 106 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=107 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 107 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=108 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 108 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=109 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 109 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=110 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 110 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=111 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 111 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=112 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 112 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=113 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 113 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=114 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 114 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=115 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 115 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=116 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 116 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=117 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 117 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=118 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 118 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=119 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 119 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=120 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 120 -gt 120 ']' + clickhouse start --user root + sleep 0.5 + counter=121 + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + '[' 121 -gt 120 ']' + echo 'Cannot start clickhouse-server' + rg --text '.*Application' /var/log/clickhouse-server/clickhouse-server.log Cannot start clickhouse-server ++ trim_server_logs application_errors.txt ++ head -n 100 /test_output/application_errors.txt ++ grep -Eo ' \[ [0-9]+ \] \{.*' ++ escaped ++ clickhouse local -S 's String' --input-format=LineAsString -q 'select substr(s, 1, 300) from table format CustomSeparated settings format_custom_row_after_delimiter='\''\\\\n'\''' + echo -e 'Cannot start clickhouse-server\tFAIL\t\N\t [ 20358 ] {} Application: Caught exception while loading metadata: Code: 695. DB::Exception: Load job \'\''load table test_1.wv\'\'' failed: Code: 1. DB::Exception: Experimental WINDOW VIEW feature is not supported with new infrastructure for query analysis (the setting \'\''allow_experimental_analyzer\\n [ 20358 ] {} Application: Code: 695. DB::Exception: Load job \'\''load table test_1.wv\'\'' failed: Code: 1. DB::Exception: Experimental WINDOW VIEW feature is not supported with new infrastructure for query analysis (the setting \'\''allow_experimental_analyzer\'\''): Cannot attach table `test_1`.`wv` fro\\n [ 21274 ] {} Application: Caught exception while loading metadata: Code: 695. DB::Exception: Load job \'\''load table test_1.wv\'\'' failed: Code: 1. DB::Exception: Experimental WINDOW VIEW feature is not supported with new infrastructure for query analysis (the setting \'\''allow_experimental_analyzer\\n [ 21274 ] {} Application: Code: 695. DB::Exception: Load job \'\''load table test_1.wv\'\'' failed: Code: 1. DB::Exception: Experimental WINDOW VIEW feature is not supported with new infrastructure for query analysis (the setting \'\''allow_experimental_analyzer\'\''): Cannot attach table `test_1`.`wv` fro\\n [ 22113 ] {} Application: Caught exception while loading metadata: Code: 695. DB::Exception: Load job \'\''load table test_1.wv\'\'' failed: Code: 1. DB::Exception: Experimental WINDOW VIEW feature is not supported with new infrastructure for query analysis (the setting \'\''allow_experimental_analyzer\\n [ 22113 ] {} Application: Code: 695. DB::Exception: Load job \'\''load table test_1.wv\'\'' failed: Code: 1. DB::Exception: Experimental WINDOW VIEW feature is not supported with new infrastructure for query analysis (the setting \'\''allow_experimental_analyzer\'\''): Cannot attach table `test_1`.`wv` fro\\n [ 22986 ] {} Application: Caught exception while loading metadata: Code: 695. DB::Exception: Load job \'\''load table test_1.wv\'\'' failed: Code: 1. DB::Exception: Experimental WINDOW VIEW feature is not supported with new infrastructure for query analysis (the setting \'\''allow_experimental_analyzer\\n [ 22986 ] {} Application: Code: 695. DB::Exception: Load job \'\''load table test_1.wv\'\'' failed: Code: 1. DB::Exception: Experimental WINDOW VIEW feature is not supported with new infrastructure for query analysis (the setting \'\''allow_experimental_analyzer\'\''): Cannot attach table `test_1`.`wv` fro\\n [ 23831 ] {} Application: Caught exception while loading metadata: Code: 695. DB::Exception: Load job \'\''load table test_1.wv\'\'' failed: Code: 1. DB::Exception: Experimental WINDOW VIEW feature is not supported with new infrastructure for query analysis (the setting \'\''allow_experimental_analyzer\\n [ 23831 ] {} Application: Code: 695. DB::Exception: Load job \'\''load table test_1.wv\'\'' failed: Code: 1. DB::Exception: Experimental WINDOW VIEW feature is not supported with new infrastructure for query analysis (the setting \'\''allow_experimental_analyzer\'\''): Cannot attach table `test_1`.`wv` fro\\n [ 24679 ] {} Application: Caught exception while loading metadata: Code: 695. DB::Exception: Load job \'\''load table test_1.wv\'\'' failed: Code: 1. DB::Exception: Experimental WINDOW VIEW feature is not supported with new infrastructure for query analysis (the setting \'\''allow_experimental_analyzer\\n' + cat /var/log/clickhouse-server/stdout.log /var/run/clickhouse-server/clickhouse-server.pid file exists and contains pid = 24679. The process with pid = 24679 is already running. + tail -n100 /var/log/clickhouse-server/stderr.log Logging trace to /var/log/clickhouse-server/clickhouse-server.log Logging errors to /var/log/clickhouse-server/clickhouse-server.err.log Logging trace to /var/log/clickhouse-server/clickhouse-server.log Logging errors to /var/log/clickhouse-server/clickhouse-server.err.log Logging test to /var/log/clickhouse-server/clickhouse-server.log Logging errors to /var/log/clickhouse-server/clickhouse-server.err.log Logging test to /var/log/clickhouse-server/clickhouse-server.log Logging errors to /var/log/clickhouse-server/clickhouse-server.err.log Logging test to /var/log/clickhouse-server/clickhouse-server.log Logging errors to /var/log/clickhouse-server/clickhouse-server.err.log Logging test to /var/log/clickhouse-server/clickhouse-server.log Logging errors to /var/log/clickhouse-server/clickhouse-server.err.log Logging test to /var/log/clickhouse-server/clickhouse-server.log Logging errors to /var/log/clickhouse-server/clickhouse-server.err.log Logging test to /var/log/clickhouse-server/clickhouse-server.log Logging errors to /var/log/clickhouse-server/clickhouse-server.err.log Logging test to /var/log/clickhouse-server/clickhouse-server.log Logging errors to /var/log/clickhouse-server/clickhouse-server.err.log + tail -n100000 /var/log/clickhouse-server/clickhouse-server.log + rg -F -v -e ' RaftInstance:' -e ' RaftInstance' + tail -n100 2024.12.16 10:32:08.113435 [ 24679 ] {} DiskObjectStorage(s3_disk): Shutting down disk s3_disk 2024.12.16 10:32:08.113473 [ 24679 ] {} DiskObjectStorage(s3_disk): Disk s3_disk shut down 2024.12.16 10:32:08.113496 [ 24679 ] {} Context: Shutdown disk s3_cache_02933 2024.12.16 10:32:08.113518 [ 24679 ] {} DiskObjectStorage(s3_disk): Shutting down disk s3_disk 2024.12.16 10:32:08.113540 [ 24679 ] {} DiskObjectStorage(s3_disk): Disk s3_disk shut down 2024.12.16 10:32:08.113561 [ 24679 ] {} Context: Shutdown disk s3_cache_02944 2024.12.16 10:32:08.113584 [ 24679 ] {} DiskObjectStorage(s3_disk): Shutting down disk s3_disk 2024.12.16 10:32:08.113605 [ 24679 ] {} DiskObjectStorage(s3_disk): Disk s3_disk shut down 2024.12.16 10:32:08.113626 [ 24679 ] {} Context: Shutdown disk s3_cache_multi 2024.12.16 10:32:08.113648 [ 24679 ] {} DiskObjectStorage(s3_disk): Shutting down disk s3_disk 2024.12.16 10:32:08.113671 [ 24679 ] {} DiskObjectStorage(s3_disk): Disk s3_disk shut down 2024.12.16 10:32:08.113692 [ 24679 ] {} Context: Shutdown disk s3_cache_multi_2 2024.12.16 10:32:08.113714 [ 24679 ] {} DiskObjectStorage(s3_disk): Shutting down disk s3_disk 2024.12.16 10:32:08.113738 [ 24679 ] {} DiskObjectStorage(s3_disk): Disk s3_disk shut down 2024.12.16 10:32:08.113758 [ 24679 ] {} Context: Shutdown disk s3_disk 2024.12.16 10:32:08.113780 [ 24679 ] {} DiskObjectStorage(s3_disk): Shutting down disk s3_disk 2024.12.16 10:32:08.113802 [ 24679 ] {} DiskObjectStorage(s3_disk): Disk s3_disk shut down 2024.12.16 10:32:08.113823 [ 24679 ] {} Context: Shutdown disk s3_disk_02944 2024.12.16 10:32:08.113847 [ 24679 ] {} DiskObjectStorage(s3_disk_02944): Shutting down disk s3_disk_02944 2024.12.16 10:32:08.113872 [ 24679 ] {} DiskObjectStorage(s3_disk_02944): Disk s3_disk_02944 shut down 2024.12.16 10:32:08.113893 [ 24679 ] {} Context: Shutdown disk s3_disk_02963 2024.12.16 10:32:08.113917 [ 24679 ] {} DiskObjectStorage(s3_disk_02963): Shutting down disk s3_disk_02963 2024.12.16 10:32:08.113951 [ 24679 ] {} DiskObjectStorage(s3_disk_02963): Disk s3_disk_02963 shut down 2024.12.16 10:32:08.113985 [ 24679 ] {} Context: Shutdown disk s3_plain_disk 2024.12.16 10:32:08.114021 [ 24679 ] {} DiskObjectStorage(s3_plain_disk): Shutting down disk s3_plain_disk 2024.12.16 10:32:08.114055 [ 24679 ] {} DiskObjectStorage(s3_plain_disk): Disk s3_plain_disk shut down 2024.12.16 10:32:08.114091 [ 24679 ] {} Context: Shutting down AccessControl 2024.12.16 10:32:08.114517 [ 24679 ] {} ZooKeeperClient: Finalizing session 32. finalization_started: false, queue_finished: false, reason: 'Destructor called' 2024.12.16 10:32:08.114754 [ 24682 ] {} KeeperTCPHandler: Polling socket for session 32 took 825 ms 2024.12.16 10:32:08.114897 [ 24682 ] {} KeeperTCPHandler: Received close event with xid 2147483647 for session id #32 2024.12.16 10:32:08.114937 [ 25313 ] {} KeeperDispatcher: Processing requests batch, size: 1, bytes: 0 2024.12.16 10:32:08.115592 [ 25320 ] {} KeeperLogStore: Flushing 1 logs 2024.12.16 10:32:08.116091 [ 24682 ] {} KeeperTCPHandler: Session #32 successfully closed 2024.12.16 10:32:08.118975 [ 24679 ] {} BackgroundSchedulePool/BgBufSchPool: Waiting for threads to finish. 2024.12.16 10:32:08.120893 [ 24679 ] {} BackgroundSchedulePool/BgSchPool: Waiting for threads to finish. 2024.12.16 10:32:08.121479 [ 24679 ] {} BackgroundSchedulePool/BgDistSchPool: Waiting for threads to finish. 2024.12.16 10:32:08.121666 [ 24679 ] {} BackgroundSchedulePool/BgMBSchPool: Waiting for threads to finish. 2024.12.16 10:32:08.122486 [ 24679 ] {} Application: Shut down storages. 2024.12.16 10:32:08.122556 [ 24679 ] {} Application: Waiting for current connections to servers for tables to finish. 2024.12.16 10:32:08.342153 [ 24679 ] {} Application: Closed all listening sockets. 2024.12.16 10:32:08.342313 [ 24679 ] {} Application: Closed connections to servers for tables. 2024.12.16 10:32:08.342408 [ 24679 ] {} KeeperDispatcher: Shutting down storage dispatcher 2024.12.16 10:32:08.585739 [ 25324 ] {} RaftInstance: failed to accept a rpc connection due to error 125, Operation canceled 2024.12.16 10:32:08.597316 [ 24679 ] {} KeeperLogStore: Shutting down Changelog 2024.12.16 10:32:08.597372 [ 25320 ] {} KeeperLogStore: Flushing 0 logs 2024.12.16 10:32:08.597582 [ 25319 ] {} KeeperLogStore: Raft server is not set in LogStore. 2024.12.16 10:32:08.598295 [ 24679 ] {} KeeperSnapshotManagerS3: Shutting down KeeperSnapshotManagerS3 2024.12.16 10:32:08.598471 [ 24679 ] {} KeeperSnapshotManagerS3: KeeperSnapshotManagerS3 shut down 2024.12.16 10:32:08.598562 [ 24679 ] {} KeeperDispatcher: Dispatcher shut down 2024.12.16 10:32:08.598662 [ 24679 ] {} Context: Destructing remote fs threadpool reader 2024.12.16 10:32:08.598750 [ 24679 ] {} Context: Destructing local fs threadpool reader 2024.12.16 10:32:08.598825 [ 24679 ] {} Context: Destructing local fs threadpool reader 2024.12.16 10:32:08.598893 [ 24679 ] {} Context: Destructing threadpool writer 2024.12.16 10:32:08.599133 [ 24679 ] {} Context: Destructing marks loader 2024.12.16 10:32:08.841809 [ 24679 ] {} Application: Destroyed global context. 2024.12.16 10:32:08.841893 [ 24679 ] {} Application: Waiting for background threads 2024.12.16 10:32:08.937223 [ 24679 ] {} Application: Background threads finished in 95 ms 2024.12.16 10:32:08.938148 [ 24679 ] {} KeeperLogStore: Changelog is shut down 2024.12.16 10:32:08.938212 [ 24679 ] {} KeeperLogStore: Shutting down Changelog 2024.12.16 10:32:08.963582 [ 24679 ] {} Application: Code: 695. DB::Exception: Load job 'load table test_1.wv' failed: Code: 1. DB::Exception: Experimental WINDOW VIEW feature is not supported with new infrastructure for query analysis (the setting 'allow_experimental_analyzer'): Cannot attach table `test_1`.`wv` from metadata file /var/lib/clickhouse/store/528/52833ab6-63d7-4ec0-916c-3df6b5a61725/wv.sql from query ATTACH WINDOW VIEW test_1.wv UUID '369339d6-7685-4c8b-a84c-2e8808312d3e' TO test_1.dst (`count` UInt64, `w_end` DateTime('US/Samoa')) WATERMARK toIntervalSecond('2') ALLOWED_LATENESS toIntervalSecond('2') AS SELECT count(a) AS count, tumbleEnd(wid) AS w_end FROM test_1.mt GROUP BY tumble(timestamp, toIntervalSecond('5'), 'US/Samoa') AS wid. (UNSUPPORTED_METHOD), Stack trace (when copying this message, always include the lines below): 0. ./contrib/llvm-project/libcxx/include/exception:141: Poco::Exception::Exception(String const&, int) @ 0x000000003e9e3849 1. ./build_docker/./src/Common/Exception.cpp:96: DB::Exception::Exception(DB::Exception::MessageMasked&&, int, bool) @ 0x000000001f110cd3 2. DB::Exception::Exception<>(int, FormatStringHelperImpl<>) @ 0x0000000007c8764d 3. ./build_docker/./src/Storages/WindowView/StorageWindowView.cpp:1181: DB::StorageWindowView::StorageWindowView(DB::StorageID const&, std::shared_ptr, DB::ASTCreateQuery const&, DB::ColumnsDescription const&, DB::LoadingStrictnessLevel) @ 0x0000000039e68c73 4. ./contrib/llvm-project/libcxx/include/__memory/allocation_guard.h:66: std::shared_ptr std::allocate_shared[abi:v15000], DB::StorageID const&, std::shared_ptr, DB::ASTCreateQuery const&, DB::ColumnsDescription const&, DB::LoadingStrictnessLevel const&, void>(std::allocator const&, DB::StorageID const&, std::shared_ptr&&, DB::ASTCreateQuery const&, DB::ColumnsDescription const&, DB::LoadingStrictnessLevel const&) @ 0x0000000039ea5b32 5. ./contrib/llvm-project/libcxx/include/__memory/shared_ptr.h:0: std::shared_ptr std::__function::__policy_invoker (DB::StorageFactory::Arguments const&)>::__call_impl (DB::StorageFactory::Arguments const&)>>(std::__function::__policy_storage const*, DB::StorageFactory::Arguments const&) @ 0x0000000039e88c02 6. ./build_docker/./src/Storages/StorageFactory.cpp:224: DB::StorageFactory::get(DB::ASTCreateQuery const&, String const&, std::shared_ptr, std::shared_ptr, DB::ColumnsDescription const&, DB::ConstraintsDescription const&, DB::LoadingStrictnessLevel) const @ 0x0000000037c955b6 7. ./build_docker/./src/Databases/DatabaseOnDisk.cpp:0: DB::createTableFromAST(DB::ASTCreateQuery, String const&, String const&, std::shared_ptr, DB::LoadingStrictnessLevel) @ 0x0000000031dd0ccf 8. ./contrib/llvm-project/libcxx/include/string:1499: DB::DatabaseOrdinary::loadTableFromMetadata(std::shared_ptr, String const&, DB::QualifiedTableName const&, std::shared_ptr const&, DB::LoadingStrictnessLevel) @ 0x0000000031e5789d 9. ./contrib/llvm-project/libcxx/include/__functional/function.h:717: ? @ 0x0000000031e725fb 10. ./contrib/llvm-project/libcxx/include/__functional/function.h:0: ? @ 0x000000001f737840 11. ./contrib/llvm-project/libcxx/include/__functional/function.h:717: ? @ 0x000000001f739e00 12. ./src/Common/Stopwatch.h:26: ThreadPoolImpl>::worker(std::__list_iterator, void*>) @ 0x000000001f34b2c4 13. ./build_docker/./src/Common/ThreadPool.cpp:0: ThreadFromGlobalPoolImpl::ThreadFromGlobalPoolImpl>::scheduleImpl(std::function, Priority, std::optional, bool)::'lambda0'()>(void&&)::'lambda'()::operator()() @ 0x000000001f3578b3 14. ./contrib/llvm-project/libcxx/include/__functional/function.h:717: ? @ 0x000000001f35775f 15. ./base/base/../base/wide_integer_impl.h:810: ThreadPoolImpl::worker(std::__list_iterator) @ 0x000000001f341d55 16. ./contrib/llvm-project/libcxx/include/__memory/unique_ptr.h:303: void* std::__thread_proxy[abi:v15000]>, void ThreadPoolImpl::scheduleImpl(std::function, Priority, std::optional, bool)::'lambda0'()>>(void*) @ 0x000000001f34fe4b 17. ? @ 0x00007f61f0230ac3 18. ? @ 0x00007f61f02c2850 (version 24.3.12.76.altinitystable). (ASYNC_LOAD_FAILED), Stack trace (when copying this message, always include the lines below): 0. ./contrib/llvm-project/libcxx/include/exception:141: Poco::Exception::Exception(String const&, int) @ 0x000000003e9e3849 1. ./build_docker/./src/Common/Exception.cpp:96: DB::Exception::Exception(DB::Exception::MessageMasked&&, int, bool) @ 0x000000001f110cd3 2. DB::Exception::Exception(int, FormatStringHelperImpl::type, std::type_identity::type>, String const&, String&&) @ 0x000000000a109295 3. ./build_docker/./src/Common/AsyncLoader.cpp:942: DB::AsyncLoader::worker(DB::AsyncLoader::Pool&) @ 0x000000001f7391ec 4. ./contrib/llvm-project/libcxx/include/__functional/function.h:717: ? @ 0x000000001f739e00 5. ./src/Common/Stopwatch.h:26: ThreadPoolImpl>::worker(std::__list_iterator, void*>) @ 0x000000001f34b2c4 6. ./build_docker/./src/Common/ThreadPool.cpp:0: ThreadFromGlobalPoolImpl::ThreadFromGlobalPoolImpl>::scheduleImpl(std::function, Priority, std::optional, bool)::'lambda0'()>(void&&)::'lambda'()::operator()() @ 0x000000001f3578b3 7. ./contrib/llvm-project/libcxx/include/__functional/function.h:717: ? @ 0x000000001f35775f 8. ./base/base/../base/wide_integer_impl.h:810: ThreadPoolImpl::worker(std::__list_iterator) @ 0x000000001f341d55 9. ./contrib/llvm-project/libcxx/include/__memory/unique_ptr.h:303: void* std::__thread_proxy[abi:v15000]>, void ThreadPoolImpl::scheduleImpl(std::function, Priority, std::optional, bool)::'lambda0'()>>(void*) @ 0x000000001f34fe4b 10. ? @ 0x00007f61f0230ac3 11. ? @ 0x00007f61f02c2850 (version 24.3.12.76.altinitystable) 2024.12.16 10:32:08.963913 [ 24679 ] {} Application: shutting down 2024.12.16 10:32:08.964024 [ 24679 ] {} Application: Uninitializing subsystem: Logging Subsystem 2024.12.16 10:32:08.983978 [ 24680 ] {} BaseDaemon: Received signal -2 2024.12.16 10:32:08.984297 [ 24680 ] {} BaseDaemon: Stop SignalListener thread 2024.12.16 10:32:09.149634 [ 24678 ] {} Application: Child process exited normally with code 183. + break + attach_gdb_to_clickhouse ++ kill -l SIGRTMIN + RTMIN=34 + echo ' set follow-fork-mode parent handle SIGHUP nostop noprint pass handle SIGINT nostop noprint pass handle SIGQUIT nostop noprint pass handle SIGPIPE nostop noprint pass handle SIGTERM nostop noprint pass handle SIGUSR1 nostop noprint pass handle SIGUSR2 nostop noprint pass handle SIG34 nostop noprint pass info signals continue backtrace full thread apply all backtrace full info registers disassemble /s up disassemble /s up disassemble /s p "done" detach quit ' + sleep 5 + ts '%Y-%m-%d %H:%M:%S' ++ cat /var/run/clickhouse-server/clickhouse-server.pid cat: /var/run/clickhouse-server/clickhouse-server.pid: No such file or directory + gdb -batch -command script.gdb -p '' Argument required (process-id to attach). script.gdb:12: Error in sourced command file: The program is not being run. + run_with_retry 60 clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' + [[ hxB =~ e ]] + set_e=false + set +e + local total_retries=60 + shift + local retry=0 + '[' 0 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=1 + sleep 5 + '[' 1 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=2 + sleep 5 + '[' 2 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=3 + sleep 5 + '[' 3 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=4 + sleep 5 + '[' 4 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=5 + sleep 5 + '[' 5 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=6 + sleep 5 + '[' 6 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=7 + sleep 5 + '[' 7 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=8 + sleep 5 + '[' 8 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=9 + sleep 5 + '[' 9 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=10 + sleep 5 + '[' 10 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=11 + sleep 5 + '[' 11 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=12 + sleep 5 + '[' 12 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=13 + sleep 5 + '[' 13 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=14 + sleep 5 + '[' 14 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=15 + sleep 5 + '[' 15 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=16 + sleep 5 + '[' 16 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=17 + sleep 5 + '[' 17 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=18 + sleep 5 + '[' 18 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=19 + sleep 5 + '[' 19 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=20 + sleep 5 + '[' 20 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=21 + sleep 5 + '[' 21 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=22 + sleep 5 + '[' 22 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=23 + sleep 5 + '[' 23 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=24 + sleep 5 + '[' 24 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=25 + sleep 5 + '[' 25 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=26 + sleep 5 + '[' 26 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=27 + sleep 5 + '[' 27 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=28 + sleep 5 + '[' 28 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=29 + sleep 5 + '[' 29 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=30 + sleep 5 + '[' 30 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=31 + sleep 5 + '[' 31 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=32 + sleep 5 + '[' 32 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=33 + sleep 5 + '[' 33 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=34 + sleep 5 + '[' 34 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=35 + sleep 5 + '[' 35 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=36 + sleep 5 + '[' 36 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=37 + sleep 5 + '[' 37 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=38 + sleep 5 + '[' 38 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=39 + sleep 5 + '[' 39 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=40 + sleep 5 + '[' 40 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=41 + sleep 5 + '[' 41 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=42 + sleep 5 + '[' 42 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=43 + sleep 5 + '[' 43 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=44 + sleep 5 + '[' 44 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=45 + sleep 5 + '[' 45 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=46 + sleep 5 + '[' 46 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=47 + sleep 5 + '[' 47 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=48 + sleep 5 + '[' 48 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=49 + sleep 5 + '[' 49 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=50 + sleep 5 + '[' 50 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=51 + sleep 5 + '[' 51 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=52 + sleep 5 + '[' 52 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=53 + sleep 5 + '[' 53 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=54 + sleep 5 + '[' 54 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=55 + sleep 5 + '[' 55 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=56 + sleep 5 + '[' 56 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=57 + sleep 5 + '[' 57 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=58 + sleep 5 + '[' 58 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=59 + sleep 5 + '[' 59 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + retry=60 + sleep 5 + '[' 60 -ge 60 ']' + echo 'Command '\''clickhouse-client --query SELECT '\''Connected to clickhouse-server after attaching gdb'\'''\'' failed after 60 retries, exiting' + exit 1 Command 'clickhouse-client --query SELECT 'Connected to clickhouse-server after attaching gdb'' failed after 60 retries, exiting