Document status - Public


QA Software Build Report

ClickHouse 22.8.13.21 / x86_64

(c) 2023 Altinity Inc. All Rights Reserved.

Approval

Status: Approved for release by QA [vzakaznikov@altinity.com Tue 14 Feb 2023 06:47:36 AM EST]

Reviewed by: azvonov@altinity.com

Date: 13 February 2023

Table of Contents

Test Results

Stage Status
Integration Fail
Stateful Pass
Stateless Fail
TestFlows Fail

Results https://altinity-test-reports.s3.amazonaws.com/index.html#builds/stable/v22.8.13.21/2023-02-10T17-31-26.183/
GitLab Pipeline https://gitlab.com/altinity-qa/clickhouse/cicd/release/-/pipelines/774258191
GitHub Actions https://github.com/Altinity/ClickHouse/actions/runs/4142324255

Results Analysis

Integration Results

Results
https://altinity-test-reports.s3.amazonaws.com/builds/stable/v22.8.13.21/2023-02-10T17-31-26.183/integration/integration_results_1.html
https://altinity-test-reports.s3.amazonaws.com/builds/stable/v22.8.13.21/2023-02-10T17-31-26.183/integration/integration_results_2.html

Test: /integration/test_cgroup_limit/test.py::test_cgroup_cpu_limit

Reason:

[gw0] linux -- Python 3.8.10 /usr/bin/python3

    def test_cgroup_cpu_limit():
        for num_cpus in (1, 2, 4, 2.8):
>           result = run_with_cpu_limit(
                "clickhouse local -q \"select value from system.settings where name='max_threads'\"",
                num_cpus,
            )

test_cgroup_limit/test.py:43: 
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
test_cgroup_limit/test.py:38: in run_with_cpu_limit
    return run_command_in_container(cmd, *args)
test_cgroup_limit/test.py:19: in run_command_in_container
    return subprocess.check_output(
/usr/lib/python3.8/subprocess.py:415: in check_output
    return run(*popenargs, stdout=PIPE, timeout=timeout, check=True,
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 

input = None, capture_output = False, timeout = None, check = True
popenargs = (['docker', 'run', '--rm', '--cpus', '1', '--volume', ...],)
kwargs = {'stdout': -1}, process = <subprocess.Popen object at 0x7fb308e068e0>
stdout = b'', stderr = None, retcode = 125

    def run(*popenargs,
            input=None, capture_output=False, timeout=None, check=False, **kwargs):
        """Run command with arguments and return a CompletedProcess instance.

        The returned instance will have attributes args, returncode, stdout and
        stderr. By default, stdout and stderr are not captured, and those attributes
        will be None. Pass stdout=PIPE and/or stderr=PIPE in order to capture them.

        If check is True and the exit code was non-zero, it raises a
        CalledProcessError. The CalledProcessError object will have the return code
        in the returncode attribute, and output & stderr attributes if those streams
        were captured.

        If timeout is given, and the process takes too long, a TimeoutExpired
        exception will be raised.

        There is an optional argument "input", allowing you to
        pass bytes or a string to the subprocess's stdin.  If you use this argument
        you may not also use the Popen constructor's "stdin" argument, as
        it will be used internally.

        By default, all communication is in bytes, and therefore any "input" should
        be bytes, and the stdout and stderr will be bytes. If in text mode, any
        "input" should be a string, and stdout and stderr will be strings decoded
        according to locale encoding, or by "encoding" if set. Text mode is
        triggered by setting any of text, encoding, errors or universal_newlines.

        The other arguments are the same as for the Popen constructor.
        """
        if input is not None:
            if kwargs.get('stdin') is not None:
                raise ValueError('stdin and input arguments may not both be used.')
            kwargs['stdin'] = PIPE

        if capture_output:
            if kwargs.get('stdout') is not None or kwargs.get('stderr') is not None:
                raise ValueError('stdout and stderr arguments may not be used '
                                 'with capture_output.')
            kwargs['stdout'] = PIPE
            kwargs['stderr'] = PIPE

        with Popen(*popenargs, **kwargs) as process:
            try:
                stdout, stderr = process.communicate(input, timeout=timeout)
            except TimeoutExpired as exc:
                process.kill()
                if _mswindows:
                    # Windows accumulates the output in a single blocking
                    # read() call run on child threads, with the timeout
                    # being done in a join() on those threads.  communicate()
                    # _after_ kill() is required to collect that and add it
                    # to the exception.
                    exc.stdout, exc.stderr = process.communicate()
                else:
                    # POSIX _communicate already populated the output so
                    # far into the TimeoutExpired exception.
                    process.wait()
                raise
            except:  # Including KeyboardInterrupt, communicate handled that.
                process.kill()
                # We don't call process.wait() as .__exit__ does that for us.
                raise
            retcode = process.poll()
            if check and retcode:
>               raise CalledProcessError(retcode, process.args,
                                         output=stdout, stderr=stderr)
E               subprocess.CalledProcessError: Command '['docker', 'run', '--rm', '--cpus', '1', '--volume', '/clickhouse:/usr/bin/clickhouse', 'ubuntu:20.04', 'sh', '-c', 'clickhouse local -q "select value from system.settings where name=\'max_threads\'"']' returned non-zero exit status 125.

/usr/lib/python3.8/subprocess.py:516: CalledProcessError

Comment: Misconfiguration: 1 CPU/hardware thread available to the CH instead of at least 2.
Status: FAIL (OK to fail)

Stateful Results

Results https://altinity-test-reports.s3.amazonaws.com/builds/stable/v22.8.13.21/2023-02-10T17-31-26.183/stateful/stateful_results.html

Stateless Results

Results https://altinity-test-reports.s3.amazonaws.com/builds/stable/v22.8.13.21/2023-02-10T17-31-26.183/stateless/stateless_results.html

Test: /stateless/02354_distributed_with_external_aggregation_memory_usage

Reason:

2023-02-10 21:28:30 [293268ce4e65] 2023.02.10 21:28:30.048134 [ 4270 ] {e4a83ceb-4450-46c3-a7b0-c2eb11f5a323}  executeQuery: Code: 241. DB::Exception: Memory limit (for query) exceeded: would use 4.03 GiB (attempt to allocate chunk of 33554672 bytes), maximum: 4.00 GiB. OvercommitTracker decision: Memory overcommit isn't used. OvercommitTracker isn't set.: while receiving packet from 127.0.0.2:9000: While executing Remote. (MEMORY_LIMIT_EXCEEDED) (version 22.8.13.21.altinitystable (altinity build)) (from [::1]:41360) (comment: 02354_distributed_with_external_aggregation_memory_usage.sql) (in query: -- whole aggregation state of local aggregation uncompressed is 5.8G
2023-02-10 21:28:30  -- it is hard to provide an accurate estimation for memory usage, so 4G is just the actual value taken from the logs + delta
2023-02-10 21:28:30  select a, b, c, sum(a) as s from remote('127.0.0.{1,2}', currentDatabase(), t_2354_dist_with_external_aggr) group by a, b, c format Null settings max_memory_usage = '4Gi';), Stack trace (when copying this message, always include the lines below):
2023-02-10 21:28:30 
2023-02-10 21:28:30 0. DB::Exception::Exception(std::__1::basic_string, std::__1::allocator > const&, int, bool) @ 0xa3fcd9a in /usr/bin/clickhouse
2023-02-10 21:28:30 1. DB::Exception::Exception, std::__1::allocator >, long&, std::__1::basic_string, std::__1::allocator >, std::__1::basic_string_view > >(int, fmt::v8::basic_format_string::type, fmt::v8::type_identity::type, fmt::v8::type_identity, std::__1::allocator > >::type, fmt::v8::type_identity::type, fmt::v8::type_identity, std::__1::allocator > >::type, fmt::v8::type_identity > >::type>, char const*&&, char const*&&, std::__1::basic_string, std::__1::allocator >&&, long&, std::__1::basic_string, std::__1::allocator >&&, std::__1::basic_string_view >&&) @ 0xa3ef417 in /usr/bin/clickhouse
2023-02-10 21:28:30 2. MemoryTracker::allocImpl(long, bool, MemoryTracker*) @ 0xa3eee01 in /usr/bin/clickhouse
2023-02-10 21:28:30 3. MemoryTracker::allocImpl(long, bool, MemoryTracker*) @ 0xa3ee832 in /usr/bin/clickhouse
2023-02-10 21:28:30 4. void DB::PODArrayBase<1ul, 4096ul, Allocator, 15ul, 16ul>::alloc<>(unsigned long) @ 0xa47107b in /usr/bin/clickhouse
2023-02-10 21:28:30 5. DB::SerializationFixedString::deserializeBinaryBulk(DB::IColumn&, DB::ReadBuffer&, unsigned long, double) const @ 0x142325d1 in /usr/bin/clickhouse
2023-02-10 21:28:30 6. DB::ISerialization::deserializeBinaryBulkWithMultipleStreams(COW::immutable_ptr&, unsigned long, DB::ISerialization::DeserializeBinaryBulkSettings&, std::__1::shared_ptr&, std::__1::unordered_map, std::__1::allocator >, COW::immutable_ptr, std::__1::hash, std::__1::allocator > >, std::__1::equal_to, std::__1::allocator > >, std::__1::allocator, std::__1::allocator > const, COW::immutable_ptr > > >*) const @ 0x1421e276 in /usr/bin/clickhouse
2023-02-10 21:28:30 7. DB::NativeReader::readData(DB::ISerialization const&, COW::immutable_ptr&, DB::ReadBuffer&, unsigned long, double) @ 0x15cc96e0 in /usr/bin/clickhouse
2023-02-10 21:28:30 8. DB::NativeReader::read() @

Comment: New fail (not seen before).
Status: FAIL

TestFlows Results

Passed:
- AES Encryption
- Aggregate Functions
- Base58
- ClickHouse Keeper
- DateTime64 Extended Range
- Disk Level Encryption
- Example
- Extended Precision Data Types
- Kafka
- Kerberos
- LDAP
- Lightweight Delete
- Map Type
- Part Moves Between Shards
- RBAC
- S3 AWS
- S3 GCS
- SSL Server
- Tiered Storage GCS
- Tiered Storage AWS
- Tiered Storage original
- Tiered Storage Minio
- Window Functions

Failed:

/parquet/postgresql/compression_type='NONE'/postgresql engine to parquet file to postgresql engine
/parquet/postgresql/compression_type='NONE'/postgresql function to parquet file to postgresql function
/parquet/postgresql/compression_type='GZIP'/postgresql engine to parquet file to postgresql engine
/parquet/postgresql/compression_type='GZIP'/postgresql function to parquet file to postgresql function
/parquet/postgresql/compression_type='LZ4'/postgresql engine to parquet file to postgresql engine
/parquet/postgresql/compression_type='LZ4'/postgresql function to parquet file to postgresql function
/s3/minio zero copy replication/lost data during mutation 

Feb 10,2023 22:32:16           ⟥  And executing command
    echo -e "
    CREATE TABLE IF NOT EXISTS table_39e94fee_a988_11ed_b802_0242ac110003 ON CLUSTER 'sharded_cluster' (key UInt32, value1 String, value2 String, value3 String) engine=ReplicatedMergeTree('/table_39e94fee_a988_11ed_b802_0242ac110003', '{replica}')
    ORDER BY key
    PARTITION BY (key % 4)
    SETTINGS storage_policy='external'
    " | clickhouse client -n 2>&1
141ms                [clickhouse1] clickhouse1  9000    0       2   0
3m 0s                [clickhouse1] Received exception from server (version 22.8.13):
3m 0s                [clickhouse1] Code: 159. DB::Exception: Received from localhost:9000. DB::Exception: Watching task /clickhouse/task_queue/ddl/query-0000000006 is executing longer than distributed_ddl_task_timeout (=180) seconds. There are 2 unfinished hosts (0 of them are currently active), they are going to execute the query in background. (TIMEOUT_EXCEEDED)
3m 0s                [clickhouse1] (query: CREATE TABLE IF NOT EXISTS table_39e94fee_a988_11ed_b802_0242ac110003 ON CLUSTER 'sharded_cluster' (key UInt32, value1 String, value2 String, value3 String) engine=ReplicatedMergeTree('/table_39e94fee_a988_11ed_b802_0242ac110003', '{replica}')
3m 0s                [clickhouse1]             ORDER BY key
3m 0s                [clickhouse1]             PARTITION BY (key % 4)
3m 0s                [clickhouse1]             SETTINGS storage_policy='external'
3m 0s                [clickhouse1] 
3m 0s                [clickhouse1] )
3m 0s                [clickhouse1] bash# echo $?
3m 0s                [clickhouse1] 159
3m 0s                [clickhouse1] bash#
3m 0s           ⟥⟤ OK executing command, /s3/minio zero copy replication/lost data during mutation/I have a table/executing command