S3 Test Run Report

DateMay 25, 2023 20:24
Duration1h 5m
Framework TestFlows 1.9.230315.1003122

Artifacts

Test artifacts can be found at https://altinity-test-reports.s3.amazonaws.com/index.html#clickhouse/22.8.15.25.altinityfips/5080932159/testflows/

Attributes

projectAltinity/ClickHouse
project.id159717931
packagehttps://s3.amazonaws.com/altinity-build-artifacts/22.8/ef5580902f444e6a5d617e7c7a6e8d7a1f5fc9fb/package_release/clickhouse-common-static_22.8.15.25.altinityfips_amd64.deb
version22.8.15.25.altinityfips
user.nameEnmk
repositoryhttps://github.com/Altinity/clickhouse-regression
commit.hash63a15b5dfc55badefcf4b869296e3ec99ca08141
job.id5080932159
job.urlhttps://github.com/Altinity/ClickHouse/actions/runs/5080932159
archx86_64
localTrue
clickhouse_versionNone
clickhouse_binary_pathhttps://s3.amazonaws.com/altinity-build-artifacts/22.8/ef5580902f444e6a5d617e7c7a6e8d7a1f5fc9fb/package_release/clickhouse-common-static_22.8.15.25.altinityfips_amd64.deb
stressFalse
collect_service_logsTrue
storages['aws_s3']
minio_urihttp://minio1:9001
minio_root_userminio
minio_root_passwordminio123
aws_s3_bucketSecret(name='aws_s3_bucket')
aws_s3_regionSecret(name='aws_s3_region')
aws_s3_key_idSecret(name='aws_s3_key_id')
aws_s3_access_keySecret(name='aws_s3_access_key')
gcs_uriSecret(name='gcs_uri')
gcs_key_idSecret(name='gcs_key_id')
gcs_key_secretSecret(name='gcs_key_secret')

Summary

77.3%OK
19.5%Known

Statistics

Units Skip OK Fail XFail
Modules
1
1
Suites
11
3
8
Features
11
3
8
Scenarios
102
90
12
Examples
60
36
24
Steps
13057
12884
151
22

Known Fails

Test NameResultMessage
/s3/aws s3 disk/cacheXFail 1ms
Under development for 22.8 and newer.
None
/s3/aws s3 disk/cache defaultXFail 719us
Under development for 22.8 and newer.
None
/s3/aws s3 disk/cache pathXFail 739us
Under development for 22.8 and newer.
None
/s3/aws s3 disk/generic urlXFail 1ms
not yet supported
Generic URL is treated as invalid configuration, ClickHouse
             will not start if config is added
/s3/aws s3 disk/low cardinality offsetXFail 40s 850ms
https://github.com/ClickHouse/ClickHouse/pull/44875
AssertionError
Traceback (most recent call last):
  File "/home/ubuntu/actions-runner/_work/ClickHouse/ClickHouse/s3/regression.py", line 486, in 
    regression()
  File "/home/ubuntu/actions-runner/_work/ClickHouse/ClickHouse/s3/regression.py", line 464, in regression
    aws_s3_regression(
  File "/home/ubuntu/actions-runner/_work/ClickHouse/ClickHouse/s3/regression.py", line 344, in aws_s3_regression
    Feature(test=load("s3.tests.disk", "aws_s3"))(
  File "/home/ubuntu/actions-runner/_work/ClickHouse/ClickHouse/s3/../s3/tests/disk.py", line 2755, in aws_s3
    disk_tests()
  File "/home/ubuntu/actions-runner/_work/ClickHouse/ClickHouse/s3/../s3/tests/disk.py", line 2742, in disk_tests
    scenario()
  File "/home/ubuntu/actions-runner/_work/ClickHouse/ClickHouse/s3/../s3/tests/disk.py", line 2711, in low_cardinality_offset
    assert output == "23999\n", error()
AssertionError: Oops! Assertion failed

The following assertion was not satisfied
  assert output == "23999\n", error()

Assertion values
  assert output == "23999\n", error()
         ^ is '23999'
  assert output == "23999\n", error()
                ^ is = False

  assert output == "23999\n", error()
  ^ is False

Where
  File '/home/ubuntu/actions-runner/_work/ClickHouse/ClickHouse/s3/../s3/tests/disk.py', line 2711 in 'low_cardinality_offset'

2703\|                          "1",
2704\|                      ),
2705\|                      (
2706\|                          "merge_tree_min_bytes_for_concurrent_read_for_remote_filesystem",
2707\|                          "1",
2708\|                      ),
2709\|                  ],
2710\|              ).output
2711\|>             assert output == "23999\n", error()
2712\|  
2713\|      finally:
2714\|          with Finally(f"I remove the table {name}"):
/s3/aws s3 invalid disk/cache path conflictXFail 1ms
Under development for 22.8 and newer.
None
/s3/aws s3 zero copy replication/add replicaXFail 52s 293ms
Under investigation
AssertionError
Traceback (most recent call last):
  File "/home/ubuntu/actions-runner/_work/ClickHouse/ClickHouse/s3/regression.py", line 486, in 
    regression()
  File "/home/ubuntu/actions-runner/_work/ClickHouse/ClickHouse/s3/regression.py", line 464, in regression
    aws_s3_regression(
  File "/home/ubuntu/actions-runner/_work/ClickHouse/ClickHouse/s3/regression.py", line 353, in aws_s3_regression
    Feature(test=load("s3.tests.zero_copy_replication", "aws_s3"))(
  File "/home/ubuntu/actions-runner/_work/ClickHouse/ClickHouse/s3/../s3/tests/zero_copy_replication.py", line 2148, in aws_s3
    outline()
  File "/home/ubuntu/actions-runner/_work/ClickHouse/ClickHouse/s3/../s3/tests/zero_copy_replication.py", line 2133, in outline
    scenario()
  File "/home/ubuntu/actions-runner/_work/ClickHouse/ClickHouse/s3/../s3/tests/zero_copy_replication.py", line 382, in add_replica
    assert size_after + 1 == get_bucket_size(
AssertionError: Oops! Assertion failed

The following assertion was not satisfied
  assert size_after + 1 == get_bucket_size(
      name=bucket_name,
      prefix=bucket_path,
      minio_enabled=minio_enabled,
      access_key=self.context.secret_access_key,
      key_id=self.context.access_key_id,
  ), error()

Assertion values
  assert size_after + 1 == get_bucket_size(
         ^ is 0
      name=bucket_name,
      prefix=bucket_path,
      minio_enabled=minio_enabled,
      access_key=self.context.secret_access_key,
      key_id=self.context.access_key_id,
  ), error()
  assert size_after + 1 == get_bucket_size(
                    ^ is = 1
      name=bucket_name,
      prefix=bucket_path,
      minio_enabled=minio_enabled,
      access_key=self.context.secret_access_key,
      key_id=self.context.access_key_id,
  ), error()
  assert size_after + 1 == get_bucket_size(
      name=bucket_name,
           ^ is 'altinity-qa-test'
      prefix=bucket_path,
      minio_enabled=minio_enabled,
      access_key=self.context.secret_access_key,
      key_id=self.context.access_key_id,
  ), error()
  assert size_after + 1 == get_bucket_size(
      name=bucket_name,
      prefix=bucket_path,
             ^ is 'data/zero-copy-replication'
      minio_enabled=minio_enabled,
      access_key=self.context.secret_access_key,
      key_id=self.context.access_key_id,
  ), error()
  assert size_after + 1 == get_bucket_size(
      name=bucket_name,
      prefix=bucket_path,
      minio_enabled=minio_enabled,
                    ^ is False
      access_key=self.context.secret_access_key,
      key_id=self.context.access_key_id,
  ), error()
  assert size_after + 1 == get_bucket_size(
      name=bucket_name,
      prefix=bucket_path,
      minio_enabled=minio_enabled,
      access_key=self.context.secret_access_key,
                 ^ is 
      key_id=self.context.access_key_id,
  ), error()
  assert size_after + 1 == get_bucket_size(
      name=bucket_name,
      prefix=bucket_path,
      minio_enabled=minio_enabled,
      access_key=self.context.secret_access_key,
                 ^ is = 
      key_id=self.context.access_key_id,
  ), error()
  assert size_after + 1 == get_bucket_size(
      name=bucket_name,
      prefix=bucket_path,
      minio_enabled=minio_enabled,
      access_key=self.context.secret_access_key,
                 ^ is 
      key_id=self.context.access_key_id,
  ), error()
  assert size_after + 1 == get_bucket_size(
      name=bucket_name,
      prefix=bucket_path,
      minio_enabled=minio_enabled,
      access_key=self.context.secret_access_key,
                 ^ is = '[masked]:Secret(name='aws_s3_access_key')'
      key_id=self.context.access_key_id,
  ), error()
  assert size_after + 1 == get_bucket_size(
      name=bucket_name,
      prefix=bucket_path,
      minio_enabled=minio_enabled,
      access_key=self.context.secret_access_key,
      key_id=self.context.access_key_id,
             ^ is 
  ), error()
  assert size_after + 1 == get_bucket_size(
      name=bucket_name,
      prefix=bucket_path,
      minio_enabled=minio_enabled,
      access_key=self.context.secret_access_key,
      key_id=self.context.access_key_id,
             ^ is = 
  ), error()
  assert size_after + 1 == get_bucket_size(
      name=bucket_name,
      prefix=bucket_path,
      minio_enabled=minio_enabled,
      access_key=self.context.secret_access_key,
      key_id=self.context.access_key_id,
             ^ is 
  ), error()
  assert size_after + 1 == get_bucket_size(
      name=bucket_name,
      prefix=bucket_path,
      minio_enabled=minio_enabled,
      access_key=self.context.secret_access_key,
      key_id=self.context.access_key_id,
             ^ is = '[masked]:Secret(name='aws_s3_key_id')'
  ), error()
  assert size_after + 1 == get_bucket_size(
                           ^ is = 0
      name=bucket_name,
      prefix=bucket_path,
      minio_enabled=minio_enabled,
      access_key=self.context.secret_access_key,
      key_id=self.context.access_key_id,
  ), error()
  assert size_after + 1 == get_bucket_size(
                        ^ is = False
      name=bucket_name,
      prefix=bucket_path,
      minio_enabled=minio_enabled,
      access_key=self.context.secret_access_key,
      key_id=self.context.access_key_id,
  ), error()
  assert size_after + 1 == get_bucket_size(
  ^ is False
      name=bucket_name,
      prefix=bucket_path,
      minio_enabled=minio_enabled,
      access_key=self.context.secret_access_key,
      key_id=self.context.access_key_id,
  ), error()

Where
  File '/home/ubuntu/actions-runner/_work/ClickHouse/ClickHouse/s3/../s3/tests/zero_copy_replication.py', line 388 in 'add_replica'

380\|                        than previously because of the additional replica"""
381\|              ):
382\|                  assert size_after + 1 == get_bucket_size(
383\|                      name=bucket_name,
384\|                      prefix=bucket_path,
385\|                      minio_enabled=minio_enabled,
386\|                      access_key=self.context.secret_access_key,
387\|                      key_id=self.context.access_key_id,
388\|>                 ), error()
389\|  
390\|              with And("I check simple queries on the first node"):
391\|                  check_query_node(
/s3/aws s3 zero copy replication/alterXFail 1m 55s
https://github.com/ClickHouse/ClickHouse/issues/22516
AssertionError
Traceback (most recent call last):
  File "/home/ubuntu/actions-runner/_work/ClickHouse/ClickHouse/s3/regression.py", line 486, in 
    regression()
  File "/home/ubuntu/actions-runner/_work/ClickHouse/ClickHouse/s3/regression.py", line 464, in regression
    aws_s3_regression(
  File "/home/ubuntu/actions-runner/_work/ClickHouse/ClickHouse/s3/regression.py", line 353, in aws_s3_regression
    Feature(test=load("s3.tests.zero_copy_replication", "aws_s3"))(
  File "/home/ubuntu/actions-runner/_work/ClickHouse/ClickHouse/s3/../s3/tests/zero_copy_replication.py", line 2148, in aws_s3
    outline()
  File "/home/ubuntu/actions-runner/_work/ClickHouse/ClickHouse/s3/../s3/tests/zero_copy_replication.py", line 2133, in outline
    scenario()
  File "/home/ubuntu/actions-runner/_work/ClickHouse/ClickHouse/s3/../s3/tests/zero_copy_replication.py", line 863, in alter
    assert False, "data in S3 has grown by more than 50%"
AssertionError: data in S3 has grown by more than 50%
/s3/aws s3 zero copy replication/alter repeatXFail 1m 52s
https://github.com/ClickHouse/ClickHouse/issues/22516
AssertionError
Traceback (most recent call last):
  File "/home/ubuntu/actions-runner/_work/ClickHouse/ClickHouse/s3/regression.py", line 486, in 
    regression()
  File "/home/ubuntu/actions-runner/_work/ClickHouse/ClickHouse/s3/regression.py", line 464, in regression
    aws_s3_regression(
  File "/home/ubuntu/actions-runner/_work/ClickHouse/ClickHouse/s3/regression.py", line 353, in aws_s3_regression
    Feature(test=load("s3.tests.zero_copy_replication", "aws_s3"))(
  File "/home/ubuntu/actions-runner/_work/ClickHouse/ClickHouse/s3/../s3/tests/zero_copy_replication.py", line 2148, in aws_s3
    outline()
  File "/home/ubuntu/actions-runner/_work/ClickHouse/ClickHouse/s3/../s3/tests/zero_copy_replication.py", line 2133, in outline
    scenario()
  File "/home/ubuntu/actions-runner/_work/ClickHouse/ClickHouse/s3/../s3/tests/zero_copy_replication.py", line 1025, in alter_repeat
    assert False, "data in S3 has grown by more than 50%"
AssertionError: data in S3 has grown by more than 50%
/s3/aws s3 zero copy replication/delete allXFail 57s 188ms
Under investigation
AssertionError
Traceback (most recent call last):
  File "/home/ubuntu/actions-runner/_work/ClickHouse/ClickHouse/s3/regression.py", line 486, in 
    regression()
  File "/home/ubuntu/actions-runner/_work/ClickHouse/ClickHouse/s3/regression.py", line 464, in regression
    aws_s3_regression(
  File "/home/ubuntu/actions-runner/_work/ClickHouse/ClickHouse/s3/regression.py", line 353, in aws_s3_regression
    Feature(test=load("s3.tests.zero_copy_replication", "aws_s3"))(
  File "/home/ubuntu/actions-runner/_work/ClickHouse/ClickHouse/s3/../s3/tests/zero_copy_replication.py", line 2148, in aws_s3
    outline()
  File "/home/ubuntu/actions-runner/_work/ClickHouse/ClickHouse/s3/../s3/tests/zero_copy_replication.py", line 2133, in outline
    scenario()
  File "/home/ubuntu/actions-runner/_work/ClickHouse/ClickHouse/s3/../s3/tests/zero_copy_replication.py", line 1385, in delete_all
    assert (
AssertionError: Oops! Assertion failed

The following assertion was not satisfied
  assert (
      get_bucket_size(
          name=bucket_name,
          prefix=bucket_path,
          minio_enabled=minio_enabled,
          access_key=self.context.secret_access_key,
          key_id=self.context.access_key_id,
      )
      > size_before
  ), error()

Assertion values
  assert (
      get_bucket_size(
          name=bucket_name,
               ^ is 'altinity-qa-test'
          prefix=bucket_path,
          minio_enabled=minio_enabled,
          access_key=self.context.secret_access_key,
          key_id=self.context.access_key_id,
      )
      > size_before
  ), error()
  assert (
      get_bucket_size(
          name=bucket_name,
          prefix=bucket_path,
                 ^ is 'data/zero-copy-replication'
          minio_enabled=minio_enabled,
          access_key=self.context.secret_access_key,
          key_id=self.context.access_key_id,
      )
      > size_before
  ), error()
  assert (
      get_bucket_size(
          name=bucket_name,
          prefix=bucket_path,
          minio_enabled=minio_enabled,
                        ^ is False
          access_key=self.context.secret_access_key,
          key_id=self.context.access_key_id,
      )
      > size_before
  ), error()
  assert (
      get_bucket_size(
          name=bucket_name,
          prefix=bucket_path,
          minio_enabled=minio_enabled,
          access_key=self.context.secret_access_key,
                     ^ is 
          key_id=self.context.access_key_id,
      )
      > size_before
  ), error()
  assert (
      get_bucket_size(
          name=bucket_name,
          prefix=bucket_path,
          minio_enabled=minio_enabled,
          access_key=self.context.secret_access_key,
                     ^ is = 
          key_id=self.context.access_key_id,
      )
      > size_before
  ), error()
  assert (
      get_bucket_size(
          name=bucket_name,
          prefix=bucket_path,
          minio_enabled=minio_enabled,
          access_key=self.context.secret_access_key,
                     ^ is 
          key_id=self.context.access_key_id,
      )
      > size_before
  ), error()
  assert (
      get_bucket_size(
          name=bucket_name,
          prefix=bucket_path,
          minio_enabled=minio_enabled,
          access_key=self.context.secret_access_key,
                     ^ is = '[masked]:Secret(name='aws_s3_access_key')'
          key_id=self.context.access_key_id,
      )
      > size_before
  ), error()
  assert (
      get_bucket_size(
          name=bucket_name,
          prefix=bucket_path,
          minio_enabled=minio_enabled,
          access_key=self.context.secret_access_key,
          key_id=self.context.access_key_id,
                 ^ is 
      )
      > size_before
  ), error()
  assert (
      get_bucket_size(
          name=bucket_name,
          prefix=bucket_path,
          minio_enabled=minio_enabled,
          access_key=self.context.secret_access_key,
          key_id=self.context.access_key_id,
                 ^ is = 
      )
      > size_before
  ), error()
  assert (
      get_bucket_size(
          name=bucket_name,
          prefix=bucket_path,
          minio_enabled=minio_enabled,
          access_key=self.context.secret_access_key,
          key_id=self.context.access_key_id,
                 ^ is 
      )
      > size_before
  ), error()
  assert (
      get_bucket_size(
          name=bucket_name,
          prefix=bucket_path,
          minio_enabled=minio_enabled,
          access_key=self.context.secret_access_key,
          key_id=self.context.access_key_id,
                 ^ is = '[masked]:Secret(name='aws_s3_key_id')'
      )
      > size_before
  ), error()
  assert (
      get_bucket_size(
      ^ is = 0
          name=bucket_name,
          prefix=bucket_path,
          minio_enabled=minio_enabled,
          access_key=self.context.secret_access_key,
          key_id=self.context.access_key_id,
      )
      > size_before
  ), error()
  assert (
      get_bucket_size(
          name=bucket_name,
          prefix=bucket_path,
          minio_enabled=minio_enabled,
          access_key=self.context.secret_access_key,
          key_id=self.context.access_key_id,
      )
      > size_before
        ^ is 0
  ), error()
  assert (
      get_bucket_size(
          name=bucket_name,
          prefix=bucket_path,
          minio_enabled=minio_enabled,
          access_key=self.context.secret_access_key,
          key_id=self.context.access_key_id,
      )
      > size_before
      ^ is = False
  ), error()
  assert (
  ^ is False
      get_bucket_size(
          name=bucket_name,
          prefix=bucket_path,
          minio_enabled=minio_enabled,
          access_key=self.context.secret_access_key,
          key_id=self.context.access_key_id,
      )
      > size_before
  ), error()

Where
  File '/home/ubuntu/actions-runner/_work/ClickHouse/ClickHouse/s3/../s3/tests/zero_copy_replication.py', line 1394 in 'delete_all'

1386\|                      get_bucket_size(
1387\|                          name=bucket_name,
1388\|                          prefix=bucket_path,
1389\|                          minio_enabled=minio_enabled,
1390\|                          access_key=self.context.secret_access_key,
1391\|                          key_id=self.context.access_key_id,
1392\|                      )
1393\|                      > size_before
1394\|>                 ), error()
1395\|  
1396\|          finally:
1397\|              with Finally("I drop the table on each node"):
/s3/aws s3 zero copy replication/insert multiple replicasXFail 54s 263ms
Under investigation
AssertionError
Traceback (most recent call last):
  File "/home/ubuntu/actions-runner/_work/ClickHouse/ClickHouse/s3/regression.py", line 486, in 
    regression()
  File "/home/ubuntu/actions-runner/_work/ClickHouse/ClickHouse/s3/regression.py", line 464, in regression
    aws_s3_regression(
  File "/home/ubuntu/actions-runner/_work/ClickHouse/ClickHouse/s3/regression.py", line 353, in aws_s3_regression
    Feature(test=load("s3.tests.zero_copy_replication", "aws_s3"))(
  File "/home/ubuntu/actions-runner/_work/ClickHouse/ClickHouse/s3/../s3/tests/zero_copy_replication.py", line 2148, in aws_s3
    outline()
  File "/home/ubuntu/actions-runner/_work/ClickHouse/ClickHouse/s3/../s3/tests/zero_copy_replication.py", line 2133, in outline
    scenario()
  File "/home/ubuntu/actions-runner/_work/ClickHouse/ClickHouse/s3/../s3/tests/zero_copy_replication.py", line 1195, in insert_multiple_replicas
    assert added_size >= expected * 0.99, error()
AssertionError: Oops! Assertion failed

The following assertion was not satisfied
  assert added_size >= expected * 0.99, error()

Assertion values
  assert added_size >= expected * 0.99, error()
         ^ is 0
  assert added_size >= expected * 0.99, error()
                       ^ is 6306510
  assert added_size >= expected * 0.99, error()
                                ^ is = 6243444.9
  assert added_size >= expected * 0.99, error()
                    ^ is = False
  assert added_size >= expected * 0.99, error()
  ^ is False

Where
  File '/home/ubuntu/actions-runner/_work/ClickHouse/ClickHouse/s3/../s3/tests/zero_copy_replication.py', line 1195 in 'insert_multiple_replicas'

1187\|                      name=bucket_name,
1188\|                      prefix=bucket_path,
1189\|                      minio_enabled=minio_enabled,
1190\|                      access_key=self.context.secret_access_key,
1191\|                      key_id=self.context.access_key_id,
1192\|                  )
1193\|                  added_size = current_size - size_before
1194\|  
1195\|>                 assert added_size >= expected * 0.99, error()
1196\|                  assert added_size <= expected * 1.01, error()
1197\|  
1198\|          finally:

Results

Test Name Result Duration
/s3 OK 1h 5m
/s3/aws s3 table function OK 3m 44s
/s3/aws s3 table function/auto OK 11s 752ms
/s3/aws s3 table function/compression OK 11s 617ms
/s3/aws s3 table function/credentials OK 1s 301ms
/s3/aws s3 table function/data format OK 10s 447ms
/s3/aws s3 table function/multipart OK 5s 643ms
/s3/aws s3 table function/multiple columns OK 1s 539ms
/s3/aws s3 table function/partition OK 1s 245ms
/s3/aws s3 table function/remote host filter OK 40s 476ms
/s3/aws s3 table function/syntax OK 1s 285ms
/s3/aws s3 table function/wildcard OK 17s 498ms
/s3/aws s3 table function/ssec encryption check Skip 1ms
/s3/aws s3 table function/ssec OK 2m 1s
/s3/aws s3 table function/ssec/auto OK 8s 34ms
/s3/aws s3 table function/ssec/compression OK 7s 956ms
/s3/aws s3 table function/ssec/credentials OK 2s 259ms
/s3/aws s3 table function/ssec/data format OK 29s 760ms
/s3/aws s3 table function/ssec/multipart OK 5s 394ms
/s3/aws s3 table function/ssec/multiple columns OK 2s 576ms
/s3/aws s3 table function/ssec/partition OK 4s 329ms
/s3/aws s3 table function/ssec/remote host filter OK 39s 509ms
/s3/aws s3 table function/ssec/syntax OK 2s 336ms
/s3/aws s3 table function/ssec/wildcard OK 3s 546ms
/s3/aws s3 invalid table function OK 2s 675ms
/s3/aws s3 invalid table function/empty path OK 284ms
/s3/aws s3 invalid table function/empty structure OK 324ms
/s3/aws s3 invalid table function/invalid compression OK 276ms
/s3/aws s3 invalid table function/invalid credentials OK 415ms
/s3/aws s3 invalid table function/invalid format OK 776ms
/s3/aws s3 invalid table function/invalid path OK 288ms
/s3/aws s3 invalid table function/invalid structure OK 290ms
/s3/aws s3 disk OK 34m 2s
/s3/aws s3 disk/access OK 42s 152ms
/s3/aws s3 disk/access skip check OK 42s 304ms
/s3/aws s3 disk/add storage OK 1m 27s
/s3/aws s3 disk/alter move OK 57s 16ms
/s3/aws s3 disk/alter on cluster modify ttl OK 1m 38s
/s3/aws s3 disk/cache XFail 1ms
/s3/aws s3 disk/cache default XFail 719us
/s3/aws s3 disk/cache path XFail 739us
/s3/aws s3 disk/compact parts OK 42s 917ms
/s3/aws s3 disk/config over restart OK 1m 19s
/s3/aws s3 disk/default move factor OK 1m 4s
/s3/aws s3 disk/delete OK 4m 0s
/s3/aws s3 disk/download appropriate disk OK 1m 14s
/s3/aws s3 disk/environment credentials OK 1m 19s
/s3/aws s3 disk/exports OK 42s 683ms
/s3/aws s3 disk/generic url XFail 1ms
/s3/aws s3 disk/imports OK 42s 580ms
/s3/aws s3 disk/low cardinality offset XFail 40s 850ms
/s3/aws s3 disk/max single part upload size syntax OK 49s 832ms
/s3/aws s3 disk/mergetree OK 3m 29s
/s3/aws s3 disk/mergetree collapsing OK 52s 665ms
/s3/aws s3 disk/mergetree versionedcollapsing OK 54s 400ms
/s3/aws s3 disk/metadata OK 50s 932ms
/s3/aws s3 disk/min bytes for seek syntax OK 44s 356ms
/s3/aws s3 disk/multiple storage OK 54s 910ms
/s3/aws s3 disk/multiple storage query OK 55s 443ms
/s3/aws s3 disk/perform ttl move on insert OK 1m 25s
/s3/aws s3 disk/perform ttl move on insert default OK 50s 577ms
/s3/aws s3 disk/performance ttl move OK 1m 8s
/s3/aws s3 disk/remote host filter OK 1m 21s
/s3/aws s3 disk/restart OK 18s 146ms
/s3/aws s3 disk/specific url OK 39s 445ms
/s3/aws s3 disk/syntax OK 48s 897ms
/s3/aws s3 disk/wide parts OK 42s 319ms
/s3/aws s3 disk/ssec Skip 567us
/s3/aws s3 sanity OK 54s 619ms
/s3/aws s3 sanity/sanity OK 12s 801ms
/s3/aws s3 invalid disk OK 1m 57s
/s3/aws s3 invalid disk/access default OK 17s 705ms
/s3/aws s3 invalid disk/access failed OK 18s 723ms
/s3/aws s3 invalid disk/access failed skip check OK 42s 274ms
/s3/aws s3 invalid disk/cache path conflict XFail 1ms
/s3/aws s3 invalid disk/empty endpoint OK 9s 651ms
/s3/aws s3 invalid disk/invalid endpoint OK 9s 639ms
/s3/aws s3 invalid disk/invalid type OK 19s 337ms
/s3/aws s3 zero copy replication OK 18m 52s
/s3/aws s3 zero copy replication/add replica XFail 52s 293ms
/s3/aws s3 zero copy replication/alter XFail 1m 55s
/s3/aws s3 zero copy replication/alter repeat XFail 1m 52s
/s3/aws s3 zero copy replication/default value OK 12s 654ms
/s3/aws s3 zero copy replication/delete OK 51s 944ms
/s3/aws s3 zero copy replication/delete all XFail 57s 188ms
/s3/aws s3 zero copy replication/drop alter replica OK 1m 8s
/s3/aws s3 zero copy replication/drop replica OK 1m 8s
/s3/aws s3 zero copy replication/global setting OK 53s 602ms
/s3/aws s3 zero copy replication/insert multiple replicas XFail 54s 263ms
/s3/aws s3 zero copy replication/lost data during mutation OK 3s 944ms
/s3/aws s3 zero copy replication/metadata OK 49s 514ms
/s3/aws s3 zero copy replication/performance alter OK 1m 32s
/s3/aws s3 zero copy replication/performance insert OK 1m 27s
/s3/aws s3 zero copy replication/performance select OK 1m 32s
/s3/aws s3 zero copy replication/ttl delete OK 57s 604ms
/s3/aws s3 zero copy replication/ttl move OK 59s 949ms
/s3/aws s3 reconnect OK 2m 33s
/s3/aws s3 reconnect/local and s3 disk OK 52s 864ms
/s3/aws s3 reconnect/local and s3 volumes OK 48s 60ms
/s3/aws s3 reconnect/s3 disk OK 52s 835ms
/s3/aws s3 backup Skip 1ms

Generated by TestFlows Open-Source Test Framework v1.9.230315.1003122