* master: (43 commits)
Publish list of tests that failed the concurrent fast test
fix test
Use gnu++2a instead of c++2a for unbundled build to fix numeric_limits<__int128>
fix test
Update 01453_fixsed_string_sort.sql
Added test.
Fix ColumnString::updatePermutationWithCollation.
Add support for extended precision integers and decimals (#13097)
remove retries
Check that ya.make files are auto-generated
Fix "Arcadia"
ISSUES-4006 trigger CI again
fixed typo arrayCompact
Stratify nans comparison in arrayCompact function
Remove even more useless code
Remove useless code around zkutil
Fix 00956_sensitive_data_masking flackiness
Update docker/test/stress/run.sh
add exclusive DDLGuard for database
Fix handling embedded config.
...
00956_sensitive_data_masking is still flacky even after #13748 [1].
[1]: https://clickhouse-test-reports.s3.yandex.net/10373/348ef1256ea8fb8f61109c33bbdd28daf46bdc8e/functional_stateless_tests_(debug).html#fail1
The problem is that it uses the following pattern:
clickhouse-client -q ... & # run some query in background
clickhouse-client -q 'show processlist' > log
grep background-query log
But there is no guarantee that the query in background will be executed before `show processlist`:
2020.08.18 02:52:47.916386 [ 26788 ] {98c36d38-f710-4dfb-af8f-61906abc163c} <Debug> executeQuery: (from [::1]:51650) SHOW PROCESSLIST
...
2020.08.18 02:52:47.926854 [ 26756 ] {086c64fa-713b-4f8c-b702-23bfea10a49c} <Debug> executeQuery: (from [::1]:51652) select count() from system.numbers where ignore('find_me_[hidden]')=0 and ignore('fwerkh_that_magic_string_make_me_unique') = 0 FORMAT Null
Fix the test by waiting until the query in backgroud will start, and use
limited numbers + sleepEachRow over system.numbers to reduce CPU usage.
With default index_granularity=8096 CH always reads at least that number
of rows per selected part, but limit is checked afterwards. The optimization
that interrupts execution of queries based on approx number of rows to read
breaks the test. This means that test case is potetntially contains incorrect
limits.