From 749610acc297acd2825a90b0d310b36cc215393a Mon Sep 17 00:00:00 2001 From: sundyli <543950155@qq.com> Date: Thu, 9 Jan 2025 15:05:06 +0800 Subject: [PATCH] fix(query): backport to 680 stick the created_by infos in parquet writer (#17223) * fix(query): stick the created_by infos in parquet writer (#17220) * fix(query): stick the created_by infos in parquet writer * fix(query): stick the created_by infos in parquet writer * fix(query): stick the created_by infos in parquet writer * fix(query): stick the created_by infos in parquet writer * fix(query): stick the created_by infos in parquet writer * fix(query): fix register function working with nullable scalar (#17217) * fix(query): fix register function working with nullable scalar * fix(query): fix register function working with nullable scalar * fix(query): increase pool * Update 19_0005_fuzz_cte.sh * Update mysql_source.rs * fix(query): fix register function working with nullable scalar * fix(query): rollback tests --- .../expression/src/register_vectorize.rs | 25 +++----- src/query/functions/tests/it/scalars/mod.rs | 40 ++++++++++++ .../append/parquet_file/writer_processor.rs | 18 +++++- .../suites/query/cte/basic_r_cte.test | 61 ++++++++++++++++++- .../formats/parquet/options/null_if.test | 2 +- .../parquet/options/parquet_missing_uuid.test | 4 +- .../05_hints/05_0001_set_var.result | 2 +- .../18_rbac/18_0002_ownership_cover.result | 2 +- .../18_rbac/18_0007_privilege_access.result | 2 +- .../19_fuzz/19_0005_fuzz_cte.result | 1 + .../0_stateless/19_fuzz/19_0005_fuzz_cte.sh | 25 ++++++++ ..._0015_compact_hook_stas_issue_13947.result | 2 +- .../00_stage/00_0001_copy_into_stage.result | 2 +- .../00_0005_copy_into_location.result | 2 +- .../00_stage/00_0012_stage_priv.result | 2 +- .../00_0012_stage_with_connection.result | 2 +- .../00_stage/00_0015_unload_output.result | 32 +++++----- .../05_05_01_parquet_load_unload.result | 4 +- .../08_00_parquet/08_00_00_basic.result | 6 +- .../08_00_parquet/08_00_06_transform.result | 2 +- 20 files changed, 185 insertions(+), 51 deletions(-) create mode 100755 tests/suites/0_stateless/19_fuzz/19_0005_fuzz_cte.result create mode 100755 tests/suites/0_stateless/19_fuzz/19_0005_fuzz_cte.sh diff --git a/src/query/expression/src/register_vectorize.rs b/src/query/expression/src/register_vectorize.rs index 794e2c1dc1ee4..a5332be3a084d 100755 --- a/src/query/expression/src/register_vectorize.rs +++ b/src/query/expression/src/register_vectorize.rs @@ -283,8 +283,7 @@ pub fn passthrough_nullable_1_arg( match out { Value::Column(out) => Value::Column(NullableColumn::new(out, args_validity)), - Value::Scalar(out) if args_validity.get_bit(0) => Value::Scalar(Some(out)), - _ => Value::Scalar(None), + Value::Scalar(out) => Value::Scalar(Some(out)), } } _ => Value::Scalar(None), @@ -308,6 +307,7 @@ pub fn passthrough_nullable_2_arg( if let Some(validity) = ctx.validity.as_ref() { args_validity = &args_validity & validity; } + ctx.validity = Some(args_validity.clone()); match (arg1.value(), arg2.value()) { (Some(arg1), Some(arg2)) => { @@ -315,8 +315,7 @@ pub fn passthrough_nullable_2_arg( match out { Value::Column(out) => Value::Column(NullableColumn::new(out, args_validity)), - Value::Scalar(out) if args_validity.get_bit(0) => Value::Scalar(Some(out)), - _ => Value::Scalar(None), + Value::Scalar(out) => Value::Scalar(Some(out)), } } _ => Value::Scalar(None), @@ -352,8 +351,7 @@ pub fn passthrough_nullable_3_arg Value::Column(NullableColumn::new(out, args_validity)), - Value::Scalar(out) if args_validity.get_bit(0) => Value::Scalar(Some(out)), - _ => Value::Scalar(None), + Value::Scalar(out) => Value::Scalar(Some(out)), } } _ => Value::Scalar(None), @@ -397,8 +395,7 @@ pub fn passthrough_nullable_4_arg< match out { Value::Column(out) => Value::Column(NullableColumn::new(out, args_validity)), - Value::Scalar(out) if args_validity.get_bit(0) => Value::Scalar(Some(out)), - _ => Value::Scalar(None), + Value::Scalar(out) => Value::Scalar(Some(out)), } } _ => Value::Scalar(None), @@ -427,8 +424,7 @@ pub fn combine_nullable_1_arg( out.column, &args_validity & &out.validity, )), - Value::Scalar(out) if args_validity.get_bit(0) => Value::Scalar(out), - _ => Value::Scalar(None), + Value::Scalar(out) => Value::Scalar(out), } } _ => Value::Scalar(None), @@ -465,8 +461,7 @@ pub fn combine_nullable_2_arg( out.column, &args_validity & &out.validity, )), - Value::Scalar(out) if args_validity.get_bit(0) => Value::Scalar(out), - _ => Value::Scalar(None), + Value::Scalar(out) => Value::Scalar(out), } } _ => Value::Scalar(None), @@ -505,8 +500,7 @@ pub fn combine_nullable_3_arg out.column, &args_validity & &out.validity, )), - Value::Scalar(out) if args_validity.get_bit(0) => Value::Scalar(out), - _ => Value::Scalar(None), + Value::Scalar(out) => Value::Scalar(out), } } _ => Value::Scalar(None), @@ -552,8 +546,7 @@ pub fn combine_nullable_4_arg Value::Scalar(out), - _ => Value::Scalar(None), + Value::Scalar(out) => Value::Scalar(out), } } _ => Value::Scalar(None), diff --git a/src/query/functions/tests/it/scalars/mod.rs b/src/query/functions/tests/it/scalars/mod.rs index 7f8c5344df4c0..770961db74f04 100644 --- a/src/query/functions/tests/it/scalars/mod.rs +++ b/src/query/functions/tests/it/scalars/mod.rs @@ -271,3 +271,43 @@ fn list_all_builtin_functions() { fn check_ambiguity() { BUILTIN_FUNCTIONS.check_ambiguity() } + +#[test] +fn test_if_function() -> Result<()> { + use databend_common_expression::types::*; + use databend_common_expression::FromData; + use databend_common_expression::Scalar; + let raw_expr = parser::parse_raw_expr("if(eq(n,1), sum_sid + 1,100)", &[ + ("n", UInt8Type::data_type()), + ("sum_sid", Int32Type::data_type().wrap_nullable()), + ]); + let expr = type_check::check(&raw_expr, &BUILTIN_FUNCTIONS)?; + let block = DataBlock::new( + vec![ + BlockEntry { + data_type: UInt8Type::data_type(), + value: Value::Column(UInt8Type::from_data(vec![2_u8, 1])), + }, + BlockEntry { + data_type: Int32Type::data_type().wrap_nullable(), + value: Value::Scalar(Scalar::Number(NumberScalar::Int32(2400_i32))), + }, + ], + 2, + ); + let func_ctx = FunctionContext::default(); + let evaluator = Evaluator::new(&block, &func_ctx, &BUILTIN_FUNCTIONS); + let result = evaluator.run(&expr).unwrap(); + let result = result + .as_column() + .unwrap() + .clone() + .as_nullable() + .unwrap() + .clone(); + + let bm = Bitmap::from_iter([true, true]); + assert_eq!(result.validity, bm); + assert_eq!(result.column, Int64Type::from_data(vec![100, 2401])); + Ok(()) +} diff --git a/src/query/storages/stage/src/append/parquet_file/writer_processor.rs b/src/query/storages/stage/src/append/parquet_file/writer_processor.rs index 40101d668edcb..54b1d11d19052 100644 --- a/src/query/storages/stage/src/append/parquet_file/writer_processor.rs +++ b/src/query/storages/stage/src/append/parquet_file/writer_processor.rs @@ -71,11 +71,27 @@ pub struct ParquetFileWriter { const MAX_BUFFER_SIZE: usize = 64 * 1024 * 1024; // this is number of rows, not size const MAX_ROW_GROUP_SIZE: usize = 1024 * 1024; +const CREATE_BY_LEN: usize = 24; // "Databend 1.2.333-nightly".len(); fn create_writer( arrow_schema: Arc, targe_file_size: Option, ) -> Result>> { + // example: 1.2.333-nightly + // tags may contain other items like `1.2.680-p2`, we will fill it with `1.2.680-p2.....` + let mut create_by = format!( + "Databend {}.{}.{}-{:.<7}", + DATABEND_SEMVER.major, + DATABEND_SEMVER.minor, + DATABEND_SEMVER.patch, + DATABEND_SEMVER.pre.as_str() + ); + + if create_by.len() != CREATE_BY_LEN { + create_by = format!("{:.<24}", create_by); + create_by.truncate(24); + } + let props = WriterProperties::builder() .set_compression(TableCompression::Zstd.into()) .set_max_row_group_size(MAX_ROW_GROUP_SIZE) @@ -83,7 +99,7 @@ fn create_writer( .set_dictionary_enabled(false) .set_statistics_enabled(EnabledStatistics::Chunk) .set_bloom_filter_enabled(false) - .set_created_by(format!("Databend {}", *DATABEND_SEMVER)) + .set_created_by(create_by) .build(); let buf_size = match targe_file_size { Some(n) if n < MAX_BUFFER_SIZE => n, diff --git a/tests/sqllogictests/suites/query/cte/basic_r_cte.test b/tests/sqllogictests/suites/query/cte/basic_r_cte.test index 12cd5a84b74b1..1d4ce93efcd9d 100644 --- a/tests/sqllogictests/suites/query/cte/basic_r_cte.test +++ b/tests/sqllogictests/suites/query/cte/basic_r_cte.test @@ -227,5 +227,64 @@ select cte1.a from cte1; 8 9 + +statement ok +create table train( +train_id varchar(8) not null , +departure_station varchar(32) not null, +arrival_station varchar(32) not null, +seat_count int not null +); + +statement ok +create table passenger( +passenger_id varchar(16) not null, +departure_station varchar(32) not null, +arrival_station varchar(32) not null +); + +statement ok +create table city(city varchar(32)); + +statement ok +insert into city +with t as (select 1 n union select 2 union select 3 union select 4 union select 5) +,t1 as(select row_number()over() rn from t ,t t2,t t3) +select concat('城市',rn::varchar) city from t1 where rn<=5; + +statement ok +insert into train +select concat('G',row_number()over()::varchar),c1.city,c2.city, n from city c1, city c2, (select 600 n union select 800 union select 1200 union select 1600) a ; + +statement ok +insert into passenger +select concat('P',substr((100000000+row_number()over())::varchar,2)),c1.city,c2.city from city c1, city c2 ,city c3, city c4, city c5, +city c6, (select 1 n union select 2 union select 3 union select 4) c7,(select 1 n union select 2) c8; + + +query III +with +t0 as ( +select + train_id, + seat_count, + sum(seat_count) over ( + partition by departure_station, arrival_station order by train_id + ) ::int sum_sid +from + train +) +select + sum(case when n=1 then sum_sid+1 else 0 end::int), + sum(sum_sid), + sum(seat_count) +from + t0,(select 1 n union all select 2); +---- +261700 523200 210000 + +statement ok +use default; + statement ok -drop table t1; +drop database db; diff --git a/tests/sqllogictests/suites/stage/formats/parquet/options/null_if.test b/tests/sqllogictests/suites/stage/formats/parquet/options/null_if.test index 821b66324545e..fee31e3da19b3 100644 --- a/tests/sqllogictests/suites/stage/formats/parquet/options/null_if.test +++ b/tests/sqllogictests/suites/stage/formats/parquet/options/null_if.test @@ -26,7 +26,7 @@ remove @data/unload/parquet/null_if/ query copy into @data/unload/parquet/null_if from string ---- -3 56 379 +3 56 387 statement ok drop file format if exists parquet_null_if diff --git a/tests/sqllogictests/suites/stage/formats/parquet/options/parquet_missing_uuid.test b/tests/sqllogictests/suites/stage/formats/parquet/options/parquet_missing_uuid.test index faaed4301cf0a..f7079fb7435f9 100644 --- a/tests/sqllogictests/suites/stage/formats/parquet/options/parquet_missing_uuid.test +++ b/tests/sqllogictests/suites/stage/formats/parquet/options/parquet_missing_uuid.test @@ -10,7 +10,7 @@ remove @data/parquet/unload/uuid query copy into @data/parquet/unload/uuid/ from (select 1 as a) file_format = (type = parquet) ---- -1 1 366 +1 1 374 query error column id doesn't exist copy into t_uuid from @data/parquet/unload/uuid file_format = (type = parquet) RETURN_FAILED_ONLY=TRUE @@ -22,7 +22,7 @@ select * from t_uuid query copy into @data/parquet/unload/uuid/ from (select 1 as a) file_format = (type = parquet) ---- -1 1 366 +1 1 374 statement ok truncate table t_uuid diff --git a/tests/suites/0_stateless/05_hints/05_0001_set_var.result b/tests/suites/0_stateless/05_hints/05_0001_set_var.result index 2393abe29bea2..45cbaeaebadff 100644 --- a/tests/suites/0_stateless/05_hints/05_0001_set_var.result +++ b/tests/suites/0_stateless/05_hints/05_0001_set_var.result @@ -23,5 +23,5 @@ America/Toronto 1 2022-02-02 03:00:00 2022-02-02 03:00:00 -1 13 419 +1 13 427 Asia/Shanghai diff --git a/tests/suites/0_stateless/18_rbac/18_0002_ownership_cover.result b/tests/suites/0_stateless/18_rbac/18_0002_ownership_cover.result index 3efd2f9f36662..bf32053ef5ed1 100644 --- a/tests/suites/0_stateless/18_rbac/18_0002_ownership_cover.result +++ b/tests/suites/0_stateless/18_rbac/18_0002_ownership_cover.result @@ -2,7 +2,7 @@ 1 200 === test stage === -1 8 392 +1 8 400 0 === test udf === 2 diff --git a/tests/suites/0_stateless/18_rbac/18_0007_privilege_access.result b/tests/suites/0_stateless/18_rbac/18_0007_privilege_access.result index 17ca3c4e09d76..fd9a369f184f9 100644 --- a/tests/suites/0_stateless/18_rbac/18_0007_privilege_access.result +++ b/tests/suites/0_stateless/18_rbac/18_0007_privilege_access.result @@ -101,7 +101,7 @@ Error: APIError: QueryFailed: [1063]Permission denied: privilege READ is require Error: APIError: QueryFailed: [1063]Permission denied: No privilege on database root_db for user b. Error: APIError: QueryFailed: [1063]Permission denied: No privilege on table root_table for user b. Error: APIError: QueryFailed: [1063]Permission denied: No privilege on table root_table for user b. -1 1 366 +1 1 374 Error: APIError: QueryFailed: [1063]Permission denied: privilege [Select] is required on 'default'.'default'.'t1' for user 'b'@'%' with roles [public] Error: APIError: QueryFailed: [1063]Permission denied: privilege [Read] is required on STAGE s3 for user 'b'@'%' with roles [public]. Note: Please ensure that your current role have the appropriate permissions to create a new Warehouse|Database|Table|UDF|Stage. Error: APIError: QueryFailed: [1063]Permission denied: privilege [Select] is required on 'default'.'default'.'t' for user 'b'@'%' with roles [public] diff --git a/tests/suites/0_stateless/19_fuzz/19_0005_fuzz_cte.result b/tests/suites/0_stateless/19_fuzz/19_0005_fuzz_cte.result new file mode 100755 index 0000000000000..d86bac9de59ab --- /dev/null +++ b/tests/suites/0_stateless/19_fuzz/19_0005_fuzz_cte.result @@ -0,0 +1 @@ +OK diff --git a/tests/suites/0_stateless/19_fuzz/19_0005_fuzz_cte.sh b/tests/suites/0_stateless/19_fuzz/19_0005_fuzz_cte.sh new file mode 100755 index 0000000000000..8b3d81efb4a20 --- /dev/null +++ b/tests/suites/0_stateless/19_fuzz/19_0005_fuzz_cte.sh @@ -0,0 +1,25 @@ +#!/usr/bin/env bash + +CURDIR=$(cd "$(dirname "${BASH_SOURCE[0]}")" && pwd) +. "$CURDIR"/../../../shell_env.sh + + +times=256 + +echo "" > /tmp/fuzz_a.txt +echo "" > /tmp/fuzz_b.txt + +for i in `seq 1 ${times}`;do + echo """with t0(sum_sid) as (select sum(number) over(partition by number order by number) + from numbers(3)) select n, if(n =1, sum_sid +1, 0) from t0, (select 1 n union all select 2) order by 1,2; + """ | $BENDSQL_CLIENT_CONNECT >> /tmp/fuzz_a.txt +done + + +for i in `seq 1 ${times}`;do + echo """with t0(sum_sid) as (select sum(number) over(partition by number order by number) + from numbers(3)) select n, if(n =1, sum_sid +1, 0) from t0, (select 1 n union all select 2) order by 1,2; + """ | $BENDSQL_CLIENT_CONNECT >> /tmp/fuzz_b.txt +done + +diff /tmp/fuzz_a.txt /tmp/fuzz_b.txt && echo "OK" diff --git a/tests/suites/0_stateless/20+_others/20_0015_compact_hook_stas_issue_13947.result b/tests/suites/0_stateless/20+_others/20_0015_compact_hook_stas_issue_13947.result index 9bbaafbbe5a52..6887d3cff3d73 100644 --- a/tests/suites/0_stateless/20+_others/20_0015_compact_hook_stas_issue_13947.result +++ b/tests/suites/0_stateless/20+_others/20_0015_compact_hook_stas_issue_13947.result @@ -1,6 +1,6 @@ 1 1 -2 10 391 +2 10 399 expects .stats.write_progress.rows be 2 expects .error be null 2 diff --git a/tests/suites/1_stateful/00_stage/00_0001_copy_into_stage.result b/tests/suites/1_stateful/00_stage/00_0001_copy_into_stage.result index 4bc3917c2c82f..b4beef56607e2 100644 --- a/tests/suites/1_stateful/00_stage/00_0001_copy_into_stage.result +++ b/tests/suites/1_stateful/00_stage/00_0001_copy_into_stage.result @@ -9,6 +9,6 @@ 2 2 20 160 160 -20 530 810 +20 530 818 2 20 160 160 diff --git a/tests/suites/1_stateful/00_stage/00_0005_copy_into_location.result b/tests/suites/1_stateful/00_stage/00_0005_copy_into_location.result index 62407671ba6f3..efa2237d17e29 100755 --- a/tests/suites/1_stateful/00_stage/00_0005_copy_into_location.result +++ b/tests/suites/1_stateful/00_stage/00_0005_copy_into_location.result @@ -1,4 +1,4 @@ >>>> create or replace connection c_00_0005 storage_type='s3' access_key_id = 'minioadmin' endpoint_url = 'http://127.0.0.1:9900' secret_access_key = 'minioadmin' >>>> copy into 's3://testbucket/c_00_0005/ab de/f' connection=(connection_name='c_00_0005') from (select 1) detailed_output=true use_raw_path=true single=true overwrite=true -c_00_0005/ab de/f 366 1 +c_00_0005/ab de/f 374 1 <<<< diff --git a/tests/suites/1_stateful/00_stage/00_0012_stage_priv.result b/tests/suites/1_stateful/00_stage/00_0012_stage_priv.result index a82d89a6222fa..44ed5f3f4f70d 100644 --- a/tests/suites/1_stateful/00_stage/00_0012_stage_priv.result +++ b/tests/suites/1_stateful/00_stage/00_0012_stage_priv.result @@ -27,7 +27,7 @@ Error: APIError: QueryFailed: [1063]Permission denied: privilege [Write] is requ Error: APIError: QueryFailed: [1063]Permission denied: privilege [Read] is required on STAGE presign_stage for user 'u1'@'%' with roles [public]. Note: Please ensure that your current role have the appropriate permissions to create a new Warehouse|Database|Table|UDF|Stage. 000 Error: APIError: QueryFailed: [1063]Permission denied: privilege [Write] is required on STAGE s3 for user 'u1'@'%' with roles [public]. Note: Please ensure that your current role have the appropriate permissions to create a new Warehouse|Database|Table|UDF|Stage. -1 1 366 +1 1 374 Error: APIError: QueryFailed: [1063]Permission denied: privilege [Read] is required on STAGE s3 for user 'u1'@'%' with roles [public]. Note: Please ensure that your current role have the appropriate permissions to create a new Warehouse|Database|Table|UDF|Stage. Error: APIError: QueryFailed: [1063]Permission denied: privilege READ is required on stage s3 for user 'u1'@'%' Error: APIError: QueryFailed: [1063]Permission denied: privilege READ is required on stage s3 for user 'u1'@'%' diff --git a/tests/suites/1_stateful/00_stage/00_0012_stage_with_connection.result b/tests/suites/1_stateful/00_stage/00_0012_stage_with_connection.result index 49d22b11010bc..5fb48244de4a6 100644 --- a/tests/suites/1_stateful/00_stage/00_0012_stage_with_connection.result +++ b/tests/suites/1_stateful/00_stage/00_0012_stage_with_connection.result @@ -8,7 +8,7 @@ >>>> create stage my_stage url= 's3://testbucket/admin/tempdata/' connection = (connection_name='my_conn'); >>>> remove @my_stage; >>>> copy into @my_stage/a.csv from my_table -3 13 393 +3 13 401 >>>> select * from @my_stage order by a; 1 2 diff --git a/tests/suites/1_stateful/00_stage/00_0015_unload_output.result b/tests/suites/1_stateful/00_stage/00_0015_unload_output.result index 24a187fd5e3bd..56294498558f3 100644 --- a/tests/suites/1_stateful/00_stage/00_0015_unload_output.result +++ b/tests/suites/1_stateful/00_stage/00_0015_unload_output.result @@ -40,34 +40,34 @@ copy2 <<<< copy3 <<<< -a/bc/data_UUID_0000_00000000.parquet 385 1 -a/bc/data_UUID_0000_00000001.parquet 385 1 -a/bc/data_UUID_0000_00000002.parquet 385 1 -a/bc/data_UUID_0000_00000003.parquet 385 1 -a/bc/data_UUID_0000_00000004.parquet 385 1 -a/bc/data_UUID_0000_00000005.parquet 385 1 -a/bc/data_UUID_0000_00000006.parquet 385 1 -a/bc/data_UUID_0000_00000007.parquet 385 1 -a/bc/data_UUID_0000_00000008.parquet 385 1 -a/bc/data_UUID_0000_00000009.parquet 385 1 +a/bc/data_UUID_0000_00000000.parquet 393 1 +a/bc/data_UUID_0000_00000001.parquet 393 1 +a/bc/data_UUID_0000_00000002.parquet 393 1 +a/bc/data_UUID_0000_00000003.parquet 393 1 +a/bc/data_UUID_0000_00000004.parquet 393 1 +a/bc/data_UUID_0000_00000005.parquet 393 1 +a/bc/data_UUID_0000_00000006.parquet 393 1 +a/bc/data_UUID_0000_00000007.parquet 393 1 +a/bc/data_UUID_0000_00000008.parquet 393 1 +a/bc/data_UUID_0000_00000009.parquet 393 1 >>>> unload path >>>> copy /*+ set_var(max_threads=1) */ into @s1 from (select 1) detailed_output=true -data_UUID_0000_00000000.parquet 366 1 +data_UUID_0000_00000000.parquet 374 1 <<<< >>>> copy /*+ set_var(max_threads=1) */ into @s1/ from (select 1) detailed_output=true -data_UUID_0000_00000000.parquet 366 1 +data_UUID_0000_00000000.parquet 374 1 <<<< >>>> copy /*+ set_var(max_threads=1) */ into @s1/a from (select 1) detailed_output=true -a/data_UUID_0000_00000000.parquet 366 1 +a/data_UUID_0000_00000000.parquet 374 1 <<<< >>>> copy /*+ set_var(max_threads=1) */ into @s1/a/ from (select 1) detailed_output=true -a/data_UUID_0000_00000000.parquet 366 1 +a/data_UUID_0000_00000000.parquet 374 1 <<<< >>>> copy /*+ set_var(max_threads=1) */ into @s1/a/bc from (select 1) detailed_output=true -a/bc/data_UUID_0000_00000000.parquet 366 1 +a/bc/data_UUID_0000_00000000.parquet 374 1 <<<< >>>> copy /*+ set_var(max_threads=1) */ into @s1/a/data_ from (select 1) detailed_output=true -a/data_UUID_0000_00000000.parquet 366 1 +a/data_UUID_0000_00000000.parquet 374 1 <<<< >>>> drop stage if exists s1 >>>> drop table if exists t1 diff --git a/tests/suites/1_stateful/05_formats/05_05_parquet/05_05_01_parquet_load_unload.result b/tests/suites/1_stateful/05_formats/05_05_parquet/05_05_01_parquet_load_unload.result index 2bca090c6ecde..0066945e036c6 100644 --- a/tests/suites/1_stateful/05_formats/05_05_parquet/05_05_01_parquet_load_unload.result +++ b/tests/suites/1_stateful/05_formats/05_05_parquet/05_05_01_parquet_load_unload.result @@ -29,14 +29,14 @@ a"b 1 ['a"b'] {"k":"v"} 2044-05-06 10:25:02.868894 10.01 ('a',5) ['{"k":"v"}'] [ NULL 2 ['a'b'] [1] 2044-05-06 10:25:02.868894 -10.01 ('b',10) ['[1]'] [('b',10)] <<<< >>>> copy into @s1/unload1/ from test_load_unload -2 390 3088 +2 390 3096 >>>> truncate table test_load_unload >>>> copy into test_load_unload from @s1/unload1.parquet force=true; unload1.parquet 2 0 NULL NULL begin diff select end diff >>>> copy into @s1/unload2/ from test_load_unload -2 390 3088 +2 390 3096 begin diff parquet end diff >>>> truncate table test_load_unload diff --git a/tests/suites/1_stateful/08_select_stage/08_00_parquet/08_00_00_basic.result b/tests/suites/1_stateful/08_select_stage/08_00_parquet/08_00_00_basic.result index 2d0cfd3c5d16f..6d16d279d5e2e 100644 --- a/tests/suites/1_stateful/08_select_stage/08_00_parquet/08_00_00_basic.result +++ b/tests/suites/1_stateful/08_select_stage/08_00_parquet/08_00_00_basic.result @@ -1,9 +1,9 @@ 2 --- named internal stage -2 53 783 +2 53 791 1 2 3 4 5 6 -2 53 783 +2 53 791 --- external stage 1 2 3 4 5 6 @@ -12,6 +12,6 @@ 4 5 6 2 --- variant named internal stage -2 70 759 +2 70 767 1 [1,2,3] 2 {"k":"v"} diff --git a/tests/suites/1_stateful/08_select_stage/08_00_parquet/08_00_06_transform.result b/tests/suites/1_stateful/08_select_stage/08_00_parquet/08_00_06_transform.result index 273439b5dfd9a..a560d769a5c97 100755 --- a/tests/suites/1_stateful/08_select_stage/08_00_parquet/08_00_06_transform.result +++ b/tests/suites/1_stateful/08_select_stage/08_00_parquet/08_00_06_transform.result @@ -1,5 +1,5 @@ 2 -2 18 588 +2 18 596 --- copy from uri with transform 2 5