[branch-2.1][test](external)move hive cases from p2 to p0 (#37149)
pk (#36855) test_hive_same_db_table_name test_hive_special_char_partition test_complex_types test_wide_table
This commit is contained in:
@ -0,0 +1,138 @@
|
||||
create database if not exists multi_catalog;
|
||||
|
||||
use multi_catalog;
|
||||
|
||||
CREATE TABLE `byd`(
|
||||
`id` int,
|
||||
`capacity` array<double>,
|
||||
`singles` map<string,double>,
|
||||
`favor` struct<name:string,age:int,tip:timestamp>)
|
||||
ROW FORMAT SERDE
|
||||
'org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe'
|
||||
STORED AS INPUTFORMAT
|
||||
'org.apache.hadoop.hive.ql.io.parquet.MapredParquetInputFormat'
|
||||
OUTPUTFORMAT
|
||||
'org.apache.hadoop.hive.ql.io.parquet.MapredParquetOutputFormat'
|
||||
LOCATION
|
||||
'/user/doris/suites/multi_catalog/byd'
|
||||
TBLPROPERTIES (
|
||||
'transient_lastDdlTime'='1690356922');
|
||||
|
||||
msck repair table byd;
|
||||
|
||||
CREATE TABLE `complex_offsets_check`(
|
||||
`id` int,
|
||||
`array1` array<int>,
|
||||
`array2` array<array<int>>,
|
||||
`map1` map<string,int>,
|
||||
`struct1` struct<s1:string,s2:int>)
|
||||
ROW FORMAT SERDE
|
||||
'org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe'
|
||||
STORED AS INPUTFORMAT
|
||||
'org.apache.hadoop.hive.ql.io.parquet.MapredParquetInputFormat'
|
||||
OUTPUTFORMAT
|
||||
'org.apache.hadoop.hive.ql.io.parquet.MapredParquetOutputFormat'
|
||||
LOCATION
|
||||
'/user/doris/suites/multi_catalog/complex_offsets_check'
|
||||
TBLPROPERTIES (
|
||||
'transient_lastDdlTime'='1690974653');
|
||||
|
||||
msck repair table complex_offsets_check;
|
||||
|
||||
CREATE TABLE `parquet_all_types`(
|
||||
`t_null_string` string,
|
||||
`t_null_varchar` varchar(65535),
|
||||
`t_null_char` char(10),
|
||||
`t_null_decimal_precision_2` decimal(2,1),
|
||||
`t_null_decimal_precision_4` decimal(4,2),
|
||||
`t_null_decimal_precision_8` decimal(8,4),
|
||||
`t_null_decimal_precision_17` decimal(17,8),
|
||||
`t_null_decimal_precision_18` decimal(18,8),
|
||||
`t_null_decimal_precision_38` decimal(38,16),
|
||||
`t_empty_string` string,
|
||||
`t_string` string,
|
||||
`t_empty_varchar` varchar(65535),
|
||||
`t_varchar` varchar(65535),
|
||||
`t_varchar_max_length` varchar(65535),
|
||||
`t_char` char(10),
|
||||
`t_int` int,
|
||||
`t_bigint` bigint,
|
||||
`t_float` float,
|
||||
`t_double` double,
|
||||
`t_boolean_true` boolean,
|
||||
`t_boolean_false` boolean,
|
||||
`t_decimal_precision_2` decimal(2,1),
|
||||
`t_decimal_precision_4` decimal(4,2),
|
||||
`t_decimal_precision_8` decimal(8,4),
|
||||
`t_decimal_precision_17` decimal(17,8),
|
||||
`t_decimal_precision_18` decimal(18,8),
|
||||
`t_decimal_precision_38` decimal(38,16),
|
||||
`t_binary` binary,
|
||||
`t_map_string` map<string,string>,
|
||||
`t_map_varchar` map<varchar(65535),varchar(65535)>,
|
||||
`t_map_char` map<char(10),char(10)>,
|
||||
`t_map_int` map<int,int>,
|
||||
`t_map_bigint` map<bigint,bigint>,
|
||||
`t_map_float` map<float,float>,
|
||||
`t_map_double` map<double,double>,
|
||||
`t_map_boolean` map<boolean,boolean>,
|
||||
`t_map_decimal_precision_2` map<decimal(2,1),decimal(2,1)>,
|
||||
`t_map_decimal_precision_4` map<decimal(4,2),decimal(4,2)>,
|
||||
`t_map_decimal_precision_8` map<decimal(8,4),decimal(8,4)>,
|
||||
`t_map_decimal_precision_17` map<decimal(17,8),decimal(17,8)>,
|
||||
`t_map_decimal_precision_18` map<decimal(18,8),decimal(18,8)>,
|
||||
`t_map_decimal_precision_38` map<decimal(38,16),decimal(38,16)>,
|
||||
`t_array_string` array<string>,
|
||||
`t_array_int` array<int>,
|
||||
`t_array_bigint` array<bigint>,
|
||||
`t_array_float` array<float>,
|
||||
`t_array_double` array<double>,
|
||||
`t_array_boolean` array<boolean>,
|
||||
`t_array_varchar` array<varchar(65535)>,
|
||||
`t_array_char` array<char(10)>,
|
||||
`t_array_decimal_precision_2` array<decimal(2,1)>,
|
||||
`t_array_decimal_precision_4` array<decimal(4,2)>,
|
||||
`t_array_decimal_precision_8` array<decimal(8,4)>,
|
||||
`t_array_decimal_precision_17` array<decimal(17,8)>,
|
||||
`t_array_decimal_precision_18` array<decimal(18,8)>,
|
||||
`t_array_decimal_precision_38` array<decimal(38,16)>,
|
||||
`t_struct_bigint` struct<s_bigint:bigint>,
|
||||
`t_complex` map<string,array<struct<s_int:int>>>,
|
||||
`t_struct_nested` struct<struct_field:array<string>>,
|
||||
`t_struct_null` struct<struct_field_null:string,struct_field_null2:string>,
|
||||
`t_struct_non_nulls_after_nulls` struct<struct_non_nulls_after_nulls1:int,struct_non_nulls_after_nulls2:string>,
|
||||
`t_nested_struct_non_nulls_after_nulls` struct<struct_field1:int,struct_field2:string,strict_field3:struct<nested_struct_field1:int,nested_struct_field2:string>>,
|
||||
`t_map_null_value` map<string,string>,
|
||||
`t_array_string_starting_with_nulls` array<string>,
|
||||
`t_array_string_with_nulls_in_between` array<string>,
|
||||
`t_array_string_ending_with_nulls` array<string>,
|
||||
`t_array_string_all_nulls` array<string>)
|
||||
ROW FORMAT SERDE
|
||||
'org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe'
|
||||
STORED AS INPUTFORMAT
|
||||
'org.apache.hadoop.hive.ql.io.parquet.MapredParquetInputFormat'
|
||||
OUTPUTFORMAT
|
||||
'org.apache.hadoop.hive.ql.io.parquet.MapredParquetOutputFormat'
|
||||
LOCATION
|
||||
'/user/doris/suites/multi_catalog/parquet_all_types'
|
||||
TBLPROPERTIES (
|
||||
'transient_lastDdlTime'='1692347490');
|
||||
|
||||
msck repair table parquet_all_types;
|
||||
|
||||
CREATE TABLE `date_dict`(
|
||||
`date1` date,
|
||||
`date2` date,
|
||||
`date3` date)
|
||||
ROW FORMAT SERDE
|
||||
'org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe'
|
||||
STORED AS INPUTFORMAT
|
||||
'org.apache.hadoop.hive.ql.io.parquet.MapredParquetInputFormat'
|
||||
OUTPUTFORMAT
|
||||
'org.apache.hadoop.hive.ql.io.parquet.MapredParquetOutputFormat'
|
||||
LOCATION
|
||||
'/user/doris/suites/multi_catalog/date_dict'
|
||||
TBLPROPERTIES (
|
||||
'transient_lastDdlTime'='1693396885');
|
||||
|
||||
msck repair table date_dict;
|
||||
@ -0,0 +1,22 @@
|
||||
#!/bin/bash
|
||||
set -x
|
||||
|
||||
CUR_DIR="$(cd "$(dirname "${BASH_SOURCE[0]}")" &>/dev/null && pwd)"
|
||||
|
||||
if [[ ! -d "${CUR_DIR}/data" ]]; then
|
||||
echo "${CUR_DIR}/data does not exist"
|
||||
cd "${CUR_DIR}" && rm -f data.tar.gz \
|
||||
&& curl -O https://s3BucketName.s3Endpoint/regression/datalake/pipeline_data/multi_catalog/test_complex_types/data.tar.gz \
|
||||
&& tar xzf data.tar.gz
|
||||
cd -
|
||||
else
|
||||
echo "${CUR_DIR}/data exist, continue !"
|
||||
fi
|
||||
|
||||
## mkdir and put data to hdfs
|
||||
hadoop fs -mkdir -p /user/doris/suites/multi_catalog/
|
||||
hadoop fs -put "${CUR_DIR}"/data/* /user/doris/suites/multi_catalog/
|
||||
|
||||
# create table
|
||||
hive -f "${CUR_DIR}"/create_table.hql
|
||||
|
||||
@ -0,0 +1,22 @@
|
||||
create database if not exists multi_catalog;
|
||||
|
||||
use multi_catalog;
|
||||
|
||||
CREATE TABLE `region`(
|
||||
`r_regionkey` int,
|
||||
`r_name` char(25))
|
||||
ROW FORMAT SERDE
|
||||
'org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
|
||||
WITH SERDEPROPERTIES (
|
||||
'field.delim'='|',
|
||||
'serialization.format'='|')
|
||||
STORED AS INPUTFORMAT
|
||||
'org.apache.hadoop.mapred.TextInputFormat'
|
||||
OUTPUTFORMAT
|
||||
'org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
|
||||
LOCATION
|
||||
'/user/doris/suites/multi_catalog/region'
|
||||
TBLPROPERTIES (
|
||||
'transient_lastDdlTime'='1670483235');
|
||||
|
||||
msck repair table region;
|
||||
Binary file not shown.
@ -0,0 +1,12 @@
|
||||
#!/bin/bash
|
||||
set -x
|
||||
|
||||
CUR_DIR="$(cd "$(dirname "${BASH_SOURCE[0]}")" &>/dev/null && pwd)"
|
||||
|
||||
## mkdir and put data to hdfs
|
||||
cd "${CUR_DIR}" && rm -rf data/ && tar xzf data.tar.gz
|
||||
hadoop fs -mkdir -p /user/doris/suites/multi_catalog/
|
||||
hadoop fs -put "${CUR_DIR}"/data/* /user/doris/suites/multi_catalog/
|
||||
|
||||
# create table
|
||||
hive -f "${CUR_DIR}"/create_table.hql
|
||||
@ -0,0 +1,22 @@
|
||||
create database if not exists multi_catalog;
|
||||
|
||||
use multi_catalog;
|
||||
|
||||
CREATE TABLE `special_character_1_partition`(
|
||||
`name` string)
|
||||
PARTITIONED BY (
|
||||
`part` string)
|
||||
ROW FORMAT SERDE
|
||||
'org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe'
|
||||
STORED AS INPUTFORMAT
|
||||
'org.apache.hadoop.hive.ql.io.parquet.MapredParquetInputFormat'
|
||||
OUTPUTFORMAT
|
||||
'org.apache.hadoop.hive.ql.io.parquet.MapredParquetOutputFormat'
|
||||
LOCATION
|
||||
'/user/doris/suites/multi_catalog/special_character_1_partition'
|
||||
TBLPROPERTIES (
|
||||
'transient_lastDdlTime'='1689575322');
|
||||
|
||||
set hive.msck.path.validation=ignore;
|
||||
|
||||
msck repair table special_character_1_partition;
|
||||
Binary file not shown.
@ -0,0 +1,12 @@
|
||||
#!/bin/bash
|
||||
set -x
|
||||
|
||||
CUR_DIR="$(cd "$(dirname "${BASH_SOURCE[0]}")" &>/dev/null && pwd)"
|
||||
|
||||
## mkdir and put data to hdfs
|
||||
cd "${CUR_DIR}" && rm -rf data/ && tar xzf data.tar.gz
|
||||
hadoop fs -mkdir -p /user/doris/suites/multi_catalog/
|
||||
hadoop fs -put "${CUR_DIR}"/data/* /user/doris/suites/multi_catalog/
|
||||
|
||||
# create table
|
||||
hive -f "${CUR_DIR}"/create_table.hql
|
||||
@ -0,0 +1,654 @@
|
||||
create database if not exists multi_catalog;
|
||||
|
||||
use multi_catalog;
|
||||
|
||||
CREATE TABLE `wide_table1_orc`(
|
||||
`col1` decimal(16,0),
|
||||
`col2` string,
|
||||
`col3` string,
|
||||
`col4` string,
|
||||
`col5` string,
|
||||
`col6` string,
|
||||
`col7` string,
|
||||
`col8` string,
|
||||
`col9` string,
|
||||
`col10` string,
|
||||
`col11` string,
|
||||
`col12` string,
|
||||
`col13` string,
|
||||
`col14` string,
|
||||
`col15` decimal(2,0),
|
||||
`col16` decimal(2,0),
|
||||
`col17` decimal(2,0),
|
||||
`col18` decimal(2,0),
|
||||
`col19` decimal(2,0),
|
||||
`col20` decimal(10,0),
|
||||
`col21` string,
|
||||
`col22` decimal(10,0),
|
||||
`col23` string,
|
||||
`col24` decimal(12,0),
|
||||
`col25` string,
|
||||
`col26` decimal(8,0),
|
||||
`col27` decimal(8,0),
|
||||
`col28` decimal(8,0),
|
||||
`col29` decimal(8,0),
|
||||
`col30` decimal(2,0),
|
||||
`col31` string,
|
||||
`col32` decimal(12,0),
|
||||
`col33` string,
|
||||
`col34` decimal(8,0),
|
||||
`col35` decimal(12,0),
|
||||
`col36` string,
|
||||
`col37` decimal(12,0),
|
||||
`col38` string,
|
||||
`col39` decimal(12,0),
|
||||
`col40` string,
|
||||
`col41` int,
|
||||
`col42` int,
|
||||
`col43` string,
|
||||
`col44` decimal(12,0),
|
||||
`col45` string,
|
||||
`col46` int,
|
||||
`col47` decimal(12,0),
|
||||
`col48` string,
|
||||
`col49` decimal(8,0),
|
||||
`col50` string,
|
||||
`col51` string,
|
||||
`col52` string,
|
||||
`col53` string,
|
||||
`col54` string,
|
||||
`col55` string,
|
||||
`col56` string,
|
||||
`col57` string,
|
||||
`col58` string,
|
||||
`col59` string,
|
||||
`col60` string,
|
||||
`col61` string,
|
||||
`col62` string,
|
||||
`col63` string,
|
||||
`col64` decimal(12,0),
|
||||
`col65` string,
|
||||
`col66` string,
|
||||
`col67` string,
|
||||
`col68` decimal(12,0),
|
||||
`col69` string,
|
||||
`col70` decimal(16,2),
|
||||
`col71` decimal(8,0),
|
||||
`col72` string,
|
||||
`col73` decimal(16,2),
|
||||
`col74` decimal(16,2),
|
||||
`col75` decimal(16,2),
|
||||
`col76` decimal(16,2),
|
||||
`col77` string,
|
||||
`col78` string,
|
||||
`col79` string,
|
||||
`col80` decimal(16,2),
|
||||
`col81` decimal(16,2),
|
||||
`col82` decimal(16,2),
|
||||
`col83` decimal(16,2),
|
||||
`col84` decimal(16,2),
|
||||
`col85` decimal(16,2),
|
||||
`col86` decimal(16,2),
|
||||
`col87` decimal(16,2),
|
||||
`col88` decimal(16,2),
|
||||
`col89` decimal(16,2),
|
||||
`col90` decimal(16,2),
|
||||
`col91` decimal(16,2),
|
||||
`col92` decimal(16,2),
|
||||
`col93` decimal(16,2),
|
||||
`col94` decimal(16,2),
|
||||
`col95` decimal(16,2),
|
||||
`col96` decimal(16,2),
|
||||
`col97` decimal(19,6),
|
||||
`col98` decimal(12,5),
|
||||
`col99` decimal(12,5),
|
||||
`col100` decimal(12,5),
|
||||
`col101` decimal(16,2),
|
||||
`col102` decimal(16,2),
|
||||
`col103` decimal(16,2),
|
||||
`col104` decimal(16,2),
|
||||
`col105` decimal(16,2),
|
||||
`col106` decimal(16,2),
|
||||
`col107` decimal(16,2),
|
||||
`col108` decimal(16,2),
|
||||
`col109` decimal(16,2),
|
||||
`col110` decimal(16,2),
|
||||
`col111` decimal(16,2),
|
||||
`col112` decimal(16,2),
|
||||
`col113` decimal(16,2),
|
||||
`col114` decimal(16,2),
|
||||
`col115` decimal(16,2),
|
||||
`col116` decimal(16,2),
|
||||
`col117` decimal(8,0),
|
||||
`col118` decimal(8,0),
|
||||
`col119` decimal(8,0),
|
||||
`col120` decimal(8,0),
|
||||
`col121` decimal(8,0),
|
||||
`col122` decimal(8,0),
|
||||
`col123` decimal(8,0),
|
||||
`col124` decimal(8,0),
|
||||
`col125` decimal(8,0),
|
||||
`col126` decimal(8,0),
|
||||
`col127` decimal(8,0),
|
||||
`col128` decimal(8,0),
|
||||
`col129` decimal(8,0),
|
||||
`col130` decimal(8,0),
|
||||
`col131` decimal(8,0),
|
||||
`col132` decimal(16,2),
|
||||
`col133` decimal(16,2),
|
||||
`col134` decimal(16,2),
|
||||
`col135` decimal(16,2),
|
||||
`col136` decimal(16,2),
|
||||
`col137` decimal(16,2),
|
||||
`col138` decimal(16,2),
|
||||
`col139` decimal(16,2),
|
||||
`col140` decimal(16,2),
|
||||
`col141` decimal(16,2),
|
||||
`col142` decimal(16,2),
|
||||
`col143` decimal(16,2),
|
||||
`col144` decimal(16,2),
|
||||
`col145` decimal(16,2),
|
||||
`col146` decimal(16,2),
|
||||
`col147` decimal(16,2),
|
||||
`col148` decimal(16,2),
|
||||
`col149` decimal(16,2),
|
||||
`col150` decimal(16,2),
|
||||
`col151` decimal(16,2),
|
||||
`col152` decimal(16,2),
|
||||
`col153` decimal(16,2),
|
||||
`col154` decimal(16,2),
|
||||
`col155` decimal(16,2),
|
||||
`col156` decimal(16,2),
|
||||
`col157` decimal(16,2),
|
||||
`col158` decimal(16,2),
|
||||
`col159` decimal(16,2),
|
||||
`col160` decimal(16,2),
|
||||
`col161` decimal(16,2),
|
||||
`col162` decimal(16,2),
|
||||
`col163` decimal(16,2),
|
||||
`col164` decimal(16,2),
|
||||
`col165` decimal(16,2),
|
||||
`col166` decimal(16,2),
|
||||
`col167` decimal(16,2),
|
||||
`col168` decimal(16,2),
|
||||
`col169` decimal(16,2),
|
||||
`col170` decimal(16,2),
|
||||
`col171` decimal(16,2),
|
||||
`col172` decimal(16,2),
|
||||
`col173` decimal(16,2),
|
||||
`col174` decimal(16,2),
|
||||
`col175` decimal(16,2),
|
||||
`col176` decimal(16,2),
|
||||
`col177` decimal(16,2),
|
||||
`col178` decimal(16,2),
|
||||
`col179` decimal(16,2),
|
||||
`col180` decimal(16,2),
|
||||
`col181` decimal(16,2),
|
||||
`col182` decimal(16,2),
|
||||
`col183` decimal(16,2),
|
||||
`col184` decimal(16,2),
|
||||
`col185` decimal(16,2),
|
||||
`col186` decimal(16,2),
|
||||
`col187` decimal(16,2),
|
||||
`col188` decimal(16,2),
|
||||
`col189` decimal(16,2),
|
||||
`col190` decimal(16,2),
|
||||
`col191` decimal(16,2),
|
||||
`col192` decimal(16,2),
|
||||
`col193` decimal(16,2),
|
||||
`col194` decimal(16,2),
|
||||
`col195` decimal(16,2),
|
||||
`col196` decimal(16,2),
|
||||
`col197` decimal(16,2),
|
||||
`col198` decimal(16,2),
|
||||
`col199` decimal(16,2),
|
||||
`col200` decimal(16,2),
|
||||
`col201` decimal(16,2),
|
||||
`col202` decimal(16,2),
|
||||
`col203` decimal(16,2),
|
||||
`col204` decimal(16,2),
|
||||
`col205` decimal(16,2),
|
||||
`col206` decimal(16,2),
|
||||
`col207` decimal(16,2),
|
||||
`col208` decimal(16,2),
|
||||
`col209` decimal(16,2),
|
||||
`col210` decimal(16,2),
|
||||
`col211` decimal(16,2),
|
||||
`col212` decimal(16,2),
|
||||
`col213` decimal(16,2),
|
||||
`col214` decimal(16,2),
|
||||
`col215` decimal(16,2),
|
||||
`col216` decimal(16,2),
|
||||
`col217` decimal(16,2),
|
||||
`col218` decimal(16,2),
|
||||
`col219` decimal(16,2),
|
||||
`col220` decimal(16,2),
|
||||
`col221` decimal(16,2),
|
||||
`col222` decimal(16,2),
|
||||
`col223` decimal(16,2),
|
||||
`col224` decimal(16,2),
|
||||
`col225` decimal(16,2),
|
||||
`col226` decimal(16,2),
|
||||
`col227` decimal(16,2),
|
||||
`col228` decimal(16,2),
|
||||
`col229` decimal(16,2),
|
||||
`col230` decimal(16,2),
|
||||
`col231` decimal(22,6),
|
||||
`col232` decimal(22,6),
|
||||
`col233` decimal(22,6),
|
||||
`col234` decimal(22,6),
|
||||
`col235` decimal(22,6),
|
||||
`col236` decimal(22,6),
|
||||
`col237` decimal(22,6),
|
||||
`col238` decimal(22,6),
|
||||
`col239` decimal(22,6),
|
||||
`col240` decimal(22,6),
|
||||
`col241` decimal(22,6),
|
||||
`col242` decimal(22,6),
|
||||
`col243` decimal(22,6),
|
||||
`col244` decimal(22,6),
|
||||
`col245` decimal(22,6),
|
||||
`col246` decimal(22,6),
|
||||
`col247` decimal(22,6),
|
||||
`col248` decimal(22,6),
|
||||
`col249` decimal(22,6),
|
||||
`col250` decimal(22,6),
|
||||
`col251` decimal(22,6),
|
||||
`col252` decimal(16,2),
|
||||
`col253` decimal(16,2),
|
||||
`col254` decimal(16,2),
|
||||
`col255` decimal(16,2),
|
||||
`col256` decimal(16,2),
|
||||
`col257` decimal(16,2),
|
||||
`col258` decimal(16,2),
|
||||
`col259` decimal(16,2),
|
||||
`col260` decimal(16,2),
|
||||
`col261` decimal(16,2),
|
||||
`col262` decimal(16,2),
|
||||
`col263` decimal(16,2),
|
||||
`col264` decimal(16,2),
|
||||
`col265` decimal(16,2),
|
||||
`col266` decimal(16,2),
|
||||
`col267` decimal(16,2),
|
||||
`col268` decimal(16,2),
|
||||
`col269` decimal(16,2),
|
||||
`col270` decimal(16,2),
|
||||
`col271` decimal(16,2),
|
||||
`col272` decimal(16,2),
|
||||
`col273` decimal(16,2),
|
||||
`col274` decimal(16,2),
|
||||
`col275` decimal(16,2),
|
||||
`col276` decimal(16,2),
|
||||
`col277` decimal(16,2),
|
||||
`col278` decimal(16,2),
|
||||
`col279` decimal(16,2),
|
||||
`col280` decimal(16,2),
|
||||
`col281` decimal(16,2),
|
||||
`col282` decimal(16,2),
|
||||
`col283` decimal(16,2),
|
||||
`col284` decimal(16,2),
|
||||
`col285` decimal(16,2),
|
||||
`col286` decimal(16,2),
|
||||
`col287` decimal(16,2),
|
||||
`col288` decimal(16,2),
|
||||
`col289` decimal(16,2),
|
||||
`col290` decimal(16,2),
|
||||
`col291` decimal(16,2),
|
||||
`col292` decimal(16,2),
|
||||
`col293` decimal(16,2),
|
||||
`col294` decimal(16,2),
|
||||
`col295` decimal(16,2),
|
||||
`col296` decimal(16,2),
|
||||
`col297` decimal(16,2),
|
||||
`col298` decimal(16,2),
|
||||
`col299` decimal(16,2),
|
||||
`col300` decimal(16,2),
|
||||
`col301` decimal(16,2),
|
||||
`col302` decimal(16,2),
|
||||
`col303` decimal(16,2),
|
||||
`col304` decimal(16,2),
|
||||
`col305` decimal(16,2),
|
||||
`col306` decimal(16,2),
|
||||
`col307` decimal(16,2),
|
||||
`col308` decimal(16,2),
|
||||
`col309` decimal(16,2),
|
||||
`col310` decimal(16,2),
|
||||
`col311` decimal(16,2),
|
||||
`col312` decimal(16,2),
|
||||
`col313` decimal(16,2),
|
||||
`col314` decimal(16,2),
|
||||
`col315` decimal(16,2),
|
||||
`col316` decimal(16,2),
|
||||
`col317` decimal(16,2),
|
||||
`col318` decimal(16,2),
|
||||
`col319` decimal(16,2),
|
||||
`col320` decimal(16,2),
|
||||
`col321` decimal(16,2),
|
||||
`col322` decimal(16,2),
|
||||
`col323` decimal(16,2),
|
||||
`col324` decimal(16,2),
|
||||
`col325` decimal(16,2),
|
||||
`col326` decimal(16,2),
|
||||
`col327` decimal(16,2),
|
||||
`col328` decimal(16,2),
|
||||
`col329` decimal(16,2),
|
||||
`col330` decimal(16,2),
|
||||
`col331` decimal(16,2),
|
||||
`col332` decimal(16,2),
|
||||
`col333` decimal(16,2),
|
||||
`col334` decimal(16,2),
|
||||
`col335` decimal(16,2),
|
||||
`col336` decimal(16,2),
|
||||
`col337` decimal(16,2),
|
||||
`col338` decimal(16,2),
|
||||
`col339` decimal(22,6),
|
||||
`col340` decimal(22,6),
|
||||
`col341` decimal(22,6),
|
||||
`col342` decimal(22,6),
|
||||
`col343` decimal(22,6),
|
||||
`col344` decimal(22,6),
|
||||
`col345` decimal(22,6),
|
||||
`col346` decimal(22,6),
|
||||
`col347` decimal(22,6),
|
||||
`col348` decimal(22,6),
|
||||
`col349` decimal(22,6),
|
||||
`col350` decimal(22,6),
|
||||
`col351` decimal(22,6),
|
||||
`col352` decimal(22,6),
|
||||
`col353` decimal(16,2),
|
||||
`col354` decimal(16,2),
|
||||
`col355` decimal(16,2),
|
||||
`col356` decimal(16,2),
|
||||
`col357` decimal(16,2),
|
||||
`col358` decimal(16,2),
|
||||
`col359` decimal(16,2),
|
||||
`col360` decimal(16,2),
|
||||
`col361` decimal(16,2),
|
||||
`col362` decimal(16,2),
|
||||
`col363` decimal(16,2),
|
||||
`col364` decimal(16,2),
|
||||
`col365` decimal(16,2),
|
||||
`col366` decimal(16,2),
|
||||
`col367` decimal(16,2),
|
||||
`col368` decimal(16,2),
|
||||
`col369` decimal(16,2),
|
||||
`col370` decimal(16,2),
|
||||
`col371` decimal(16,2),
|
||||
`col372` decimal(16,2),
|
||||
`col373` decimal(16,2),
|
||||
`col374` decimal(16,2),
|
||||
`col375` decimal(16,2),
|
||||
`col376` decimal(16,2),
|
||||
`col377` decimal(16,2),
|
||||
`col378` decimal(16,2),
|
||||
`col379` decimal(16,2),
|
||||
`col380` decimal(16,2),
|
||||
`col381` decimal(16,2),
|
||||
`col382` decimal(16,2),
|
||||
`col383` decimal(16,2),
|
||||
`col384` decimal(16,2),
|
||||
`col385` decimal(16,2),
|
||||
`col386` decimal(16,2),
|
||||
`col387` decimal(16,2),
|
||||
`col388` decimal(16,2),
|
||||
`col389` decimal(16,2),
|
||||
`col390` decimal(16,2),
|
||||
`col391` decimal(16,2),
|
||||
`col392` decimal(16,2),
|
||||
`col393` decimal(16,2),
|
||||
`col394` decimal(16,2),
|
||||
`col395` decimal(8,0),
|
||||
`col396` decimal(8,0),
|
||||
`col397` string,
|
||||
`col398` string,
|
||||
`col399` decimal(16,2),
|
||||
`col400` decimal(16,2),
|
||||
`col401` decimal(16,2),
|
||||
`col402` decimal(16,2),
|
||||
`col403` decimal(16,2),
|
||||
`col404` decimal(16,2),
|
||||
`col405` decimal(16,2),
|
||||
`col406` decimal(16,2),
|
||||
`col407` decimal(16,2),
|
||||
`col408` decimal(16,2),
|
||||
`col409` decimal(16,2),
|
||||
`col410` decimal(16,2),
|
||||
`col411` decimal(16,2),
|
||||
`col412` decimal(16,2),
|
||||
`col413` decimal(16,2),
|
||||
`col414` decimal(16,2),
|
||||
`col415` decimal(16,2),
|
||||
`col416` decimal(16,2),
|
||||
`col417` decimal(16,2),
|
||||
`col418` decimal(16,2),
|
||||
`col419` decimal(16,2),
|
||||
`col420` int,
|
||||
`col421` int,
|
||||
`col422` int,
|
||||
`col423` decimal(16,2),
|
||||
`col424` decimal(16,2),
|
||||
`col425` decimal(16,2),
|
||||
`col426` decimal(16,2),
|
||||
`col427` decimal(16,2),
|
||||
`col428` decimal(16,2),
|
||||
`col429` string,
|
||||
`col430` string,
|
||||
`col431` string,
|
||||
`col432` decimal(16,2),
|
||||
`col433` decimal(16,2),
|
||||
`col434` decimal(16,2),
|
||||
`col435` decimal(16,2),
|
||||
`col436` decimal(16,2),
|
||||
`col437` decimal(16,2),
|
||||
`col438` decimal(16,2),
|
||||
`col439` decimal(16,2),
|
||||
`col440` decimal(22,6),
|
||||
`col441` decimal(22,6),
|
||||
`col442` decimal(16,3),
|
||||
`col443` decimal(16,3),
|
||||
`col444` decimal(16,3),
|
||||
`col445` decimal(16,2),
|
||||
`col446` decimal(16,2),
|
||||
`col447` decimal(9,6),
|
||||
`col448` decimal(9,6),
|
||||
`col449` decimal(16,2),
|
||||
`col450` decimal(16,2),
|
||||
`col451` decimal(16,2),
|
||||
`col452` decimal(16,2),
|
||||
`col453` decimal(16,2),
|
||||
`col454` int,
|
||||
`col455` decimal(16,2),
|
||||
`col456` decimal(16,2),
|
||||
`col457` decimal(16,2),
|
||||
`col458` decimal(16,2),
|
||||
`col459` decimal(16,2),
|
||||
`col460` decimal(16,2),
|
||||
`col461` decimal(16,2),
|
||||
`col462` decimal(16,2),
|
||||
`col463` decimal(22,6),
|
||||
`col464` decimal(22,6),
|
||||
`col465` decimal(22,6),
|
||||
`col466` decimal(22,6),
|
||||
`col467` decimal(22,6),
|
||||
`col468` decimal(22,6),
|
||||
`col469` decimal(22,6),
|
||||
`col470` decimal(22,6),
|
||||
`col471` decimal(16,2),
|
||||
`col472` decimal(16,2),
|
||||
`col473` decimal(16,2),
|
||||
`col474` decimal(16,2),
|
||||
`col475` decimal(16,2),
|
||||
`col476` decimal(16,2),
|
||||
`col477` decimal(16,2),
|
||||
`col478` decimal(16,2),
|
||||
`col479` decimal(16,2),
|
||||
`col480` decimal(16,2),
|
||||
`col481` decimal(16,2),
|
||||
`col482` decimal(16,2),
|
||||
`col483` decimal(16,2),
|
||||
`col484` decimal(16,2),
|
||||
`col485` decimal(16,2),
|
||||
`col486` decimal(16,2),
|
||||
`col487` decimal(16,2),
|
||||
`col488` decimal(16,2),
|
||||
`col489` decimal(16,2),
|
||||
`col490` decimal(16,2),
|
||||
`col491` decimal(16,2),
|
||||
`col492` decimal(16,2),
|
||||
`col493` decimal(16,2),
|
||||
`col494` decimal(16,2),
|
||||
`col495` decimal(16,2),
|
||||
`col496` decimal(16,2),
|
||||
`col497` decimal(16,2),
|
||||
`col498` decimal(16,2),
|
||||
`col499` decimal(16,2),
|
||||
`col500` decimal(16,2),
|
||||
`col501` decimal(16,2),
|
||||
`col502` decimal(16,2),
|
||||
`col503` decimal(16,2),
|
||||
`col504` decimal(16,2),
|
||||
`col505` decimal(16,2),
|
||||
`col506` decimal(16,2),
|
||||
`col507` decimal(16,2),
|
||||
`col508` decimal(8,0),
|
||||
`col509` decimal(8,0),
|
||||
`col510` decimal(8,0),
|
||||
`col511` decimal(8,0),
|
||||
`col512` decimal(8,0),
|
||||
`col513` decimal(8,0),
|
||||
`col514` decimal(8,0),
|
||||
`col515` decimal(8,0),
|
||||
`col516` decimal(8,0),
|
||||
`col517` decimal(8,0),
|
||||
`col518` decimal(8,0),
|
||||
`col519` decimal(8,0),
|
||||
`col520` decimal(8,0),
|
||||
`col521` decimal(8,0),
|
||||
`col522` decimal(8,0),
|
||||
`col523` decimal(8,0),
|
||||
`col524` decimal(8,0),
|
||||
`col525` decimal(8,0),
|
||||
`col526` decimal(8,0),
|
||||
`col527` decimal(8,0),
|
||||
`col528` decimal(8,0),
|
||||
`col529` decimal(8,0),
|
||||
`col530` decimal(16,2),
|
||||
`col531` decimal(16,2),
|
||||
`col532` decimal(16,2),
|
||||
`col533` decimal(16,2),
|
||||
`col534` decimal(22,6),
|
||||
`col535` decimal(16,2),
|
||||
`col536` decimal(16,2),
|
||||
`col537` decimal(16,2),
|
||||
`col538` decimal(16,2),
|
||||
`col539` decimal(22,6),
|
||||
`col540` decimal(22,6),
|
||||
`col541` decimal(22,6),
|
||||
`col542` string,
|
||||
`col543` decimal(16,2),
|
||||
`col544` decimal(16,2),
|
||||
`col545` decimal(16,2),
|
||||
`col546` decimal(16,2),
|
||||
`col547` decimal(16,2),
|
||||
`col548` decimal(16,2),
|
||||
`col549` decimal(16,2),
|
||||
`col550` decimal(16,2),
|
||||
`col551` decimal(16,2),
|
||||
`col552` decimal(16,2),
|
||||
`col553` decimal(16,2),
|
||||
`col554` decimal(16,2),
|
||||
`col555` decimal(16,2),
|
||||
`col556` decimal(16,2),
|
||||
`col557` string,
|
||||
`col558` string,
|
||||
`col559` string,
|
||||
`col560` string,
|
||||
`col561` string,
|
||||
`col562` string,
|
||||
`col563` string,
|
||||
`col564` string,
|
||||
`col565` decimal(16,2),
|
||||
`col566` decimal(22,6),
|
||||
`col567` decimal(22,6),
|
||||
`col568` decimal(22,6),
|
||||
`col569` string,
|
||||
`col570` string,
|
||||
`col571` decimal(16,2),
|
||||
`col572` decimal(12,0),
|
||||
`col573` string,
|
||||
`col574` string,
|
||||
`col575` decimal(22,6),
|
||||
`col576` decimal(22,6),
|
||||
`col577` decimal(22,6),
|
||||
`col578` decimal(16,2),
|
||||
`col579` decimal(22,2),
|
||||
`col580` decimal(22,2),
|
||||
`col581` decimal(22,2),
|
||||
`col582` decimal(16,2),
|
||||
`col583` decimal(16,2),
|
||||
`col584` decimal(16,2),
|
||||
`col585` decimal(16,2),
|
||||
`col586` decimal(16,2),
|
||||
`col587` decimal(16,2),
|
||||
`col588` decimal(16,2),
|
||||
`col589` decimal(16,2),
|
||||
`col590` decimal(16,2),
|
||||
`col591` decimal(16,2),
|
||||
`col592` decimal(16,2),
|
||||
`col593` decimal(16,2),
|
||||
`col594` decimal(16,2),
|
||||
`col595` decimal(16,2),
|
||||
`col596` decimal(16,2),
|
||||
`col597` string,
|
||||
`col598` string,
|
||||
`col599` decimal(16,2),
|
||||
`col600` decimal(16,2),
|
||||
`col601` decimal(16,2),
|
||||
`col602` decimal(16,2),
|
||||
`col603` decimal(16,2),
|
||||
`col604` decimal(16,2),
|
||||
`col605` decimal(16,2),
|
||||
`col606` decimal(16,2),
|
||||
`col607` int,
|
||||
`col608` string,
|
||||
`col609` string,
|
||||
`col610` string,
|
||||
`col611` decimal(22,2),
|
||||
`col612` string,
|
||||
`col613` string,
|
||||
`col614` string,
|
||||
`col615` string,
|
||||
`col616` string,
|
||||
`col617` string,
|
||||
`col618` string,
|
||||
`col619` string,
|
||||
`col620` string,
|
||||
`col621` string,
|
||||
`col622` string,
|
||||
`col623` string,
|
||||
`col624` string,
|
||||
`col625` string,
|
||||
`col626` string,
|
||||
`col627` string,
|
||||
`col628` string,
|
||||
`col629` decimal(16,2),
|
||||
`col630` decimal(16,2),
|
||||
`col631` string,
|
||||
`col632` string,
|
||||
`col633` string,
|
||||
`col634` string,
|
||||
`col635` string,
|
||||
`col636` string,
|
||||
`col637` decimal(16,2))
|
||||
ROW FORMAT SERDE
|
||||
'org.apache.hadoop.hive.ql.io.orc.OrcSerde'
|
||||
STORED AS INPUTFORMAT
|
||||
'org.apache.hadoop.hive.ql.io.orc.OrcInputFormat'
|
||||
OUTPUTFORMAT
|
||||
'org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat'
|
||||
LOCATION
|
||||
'/user/doris/suites/multi_catalog/wide_table1_orc'
|
||||
TBLPROPERTIES (
|
||||
'transient_lastDdlTime'='1680503244');
|
||||
|
||||
msck repair table wide_table1_orc;
|
||||
@ -0,0 +1,22 @@
|
||||
#!/bin/bash
|
||||
set -x
|
||||
|
||||
CUR_DIR="$(cd "$(dirname "${BASH_SOURCE[0]}")" &>/dev/null && pwd)"
|
||||
|
||||
if [[ ! -d "${CUR_DIR}/data" ]]; then
|
||||
echo "${CUR_DIR}/data does not exist"
|
||||
cd "${CUR_DIR}" && rm -f data.tar.gz \
|
||||
&& curl -O https://s3BucketName.s3Endpoint/regression/datalake/pipeline_data/multi_catalog/test_wide_table/data.tar.gz \
|
||||
&& tar xzf data.tar.gz
|
||||
cd -
|
||||
else
|
||||
echo "${CUR_DIR}/data exist, continue !"
|
||||
fi
|
||||
|
||||
## mkdir and put data to hdfs
|
||||
hadoop fs -mkdir -p /user/doris/suites/multi_catalog/
|
||||
hadoop fs -put "${CUR_DIR}"/data/* /user/doris/suites/multi_catalog/
|
||||
|
||||
# create table
|
||||
hive -f "${CUR_DIR}"/create_table.hql
|
||||
|
||||
File diff suppressed because one or more lines are too long
@ -14,3 +14,18 @@
|
||||
-- !6 --
|
||||
1 name1
|
||||
|
||||
-- !1 --
|
||||
|
||||
-- !2 --
|
||||
|
||||
-- !3 --
|
||||
|
||||
-- !4 --
|
||||
1 name1
|
||||
|
||||
-- !5 --
|
||||
1 name1
|
||||
|
||||
-- !6 --
|
||||
1 name1
|
||||
|
||||
@ -49,3 +49,53 @@ name6 2023%01%01
|
||||
-- !13 --
|
||||
name# 2023#01#01
|
||||
|
||||
-- !1 --
|
||||
name# 2023#01#01
|
||||
name1 2023/01/01
|
||||
name10 2023<01><01>
|
||||
name11 2023\\01\\01
|
||||
name12 2023.01.01
|
||||
name2 2023 01 01
|
||||
name3 2023:01:01
|
||||
name4 2023?01?01
|
||||
name5 2023=01=01
|
||||
name6 2023%01%01
|
||||
name8 2023"01"01
|
||||
name9 2023'01'01
|
||||
|
||||
-- !2 --
|
||||
name2
|
||||
|
||||
-- !3 --
|
||||
name1
|
||||
|
||||
-- !4 --
|
||||
name4 2023?01?01
|
||||
|
||||
-- !5 --
|
||||
name12 2023.01.01
|
||||
|
||||
-- !6 --
|
||||
name10 2023<01><01>
|
||||
|
||||
-- !7 --
|
||||
name3 2023:01:01
|
||||
|
||||
-- !8 --
|
||||
name5 2023=01=01
|
||||
|
||||
-- !9 --
|
||||
name8 2023"01"01
|
||||
|
||||
-- !10 --
|
||||
name9 2023'01'01
|
||||
|
||||
-- !11 --
|
||||
name11 2023\\01\\01
|
||||
|
||||
-- !12 --
|
||||
name6 2023%01%01
|
||||
|
||||
-- !13 --
|
||||
name# 2023#01#01
|
||||
|
||||
@ -0,0 +1,45 @@
|
||||
-- This file is automatically generated. You should know what you did if you want to edit this
|
||||
-- !01 --
|
||||
6117920261 28156890937818.64 11058113 84788841307158.93 9988065.83660 8116313253956313.527443
|
||||
|
||||
-- !02 --
|
||||
6117920261 28156890937818.64 11058113 84788841307158.93 9988065.83660 8116313253956313.527443
|
||||
|
||||
-- !03 --
|
||||
|
||||
-- !04 --
|
||||
|
||||
-- !05 --
|
||||
6117920261 28156890937818.64 11058113 84788841307158.93 9988065.83660 8116313253956313.527443
|
||||
|
||||
-- !06 --
|
||||
6117920261 28156890937818.64 11058113 84788841307158.93 9988065.83660 8116313253956313.527443
|
||||
|
||||
-- !07 --
|
||||
6117920261 28156890937818.64 11058113 84788841307158.93 9988065.83660 8116313253956313.527443
|
||||
|
||||
-- !08 --
|
||||
9999999541515682 99999218685068.86 99999869 221095586.11 27.54254 61077635638.763621
|
||||
|
||||
-- !01 --
|
||||
6117920261 28156890937818.64 11058113 84788841307158.93 9988065.83660 8116313253956313.527443
|
||||
|
||||
-- !02 --
|
||||
6117920261 28156890937818.64 11058113 84788841307158.93 9988065.83660 8116313253956313.527443
|
||||
|
||||
-- !03 --
|
||||
|
||||
-- !04 --
|
||||
|
||||
-- !05 --
|
||||
6117920261 28156890937818.64 11058113 84788841307158.93 9988065.83660 8116313253956313.527443
|
||||
|
||||
-- !06 --
|
||||
6117920261 28156890937818.64 11058113 84788841307158.93 9988065.83660 8116313253956313.527443
|
||||
|
||||
-- !07 --
|
||||
6117920261 28156890937818.64 11058113 84788841307158.93 9988065.83660 8116313253956313.527443
|
||||
|
||||
-- !08 --
|
||||
9999999541515682 99999218685068.86 99999869 221095586.11 27.54254 61077635638.763621
|
||||
|
||||
@ -1,23 +0,0 @@
|
||||
-- This file is automatically generated. You should know what you did if you want to edit this
|
||||
-- !01 --
|
||||
6117920261 28156890937818.64 11058113 84788841307158.93 9988065.83660 8116313253956313.527443
|
||||
|
||||
-- !02 --
|
||||
6117920261 28156890937818.64 11058113 84788841307158.93 9988065.83660 8116313253956313.527443
|
||||
|
||||
-- !03 --
|
||||
|
||||
-- !04 --
|
||||
|
||||
-- !05 --
|
||||
6117920261 28156890937818.64 11058113 84788841307158.93 9988065.83660 8116313253956313.527443
|
||||
|
||||
-- !06 --
|
||||
6117920261 28156890937818.64 11058113 84788841307158.93 9988065.83660 8116313253956313.527443
|
||||
|
||||
-- !07 --
|
||||
6117920261 28156890937818.64 11058113 84788841307158.93 9988065.83660 8116313253956313.527443
|
||||
|
||||
-- !08 --
|
||||
9999999541515682.000000000 99999218685068.860000000 99999869.000000000 221095586.110000000 27.542540000 61077635638.763621000
|
||||
|
||||
@ -15,20 +15,23 @@
|
||||
// specific language governing permissions and limitations
|
||||
// under the License.
|
||||
|
||||
suite("test_complex_types", "p2,external,hive,external_remote,external_remote_hive") {
|
||||
String enabled = context.config.otherConfigs.get("enableExternalHiveTest")
|
||||
if (enabled != null && enabled.equalsIgnoreCase("true")) {
|
||||
String extHiveHmsHost = context.config.otherConfigs.get("extHiveHmsHost")
|
||||
String extHiveHmsPort = context.config.otherConfigs.get("extHiveHmsPort")
|
||||
String catalog_name = "test_complex_types"
|
||||
sql """drop catalog if exists ${catalog_name};"""
|
||||
sql """
|
||||
create catalog if not exists ${catalog_name} properties (
|
||||
'type'='hms',
|
||||
'hadoop.username' = 'hadoop',
|
||||
'hive.metastore.uris' = 'thrift://${extHiveHmsHost}:${extHiveHmsPort}'
|
||||
);
|
||||
"""
|
||||
suite("test_complex_types", "p0,external,hive,external_docker,external_docker_hive") {
|
||||
String enabled = context.config.otherConfigs.get("enableHiveTest")
|
||||
if (enabled == null || !enabled.equalsIgnoreCase("true")) {
|
||||
logger.info("diable Hive test.")
|
||||
return;
|
||||
}
|
||||
|
||||
for (String hivePrefix : ["hive2", "hive3"]) {
|
||||
String hms_port = context.config.otherConfigs.get(hivePrefix + "HmsPort")
|
||||
String catalog_name = "${hivePrefix}_test_complex_types"
|
||||
String externalEnvIp = context.config.otherConfigs.get("externalEnvIp")
|
||||
|
||||
sql """drop catalog if exists ${catalog_name}"""
|
||||
sql """create catalog if not exists ${catalog_name} properties (
|
||||
"type"="hms",
|
||||
'hive.metastore.uris' = 'thrift://${externalEnvIp}:${hms_port}'
|
||||
);"""
|
||||
logger.info("catalog " + catalog_name + " created")
|
||||
sql """switch ${catalog_name};"""
|
||||
logger.info("switched to catalog " + catalog_name)
|
||||
@ -47,9 +50,9 @@ suite("test_complex_types", "p2,external,hive,external_remote,external_remote_hi
|
||||
|
||||
qt_array_max """select count(array_max(capacity)) from byd where array_max(capacity) > 0.99"""
|
||||
|
||||
qt_array_filter """select count(array_size(array_filter(i -> (i > 0.99), capacity))) from byd where array_size(array_filter(i -> (i > 0.99), capacity))"""
|
||||
// qt_array_filter """select count(array_size(array_filter(i -> (i > 0.99), capacity))) from byd where array_size(array_filter(i -> (i > 0.99), capacity))"""
|
||||
|
||||
qt_array_last """select max(array_last(i -> i > 0, capacity)) from byd where array_last(i -> i > 0, capacity) < 0.99"""
|
||||
// qt_array_last """select max(array_last(i -> i > 0, capacity)) from byd where array_last(i -> i > 0, capacity) < 0.99"""
|
||||
|
||||
qt_null_struct_element_orc """select count(struct_element(favor, 'tip')) from byd where id % 13 = 0"""
|
||||
|
||||
@ -63,9 +66,9 @@ suite("test_complex_types", "p2,external,hive,external_remote,external_remote_hi
|
||||
|
||||
qt_array_max_orc """select count(array_max(capacity)) from byd where array_max(capacity) > 0.99"""
|
||||
|
||||
qt_array_filter_orc """select count(array_size(array_filter(i -> (i > 0.99), capacity))) from byd where array_size(array_filter(i -> (i > 0.99), capacity))"""
|
||||
// qt_array_filter_orc """select count(array_size(array_filter(i -> (i > 0.99), capacity))) from byd where array_size(array_filter(i -> (i > 0.99), capacity))"""
|
||||
|
||||
qt_array_last_orc """select max(array_last(i -> i > 0, capacity)) from byd where array_last(i -> i > 0, capacity) < 0.99"""
|
||||
// qt_array_last_orc """select max(array_last(i -> i > 0, capacity)) from byd where array_last(i -> i > 0, capacity) < 0.99"""
|
||||
|
||||
qt_offsets_check """select * from complex_offsets_check order by id"""
|
||||
|
||||
@ -15,20 +15,23 @@
|
||||
// specific language governing permissions and limitations
|
||||
// under the License.
|
||||
|
||||
suite("test_hive_same_db_table_name", "p2,external,hive,external_remote,external_remote_hive") {
|
||||
String enabled = context.config.otherConfigs.get("enableExternalHiveTest")
|
||||
if (enabled != null && enabled.equalsIgnoreCase("true")) {
|
||||
String extHiveHmsHost = context.config.otherConfigs.get("extHiveHmsHost")
|
||||
String extHiveHmsPort = context.config.otherConfigs.get("extHiveHmsPort")
|
||||
String catalog_name = "test_hive_same_db_table_name"
|
||||
sql """drop catalog if exists ${catalog_name};"""
|
||||
sql """
|
||||
create catalog if not exists ${catalog_name} properties (
|
||||
'type'='hms',
|
||||
'hadoop.username' = 'hadoop',
|
||||
'hive.metastore.uris' = 'thrift://${extHiveHmsHost}:${extHiveHmsPort}'
|
||||
);
|
||||
"""
|
||||
suite("test_hive_same_db_table_name", "p0,external,hive,external_docker,external_docker_hive") {
|
||||
String enabled = context.config.otherConfigs.get("enableHiveTest")
|
||||
if (enabled == null || !enabled.equalsIgnoreCase("true")) {
|
||||
logger.info("diable Hive test.")
|
||||
return;
|
||||
}
|
||||
|
||||
for (String hivePrefix : ["hive2", "hive3"]) {
|
||||
String hms_port = context.config.otherConfigs.get(hivePrefix + "HmsPort")
|
||||
String catalog_name = "${hivePrefix}_test_hive_same_db_table_name"
|
||||
String externalEnvIp = context.config.otherConfigs.get("externalEnvIp")
|
||||
|
||||
sql """drop catalog if exists ${catalog_name}"""
|
||||
sql """create catalog if not exists ${catalog_name} properties (
|
||||
"type"="hms",
|
||||
'hive.metastore.uris' = 'thrift://${externalEnvIp}:${hms_port}'
|
||||
);"""
|
||||
sql """switch internal;"""
|
||||
sql """create database if not exists multi_catalog;"""
|
||||
sql """use multi_catalog;"""
|
||||
@ -15,20 +15,23 @@
|
||||
// specific language governing permissions and limitations
|
||||
// under the License.
|
||||
|
||||
suite("test_hive_special_char_partition", "p2,external,hive,external_remote,external_remote_hive") {
|
||||
String enabled = context.config.otherConfigs.get("enableExternalHiveTest")
|
||||
if (enabled != null && enabled.equalsIgnoreCase("true")) {
|
||||
String extHiveHmsHost = context.config.otherConfigs.get("extHiveHmsHost")
|
||||
String extHiveHmsPort = context.config.otherConfigs.get("extHiveHmsPort")
|
||||
String catalog_name = "test_hive_special_char_partition"
|
||||
sql """drop catalog if exists ${catalog_name};"""
|
||||
sql """
|
||||
create catalog if not exists ${catalog_name} properties (
|
||||
'type'='hms',
|
||||
'hadoop.username' = 'hadoop',
|
||||
'hive.metastore.uris' = 'thrift://${extHiveHmsHost}:${extHiveHmsPort}'
|
||||
);
|
||||
"""
|
||||
suite("test_hive_special_char_partition", "p0,external,hive,external_docker,external_docker_hive") {
|
||||
String enabled = context.config.otherConfigs.get("enableHiveTest")
|
||||
if (enabled == null || !enabled.equalsIgnoreCase("true")) {
|
||||
logger.info("diable Hive test.")
|
||||
return;
|
||||
}
|
||||
|
||||
for (String hivePrefix : ["hive2", "hive3"]) {
|
||||
String hms_port = context.config.otherConfigs.get(hivePrefix + "HmsPort")
|
||||
String catalog_name = "${hivePrefix}_test_hive_special_char_partition"
|
||||
String externalEnvIp = context.config.otherConfigs.get("externalEnvIp")
|
||||
|
||||
sql """drop catalog if exists ${catalog_name}"""
|
||||
sql """create catalog if not exists ${catalog_name} properties (
|
||||
"type"="hms",
|
||||
'hive.metastore.uris' = 'thrift://${externalEnvIp}:${hms_port}'
|
||||
);"""
|
||||
logger.info("catalog " + catalog_name + " created")
|
||||
sql """switch ${catalog_name};"""
|
||||
logger.info("switched to catalog " + catalog_name)
|
||||
@ -15,7 +15,7 @@
|
||||
// specific language governing permissions and limitations
|
||||
// under the License.
|
||||
|
||||
suite("test_wide_table", "p2,external,hive,external_remote,external_remote_hive") {
|
||||
suite("test_wide_table", "p0,external,hive,external_docker,external_docker_hive") {
|
||||
|
||||
def formats = ["_orc"]
|
||||
def decimal_test1 = """select col1, col70, col71, col81, col100, col534 from wide_table1SUFFIX where col1 is not null order by col1 limit 1;"""
|
||||
@ -39,24 +39,27 @@ suite("test_wide_table", "p2,external,hive,external_remote,external_remote_hive"
|
||||
"""
|
||||
def decimal_test8 = """select max(col1), max(col70), max(col71), min(col81), min(col100), min(col534) from wide_table1SUFFIX;"""
|
||||
|
||||
String enabled = context.config.otherConfigs.get("enableExternalHiveTest")
|
||||
if (enabled != null && enabled.equalsIgnoreCase("true")) {
|
||||
String extHiveHmsHost = context.config.otherConfigs.get("extHiveHmsHost")
|
||||
String extHiveHmsPort = context.config.otherConfigs.get("extHiveHmsPort")
|
||||
String catalog_name = "external_wide_table"
|
||||
String enabled = context.config.otherConfigs.get("enableHiveTest")
|
||||
if (enabled == null || !enabled.equalsIgnoreCase("true")) {
|
||||
logger.info("diable Hive test.")
|
||||
return;
|
||||
}
|
||||
|
||||
sql """drop catalog if exists ${catalog_name};"""
|
||||
sql """
|
||||
create catalog if not exists ${catalog_name} properties (
|
||||
'type'='hms',
|
||||
'hive.metastore.uris' = 'thrift://${extHiveHmsHost}:${extHiveHmsPort}'
|
||||
);
|
||||
"""
|
||||
for (String hivePrefix : ["hive2", "hive3"]) {
|
||||
String hms_port = context.config.otherConfigs.get(hivePrefix + "HmsPort")
|
||||
String catalog_name = "${hivePrefix}_test_complex_types"
|
||||
String externalEnvIp = context.config.otherConfigs.get("externalEnvIp")
|
||||
|
||||
sql """drop catalog if exists ${catalog_name}"""
|
||||
sql """create catalog if not exists ${catalog_name} properties (
|
||||
"type"="hms",
|
||||
'hive.metastore.uris' = 'thrift://${externalEnvIp}:${hms_port}'
|
||||
);"""
|
||||
logger.info("catalog " + catalog_name + " created")
|
||||
sql """switch ${catalog_name};"""
|
||||
logger.info("switched to catalog " + catalog_name)
|
||||
sql """use wide_tables;"""
|
||||
logger.info("use wide_tables")
|
||||
sql """use multi_catalog;"""
|
||||
logger.info("use multi_catalog")
|
||||
|
||||
for (String format in formats) {
|
||||
logger.info("Process format " + format)
|
||||
Reference in New Issue
Block a user