290 lines
9.9 KiB
Python
290 lines
9.9 KiB
Python
#!/bin/env python
|
|
# -*- coding: utf-8 -*-
|
|
# Licensed to the Apache Software Foundation (ASF) under one
|
|
# or more contributor license agreements. See the NOTICE file
|
|
# distributed with this work for additional information
|
|
# regarding copyright ownership. The ASF licenses this file
|
|
# to you under the Apache License, Version 2.0 (the
|
|
# "License"); you may not use this file except in compliance
|
|
# with the License. You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing,
|
|
# software distributed under the License is distributed on an
|
|
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
|
# KIND, either express or implied. See the License for the
|
|
# specific language governing permissions and limitations
|
|
# under the License.
|
|
|
|
############################################################################
|
|
#
|
|
# @file test_sys_verify.py
|
|
# @date 2015/02/04 15:26:21
|
|
# @brief This file is a test file for palo data loading and verifying.
|
|
#
|
|
#############################################################################
|
|
|
|
"""
|
|
测试各种数据类型和存储方式的数据正确性
|
|
"""
|
|
import pytest
|
|
from data import verify as VERIFY_DATA
|
|
from lib import palo_config
|
|
from lib import palo_client
|
|
from lib import util
|
|
from lib import common
|
|
|
|
config = palo_config.config
|
|
LOG = palo_client.LOG
|
|
L = palo_client.L
|
|
|
|
|
|
broker_name = config.broker_name
|
|
broker_info = palo_config.broker_info
|
|
|
|
|
|
def setup_module():
|
|
"""
|
|
setUp
|
|
"""
|
|
global client
|
|
client = palo_client.PaloClient(config.fe_host, config.fe_query_port, user=config.fe_user, password=config.fe_password)
|
|
client.init()
|
|
try:
|
|
is_exist = False
|
|
brokers = client.get_broker_list()
|
|
for br in brokers:
|
|
if broker_name == br[0]:
|
|
is_exist = True
|
|
break
|
|
if not is_exist:
|
|
raise pytest.skip('no broker')
|
|
except:
|
|
pass
|
|
|
|
|
|
def check_table_load_and_verify(table_name):
|
|
"""
|
|
验证表是否创建成功,导入数据,校验
|
|
"""
|
|
assert client.show_tables(table_name)
|
|
assert client.get_index(table_name)
|
|
|
|
data_desc_list = palo_client.LoadDataInfo(VERIFY_DATA.file_path_1, table_name)
|
|
label_1 = util.get_label()
|
|
client.batch_load(label_1, data_desc_list, is_wait=True, broker=broker_info)
|
|
|
|
assert client.verify(VERIFY_DATA.expected_data_file_list_1, table_name)
|
|
|
|
data_desc_list = palo_client.LoadDataInfo(VERIFY_DATA.file_path_2, table_name)
|
|
label_2 = util.get_label()
|
|
client.batch_load(label_2, data_desc_list, is_wait=True, broker=broker_info)
|
|
|
|
assert client.verify(VERIFY_DATA.expected_data_file_list_2, table_name)
|
|
|
|
|
|
def test_column_sum():
|
|
"""
|
|
{
|
|
"title": "test_sys_verify.test_column_sum",
|
|
"describe": "测试列存储sum聚合方式",
|
|
"tag": "system,p0"
|
|
}
|
|
"""
|
|
"""
|
|
测试列存储sum聚合方式
|
|
"""
|
|
database_name, table_name, index_name = util.gen_num_format_name_list()
|
|
LOG.info(L('', database_name=database_name, \
|
|
table_name=table_name, index_name=index_name))
|
|
client.clean(database_name)
|
|
client.create_database(database_name)
|
|
client.create_table(table_name, \
|
|
VERIFY_DATA.schema_1, storage_type='column', keys_desc='AGGREGATE KEY (K1)')
|
|
check_table_load_and_verify(table_name)
|
|
client.clean(database_name)
|
|
|
|
|
|
def test_column_max():
|
|
"""
|
|
{
|
|
"title": "test_sys_verify.test_column_max",
|
|
"describe": "测试列存储max聚合方式",
|
|
"tag": "system,p0"
|
|
}
|
|
"""
|
|
"""
|
|
测试列存储max聚合方式
|
|
"""
|
|
database_name, table_name, index_name = util.gen_num_format_name_list()
|
|
LOG.info(L('', database_name=database_name, \
|
|
table_name=table_name, index_name=index_name))
|
|
client.clean(database_name)
|
|
client.create_database(database_name)
|
|
client.create_table(table_name, \
|
|
VERIFY_DATA.schema_2, storage_type='column', keys_desc='AGGREGATE KEY (K1)')
|
|
check_table_load_and_verify(table_name)
|
|
client.clean(database_name)
|
|
|
|
|
|
def test_column_min():
|
|
"""
|
|
{
|
|
"title": "test_sys_verify.test_column_min",
|
|
"describe": "测试列存储min聚合方式",
|
|
"tag": "system,p0"
|
|
}
|
|
"""
|
|
"""
|
|
测试列存储min聚合方式
|
|
"""
|
|
database_name, table_name, index_name = util.gen_num_format_name_list()
|
|
LOG.info(L('', database_name=database_name, \
|
|
table_name=table_name, index_name=index_name))
|
|
client.clean(database_name)
|
|
client.create_database(database_name)
|
|
client.create_table(table_name, \
|
|
VERIFY_DATA.schema_3, storage_type='column', keys_desc='AGGREGATE KEY (K1)')
|
|
check_table_load_and_verify(table_name)
|
|
client.clean(database_name)
|
|
|
|
|
|
def test_column_replace():
|
|
"""
|
|
{
|
|
"title": "test_sys_verify.test_column_replace",
|
|
"describe": "测试列存储replace聚合方式",
|
|
"tag": "system,p0"
|
|
}
|
|
"""
|
|
"""
|
|
测试列存储replace聚合方式
|
|
"""
|
|
database_name, table_name, index_name = util.gen_num_format_name_list()
|
|
LOG.info(L('', database_name=database_name, \
|
|
table_name=table_name, index_name=index_name))
|
|
client.clean(database_name)
|
|
client.create_database(database_name)
|
|
client.create_table(table_name, \
|
|
VERIFY_DATA.schema_4, storage_type='column', keys_desc='AGGREGATE KEY (k1)')
|
|
check_table_load_and_verify(table_name)
|
|
client.clean(database_name)
|
|
|
|
|
|
def test_same_name_diff_len():
|
|
"""
|
|
{
|
|
"title": "test_sys_verify.test_same_name_diff_len",
|
|
"describe": "查询的数据越界,tinyint的大小为[-128, 127]你们可以查小于这个最小值,以及大于这个最大值的查询",
|
|
"tag": "system,p0"
|
|
}
|
|
"""
|
|
"""
|
|
1.
|
|
针对上次的nmga的core你们可以加个case,就是查询的数据越界
|
|
就像tinyint的大小为[-128, 127]你们可以查小于这个最小值,以及大于这个最大值的查询
|
|
都是字段f,一个是int,一个是tinyint;或者一个是varchar(10),一个是varchar(20)
|
|
2.
|
|
char或varchar长度溢出时需要进行处理
|
|
改写逻辑如下: 对于varchar(5)
|
|
k1 < 'aaaaab' --> k1 <= 'aaaaa' 截断、同时改写
|
|
k1 <= 'aaaaab' --> k1 <= 'aaaaa' 只截断,不改写
|
|
k1 = 'aaaaab' --> 不下推
|
|
k1 > 'aaaaab' --> k1 > 'aaaaa' 只截断,不改写
|
|
k1 >= 'aaaaab' --> k1 > 'aaaaa' 截断、同时改写
|
|
"""
|
|
database_name, table_name, index_name = util.gen_num_format_name_list()
|
|
LOG.info(L('', database_name=database_name, \
|
|
table_name=table_name, index_name=index_name))
|
|
client.clean(database_name)
|
|
client.create_database(database_name)
|
|
client.use(database_name)
|
|
table_name_a = 'table_a'
|
|
table_name_b = 'table_b'
|
|
client.create_table(table_name_a, VERIFY_DATA.schema_5, storage_type='column', \
|
|
keys_desc='AGGREGATE KEY (K1, k2, k3, k4, k5, k6, k7, k8, k9, k10)')
|
|
client.create_table(table_name_b, VERIFY_DATA.schema_6, storage_type='column', \
|
|
keys_desc='AGGREGATE KEY (K1, k2, k3, k4, k5, k6, k7, k8, k9, k10)')
|
|
|
|
assert client.show_tables(table_name_a)
|
|
assert client.show_tables(table_name_b)
|
|
|
|
data_desc_list = palo_client.LoadDataInfo(VERIFY_DATA.file_path_3, table_name_a)
|
|
assert client.batch_load(util.get_label(), data_desc_list, is_wait=True, broker=broker_info)
|
|
|
|
data_desc_list = palo_client.LoadDataInfo(VERIFY_DATA.file_path_3, table_name_b)
|
|
assert client.batch_load(util.get_label(), data_desc_list, is_wait=True, broker=broker_info)
|
|
|
|
sql = 'select * from table_a where k1 > 130'
|
|
assert () == client.execute(sql)
|
|
sql = 'select * from table_a where k1 < -130'
|
|
assert () == client.execute(sql)
|
|
|
|
sql = 'select a.k1 as a_k1, b.k1 as b_k1, a.k9 as a_k9, b.k9 as b_k9 from table_a as a, ' \
|
|
'table_b as b where a.k9 = b.k9 order by a_k1, b_k1, a_k9, b_k9'
|
|
assert common.check_by_file(VERIFY_DATA.expected_file_1, sql=sql, client=client)
|
|
|
|
sql = "select a.k1 as a_k1, b.k1 as b_k1, a.k9 as a_k9, b.k9 as b_k9 " \
|
|
"from table_a as a, table_b as b " \
|
|
"where a.k9 > b.k9 and a.k9 = 'vzb' and b.k9 = 'ddsc' order by a_k1, b_k1, a_k9, b_k9"
|
|
assert common.check_by_file(VERIFY_DATA.expected_file_2, sql=sql, client=client)
|
|
|
|
sql = "select a.k1 as a_k1, b.k1 as b_k1, a.k9 as a_k9, b.k9 as b_k9 " \
|
|
"from table_a as a, table_b as b " \
|
|
"where a.k9 < b.k9 and a.k9 = 'ddsc' and b.k9 = 'vzb' order by a_k1, b_k1, a_k9, b_k9"
|
|
assert common.check_by_file(VERIFY_DATA.expected_file_3, sql=sql, client=client)
|
|
|
|
#NOTICE K9 varchar(10)
|
|
#TODO
|
|
sql = "select k9 from table_a where k9 = 'aaaaaaaaaa'"
|
|
assert () == client.execute(sql)
|
|
sql = "select k9 from table_a where k9 = 'aaaaaaaaaaa'"
|
|
print(client.execute(sql))
|
|
sql = "select k9 from table_a where k9 > 'aaaaaaaaaaa'"
|
|
print(client.execute(sql))
|
|
sql = "select k9 from table_a where k9 < 'aaaaaaaaaaa'"
|
|
print(client.execute(sql))
|
|
client.clean(database_name)
|
|
|
|
|
|
def test_not_support_row():
|
|
"""
|
|
{
|
|
"title": "test_sys_verify.test_not_support_row",
|
|
"describe": "不支持row存储,目前client中create_table,统一建column,不支持指定storage_type",
|
|
"tag": "system,p0,fuzz"
|
|
}
|
|
"""
|
|
"""
|
|
不支持row存储
|
|
目前client中create_table,统一建column,不支持指定storage_type
|
|
"""
|
|
database_name, table_name, index_name = util.gen_num_format_name_list()
|
|
LOG.info(L('', database_name=database_name, \
|
|
table_name=table_name, index_name=index_name))
|
|
client.clean(database_name)
|
|
client.create_database(database_name)
|
|
try:
|
|
ret = client.create_table(table_name,
|
|
VERIFY_DATA.schema_4, storage_type='row',
|
|
keys_desc='AGGREGATE KEY (K1)')
|
|
assert not ret
|
|
except Exception as e:
|
|
pass
|
|
client.clean(database_name)
|
|
|
|
|
|
def teardown_module():
|
|
"""
|
|
tearDown
|
|
"""
|
|
pass
|
|
|
|
|
|
if __name__ == '__main__':
|
|
import pdb
|
|
pdb.set_trace()
|
|
setup_module()
|
|
|