// Copyright 2017 PingCAP, Inc. // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // See the License for the specific language governing permissions and // limitations under the License. package ast_test import ( "strings" "testing" "github.com/pingcap/tidb/pkg/parser" . "github.com/pingcap/tidb/pkg/parser/ast" "github.com/pingcap/tidb/pkg/parser/format" "github.com/pingcap/tidb/pkg/parser/mysql" "github.com/pingcap/tidb/pkg/parser/test_driver" "github.com/stretchr/testify/require" ) func TestFunctionsVisitorCover(t *testing.T) { valueExpr := NewValueExpr(42, mysql.DefaultCharset, mysql.DefaultCollationName) stmts := []Node{ &AggregateFuncExpr{Args: []ExprNode{valueExpr}}, &FuncCallExpr{Args: []ExprNode{valueExpr}}, &FuncCastExpr{Expr: valueExpr}, &WindowFuncExpr{Spec: WindowSpec{}}, } for _, stmt := range stmts { stmt.Accept(visitor{}) stmt.Accept(visitor1{}) } } func TestFuncCallExprRestore(t *testing.T) { testCases := []NodeRestoreTestCase{ {"JSON_ARRAYAGG(attribute)", "JSON_ARRAYAGG(`attribute`)"}, {"JSON_OBJECTAGG(attribute, value)", "JSON_OBJECTAGG(`attribute`, `value`)"}, {"ABS(-1024)", "ABS(-1024)"}, {"ACOS(3.14)", "ACOS(3.14)"}, {"CONV('a',16,2)", "CONV(_UTF8MB4'a', 16, 2)"}, {"COS(PI())", "COS(PI())"}, {"RAND()", "RAND()"}, {"ADDDATE('2000-01-01', 1)", "ADDDATE(_UTF8MB4'2000-01-01', INTERVAL 1 DAY)"}, {"DATE_ADD('2000-01-01', INTERVAL 1 DAY)", "DATE_ADD(_UTF8MB4'2000-01-01', INTERVAL 1 DAY)"}, {"DATE_ADD('2000-01-01', INTERVAL '1 1:12:23.100000' DAY_MICROSECOND)", "DATE_ADD(_UTF8MB4'2000-01-01', INTERVAL _UTF8MB4'1 1:12:23.100000' DAY_MICROSECOND)"}, {"EXTRACT(DAY FROM '2000-01-01')", "EXTRACT(DAY FROM _UTF8MB4'2000-01-01')"}, {"extract(day from '1999-01-01')", "EXTRACT(DAY FROM _UTF8MB4'1999-01-01')"}, {"GET_FORMAT(DATE, 'EUR')", "GET_FORMAT(DATE, _UTF8MB4'EUR')"}, {"POSITION('a' IN 'abc')", "POSITION(_UTF8MB4'a' IN _UTF8MB4'abc')"}, {"TRIM(' bar ')", "TRIM(_UTF8MB4' bar ')"}, {"TRIM('a' FROM ' bar ')", "TRIM(_UTF8MB4'a' FROM _UTF8MB4' bar ')"}, {"TRIM(LEADING FROM ' bar ')", "TRIM(LEADING _UTF8MB4' ' FROM _UTF8MB4' bar ')"}, {"TRIM(BOTH FROM ' bar ')", "TRIM(BOTH _UTF8MB4' ' FROM _UTF8MB4' bar ')"}, {"TRIM(TRAILING FROM ' bar ')", "TRIM(TRAILING _UTF8MB4' ' FROM _UTF8MB4' bar ')"}, {"TRIM(LEADING 'x' FROM 'xxxyxxx')", "TRIM(LEADING _UTF8MB4'x' FROM _UTF8MB4'xxxyxxx')"}, {"TRIM(BOTH 'x' FROM 'xxxyxxx')", "TRIM(BOTH _UTF8MB4'x' FROM _UTF8MB4'xxxyxxx')"}, {"TRIM(TRAILING 'x' FROM 'xxxyxxx')", "TRIM(TRAILING _UTF8MB4'x' FROM _UTF8MB4'xxxyxxx')"}, {"TRIM(BOTH col1 FROM col2)", "TRIM(BOTH `col1` FROM `col2`)"}, {"DATE_ADD('2008-01-02', INTERVAL INTERVAL(1, 0, 1) DAY)", "DATE_ADD(_UTF8MB4'2008-01-02', INTERVAL INTERVAL(1, 0, 1) DAY)"}, {"BENCHMARK(1000000, AES_ENCRYPT('text', UNHEX('F3229A0B371ED2D9441B830D21A390C3')))", "BENCHMARK(1000000, AES_ENCRYPT(_UTF8MB4'text', UNHEX(_UTF8MB4'F3229A0B371ED2D9441B830D21A390C3')))"}, {"SUBSTRING('Quadratically', 5)", "SUBSTRING(_UTF8MB4'Quadratically', 5)"}, {"SUBSTRING('Quadratically' FROM 5)", "SUBSTRING(_UTF8MB4'Quadratically', 5)"}, {"SUBSTRING('Quadratically', 5, 6)", "SUBSTRING(_UTF8MB4'Quadratically', 5, 6)"}, {"SUBSTRING('Quadratically' FROM 5 FOR 6)", "SUBSTRING(_UTF8MB4'Quadratically', 5, 6)"}, {"JSON_TYPE('[123]')", "JSON_TYPE(_UTF8MB4'[123]')"}, {"bit_and(all c1)", "BIT_AND(`c1`)"}, {"nextval(seq)", "NEXTVAL(`seq`)"}, {"nextval(test.seq)", "NEXTVAL(`test`.`seq`)"}, {"lastval(seq)", "LASTVAL(`seq`)"}, {"lastval(test.seq)", "LASTVAL(`test`.`seq`)"}, {"setval(seq, 100)", "SETVAL(`seq`, 100)"}, {"setval(test.seq, 100)", "SETVAL(`test`.`seq`, 100)"}, {"next value for seq", "NEXTVAL(`seq`)"}, {"next value for test.seq", "NEXTVAL(`test`.`seq`)"}, {"next value for sequence", "NEXTVAL(`sequence`)"}, {"NeXt vAluE for seQuEncE2", "NEXTVAL(`seQuEncE2`)"}, {"NeXt vAluE for test.seQuEncE2", "NEXTVAL(`test`.`seQuEncE2`)"}, {"weight_string(a)", "WEIGHT_STRING(`a`)"}, {"Weight_stRing(test.a)", "WEIGHT_STRING(`test`.`a`)"}, {"weight_string('a')", "WEIGHT_STRING(_UTF8MB4'a')"}, // Expressions with collations of different charsets will lead to an error in MySQL, but the error check should be done in TiDB, so it's valid here. {"weight_string('a' collate utf8_general_ci collate utf8mb4_general_ci)", "WEIGHT_STRING(_UTF8MB4'a' COLLATE utf8_general_ci COLLATE utf8mb4_general_ci)"}, {"weight_string(_utf8 'a' collate utf8_general_ci)", "WEIGHT_STRING(_UTF8'a' COLLATE utf8_general_ci)"}, {"weight_string(_utf8 'a')", "WEIGHT_STRING(_UTF8'a')"}, {"weight_string(a as char(5))", "WEIGHT_STRING(`a` AS CHAR(5))"}, {"weight_string(a as character(5))", "WEIGHT_STRING(`a` AS CHAR(5))"}, {"weight_string(a as binary(5))", "WEIGHT_STRING(`a` AS BINARY(5))"}, {"hex(weight_string('abc' as binary(5)))", "HEX(WEIGHT_STRING(_UTF8MB4'abc' AS BINARY(5)))"}, {"soundex(attr)", "SOUNDEX(`attr`)"}, {"soundex('string')", "SOUNDEX(_UTF8MB4'string')"}, } extractNodeFunc := func(node Node) Node { return node.(*SelectStmt).Fields.Fields[0].Expr } runNodeRestoreTest(t, testCases, "select %s", extractNodeFunc) } func TestFuncCastExprRestore(t *testing.T) { testCases := []NodeRestoreTestCase{ {"CONVERT('Müller' USING UtF8)", "CONVERT(_UTF8MB4'Müller' USING 'utf8')"}, {"CONVERT('Müller' USING UtF8Mb4)", "CONVERT(_UTF8MB4'Müller' USING 'utf8mb4')"}, {"CONVERT('Müller', CHAR(32) CHARACTER SET UtF8)", "CONVERT(_UTF8MB4'Müller', CHAR(32) CHARSET UTF8)"}, {"CAST('test' AS CHAR CHARACTER SET UtF8)", "CAST(_UTF8MB4'test' AS CHAR CHARSET UTF8)"}, {"BINARY 'New York'", "BINARY _UTF8MB4'New York'"}, } extractNodeFunc := func(node Node) Node { return node.(*SelectStmt).Fields.Fields[0].Expr } runNodeRestoreTest(t, testCases, "select %s", extractNodeFunc) } func TestAggregateFuncExprRestore(t *testing.T) { testCases := []NodeRestoreTestCase{ {"AVG(test_score)", "AVG(`test_score`)"}, {"AVG(distinct test_score)", "AVG(DISTINCT `test_score`)"}, {"BIT_AND(test_score)", "BIT_AND(`test_score`)"}, {"BIT_OR(test_score)", "BIT_OR(`test_score`)"}, {"BIT_XOR(test_score)", "BIT_XOR(`test_score`)"}, {"COUNT(test_score)", "COUNT(`test_score`)"}, {"COUNT(*)", "COUNT(1)"}, {"COUNT(DISTINCT scores, results)", "COUNT(DISTINCT `scores`, `results`)"}, {"MIN(test_score)", "MIN(`test_score`)"}, {"MIN(DISTINCT test_score)", "MIN(DISTINCT `test_score`)"}, {"MAX(test_score)", "MAX(`test_score`)"}, {"MAX(DISTINCT test_score)", "MAX(DISTINCT `test_score`)"}, {"STD(test_score)", "STDDEV_POP(`test_score`)"}, {"STDDEV(test_score)", "STDDEV_POP(`test_score`)"}, {"STDDEV_POP(test_score)", "STDDEV_POP(`test_score`)"}, {"STDDEV_SAMP(test_score)", "STDDEV_SAMP(`test_score`)"}, {"SUM(test_score)", "SUM(`test_score`)"}, {"SUM(DISTINCT test_score)", "SUM(DISTINCT `test_score`)"}, {"VAR_POP(test_score)", "VAR_POP(`test_score`)"}, {"VAR_SAMP(test_score)", "VAR_SAMP(`test_score`)"}, {"VARIANCE(test_score)", "VAR_POP(`test_score`)"}, {"JSON_OBJECTAGG(test_score, results)", "JSON_OBJECTAGG(`test_score`, `results`)"}, {"GROUP_CONCAT(a)", "GROUP_CONCAT(`a` SEPARATOR ',')"}, {"GROUP_CONCAT(a separator '--')", "GROUP_CONCAT(`a` SEPARATOR '--')"}, {"GROUP_CONCAT(a order by b desc, c)", "GROUP_CONCAT(`a` ORDER BY `b` DESC,`c` SEPARATOR ',')"}, {"GROUP_CONCAT(a order by b desc, c separator '--')", "GROUP_CONCAT(`a` ORDER BY `b` DESC,`c` SEPARATOR '--')"}, } extractNodeFunc := func(node Node) Node { return node.(*SelectStmt).Fields.Fields[0].Expr } runNodeRestoreTest(t, testCases, "select %s", extractNodeFunc) } func TestConvert(t *testing.T) { // Test case for CONVERT(expr USING transcoding_name). cases := []struct { SQL string CharsetName string ErrorMessage string }{ {`SELECT CONVERT("abc" USING "latin1")`, "latin1", ""}, {`SELECT CONVERT("abc" USING laTiN1)`, "latin1", ""}, {`SELECT CONVERT("abc" USING "binary")`, "binary", ""}, {`SELECT CONVERT("abc" USING biNaRy)`, "binary", ""}, {`SELECT CONVERT(a USING a)`, "", `[parser:1115]Unknown character set: 'a'`}, // TiDB issue #4436. {`SELECT CONVERT("abc" USING CONCAT("utf", "8"))`, "", `[parser:1115]Unknown character set: 'CONCAT'`}, } for _, testCase := range cases { stmt, err := parser.New().ParseOneStmt(testCase.SQL, "", "") if testCase.ErrorMessage != "" { require.EqualError(t, err, testCase.ErrorMessage) continue } require.NoError(t, err) st := stmt.(*SelectStmt) expr := st.Fields.Fields[0].Expr.(*FuncCallExpr) charsetArg := expr.Args[1].(*test_driver.ValueExpr) require.Equal(t, testCase.CharsetName, charsetArg.GetString()) } } func TestChar(t *testing.T) { // Test case for CHAR(N USING charset_name) cases := []struct { SQL string CharsetName string ErrorMessage string }{ {`SELECT CHAR("abc" USING "latin1")`, "latin1", ""}, {`SELECT CHAR("abc" USING laTiN1)`, "latin1", ""}, {`SELECT CHAR("abc" USING "binary")`, "binary", ""}, {`SELECT CHAR("abc" USING binary)`, "binary", ""}, {`SELECT CHAR(a USING a)`, "", `[parser:1115]Unknown character set: 'a'`}, {`SELECT CHAR("abc" USING CONCAT("utf", "8"))`, "", `[parser:1115]Unknown character set: 'CONCAT'`}, } for _, testCase := range cases { stmt, err := parser.New().ParseOneStmt(testCase.SQL, "", "") if testCase.ErrorMessage != "" { require.EqualError(t, err, testCase.ErrorMessage) continue } require.NoError(t, err) st := stmt.(*SelectStmt) expr := st.Fields.Fields[0].Expr.(*FuncCallExpr) charsetArg := expr.Args[1].(*test_driver.ValueExpr) require.Equal(t, testCase.CharsetName, charsetArg.GetString()) } } func TestWindowFuncExprRestore(t *testing.T) { testCases := []NodeRestoreTestCase{ {"RANK() OVER w", "RANK() OVER `w`"}, {"RANK() OVER (PARTITION BY a)", "RANK() OVER (PARTITION BY `a`)"}, {"MAX(DISTINCT a) OVER (PARTITION BY a)", "MAX(DISTINCT `a`) OVER (PARTITION BY `a`)"}, {"MAX(DISTINCTROW a) OVER (PARTITION BY a)", "MAX(DISTINCT `a`) OVER (PARTITION BY `a`)"}, {"MAX(DISTINCT ALL a) OVER (PARTITION BY a)", "MAX(DISTINCT `a`) OVER (PARTITION BY `a`)"}, {"MAX(ALL a) OVER (PARTITION BY a)", "MAX(`a`) OVER (PARTITION BY `a`)"}, {"FIRST_VALUE(val) IGNORE NULLS OVER (w)", "FIRST_VALUE(`val`) IGNORE NULLS OVER (`w`)"}, {"FIRST_VALUE(val) RESPECT NULLS OVER w", "FIRST_VALUE(`val`) OVER `w`"}, {"NTH_VALUE(val, 233) FROM LAST IGNORE NULLS OVER w", "NTH_VALUE(`val`, 233) FROM LAST IGNORE NULLS OVER `w`"}, {"NTH_VALUE(val, 233) FROM FIRST IGNORE NULLS OVER (w)", "NTH_VALUE(`val`, 233) IGNORE NULLS OVER (`w`)"}, } extractNodeFunc := func(node Node) Node { return node.(*SelectStmt).Fields.Fields[0].Expr } runNodeRestoreTest(t, testCases, "select %s from t", extractNodeFunc) } func TestGenericFuncRestore(t *testing.T) { testCases := []NodeRestoreTestCase{ {"s.a()", "`s`.`a`()"}, {"`s`.`a`()", "`s`.`a`()"}, {"now()", "NOW()"}, {"`s`.`now`()", "`s`.`now`()"}, // FIXME: expectSQL should be `generic_func()`. {"generic_func()", "GENERIC_FUNC()"}, {"`ident.1`.`ident.2`()", "`ident.1`.`ident.2`()"}, } extractNodeFunc := func(node Node) Node { return node.(*SelectStmt).Fields.Fields[0].Expr } runNodeRestoreTest(t, testCases, "select %s from t", extractNodeFunc) } func TestRestoreWithError(t *testing.T) { testCases := []string{ "json_memberof()", } extractNodeFunc := func(node Node) Node { return node.(*SelectStmt).Fields.Fields[0].Expr } for _, c := range testCases { sql := "select " + c p := parser.New() stmt, err := p.ParseOneStmt(sql, "", "") require.NoError(t, err) var sb strings.Builder require.Error(t, extractNodeFunc(stmt).Restore(format.NewRestoreCtx(format.DefaultRestoreFlags, &sb))) } }