| // Licensed to the Apache Software Foundation (ASF) under one |
| // or more contributor license agreements. See the NOTICE file |
| // distributed with this work for additional information |
| // regarding copyright ownership. The ASF licenses this file |
| // to you under the Apache License, Version 2.0 (the |
| // "License"); you may not use this file except in compliance |
| // with the License. You may obtain a copy of the License at |
| // |
| // http://www.apache.org/licenses/LICENSE-2.0 |
| // |
| // Unless required by applicable law or agreed to in writing, |
| // software distributed under the License is distributed on an |
| // "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY |
| // KIND, either express or implied. See the License for the |
| // specific language governing permissions and limitations |
| // under the License. |
| |
| import org.codehaus.groovy.runtime.IOGroovyMethods |
| |
| suite("test_compaction_dup_keys") { |
| def tableName = "compaction_dup_keys_regression_test" |
| |
| try { |
| //BackendId,Cluster,IP,HeartbeatPort,BePort,HttpPort,BrpcPort,LastStartTime,LastHeartbeat,Alive,SystemDecommissioned,ClusterDecommissioned,TabletNum,DataUsedCapacity,AvailCapacity,TotalCapacity,UsedPct,MaxDiskUsedPct,Tag,ErrMsg,Version,Status |
| String[][] backends = sql """ show backends; """ |
| assertTrue(backends.size() > 0) |
| String backend_id; |
| def backendId_to_backendIP = [:] |
| def backendId_to_backendHttpPort = [:] |
| for (String[] backend in backends) { |
| backendId_to_backendIP.put(backend[0], backend[2]) |
| backendId_to_backendHttpPort.put(backend[0], backend[5]) |
| } |
| |
| backend_id = backendId_to_backendIP.keySet()[0] |
| StringBuilder showConfigCommand = new StringBuilder(); |
| showConfigCommand.append("curl -X GET http://") |
| showConfigCommand.append(backendId_to_backendIP.get(backend_id)) |
| showConfigCommand.append(":") |
| showConfigCommand.append(backendId_to_backendHttpPort.get(backend_id)) |
| showConfigCommand.append("/api/show_config") |
| logger.info(showConfigCommand.toString()) |
| def process = showConfigCommand.toString().execute() |
| int code = process.waitFor() |
| String err = IOGroovyMethods.getText(new BufferedReader(new InputStreamReader(process.getErrorStream()))); |
| String out = process.getText() |
| logger.info("Show config: code=" + code + ", out=" + out + ", err=" + err) |
| assertEquals(code, 0) |
| def configList = parseJson(out.trim()) |
| assert configList instanceof List |
| |
| boolean disableAutoCompaction = true |
| for (Object ele in (List) configList) { |
| assert ele instanceof List<String> |
| if (((List<String>) ele)[0] == "disable_auto_compaction") { |
| disableAutoCompaction = Boolean.parseBoolean(((List<String>) ele)[2]) |
| } |
| } |
| |
| sql """ DROP TABLE IF EXISTS ${tableName} """ |
| sql """ |
| CREATE TABLE IF NOT EXISTS ${tableName} ( |
| `user_id` LARGEINT NOT NULL COMMENT "用户id", |
| `date` DATE NOT NULL COMMENT "数据灌入日期时间", |
| `datev2` DATEV2 NOT NULL COMMENT "数据灌入日期时间", |
| `datetimev2_1` DATETIMEV2(3) NOT NULL COMMENT "数据灌入日期时间", |
| `datetimev2_2` DATETIMEV2(6) NOT NULL COMMENT "数据灌入日期时间", |
| `city` VARCHAR(20) COMMENT "用户所在城市", |
| `age` SMALLINT COMMENT "用户年龄", |
| `sex` TINYINT COMMENT "用户性别", |
| `last_visit_date` DATETIME DEFAULT "1970-01-01 00:00:00" COMMENT "用户最后一次访问时间", |
| `last_update_date` DATETIME DEFAULT "1970-01-01 00:00:00" COMMENT "用户最后一次更新时间", |
| `datetime_val1` DATETIMEV2(3) DEFAULT "1970-01-01 00:00:00.111" COMMENT "用户最后一次访问时间", |
| `datetime_val2` DATETIME(6) DEFAULT "1970-01-01 00:00:00" COMMENT "用户最后一次更新时间", |
| `last_visit_date_not_null` DATETIME NOT NULL DEFAULT "1970-01-01 00:00:00" COMMENT "用户最后一次访问时间", |
| `cost` BIGINT DEFAULT "0" COMMENT "用户总消费", |
| `max_dwell_time` INT DEFAULT "0" COMMENT "用户最大停留时间", |
| `min_dwell_time` INT DEFAULT "99999" COMMENT "用户最小停留时间") |
| DUPLICATE KEY(`user_id`, `date`, `datev2`, `datetimev2_1`, `datetimev2_2`, `city`, `age`, `sex`) DISTRIBUTED BY HASH(`user_id`) |
| PROPERTIES ( "replication_num" = "1" ); |
| """ |
| |
| sql """ INSERT INTO ${tableName} VALUES |
| (1, '2017-10-01', '2017-10-01', '2017-10-01 11:11:11.110000', '2017-10-01 11:11:11.110111', 'Beijing', 10, 1, '2020-01-01', '2020-01-01', '2017-10-01 11:11:11.170000', '2017-10-01 11:11:11.110111', '2020-01-01', 1, 30, 20) |
| """ |
| |
| sql """ INSERT INTO ${tableName} VALUES |
| (1, '2017-10-01', '2017-10-01', '2017-10-01 11:11:11.110000', '2017-10-01 11:11:11.110111', 'Beijing', 10, 1, '2020-01-02', '2020-01-02', '2017-10-01 11:11:11.160000', '2017-10-01 11:11:11.100111', '2020-01-02', 1, 31, 19) |
| """ |
| |
| sql """ INSERT INTO ${tableName} VALUES |
| (2, '2017-10-01', '2017-10-01', '2017-10-01 11:11:11.110000', '2017-10-01 11:11:11.110111', 'Beijing', 10, 1, '2020-01-02', '2020-01-02', '2017-10-01 11:11:11.150000', '2017-10-01 11:11:11.130111', '2020-01-02', 1, 31, 21) |
| """ |
| |
| sql """ INSERT INTO ${tableName} VALUES |
| (2, '2017-10-01', '2017-10-01', '2017-10-01 11:11:11.110000', '2017-10-01 11:11:11.110111', 'Beijing', 10, 1, '2020-01-03', '2020-01-03', '2017-10-01 11:11:11.140000', '2017-10-01 11:11:11.120111', '2020-01-03', 1, 32, 20) |
| """ |
| |
| sql """ INSERT INTO ${tableName} VALUES |
| (3, '2017-10-01', '2017-10-01', '2017-10-01 11:11:11.110000', '2017-10-01 11:11:11.110111', 'Beijing', 10, 1, '2020-01-03', '2020-01-03', '2017-10-01 11:11:11.100000', '2017-10-01 11:11:11.140111', '2020-01-03', 1, 32, 22) |
| """ |
| |
| sql """ INSERT INTO ${tableName} VALUES |
| (3, '2017-10-01', '2017-10-01', '2017-10-01 11:11:11.110000', '2017-10-01 11:11:11.110111', 'Beijing', 10, 1, '2020-01-04', '2020-01-04', '2017-10-01 11:11:11.110000', '2017-10-01 11:11:11.150111', '2020-01-04', 1, 33, 21) |
| """ |
| |
| sql """ INSERT INTO ${tableName} VALUES |
| (3, '2017-10-01', '2017-10-01', '2017-10-01 11:11:11.110000', '2017-10-01 11:11:11.110111', 'Beijing', 10, 1, NULL, NULL, NULL, NULL, '2020-01-05', 1, 34, 20) |
| """ |
| |
| sql """ INSERT INTO ${tableName} VALUES |
| (4, '2017-10-01', '2017-10-01', '2017-10-01 11:11:11.110000', '2017-10-01 11:11:11.110111', 'Beijing', 10, 1, NULL, NULL, NULL, NULL, '2020-01-05', 1, 34, 20) |
| """ |
| |
| qt_select_default """ SELECT * FROM ${tableName} t ORDER BY user_id,date,city,age,sex,last_visit_date,last_update_date,last_visit_date_not_null,cost,max_dwell_time,min_dwell_time; """ |
| |
| //TabletId,ReplicaId,BackendId,SchemaHash,Version,LstSuccessVersion,LstFailedVersion,LstFailedTime,LocalDataSize,RemoteDataSize,RowCount,State,LstConsistencyCheckTime,CheckVersion,VersionCount,PathHash,MetaUrl,CompactionStatus |
| String[][] tablets = sql """ show tablets from ${tableName}; """ |
| |
| // trigger compactions for all tablets in ${tableName} |
| for (String[] tablet in tablets) { |
| String tablet_id = tablet[0] |
| backend_id = tablet[2] |
| StringBuilder sb = new StringBuilder(); |
| sb.append("curl -X POST http://") |
| sb.append(backendId_to_backendIP.get(backend_id)) |
| sb.append(":") |
| sb.append(backendId_to_backendHttpPort.get(backend_id)) |
| sb.append("/api/compaction/run?tablet_id=") |
| sb.append(tablet_id) |
| sb.append("&compact_type=cumulative") |
| |
| String command = sb.toString() |
| process = command.execute() |
| code = process.waitFor() |
| err = IOGroovyMethods.getText(new BufferedReader(new InputStreamReader(process.getErrorStream()))); |
| out = process.getText() |
| logger.info("Run compaction: code=" + code + ", out=" + out + ", err=" + err) |
| assertEquals(code, 0) |
| def compactJson = parseJson(out.trim()) |
| if (compactJson.status.toLowerCase() == "fail") { |
| assertEquals(disableAutoCompaction, false) |
| logger.info("Compaction was done automatically!") |
| } |
| if (disableAutoCompaction) { |
| assertEquals("success", compactJson.status.toLowerCase()) |
| } |
| } |
| |
| // wait for all compactions done |
| for (String[] tablet in tablets) { |
| boolean running = true |
| do { |
| Thread.sleep(1000) |
| String tablet_id = tablet[0] |
| backend_id = tablet[2] |
| StringBuilder sb = new StringBuilder(); |
| sb.append("curl -X GET http://") |
| sb.append(backendId_to_backendIP.get(backend_id)) |
| sb.append(":") |
| sb.append(backendId_to_backendHttpPort.get(backend_id)) |
| sb.append("/api/compaction/run_status?tablet_id=") |
| sb.append(tablet_id) |
| |
| String command = sb.toString() |
| logger.info(command) |
| process = command.execute() |
| code = process.waitFor() |
| err = IOGroovyMethods.getText(new BufferedReader(new InputStreamReader(process.getErrorStream()))); |
| out = process.getText() |
| logger.info("Get compaction status: code=" + code + ", out=" + out + ", err=" + err) |
| assertEquals(code, 0) |
| def compactionStatus = parseJson(out.trim()) |
| assertEquals("success", compactionStatus.status.toLowerCase()) |
| running = compactionStatus.run_status |
| } while (running) |
| } |
| |
| int rowCount = 0 |
| for (String[] tablet in tablets) { |
| String tablet_id = tablet[0] |
| StringBuilder sb = new StringBuilder(); |
| def compactionStatusUrlIndex = 17 |
| sb.append("curl -X GET ") |
| sb.append(tablet[compactionStatusUrlIndex]) |
| String command = sb.toString() |
| // wait for cleaning stale_rowsets |
| process = command.execute() |
| code = process.waitFor() |
| err = IOGroovyMethods.getText(new BufferedReader(new InputStreamReader(process.getErrorStream()))); |
| out = process.getText() |
| logger.info("Show tablets status: code=" + code + ", out=" + out + ", err=" + err) |
| assertEquals(code, 0) |
| def tabletJson = parseJson(out.trim()) |
| assert tabletJson.rowsets instanceof List |
| for (String rowset in (List<String>) tabletJson.rowsets) { |
| rowCount += Integer.parseInt(rowset.split(" ")[1]) |
| } |
| } |
| assert (rowCount <= 8) |
| qt_select_default2 """ SELECT * FROM ${tableName} t ORDER BY user_id,date,city,age,sex,last_visit_date,last_update_date,last_visit_date_not_null,cost,max_dwell_time,min_dwell_time; """ |
| } finally { |
| try_sql("DROP TABLE IF EXISTS ${tableName}") |
| } |
| } |