| /* |
| * Licensed to the Apache Software Foundation (ASF) under one or more |
| * contributor license agreements. See the NOTICE file distributed with |
| * this work for additional information regarding copyright ownership. |
| * The ASF licenses this file to You under the Apache License, Version 2.0 |
| * (the "License"); you may not use this file except in compliance with |
| * the License. You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| |
| package org.apache.spark.util |
| |
| import java.io.File |
| import java.sql.Timestamp |
| import java.util.Date |
| |
| import org.apache.spark.sql.CarbonEnv |
| import org.apache.spark.sql.common.util.Spark2QueryTest |
| import org.scalatest.BeforeAndAfterAll |
| |
| import org.apache.carbondata.api.CarbonStore |
| import org.apache.carbondata.common.constants.LoggerAction |
| import org.apache.carbondata.core.constants.CarbonCommonConstants |
| import org.apache.carbondata.core.metadata.CarbonMetadata |
| import org.apache.carbondata.core.statusmanager.SegmentStatusManager |
| import org.apache.carbondata.core.util.CarbonProperties |
| |
| class CarbonCommandSuite extends Spark2QueryTest with BeforeAndAfterAll { |
| |
| val bad_records_action = CarbonProperties.getInstance() |
| .getProperty(CarbonCommonConstants.CARBON_BAD_RECORDS_ACTION) |
| |
| protected def createAndLoadInputTable(inputTableName: String, inputPath: String): Unit = { |
| sql( |
| s""" |
| | CREATE TABLE $inputTableName |
| | ( shortField short, |
| | intField int, |
| | bigintField long, |
| | doubleField double, |
| | stringField string, |
| | timestampField string, |
| | decimalField decimal(18,2), |
| | dateField string, |
| | charField char(5) |
| | ) |
| | ROW FORMAT DELIMITED FIELDS TERMINATED BY ',' |
| """.stripMargin) |
| |
| sql( |
| s""" |
| | LOAD DATA LOCAL INPATH '$inputPath' |
| | INTO TABLE $inputTableName |
| """.stripMargin) |
| } |
| |
| protected def createAndLoadTestTable(tableName: String, inputTableName: String): Unit = { |
| sql( |
| s""" |
| | CREATE TABLE $tableName( |
| | shortField short, |
| | intField int, |
| | bigintField long, |
| | doubleField double, |
| | stringField string, |
| | timestampField timestamp, |
| | decimalField decimal(18,2), |
| | dateField date, |
| | charField char(5) |
| | ) |
| | USING org.apache.spark.sql.CarbonSource |
| | OPTIONS ('tableName' '$tableName') |
| """.stripMargin) |
| sql( |
| s""" |
| | INSERT INTO TABLE $tableName |
| | SELECT shortField, intField, bigintField, doubleField, stringField, |
| | from_unixtime(unix_timestamp(timestampField,'yyyy/M/dd')) timestampField, decimalField, |
| | cast(to_date(from_unixtime(unix_timestamp(dateField,'yyyy/M/dd'))) as date), charField |
| | FROM $inputTableName |
| """.stripMargin) |
| } |
| |
| override def beforeAll(): Unit = { |
| CarbonProperties.getInstance().addProperty( |
| CarbonCommonConstants.CARBON_BAD_RECORDS_ACTION, LoggerAction.FORCE.name()) |
| dropTable("csv_table") |
| dropTable("carbon_table") |
| dropTable("carbon_table2") |
| createAndLoadInputTable("csv_table", s"$resourcesPath/data_alltypes.csv") |
| createAndLoadTestTable("carbon_table", "csv_table") |
| } |
| |
| override def afterAll(): Unit = { |
| dropTable("csv_table") |
| dropTable("carbon_table") |
| CarbonProperties.getInstance().addProperty( |
| CarbonCommonConstants.CARBON_BAD_RECORDS_ACTION, |
| bad_records_action) |
| |
| } |
| |
| private lazy val location = |
| CarbonProperties.getInstance().getProperty(CarbonCommonConstants.STORE_LOCATION) |
| |
| |
| test("delete segment by id") { |
| DeleteSegmentById.main(Array(s"${location}", "carbon_table", "0")) |
| assert(!CarbonStore.isSegmentValid("default", "carbon_table", location, "0")) |
| } |
| |
| test("delete segment by date") { |
| createAndLoadTestTable("carbon_table2", "csv_table") |
| val time = new Timestamp(new Date().getTime) |
| DeleteSegmentByDate.main(Array(s"${location}", "carbon_table2", time.toString)) |
| assert(!CarbonStore.isSegmentValid("default", "carbon_table2", location, "0")) |
| dropTable("carbon_table2") |
| } |
| |
| test("clean files") { |
| val table = "carbon_table3" |
| createAndLoadTestTable(table, "csv_table") |
| DeleteSegmentById.main(Array(s"${location}", table, "0")) |
| CleanFiles.main(Array(s"${location}", table)) |
| val carbonTable = CarbonMetadata.getInstance().getCarbonTable("default", table) |
| val tablePath = carbonTable.getAbsoluteTableIdentifier.getTablePath |
| val f = new File(s"$tablePath/Fact/Part0") |
| assert(f.isDirectory) |
| |
| // all segment folders should be deleted after CleanFiles command |
| assert(f.list().length == 0) |
| dropTable(table) |
| } |
| |
| test("clean files with force clean option") { |
| val table = "carbon_table4" |
| dropTable(table) |
| createAndLoadTestTable(table, "csv_table") |
| CleanFiles.main(Array(s"${location}", table, "true")) |
| val carbonTable = CarbonMetadata.getInstance().getCarbonTable("default", table) |
| val tablePath = carbonTable.getTablePath |
| val f = new File(tablePath) |
| assert(!f.exists()) |
| |
| dropTable(table) |
| } |
| |
| test("test if delete segments by id is unsupported for pre-aggregate tables") { |
| dropTable("preaggMain") |
| dropTable("preaggMain_preagg1") |
| sql("create table preaggMain (a string, b string, c string) stored by 'carbondata'") |
| sql("create datamap preagg1 on table PreAggMain using 'preaggregate' as select a,sum(b) from PreAggMain group by a") |
| intercept[UnsupportedOperationException] { |
| sql("delete from table preaggMain where segment.id in (1,2)") |
| }.getMessage.contains("Delete segment operation is not supported on tables") |
| intercept[UnsupportedOperationException] { |
| sql("delete from table preaggMain_preagg1 where segment.id in (1,2)") |
| }.getMessage.contains("Delete segment operation is not supported on pre-aggregate tables") |
| dropTable("preaggMain") |
| dropTable("preagg1") |
| } |
| |
| test("separate visible and invisible segments info into two files") { |
| val tableName = "test_tablestatus_history" |
| sql(s"drop table if exists ${tableName}") |
| sql(s"create table ${tableName} (name String, age int) stored by 'carbondata' " |
| + "TBLPROPERTIES('AUTO_LOAD_MERGE'='true','COMPACTION_LEVEL_THRESHOLD'='2,2')") |
| val carbonTable = CarbonEnv.getCarbonTable(Some("default"), tableName)(sqlContext.sparkSession) |
| sql(s"insert into ${tableName} select 'abc1',1") |
| sql(s"insert into ${tableName} select 'abc2',2") |
| sql(s"insert into ${tableName} select 'abc3',3") |
| sql(s"insert into ${tableName} select 'abc4',4") |
| sql(s"insert into ${tableName} select 'abc5',5") |
| sql(s"insert into ${tableName} select 'abc6',6") |
| assert(sql(s"show segments for table ${tableName}").collect().length == 10) |
| var detail = SegmentStatusManager.readLoadMetadata(carbonTable.getMetadataPath) |
| var historyDetail = SegmentStatusManager.readLoadHistoryMetadata(carbonTable.getMetadataPath) |
| assert(detail.length == 10) |
| assert(historyDetail.length == 0) |
| sql(s"clean files for table ${tableName}") |
| assert(sql(s"show segments for table ${tableName}").collect().length == 2) |
| detail = SegmentStatusManager.readLoadMetadata(carbonTable.getMetadataPath) |
| historyDetail = SegmentStatusManager.readLoadHistoryMetadata(carbonTable.getMetadataPath) |
| assert(detail.length == 4) |
| assert(historyDetail.length == 6) |
| dropTable(tableName) |
| } |
| |
| test("show history segments") { |
| val tableName = "test_tablestatus_history" |
| sql(s"drop table if exists ${tableName}") |
| sql(s"create table ${tableName} (name String, age int) stored by 'carbondata' " |
| + "TBLPROPERTIES('AUTO_LOAD_MERGE'='true','COMPACTION_LEVEL_THRESHOLD'='2,2')") |
| val carbonTable = CarbonMetadata.getInstance().getCarbonTable("default", tableName) |
| sql(s"insert into ${tableName} select 'abc1',1") |
| sql(s"insert into ${tableName} select 'abc2',2") |
| sql(s"insert into ${tableName} select 'abc3',3") |
| sql(s"insert into ${tableName} select 'abc4',4") |
| sql(s"insert into ${tableName} select 'abc5',5") |
| sql(s"insert into ${tableName} select 'abc6',6") |
| assert(sql(s"show segments for table ${tableName}").collect().length == 10) |
| assert(sql(s"show history segments for table ${tableName}").collect().length == 10) |
| sql(s"clean files for table ${tableName}") |
| assert(sql(s"show segments for table ${tableName}").collect().length == 2) |
| val segmentsHisotryList = sql(s"show history segments for table ${tableName}").collect() |
| assert(segmentsHisotryList.length == 10) |
| assert(segmentsHisotryList(0).getString(0).equalsIgnoreCase("5")) |
| assert(segmentsHisotryList(0).getString(6).equalsIgnoreCase("false")) |
| assert(segmentsHisotryList(0).getString(1).equalsIgnoreCase("Compacted")) |
| assert(segmentsHisotryList(1).getString(0).equalsIgnoreCase("4.1")) |
| assert(segmentsHisotryList(1).getString(6).equalsIgnoreCase("true")) |
| assert(segmentsHisotryList(1).getString(1).equalsIgnoreCase("Success")) |
| assert(segmentsHisotryList(3).getString(0).equalsIgnoreCase("3")) |
| assert(segmentsHisotryList(3).getString(6).equalsIgnoreCase("false")) |
| assert(segmentsHisotryList(3).getString(1).equalsIgnoreCase("Compacted")) |
| assert(segmentsHisotryList(7).getString(0).equalsIgnoreCase("0.2")) |
| assert(segmentsHisotryList(7).getString(6).equalsIgnoreCase("true")) |
| assert(segmentsHisotryList(7).getString(1).equalsIgnoreCase("Success")) |
| assert(sql(s"show history segments for table ${tableName} limit 3").collect().length == 3) |
| dropTable(tableName) |
| } |
| } |