Configuring CarbonData

This tutorial guides you through the advanced configurations of CarbonData :

System Configuration

This section provides the details of all the configurations required for the CarbonData System.

System Configuration in carbon.properties

PropertyDefault ValueDescription
carbon.storelocation/user/hive/warehouse/carbon.storeLocation where CarbonData will create the store, and write the data in its own format. NOTE: Store location should be in HDFS.
carbon.ddl.base.hdfs.urlhdfs://hacluster/opt/dataThis property is used to configure the HDFS relative path, the path configured in carbon.ddl.base.hdfs.url will be appended to the HDFS path configured in fs.defaultFS. If this path is configured, then user need not pass the complete path while dataload. For example: If absolute path of the csv file is hdfs://10.18.101.155:54310/data/cnbc/2016/xyz.csv, the path “hdfs://10.18.101.155:54310” will come from property fs.defaultFS and user can configure the /data/cnbc/ as carbon.ddl.base.hdfs.url. Now while dataload user can specify the csv path as /2016/xyz.csv.
carbon.badRecords.location/opt/Carbon/Spark/badrecordsPath where the bad records are stored.
carbon.data.file.version2If this parameter value is set to 1, then CarbonData will support the data load which is in old format(0.x version). If the value is set to 2(1.x onwards version), then CarbonData will support the data load of new format only.

Performance Configuration

This section provides the details of all the configurations required for CarbonData Performance Optimization.

Performance Configuration in carbon.properties

  • Data Loading Configuration
ParameterDefault ValueDescriptionRange
carbon.sort.file.buffer.size20File read buffer size used during sorting. This value is expressed in MB.Min=1 and Max=100
carbon.graph.rowset.size100000Rowset size exchanged between data load graph steps.Min=500 and Max=1000000
carbon.number.of.cores.while.loading6Number of cores to be used while loading data.
carbon.sort.size500000Record count to sort and write intermediate files to temp.
carbon.enableXXHashtrueAlgorithm for hashmap for hashkey calculation.
carbon.number.of.cores.block.sort7Number of cores to use for block sort while loading data.
carbon.max.driver.lru.cache.size-1Max LRU cache size upto which data will be loaded at the driver side. This value is expressed in MB. Default value of -1 means there is no memory limit for caching. Only integer values greater than 0 are accepted.
carbon.max.executor.lru.cache.size-1Max LRU cache size upto which data will be loaded at the executor side. This value is expressed in MB. Default value of -1 means there is no memory limit for caching. Only integer values greater than 0 are accepted. If this parameter is not configured, then the carbon.max.driver.lru.cache.size value will be considered.
carbon.merge.sort.prefetchtrueEnable prefetch of data during merge sort while reading data from sort temp files in data loading.
carbon.update.persist.enabletrueEnabling this parameter considers persistent data. Enabling this will reduce the execution time of UPDATE operation.
carbon.load.global.sort.partitions0The Number of partitions to use when shuffling data for sort. If user don't configurate or configurate it less than 1, it uses the number of map tasks as reduce tasks. In general, we recommend 2-3 tasks per CPU core in your cluster.
carbon.options.bad.records.logger.enablefalseWhether to create logs with details about bad records.
carbon.bad.records.actionfailThis property can have four types of actions for bad records FORCE, REDIRECT, IGNORE and FAIL. If set to FORCE then it auto-corrects the data by storing the bad records as NULL. If set to REDIRECT then bad records are written to the raw CSV instead of being loaded. If set to IGNORE then bad records are neither loaded nor written to the raw CSV. If set to FAIL then data loading fails if any bad records are found.
carbon.options.is.empty.data.bad.recordfalseIf false, then empty ("" or '' or ,,) data will not be considered as bad record and vice versa.
carbon.options.bad.record.pathSpecifies the HDFS path where bad records are stored. By default the value is Null. This path must to be configured by the user if bad record logger is enabled or bad record action redirect.
carbon.enable.vector.readertrueThis parameter increases the performance of select queries as it fetch columnar batch of size 4*1024 rows instead of fetching data row by row.
  • Compaction Configuration
ParameterDefault ValueDescriptionRange
carbon.number.of.cores.while.compacting2Number of cores which are used to write data during compaction.
carbon.compaction.level.threshold4, 3This property is for minor compaction which decides how many segments to be merged. Example: If it is set as 2, 3 then minor compaction will be triggered for every 2 segments. 3 is the number of level 1 compacted segment which is further compacted to new segment.Valid values are from 0-100.
carbon.major.compaction.size1024Major compaction size can be configured using this parameter. Sum of the segments which is below this threshold will be merged. This value is expressed in MB.
carbon.horizontal.compaction.enabletrueThis property is used to turn ON/OFF horizontal compaction. After every DELETE and UPDATE statement, horizontal compaction may occur in case the delta (DELETE/ UPDATE) files becomes more than specified threshold.
carbon.horizontal.UPDATE.compaction.threshold1This property specifies the threshold limit on number of UPDATE delta files within a segment. In case the number of delta files goes beyond the threshold, the UPDATE delta files within the segment becomes eligible for horizontal compaction and compacted into single UPDATE delta file.Values between 1 to 10000.
carbon.horizontal.DELETE.compaction.threshold1This property specifies the threshold limit on number of DELETE delta files within a block of a segment. In case the number of delta files goes beyond the threshold, the DELETE delta files for the particular block of the segment becomes eligible for horizontal compaction and compacted into single DELETE delta file.Values between 1 to 10000.
carbon.update.segment.parallelism1This property specifies the parallelism for each segment during update. If there are segments that contain too many records to update and the spark job encounter data-spill related errors, it is better to increase this property value. It is recommended to set this value to a multiple of the number of executors for balance.Values between 1 to 1000.
  • Query Configuration
ParameterDefault ValueDescriptionRange
carbon.number.of.cores4Number of cores to be used while querying.
carbon.enable.quick.filterfalseImproves the performance of filter query.

Miscellaneous Configuration

Extra Configuration in carbon.properties

  • Time format for CarbonData
ParameterDefault FormatDescription
carbon.timestamp.formatyyyy-MM-dd HH:mm:ssTimestamp format of input data used for timestamp data type.
  • Dataload Configuration
ParameterDefault ValueDescription
carbon.sort.file.write.buffer.size10485760File write buffer size used during sorting.
carbon.lock.typeLOCALLOCKThis configuration specifies the type of lock to be acquired during concurrent operations on table. There are following types of lock implementation: - LOCALLOCK: Lock is created on local file system as file. This lock is useful when only one spark driver (thrift server) runs on a machine and no other CarbonData spark application is launched concurrently. - HDFSLOCK: Lock is created on HDFS file system as file. This lock is useful when multiple CarbonData spark applications are launched and no ZooKeeper is running on cluster and HDFS supports file based locking.
carbon.sort.intermediate.files.limit20Minimum number of intermediate files after which merged sort can be started.
carbon.block.meta.size.reserved.percentage10Space reserved in percentage for writing block meta data in CarbonData file.
carbon.csv.read.buffersize.byte1048576csv reading buffer size.
high.cardinality.value100000To identify and apply compression for non-high cardinality columns.
carbon.merge.sort.reader.thread3Maximum no of threads used for reading intermediate files for final merging.
carbon.load.metadata.lock.retries3Maximum number of retries to get the metadata lock for loading data to table.
carbon.load.metadata.lock.retry.timeout.sec5Interval between the retries to get the lock.
carbon.tempstore.location/opt/Carbon/TempStoreLocTemporary store location. By default it takes System.getProperty(“java.io.tmpdir”).
carbon.load.log.counter500000Data loading records count logger.
  • Compaction Configuration
ParameterDefault ValueDescription
carbon.numberof.preserve.segments0If the user wants to preserve some number of segments from being compacted then he can set this property. Example: carbon.numberof.preserve.segments = 2 then 2 latest segments will always be excluded from the compaction. No segments will be preserved by default.
carbon.allowed.compaction.days0Compaction will merge the segments which are loaded with in the specific number of days configured. Example: If the configuration is 2, then the segments which are loaded in the time frame of 2 days only will get merged. Segments which are loaded 2 days apart will not be merged. This is disabled by default.
carbon.enable.auto.load.mergefalseTo enable compaction while data loading.
  • Query Configuration
ParameterDefault ValueDescription
max.query.execution.time60Maximum time allowed for one query to be executed. The value is in minutes.
carbon.enableMinMaxtrueMin max is feature added to enhance query performance. To disable this feature, set it false.
  • Global Dictionary Configurations
ParameterDefault ValueDescription
high.cardinality.identify.enabletrueIf the parameter is true, the high cardinality columns of the dictionary code are automatically recognized and these columns will not be used as global dictionary encoding. If the parameter is false, all dictionary encoding columns are used as dictionary encoding. The high cardinality column must meet the following requirements: value of cardinality > configured value of high.cardinality. Note: If SINGLE_PASS is used during data load, then this property will be disabled.
high.cardinality.threshold1000000It is a threshold to identify high cardinality of the columns.If the value of columns' cardinality > the configured value, then the columns are excluded from dictionary encoding.
carbon.cutOffTimestamp1970-01-01 05:30:00Sets the start date for calculating the timestamp. Java counts the number of milliseconds from start of “1970-01-01 00:00:00”. This property is used to customize the start of position. For example “2000-01-01 00:00:00”. The date must be in the form “carbon.timestamp.format”. NOTE: The CarbonData supports data store up to 68 years from the cut-off time defined. For example, if the cut-off time is 1970-01-01 05:30:00, then the data can be stored up to 2038-01-01 05:30:00.
carbon.timegranularitySECONDThe property used to set the data granularity level DAY, HOUR, MINUTE, or SECOND.

Spark Configuration

Spark Configuration Reference in spark-defaults.conf

ParameterDefault ValueDescription
spark.driver.memory1gAmount of memory to be used by the driver process.
spark.executor.memory1gAmount of memory to be used per executor process.

Dynamic Configuration In CarbonData Using SET-RESET

SET/RESET commands are used to add, update, display, or reset the carbondata properties dynamically without restarting the driver.

Syntax

  • Add or Update : This command adds or updates the value of parameter_name.
SET parameter_name=parameter_value
  • Display Property Value: This command displays the value of the specified parameter_name.
SET parameter_name
  • Display Session Parameters: This command displays all the supported session parameters.
SET
  • Display Session Parameters along with usage details: This command displays all the supported session parameters along with their usage details.
SET -v
  • Reset: This command clears all the session parameters.
RESET

Parameter Description:

ParameterDescription
parameter_nameName of the property whose value needs to be dynamically added, updated, or displayed.
parameter_valueNew value of the parameter_name to be set.

Dynamically Configurable Properties of CarbonData

PropertiesDescription
carbon.options.bad.records.logger.enableTo enable or disable bad record logger.
carbon.options.bad.records.actionThis property can have four types of actions for bad records FORCE, REDIRECT, IGNORE and FAIL. If set to FORCE then it auto-corrects the data by storing the bad records as NULL. If set to REDIRECT then bad records are written to the raw CSV instead of being loaded. If set to IGNORE then bad records are neither loaded nor written to the raw CSV. If set to FAIL then data loading fails if any bad records are found.
carbon.options.is.empty.data.bad.recordIf false, then empty ("" or '' or ,,) data will not be considered as bad record and vice versa.
carbon.options.sort.scopeThis property can have four possible values BATCH_SORT, LOCAL_SORT, GLOBAL_SORT and NO_SORT. If set to BATCH_SORT, the sorting scope is smaller and more index tree will be created,thus loading is faster but query maybe slower. If set to LOCAL_SORT, the sorting scope is bigger and one index tree per data node will be created, thus loading is slower but query is faster. If set to GLOBAL_SORT, the sorting scope is bigger and one index tree per task will be created, thus loading is slower but query is faster. If set to NO_SORT data will be loaded in unsorted manner.
carbon.options.batch.sort.size.inmbSize of batch data to keep in memory, as a thumb rule it supposed to be less than 45% of sort.inmemory.size.inmb otherwise it may spill intermediate data to disk.
carbon.options.single.passSingle Pass Loading enables single job to finish data loading with dictionary generation on the fly. It enhances performance in the scenarios where the subsequent data loading after initial load involves fewer incremental updates on the dictionary. This option specifies whether to use single pass for loading data or not. By default this option is set to FALSE.
carbon.options.bad.record.pathSpecifies the HDFS path where bad records needs to be stored.
carbon.options.global.sort.partitionsThe Number of partitions to use when shuffling data for sort. If user don't configurate or configurate it less than 1, it uses the number of map tasks as reduce tasks. In general, we recommend 2-3 tasks per CPU core in your cluster.
carbon.custom.block.distributionSpecifies whether to use the Spark or Carbon block distribution feature.
enable.unsafe.sortSpecifies whether to use unsafe sort during data loading. Unsafe sort reduces the garbage collection during data load operation, resulting in better performance.

Examples:

  • Add or Update:
SET enable.unsafe.sort =true
  • Display Property Value:
SET enable.unsafe.sort
  • Reset:
RESET

System Response:

  • Success will be recorded in the driver log.

  • Failure will be displayed in the UI.