tree: 8c90b507ff86367f38c4296c84cbdc216c762996 [path history] [tgz]
  1. src/
  2. pom.xml
  3. Presto_Cluster_Setup_For_Carbondata.md
  4. README.md
integration/presto/README.md

Please follow the below steps to query carbondata in presto

Config presto server

  • Download presto server (0.210 is suggested and supported) : https://repo1.maven.org/maven2/com/facebook/presto/presto-server/

  • Finish presto configuration following https://prestodb.io/docs/current/installation/deployment.html. A configuration example:

    config.properties:
    coordinator=true
    node-scheduler.include-coordinator=true
    http-server.http.port=8086
    query.max-memory=5GB
    query.max-total-memory-per-node=5GB
    query.max-memory-per-node=3GB
    memory.heap-headroom-per-node=1GB
    discovery-server.enabled=true
    discovery.uri=http://localhost:8086
    task.max-worker-threads=4
    optimizer.dictionary-aggregation=true
    optimizer.optimize-hash-generation = false
    
    
    jvm.config:
    -server
    -Xmx4G
    -XX:+UseG1GC
    -XX:G1HeapRegionSize=32M
    -XX:+UseGCOverheadLimit
    -XX:+ExplicitGCInvokesConcurrent
    -XX:+HeapDumpOnOutOfMemoryError
    -XX:OnOutOfMemoryError=kill -9 %p
    -XX:+TraceClassLoading
    
    log.properties:
    com.facebook.presto=DEBUG
    com.facebook.presto.server.PluginManager=DEBUG
    
    node.properties:
    node.environment=carbondata
    node.id=ffffffff-ffff-ffff-ffff-ffffffffffff
    node.data-dir=/Users/apple/DEMO/presto_test/data
    
  • Config carbondata-connector for presto

    Firstly: Compile carbondata, including carbondata-presto integration module

    $ git clone https://github.com/apache/carbondata
    $ cd carbondata
    $ mvn -DskipTests -P{spark-version} -Dspark.version={spark-version-number} -Dhadoop.version={hadoop-version-number} clean package
    

    Replace the spark and hadoop version with the version used in your cluster. For example, if you are using Spark 2.2.1 and Hadoop 2.7.2, you would like to compile using:

    mvn -DskipTests -Pspark-2.2 -Dspark.version=2.2.1 -Dhadoop.version=2.7.2 clean package
    

    Secondly: Create a folder named ‘carbondata’ under $PRESTO_HOME$/plugin and copy all jars from carbondata/integration/presto/target/carbondata-presto-x.x.x-SNAPSHOT to $PRESTO_HOME$/plugin/carbondata

    Thirdly: Create a carbondata.properties file under $PRESTO_HOME$/etc/catalog/ containing the following contents:

    connector.name=carbondata
    carbondata-store={schema-store-path}
    enable.unsafe.in.query.processing=false
    carbon.unsafe.working.memory.in.mb={value}
    enable.unsafe.columnpage=false
    enable.unsafe.sort=false
    
    

    Replace the schema-store-path with the absolute path of the parent directory of the schema. For example, if you have a schema named ‘default’ stored in hdfs://namenode:9000/test/carbondata/, Then set carbondata-store=hdfs://namenode:9000/test/carbondata

Connecting to carbondata store on s3

  • In case you want to query carbonstore on S3 using S3A api put following additional properties inside $PRESTO_HOME$/etc/catalog/carbondata.properties

     Required properties
    
     fs.s3a.access.key={value}
     fs.s3a.secret.key={value}
    
     Optional properties
    
     fs.s3a.endpoint={value}
    
  • In case you want to query carbonstore on s3 using S3 api put following additional properties inside $PRESTO_HOME$/etc/catalog/carbondata.properties

      fs.s3.awsAccessKeyId={value}
      fs.s3.awsSecretAccessKey={value}
    
  • In case You want to query carbonstore on s3 using S3N api put following additional properties inside $PRESTO_HOME$/etc/catalog/carbondata.properties

        fs.s3n.awsAccessKeyId={value}
        fs.s3n.awsSecretAccessKey={value}
    

    Replace the schema-store-path with the absolute path of the parent directory of the schema. For example, if you have a schema named ‘default’ stored in a bucket s3a://s3-carbon/store, Then set carbondata-store=s3a://s3-carbon/store

Unsafe Properties

enable.unsafe.in.query.processing property by default is true in CarbonData system, the carbon.unsafe.working.memory.in.mb property defines the limit for Unsafe Memory usage in Mega Bytes, the default value is 512 MB. Currently Presto does not support Unsafe Memory so we have to disable the unsafe feature by setting below properties to false.

enable.unsafe.in.query.processing=false. enable.unsafe.columnpage=false enable.unsafe.sort=false

If you updated the jar balls or configuration files, make sure you have dispatched them to all the presto nodes and restarted the presto servers on the nodes. The updates will not take effect before restarting.

Generate CarbonData file

Please refer to quick start: https://github.com/apache/carbondata/blob/master/docs/quick-start-guide.md. Load data statement in Spark can be used to create carbondata tables. And then you can easily find the created carbondata files.

Query carbondata in CLI of presto