blob: c1319f0fd07feb00c87aadc78d10426eb74b0d66 [file] [log] [blame]
<?xml version="1.0" encoding="UTF-8"?>
<!--
Licensed to the Apache Software Foundation (ASF) under one or more
contributor license agreements. See the NOTICE file distributed with
this work for additional information regarding copyright ownership.
The ASF licenses this file to You under the Apache License, Version 2.0
(the "License"); you may not use this file except in compliance with
the License. You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
-->
<project xmlns="http://maven.apache.org/POM/4.0.0" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 http://maven.apache.org/xsd/maven-4.0.0.xsd">
<modelVersion>4.0.0</modelVersion>
<parent>
<groupId>org.apache</groupId>
<artifactId>apache</artifactId>
<version>17</version>
</parent>
<groupId>org.apache.carbondata</groupId>
<artifactId>carbondata-parent</artifactId>
<name>Apache CarbonData :: Parent</name>
<description>Apache CarbonData(incubating) is an indexed columnar data format for fast analytics
on big data platform, e.g.Apache Hadoop, Apache Spark, etc.
</description>
<url>http://carbondata.incubator.apache.org</url>
<inceptionYear>2016</inceptionYear>
<packaging>pom</packaging>
<version>1.0.1-incubating-SNAPSHOT</version>
<licenses>
<license>
<name>Apache License, Version 2.0</name>
<url>http://www.apache.org/licenses/LICENSE-2.0.txt</url>
<distribution>repo</distribution>
</license>
</licenses>
<scm>
<connection>scm:git:https://git-wip-us.apache.org/repos/asf/incubator-carbondata.git</connection>
<developerConnection>scm:git:https://git-wip-us.apache.org/repos/asf/incubator-carbondata.git</developerConnection>
<url>https://git-wip-us.apache.org/repos/asf/incubator-carbondata.git</url>
<tag>HEAD</tag>
</scm>
<issueManagement>
<system>jira</system>
<url>https://issues.apache.org/jira/browse/CARBONDATA</url>
</issueManagement>
<mailingLists>
<mailingList>
<name>CarbonData Dev</name>
<subscribe>dev-subscribe@carbondata.incubator.apache.org</subscribe>
<unsubscribe>dev-unsubscribe@carbondata.incubator.apache.org</unsubscribe>
<post>dev@carbondata.incubator.apache.org</post>
<archive>http://www.mail-archive.com/dev%carbondata.incubator.apache.org</archive>
</mailingList>
<mailingList>
<name>CarbonData User</name>
<subscribe>user-subscribe@carbondata.incubator.apache.org</subscribe>
<unsubscribe>user-unsubscribe@carbondata.incubator.apache.org</unsubscribe>
<post>user@carbondata.incubator.apache.org</post>
<archive>http://www.mail-archive.com/user%carbondata.incubator.apache.org</archive>
</mailingList>
<mailingList>
<name>CarbonData Commits</name>
<subscribe>commits-subscribe@carbondata.incubator.apache.org</subscribe>
<unsubscribe>commits-unsubscribe@carbondata.incubator.apache.org</unsubscribe>
<post>commits@carbondata.incubator.apache.org</post>
<archive>http://www.mail-archive.com/commits%carbondata.incubator.apache.org</archive>
</mailingList>
</mailingLists>
<developers>
<developer>
<name>The Apache CarbonData Team</name>
<email>dev@carbondata.incubator.apache.org</email>
<url>http://carbondata.incubator.apache.org</url>
<organization>Apache Software Foundation</organization>
<organizationUrl>http://www.apache.org</organizationUrl>
</developer>
</developers>
<modules>
<module>common</module>
<module>core</module>
<module>processing</module>
<module>hadoop</module>
<module>integration/spark-common</module>
<module>integration/spark-common-test</module>
<module>assembly</module>
</modules>
<properties>
<project.build.sourceEncoding>UTF-8</project.build.sourceEncoding>
<snappy.version>1.1.2.6</snappy.version>
<hadoop.version>2.2.0</hadoop.version>
<kettle.version>4.4.0-stable</kettle.version>
<use.kettle>true</use.kettle>
<hadoop.deps.scope>compile</hadoop.deps.scope>
<spark.deps.scope>compile</spark.deps.scope>
<scala.deps.scope>compile</scala.deps.scope>
<dev.path>${basedir}/dev</dev.path>
</properties>
<repositories>
<repository>
<id>central</id>
<!-- This should be at top, it makes maven try the central repo first and then others and hence faster dep resolution -->
<name>Maven Repository</name>
<url>http://repo1.maven.org/maven2</url>
<releases>
<enabled>true</enabled>
</releases>
</repository>
<repository>
<id>pentaho-releases</id>
<url>http://repository.pentaho.org/artifactory/repo/</url>
</repository>
<repository>
<id>carbondata-releases</id>
<url>http://136.243.101.176:9091/repository/carbondata/</url>
</repository>
</repositories>
<dependencyManagement>
<dependencies>
<dependency>
<groupId>org.apache.hadoop</groupId>
<artifactId>hadoop-common</artifactId>
<version>${hadoop.version}</version>
<scope>${hadoop.deps.scope}</scope>
</dependency>
<dependency>
<groupId>org.apache.hadoop</groupId>
<artifactId>hadoop-hdfs</artifactId>
<version>${hadoop.version}</version>
<scope>${hadoop.deps.scope}</scope>
</dependency>
<dependency>
<groupId>org.apache.hadoop</groupId>
<artifactId>hadoop-mapreduce-client-core</artifactId>
<version>${hadoop.version}</version>
<scope>${hadoop.deps.scope}</scope>
</dependency>
<dependency>
<groupId>org.apache.spark</groupId>
<artifactId>spark-sql_${scala.binary.version}</artifactId>
<version>${spark.version}</version>
<scope>${spark.deps.scope}</scope>
</dependency>
<dependency>
<groupId>org.apache.spark</groupId>
<artifactId>spark-hive-thriftserver_${scala.binary.version}</artifactId>
<version>${spark.version}</version>
<scope>${spark.deps.scope}</scope>
</dependency>
<dependency>
<groupId>org.apache.spark</groupId>
<artifactId>spark-repl_${scala.binary.version}</artifactId>
<version>${spark.version}</version>
<scope>${spark.deps.scope}</scope>
</dependency>
<dependency>
<groupId>org.scala-lang</groupId>
<artifactId>scala-compiler</artifactId>
<version>${scala.version}</version>
<scope>${scala.deps.scope}</scope>
</dependency>
<dependency>
<groupId>org.scala-lang</groupId>
<artifactId>scala-reflect</artifactId>
<version>${scala.version}</version>
<scope>${scala.deps.scope}</scope>
</dependency>
<dependency>
<groupId>org.scala-lang</groupId>
<artifactId>scala-library</artifactId>
<version>${scala.version}</version>
<scope>${scala.deps.scope}</scope>
</dependency>
<dependency>
<groupId>org.scala-lang</groupId>
<artifactId>scala-actors</artifactId>
<version>${scala.version}</version>
<scope>${scala.deps.scope}</scope>
</dependency>
<dependency>
<groupId>org.scala-lang</groupId>
<artifactId>scalap</artifactId>
<version>${scala.version}</version>
<scope>${scala.deps.scope}</scope>
</dependency>
<dependency>
<groupId>org.jmockit</groupId>
<artifactId>jmockit</artifactId>
<version>1.10</version>
<exclusions>
<exclusion>
<groupId>*</groupId>
<artifactId>*</artifactId>
</exclusion>
</exclusions>
<scope>test</scope>
</dependency>
<dependency>
<groupId>junit</groupId>
<artifactId>junit</artifactId>
<version>4.11</version>
<scope>test</scope>
</dependency>
</dependencies>
</dependencyManagement>
<build>
<pluginManagement>
<plugins>
<plugin>
<groupId>org.apache.maven.plugins</groupId>
<artifactId>maven-javadoc-plugin</artifactId>
<configuration>
<additionalparam>-Xdoclint:missing</additionalparam>
</configuration>
</plugin>
<plugin>
<groupId>org.apache.maven.plugins</groupId>
<artifactId>maven-shade-plugin</artifactId>
<version>2.4.3</version>
</plugin>
</plugins>
</pluginManagement>
<plugins>
<plugin>
<groupId>org.apache.maven.plugins</groupId>
<artifactId>maven-checkstyle-plugin</artifactId>
<version>2.17</version>
<configuration>
<configLocation>${dev.path}/javastyle-config.xml</configLocation>
<suppressionsLocation>${dev.path}/javastyle-suppressions.xml</suppressionsLocation>
<headerLocation>${dev.path}/java.header</headerLocation>
<consoleOutput>true</consoleOutput>
<failsOnError>true</failsOnError>
<linkXRef>false</linkXRef>
<failOnViolation>true</failOnViolation>
<includeTestSourceDirectory>false</includeTestSourceDirectory>
<sourceDirectory>${basedir}/src/main/java</sourceDirectory>
<testSourceDirectory>${basedir}/src/test/java</testSourceDirectory>
<outputFile>${basedir}/target/checkstyle-output.xml</outputFile>
</configuration>
<executions>
<execution>
<goals>
<goal>check</goal>
</goals>
</execution>
</executions>
</plugin>
<plugin>
<groupId>org.scalastyle</groupId>
<artifactId>scalastyle-maven-plugin</artifactId>
<version>0.8.0</version>
<executions>
<execution>
<goals>
<goal>check</goal>
</goals>
</execution>
</executions>
<configuration>
<verbose>false</verbose>
<failOnViolation>true</failOnViolation>
<includeTestSourceDirectory>false</includeTestSourceDirectory>
<failOnWarning>false</failOnWarning>
<sourceDirectory>${basedir}/src/main/scala</sourceDirectory>
<testSourceDirectory>${basedir}/src/test/scala</testSourceDirectory>
<configLocation>${dev.path}/scalastyle-config.xml</configLocation>
<outputFile>${basedir}/target/scalastyle-output.xml</outputFile>
<outputEncoding>${project.build.sourceEncoding}</outputEncoding>
</configuration>
</plugin>
</plugins>
</build>
<profiles>
<profile>
<!--This profile does not build spark module, so user should explicitly give spark profile also like spark-1.6 -->
<id>build-with-format</id>
<modules>
<module>format</module>
</modules>
</profile>
<profile>
<!-- This profile only should be used for release prepare to cover all the modules -->
<id>build-all</id>
<properties>
<spark.version>1.6.2</spark.version>
<scala.binary.version>2.10</scala.binary.version>
<scala.version>2.10.4</scala.version>
<maven.test.skip>true</maven.test.skip>
<flink.version>1.1.4</flink.version>
</properties>
<modules>
<module>format</module>
<module>integration/spark</module>
<module>examples/spark</module>
<module>integration/spark2</module>
<module>examples/spark2</module>
<module>examples/flink</module>
</modules>
</profile>
<profile>
<id>hadoop-2.2.0</id>
<!-- default -->
<properties>
<hadoop.version>2.2.0</hadoop.version>
</properties>
</profile>
<profile>
<id>hadoop-2.7.2</id>
<properties>
<hadoop.version>2.7.2</hadoop.version>
</properties>
</profile>
<profile>
<id>spark-1.5</id>
<properties>
<spark.version>1.5.2</spark.version>
<scala.binary.version>2.10</scala.binary.version>
<scala.version>2.10.4</scala.version>
</properties>
<modules>
<module>integration/spark</module>
<module>examples/spark</module>
</modules>
</profile>
<profile>
<id>spark-1.6</id>
<activation>
<activeByDefault>true</activeByDefault>
</activation>
<properties>
<spark.version>1.6.2</spark.version>
<scala.binary.version>2.10</scala.binary.version>
<scala.version>2.10.4</scala.version>
</properties>
<modules>
<module>integration/spark</module>
<module>examples/spark</module>
</modules>
</profile>
<profile>
<id>spark-2.1</id>
<properties>
<spark.version>2.1.0</spark.version>
<scala.binary.version>2.11</scala.binary.version>
<scala.version>2.11.8</scala.version>
</properties>
<modules>
<module>integration/spark2</module>
<module>examples/spark2</module>
</modules>
</profile>
<profile>
<id>flink</id>
<properties>
<flink.version>1.1.4</flink.version>
<scala.binary.version>2.10</scala.binary.version>
</properties>
<modules>
<module>examples/flink</module>
</modules>
</profile>
<profile>
<id>findbugs</id>
<build>
<plugins>
<plugin>
<groupId>org.codehaus.mojo</groupId>
<artifactId>findbugs-maven-plugin</artifactId>
<version>3.0.0</version>
<configuration>
<excludeFilterFile>${dev.path}/findbugs-exclude.xml</excludeFilterFile>
<failOnError>true</failOnError>
<findbugsXmlOutput>true</findbugsXmlOutput>
<xmlOutput>true</xmlOutput>
<effort>Max</effort>
</configuration>
<executions>
<execution>
<id>analyze-compile</id>
<phase>compile</phase>
<goals>
<goal>check</goal>
</goals>
</execution>
</executions>
</plugin>
</plugins>
</build>
</profile>
<profile>
<id>include-all</id>
</profile>
<profile>
<id>no-kettle</id>
<properties>
<use.kettle>false</use.kettle>
</properties>
</profile>
<profile>
<id>rat</id>
<build>
<plugins>
<plugin>
<groupId>org.apache.rat</groupId>
<artifactId>apache-rat-plugin</artifactId>
<version>0.11</version>
<executions>
<execution>
<phase>verify</phase>
<goals>
<goal>check</goal>
</goals>
</execution>
</executions>
<configuration>
<reportFile>${project.build.directory}/${project.build.finalName}.rat</reportFile>
<excludeSubProjects>false</excludeSubProjects>
<useDefaultExcludes>true</useDefaultExcludes>
<excludes>
<exclude>**/*.iml</exclude>
<exclude>**/*.csv</exclude>
<exclude>**/*.dictionary</exclude>
<exclude>**/*.ktr</exclude>
<exclude>**/*.rat</exclude>
<exclude>**/_SUCCESS</exclude>
<exclude>**/non-csv</exclude>
<exclude>**/.invisibilityfile</exclude>
<exclude>**/noneCsvFormat.cs</exclude>
<exclude>**/org.apache.spark.sql.sources.DataSourceRegister</exclude>
<exclude>**/org.apache.spark.sql.test.TestQueryExecutorRegister</exclude>
<exclude>**/derby.log</exclude>
<exclude>**/meta.lock</exclude>
<exclude>**/loadmetadata.metadata</exclude>
<exclude>**/modifiedTime.mdt</exclude>
<exclude>**/PULL_REQUEST_TEMPLATE.md</exclude>
<exclude>**/dict.txt</exclude>
<exclude>**/sample</exclude>
<exclude>**/test.json</exclude>
</excludes>
</configuration>
</plugin>
</plugins>
</build>
</profile>
</profiles>
</project>