| <?xml version="1.0" encoding="UTF-8"?> |
| <!-- |
| Licensed to the Apache Software Foundation (ASF) under one or more |
| contributor license agreements. See the NOTICE file distributed with |
| this work for additional information regarding copyright ownership. |
| The ASF licenses this file to You under the Apache License, Version 2.0 |
| (the "License"); you may not use this file except in compliance with |
| the License. You may obtain a copy of the License at |
| |
| http://www.apache.org/licenses/LICENSE-2.0 |
| |
| Unless required by applicable law or agreed to in writing, software |
| distributed under the License is distributed on an "AS IS" BASIS, |
| WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| See the License for the specific language governing permissions and |
| limitations under the License. |
| --> |
| <project xmlns="http://maven.apache.org/POM/4.0.0" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 https://maven.apache.org/xsd/maven-4.0.0.xsd"> |
| <modelVersion>4.0.0</modelVersion> |
| <parent> |
| <groupId>org.apache.accumulo</groupId> |
| <artifactId>accumulo-project</artifactId> |
| <version>2.0.0</version> |
| </parent> |
| <artifactId>accumulo-hadoop-mapreduce</artifactId> |
| <name>Apache Accumulo Hadoop MapReduce</name> |
| <description>Apache Accumulo MapReduce bindings.</description> |
| <dependencies> |
| <dependency> |
| <groupId>com.beust</groupId> |
| <artifactId>jcommander</artifactId> |
| </dependency> |
| <dependency> |
| <groupId>com.google.guava</groupId> |
| <artifactId>guava</artifactId> |
| </dependency> |
| <dependency> |
| <groupId>org.apache.accumulo</groupId> |
| <artifactId>accumulo-core</artifactId> |
| </dependency> |
| <dependency> |
| <groupId>org.apache.hadoop</groupId> |
| <artifactId>hadoop-client-api</artifactId> |
| </dependency> |
| <dependency> |
| <groupId>org.slf4j</groupId> |
| <artifactId>slf4j-api</artifactId> |
| </dependency> |
| <dependency> |
| <groupId>org.apache.hadoop</groupId> |
| <artifactId>hadoop-client-runtime</artifactId> |
| <scope>runtime</scope> |
| </dependency> |
| <dependency> |
| <groupId>junit</groupId> |
| <artifactId>junit</artifactId> |
| <scope>test</scope> |
| </dependency> |
| <dependency> |
| <groupId>org.apache.accumulo</groupId> |
| <artifactId>accumulo-minicluster</artifactId> |
| <scope>test</scope> |
| </dependency> |
| <dependency> |
| <groupId>org.apache.accumulo</groupId> |
| <artifactId>accumulo-server-base</artifactId> |
| <scope>test</scope> |
| </dependency> |
| <dependency> |
| <groupId>org.apache.accumulo</groupId> |
| <artifactId>accumulo-test</artifactId> |
| <scope>test</scope> |
| </dependency> |
| <dependency> |
| <groupId>org.slf4j</groupId> |
| <artifactId>slf4j-log4j12</artifactId> |
| <scope>test</scope> |
| </dependency> |
| </dependencies> |
| <build> |
| <plugins> |
| <plugin> |
| <groupId>net.revelc.code</groupId> |
| <artifactId>apilyzer-maven-plugin</artifactId> |
| <executions> |
| <execution> |
| <id>apilyzer</id> |
| <goals> |
| <goal>analyze</goal> |
| </goals> |
| <configuration> |
| <includes> |
| <include>org[.]apache[.]accumulo[.]core[.]client[.]mapred(?:uce)?[.].*</include> |
| </includes> |
| <excludes> |
| <exclude>.*[.]impl[.].*</exclude> |
| </excludes> |
| <allows> |
| <allow>org[.]apache[.]accumulo[.]core[.](?:client|data|security)[.](?!.*(impl|thrift|crypto).*).*</allow> |
| <!-- Not public API, but hard to get rid of. Pair is just so useful. --> |
| <allow>org[.]apache[.]accumulo[.]core[.]util[.]Pair</allow> |
| <!--Types from hadoop used in API. If adding a new type from |
| Hadoop to the Accumulo API ensure its annotated as stable.--> |
| <allow>org[.]apache[.]hadoop[.]conf[.]Configuration</allow> |
| <allow>org[.]apache[.]hadoop[.]fs[.](FileSystem|Path)</allow> |
| <allow>org[.]apache[.]hadoop[.]io[.](Text|Writable|WritableComparable|WritableComparator)</allow> |
| <allow>org[.]apache[.]hadoop[.]mapred[.](JobConf|RecordReader|InputSplit|RecordWriter|Reporter)</allow> |
| <allow>org[.]apache[.]hadoop[.]mapred[.]FileOutputFormat[$]Counter</allow> |
| <allow>org[.]apache[.]hadoop[.]mapreduce[.](Job|JobContext|RecordReader|InputSplit|TaskAttemptContext|RecordWriter|OutputCommitter|TaskInputOutputContext)</allow> |
| <allow>org[.]apache[.]hadoop[.]mapreduce[.]lib[.]output[.]FileOutputFormat[$]Counter</allow> |
| <allow>org[.]apache[.]hadoop[.]util[.]Progressable</allow> |
| <allow>org[.]apache[.]hadoop[.]mapred[.](FileAlreadyExistsException|InvalidJobConfException)</allow> |
| </allows> |
| </configuration> |
| </execution> |
| </executions> |
| </plugin> |
| </plugins> |
| </build> |
| </project> |