blob: d2522eabcc83aa2e1ad7c7257a348c12db184f90 [file] [log] [blame]
/**
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
import com.jayway.jsonpath.JsonPath
import org.apache.hadoop.gateway.shell.Hadoop
import org.apache.hadoop.gateway.shell.hdfs.Hdfs
import org.apache.hadoop.gateway.shell.job.Job
import static java.util.concurrent.TimeUnit.SECONDS
gateway = "https://localhost:8443/gateway/sample"
username = "mapred"
password = "mapred-password"
dataFile = "LICENSE"
jarFile = "samples/hadoop-examples.jar"
hadoop = Hadoop.login( gateway, username, password )
println "Delete /tmp/test " + Hdfs.rm(hadoop).file( "/tmp/test" ).recursive().now().statusCode
println "Create /tmp/test " + Hdfs.mkdir(hadoop).dir( "/tmp/test").now().statusCode
putData = Hdfs.put(hadoop).file( dataFile ).to( "/tmp/test/input/FILE" ).later() {
println "Put /tmp/test/input/FILE " + it.statusCode }
putJar = Hdfs.put(hadoop).file( jarFile ).to( "/tmp/test/hadoop-examples.jar" ).later() {
println "Put /tmp/test/hadoop-examples.jar " + it.statusCode }
hadoop.waitFor( putData, putJar )
jobId = Job.submitJava(hadoop) \
.jar( "/tmp/test/hadoop-examples.jar" ) \
.app( "wordcount" ) \
.input( "/tmp/test/input" ) \
.output( "/tmp/test/output" ) \
.now().jobId
println "Submitted job " + jobId
println "Polling for completion..."
done = false
count = 0
while( !done && count++ < 60 ) {
sleep( 1000 )
json = Job.queryStatus(hadoop).jobId(jobId).now().string
done = JsonPath.read( json, "\$.status.jobComplete" )
}
println "Done " + done
println "Shutdown " + hadoop.shutdown( 10, SECONDS )