blob: d1439bafb7488c2b299b9bac4de73b9273945d2f [file] [log] [blame]
/*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* License); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an AS IS BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
plugins { id 'org.apache.beam.module' }
applyJavaNature(
publish: false,
archivesBaseName: 'beam-sdks-java-load-tests',
exportJavadoc: false
)
description = "Apache Beam :: SDKs :: Java :: Load Tests"
def mainClassProperty = "loadTest.mainClass"
// When running via Gradle, this property can be used to pass commandline arguments
// to the load-tests launch
def loadTestArgsProperty = "loadTest.args"
// When running via Gradle, this property sets the runner dependency
def runnerProperty = "runner"
def runnerDependency = (project.hasProperty(runnerProperty)
? project.getProperty(runnerProperty)
: ":runners:direct-java")
def loadTestRunnerVersionProperty = "runner.version"
def loadTestRunnerVersion = project.findProperty(loadTestRunnerVersionProperty)
def isSparkRunner = runnerDependency.startsWith(":runners:spark:")
def isDataflowRunner = ":runners:google-cloud-dataflow-java".equals(runnerDependency)
def isDataflowRunnerV2 = isDataflowRunner && "V2".equals(loadTestRunnerVersion)
def runnerConfiguration = ":runners:direct-java".equals(runnerDependency) ? "shadow" : null
if (isDataflowRunner) {
/*
* We need to rely on manually specifying these evaluationDependsOn to ensure that
* the following projects are evaluated before we evaluate this project. This is because
* we are attempting to reference a property from the project directly.
*/
if (isDataflowRunnerV2) {
evaluationDependsOn(":runners:google-cloud-dataflow-java")
} else {
evaluationDependsOn(":runners:google-cloud-dataflow-java:worker")
}
}
configurations {
// A configuration for running the Load testlauncher directly from Gradle, which
// uses Gradle to put the appropriate dependencies on the Classpath rather than
// bundling them into a fat jar
gradleRun
}
dependencies {
implementation enforcedPlatform(library.java.google_cloud_platform_libraries_bom)
implementation library.java.kafka_clients
implementation project(path: ":sdks:java:core", configuration: "shadow")
implementation project(":sdks:java:io:synthetic")
implementation project(":sdks:java:testing:test-utils")
implementation project(":sdks:java:io:google-cloud-platform")
implementation project(":sdks:java:io:kafka")
implementation project(":sdks:java:io:kinesis")
implementation library.java.aws_java_sdk_core
implementation library.java.google_cloud_core
implementation library.java.joda_time
implementation library.java.vendored_guava_32_1_2_jre
implementation library.java.slf4j_api
gradleRun project(project.path)
gradleRun project(path: runnerDependency, configuration: runnerConfiguration)
}
if (isSparkRunner) {
configurations.gradleRun {
// Using Spark runner causes a StackOverflowError if slf4j-jdk14 is on the classpath
exclude group: "org.slf4j", module: "slf4j-jdk14"
}
}
def getLoadTestArgs = {
def loadTestArgs = project.findProperty(loadTestArgsProperty) ?: ""
def loadTestArgsList = new ArrayList<String>()
Collections.addAll(loadTestArgsList as Collection<? super Collection>, loadTestArgs.split())
if (isDataflowRunner) {
if (isDataflowRunnerV2) {
loadTestArgsList.add("--experiments=beam_fn_api,use_unified_worker,use_runner_v2,shuffle_mode=service")
def sdkContainerImage = project.findProperty('sdkContainerImage') ?: project(":runners:google-cloud-dataflow-java").dockerJavaImageName
loadTestArgsList.add("--sdkContainerImage=${sdkContainerImage}")
} else {
def dataflowWorkerJar = project.findProperty('dataflowWorkerJar') ?: project(":runners:google-cloud-dataflow-java:worker").shadowJar.archivePath
// Provide job with a customizable worker jar.
// With legacy worker jar, containerImage is set to empty (i.e. to use the internal build).
// More context and discussions can be found in PR#6694.
loadTestArgsList.add("--dataflowWorkerJar=${dataflowWorkerJar}")
loadTestArgsList.add("--workerHarnessContainerImage=")
}
}
return loadTestArgsList
}
task run(type: JavaExec) {
def loadTestArgsList = getLoadTestArgs()
if (isDataflowRunner) {
if (isDataflowRunnerV2){
dependsOn ":runners:google-cloud-dataflow-java:buildAndPushDockerJavaContainer"
finalizedBy ":runners:google-cloud-dataflow-java:cleanUpDockerJavaImages"
} else {
dependsOn ":runners:google-cloud-dataflow-java:worker:shadowJar"
}
}
if(isSparkRunner){
// Disable UI
systemProperty "spark.ui.enabled", "false"
systemProperty "spark.ui.showConsoleProgress", "false"
}
mainClass = project.findProperty(mainClassProperty)
classpath = configurations.gradleRun
args loadTestArgsList.toArray()
}