Merge pull request #14769: [BEAM-12325] Remove use of beam_fn_api experiment from Dataflow tests
diff --git a/.test-infra/jenkins/job_PerformanceTests_KafkaIO_IT.groovy b/.test-infra/jenkins/job_PerformanceTests_KafkaIO_IT.groovy
index 59cb4e9..87767f8 100644
--- a/.test-infra/jenkins/job_PerformanceTests_KafkaIO_IT.groovy
+++ b/.test-infra/jenkins/job_PerformanceTests_KafkaIO_IT.groovy
@@ -75,7 +75,7 @@
bigQueryTable : 'kafkaioit_results_sdf_wrapper',
influxMeasurement : 'kafkaioit_results_sdf_wrapper',
// TODO(BEAM-11779) remove shuffle_mode=appliance with runner v2 once issue is resolved.
- experiments : 'beam_fn_api,use_runner_v2,shuffle_mode=appliance,use_unified_worker',
+ experiments : 'use_runner_v2,shuffle_mode=appliance,use_unified_worker',
]
Map dataflowRunnerV2SdfPipelineOptions = pipelineOptions + [
@@ -90,7 +90,7 @@
bigQueryTable : 'kafkaioit_results_runner_v2',
influxMeasurement : 'kafkaioit_results_runner_v2',
// TODO(BEAM-11779) remove shuffle_mode=appliance with runner v2 once issue is resolved.
- experiments : 'beam_fn_api,use_runner_v2,shuffle_mode=appliance,use_unified_worker',
+ experiments : 'use_runner_v2,shuffle_mode=appliance,use_unified_worker',
]
steps {
diff --git a/runners/google-cloud-dataflow-java/build.gradle b/runners/google-cloud-dataflow-java/build.gradle
index 43f507d..e8cb290 100644
--- a/runners/google-cloud-dataflow-java/build.gradle
+++ b/runners/google-cloud-dataflow-java/build.gradle
@@ -151,7 +151,7 @@
"--tempRoot=${dataflowValidatesTempRoot}",
"--sdkContainerImage=${dockerImageContainer}:${dockerTag}",
// TODO(BEAM-11779) remove shuffle_mode=appliance with runner v2 once issue is resolved.
- "--experiments=beam_fn_api,use_unified_worker,use_runner_v2,shuffle_mode=appliance",
+ "--experiments=use_unified_worker,use_runner_v2,shuffle_mode=appliance",
]
def commonLegacyExcludeCategories = [
diff --git a/runners/google-cloud-dataflow-java/examples/build.gradle b/runners/google-cloud-dataflow-java/examples/build.gradle
index b128bc1..b52469b 100644
--- a/runners/google-cloud-dataflow-java/examples/build.gradle
+++ b/runners/google-cloud-dataflow-java/examples/build.gradle
@@ -41,7 +41,7 @@
def gcsTempRoot = project.findProperty('gcsTempRoot') ?: 'gs://temp-storage-for-end-to-end-tests/'
def dockerImageName = project(':runners:google-cloud-dataflow-java').ext.dockerImageName
// If -PuseExecutableStage is set, the use_executable_stage_bundle_execution wil be enabled.
-def fnapiExperiments = project.hasProperty('useExecutableStage') ? 'beam_fn_api,beam_fn_api_use_deprecated_read,use_executable_stage_bundle_execution' : "beam_fn_api,beam_fn_api_use_deprecated_read"
+def fnapiExperiments = project.hasProperty('useExecutableStage') ? 'beam_fn_api_use_deprecated_read,use_executable_stage_bundle_execution' : "beam_fn_api,beam_fn_api_use_deprecated_read"
def commonConfig = { dataflowWorkerJar, workerHarnessContainerImage = '', additionalOptions = [] ->
// return the preevaluated configuration closure
diff --git a/runners/google-cloud-dataflow-java/src/test/java/org/apache/beam/runners/dataflow/BatchStatefulParDoOverridesTest.java b/runners/google-cloud-dataflow-java/src/test/java/org/apache/beam/runners/dataflow/BatchStatefulParDoOverridesTest.java
index 12c9852..f6e683e 100644
--- a/runners/google-cloud-dataflow-java/src/test/java/org/apache/beam/runners/dataflow/BatchStatefulParDoOverridesTest.java
+++ b/runners/google-cloud-dataflow-java/src/test/java/org/apache/beam/runners/dataflow/BatchStatefulParDoOverridesTest.java
@@ -76,7 +76,7 @@
@Test
public void testFnApiSingleOutputOverrideNonCrashing() throws Exception {
- DataflowPipelineOptions options = buildPipelineOptions("--experiments=beam_fn_api");
+ DataflowPipelineOptions options = buildPipelineOptions();
options.setRunner(DataflowRunner.class);
Pipeline pipeline = Pipeline.create(options);
@@ -113,7 +113,7 @@
+ "exposes a way to know when the replacement is not required by checking that the "
+ "preceding ParDos to a GBK are key preserving.")
public void testFnApiMultiOutputOverrideNonCrashing() throws Exception {
- DataflowPipelineOptions options = buildPipelineOptions("--experiments=beam_fn_api");
+ DataflowPipelineOptions options = buildPipelineOptions();
options.setRunner(DataflowRunner.class);
Pipeline pipeline = Pipeline.create(options);