| # |
| # Licensed to the Apache Software Foundation (ASF) under one or more |
| # contributor license agreements. See the NOTICE file distributed with |
| # this work for additional information regarding copyright ownership. |
| # The ASF licenses this file to You under the Apache License, Version 2.0 |
| # (the "License"); you may not use this file except in compliance with |
| # the License. You may obtain a copy of the License at |
| # |
| # http://www.apache.org/licenses/LICENSE-2.0 |
| # |
| # Unless required by applicable law or agreed to in writing, software |
| # distributed under the License is distributed on an "AS IS" BASIS, |
| # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| # See the License for the specific language governing permissions and |
| # limitations under the License. |
| # |
| |
| """Unit tests for templated pipelines.""" |
| |
| # pytype: skip-file |
| |
| import json |
| import tempfile |
| import unittest |
| |
| import apache_beam as beam |
| from apache_beam.options.pipeline_options import PipelineOptions |
| from apache_beam.pipeline import Pipeline |
| from apache_beam.runners.dataflow.dataflow_runner import DataflowRunner |
| |
| # Protect against environments where apitools library is not available. |
| # pylint: disable=wrong-import-order, wrong-import-position |
| try: |
| from apache_beam.runners.dataflow.internal import apiclient |
| except ImportError: |
| apiclient = None # type: ignore |
| # pylint: enable=wrong-import-order, wrong-import-position |
| |
| |
| @unittest.skipIf(apiclient is None, 'GCP dependencies are not installed') |
| class TemplatingDataflowRunnerTest(unittest.TestCase): |
| """TemplatingDataflow tests.""" |
| def test_full_completion(self): |
| # Create dummy file and close it. Note that we need to do this because |
| # Windows does not allow NamedTemporaryFiles to be reopened elsewhere |
| # before the temporary file is closed. |
| dummy_file = tempfile.NamedTemporaryFile(delete=False) |
| dummy_file_name = dummy_file.name |
| dummy_file.close() |
| |
| dummy_dir = tempfile.mkdtemp() |
| |
| remote_runner = DataflowRunner() |
| options = PipelineOptions([ |
| '--sdk_location=' + dummy_file_name, |
| '--job_name=test-job', |
| '--project=apache-beam-testing', |
| '--region=us-central1', |
| '--staging_location=gs://apache-beam-testing-stg/stg/', |
| '--temp_location=gs://apache-beam-testing-temp/tmp', |
| '--template_location=' + dummy_file_name |
| ]) |
| with Pipeline(remote_runner, options) as pipeline: |
| pipeline | beam.Create([1, 2, 3]) | beam.Map(lambda x: x) # pylint: disable=expression-not-assigned |
| |
| with open(dummy_file_name) as template_file: |
| saved_job_dict = json.load(template_file) |
| self.assertEqual( |
| saved_job_dict['environment']['sdkPipelineOptions']['options'] |
| ['project'], |
| 'apache-beam-testing') |
| self.assertEqual( |
| saved_job_dict['environment']['sdkPipelineOptions']['options'] |
| ['job_name'], |
| 'test-job') |
| |
| |
| if __name__ == '__main__': |
| unittest.main() |