blob: d583a9b0d98c69bed07fa848c780ebfb918f3b86 [file] [log] [blame]
:mod:`airflow.contrib.sensors.bigquery_sensor`
==============================================
.. py:module:: airflow.contrib.sensors.bigquery_sensor
Module Contents
---------------
.. py:class:: BigQueryTableSensor(project_id, dataset_id, table_id, bigquery_conn_id='bigquery_default_conn', delegate_to=None, *args, **kwargs)
Bases: :class:`airflow.sensors.base_sensor_operator.BaseSensorOperator`
Checks for the existence of a table in Google Bigquery.
:param project_id: The Google cloud project in which to look for the table.
The connection supplied to the hook must provide
access to the specified project.
:type project_id: str
:param dataset_id: The name of the dataset in which to look for the table.
storage bucket.
:type dataset_id: str
:param table_id: The name of the table to check the existence of.
:type table_id: str
:param bigquery_conn_id: The connection ID to use when connecting to
Google BigQuery.
:type bigquery_conn_id: str
:param delegate_to: The account to impersonate, if any.
For this to work, the service account making the request must
have domain-wide delegation enabled.
:type delegate_to: str
.. attribute:: template_fields
:annotation: = ['project_id', 'dataset_id', 'table_id']
.. attribute:: ui_color
:annotation: = #f0eee4
.. method:: poke(self, context)