blob: e86b62271aac3d7a4afab68a3834364bf8b162fe [file] [log] [blame]
:py:mod:`airflow.executors.debug_executor`
==========================================
.. py:module:: airflow.executors.debug_executor
.. autoapi-nested-parse::
DebugExecutor
.. seealso::
For more information on how the DebugExecutor works, take a look at the guide:
:ref:`executor:DebugExecutor`
Module Contents
---------------
Classes
~~~~~~~
.. autoapisummary::
airflow.executors.debug_executor.DebugExecutor
.. py:class:: DebugExecutor
Bases: :py:obj:`airflow.executors.base_executor.BaseExecutor`
This executor is meant for debugging purposes. It can be used with SQLite.
It executes one task instance at time. Additionally to support working
with sensors, all sensors ``mode`` will be automatically set to "reschedule".
.. py:method:: execute_async(*args, **kwargs)
The method is replaced by custom trigger_task implementation.
.. py:method:: sync()
Sync will get called periodically by the heartbeat method.
Executors should override this to perform gather statuses.
.. py:method:: queue_task_instance(task_instance, mark_success = False, pickle_id = None, ignore_all_deps = False, ignore_depends_on_past = False, ignore_task_deps = False, ignore_ti_state = False, pool = None, cfg_path = None)
Queues task instance with empty command because we do not need it.
.. py:method:: trigger_tasks(open_slots)
Triggers tasks. Instead of calling exec_async we just
add task instance to tasks_to_run queue.
:param open_slots: Number of open slots
.. py:method:: end()
When the method is called we just set states of queued tasks
to UPSTREAM_FAILED marking them as not executed.
.. py:method:: terminate()
This method is called when the daemon receives a SIGTERM
.. py:method:: change_state(key, state, info=None)
Changes state of the task.
:param info: Executor information for the task instance
:param key: Unique key for the task instance
:param state: State to set for the task.