# # Licensed to the Apache Software Foundation (ASF) under one # or more contributor license agreements. See the NOTICE file # distributed with this work for additional information # regarding copyright ownership. The ASF licenses this file # to you under the Apache License, Version 2.0 (the # "License"); you may not use this file except in compliance # with the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, # software distributed under the License is distributed on an # "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY # KIND, either express or implied. See the License for the # specific language governing permissions and limitations # under the License. # Note: Any AirflowException raised is expected to cause the TaskInstance # to be marked in an ERROR state """Exceptions used by Airflow.""" from __future__ import annotations import warnings from http import HTTPStatus from typing import TYPE_CHECKING, Any, NamedTuple from airflow.utils.trigger_rule import TriggerRule if TYPE_CHECKING: import datetime from collections.abc import Sized from airflow.models import DAG, DagRun class AirflowException(Exception): """ Base class for all Airflow's errors. Each custom exception should be derived from this class. """ status_code = HTTPStatus.INTERNAL_SERVER_ERROR def serialize(self): cls = self.__class__ return f"{cls.__module__}.{cls.__name__}", (str(self),), {} class AirflowBadRequest(AirflowException): """Raise when the application or server cannot handle the request.""" status_code = HTTPStatus.BAD_REQUEST class AirflowNotFoundException(AirflowException): """Raise when the requested object/resource is not available in the system.""" status_code = HTTPStatus.NOT_FOUND class AirflowConfigException(AirflowException): """Raise when there is configuration problem.""" class AirflowSensorTimeout(AirflowException): """Raise when there is a timeout on sensor polling.""" class AirflowRescheduleException(AirflowException): """ Raise when the task should be re-scheduled at a later time. :param reschedule_date: The date when the task should be rescheduled """ def __init__(self, reschedule_date): super().__init__() self.reschedule_date = reschedule_date def serialize(self): cls = self.__class__ return f"{cls.__module__}.{cls.__name__}", (), {"reschedule_date": self.reschedule_date} class InvalidStatsNameException(AirflowException): """Raise when name of the stats is invalid.""" # Important to inherit BaseException instead of AirflowException->Exception, since this Exception is used # to explicitly interrupt ongoing task. Code that does normal error-handling should not treat # such interrupt as an error that can be handled normally. (Compare with KeyboardInterrupt) class AirflowTaskTimeout(BaseException): """Raise when the task execution times-out.""" class AirflowTaskTerminated(BaseException): """Raise when the task execution is terminated.""" class AirflowWebServerTimeout(AirflowException): """Raise when the web server times out.""" class AirflowSkipException(AirflowException): """Raise when the task should be skipped.""" class AirflowFailException(AirflowException): """Raise when the task should be failed without retrying.""" class AirflowOptionalProviderFeatureException(AirflowException): """Raise by providers when imports are missing for optional provider features.""" class AirflowInternalRuntimeError(BaseException): """ Airflow Internal runtime error. Indicates that something really terrible happens during the Airflow execution. :meta private: """ class XComNotFound(AirflowException): """Raise when an XCom reference is being resolved against a non-existent XCom.""" def __init__(self, dag_id: str, task_id: str, key: str) -> None: super().__init__() self.dag_id = dag_id self.task_id = task_id self.key = key def __str__(self) -> str: return f'XComArg result from {self.task_id} at {self.dag_id} with key="{self.key}" is not found!' def serialize(self): cls = self.__class__ return ( f"{cls.__module__}.{cls.__name__}", (), {"dag_id": self.dag_id, "task_id": self.task_id, "key": self.key}, ) class UnmappableOperator(AirflowException): """Raise when an operator is not implemented to be mappable.""" class XComForMappingNotPushed(AirflowException): """Raise when a mapped downstream's dependency fails to push XCom for task mapping.""" def __str__(self) -> str: return "did not push XCom for task mapping" class UnmappableXComTypePushed(AirflowException): """Raise when an unmappable type is pushed as a mapped downstream's dependency.""" def __init__(self, value: Any, *values: Any) -> None: super().__init__(value, *values) def __str__(self) -> str: typename = type(self.args[0]).__qualname__ for arg in self.args[1:]: typename = f"{typename}[{type(arg).__qualname__}]" return f"unmappable return type {typename!r}" class UnmappableXComLengthPushed(AirflowException): """Raise when the pushed value is too large to map as a downstream's dependency.""" def __init__(self, value: Sized, max_length: int) -> None: super().__init__(value) self.value = value self.max_length = max_length def __str__(self) -> str: return f"unmappable return value length: {len(self.value)} > {self.max_length}" class AirflowDagCycleException(AirflowException): """Raise when there is a cycle in DAG definition.""" class AirflowDagDuplicatedIdException(AirflowException): """Raise when a DAG's ID is already used by another DAG.""" def __init__(self, dag_id: str, incoming: str, existing: str) -> None: super().__init__(dag_id, incoming, existing) self.dag_id = dag_id self.incoming = incoming self.existing = existing def __str__(self) -> str: return f"Ignoring DAG {self.dag_id} from {self.incoming} - also found in {self.existing}" class AirflowDagInconsistent(AirflowException): """Raise when a DAG has inconsistent attributes.""" class AirflowClusterPolicyViolation(AirflowException): """Raise when there is a violation of a Cluster Policy in DAG definition.""" class AirflowClusterPolicySkipDag(AirflowException): """Raise when skipping dag is needed in Cluster Policy.""" class AirflowClusterPolicyError(AirflowException): """Raise for a Cluster Policy other than AirflowClusterPolicyViolation or AirflowClusterPolicySkipDag.""" class AirflowTimetableInvalid(AirflowException): """Raise when a DAG has an invalid timetable.""" class DagNotFound(AirflowNotFoundException): """Raise when a DAG is not available in the system.""" class DagCodeNotFound(AirflowNotFoundException): """Raise when a DAG code is not available in the system.""" class DagRunNotFound(AirflowNotFoundException): """Raise when a DAG Run is not available in the system.""" class DagRunAlreadyExists(AirflowBadRequest): """Raise when creating a DAG run for DAG which already has DAG run entry.""" def __init__(self, dag_run: DagRun, execution_date: datetime.datetime, run_id: str) -> None: super().__init__( f"A DAG Run already exists for DAG {dag_run.dag_id} at {execution_date} with run id {run_id}" ) self.dag_run = dag_run self.execution_date = execution_date self.run_id = run_id def serialize(self): cls = self.__class__ # Note the DagRun object will be detached here and fails serialization, we need to create a new one from airflow.models import DagRun dag_run = DagRun( state=self.dag_run.state, dag_id=self.dag_run.dag_id, run_id=self.dag_run.run_id, external_trigger=self.dag_run.external_trigger, run_type=self.dag_run.run_type, execution_date=self.dag_run.execution_date, ) dag_run.id = self.dag_run.id return ( f"{cls.__module__}.{cls.__name__}", (), {"dag_run": dag_run, "execution_date": self.execution_date, "run_id": self.run_id}, ) class DagFileExists(AirflowBadRequest): """Raise when a DAG ID is still in DagBag i.e., DAG file is in DAG folder.""" def __init__(self, *args, **kwargs): super().__init__(*args, **kwargs) warnings.warn("DagFileExists is deprecated and will be removed.", DeprecationWarning, stacklevel=2) class FailStopDagInvalidTriggerRule(AirflowException): """Raise when a dag has 'fail_stop' enabled yet has a non-default trigger rule.""" _allowed_rules = (TriggerRule.ALL_SUCCESS, TriggerRule.ALL_DONE_SETUP_SUCCESS) @classmethod def check(cls, *, dag: DAG | None, trigger_rule: TriggerRule): """ Check that fail_stop dag tasks have allowable trigger rules. :meta private: """ if dag is not None and dag.fail_stop and trigger_rule not in cls._allowed_rules: raise cls() def __str__(self) -> str: return f"A 'fail-stop' dag can only have {TriggerRule.ALL_SUCCESS} trigger rule" class DuplicateTaskIdFound(AirflowException): """Raise when a Task with duplicate task_id is defined in the same DAG.""" class TaskAlreadyInTaskGroup(AirflowException): """Raise when a Task cannot be added to a TaskGroup since it already belongs to another TaskGroup.""" def __init__(self, task_id: str, existing_group_id: str | None, new_group_id: str) -> None: super().__init__(task_id, new_group_id) self.task_id = task_id self.existing_group_id = existing_group_id self.new_group_id = new_group_id def __str__(self) -> str: if self.existing_group_id is None: existing_group = "the DAG's root group" else: existing_group = f"group {self.existing_group_id!r}" return f"cannot add {self.task_id!r} to {self.new_group_id!r} (already in {existing_group})" class SerializationError(AirflowException): """A problem occurred when trying to serialize something.""" class ParamValidationError(AirflowException): """Raise when DAG params is invalid.""" class TaskNotFound(AirflowNotFoundException): """Raise when a Task is not available in the system.""" class TaskInstanceNotFound(AirflowNotFoundException): """Raise when a task instance is not available in the system.""" class PoolNotFound(AirflowNotFoundException): """Raise when a Pool is not available in the system.""" class NoAvailablePoolSlot(AirflowException): """Raise when there is not enough slots in pool.""" class DagConcurrencyLimitReached(AirflowException): """Raise when DAG max_active_tasks limit is reached.""" class TaskConcurrencyLimitReached(AirflowException): """Raise when task max_active_tasks limit is reached.""" class BackfillUnfinished(AirflowException): """ Raises when not all tasks succeed in backfill. :param message: The human-readable description of the exception :param ti_status: The information about all task statuses """ def __init__(self, message, ti_status): super().__init__(message) self.ti_status = ti_status class FileSyntaxError(NamedTuple): """Information about a single error in a file.""" line_no: int | None message: str def __str__(self): return f"{self.message}. Line number: s{str(self.line_no)}," class AirflowFileParseException(AirflowException): """ Raises when connection or variable file can not be parsed. :param msg: The human-readable description of the exception :param file_path: A processed file that contains errors :param parse_errors: File syntax errors """ def __init__(self, msg: str, file_path: str, parse_errors: list[FileSyntaxError]) -> None: super().__init__(msg) self.msg = msg self.file_path = file_path self.parse_errors = parse_errors def __str__(self): from airflow.utils.code_utils import prepare_code_snippet from airflow.utils.platform import is_tty result = f"{self.msg}\nFilename: {self.file_path}\n\n" for error_no, parse_error in enumerate(self.parse_errors, 1): result += "=" * 20 + f" Parse error {error_no:3} " + "=" * 20 + "\n" result += f"{parse_error.message}\n" if parse_error.line_no: result += f"Line number: {parse_error.line_no}\n" if parse_error.line_no and is_tty(): result += "\n" + prepare_code_snippet(self.file_path, parse_error.line_no) + "\n" return result class ConnectionNotUnique(AirflowException): """Raise when multiple values are found for the same connection ID.""" class TaskDeferred(BaseException): """ Signal an operator moving to deferred state. Special exception raised to signal that the operator it was raised from wishes to defer until a trigger fires. Triggers can send execution back to task or end the task instance directly. If the trigger should end the task instance itself, ``method_name`` does not matter, and can be None; otherwise, provide the name of the method that should be used when resuming execution in the task. """ def __init__( self, *, trigger, method_name: str, kwargs: dict[str, Any] | None = None, timeout: datetime.timedelta | None = None, ): super().__init__() self.trigger = trigger self.method_name = method_name self.kwargs = kwargs self.timeout = timeout # Check timeout type at runtime if self.timeout is not None and not hasattr(self.timeout, "total_seconds"): raise ValueError("Timeout value must be a timedelta") def serialize(self): cls = self.__class__ return ( f"{cls.__module__}.{cls.__name__}", (), { "trigger": self.trigger, "method_name": self.method_name, "kwargs": self.kwargs, "timeout": self.timeout, }, ) def __repr__(self) -> str: return f"" class TaskDeferralError(AirflowException): """Raised when a task failed during deferral for some reason.""" # The try/except handling is needed after we moved all k8s classes to cncf.kubernetes provider # These two exceptions are used internally by Kubernetes Executor but also by PodGenerator, so we need # to leave them here in case older version of cncf.kubernetes provider is used to run KubernetesPodOperator # and it raises one of those exceptions. The code should be backwards compatible even if you import # and try/except the exception using direct imports from airflow.exceptions. # 1) if you have old provider, both provider and pod generator will throw the "airflow.exceptions" exception. # 2) if you have new provider, both provider and pod generator will throw the # "airflow.providers.cncf.kubernetes" as it will be imported here from the provider. try: from airflow.providers.cncf.kubernetes.pod_generator import PodMutationHookException except ImportError: class PodMutationHookException(AirflowException): # type: ignore[no-redef] """Raised when exception happens during Pod Mutation Hook execution.""" try: from airflow.providers.cncf.kubernetes.pod_generator import PodReconciliationError except ImportError: class PodReconciliationError(AirflowException): # type: ignore[no-redef] """Raised when an error is encountered while trying to merge pod configs.""" class RemovedInAirflow3Warning(DeprecationWarning): """Issued for usage of deprecated features that will be removed in Airflow3.""" deprecated_since: str | None = None "Indicates the airflow version that started raising this deprecation warning" class AirflowProviderDeprecationWarning(DeprecationWarning): """Issued for usage of deprecated features of Airflow provider.""" deprecated_provider_since: str | None = None "Indicates the provider version that started raising this deprecation warning" class DeserializingResultError(ValueError): """Raised when an error is encountered while a pickling library deserializes a pickle file.""" def __str__(self): return ( "Error deserializing result. Note that result deserialization " "is not supported across major Python versions. Cause: " + str(self.__cause__) ) class UnknownExecutorException(ValueError): """Raised when an attempt is made to load an executor which is not configured."""