123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293 |
- # Licensed to the Apache Software Foundation (ASF) under one
- # or more contributor license agreements. See the NOTICE file
- # distributed with this work for additional information
- # regarding copyright ownership. The ASF licenses this file
- # to you under the Apache License, Version 2.0 (the
- # "License"); you may not use this file except in compliance
- # with the License. You may obtain a copy of the License at
- #
- # http://www.apache.org/licenses/LICENSE-2.0
- #
- # Unless required by applicable law or agreed to in writing,
- # software distributed under the License is distributed on an
- # "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
- # KIND, either express or implied. See the License for the
- # specific language governing permissions and limitations
- # under the License.
- """Scheduler command."""
- from __future__ import annotations
- import logging
- from argparse import Namespace
- from contextlib import contextmanager
- from multiprocessing import Process
- from airflow import settings
- from airflow.cli.commands.daemon_utils import run_command_with_daemon_option
- from airflow.configuration import conf
- from airflow.executors.executor_loader import ExecutorLoader
- from airflow.jobs.job import Job, run_job
- from airflow.jobs.scheduler_job_runner import SchedulerJobRunner
- from airflow.utils import cli as cli_utils
- from airflow.utils.cli import process_subdir
- from airflow.utils.providers_configuration_loader import providers_configuration_loaded
- from airflow.utils.scheduler_health import serve_health_check
- log = logging.getLogger(__name__)
- def _run_scheduler_job(args) -> None:
- job_runner = SchedulerJobRunner(
- job=Job(), subdir=process_subdir(args.subdir), num_runs=args.num_runs, do_pickle=args.do_pickle
- )
- ExecutorLoader.validate_database_executor_compatibility(job_runner.job.executor.__class__)
- enable_health_check = conf.getboolean("scheduler", "ENABLE_HEALTH_CHECK")
- with _serve_logs(args.skip_serve_logs), _serve_health_check(enable_health_check):
- run_job(job=job_runner.job, execute_callable=job_runner._execute)
- @cli_utils.action_cli
- @providers_configuration_loaded
- def scheduler(args: Namespace):
- """Start Airflow Scheduler."""
- print(settings.HEADER)
- run_command_with_daemon_option(
- args=args,
- process_name="scheduler",
- callback=lambda: _run_scheduler_job(args),
- should_setup_logging=True,
- )
- @contextmanager
- def _serve_logs(skip_serve_logs: bool = False):
- """Start serve_logs sub-process."""
- from airflow.utils.serve_logs import serve_logs
- sub_proc = None
- executor_class, _ = ExecutorLoader.import_default_executor_cls()
- if executor_class.serve_logs:
- if skip_serve_logs is False:
- sub_proc = Process(target=serve_logs)
- sub_proc.start()
- try:
- yield
- finally:
- if sub_proc:
- sub_proc.terminate()
- @contextmanager
- def _serve_health_check(enable_health_check: bool = False):
- """Start serve_health_check sub-process."""
- sub_proc = None
- if enable_health_check:
- sub_proc = Process(target=serve_health_check)
- sub_proc.start()
- try:
- yield
- finally:
- if sub_proc:
- sub_proc.terminate()
|