airflow-scheduler

Airflow: Log file isn't local, Unsupported remote log location

强颜欢笑 提交于 2019-11-30 12:13:37
I am not able see the logs attached to the tasks from the Airflow UI: Log related settings in airflow.cfg file are: remote_base_log_folder = base_log_folder = /home/my_projects/ksaprice_project/airflow/logs worker_log_server_port = 8793 child_process_log_directory = /home/my_projects/ksaprice_project/airflow/logs/scheduler Although I am setting remote_base_log_folter it is trying to fetch the log from http://:8793/log/tutorial/print_date/2017-08-02T00:00:00 - I don't understand this behavior. According to the settings the workers should store the logs at /home/my_projects/ksaprice_project

Airflow unpause dag programmatically?

我是研究僧i 提交于 2019-11-30 03:00:56
问题 I have a dag that we'll deploy to multiple different airflow instances and in our airflow.cfg we have dags_are_paused_at_creation = True but for this specific dag we want it to be turned on without having to do so manually by clicking on the UI. Is there a way to do it programmatically? 回答1: airflow-rest-api-plugin plugin can also be used to programmatically pause tasks. Pauses a DAG Available in Airflow Version: 1.7.0 or greater GET - http://{HOST}:{PORT}/admin/rest_api/api?api=pause Query

Airflow 1.10 Installation Failing

半世苍凉 提交于 2019-11-29 16:39:53
I have a working Airflow environment using Airflow version 1.9 that is running on an Amazon EC2-Instance. I need to upgrade to the latest version of Airflow which is 1.10. I have the option of either upgrading from version 1.9 or installing 1.10 freshly on a new server. Airflow version 1.10 is not listed on Pip so I'm installing it from Git via this command, pip-3.6 install git+git://github.com/apache/incubator-airflow.git@v1-10-stable This command successfully installs Airflow version 1.10. You can see that by running the command airflow version and viewing the output, ____________ __________

Airflow : ExternalTaskSensor doesn't trigger the task

北城余情 提交于 2019-11-29 15:25:01
I have already seen this and this questions on SO and made the changes accordingly. However, my dependent DAG still gets stuck in poking state. Below is my master DAG: from airflow import DAG from airflow.operators.jdbc_operator import JdbcOperator from datetime import datetime from airflow.operators.bash_operator import BashOperator today = datetime.today() default_args = { 'depends_on_past': False, 'retries': 0, 'start_date': datetime(today.year, today.month, today.day), 'schedule_interval': '@once' } dag = DAG('call-procedure-and-bash', default_args=default_args) call_procedure =

Airflow tasks get stuck at “queued” status and never gets running

假装没事ソ 提交于 2019-11-29 03:08:51
I'm using Airflow v1.8.1 and run all components (worker, web, flower, scheduler) on kubernetes & Docker. I use Celery Executor with Redis and my tasks are looks like: (start) -> (do_work_for_product1) ├ -> (do_work_for_product2) ├ -> (do_work_for_product3) ├ … So the start task has multiple downstreams. And I setup concurrency related configuration as below: parallelism = 3 dag_concurrency = 3 max_active_runs = 1 Then when I run this DAG manually (not sure if it never happens on a scheduled task) , some downstreams get executed, but others stuck at "queued" status. If I clear the task from

Airflow 1.10 Installation Failing

不想你离开。 提交于 2019-11-28 11:34:52
问题 I have a working Airflow environment using Airflow version 1.9 that is running on an Amazon EC2-Instance. I need to upgrade to the latest version of Airflow which is 1.10. I have the option of either upgrading from version 1.9 or installing 1.10 freshly on a new server. Airflow version 1.10 is not listed on Pip so I'm installing it from Git via this command, pip-3.6 install git+git://github.com/apache/incubator-airflow.git@v1-10-stable This command successfully installs Airflow version 1.10.

DAGs not clickable on Google Cloud Composer webserver, but working fine on a local Airflow

为君一笑 提交于 2019-11-28 11:15:30
I'm using Google Cloud Composer (managed Airflow on Google Cloud Platform) with image version composer-0.5.3-airflow-1.9.0 and Python 2.7, and I'm facing a weird issue : after importing my DAGs, they are not clickable from the Web UI (and there are no buttons "Trigger DAG", "Graph view", ...), while all works perfectly when running a local Airflow. Even if non usable from the webserver on Composer, my DAGs still exist. I can list them using CLI ( list_dags ), describe them ( list_tasks ) and even trigger them ( trigger_dag ). Minimal example reproducing the issue A minimal example I used to

Airflow : ExternalTaskSensor doesn't trigger the task

你。 提交于 2019-11-28 08:32:15
问题 I have already seen this and this questions on SO and made the changes accordingly. However, my dependent DAG still gets stuck in poking state. Below is my master DAG: from airflow import DAG from airflow.operators.jdbc_operator import JdbcOperator from datetime import datetime from airflow.operators.bash_operator import BashOperator today = datetime.today() default_args = { 'depends_on_past': False, 'retries': 0, 'start_date': datetime(today.year, today.month, today.day), 'schedule_interval'

Airflow tasks get stuck at “queued” status and never gets running

六眼飞鱼酱① 提交于 2019-11-27 17:19:38
问题 I'm using Airflow v1.8.1 and run all components (worker, web, flower, scheduler) on kubernetes & Docker. I use Celery Executor with Redis and my tasks are looks like: (start) -> (do_work_for_product1) ├ -> (do_work_for_product2) ├ -> (do_work_for_product3) ├ … So the start task has multiple downstreams. And I setup concurrency related configuration as below: parallelism = 3 dag_concurrency = 3 max_active_runs = 1 Then when I run this DAG manually (not sure if it never happens on a scheduled

Airflow 1.9.0 is queuing but not launching tasks

时间秒杀一切 提交于 2019-11-26 16:19:20
Airflow is randomly not running queued tasks some tasks dont even get queued status. I keep seeing below in the scheduler logs [2018-02-28 02:24:58,780] {jobs.py:1077} INFO - No tasks to consider for execution. I do see tasks in database that either have no status or queued status but they never get started. The airflow setup is running https://github.com/puckel/docker-airflow on ECS with Redis. There are 4 scheduler threads and 4 Celery worker tasks. For the tasks that are not running are showing in queued state (grey icon) when hovering over the task icon operator is null and task details