airflow-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "wei.he (JIRA)" <j...@apache.org>
Subject [jira] [Comment Edited] (AIRFLOW-357) how should I use the right owner task in airflow?
Date Tue, 26 Jul 2016 03:21:20 GMT

    [ https://issues.apache.org/jira/browse/AIRFLOW-357?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15393130#comment-15393130
] 

wei.he edited comment on AIRFLOW-357 at 7/26/16 3:21 AM:
---------------------------------------------------------

I don't  fix this problem.

My dag code is the following.
{code:title=test3.py|borderStyle=solid}
from airflow.models import DAG
from airflow.operators import BashOperator
from datetime import datetime, timedelta
rootdir = "/tmp/airflow"
default_args = {
    'owner': 'max',
    'depends_on_past': False,
    'start_date': datetime.now(),
    'email': ['max@test.com'],
    'email_on_failure': False,
    'email_on_retry': False,
    'retries': 1,
    'retry_delay': timedelta(minutes=5),
}
dag = DAG('test3', default_args=default_args,
                    schedule_interval='*/2 * * * *')
t1 = BashOperator(
    task_id='test3-task1',
    bash_command='date  >> {rootdir}/test3-task1.out'.format(rootdir=rootdir),
    owner='max',
    dag=dag)
t2 = BashOperator(
    task_id='test3-task2',
    bash_command='whoami',
    retries=3,
    owner='max',
    dag=dag)
{code}

Then I run the command "airflow test test3 test3-task2  2016-07-25" with 'airflow' user of
linux.   The result of output "whoami" is "airflow".
But I hope that the output result is "owner" of task.

[2016-07-25 11:22:37,716] {bash_operator.py:64} INFO - Temporary script location :/tmp/airflowtmpoYNJE8//tmp/airflowtmpoYNJE8/test3-task2U1lpom

[2016-07-25 11:22:37,716] {bash_operator.py:65} INFO - Running command: whoami

[2016-07-25 11:22:37,722] {bash_operator.py:73} INFO - Output:

[2016-07-25 11:22:37,725] {bash_operator.py:77} INFO - **airflow**

[2016-07-25 11:22:37,725] {bash_operator.py:80} INFO - Command exited with return code 0

What is my wrong ?  

Thanks



was (Author: hwbj):
I don't  fix this problem.

My dag code is the following.

from airflow.models import DAG
from airflow.operators import BashOperator
from datetime import datetime, timedelta
rootdir = "/tmp/airflow"
default_args = {
    'owner': 'max',
    'depends_on_past': False,
    'start_date': datetime.now(),
    'email': ['max@test.com'],
    'email_on_failure': False,
    'email_on_retry': False,
    'retries': 1,
    'retry_delay': timedelta(minutes=5),
}
dag = DAG('test3', default_args=default_args,
                    schedule_interval='*/2 * * * *')
t1 = BashOperator(
    task_id='test3-task1',
    bash_command='date  >> {rootdir}/test3-task1.out'.format(rootdir=rootdir),
    owner='max',
    dag=dag)
t2 = BashOperator(
    task_id='test3-task2',
    bash_command='whoami',
    retries=3,
    owner='max',
    dag=dag)

Then I run the command "airflow test test3 test3-task2  2016-07-25" with 'airflow' user of
linux.   The result of output "whoami" is "airflow".
But I hope that the output result is "owner" of task.

[2016-07-25 11:22:37,716] {bash_operator.py:64} INFO - Temporary script location :/tmp/airflowtmpoYNJE8//tmp/airflowtmpoYNJE8/test3-task2U1lpom

[2016-07-25 11:22:37,716] {bash_operator.py:65} INFO - Running command: whoami

[2016-07-25 11:22:37,722] {bash_operator.py:73} INFO - Output:

[2016-07-25 11:22:37,725] {bash_operator.py:77} INFO - **airflow**

[2016-07-25 11:22:37,725] {bash_operator.py:80} INFO - Command exited with return code 0

What is my wrong ?  

Thanks


> how should I use the right owner task in airflow?
> -------------------------------------------------
>
>                 Key: AIRFLOW-357
>                 URL: https://issues.apache.org/jira/browse/AIRFLOW-357
>             Project: Apache Airflow
>          Issue Type: Bug
>    Affects Versions: Airflow 1.7.1
>            Reporter: wei.he
>
> I dont understand the "owner" in airflow. the comment of ower is "the owner of the task,
using the unix username is recommended". I wrote some the following code.
> Default_args = {
> 'owner': 'max',
> 'depends_on_past': False,
> 'start_date': datetime(2016, 7, 14),
> 'email': ['max@test.com'],
> 'email_on_failure': False,
> 'email_on_retry': False,
> 'retries': 1,
> 'retry_delay': timedelta(minutes=5), 
> dag = DAG('dmp-annalect', default_args=default_args,
>                     schedule_interval='30 0 * * *')
> task1_pigjob_basedata = """
> {local_dir}/src/basedata/basedata.sh > {local_dir}/log/basedata/run_log &
> """.format(local_dir=WORKSPACE)
> task1_pigjob_basedata = BashOperator(
> task_id='task1_pigjob_basedata_impclk',owner='max',
> bash_command=pigjob_basedata_impclk,
> dag=dag)
>  I used the command "airflow test dagid taskid 2016-07-20" , 
> But I got a error, 
> ... {bash_operator.py:77} INFO - put: Permission denied: user=airflow, ....
> I thought that my job ran with "max" user, but apperently , ran test using 'airflow'
user .
> I hope if I run my task using 'max' user, how should I do.
> Thanks



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Mime
View raw message