hadoop-common-dev mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Dmitriy V. Ryaboy (JIRA)" <j...@apache.org>
Subject [jira] Commented: (HADOOP-5303) Hadoop Workflow System (HWS)
Date Tue, 09 Jun 2009 00:56:07 GMT

    [ https://issues.apache.org/jira/browse/HADOOP-5303?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12717503#action_12717503
] 

Dmitriy V. Ryaboy commented on HADOOP-5303:
-------------------------------------------

Alejandro,
I am trying to run the tests for this on Hadoop 18.3 and several things are failing:

1) I get repeated warnings along the lines of 
17:32:50,054  WARN PigActionExecutor:547 - USER[-] GROUP[-] TOKEN[-] APP[-] JOB[-] ACTION[-]
Exception [org.apache.hadoop.hdfs.server.namenode.SafeModeException] no in classpath, ActionExecutor
[pig] will handled it as ERROR
17:32:50,051  WARN FsActionExecutor:547 - USER[-] GROUP[-] TOKEN[-] APP[-] JOB[-] ACTION[-]
Exception [org.apache.hadoop.hdfs.protocol.QuotaExceededException] no in classpath, ActionExecutor
[fs] will handled it as ERROR

The package name was org.apache.hadoop.dfs in 18. I am using the -Dh18 flag.

2) I also get the following Failed and Errored out tests:

Failed tests: 
  testSourceNotFoundException(org.apache.oozie.dag.action.hadoop.TestFsActionExecutor)
  testAccessDeniedException(org.apache.oozie.dag.action.hadoop.TestFsActionExecutor)
  testAuthorizationService(org.apache.oozie.service.TestAuthorizationService)
  testFsDir(org.apache.oozie.test.TestXFsTestCase)

Tests in error: 
  testJobStart(org.apache.oozie.dag.action.ssh.TestSshActionExecutor)
  testJobRecover(org.apache.oozie.dag.action.ssh.TestSshActionExecutor)
  testConnectionErrors(org.apache.oozie.dag.action.ssh.TestSshActionExecutor)

These are along the lines of 
testAccessDeniedException(org.apache.oozie.dag.action.hadoop.TestFsActionExecutor)  Time elapsed:
0.136sec  <<< FAILURE!
junit.framework.AssertionFailedError
        at junit.framework.Assert.fail(Assert.java:47)
        at junit.framework.Assert.fail(Assert.java:53)
        at org.apache.oozie.dag.action.hadoop.TestFsActionExecutor.testAccessDeniedException(TestFsActionExecutor.java:276)

Thoughts?

> Hadoop Workflow System (HWS)
> ----------------------------
>
>                 Key: HADOOP-5303
>                 URL: https://issues.apache.org/jira/browse/HADOOP-5303
>             Project: Hadoop Core
>          Issue Type: New Feature
>            Reporter: Alejandro Abdelnur
>            Assignee: Alejandro Abdelnur
>         Attachments: hws-preso-v1_0_2009FEB22.pdf, hws-spec2009MAR09.pdf, hws-v1_0_2009FEB22.pdf,
oozie-0.18.3.o0.1-SNAPSHOT-distro.tar.gz, oozie-spec-20090521.pdf, oozie-src-20090605.tar.gz
>
>
> This is a proposal for a system specialized in running Hadoop/Pig jobs in a control dependency
DAG (Direct Acyclic Graph), a Hadoop workflow application.
> Attached there is a complete specification and a high level overview presentation.
> ----
> *Highlights* 
> A Workflow application is DAG that coordinates the following types of actions: Hadoop,
Pig, Ssh, Http, Email and sub-workflows. 
> Flow control operations within the workflow applications can be done using decision,
fork and join nodes. Cycles in workflows are not supported.
> Actions and decisions can be parameterized with job properties, actions output (i.e.
Hadoop counters, Ssh key/value pairs output) and file information (file exists, file size,
etc). Formal parameters are expressed in the workflow definition as {{${VAR}}} variables.
> A Workflow application is a ZIP file that contains the workflow definition (an XML file),
all the necessary files to run all the actions: JAR files for Map/Reduce jobs, shells for
streaming Map/Reduce jobs, native libraries, Pig scripts, and other resource files.
> Before running a workflow job, the corresponding workflow application must be deployed
in HWS.
> Deploying workflow application and running workflow jobs can be done via command line
tools, a WS API and a Java API.
> Monitoring the system and workflow jobs can be done via a web console, command line tools,
a WS API and a Java API.
> When submitting a workflow job, a set of properties resolving all the formal parameters
in the workflow definitions must be provided. This set of properties is a Hadoop configuration.
> Possible states for a workflow jobs are: {{CREATED}}, {{RUNNING}}, {{SUSPENDED}}, {{SUCCEEDED}},
{{KILLED}} and {{FAILED}}.
> In the case of a action failure in a workflow job, depending on the type of failure,
HWS will attempt automatic retries, it will request a manual retry or it will fail the workflow
job.
> HWS can make HTTP callback notifications on action start/end/failure events and workflow
end/failure events.
> In the case of workflow job failure, the workflow job can be resubmitted skipping previously
completed actions. Before doing a resubmission the workflow application could be updated with
a patch to fix a problem in the workflow application code.
> ----

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.


Mime
View raw message