Return-Path: X-Original-To: apmail-hadoop-common-commits-archive@www.apache.org Delivered-To: apmail-hadoop-common-commits-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 776854E06 for ; Mon, 6 Jun 2011 21:33:57 +0000 (UTC) Received: (qmail 42225 invoked by uid 500); 6 Jun 2011 21:33:57 -0000 Delivered-To: apmail-hadoop-common-commits-archive@hadoop.apache.org Received: (qmail 42200 invoked by uid 500); 6 Jun 2011 21:33:57 -0000 Mailing-List: contact common-commits-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: common-dev@hadoop.apache.org Delivered-To: mailing list common-commits@hadoop.apache.org Received: (qmail 42193 invoked by uid 99); 6 Jun 2011 21:33:57 -0000 Received: from nike.apache.org (HELO nike.apache.org) (192.87.106.230) by apache.org (qpsmtpd/0.29) with ESMTP; Mon, 06 Jun 2011 21:33:57 +0000 X-ASF-Spam-Status: No, hits=-2000.0 required=5.0 tests=ALL_TRUSTED X-Spam-Check-By: apache.org Received: from [140.211.11.4] (HELO eris.apache.org) (140.211.11.4) by apache.org (qpsmtpd/0.29) with ESMTP; Mon, 06 Jun 2011 21:33:53 +0000 Received: by eris.apache.org (Postfix, from userid 65534) id 077A323889D5; Mon, 6 Jun 2011 21:33:32 +0000 (UTC) Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit Subject: svn commit: r1132788 - in /hadoop/common/branches/branch-0.20-security-205/src: contrib/streaming/src/test/system/org/apache/hadoop/mapred/ test/system/java/org/apache/hadoop/mapred/ test/system/java/org/apache/hadoop/test/system/process/ Date: Mon, 06 Jun 2011 21:33:31 -0000 To: common-commits@hadoop.apache.org From: cdouglas@apache.org X-Mailer: svnmailer-1.0.8 Message-Id: <20110606213332.077A323889D5@eris.apache.org> X-Virus-Checked: Checked by ClamAV on apache.org Author: cdouglas Date: Mon Jun 6 21:33:31 2011 New Revision: 1132788 URL: http://svn.apache.org/viewvc?rev=1132788&view=rev Log: MAPREDUCE-2535. Fix system tests affected by previous commit. (Robert Joseph Evans via cdouglas) Modified: hadoop/common/branches/branch-0.20-security-205/src/contrib/streaming/src/test/system/org/apache/hadoop/mapred/TestTaskKillingOfStreamingJob.java hadoop/common/branches/branch-0.20-security-205/src/test/system/java/org/apache/hadoop/mapred/TestJobCacheDirectoriesCleanUp.java hadoop/common/branches/branch-0.20-security-205/src/test/system/java/org/apache/hadoop/mapred/TestTaskChildsKilling.java hadoop/common/branches/branch-0.20-security-205/src/test/system/java/org/apache/hadoop/mapred/TestTaskKilling.java hadoop/common/branches/branch-0.20-security-205/src/test/system/java/org/apache/hadoop/mapred/TestTaskTrackerInfoSuccessfulFailedJobs.java hadoop/common/branches/branch-0.20-security-205/src/test/system/java/org/apache/hadoop/test/system/process/HadoopDaemonRemoteCluster.java Modified: hadoop/common/branches/branch-0.20-security-205/src/contrib/streaming/src/test/system/org/apache/hadoop/mapred/TestTaskKillingOfStreamingJob.java URL: http://svn.apache.org/viewvc/hadoop/common/branches/branch-0.20-security-205/src/contrib/streaming/src/test/system/org/apache/hadoop/mapred/TestTaskKillingOfStreamingJob.java?rev=1132788&r1=1132787&r2=1132788&view=diff ============================================================================== --- hadoop/common/branches/branch-0.20-security-205/src/contrib/streaming/src/test/system/org/apache/hadoop/mapred/TestTaskKillingOfStreamingJob.java (original) +++ hadoop/common/branches/branch-0.20-security-205/src/contrib/streaming/src/test/system/org/apache/hadoop/mapred/TestTaskKillingOfStreamingJob.java Mon Jun 6 21:33:31 2011 @@ -27,7 +27,6 @@ import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; import org.apache.hadoop.fs.permission.FsAction; import org.apache.hadoop.fs.permission.FsPermission; -import org.apache.hadoop.mapred.JobClient.NetworkedJob; import org.apache.hadoop.mapreduce.test.system.JTProtocol; import org.apache.hadoop.mapreduce.test.system.JobInfo; import org.apache.hadoop.mapreduce.test.system.MRCluster; @@ -94,7 +93,7 @@ public class TestTaskKillingOfStreamingJ jtClient.isTaskStarted(taskInfo)); JobInfo jInfo = wovenClient.getJobInfo(jobId); - NetworkedJob networkJob = client.new NetworkedJob(jInfo.getStatus()); + RunningJob networkJob = client.getJob(jobId); TaskID tID = TaskID.downgrade(taskInfo.getTaskID()); TaskAttemptID taskAttID = new TaskAttemptID(tID, 0); networkJob.killTask(taskAttID, false); @@ -153,7 +152,7 @@ public class TestTaskKillingOfStreamingJ jtClient.isTaskStarted(taskInfo)); JobInfo jInfo = wovenClient.getJobInfo(jobId); - NetworkedJob networkJob = client.new NetworkedJob(jInfo.getStatus()); + RunningJob networkJob = client.getJob(jobId); TaskID tID = TaskID.downgrade(taskInfo.getTaskID()); TaskAttemptID taskAttID = new TaskAttemptID(tID, 0); networkJob.killTask(taskAttID, true); Modified: hadoop/common/branches/branch-0.20-security-205/src/test/system/java/org/apache/hadoop/mapred/TestJobCacheDirectoriesCleanUp.java URL: http://svn.apache.org/viewvc/hadoop/common/branches/branch-0.20-security-205/src/test/system/java/org/apache/hadoop/mapred/TestJobCacheDirectoriesCleanUp.java?rev=1132788&r1=1132787&r2=1132788&view=diff ============================================================================== --- hadoop/common/branches/branch-0.20-security-205/src/test/system/java/org/apache/hadoop/mapred/TestJobCacheDirectoriesCleanUp.java (original) +++ hadoop/common/branches/branch-0.20-security-205/src/test/system/java/org/apache/hadoop/mapred/TestJobCacheDirectoriesCleanUp.java Mon Jun 6 21:33:31 2011 @@ -28,7 +28,6 @@ import org.apache.hadoop.mapreduce.test. import org.apache.hadoop.mapreduce.test.system.JTProtocol; import org.apache.hadoop.mapreduce.test.system.TaskInfo; import org.apache.hadoop.mapreduce.test.system.JobInfo; -import org.apache.hadoop.mapred.JobClient.NetworkedJob; import org.apache.hadoop.fs.Path; import org.apache.hadoop.fs.FileStatus; import org.apache.hadoop.fs.permission.FsAction; @@ -189,8 +188,8 @@ public class TestJobCacheDirectoriesClea int MAX_MAP_TASK_ATTEMPTS = Integer. parseInt(jobConf.get("mapred.map.max.attempts")); while(taskinfo.numFailedAttempts() < MAX_MAP_TASK_ATTEMPTS) { - NetworkedJob networkJob = jtClient.getClient(). - new NetworkedJob(jobInfo.getStatus()); + org.apache.hadoop.mapreduce.JobID temp = jobInfo.getID(); + RunningJob networkJob = client.getJob(new JobID(temp.getJtIdentifier(), temp.getId())); networkJob.killTask(taskAttID, true); taskinfo = rtClient.getTaskInfo(taskinfo.getTaskID()); taskAttID = new TaskAttemptID(taskId, taskinfo.numFailedAttempts()); Modified: hadoop/common/branches/branch-0.20-security-205/src/test/system/java/org/apache/hadoop/mapred/TestTaskChildsKilling.java URL: http://svn.apache.org/viewvc/hadoop/common/branches/branch-0.20-security-205/src/test/system/java/org/apache/hadoop/mapred/TestTaskChildsKilling.java?rev=1132788&r1=1132787&r2=1132788&view=diff ============================================================================== --- hadoop/common/branches/branch-0.20-security-205/src/test/system/java/org/apache/hadoop/mapred/TestTaskChildsKilling.java (original) +++ hadoop/common/branches/branch-0.20-security-205/src/test/system/java/org/apache/hadoop/mapred/TestTaskChildsKilling.java Mon Jun 6 21:33:31 2011 @@ -35,7 +35,6 @@ import org.apache.hadoop.mapreduce.test. import org.apache.hadoop.mapreduce.test.system.TTTaskInfo; import org.apache.hadoop.mapreduce.test.system.TaskInfo; import org.apache.hadoop.mapreduce.test.system.FinishTaskControlAction; -import org.apache.hadoop.mapred.JobClient.NetworkedJob; import org.apache.hadoop.util.ToolRunner; import org.apache.hadoop.util.Tool; import org.junit.AfterClass; @@ -271,8 +270,7 @@ public class TestTaskChildsKilling { Assert.assertTrue("Map process is not alive before task kills.", ttIns.isProcessTreeAlive(pid)); - NetworkedJob networkJob = client.new NetworkedJob(jInfo.getStatus()); - networkJob.killTask(taskAttID, false); + runJob.killTask(taskAttID, false); LOG.info("Waiting till the task is killed..."); taskInfo = wovenClient.getTaskInfo(taskInfo.getTaskID()); Modified: hadoop/common/branches/branch-0.20-security-205/src/test/system/java/org/apache/hadoop/mapred/TestTaskKilling.java URL: http://svn.apache.org/viewvc/hadoop/common/branches/branch-0.20-security-205/src/test/system/java/org/apache/hadoop/mapred/TestTaskKilling.java?rev=1132788&r1=1132787&r2=1132788&view=diff ============================================================================== --- hadoop/common/branches/branch-0.20-security-205/src/test/system/java/org/apache/hadoop/mapred/TestTaskKilling.java (original) +++ hadoop/common/branches/branch-0.20-security-205/src/test/system/java/org/apache/hadoop/mapred/TestTaskKilling.java Mon Jun 6 21:33:31 2011 @@ -34,7 +34,6 @@ import org.apache.hadoop.mapreduce.test. import org.apache.hadoop.mapreduce.test.system.TTClient; import org.apache.hadoop.mapreduce.test.system.JTClient; import org.apache.hadoop.mapreduce.test.system.FinishTaskControlAction; -import org.apache.hadoop.mapred.JobClient.NetworkedJob; import org.apache.hadoop.io.NullWritable; import org.apache.hadoop.io.IntWritable; import org.apache.hadoop.io.Text; @@ -100,7 +99,7 @@ public class TestTaskKilling { jtClient.isTaskStarted(taskInfo)); // Fail the running task. - NetworkedJob networkJob = jobClient.new NetworkedJob(jInfo.getStatus()); + RunningJob networkJob = jobClient.getJob(jobId); TaskID tID = TaskID.downgrade(taskInfo.getTaskID()); TaskAttemptID taskAttID = new TaskAttemptID(tID , 0); networkJob.killTask(taskAttID, true); @@ -206,7 +205,7 @@ public class TestTaskKilling { taskAttID + " has not been found while task was running.", isTempFolderExists); - NetworkedJob networkJob = jobClient.new NetworkedJob(jInfo.getStatus()); + RunningJob networkJob = jobClient.getJob(id); networkJob.killTask(taskAttID, false); ttClient.getProxy().sendAction(action); taskInfo = remoteJTClient.getTaskInfo(tID); @@ -353,8 +352,7 @@ public class TestTaskKilling { TaskAttemptID tAttID = new TaskAttemptID(taskId, taskInfo.numFailedAttempts()); while(taskInfo.numFailedAttempts() < MAX_MAP_TASK_ATTEMPTS) { - NetworkedJob networkJob = jtClient.getClient(). - new NetworkedJob(jobInfo.getStatus()); + RunningJob networkJob = jobClient.getJob(id); networkJob.killTask(taskAttID, true); taskInfo = remoteJTClient.getTaskInfo(taskInfo.getTaskID()); taskAttID = new TaskAttemptID(taskId, taskInfo.numFailedAttempts()); @@ -484,8 +482,7 @@ public class TestTaskKilling { taskIdKilled = taskid.toString(); taskAttemptID = new TaskAttemptID(taskid, i); LOG.info("taskAttemptid going to be killed is : " + taskAttemptID); - (jobClient.new NetworkedJob(jInfo.getStatus())). - killTask(taskAttemptID,true); + rJob.killTask(taskAttemptID,true); checkTaskCompletionEvent(taskAttemptID, jInfo); break; } else { @@ -495,8 +492,7 @@ public class TestTaskKilling { UtilsForTests.waitFor(20000); LOG.info("taskAttemptid going to be killed is : " + taskAttemptID); - (jobClient.new NetworkedJob(jInfo.getStatus())). - killTask(taskAttemptID,true); + rJob.killTask(taskAttemptID,true); checkTaskCompletionEvent(taskAttemptID,jInfo); break; } @@ -536,8 +532,10 @@ public class TestTaskKilling { boolean match = false; int count = 0; while (!match) { - TaskCompletionEvent[] taskCompletionEvents = jobClient.new - NetworkedJob(jInfo.getStatus()).getTaskCompletionEvents(0); + org.apache.hadoop.mapreduce.JobID temp = jInfo.getID(); + RunningJob rJob = jobClient.getJob(new JobID(temp.getJtIdentifier(), temp.getId())); + + TaskCompletionEvent[] taskCompletionEvents = rJob.getTaskCompletionEvents(0); for (TaskCompletionEvent taskCompletionEvent : taskCompletionEvents) { LOG.info("taskCompletionEvent.getTaskAttemptId().toString() is : " + taskCompletionEvent.getTaskAttemptId().toString()); Modified: hadoop/common/branches/branch-0.20-security-205/src/test/system/java/org/apache/hadoop/mapred/TestTaskTrackerInfoSuccessfulFailedJobs.java URL: http://svn.apache.org/viewvc/hadoop/common/branches/branch-0.20-security-205/src/test/system/java/org/apache/hadoop/mapred/TestTaskTrackerInfoSuccessfulFailedJobs.java?rev=1132788&r1=1132787&r2=1132788&view=diff ============================================================================== --- hadoop/common/branches/branch-0.20-security-205/src/test/system/java/org/apache/hadoop/mapred/TestTaskTrackerInfoSuccessfulFailedJobs.java (original) +++ hadoop/common/branches/branch-0.20-security-205/src/test/system/java/org/apache/hadoop/mapred/TestTaskTrackerInfoSuccessfulFailedJobs.java Mon Jun 6 21:33:31 2011 @@ -30,7 +30,6 @@ import org.apache.hadoop.mapreduce.test. import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.mapred.UtilsForTests; -import org.apache.hadoop.mapred.JobClient.NetworkedJob; import org.apache.hadoop.examples.SleepJob; import org.apache.hadoop.fs.permission.FsPermission; import org.apache.hadoop.fs.permission.FsAction; @@ -340,11 +339,9 @@ public class TestTaskTrackerInfoSuccessf Assert.assertTrue("Task has not been started for 1 min.", count != 60); - NetworkedJob networkJob = (cluster.getJTClient().getClient()).new - NetworkedJob(jInfo.getStatus()); TaskID tID = TaskID.downgrade(taskInfo.getTaskID()); TaskAttemptID taskAttID = new TaskAttemptID(tID , 0); - networkJob.killTask(taskAttID, false); + rJob.killTask(taskAttID, false); count = 0; LOG.info("Waiting till the job is completed..."); Modified: hadoop/common/branches/branch-0.20-security-205/src/test/system/java/org/apache/hadoop/test/system/process/HadoopDaemonRemoteCluster.java URL: http://svn.apache.org/viewvc/hadoop/common/branches/branch-0.20-security-205/src/test/system/java/org/apache/hadoop/test/system/process/HadoopDaemonRemoteCluster.java?rev=1132788&r1=1132787&r2=1132788&view=diff ============================================================================== --- hadoop/common/branches/branch-0.20-security-205/src/test/system/java/org/apache/hadoop/test/system/process/HadoopDaemonRemoteCluster.java (original) +++ hadoop/common/branches/branch-0.20-security-205/src/test/system/java/org/apache/hadoop/test/system/process/HadoopDaemonRemoteCluster.java Mon Jun 6 21:33:31 2011 @@ -124,9 +124,11 @@ public abstract class HadoopDaemonRemote } } finally { try { - reader.close(); + if(reader != null) { + reader.close(); + } } catch (IOException e) { - LOG.warn("Could not close reader"); + LOG.warn("Could not close reader", e); } } LOG.info("Created HadoopDaemonInfo for " + cmd + " " + role + " from "