hadoop-common-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From d...@apache.org
Subject svn commit: r712099 - /hadoop/core/trunk/src/test/org/apache/hadoop/mapred/TestJobHistoryVersion.java
Date Fri, 07 Nov 2008 09:54:21 GMT
Author: ddas
Date: Fri Nov  7 01:54:19 2008
New Revision: 712099

URL: http://svn.apache.org/viewvc?rev=712099&view=rev
Log:
HADOOP-4190. Adding testcase that I forgot to add earlier.

Added:
    hadoop/core/trunk/src/test/org/apache/hadoop/mapred/TestJobHistoryVersion.java

Added: hadoop/core/trunk/src/test/org/apache/hadoop/mapred/TestJobHistoryVersion.java
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/mapred/TestJobHistoryVersion.java?rev=712099&view=auto
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/mapred/TestJobHistoryVersion.java (added)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/mapred/TestJobHistoryVersion.java Fri Nov
 7 01:54:19 2008
@@ -0,0 +1,154 @@
+/**
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ *     http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.hadoop.mapred;
+
+import java.io.IOException;
+import junit.framework.TestCase;
+
+import org.apache.hadoop.fs.FSDataOutputStream;
+import org.apache.hadoop.fs.FileSystem;
+import org.apache.hadoop.fs.Path;
+import org.apache.hadoop.mapred.JobHistory.JobInfo;
+import org.apache.hadoop.mapred.JobHistory.RecordTypes;
+
+public class TestJobHistoryVersion extends TestCase {
+  private static final String HOSTNAME = "localhost";
+  private static final String TIME= "1234567890123";
+  private static final String USER = "user";
+  private static final String JOBNAME = "job";
+  private static final String JOB = "job_200809180000_0001";
+  private static final String FILENAME = 
+    HOSTNAME + "_" + TIME + "_" +  JOB + "_" + USER + "_" + JOBNAME;
+  private static final String TASK_ID = "tip_200809180000_0001_0";
+  private static final String TASK_ATTEMPT_ID = 
+    "attempt_200809180000_0001_0_1234567890123";
+  private static final String COUNTERS = 
+    "Job Counters.Launched map tasks:1," 
+    + "Map-Reduce Framework.Map input records:0,"
+    + "Map-Reduce Framework.Map input bytes:0,"
+    + "File Systems.HDFS bytes written:0,";
+  private static final String TEST_DIR = "test-history-version";
+  private static final String DELIM = ".";
+  
+  
+  private void writeHistoryFile(FSDataOutputStream out, boolean old)
+  throws IOException {
+    String delim = "\n";
+    String counters = COUNTERS;
+    String jobConf = "job.xml";
+    if (!old) {
+      // Change the delimiter
+      delim = DELIM + delim;
+      
+      // Write the version line
+      out.writeBytes(RecordTypes.Meta.name() + " VERSION=\"" 
+                     + JobHistory.VERSION + "\" " + delim);
+      jobConf = JobHistory.escapeString(jobConf);
+      counters = JobHistory.escapeString(counters);
+    }
+    
+    // Write the job-start line
+    
+    
+    out.writeBytes("Job JOBID=\"" + JOB + "\" JOBNAME=\"" + JOBNAME + "\"" 
+                   + " USER=\"" + USER + "\" SUBMIT_TIME=\"" + TIME + "\"" 
+                   + " JOBCONF=\"" + jobConf + "\" " + delim);
+    
+    // Write the job-launch line
+    out.writeBytes("Job JOBID=\"" + JOB + "\" LAUNCH_TIME=\"" + TIME + "\"" 
+                   + " TOTAL_MAPS=\"1\" TOTAL_REDUCES=\"0\" " + delim);
+    
+    // Write the task start line
+    out.writeBytes("Task TASKID=\"" + TASK_ID + "\" TASK_TYPE=\"MAP\"" 
+                   + " START_TIME=\"" + TIME + "\" SPLITS=\"\"" 
+                   + " TOTAL_MAPS=\"1\" TOTAL_REDUCES=\"0\" " + delim);
+    
+    // Write the attempt start line
+    out.writeBytes("MapAttempt TASK_TYPE=\"MAP\" TASKID=\"" + TASK_ID + "\"" 
+                   + " TASK_ATTEMPT_ID=\"" + TASK_ATTEMPT_ID + "\"" 
+                   + " START_TIME=\"" + TIME + "\"" 
+                   + " HOSTNAME=\"" + HOSTNAME + "\" " + delim);
+    
+    // Write the attempt finish line
+    out.writeBytes("MapAttempt TASK_TYPE=\"MAP\" TASKID=\"" + TASK_ID + "\"" 
+                   + " TASK_ATTEMPT_ID=\"" + TASK_ATTEMPT_ID + "\"" 
+                   + " FINISH_TIME=\"" + TIME + "\""
+                   + " TASK_STATUS=\"SUCCESS\""
+                   + " HOSTNAME=\"" + HOSTNAME + "\" " + delim);
+    
+    // Write the task finish line
+    out.writeBytes("Task TASKID=\"" + TASK_ID + "\" TASK_TYPE=\"MAP\""
+                   + " TASK_STATUS=\"SUCCESS\""
+                   + " FINISH_TIME=\"" + TIME + "\""
+                   + " COUNTERS=\"" + counters + "\" " + delim);
+    
+    // Write the job-finish line
+    out.writeBytes("Job JOBID=\"" + JOB + "\" FINISH_TIME=\"" + TIME + "\"" 
+                   + " TOTAL_MAPS=\"1\" TOTAL_REDUCES=\"0\""
+                   + " JOB_STATUS=\"SUCCESS\" FINISHED_MAPS=\"1\""
+                   + " FINISHED_REDUCES=\"0\" FAILED_MAPS=\"0\""
+                   + " FAILED_REDUCES=\"0\""
+                   + " COUNTERS=\"" + counters + "\" " + delim);
+    
+  }
+  
+  /**
+   * Tests the JobHistory parser with old files
+   */
+  public void testJobHistoryWithoutVersion() throws IOException {
+    JobConf conf = new JobConf();
+    FileSystem fs = FileSystem.getLocal(conf);
+    
+    Path historyPath = new Path(TEST_DIR + "/_logs/history/" + FILENAME);
+    
+    fs.delete(historyPath, false);
+    
+    FSDataOutputStream out = fs.create(historyPath);
+    writeHistoryFile(out, true);
+    out.close();
+    
+    JobInfo job = new JobHistory.JobInfo(JOB); 
+    DefaultJobHistoryParser.parseJobTasks(historyPath.toString(), job, fs);
+    
+    assertTrue("Failed to parse old jobhistory files", 
+               job.getAllTasks().size() > 0);
+  }
+  
+  /**
+   * Tests the JobHistory parser with new file
+   */
+  public void testJobHistoryWithVersion() throws IOException {
+    JobConf conf = new JobConf();
+    FileSystem fs = FileSystem.getLocal(conf);
+    
+    Path historyPath = new Path(TEST_DIR + "/_logs/history/" + FILENAME);
+    
+    fs.delete(historyPath, false);
+    
+    FSDataOutputStream out = fs.create(historyPath);
+    writeHistoryFile(out, false);
+    out.close();
+    
+    JobInfo job = new JobHistory.JobInfo(JOB); 
+    DefaultJobHistoryParser.parseJobTasks(historyPath.toString(), job, fs);
+    
+    assertTrue("Failed to parse old jobhistory files", 
+               job.getAllTasks().size() > 0);
+  }
+}



Mime
View raw message