incubator-hcatalog-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From hashut...@apache.org
Subject svn commit: r1233546 - in /incubator/hcatalog/branches/branch-0.3: ./ shims/ shims/src/ shims/src/20S/ shims/src/20S/java/ shims/src/20S/java/org/ shims/src/20S/java/org/apache/ shims/src/20S/java/org/apache/hcatalog/ shims/src/20S/java/org/apache/hcat...
Date Thu, 19 Jan 2012 19:28:06 GMT
Author: hashutosh
Date: Thu Jan 19 19:28:05 2012
New Revision: 1233546

URL: http://svn.apache.org/viewvc?rev=1233546&view=rev
Log:
Merged HCATALOG-179 from trunk into 0.3 branch (thw via hashutosh)

Added:
    incubator/hcatalog/branches/branch-0.3/shims/
      - copied from r1231282, incubator/hcatalog/trunk/shims/
    incubator/hcatalog/branches/branch-0.3/shims/src/
      - copied from r1231282, incubator/hcatalog/trunk/shims/src/
    incubator/hcatalog/branches/branch-0.3/shims/src/20S/
      - copied from r1231282, incubator/hcatalog/trunk/shims/src/20S/
    incubator/hcatalog/branches/branch-0.3/shims/src/20S/java/
      - copied from r1231282, incubator/hcatalog/trunk/shims/src/20S/java/
    incubator/hcatalog/branches/branch-0.3/shims/src/20S/java/org/
      - copied from r1231282, incubator/hcatalog/trunk/shims/src/20S/java/org/
    incubator/hcatalog/branches/branch-0.3/shims/src/20S/java/org/apache/
      - copied from r1231282, incubator/hcatalog/trunk/shims/src/20S/java/org/apache/
    incubator/hcatalog/branches/branch-0.3/shims/src/20S/java/org/apache/hcatalog/
      - copied from r1231282, incubator/hcatalog/trunk/shims/src/20S/java/org/apache/hcatalog/
    incubator/hcatalog/branches/branch-0.3/shims/src/20S/java/org/apache/hcatalog/shims/
      - copied from r1231282, incubator/hcatalog/trunk/shims/src/20S/java/org/apache/hcatalog/shims/
    incubator/hcatalog/branches/branch-0.3/shims/src/20S/java/org/apache/hcatalog/shims/HCatHadoopShims20S.java
      - copied unchanged from r1231282, incubator/hcatalog/trunk/shims/src/20S/java/org/apache/hcatalog/shims/HCatHadoopShims20S.java
    incubator/hcatalog/branches/branch-0.3/shims/src/23/
      - copied from r1231282, incubator/hcatalog/trunk/shims/src/23/
    incubator/hcatalog/branches/branch-0.3/shims/src/23/java/
      - copied from r1231282, incubator/hcatalog/trunk/shims/src/23/java/
    incubator/hcatalog/branches/branch-0.3/shims/src/23/java/org/
      - copied from r1231282, incubator/hcatalog/trunk/shims/src/23/java/org/
    incubator/hcatalog/branches/branch-0.3/shims/src/23/java/org/apache/
      - copied from r1231282, incubator/hcatalog/trunk/shims/src/23/java/org/apache/
    incubator/hcatalog/branches/branch-0.3/shims/src/23/java/org/apache/hcatalog/
      - copied from r1231282, incubator/hcatalog/trunk/shims/src/23/java/org/apache/hcatalog/
    incubator/hcatalog/branches/branch-0.3/shims/src/23/java/org/apache/hcatalog/shims/
      - copied from r1231282, incubator/hcatalog/trunk/shims/src/23/java/org/apache/hcatalog/shims/
    incubator/hcatalog/branches/branch-0.3/shims/src/23/java/org/apache/hcatalog/shims/HCatHadoopShims23.java
      - copied unchanged from r1231282, incubator/hcatalog/trunk/shims/src/23/java/org/apache/hcatalog/shims/HCatHadoopShims23.java
    incubator/hcatalog/branches/branch-0.3/src/java/org/apache/hcatalog/shims/
      - copied from r1231378, incubator/hcatalog/trunk/src/java/org/apache/hcatalog/shims/
    incubator/hcatalog/branches/branch-0.3/src/java/org/apache/hcatalog/shims/HCatHadoopShims.java
      - copied unchanged from r1231378, incubator/hcatalog/trunk/src/java/org/apache/hcatalog/shims/HCatHadoopShims.java
Modified:
    incubator/hcatalog/branches/branch-0.3/   (props changed)
    incubator/hcatalog/branches/branch-0.3/CHANGES.txt
    incubator/hcatalog/branches/branch-0.3/build-common.xml
    incubator/hcatalog/branches/branch-0.3/build.xml
    incubator/hcatalog/branches/branch-0.3/src/java/org/apache/hcatalog/mapreduce/FileOutputCommitterContainer.java
    incubator/hcatalog/branches/branch-0.3/src/java/org/apache/hcatalog/pig/HCatStorer.java
    incubator/hcatalog/branches/branch-0.3/src/test/org/apache/hcatalog/rcfile/TestRCFileInputStorageDriver.java
    incubator/hcatalog/branches/branch-0.3/src/test/org/apache/hcatalog/rcfile/TestRCFileMapReduceInputFormat.java
    incubator/hcatalog/branches/branch-0.3/src/test/org/apache/hcatalog/rcfile/TestRCFileOutputStorageDriver.java

Propchange: incubator/hcatalog/branches/branch-0.3/
------------------------------------------------------------------------------
    svn:mergeinfo = /incubator/hcatalog/trunk:1231282,1231378

Modified: incubator/hcatalog/branches/branch-0.3/CHANGES.txt
URL: http://svn.apache.org/viewvc/incubator/hcatalog/branches/branch-0.3/CHANGES.txt?rev=1233546&r1=1233545&r2=1233546&view=diff
==============================================================================
--- incubator/hcatalog/branches/branch-0.3/CHANGES.txt (original)
+++ incubator/hcatalog/branches/branch-0.3/CHANGES.txt Thu Jan 19 19:28:05 2012
@@ -58,6 +58,8 @@ Release 0.3.0 (unreleased changes)
   HCAT-63. RPM package integration with Hadoop (khorgath via hashutosh)
 
   IMPROVEMENTS
+  HCAT-179. Make HCatalog compile against Hadoop 0.23 (thw via khorgath)
+
   HCAT-166. Making configuration/installation automatable (vikram.dixit via khorgath)
 
   HCAT-184. Optionally do not generate forrest docs (traviscrawford via hashutosh)

Modified: incubator/hcatalog/branches/branch-0.3/build-common.xml
URL: http://svn.apache.org/viewvc/incubator/hcatalog/branches/branch-0.3/build-common.xml?rev=1233546&r1=1233545&r2=1233546&view=diff
==============================================================================
--- incubator/hcatalog/branches/branch-0.3/build-common.xml (original)
+++ incubator/hcatalog/branches/branch-0.3/build-common.xml Thu Jan 19 19:28:05 2012
@@ -11,6 +11,22 @@
 
     <!-- hive -->
     <property name="hive.root" value="${common.basedir}/hive/external"/>
+    <property file="${hive.root}/build.properties" prefix="hive."/>
+    <property name="shims.name" value="20S" />
+    <property name="shims.20S.hive.shims.include" value="0.20,0.20S" />
+    <property name="shims.20S.hadoop.version" value="${hive.hadoop-0.20S.version}" />
+    <property name="shims.23.hive.shims.include" value="0.23" />
+    <property name="shims.23.hadoop.version" value="${hive.hadoop-0.23.version}" />
+
+    <!-- macro to accomplish nested expansion like ${p1.${p2}.name} -->
+    <macrodef name="expandToProperty">
+     <attribute name="name"/>
+     <attribute name="value"/>
+     <sequential>
+        <property name="@{name}" value="${@{value}}" />
+     </sequential>
+    </macrodef>
+    <expandToProperty name="hadoop.version" value="shims.${shims.name}.hadoop.version"/>
 
     <!-- common classpaths for various builds -->
     <path id="common.classpath">
@@ -22,8 +38,14 @@
         <fileset dir="${hive.root}/build/serde" includes="*.jar"/>
         <fileset dir="${hive.root}/build/metastore" includes="*.jar"/>
         <fileset dir="${hive.root}/build/ql" includes="*.jar"/>
-        <fileset dir="${hive.root}/build/hadoopcore/hadoop-0.20.3-CDH3-SNAPSHOT/"
-                 includes="hadoop-core-0.20.3-CDH3-SNAPSHOT.jar"/>
+        <fileset dir="${hive.root}/build/hadoopcore/hadoop-${hadoop.version}/">
+          <include name="**/hadoop-*.jar" />
+          <exclude name="**/*test*.jar" />
+          <!-- below is for 0.23 onwards -->
+          <!--include name="share/hadoop/common/lib/*.jar" /-->
+          <exclude name="share/hadoop/common/lib/hadoop-mapreduce-*.jar" />
+          <exclude name="share/hadoop/common/lib/hadoop-yarn-*.jar" />
+        </fileset>
         <fileset dir="${common.ivy.lib.dir}" includes="*.jar"/>
         <fileset dir="${hive.root}/build/ivy/lib/default" includes="antlr-3.0.1.jar"/>
         <fileset dir="${hive.root}/build/ivy/lib/default" includes="commons-lang-*.jar"/>
@@ -35,5 +57,4 @@
         <fileset dir="${hive.root}/build/ivy/lib/default" includes="libfb303-*.jar"/>
         <fileset dir="${hive.root}/lib" includes="asm-3.1.jar"/>
     </path>
-
 </project>

Modified: incubator/hcatalog/branches/branch-0.3/build.xml
URL: http://svn.apache.org/viewvc/incubator/hcatalog/branches/branch-0.3/build.xml?rev=1233546&r1=1233545&r2=1233546&view=diff
==============================================================================
--- incubator/hcatalog/branches/branch-0.3/build.xml (original)
+++ incubator/hcatalog/branches/branch-0.3/build.xml Thu Jan 19 19:28:05 2012
@@ -125,23 +125,17 @@
       <include name="**/*.jar" />
     </fileset>
     <!-- jars Hadoop depends on -->
-    <fileset dir="${hive.root}/build/hadoopcore/hadoop-0.20.3-CDH3-SNAPSHOT/lib/" >
-      <include name="**/*.jar" />
-    </fileset>
-    <!--
-    <pathelement location="${test.src.data.dir}/conf"/>
-    <pathelement location="${hadoop.oldstyle-name.test.jar}"/>
-    <pathelement location="${hadoop.newstyle-name.test.jar}"/>
-    <pathelement location="${common.jar}"/>
-    <pathelement location="${jsp.test.jar}"/>
-    -->
     <pathelement location="${hcatalog.jar}"/>
-    <!--
-    <pathelement location="${hadoop.root}/lib/jsp-2.1/jsp-api-2.1.jar"/>
-    -->
     <path refid="classpath"/>
-    <fileset dir="${hive.root}/build/hadoopcore/hadoop-0.20.3-CDH3-SNAPSHOT/"
-      includes="hadoop-test-0.20.3-CDH3-SNAPSHOT.jar"/>
+    <fileset dir="${hive.root}/build/hadoopcore/hadoop-${hadoop.version}/">
+      <include name="**/hadoop-*.jar" />
+      <include name="lib/**/*.jar" />
+      <exclude name="lib/**/excluded/" />
+      <!-- below is for 0.23 onwards -->
+      <include name="share/hadoop/common/lib/*.jar" />
+      <exclude name="share/hadoop/common/lib/hadoop-mapreduce-*.jar" />
+      <exclude name="share/hadoop/common/lib/hadoop-yarn-*.jar" />
+    </fileset>
   </path>
 
   <!--
@@ -226,7 +220,13 @@
 
   <!-- Build the external hive code -->
   <target name="hive.jar">
-    <ant antfile="build.xml" dir="${hive.root}" target="package" useNativeBasedir='true'/>
+    <echo message="Building hive with hadoop.version ${hadoop.version}" />
+    <local name="param.shims.include"/>
+    <expandToProperty name="param.shims.include" value="shims.${shims.name}.hive.shims.include"/>
+    <ant antfile="build.xml" dir="${hive.root}" target="package" useNativeBasedir='true'>
+       <property name="shims.include" value="${param.shims.include}"/>
+       <property name="hadoop.version" value="${hadoop.version}"/>
+    </ant>
   </target>
 
   <!--
@@ -244,6 +244,17 @@
       <compilerarg line="${javac.args}"/>
       <classpath refid="classpath" />
     </javac>
+    <!-- compile shim for selected hadoop version -->
+    <!--property name="debugclasspath" refid="classpath"/>
+    <echo message="classpath = ${debugclasspath}"/-->
+    <javac encoding="${build.encoding}" srcdir="${basedir}/shims/src/${shims.name}/java"
excludes="${excludes}"
+        includes="**/*.java" destdir="${build.classes}" debug="${javac.debug}"
+        optimize="${javac.optimize}" target="${javac.version}"
+        source="${javac.version}" deprecation="${javac.deprecation}"
+        includeantruntime="false">
+      <compilerarg line="${javac.args}"/>
+      <classpath refid="classpath" />
+    </javac>
   </target>
  
   <!-- Build the hcatalog client jar -->

Modified: incubator/hcatalog/branches/branch-0.3/src/java/org/apache/hcatalog/mapreduce/FileOutputCommitterContainer.java
URL: http://svn.apache.org/viewvc/incubator/hcatalog/branches/branch-0.3/src/java/org/apache/hcatalog/mapreduce/FileOutputCommitterContainer.java?rev=1233546&r1=1233545&r2=1233546&view=diff
==============================================================================
--- incubator/hcatalog/branches/branch-0.3/src/java/org/apache/hcatalog/mapreduce/FileOutputCommitterContainer.java
(original)
+++ incubator/hcatalog/branches/branch-0.3/src/java/org/apache/hcatalog/mapreduce/FileOutputCommitterContainer.java
Thu Jan 19 19:28:05 2012
@@ -47,6 +47,7 @@ import org.apache.hcatalog.data.schema.H
 import org.apache.hcatalog.data.schema.HCatSchema;
 import org.apache.hcatalog.data.schema.HCatSchemaUtils;
 import org.apache.hcatalog.har.HarOutputCommitterPostProcessor;
+import org.apache.hcatalog.shims.HCatHadoopShims;
 import org.apache.thrift.TException;
 
 import java.io.IOException;
@@ -146,7 +147,7 @@ class FileOutputCommitterContainer exten
             for(HCatOutputStorageDriver baseOsd : storageDriversDiscoveredByPath.values()){
                 try {
                     baseOsd.abortOutputCommitterJob(
-                            new TaskAttemptContext(
+                            HCatHadoopShims.Instance.get().createTaskAttemptContext(
                                     jobContext.getConfiguration(), TaskAttemptID.forName(ptnRootLocation)
                             ),state);
                 } catch (Exception e) {
@@ -256,7 +257,7 @@ class FileOutputCommitterContainer exten
                 for(HCatOutputStorageDriver baseOsd : storageDriversDiscoveredByPath.values()){
                     try {
                         baseOsd.cleanupOutputCommitterJob(
-                                new TaskAttemptContext(
+                                HCatHadoopShims.Instance.get().createTaskAttemptContext(
                                         context.getConfiguration(), TaskAttemptID.forName(ptnRootLocation)
                                 ));
                     } catch (Exception e) {

Modified: incubator/hcatalog/branches/branch-0.3/src/java/org/apache/hcatalog/pig/HCatStorer.java
URL: http://svn.apache.org/viewvc/incubator/hcatalog/branches/branch-0.3/src/java/org/apache/hcatalog/pig/HCatStorer.java?rev=1233546&r1=1233545&r2=1233546&view=diff
==============================================================================
--- incubator/hcatalog/branches/branch-0.3/src/java/org/apache/hcatalog/pig/HCatStorer.java
(original)
+++ incubator/hcatalog/branches/branch-0.3/src/java/org/apache/hcatalog/pig/HCatStorer.java
Thu Jan 19 19:28:05 2012
@@ -23,17 +23,15 @@ import java.util.Properties;
 
 import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.mapreduce.Job;
-import org.apache.hadoop.mapreduce.OutputCommitter;
 import org.apache.hadoop.mapreduce.OutputFormat;
-import org.apache.hadoop.mapreduce.TaskAttemptContext;
 import org.apache.hadoop.mapreduce.TaskAttemptID;
 import org.apache.hcatalog.common.HCatConstants;
 import org.apache.hcatalog.common.HCatException;
 import org.apache.hcatalog.common.HCatUtil;
 import org.apache.hcatalog.data.schema.HCatSchema;
 import org.apache.hcatalog.mapreduce.HCatOutputFormat;
-import org.apache.hcatalog.mapreduce.HCatOutputStorageDriver;
 import org.apache.hcatalog.mapreduce.OutputJobInfo;
+import org.apache.hcatalog.shims.HCatHadoopShims;
 import org.apache.pig.PigException;
 import org.apache.pig.ResourceSchema;
 import org.apache.pig.impl.logicalLayer.FrontendException;
@@ -150,7 +148,8 @@ public class HCatStorer extends HCatBase
       //In local mode, mapreduce will not call OutputCommitter.cleanupJob.
       //Calling it from here so that the partition publish happens.
       //This call needs to be removed after MAPREDUCE-1447 is fixed.
-        getOutputFormat().getOutputCommitter(new TaskAttemptContext(job.getConfiguration(),
new TaskAttemptID())).cleanupJob(job);
+        getOutputFormat().getOutputCommitter(HCatHadoopShims.Instance.get().createTaskAttemptContext(
+        		job.getConfiguration(), new TaskAttemptID())).cleanupJob(job);
       } catch (IOException e) {
         throw new IOException("Failed to cleanup job",e);
       } catch (InterruptedException e) {

Modified: incubator/hcatalog/branches/branch-0.3/src/test/org/apache/hcatalog/rcfile/TestRCFileInputStorageDriver.java
URL: http://svn.apache.org/viewvc/incubator/hcatalog/branches/branch-0.3/src/test/org/apache/hcatalog/rcfile/TestRCFileInputStorageDriver.java?rev=1233546&r1=1233545&r2=1233546&view=diff
==============================================================================
--- incubator/hcatalog/branches/branch-0.3/src/test/org/apache/hcatalog/rcfile/TestRCFileInputStorageDriver.java
(original)
+++ incubator/hcatalog/branches/branch-0.3/src/test/org/apache/hcatalog/rcfile/TestRCFileInputStorageDriver.java
Thu Jan 19 19:28:05 2012
@@ -48,12 +48,14 @@ import org.apache.hcatalog.data.DefaultH
 import org.apache.hcatalog.data.HCatRecord;
 import org.apache.hcatalog.data.schema.HCatSchema;
 import org.apache.hcatalog.rcfile.RCFileInputDriver;
+import org.apache.hcatalog.shims.HCatHadoopShims;
 
 
 public class TestRCFileInputStorageDriver extends TestCase{
   private static final Configuration conf = new Configuration();
   private static final Path dir =  new Path(System.getProperty("test.data.dir", ".") + "/mapred");
   private static final Path file = new Path(dir, "test_rcfile");
+  private final HCatHadoopShims shim = HCatHadoopShims.Instance.get();
 
   // Generate sample records to compare against
   private byte[][][] getRecords() throws UnsupportedEncodingException {
@@ -99,7 +101,7 @@ public class TestRCFileInputStorageDrive
 
     HCatSchema schema = buildHiveSchema();
     RCFileInputDriver sd = new RCFileInputDriver();
-    JobContext jc = new JobContext(conf, new JobID());
+    JobContext jc = shim.createJobContext(conf, new JobID());
     sd.setInputPath(jc, file.toString());
     InputFormat<?,?> iF = sd.getInputFormat(null);
     InputSplit split = iF.getSplits(jc).get(0);
@@ -107,7 +109,7 @@ public class TestRCFileInputStorageDrive
     sd.setOutputSchema(jc, schema);
     sd.initialize(jc, getProps());
 
-    TaskAttemptContext tac = new TaskAttemptContext(conf, new TaskAttemptID());
+    TaskAttemptContext tac = shim.createTaskAttemptContext(conf, new TaskAttemptID());
     RecordReader<?,?> rr = iF.createRecordReader(split,tac);
     rr.initialize(split, tac);
     HCatRecord[] tuples = getExpectedRecords();
@@ -125,7 +127,7 @@ public class TestRCFileInputStorageDrive
     BytesRefArrayWritable[] bytesArr = initTestEnvironment();
 
     RCFileInputDriver sd = new RCFileInputDriver();
-    JobContext jc = new JobContext(conf, new JobID());
+    JobContext jc = shim.createJobContext(conf, new JobID());
     sd.setInputPath(jc, file.toString());
     InputFormat<?,?> iF = sd.getInputFormat(null);
     InputSplit split = iF.getSplits(jc).get(0);
@@ -134,7 +136,7 @@ public class TestRCFileInputStorageDrive
 
     sd.initialize(jc, getProps());
     conf.set(ColumnProjectionUtils.READ_COLUMN_IDS_CONF_STR,jc.getConfiguration().get(ColumnProjectionUtils.READ_COLUMN_IDS_CONF_STR));
-    TaskAttemptContext tac = new TaskAttemptContext(conf, new TaskAttemptID());
+    TaskAttemptContext tac = shim.createTaskAttemptContext(conf, new TaskAttemptID());
     RecordReader<?,?> rr = iF.createRecordReader(split,tac);
     rr.initialize(split, tac);
     HCatRecord[] tuples = getPrunedRecords();
@@ -154,7 +156,7 @@ public class TestRCFileInputStorageDrive
     BytesRefArrayWritable[] bytesArr = initTestEnvironment();
 
     RCFileInputDriver sd = new RCFileInputDriver();
-    JobContext jc = new JobContext(conf, new JobID());
+    JobContext jc = shim.createJobContext(conf, new JobID());
     sd.setInputPath(jc, file.toString());
     InputFormat<?,?> iF = sd.getInputFormat(null);
     InputSplit split = iF.getSplits(jc).get(0);
@@ -166,7 +168,7 @@ public class TestRCFileInputStorageDrive
     map.put("part1", "first-part");
     sd.setPartitionValues(jc, map);
     conf.set(ColumnProjectionUtils.READ_COLUMN_IDS_CONF_STR,jc.getConfiguration().get(ColumnProjectionUtils.READ_COLUMN_IDS_CONF_STR));
-    TaskAttemptContext tac = new TaskAttemptContext(conf, new TaskAttemptID());
+    TaskAttemptContext tac = shim.createTaskAttemptContext(conf, new TaskAttemptID());
     RecordReader<?,?> rr = iF.createRecordReader(split,tac);
     rr.initialize(split, tac);
     HCatRecord[] tuples = getReorderedCols();

Modified: incubator/hcatalog/branches/branch-0.3/src/test/org/apache/hcatalog/rcfile/TestRCFileMapReduceInputFormat.java
URL: http://svn.apache.org/viewvc/incubator/hcatalog/branches/branch-0.3/src/test/org/apache/hcatalog/rcfile/TestRCFileMapReduceInputFormat.java?rev=1233546&r1=1233545&r2=1233546&view=diff
==============================================================================
--- incubator/hcatalog/branches/branch-0.3/src/test/org/apache/hcatalog/rcfile/TestRCFileMapReduceInputFormat.java
(original)
+++ incubator/hcatalog/branches/branch-0.3/src/test/org/apache/hcatalog/rcfile/TestRCFileMapReduceInputFormat.java
Thu Jan 19 19:28:05 2012
@@ -45,6 +45,7 @@ import org.apache.hadoop.mapreduce.Recor
 import org.apache.hadoop.mapreduce.TaskAttemptContext;
 import org.apache.hadoop.mapreduce.TaskAttemptID;
 import org.apache.hcatalog.rcfile.RCFileMapReduceInputFormat;
+import org.apache.hcatalog.shims.HCatHadoopShims;
 
   /**
    * TestRCFile.
@@ -232,7 +233,7 @@ import org.apache.hcatalog.rcfile.RCFile
       assertEquals("splits length should be " + splitNumber, splits.size(), splitNumber);
       int readCount = 0;
       for (int i = 0; i < splits.size(); i++) {
-        TaskAttemptContext tac = new TaskAttemptContext(jonconf, new TaskAttemptID());
+        TaskAttemptContext tac = HCatHadoopShims.Instance.get().createTaskAttemptContext(jonconf,
new TaskAttemptID());
         RecordReader<LongWritable, BytesRefArrayWritable> rr = inputFormat.createRecordReader(splits.get(i),
tac);
         rr.initialize(splits.get(i), tac);
         while (rr.nextKeyValue()) {

Modified: incubator/hcatalog/branches/branch-0.3/src/test/org/apache/hcatalog/rcfile/TestRCFileOutputStorageDriver.java
URL: http://svn.apache.org/viewvc/incubator/hcatalog/branches/branch-0.3/src/test/org/apache/hcatalog/rcfile/TestRCFileOutputStorageDriver.java?rev=1233546&r1=1233545&r2=1233546&view=diff
==============================================================================
--- incubator/hcatalog/branches/branch-0.3/src/test/org/apache/hcatalog/rcfile/TestRCFileOutputStorageDriver.java
(original)
+++ incubator/hcatalog/branches/branch-0.3/src/test/org/apache/hcatalog/rcfile/TestRCFileOutputStorageDriver.java
Thu Jan 19 19:28:05 2012
@@ -30,7 +30,6 @@ import org.apache.hadoop.hive.serde2.col
 import org.apache.hadoop.hive.serde2.columnar.BytesRefWritable;
 import org.apache.hadoop.mapreduce.JobContext;
 import org.apache.hadoop.mapreduce.JobID;
-import org.apache.hcatalog.common.ErrorType;
 import org.apache.hcatalog.common.HCatConstants;
 import org.apache.hcatalog.common.HCatException;
 import org.apache.hcatalog.common.HCatUtil;
@@ -39,14 +38,13 @@ import org.apache.hcatalog.data.schema.H
 import org.apache.hcatalog.mapreduce.HCatInputStorageDriver;
 import org.apache.hcatalog.mapreduce.HCatOutputStorageDriver;
 import org.apache.hcatalog.mapreduce.OutputJobInfo;
-import org.apache.hcatalog.rcfile.RCFileInputDriver;
-import org.apache.hcatalog.rcfile.RCFileOutputDriver;
+import org.apache.hcatalog.shims.HCatHadoopShims;
 
 public class TestRCFileOutputStorageDriver extends TestCase {
 
   public void testConversion() throws IOException {
     Configuration conf = new Configuration();
-    JobContext jc = new JobContext(conf, new JobID());
+    JobContext jc = HCatHadoopShims.Instance.get().createJobContext(conf, new JobID());
     String jobString = HCatUtil.serialize(OutputJobInfo.create(null,null,null,null,null));
     jc.getConfiguration().set(HCatConstants.HCAT_KEY_OUTPUT_INFO,jobString);
 



Mime
View raw message