hive-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From gunt...@apache.org
Subject svn commit: r1571600 [19/46] - in /hive/branches/tez: ./ ant/src/org/apache/hadoop/hive/ant/ common/src/java/org/apache/hadoop/hive/common/ common/src/java/org/apache/hadoop/hive/common/type/ common/src/java/org/apache/hadoop/hive/conf/ common/src/java...
Date Tue, 25 Feb 2014 07:59:20 GMT
Modified: hive/branches/tez/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestNewIntegerEncoding.java
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestNewIntegerEncoding.java?rev=1571600&r1=1571599&r2=1571600&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestNewIntegerEncoding.java (original)
+++ hive/branches/tez/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestNewIntegerEncoding.java Tue Feb 25 07:58:52 2014
@@ -128,7 +128,7 @@ public class TestNewIntegerEncoding {
     writer.addRow(new Row(111, 1111L));
     writer.close();
 
-    Reader reader = OrcFile.createReader(fs, testFilePath);
+    Reader reader = OrcFile.createReader(fs, testFilePath, conf);
     RecordReader rows = reader.rows(null);
     while (rows.hasNext()) {
       Object row = rows.next(null);
@@ -162,7 +162,7 @@ public class TestNewIntegerEncoding {
     }
     writer.close();
 
-    Reader reader = OrcFile.createReader(fs, testFilePath);
+    Reader reader = OrcFile.createReader(fs, testFilePath, conf);
     RecordReader rows = reader.rows(null);
     int idx = 0;
     while (rows.hasNext()) {
@@ -197,7 +197,7 @@ public class TestNewIntegerEncoding {
     }
     writer.close();
 
-    Reader reader = OrcFile.createReader(fs, testFilePath);
+    Reader reader = OrcFile.createReader(fs, testFilePath, conf);
     RecordReader rows = reader.rows(null);
     int idx = 0;
     while (rows.hasNext()) {
@@ -228,7 +228,7 @@ public class TestNewIntegerEncoding {
     }
     writer.close();
 
-    Reader reader = OrcFile.createReader(fs, testFilePath);
+    Reader reader = OrcFile.createReader(fs, testFilePath, conf);
     RecordReader rows = reader.rows(null);
     int idx = 0;
     while (rows.hasNext()) {
@@ -259,7 +259,7 @@ public class TestNewIntegerEncoding {
     }
     writer.close();
 
-    Reader reader = OrcFile.createReader(fs, testFilePath);
+    Reader reader = OrcFile.createReader(fs, testFilePath, conf);
     RecordReader rows = reader.rows(null);
     int idx = 0;
     while (rows.hasNext()) {
@@ -290,7 +290,7 @@ public class TestNewIntegerEncoding {
     }
     writer.close();
 
-    Reader reader = OrcFile.createReader(fs, testFilePath);
+    Reader reader = OrcFile.createReader(fs, testFilePath, conf);
     RecordReader rows = reader.rows(null);
     int idx = 0;
     while (rows.hasNext()) {
@@ -321,7 +321,7 @@ public class TestNewIntegerEncoding {
     }
     writer.close();
 
-    Reader reader = OrcFile.createReader(fs, testFilePath);
+    Reader reader = OrcFile.createReader(fs, testFilePath, conf);
     RecordReader rows = reader.rows(null);
     int idx = 0;
     while (rows.hasNext()) {
@@ -351,7 +351,7 @@ public class TestNewIntegerEncoding {
     }
     writer.close();
 
-    Reader reader = OrcFile.createReader(fs, testFilePath);
+    Reader reader = OrcFile.createReader(fs, testFilePath, conf);
     RecordReader rows = reader.rows(null);
     int idx = 0;
     while (rows.hasNext()) {
@@ -382,7 +382,7 @@ public class TestNewIntegerEncoding {
     }
     writer.close();
 
-    Reader reader = OrcFile.createReader(fs, testFilePath);
+    Reader reader = OrcFile.createReader(fs, testFilePath, conf);
     RecordReader rows = reader.rows(null);
     int idx = 0;
     while (rows.hasNext()) {
@@ -413,7 +413,7 @@ public class TestNewIntegerEncoding {
     }
     writer.close();
 
-    Reader reader = OrcFile.createReader(fs, testFilePath);
+    Reader reader = OrcFile.createReader(fs, testFilePath, conf);
     RecordReader rows = reader.rows(null);
     int idx = 0;
     while (rows.hasNext()) {
@@ -444,7 +444,7 @@ public class TestNewIntegerEncoding {
     }
     writer.close();
 
-    Reader reader = OrcFile.createReader(fs, testFilePath);
+    Reader reader = OrcFile.createReader(fs, testFilePath, conf);
     RecordReader rows = reader.rows(null);
     int idx = 0;
     while (rows.hasNext()) {
@@ -478,7 +478,7 @@ public class TestNewIntegerEncoding {
     }
     writer.close();
 
-    Reader reader = OrcFile.createReader(fs, testFilePath);
+    Reader reader = OrcFile.createReader(fs, testFilePath, conf);
     RecordReader rows = reader.rows(null);
     int idx = 0;
     while (rows.hasNext()) {
@@ -512,7 +512,7 @@ public class TestNewIntegerEncoding {
     }
     writer.close();
 
-    Reader reader = OrcFile.createReader(fs, testFilePath);
+    Reader reader = OrcFile.createReader(fs, testFilePath, conf);
     RecordReader rows = reader.rows(null);
     int idx = 0;
     while (rows.hasNext()) {
@@ -554,7 +554,7 @@ public class TestNewIntegerEncoding {
     }
     writer.close();
 
-    Reader reader = OrcFile.createReader(fs, testFilePath);
+    Reader reader = OrcFile.createReader(fs, testFilePath, conf);
     RecordReader rows = reader.rows(null);
     int idx = 0;
     while (rows.hasNext()) {
@@ -596,7 +596,7 @@ public class TestNewIntegerEncoding {
     }
     writer.close();
 
-    Reader reader = OrcFile.createReader(fs, testFilePath);
+    Reader reader = OrcFile.createReader(fs, testFilePath, conf);
     RecordReader rows = reader.rows(null);
     int idx = 0;
     while (rows.hasNext()) {
@@ -638,7 +638,7 @@ public class TestNewIntegerEncoding {
     }
     writer.close();
 
-    Reader reader = OrcFile.createReader(fs, testFilePath);
+    Reader reader = OrcFile.createReader(fs, testFilePath, conf);
     RecordReader rows = reader.rows(null);
     int idx = 0;
     while (rows.hasNext()) {
@@ -671,7 +671,7 @@ public class TestNewIntegerEncoding {
     }
     writer.close();
 
-    Reader reader = OrcFile.createReader(fs, testFilePath);
+    Reader reader = OrcFile.createReader(fs, testFilePath, conf);
     RecordReader rows = reader.rows(null);
     int idx = 0;
     while (rows.hasNext()) {
@@ -706,7 +706,7 @@ public class TestNewIntegerEncoding {
     }
     writer.close();
 
-    Reader reader = OrcFile.createReader(fs, testFilePath);
+    Reader reader = OrcFile.createReader(fs, testFilePath, conf);
     RecordReader rows = reader.rows(null);
     int idx = 0;
     while (rows.hasNext()) {
@@ -741,7 +741,7 @@ public class TestNewIntegerEncoding {
     }
     writer.close();
 
-    Reader reader = OrcFile.createReader(fs, testFilePath);
+    Reader reader = OrcFile.createReader(fs, testFilePath, conf);
     RecordReader rows = reader.rows(null);
     int idx = 0;
     while (rows.hasNext()) {
@@ -775,7 +775,7 @@ public class TestNewIntegerEncoding {
     }
     writer.close();
 
-    Reader reader = OrcFile.createReader(fs, testFilePath);
+    Reader reader = OrcFile.createReader(fs, testFilePath, conf);
     RecordReader rows = reader.rows(null);
     int idx = 0;
     while (rows.hasNext()) {
@@ -809,7 +809,7 @@ public class TestNewIntegerEncoding {
     }
     writer.close();
 
-    Reader reader = OrcFile.createReader(fs, testFilePath);
+    Reader reader = OrcFile.createReader(fs, testFilePath, conf);
     RecordReader rows = reader.rows(null);
     int idx = 0;
     while (rows.hasNext()) {
@@ -843,7 +843,7 @@ public class TestNewIntegerEncoding {
     }
     writer.close();
 
-    Reader reader = OrcFile.createReader(fs, testFilePath);
+    Reader reader = OrcFile.createReader(fs, testFilePath, conf);
     RecordReader rows = reader.rows(null);
     int idx = 0;
     while (rows.hasNext()) {
@@ -877,7 +877,166 @@ public class TestNewIntegerEncoding {
     }
     writer.close();
 
-    Reader reader = OrcFile.createReader(fs, testFilePath);
+    Reader reader = OrcFile.createReader(fs, testFilePath, conf);
+    RecordReader rows = reader.rows(null);
+    int idx = 0;
+    while (rows.hasNext()) {
+      Object row = rows.next(null);
+      assertEquals(input.get(idx++).longValue(), ((LongWritable) row).get());
+    }
+  }
+
+  @Test
+  public void testPatchedBaseMax1() throws Exception {
+    ObjectInspector inspector;
+    synchronized (TestOrcFile.class) {
+      inspector = ObjectInspectorFactory.getReflectionObjectInspector(Long.class,
+          ObjectInspectorFactory.ObjectInspectorOptions.JAVA);
+    }
+
+    List<Long> input = Lists.newArrayList();
+    Random rand = new Random();
+    for (int i = 0; i < 5120; i++) {
+      input.add((long) rand.nextInt(60));
+    }
+    input.set(511, Long.MAX_VALUE);
+
+    Writer writer = OrcFile.createWriter(testFilePath,
+        OrcFile.writerOptions(conf).inspector(inspector).stripeSize(100000).bufferSize(10000));
+    for (Long l : input) {
+      writer.addRow(l);
+    }
+    writer.close();
+
+    Reader reader = OrcFile.createReader(fs, testFilePath, conf);
+    RecordReader rows = reader.rows(null);
+    int idx = 0;
+    while (rows.hasNext()) {
+      Object row = rows.next(null);
+      assertEquals(input.get(idx++).longValue(), ((LongWritable) row).get());
+    }
+  }
+
+  @Test
+  public void testPatchedBaseMax2() throws Exception {
+    ObjectInspector inspector;
+    synchronized (TestOrcFile.class) {
+      inspector = ObjectInspectorFactory.getReflectionObjectInspector(Long.class,
+          ObjectInspectorFactory.ObjectInspectorOptions.JAVA);
+    }
+
+    List<Long> input = Lists.newArrayList();
+    Random rand = new Random();
+    for (int i = 0; i < 5120; i++) {
+      input.add((long) rand.nextInt(60));
+    }
+    input.set(128, Long.MAX_VALUE);
+    input.set(256, Long.MAX_VALUE);
+    input.set(511, Long.MAX_VALUE);
+
+    Writer writer = OrcFile.createWriter(testFilePath,
+        OrcFile.writerOptions(conf).inspector(inspector).stripeSize(100000).bufferSize(10000));
+    for (Long l : input) {
+      writer.addRow(l);
+    }
+    writer.close();
+
+    Reader reader = OrcFile.createReader(fs, testFilePath, conf);
+    RecordReader rows = reader.rows(null);
+    int idx = 0;
+    while (rows.hasNext()) {
+      Object row = rows.next(null);
+      assertEquals(input.get(idx++).longValue(), ((LongWritable) row).get());
+    }
+  }
+
+  @Test
+  public void testPatchedBaseMax3() throws Exception {
+    ObjectInspector inspector;
+    synchronized (TestOrcFile.class) {
+      inspector = ObjectInspectorFactory.getReflectionObjectInspector(Long.class,
+          ObjectInspectorFactory.ObjectInspectorOptions.JAVA);
+    }
+
+    List<Long> input = Lists.newArrayList();
+    input.add(371946367L);
+    input.add(11963367L);
+    input.add(68639400007L);
+    input.add(100233367L);
+    input.add(6367L);
+    input.add(10026367L);
+    input.add(3670000L);
+    input.add(3602367L);
+    input.add(4719226367L);
+    input.add(7196367L);
+    input.add(444442L);
+    input.add(210267L);
+    input.add(21033L);
+    input.add(160267L);
+    input.add(400267L);
+    input.add(23634347L);
+    input.add(16027L);
+    input.add(46026367L);
+    input.add(Long.MAX_VALUE);
+    input.add(33333L);
+
+    Writer writer = OrcFile.createWriter(testFilePath,
+        OrcFile.writerOptions(conf).inspector(inspector).stripeSize(100000).bufferSize(10000));
+    for (Long l : input) {
+      writer.addRow(l);
+    }
+    writer.close();
+
+    Reader reader = OrcFile.createReader(fs, testFilePath, conf);
+    RecordReader rows = reader.rows(null);
+    int idx = 0;
+    while (rows.hasNext()) {
+      Object row = rows.next(null);
+      assertEquals(input.get(idx++).longValue(), ((LongWritable) row).get());
+    }
+  }
+
+  @Test
+  public void testPatchedBaseMax4() throws Exception {
+    ObjectInspector inspector;
+    synchronized (TestOrcFile.class) {
+      inspector = ObjectInspectorFactory.getReflectionObjectInspector(Long.class,
+          ObjectInspectorFactory.ObjectInspectorOptions.JAVA);
+    }
+
+    List<Long> input = Lists.newArrayList();
+    for (int i = 0; i < 25; i++) {
+      input.add(371292224226367L);
+      input.add(119622332222267L);
+      input.add(686329400222007L);
+      input.add(100233333222367L);
+      input.add(636272333322222L);
+      input.add(10202633223267L);
+      input.add(36700222022230L);
+      input.add(36023226224227L);
+      input.add(47192226364427L);
+      input.add(71963622222447L);
+      input.add(22244444222222L);
+      input.add(21220263327442L);
+      input.add(21032233332232L);
+      input.add(16026322232227L);
+      input.add(40022262272212L);
+      input.add(23634342227222L);
+      input.add(16022222222227L);
+      input.add(46026362222227L);
+      input.add(46026362222227L);
+      input.add(33322222222323L);
+    }
+    input.add(Long.MAX_VALUE);
+
+    Writer writer = OrcFile.createWriter(testFilePath,
+        OrcFile.writerOptions(conf).inspector(inspector).stripeSize(100000).bufferSize(10000));
+    for (Long l : input) {
+      writer.addRow(l);
+    }
+    writer.close();
+
+    Reader reader = OrcFile.createReader(fs, testFilePath, conf);
     RecordReader rows = reader.rows(null);
     int idx = 0;
     while (rows.hasNext()) {
@@ -938,7 +1097,7 @@ public class TestNewIntegerEncoding {
 
     writer.close();
 
-    Reader reader = OrcFile.createReader(fs, testFilePath);
+    Reader reader = OrcFile.createReader(fs, testFilePath, conf);
     RecordReader rows = reader.rows(null);
     int idx = 0;
     while (rows.hasNext()) {
@@ -966,7 +1125,7 @@ public class TestNewIntegerEncoding {
     writer.addRow(-5535739865598783616L);
     writer.close();
 
-    Reader reader = OrcFile.createReader(fs, testFilePath);
+    Reader reader = OrcFile.createReader(fs, testFilePath, conf);
     RecordReader rows = reader.rows(null);
     Object row = rows.next(null);
     assertEquals(-7486502418706614742L, ((LongWritable) row).get());
@@ -1005,7 +1164,7 @@ public class TestNewIntegerEncoding {
     }
     writer.close();
 
-    Reader reader = OrcFile.createReader(fs, testFilePath);
+    Reader reader = OrcFile.createReader(fs, testFilePath, conf);
     RecordReader rows = reader.rows(null);
     int idx = 55555;
     rows.seekToRow(idx);

Modified: hive/branches/tez/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestOrcFile.java
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestOrcFile.java?rev=1571600&r1=1571599&r2=1571600&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestOrcFile.java (original)
+++ hive/branches/tez/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestOrcFile.java Tue Feb 25 07:58:52 2014
@@ -207,7 +207,7 @@ public class TestOrcFile {
   @Test
   public void testReadFormat_0_11() throws Exception {
     Path oldFilePath = new Path(HiveTestUtils.getFileFromClasspath("orc-file-11-format.orc"));
-    Reader reader = OrcFile.createReader(fs, oldFilePath);
+    Reader reader = OrcFile.createReader(fs, oldFilePath, conf);
 
     int stripeCount = 0;
     int rowCount = 0;
@@ -470,7 +470,7 @@ public class TestOrcFile {
     writer.addRow(new SimpleStruct(bytes(0,1,2,3,4,5), null));
     writer.addRow(new SimpleStruct(null, "hi"));
     writer.close();
-    Reader reader = OrcFile.createReader(fs, testFilePath);
+    Reader reader = OrcFile.createReader(fs, testFilePath, conf);
 
     // check the stats
     ColumnStatistics[] stats = reader.getStatistics();
@@ -565,7 +565,7 @@ public class TestOrcFile {
     }
 
     writer.close();
-    Reader reader = OrcFile.createReader(fs, testFilePath);
+    Reader reader = OrcFile.createReader(fs, testFilePath, conf);
     Metadata metadata = reader.getMetadata();
     int numStripes = metadata.getStripeStatistics().size();
     assertEquals(3, numStripes);
@@ -643,7 +643,7 @@ public class TestOrcFile {
         list(inner(100000000, "cat"), inner(-100000, "in"), inner(1234, "hat")),
         map(inner(5, "chani"), inner(1, "mauddib"))));
     writer.close();
-    Reader reader = OrcFile.createReader(fs, testFilePath);
+    Reader reader = OrcFile.createReader(fs, testFilePath, conf);
 
     Metadata metadata = reader.getMetadata();
 
@@ -893,7 +893,7 @@ public class TestOrcFile {
       writer.addRow(inner(x, y));
     }
     writer.close();
-    Reader reader = OrcFile.createReader(fs, testFilePath);
+    Reader reader = OrcFile.createReader(fs, testFilePath, conf);
 
     // check out the statistics
     ColumnStatistics[] stats = reader.getStatistics();
@@ -957,7 +957,7 @@ public class TestOrcFile {
                                          .compress(CompressionKind.NONE)
                                          .bufferSize(100));
     writer.close();
-    Reader reader = OrcFile.createReader(fs, testFilePath);
+    Reader reader = OrcFile.createReader(fs, testFilePath, conf);
     assertEquals(false, reader.rows(null).hasNext());
     assertEquals(CompressionKind.NONE, reader.getCompression());
     assertEquals(0, reader.getNumberOfRows());
@@ -994,7 +994,7 @@ public class TestOrcFile {
         null, null, null, null));
     writer.addUserMetadata("clobber", byteBuf(5,7,11,13,17,19));
     writer.close();
-    Reader reader = OrcFile.createReader(fs, testFilePath);
+    Reader reader = OrcFile.createReader(fs, testFilePath, conf);
     assertEquals(byteBuf(5,7,11,13,17,19), reader.getMetadataValue("clobber"));
     assertEquals(byteBuf(1,2,3,4,5,6,7,-1,-2,127,-128),
         reader.getMetadataValue("my.meta"));
@@ -1112,7 +1112,7 @@ public class TestOrcFile {
     union.set((byte) 0, new IntWritable(138));
     writer.addRow(row);
     writer.close();
-    Reader reader = OrcFile.createReader(fs, testFilePath);
+    Reader reader = OrcFile.createReader(fs, testFilePath, conf);
     assertEquals(false, reader.getMetadataKeys().iterator().hasNext());
     assertEquals(5309, reader.getNumberOfRows());
     DecimalColumnStatistics stats =
@@ -1243,7 +1243,7 @@ public class TestOrcFile {
           Integer.toHexString(rand.nextInt())));
     }
     writer.close();
-    Reader reader = OrcFile.createReader(fs, testFilePath);
+    Reader reader = OrcFile.createReader(fs, testFilePath, conf);
     RecordReader rows = reader.rows(null);
     rand = new Random(12);
     OrcStruct row = null;
@@ -1286,7 +1286,7 @@ public class TestOrcFile {
       }
     }
     writer.close();
-    Reader reader = OrcFile.createReader(fs, testFilePath);
+    Reader reader = OrcFile.createReader(fs, testFilePath, conf);
     assertEquals(50000, reader.getNumberOfRows());
     assertEquals(0, reader.getRowIndexStride());
     StripeInformation stripe = reader.getStripes().iterator().next();
@@ -1350,7 +1350,7 @@ public class TestOrcFile {
     }
     writer.close();
     writer = null;
-    Reader reader = OrcFile.createReader(fs, testFilePath);
+    Reader reader = OrcFile.createReader(fs, testFilePath, conf);
     assertEquals(COUNT, reader.getNumberOfRows());
     RecordReader rows = reader.rows(null);
     OrcStruct row = null;
@@ -1517,7 +1517,7 @@ public class TestOrcFile {
     }
     writer.close();
     assertEquals(null, memory.path);
-    Reader reader = OrcFile.createReader(fs, testFilePath);
+    Reader reader = OrcFile.createReader(fs, testFilePath, conf);
     int i = 0;
     for(StripeInformation stripe: reader.getStripes()) {
       i += 1;
@@ -1542,7 +1542,7 @@ public class TestOrcFile {
       writer.addRow(new InnerStruct(i*300, Integer.toHexString(10*i)));
     }
     writer.close();
-    Reader reader = OrcFile.createReader(fs, testFilePath);
+    Reader reader = OrcFile.createReader(fs, testFilePath, conf);
     assertEquals(3500, reader.getNumberOfRows());
 
     SearchArgument sarg = SearchArgument.FACTORY.newBuilder()

Modified: hive/branches/tez/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestOrcNullOptimization.java
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestOrcNullOptimization.java?rev=1571600&r1=1571599&r2=1571600&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestOrcNullOptimization.java (original)
+++ hive/branches/tez/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestOrcNullOptimization.java Tue Feb 25 07:58:52 2014
@@ -108,7 +108,7 @@ public class TestOrcNullOptimization {
                                Lists.newArrayList(new InnerStruct(100))));
     writer.close();
 
-    Reader reader = OrcFile.createReader(fs, testFilePath);
+    Reader reader = OrcFile.createReader(fs, testFilePath, conf);
     // check the stats
     ColumnStatistics[] stats = reader.getStatistics();
     assertEquals(20000, reader.getNumberOfRows());
@@ -212,7 +212,7 @@ public class TestOrcNullOptimization {
                                Lists.newArrayList(new InnerStruct(100))));
     writer.close();
 
-    Reader reader = OrcFile.createReader(fs, testFilePath);
+    Reader reader = OrcFile.createReader(fs, testFilePath, conf);
     // check the stats
     ColumnStatistics[] stats = reader.getStatistics();
     assertEquals(20000, reader.getNumberOfRows());
@@ -313,7 +313,7 @@ public class TestOrcNullOptimization {
                                Lists.newArrayList(new InnerStruct(100))));
     writer.close();
 
-    Reader reader = OrcFile.createReader(fs, testFilePath);
+    Reader reader = OrcFile.createReader(fs, testFilePath, conf);
     // check the stats
     ColumnStatistics[] stats = reader.getStatistics();
     assertEquals(8, reader.getNumberOfRows());

Modified: hive/branches/tez/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestOrcSerDeStats.java
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestOrcSerDeStats.java?rev=1571600&r1=1571599&r2=1571600&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestOrcSerDeStats.java (original)
+++ hive/branches/tez/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestOrcSerDeStats.java Tue Feb 25 07:58:52 2014
@@ -212,7 +212,7 @@ public class TestOrcSerDeStats {
     writer.close();
     assertEquals(4, writer.getNumberOfRows());
     assertEquals(273, writer.getRawDataSize());
-    Reader reader = OrcFile.createReader(fs, testFilePath);
+    Reader reader = OrcFile.createReader(fs, testFilePath, conf);
     assertEquals(4, reader.getNumberOfRows());
     assertEquals(273, reader.getRawDataSize());
     assertEquals(15, reader.getRawDataSizeOfColumns(Lists.newArrayList("bytes1")));
@@ -310,7 +310,7 @@ public class TestOrcSerDeStats {
     assertEquals(5000, writer.getNumberOfRows());
     assertEquals(430000000, writer.getRawDataSize());
 
-    Reader reader = OrcFile.createReader(fs, testFilePath);
+    Reader reader = OrcFile.createReader(fs, testFilePath, conf);
     // stats from reader
     assertEquals(5000, reader.getNumberOfRows());
     assertEquals(430000000, reader.getRawDataSize());
@@ -341,7 +341,7 @@ public class TestOrcSerDeStats {
     assertEquals(1000, writer.getNumberOfRows());
     assertEquals(950000, writer.getRawDataSize());
 
-    Reader reader = OrcFile.createReader(fs, testFilePath);
+    Reader reader = OrcFile.createReader(fs, testFilePath, conf);
     // stats from reader
     assertEquals(1000, reader.getNumberOfRows());
     assertEquals(950000, reader.getRawDataSize());
@@ -372,7 +372,7 @@ public class TestOrcSerDeStats {
     assertEquals(1000, writer.getNumberOfRows());
     assertEquals(44500, writer.getRawDataSize());
 
-    Reader reader = OrcFile.createReader(fs, testFilePath);
+    Reader reader = OrcFile.createReader(fs, testFilePath, conf);
     // stats from reader
     assertEquals(1000, reader.getNumberOfRows());
     assertEquals(44500, reader.getRawDataSize());
@@ -413,7 +413,7 @@ public class TestOrcSerDeStats {
     long rawDataSize = writer.getRawDataSize();
     assertEquals(2, rowCount);
     assertEquals(1740, rawDataSize);
-    Reader reader = OrcFile.createReader(fs, testFilePath);
+    Reader reader = OrcFile.createReader(fs, testFilePath, conf);
 
     assertEquals(2, reader.getNumberOfRows());
     assertEquals(1740, reader.getRawDataSize());
@@ -506,7 +506,7 @@ public class TestOrcSerDeStats {
     long rawDataSize = writer.getRawDataSize();
     assertEquals(2, rowCount);
     assertEquals(1740, rawDataSize);
-    Reader reader = OrcFile.createReader(fs, testFilePath);
+    Reader reader = OrcFile.createReader(fs, testFilePath, conf);
 
     assertEquals(2, reader.getNumberOfRows());
     assertEquals(1740, reader.getRawDataSize());
@@ -573,7 +573,7 @@ public class TestOrcSerDeStats {
   @Test(expected = ClassCastException.class)
   public void testSerdeStatsOldFormat() throws Exception {
     Path oldFilePath = new Path(HiveTestUtils.getFileFromClasspath("orc-file-11-format.orc"));
-    Reader reader = OrcFile.createReader(fs, oldFilePath);
+    Reader reader = OrcFile.createReader(fs, oldFilePath, conf);
 
     int stripeCount = 0;
     int rowCount = 0;

Modified: hive/branches/tez/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestOrcSplitElimination.java
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestOrcSplitElimination.java?rev=1571600&r1=1571599&r2=1571600&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestOrcSplitElimination.java (original)
+++ hive/branches/tez/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestOrcSplitElimination.java Tue Feb 25 07:58:52 2014
@@ -1,3 +1,21 @@
+/**
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ *     http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
 package org.apache.hadoop.hive.ql.io.orc;
 
 import static org.junit.Assert.assertEquals;

Modified: hive/branches/tez/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestVectorizedORCReader.java
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestVectorizedORCReader.java?rev=1571600&r1=1571599&r2=1571600&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestVectorizedORCReader.java (original)
+++ hive/branches/tez/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestVectorizedORCReader.java Tue Feb 25 07:58:52 2014
@@ -131,8 +131,8 @@ public class TestVectorizedORCReader {
 
   private void checkVectorizedReader() throws Exception {
 
-    Reader vreader = OrcFile.createReader(testFilePath.getFileSystem(conf), testFilePath);
-    Reader reader = OrcFile.createReader(testFilePath.getFileSystem(conf), testFilePath);
+    Reader vreader = OrcFile.createReader(testFilePath.getFileSystem(conf), testFilePath, conf);
+    Reader reader = OrcFile.createReader(testFilePath.getFileSystem(conf), testFilePath, conf);
     RecordReaderImpl vrr = (RecordReaderImpl) vreader.rows(null);
     RecordReaderImpl rr = (RecordReaderImpl) reader.rows(null);
     VectorizedRowBatch batch = null;

Modified: hive/branches/tez/ql/src/test/org/apache/hadoop/hive/ql/parse/TestQBSubQuery.java
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/org/apache/hadoop/hive/ql/parse/TestQBSubQuery.java?rev=1571600&r1=1571599&r2=1571600&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/org/apache/hadoop/hive/ql/parse/TestQBSubQuery.java (original)
+++ hive/branches/tez/ql/src/test/org/apache/hadoop/hive/ql/parse/TestQBSubQuery.java Tue Feb 25 07:58:52 2014
@@ -1,3 +1,21 @@
+/**
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ *     http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
 package org.apache.hadoop.hive.ql.parse;
 
 import java.util.ArrayList;

Modified: hive/branches/tez/ql/src/test/org/apache/hadoop/hive/ql/parse/authorization/TestSessionUserName.java
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/org/apache/hadoop/hive/ql/parse/authorization/TestSessionUserName.java?rev=1571600&r1=1571599&r2=1571600&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/org/apache/hadoop/hive/ql/parse/authorization/TestSessionUserName.java (original)
+++ hive/branches/tez/ql/src/test/org/apache/hadoop/hive/ql/parse/authorization/TestSessionUserName.java Tue Feb 25 07:58:52 2014
@@ -23,6 +23,7 @@ import org.apache.hadoop.hive.conf.HiveC
 import org.apache.hadoop.hive.metastore.api.MetaException;
 import org.apache.hadoop.hive.ql.metadata.Hive;
 import org.apache.hadoop.hive.ql.security.HiveAuthenticationProvider;
+import org.apache.hadoop.hive.ql.security.SessionStateUserAuthenticator;
 import org.apache.hadoop.hive.ql.security.authorization.plugin.HiveAuthorizer;
 import org.apache.hadoop.hive.ql.security.authorization.plugin.HiveAuthorizerFactory;
 import org.apache.hadoop.hive.ql.security.authorization.plugin.HiveAuthorizerImpl;
@@ -96,6 +97,8 @@ public class TestSessionUserName {
     HiveConf conf = new HiveConf();
     conf.setVar(HiveConf.ConfVars.HIVE_AUTHORIZATION_MANAGER,
         HiveAuthorizerStoringUserNameFactory.class.getName());
+    conf.setVar(HiveConf.ConfVars.HIVE_AUTHENTICATOR_MANAGER,
+        SessionStateUserAuthenticator.class.getName());
     return conf;
   }
 

Modified: hive/branches/tez/ql/src/test/org/apache/hadoop/hive/ql/processors/TestCompileProcessor.java
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/org/apache/hadoop/hive/ql/processors/TestCompileProcessor.java?rev=1571600&r1=1571599&r2=1571600&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/org/apache/hadoop/hive/ql/processors/TestCompileProcessor.java (original)
+++ hive/branches/tez/ql/src/test/org/apache/hadoop/hive/ql/processors/TestCompileProcessor.java Tue Feb 25 07:58:52 2014
@@ -1,3 +1,21 @@
+/**
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ *     http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
 package org.apache.hadoop.hive.ql.processors;
 
 import java.io.File;

Modified: hive/branches/tez/ql/src/test/org/apache/hadoop/hive/ql/testutil/DataBuilder.java
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/org/apache/hadoop/hive/ql/testutil/DataBuilder.java?rev=1571600&r1=1571599&r2=1571600&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/org/apache/hadoop/hive/ql/testutil/DataBuilder.java (original)
+++ hive/branches/tez/ql/src/test/org/apache/hadoop/hive/ql/testutil/DataBuilder.java Tue Feb 25 07:58:52 2014
@@ -1,3 +1,21 @@
+/**
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ *     http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
 package org.apache.hadoop.hive.ql.testutil;
 
 import java.util.ArrayList;

Modified: hive/branches/tez/ql/src/test/org/apache/hadoop/hive/ql/testutil/OperatorTestUtils.java
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/org/apache/hadoop/hive/ql/testutil/OperatorTestUtils.java?rev=1571600&r1=1571599&r2=1571600&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/org/apache/hadoop/hive/ql/testutil/OperatorTestUtils.java (original)
+++ hive/branches/tez/ql/src/test/org/apache/hadoop/hive/ql/testutil/OperatorTestUtils.java Tue Feb 25 07:58:52 2014
@@ -1,3 +1,21 @@
+/**
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ *     http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
 package org.apache.hadoop.hive.ql.testutil;
 
 import java.util.ArrayList;

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/authorization_revoke_table_priv.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/authorization_revoke_table_priv.q?rev=1571600&r1=1571599&r2=1571600&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/authorization_revoke_table_priv.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/authorization_revoke_table_priv.q Tue Feb 25 07:58:52 2014
@@ -4,7 +4,7 @@ set hive.security.authenticator.manager=
 set user.name=user1;
 -- current user has been set (comment line before the set cmd is resulting in parse error!!)
 
-CREATE TABLE  table_priv_rev(i int);
+CREATE TABLE table_priv_rev(i int);
 
 -- grant insert privilege to user2
 GRANT INSERT ON table_priv_rev TO USER user2;
@@ -48,3 +48,10 @@ SHOW GRANT USER user2 ON TABLE table_pri
 REVOKE SELECT ON TABLE table_priv_rev FROM USER user2;
 SHOW GRANT USER user2 ON TABLE table_priv_rev;
 
+
+-- grant all followed by revoke all
+GRANT ALL ON table_priv_rev TO USER user2;
+SHOW GRANT USER user2 ON TABLE table_priv_rev;
+
+REVOKE ALL ON TABLE table_priv_rev FROM USER user2;
+SHOW GRANT USER user2 ON TABLE table_priv_rev;

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/vectorization_short_regress.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/vectorization_short_regress.q?rev=1571600&r1=1571599&r2=1571600&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/vectorization_short_regress.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/vectorization_short_regress.q Tue Feb 25 07:58:52 2014
@@ -395,10 +395,8 @@ WHERE  (((197 > ctinyint)
         OR (cbigint = 359)
         OR (cboolean1 < 0)
            OR ((cstring1 LIKE '%ss')
-               AND (cfloat <= ctinyint))
-              OR ((16558 != ctimestamp1)
-                  AND ((0 >= ctimestamp2)
-                       AND (2563.58 < csmallint))));
+               AND (cfloat <= ctinyint)));
+
 SELECT cint,
        cbigint,
        cstring1,
@@ -430,10 +428,7 @@ WHERE  (((197 > ctinyint)
         OR (cbigint = 359)
         OR (cboolean1 < 0)
            OR ((cstring1 LIKE '%ss')
-               AND (cfloat <= ctinyint))
-              OR ((16558 != ctimestamp1)
-                  AND ((0 >= ctimestamp2)
-                       AND (2563.58 < csmallint))));
+               AND (cfloat <= ctinyint)));
 
 -- TargetTypeClasses: String, Bool, Double, Long, Timestamp
 -- Functions: Sum, Max, Avg, Var, StDevP, VarP

Modified: hive/branches/tez/ql/src/test/results/clientnegative/add_partition_with_whitelist.q.out
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/results/clientnegative/add_partition_with_whitelist.q.out?rev=1571600&r1=1571599&r2=1571600&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/results/clientnegative/add_partition_with_whitelist.q.out (original)
+++ hive/branches/tez/ql/src/test/results/clientnegative/add_partition_with_whitelist.q.out Tue Feb 25 07:58:52 2014
@@ -2,10 +2,12 @@ PREHOOK: query: -- This pattern matches 
 
 CREATE TABLE part_whitelist_test (key STRING, value STRING) PARTITIONED BY (ds STRING)
 PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
 POSTHOOK: query: -- This pattern matches all printable ASCII characters (disallow unicode) and disallows commas
 
 CREATE TABLE part_whitelist_test (key STRING, value STRING) PARTITIONED BY (ds STRING)
 POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@part_whitelist_test
 PREHOOK: query: SHOW PARTITIONS part_whitelist_test
 PREHOOK: type: SHOWPARTITIONS
@@ -13,5 +15,5 @@ POSTHOOK: query: SHOW PARTITIONS part_wh
 POSTHOOK: type: SHOWPARTITIONS
 PREHOOK: query: ALTER TABLE part_whitelist_test ADD PARTITION (ds='1,2,3,4')
 PREHOOK: type: ALTERTABLE_ADDPARTS
-PREHOOK: Input: default@part_whitelist_test
+PREHOOK: Output: default@part_whitelist_test
 FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.DDLTask. MetaException(message:Partition value '1,2,3,4' contains a character not matched by whitelist pattern '[\\x20-\\x7E&&[^,]]*'.  (configure with hive.metastore.partition.name.whitelist.pattern))

Modified: hive/branches/tez/ql/src/test/results/clientnegative/addpart1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/results/clientnegative/addpart1.q.out?rev=1571600&r1=1571599&r2=1571600&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/results/clientnegative/addpart1.q.out (original)
+++ hive/branches/tez/ql/src/test/results/clientnegative/addpart1.q.out Tue Feb 25 07:58:52 2014
@@ -1,14 +1,16 @@
 PREHOOK: query: create table addpart1 (a int) partitioned by (b string, c string)
 PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
 POSTHOOK: query: create table addpart1 (a int) partitioned by (b string, c string)
 POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@addpart1
 PREHOOK: query: alter table addpart1 add partition (b='f', c='s')
 PREHOOK: type: ALTERTABLE_ADDPARTS
-PREHOOK: Input: default@addpart1
+PREHOOK: Output: default@addpart1
 POSTHOOK: query: alter table addpart1 add partition (b='f', c='s')
 POSTHOOK: type: ALTERTABLE_ADDPARTS
-POSTHOOK: Input: default@addpart1
+POSTHOOK: Output: default@addpart1
 POSTHOOK: Output: default@addpart1@b=f/c=s
 PREHOOK: query: show partitions addpart1
 PREHOOK: type: SHOWPARTITIONS
@@ -17,5 +19,5 @@ POSTHOOK: type: SHOWPARTITIONS
 b=f/c=s
 PREHOOK: query: alter table addpart1 add partition (b='f', c='')
 PREHOOK: type: ALTERTABLE_ADDPARTS
-PREHOOK: Input: default@addpart1
+PREHOOK: Output: default@addpart1
 FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.DDLTask. partition spec is invalid; field c does not exist or is empty

Modified: hive/branches/tez/ql/src/test/results/clientnegative/alter_concatenate_indexed_table.q.out
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/results/clientnegative/alter_concatenate_indexed_table.q.out?rev=1571600&r1=1571599&r2=1571600&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/results/clientnegative/alter_concatenate_indexed_table.q.out (original)
+++ hive/branches/tez/ql/src/test/results/clientnegative/alter_concatenate_indexed_table.q.out Tue Feb 25 07:58:52 2014
@@ -1,25 +1,33 @@
 PREHOOK: query: create table src_rc_concatenate_test(key int, value string) stored as rcfile
 PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
 POSTHOOK: query: create table src_rc_concatenate_test(key int, value string) stored as rcfile
 POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@src_rc_concatenate_test
 PREHOOK: query: load data local inpath '../../data/files/smbbucket_1.rc' into table src_rc_concatenate_test
 PREHOOK: type: LOAD
+#### A masked pattern was here ####
 PREHOOK: Output: default@src_rc_concatenate_test
 POSTHOOK: query: load data local inpath '../../data/files/smbbucket_1.rc' into table src_rc_concatenate_test
 POSTHOOK: type: LOAD
+#### A masked pattern was here ####
 POSTHOOK: Output: default@src_rc_concatenate_test
 PREHOOK: query: load data local inpath '../../data/files/smbbucket_2.rc' into table src_rc_concatenate_test
 PREHOOK: type: LOAD
+#### A masked pattern was here ####
 PREHOOK: Output: default@src_rc_concatenate_test
 POSTHOOK: query: load data local inpath '../../data/files/smbbucket_2.rc' into table src_rc_concatenate_test
 POSTHOOK: type: LOAD
+#### A masked pattern was here ####
 POSTHOOK: Output: default@src_rc_concatenate_test
 PREHOOK: query: load data local inpath '../../data/files/smbbucket_3.rc' into table src_rc_concatenate_test
 PREHOOK: type: LOAD
+#### A masked pattern was here ####
 PREHOOK: Output: default@src_rc_concatenate_test
 POSTHOOK: query: load data local inpath '../../data/files/smbbucket_3.rc' into table src_rc_concatenate_test
 POSTHOOK: type: LOAD
+#### A masked pattern was here ####
 POSTHOOK: Output: default@src_rc_concatenate_test
 PREHOOK: query: show table extended like `src_rc_concatenate_test`
 PREHOOK: type: SHOW_TABLESTATUS

Modified: hive/branches/tez/ql/src/test/results/clientnegative/alter_non_native.q.out
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/results/clientnegative/alter_non_native.q.out?rev=1571600&r1=1571599&r2=1571600&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/results/clientnegative/alter_non_native.q.out (original)
+++ hive/branches/tez/ql/src/test/results/clientnegative/alter_non_native.q.out Tue Feb 25 07:58:52 2014
@@ -1,8 +1,10 @@
 PREHOOK: query: CREATE TABLE non_native1(key int, value string) 
 STORED BY 'org.apache.hadoop.hive.ql.metadata.DefaultStorageHandler'
 PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
 POSTHOOK: query: CREATE TABLE non_native1(key int, value string) 
 STORED BY 'org.apache.hadoop.hive.ql.metadata.DefaultStorageHandler'
 POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@non_native1
 FAILED: SemanticException [Error 10134]: ALTER TABLE cannot be used for a non-native table non_native1

Modified: hive/branches/tez/ql/src/test/results/clientnegative/alter_partition_coltype_2columns.q.out
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/results/clientnegative/alter_partition_coltype_2columns.q.out?rev=1571600&r1=1571599&r2=1571600&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/results/clientnegative/alter_partition_coltype_2columns.q.out (original)
+++ hive/branches/tez/ql/src/test/results/clientnegative/alter_partition_coltype_2columns.q.out Tue Feb 25 07:58:52 2014
@@ -1,9 +1,11 @@
 PREHOOK: query: -- create testing table
 create table alter_coltype(key string, value string) partitioned by (dt string, ts string)
 PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
 POSTHOOK: query: -- create testing table
 create table alter_coltype(key string, value string) partitioned by (dt string, ts string)
 POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@alter_coltype
 PREHOOK: query: -- insert and create a partition
 insert overwrite table alter_coltype partition(dt='100x', ts='6:30pm') select * from src1

Modified: hive/branches/tez/ql/src/test/results/clientnegative/alter_partition_coltype_invalidcolname.q.out
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/results/clientnegative/alter_partition_coltype_invalidcolname.q.out?rev=1571600&r1=1571599&r2=1571600&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/results/clientnegative/alter_partition_coltype_invalidcolname.q.out (original)
+++ hive/branches/tez/ql/src/test/results/clientnegative/alter_partition_coltype_invalidcolname.q.out Tue Feb 25 07:58:52 2014
@@ -1,9 +1,11 @@
 PREHOOK: query: -- create testing table
 create table alter_coltype(key string, value string) partitioned by (dt string, ts string)
 PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
 POSTHOOK: query: -- create testing table
 create table alter_coltype(key string, value string) partitioned by (dt string, ts string)
 POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@alter_coltype
 PREHOOK: query: -- insert and create a partition
 insert overwrite table alter_coltype partition(dt='100x', ts='6:30pm') select * from src1

Modified: hive/branches/tez/ql/src/test/results/clientnegative/alter_partition_coltype_invalidtype.q.out
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/results/clientnegative/alter_partition_coltype_invalidtype.q.out?rev=1571600&r1=1571599&r2=1571600&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/results/clientnegative/alter_partition_coltype_invalidtype.q.out (original)
+++ hive/branches/tez/ql/src/test/results/clientnegative/alter_partition_coltype_invalidtype.q.out Tue Feb 25 07:58:52 2014
@@ -1,9 +1,11 @@
 PREHOOK: query: -- create testing table
 create table alter_coltype(key string, value string) partitioned by (dt string, ts string)
 PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
 POSTHOOK: query: -- create testing table
 create table alter_coltype(key string, value string) partitioned by (dt string, ts string)
 POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@alter_coltype
 PREHOOK: query: -- insert and create a partition
 insert overwrite table alter_coltype partition(dt='100x', ts='6:30pm') select * from src1

Modified: hive/branches/tez/ql/src/test/results/clientnegative/alter_partition_invalidspec.q.out
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/results/clientnegative/alter_partition_invalidspec.q.out?rev=1571600&r1=1571599&r2=1571600&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/results/clientnegative/alter_partition_invalidspec.q.out (original)
+++ hive/branches/tez/ql/src/test/results/clientnegative/alter_partition_invalidspec.q.out Tue Feb 25 07:58:52 2014
@@ -1,24 +1,30 @@
 PREHOOK: query: -- Create table
 create table if not exists alter_part_invalidspec(key string, value string ) partitioned by (year string, month string) stored as textfile
 PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
 POSTHOOK: query: -- Create table
 create table if not exists alter_part_invalidspec(key string, value string ) partitioned by (year string, month string) stored as textfile
 POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@alter_part_invalidspec
 PREHOOK: query: -- Load data
 load data local inpath '../../data/files/T1.txt' overwrite into table alter_part_invalidspec partition (year='1996', month='10')
 PREHOOK: type: LOAD
+#### A masked pattern was here ####
 PREHOOK: Output: default@alter_part_invalidspec
 POSTHOOK: query: -- Load data
 load data local inpath '../../data/files/T1.txt' overwrite into table alter_part_invalidspec partition (year='1996', month='10')
 POSTHOOK: type: LOAD
+#### A masked pattern was here ####
 POSTHOOK: Output: default@alter_part_invalidspec
 POSTHOOK: Output: default@alter_part_invalidspec@year=1996/month=10
 PREHOOK: query: load data local inpath '../../data/files/T1.txt' overwrite into table alter_part_invalidspec partition (year='1996', month='12')
 PREHOOK: type: LOAD
+#### A masked pattern was here ####
 PREHOOK: Output: default@alter_part_invalidspec
 POSTHOOK: query: load data local inpath '../../data/files/T1.txt' overwrite into table alter_part_invalidspec partition (year='1996', month='12')
 POSTHOOK: type: LOAD
+#### A masked pattern was here ####
 POSTHOOK: Output: default@alter_part_invalidspec
 POSTHOOK: Output: default@alter_part_invalidspec@year=1996/month=12
 FAILED: SemanticException [Error 10006]: Partition not found {year=1997}

Modified: hive/branches/tez/ql/src/test/results/clientnegative/alter_partition_nodrop.q.out
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/results/clientnegative/alter_partition_nodrop.q.out?rev=1571600&r1=1571599&r2=1571600&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/results/clientnegative/alter_partition_nodrop.q.out (original)
+++ hive/branches/tez/ql/src/test/results/clientnegative/alter_partition_nodrop.q.out Tue Feb 25 07:58:52 2014
@@ -1,24 +1,30 @@
 PREHOOK: query: -- Create table
 create table if not exists alter_part_nodrop_part(key string, value string ) partitioned by (year string, month string) stored as textfile
 PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
 POSTHOOK: query: -- Create table
 create table if not exists alter_part_nodrop_part(key string, value string ) partitioned by (year string, month string) stored as textfile
 POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@alter_part_nodrop_part
 PREHOOK: query: -- Load data
 load data local inpath '../../data/files/T1.txt' overwrite into table alter_part_nodrop_part partition (year='1996', month='10')
 PREHOOK: type: LOAD
+#### A masked pattern was here ####
 PREHOOK: Output: default@alter_part_nodrop_part
 POSTHOOK: query: -- Load data
 load data local inpath '../../data/files/T1.txt' overwrite into table alter_part_nodrop_part partition (year='1996', month='10')
 POSTHOOK: type: LOAD
+#### A masked pattern was here ####
 POSTHOOK: Output: default@alter_part_nodrop_part
 POSTHOOK: Output: default@alter_part_nodrop_part@year=1996/month=10
 PREHOOK: query: load data local inpath '../../data/files/T1.txt' overwrite into table alter_part_nodrop_part partition (year='1996', month='12')
 PREHOOK: type: LOAD
+#### A masked pattern was here ####
 PREHOOK: Output: default@alter_part_nodrop_part
 POSTHOOK: query: load data local inpath '../../data/files/T1.txt' overwrite into table alter_part_nodrop_part partition (year='1996', month='12')
 POSTHOOK: type: LOAD
+#### A masked pattern was here ####
 POSTHOOK: Output: default@alter_part_nodrop_part
 POSTHOOK: Output: default@alter_part_nodrop_part@year=1996/month=12
 PREHOOK: query: alter table alter_part_nodrop_part partition (year='1996') enable no_drop

Modified: hive/branches/tez/ql/src/test/results/clientnegative/alter_partition_nodrop_table.q.out
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/results/clientnegative/alter_partition_nodrop_table.q.out?rev=1571600&r1=1571599&r2=1571600&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/results/clientnegative/alter_partition_nodrop_table.q.out (original)
+++ hive/branches/tez/ql/src/test/results/clientnegative/alter_partition_nodrop_table.q.out Tue Feb 25 07:58:52 2014
@@ -1,24 +1,30 @@
 PREHOOK: query: -- Create table
 create table if not exists alter_part_nodrop_table(key string, value string ) partitioned by (year string, month string) stored as textfile
 PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
 POSTHOOK: query: -- Create table
 create table if not exists alter_part_nodrop_table(key string, value string ) partitioned by (year string, month string) stored as textfile
 POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@alter_part_nodrop_table
 PREHOOK: query: -- Load data
 load data local inpath '../../data/files/T1.txt' overwrite into table alter_part_nodrop_table partition (year='1996', month='10')
 PREHOOK: type: LOAD
+#### A masked pattern was here ####
 PREHOOK: Output: default@alter_part_nodrop_table
 POSTHOOK: query: -- Load data
 load data local inpath '../../data/files/T1.txt' overwrite into table alter_part_nodrop_table partition (year='1996', month='10')
 POSTHOOK: type: LOAD
+#### A masked pattern was here ####
 POSTHOOK: Output: default@alter_part_nodrop_table
 POSTHOOK: Output: default@alter_part_nodrop_table@year=1996/month=10
 PREHOOK: query: load data local inpath '../../data/files/T1.txt' overwrite into table alter_part_nodrop_table partition (year='1996', month='12')
 PREHOOK: type: LOAD
+#### A masked pattern was here ####
 PREHOOK: Output: default@alter_part_nodrop_table
 POSTHOOK: query: load data local inpath '../../data/files/T1.txt' overwrite into table alter_part_nodrop_table partition (year='1996', month='12')
 POSTHOOK: type: LOAD
+#### A masked pattern was here ####
 POSTHOOK: Output: default@alter_part_nodrop_table
 POSTHOOK: Output: default@alter_part_nodrop_table@year=1996/month=12
 PREHOOK: query: alter table alter_part_nodrop_table partition (year='1996') enable no_drop

Modified: hive/branches/tez/ql/src/test/results/clientnegative/alter_partition_offline.q.out
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/results/clientnegative/alter_partition_offline.q.out?rev=1571600&r1=1571599&r2=1571600&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/results/clientnegative/alter_partition_offline.q.out (original)
+++ hive/branches/tez/ql/src/test/results/clientnegative/alter_partition_offline.q.out Tue Feb 25 07:58:52 2014
@@ -1,24 +1,30 @@
 PREHOOK: query: -- create table
 create table if not exists alter_part_offline (key string, value string ) partitioned by (year string, month string) stored as textfile
 PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
 POSTHOOK: query: -- create table
 create table if not exists alter_part_offline (key string, value string ) partitioned by (year string, month string) stored as textfile
 POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@alter_part_offline
 PREHOOK: query: -- Load data
 load data local inpath '../../data/files/T1.txt' overwrite into table alter_part_offline partition (year='1996', month='10')
 PREHOOK: type: LOAD
+#### A masked pattern was here ####
 PREHOOK: Output: default@alter_part_offline
 POSTHOOK: query: -- Load data
 load data local inpath '../../data/files/T1.txt' overwrite into table alter_part_offline partition (year='1996', month='10')
 POSTHOOK: type: LOAD
+#### A masked pattern was here ####
 POSTHOOK: Output: default@alter_part_offline
 POSTHOOK: Output: default@alter_part_offline@year=1996/month=10
 PREHOOK: query: load data local inpath '../../data/files/T1.txt' overwrite into table alter_part_offline partition (year='1996', month='12')
 PREHOOK: type: LOAD
+#### A masked pattern was here ####
 PREHOOK: Output: default@alter_part_offline
 POSTHOOK: query: load data local inpath '../../data/files/T1.txt' overwrite into table alter_part_offline partition (year='1996', month='12')
 POSTHOOK: type: LOAD
+#### A masked pattern was here ####
 POSTHOOK: Output: default@alter_part_offline
 POSTHOOK: Output: default@alter_part_offline@year=1996/month=12
 PREHOOK: query: alter table alter_part_offline partition (year='1996') disable offline

Modified: hive/branches/tez/ql/src/test/results/clientnegative/alter_partition_with_whitelist.q.out
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/results/clientnegative/alter_partition_with_whitelist.q.out?rev=1571600&r1=1571599&r2=1571600&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/results/clientnegative/alter_partition_with_whitelist.q.out (original)
+++ hive/branches/tez/ql/src/test/results/clientnegative/alter_partition_with_whitelist.q.out Tue Feb 25 07:58:52 2014
@@ -2,10 +2,12 @@ PREHOOK: query: -- This pattern matches 
 
 CREATE TABLE part_whitelist_test (key STRING, value STRING) PARTITIONED BY (ds STRING)
 PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
 POSTHOOK: query: -- This pattern matches all printable ASCII characters (disallow unicode) and disallows commas
 
 CREATE TABLE part_whitelist_test (key STRING, value STRING) PARTITIONED BY (ds STRING)
 POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@part_whitelist_test
 PREHOOK: query: SHOW PARTITIONS part_whitelist_test
 PREHOOK: type: SHOWPARTITIONS
@@ -13,10 +15,10 @@ POSTHOOK: query: SHOW PARTITIONS part_wh
 POSTHOOK: type: SHOWPARTITIONS
 PREHOOK: query: ALTER TABLE part_whitelist_test ADD PARTITION (ds='1')
 PREHOOK: type: ALTERTABLE_ADDPARTS
-PREHOOK: Input: default@part_whitelist_test
+PREHOOK: Output: default@part_whitelist_test
 POSTHOOK: query: ALTER TABLE part_whitelist_test ADD PARTITION (ds='1')
 POSTHOOK: type: ALTERTABLE_ADDPARTS
-POSTHOOK: Input: default@part_whitelist_test
+POSTHOOK: Output: default@part_whitelist_test
 POSTHOOK: Output: default@part_whitelist_test@ds=1
 PREHOOK: query: ALTER TABLE part_whitelist_test PARTITION (ds='1') rename to partition (ds='1,2,3')
 PREHOOK: type: ALTERTABLE_RENAMEPART

Modified: hive/branches/tez/ql/src/test/results/clientnegative/alter_rename_partition_failure.q.out
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/results/clientnegative/alter_rename_partition_failure.q.out?rev=1571600&r1=1571599&r2=1571600&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/results/clientnegative/alter_rename_partition_failure.q.out (original)
+++ hive/branches/tez/ql/src/test/results/clientnegative/alter_rename_partition_failure.q.out Tue Feb 25 07:58:52 2014
@@ -1,18 +1,24 @@
 PREHOOK: query: create table alter_rename_partition_src ( col1 string ) stored as textfile
 PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
 POSTHOOK: query: create table alter_rename_partition_src ( col1 string ) stored as textfile
 POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@alter_rename_partition_src
 PREHOOK: query: load data local inpath '../../data/files/test.dat' overwrite into table alter_rename_partition_src
 PREHOOK: type: LOAD
+#### A masked pattern was here ####
 PREHOOK: Output: default@alter_rename_partition_src
 POSTHOOK: query: load data local inpath '../../data/files/test.dat' overwrite into table alter_rename_partition_src
 POSTHOOK: type: LOAD
+#### A masked pattern was here ####
 POSTHOOK: Output: default@alter_rename_partition_src
 PREHOOK: query: create table alter_rename_partition ( col1 string ) partitioned by (pcol1 string , pcol2 string) stored as sequencefile
 PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
 POSTHOOK: query: create table alter_rename_partition ( col1 string ) partitioned by (pcol1 string , pcol2 string) stored as sequencefile
 POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@alter_rename_partition
 PREHOOK: query: insert overwrite table alter_rename_partition partition (pCol1='old_part1:', pcol2='old_part2:') select col1 from alter_rename_partition_src
 PREHOOK: type: QUERY

Modified: hive/branches/tez/ql/src/test/results/clientnegative/alter_rename_partition_failure2.q.out
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/results/clientnegative/alter_rename_partition_failure2.q.out?rev=1571600&r1=1571599&r2=1571600&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/results/clientnegative/alter_rename_partition_failure2.q.out (original)
+++ hive/branches/tez/ql/src/test/results/clientnegative/alter_rename_partition_failure2.q.out Tue Feb 25 07:58:52 2014
@@ -1,18 +1,24 @@
 PREHOOK: query: create table alter_rename_partition_src ( col1 string ) stored as textfile
 PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
 POSTHOOK: query: create table alter_rename_partition_src ( col1 string ) stored as textfile
 POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@alter_rename_partition_src
 PREHOOK: query: load data local inpath '../../data/files/test.dat' overwrite into table alter_rename_partition_src
 PREHOOK: type: LOAD
+#### A masked pattern was here ####
 PREHOOK: Output: default@alter_rename_partition_src
 POSTHOOK: query: load data local inpath '../../data/files/test.dat' overwrite into table alter_rename_partition_src
 POSTHOOK: type: LOAD
+#### A masked pattern was here ####
 POSTHOOK: Output: default@alter_rename_partition_src
 PREHOOK: query: create table alter_rename_partition ( col1 string ) partitioned by (pcol1 string , pcol2 string) stored as sequencefile
 PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
 POSTHOOK: query: create table alter_rename_partition ( col1 string ) partitioned by (pcol1 string , pcol2 string) stored as sequencefile
 POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@alter_rename_partition
 PREHOOK: query: insert overwrite table alter_rename_partition partition (pCol1='old_part1:', pcol2='old_part2:') select col1 from alter_rename_partition_src
 PREHOOK: type: QUERY

Modified: hive/branches/tez/ql/src/test/results/clientnegative/alter_rename_partition_failure3.q.out
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/results/clientnegative/alter_rename_partition_failure3.q.out?rev=1571600&r1=1571599&r2=1571600&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/results/clientnegative/alter_rename_partition_failure3.q.out (original)
+++ hive/branches/tez/ql/src/test/results/clientnegative/alter_rename_partition_failure3.q.out Tue Feb 25 07:58:52 2014
@@ -1,18 +1,24 @@
 PREHOOK: query: create table alter_rename_partition_src ( col1 string ) stored as textfile
 PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
 POSTHOOK: query: create table alter_rename_partition_src ( col1 string ) stored as textfile
 POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@alter_rename_partition_src
 PREHOOK: query: load data local inpath '../../data/files/test.dat' overwrite into table alter_rename_partition_src
 PREHOOK: type: LOAD
+#### A masked pattern was here ####
 PREHOOK: Output: default@alter_rename_partition_src
 POSTHOOK: query: load data local inpath '../../data/files/test.dat' overwrite into table alter_rename_partition_src
 POSTHOOK: type: LOAD
+#### A masked pattern was here ####
 POSTHOOK: Output: default@alter_rename_partition_src
 PREHOOK: query: create table alter_rename_partition ( col1 string ) partitioned by (pcol1 string , pcol2 string) stored as sequencefile
 PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
 POSTHOOK: query: create table alter_rename_partition ( col1 string ) partitioned by (pcol1 string , pcol2 string) stored as sequencefile
 POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@alter_rename_partition
 PREHOOK: query: insert overwrite table alter_rename_partition partition (pCol1='old_part1:', pcol2='old_part2:') select col1 from alter_rename_partition_src
 PREHOOK: type: QUERY

Modified: hive/branches/tez/ql/src/test/results/clientnegative/alter_table_add_partition.q.out
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/results/clientnegative/alter_table_add_partition.q.out?rev=1571600&r1=1571599&r2=1571600&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/results/clientnegative/alter_table_add_partition.q.out (original)
+++ hive/branches/tez/ql/src/test/results/clientnegative/alter_table_add_partition.q.out Tue Feb 25 07:58:52 2014
@@ -1,6 +1,8 @@
 PREHOOK: query: create table mp (a int) partitioned by (b int)
 PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
 POSTHOOK: query: create table mp (a int) partitioned by (b int)
 POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@mp
 FAILED: SemanticException Partition spec {b=1, c=1} contains non-partition columns

Modified: hive/branches/tez/ql/src/test/results/clientnegative/alter_table_wrong_regex.q.out
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/results/clientnegative/alter_table_wrong_regex.q.out?rev=1571600&r1=1571599&r2=1571600&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/results/clientnegative/alter_table_wrong_regex.q.out (original)
+++ hive/branches/tez/ql/src/test/results/clientnegative/alter_table_wrong_regex.q.out Tue Feb 25 07:58:52 2014
@@ -6,10 +6,12 @@ PREHOOK: query: create table aa ( test S
   ROW FORMAT SERDE 'org.apache.hadoop.hive.contrib.serde2.RegexSerDe'
   WITH SERDEPROPERTIES ("input.regex" = "(.*)", "output.format.string" = "$1s")
 PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
 POSTHOOK: query: create table aa ( test STRING )
   ROW FORMAT SERDE 'org.apache.hadoop.hive.contrib.serde2.RegexSerDe'
   WITH SERDEPROPERTIES ("input.regex" = "(.*)", "output.format.string" = "$1s")
 POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@aa
 PREHOOK: query: alter table aa set serdeproperties ("input.regex" = "[^\\](.*)", "output.format.string" = "$1s")
 PREHOOK: type: ALTERTABLE_SERDEPROPERTIES

Modified: hive/branches/tez/ql/src/test/results/clientnegative/alter_view_as_select_with_partition.q.out
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/results/clientnegative/alter_view_as_select_with_partition.q.out?rev=1571600&r1=1571599&r2=1571600&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/results/clientnegative/alter_view_as_select_with_partition.q.out (original)
+++ hive/branches/tez/ql/src/test/results/clientnegative/alter_view_as_select_with_partition.q.out Tue Feb 25 07:58:52 2014
@@ -18,11 +18,13 @@ ADD PARTITION (value='val_86') PARTITION
 PREHOOK: type: ALTERTABLE_ADDPARTS
 PREHOOK: Input: default@src
 PREHOOK: Input: default@testviewpart
+PREHOOK: Output: default@testviewpart
 POSTHOOK: query: ALTER VIEW testViewPart 
 ADD PARTITION (value='val_86') PARTITION (value='val_xyz')
 POSTHOOK: type: ALTERTABLE_ADDPARTS
 POSTHOOK: Input: default@src
 POSTHOOK: Input: default@testviewpart
+POSTHOOK: Output: default@testviewpart
 POSTHOOK: Output: default@testviewpart@value=val_86
 POSTHOOK: Output: default@testviewpart@value=val_xyz
 PREHOOK: query: DESCRIBE FORMATTED testViewPart

Modified: hive/branches/tez/ql/src/test/results/clientnegative/alter_view_failure8.q.out
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/results/clientnegative/alter_view_failure8.q.out?rev=1571600&r1=1571599&r2=1571600&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/results/clientnegative/alter_view_failure8.q.out (original)
+++ hive/branches/tez/ql/src/test/results/clientnegative/alter_view_failure8.q.out Tue Feb 25 07:58:52 2014
@@ -1,8 +1,10 @@
 PREHOOK: query: -- should fail:  can't use ALTER VIEW on a table
 CREATE TABLE invites (foo INT, bar STRING) PARTITIONED BY (ds STRING)
 PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
 POSTHOOK: query: -- should fail:  can't use ALTER VIEW on a table
 CREATE TABLE invites (foo INT, bar STRING) PARTITIONED BY (ds STRING)
 POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@invites
 FAILED: SemanticException [Error 10132]: To alter a base table you need to use the ALTER TABLE command.

Modified: hive/branches/tez/ql/src/test/results/clientnegative/altern1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/results/clientnegative/altern1.q.out?rev=1571600&r1=1571599&r2=1571600&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/results/clientnegative/altern1.q.out (original)
+++ hive/branches/tez/ql/src/test/results/clientnegative/altern1.q.out Tue Feb 25 07:58:52 2014
@@ -1,7 +1,9 @@
 PREHOOK: query: create table altern1(a int, b int) partitioned by (ds string)
 PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
 POSTHOOK: query: create table altern1(a int, b int) partitioned by (ds string)
 POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@altern1
 PREHOOK: query: alter table altern1 replace columns(a int, b int, ds string)
 PREHOOK: type: ALTERTABLE_REPLACECOLS

Modified: hive/branches/tez/ql/src/test/results/clientnegative/archive1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/results/clientnegative/archive1.q.out?rev=1571600&r1=1571599&r2=1571600&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/results/clientnegative/archive1.q.out (original)
+++ hive/branches/tez/ql/src/test/results/clientnegative/archive1.q.out Tue Feb 25 07:58:52 2014
@@ -3,11 +3,13 @@ PREHOOK: query: -- Tests trying to archi
 
 CREATE TABLE srcpart_archived LIKE srcpart
 PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
 POSTHOOK: query: -- Tests trying to archive a partition twice.
 -- EXCLUDE_HADOOP_MAJOR_VERSIONS(0.17, 0.18, 0.19)
 
 CREATE TABLE srcpart_archived LIKE srcpart
 POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@srcpart_archived
 PREHOOK: query: INSERT OVERWRITE TABLE srcpart_archived PARTITION (ds='2008-04-08', hr='12')
 SELECT key, value FROM srcpart WHERE ds='2008-04-08' AND hr='12'

Modified: hive/branches/tez/ql/src/test/results/clientnegative/archive2.q.out
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/results/clientnegative/archive2.q.out?rev=1571600&r1=1571599&r2=1571600&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/results/clientnegative/archive2.q.out (original)
+++ hive/branches/tez/ql/src/test/results/clientnegative/archive2.q.out Tue Feb 25 07:58:52 2014
@@ -10,8 +10,10 @@ drop table tstsrcpart
 POSTHOOK: type: DROPTABLE
 PREHOOK: query: create table tstsrcpart like srcpart
 PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
 POSTHOOK: query: create table tstsrcpart like srcpart
 POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@tstsrcpart
 PREHOOK: query: insert overwrite table tstsrcpart partition (ds='2008-04-08', hr='12')
 select key, value from srcpart where ds='2008-04-08' and hr='12'

Modified: hive/branches/tez/ql/src/test/results/clientnegative/archive_corrupt.q.out
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/results/clientnegative/archive_corrupt.q.out?rev=1571600&r1=1571599&r2=1571600&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/results/clientnegative/archive_corrupt.q.out (original)
+++ hive/branches/tez/ql/src/test/results/clientnegative/archive_corrupt.q.out Tue Feb 25 07:58:52 2014
@@ -8,8 +8,10 @@ POSTHOOK: query: drop table tstsrcpart
 POSTHOOK: type: DROPTABLE
 PREHOOK: query: create table tstsrcpart like srcpart
 PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
 POSTHOOK: query: create table tstsrcpart like srcpart
 POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@tstsrcpart
 PREHOOK: query: -- EXCLUDE_HADOOP_MAJOR_VERSIONS(0.20)
 -- The version of GzipCodec that is provided in Hadoop 0.20 silently ignores
@@ -20,6 +22,7 @@ PREHOOK: query: -- EXCLUDE_HADOOP_MAJOR_
 
 load data local inpath '../../data/files/archive_corrupt.rc' overwrite into table tstsrcpart partition (ds='2008-04-08', hr='11')
 PREHOOK: type: LOAD
+#### A masked pattern was here ####
 PREHOOK: Output: default@tstsrcpart
 Failed with exception Wrong file format. Please check the file's format.
 FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.MoveTask

Modified: hive/branches/tez/ql/src/test/results/clientnegative/archive_insert1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/results/clientnegative/archive_insert1.q.out?rev=1571600&r1=1571599&r2=1571600&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/results/clientnegative/archive_insert1.q.out (original)
+++ hive/branches/tez/ql/src/test/results/clientnegative/archive_insert1.q.out Tue Feb 25 07:58:52 2014
@@ -3,11 +3,13 @@ PREHOOK: query: -- Tests trying to inser
 
 CREATE TABLE tstsrcpart LIKE srcpart
 PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
 POSTHOOK: query: -- Tests trying to insert into archived partition.
 -- EXCLUDE_HADOOP_MAJOR_VERSIONS(0.17, 0.18, 0.19)
 
 CREATE TABLE tstsrcpart LIKE srcpart
 POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@tstsrcpart
 PREHOOK: query: INSERT OVERWRITE TABLE tstsrcpart PARTITION (ds='2008-04-08', hr='12')
 SELECT key, value FROM srcpart WHERE ds='2008-04-08' AND hr='12'

Modified: hive/branches/tez/ql/src/test/results/clientnegative/archive_insert2.q.out
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/results/clientnegative/archive_insert2.q.out?rev=1571600&r1=1571599&r2=1571600&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/results/clientnegative/archive_insert2.q.out (original)
+++ hive/branches/tez/ql/src/test/results/clientnegative/archive_insert2.q.out Tue Feb 25 07:58:52 2014
@@ -3,11 +3,13 @@ PREHOOK: query: -- Tests trying to inser
 
 CREATE TABLE tstsrcpart LIKE srcpart
 PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
 POSTHOOK: query: -- Tests trying to insert into archived partition.
 -- EXCLUDE_HADOOP_MAJOR_VERSIONS(0.17, 0.18, 0.19)
 
 CREATE TABLE tstsrcpart LIKE srcpart
 POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@tstsrcpart
 PREHOOK: query: INSERT OVERWRITE TABLE tstsrcpart PARTITION (ds='2008-04-08', hr='12')
 SELECT key, value FROM srcpart WHERE ds='2008-04-08' AND hr='12'

Modified: hive/branches/tez/ql/src/test/results/clientnegative/archive_insert3.q.out
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/results/clientnegative/archive_insert3.q.out?rev=1571600&r1=1571599&r2=1571600&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/results/clientnegative/archive_insert3.q.out (original)
+++ hive/branches/tez/ql/src/test/results/clientnegative/archive_insert3.q.out Tue Feb 25 07:58:52 2014
@@ -3,11 +3,13 @@ PREHOOK: query: -- Tests trying to creat
 
 CREATE TABLE tstsrcpart LIKE srcpart
 PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
 POSTHOOK: query: -- Tests trying to create partition inside of archived directory.
 -- EXCLUDE_HADOOP_MAJOR_VERSIONS(0.17, 0.18, 0.19)
 
 CREATE TABLE tstsrcpart LIKE srcpart
 POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@tstsrcpart
 PREHOOK: query: INSERT OVERWRITE TABLE tstsrcpart PARTITION (ds='2008-04-08', hr='12')
 SELECT key, value FROM srcpart WHERE ds='2008-04-08' AND hr='12'

Modified: hive/branches/tez/ql/src/test/results/clientnegative/archive_insert4.q.out
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/results/clientnegative/archive_insert4.q.out?rev=1571600&r1=1571599&r2=1571600&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/results/clientnegative/archive_insert4.q.out (original)
+++ hive/branches/tez/ql/src/test/results/clientnegative/archive_insert4.q.out Tue Feb 25 07:58:52 2014
@@ -3,11 +3,13 @@ PREHOOK: query: -- Tests trying to (poss
 
 CREATE TABLE tstsrcpart LIKE srcpart
 PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
 POSTHOOK: query: -- Tests trying to (possible) dynamic insert into archived partition.
 -- EXCLUDE_HADOOP_MAJOR_VERSIONS(0.17, 0.18, 0.19)
 
 CREATE TABLE tstsrcpart LIKE srcpart
 POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@tstsrcpart
 PREHOOK: query: INSERT OVERWRITE TABLE tstsrcpart PARTITION (ds='2008-04-08', hr='12')
 SELECT key, value FROM srcpart WHERE ds='2008-04-08' AND hr='12'

Modified: hive/branches/tez/ql/src/test/results/clientnegative/archive_multi1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/results/clientnegative/archive_multi1.q.out?rev=1571600&r1=1571599&r2=1571600&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/results/clientnegative/archive_multi1.q.out (original)
+++ hive/branches/tez/ql/src/test/results/clientnegative/archive_multi1.q.out Tue Feb 25 07:58:52 2014
@@ -3,11 +3,13 @@ PREHOOK: query: -- Tests trying to archi
 
 CREATE TABLE tstsrcpart LIKE srcpart
 PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
 POSTHOOK: query: -- Tests trying to archive a partition twice.
 -- EXCLUDE_HADOOP_MAJOR_VERSIONS(0.17, 0.18, 0.19)
 
 CREATE TABLE tstsrcpart LIKE srcpart
 POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@tstsrcpart
 PREHOOK: query: INSERT OVERWRITE TABLE tstsrcpart PARTITION (ds='2008-04-08', hr='11')
 SELECT key, value FROM srcpart WHERE ds='2008-04-08' AND hr='11'

Modified: hive/branches/tez/ql/src/test/results/clientnegative/archive_multi2.q.out
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/results/clientnegative/archive_multi2.q.out?rev=1571600&r1=1571599&r2=1571600&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/results/clientnegative/archive_multi2.q.out (original)
+++ hive/branches/tez/ql/src/test/results/clientnegative/archive_multi2.q.out Tue Feb 25 07:58:52 2014
@@ -10,8 +10,10 @@ drop table tstsrcpart
 POSTHOOK: type: DROPTABLE
 PREHOOK: query: create table tstsrcpart like srcpart
 PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
 POSTHOOK: query: create table tstsrcpart like srcpart
 POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@tstsrcpart
 PREHOOK: query: INSERT OVERWRITE TABLE tstsrcpart PARTITION (ds='2008-04-08', hr='11')
 SELECT key, value FROM srcpart WHERE ds='2008-04-08' AND hr='11'

Modified: hive/branches/tez/ql/src/test/results/clientnegative/archive_multi3.q.out
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/results/clientnegative/archive_multi3.q.out?rev=1571600&r1=1571599&r2=1571600&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/results/clientnegative/archive_multi3.q.out (original)
+++ hive/branches/tez/ql/src/test/results/clientnegative/archive_multi3.q.out Tue Feb 25 07:58:52 2014
@@ -3,11 +3,13 @@ PREHOOK: query: -- Tests trying to archi
 
 CREATE TABLE tstsrcpart LIKE srcpart
 PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
 POSTHOOK: query: -- Tests trying to archive outer partition group containing other partition inside.
 -- EXCLUDE_HADOOP_MAJOR_VERSIONS(0.17, 0.18, 0.19)
 
 CREATE TABLE tstsrcpart LIKE srcpart
 POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@tstsrcpart
 PREHOOK: query: INSERT OVERWRITE TABLE tstsrcpart PARTITION (ds='2008-04-08', hr='11')
 SELECT key, value FROM srcpart WHERE ds='2008-04-08' AND hr='11'

Modified: hive/branches/tez/ql/src/test/results/clientnegative/archive_multi4.q.out
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/results/clientnegative/archive_multi4.q.out?rev=1571600&r1=1571599&r2=1571600&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/results/clientnegative/archive_multi4.q.out (original)
+++ hive/branches/tez/ql/src/test/results/clientnegative/archive_multi4.q.out Tue Feb 25 07:58:52 2014
@@ -3,11 +3,13 @@ PREHOOK: query: -- Tests trying to archi
 
 CREATE TABLE tstsrcpart LIKE srcpart
 PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
 POSTHOOK: query: -- Tests trying to archive inner partition contained in archived partition group.
 -- EXCLUDE_HADOOP_MAJOR_VERSIONS(0.17, 0.18, 0.19)
 
 CREATE TABLE tstsrcpart LIKE srcpart
 POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@tstsrcpart
 PREHOOK: query: INSERT OVERWRITE TABLE tstsrcpart PARTITION (ds='2008-04-08', hr='11')
 SELECT key, value FROM srcpart WHERE ds='2008-04-08' AND hr='11'

Modified: hive/branches/tez/ql/src/test/results/clientnegative/archive_multi5.q.out
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/results/clientnegative/archive_multi5.q.out?rev=1571600&r1=1571599&r2=1571600&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/results/clientnegative/archive_multi5.q.out (original)
+++ hive/branches/tez/ql/src/test/results/clientnegative/archive_multi5.q.out Tue Feb 25 07:58:52 2014
@@ -3,11 +3,13 @@ PREHOOK: query: -- Tests trying to unarc
 
 CREATE TABLE tstsrcpart LIKE srcpart
 PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
 POSTHOOK: query: -- Tests trying to unarchive outer partition group containing other partition inside.
 -- EXCLUDE_HADOOP_MAJOR_VERSIONS(0.17, 0.18, 0.19)
 
 CREATE TABLE tstsrcpart LIKE srcpart
 POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@tstsrcpart
 PREHOOK: query: INSERT OVERWRITE TABLE tstsrcpart PARTITION (ds='2008-04-08', hr='11')
 SELECT key, value FROM srcpart WHERE ds='2008-04-08' AND hr='11'

Modified: hive/branches/tez/ql/src/test/results/clientnegative/archive_multi6.q.out
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/results/clientnegative/archive_multi6.q.out?rev=1571600&r1=1571599&r2=1571600&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/results/clientnegative/archive_multi6.q.out (original)
+++ hive/branches/tez/ql/src/test/results/clientnegative/archive_multi6.q.out Tue Feb 25 07:58:52 2014
@@ -3,11 +3,13 @@ PREHOOK: query: -- Tests trying to unarc
 
 CREATE TABLE tstsrcpart LIKE srcpart
 PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
 POSTHOOK: query: -- Tests trying to unarchive inner partition contained in archived partition group.
 -- EXCLUDE_HADOOP_MAJOR_VERSIONS(0.17, 0.18, 0.19)
 
 CREATE TABLE tstsrcpart LIKE srcpart
 POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@tstsrcpart
 PREHOOK: query: INSERT OVERWRITE TABLE tstsrcpart PARTITION (ds='2008-04-08', hr='11')
 SELECT key, value FROM srcpart WHERE ds='2008-04-08' AND hr='11'

Modified: hive/branches/tez/ql/src/test/results/clientnegative/archive_multi7.q.out
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/results/clientnegative/archive_multi7.q.out?rev=1571600&r1=1571599&r2=1571600&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/results/clientnegative/archive_multi7.q.out (original)
+++ hive/branches/tez/ql/src/test/results/clientnegative/archive_multi7.q.out Tue Feb 25 07:58:52 2014
@@ -3,11 +3,13 @@ PREHOOK: query: -- Tests trying to archi
 
 CREATE TABLE tstsrcpart LIKE srcpart
 PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
 POSTHOOK: query: -- Tests trying to archive a partition group with custom locations.
 -- EXCLUDE_HADOOP_MAJOR_VERSIONS(0.17, 0.18, 0.19)
 
 CREATE TABLE tstsrcpart LIKE srcpart
 POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@tstsrcpart
 PREHOOK: query: INSERT OVERWRITE TABLE tstsrcpart PARTITION (ds='2008-04-08', hr='11')
 SELECT key, value FROM srcpart WHERE ds='2008-04-08' AND hr='11'
@@ -26,11 +28,13 @@ POSTHOOK: Lineage: tstsrcpart PARTITION(
 PREHOOK: query: ALTER TABLE tstsrcpart ADD PARTITION (ds='2008-04-08', hr='12')
 #### A masked pattern was here ####
 PREHOOK: type: ALTERTABLE_ADDPARTS
-PREHOOK: Input: default@tstsrcpart
+#### A masked pattern was here ####
+PREHOOK: Output: default@tstsrcpart
 POSTHOOK: query: ALTER TABLE tstsrcpart ADD PARTITION (ds='2008-04-08', hr='12')
 #### A masked pattern was here ####
 POSTHOOK: type: ALTERTABLE_ADDPARTS
-POSTHOOK: Input: default@tstsrcpart
+#### A masked pattern was here ####
+POSTHOOK: Output: default@tstsrcpart
 POSTHOOK: Output: default@tstsrcpart@ds=2008-04-08/hr=12
 POSTHOOK: Lineage: tstsrcpart PARTITION(ds=2008-04-08,hr=11).key SIMPLE [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ]
 POSTHOOK: Lineage: tstsrcpart PARTITION(ds=2008-04-08,hr=11).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ]

Modified: hive/branches/tez/ql/src/test/results/clientnegative/archive_partspec1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/results/clientnegative/archive_partspec1.q.out?rev=1571600&r1=1571599&r2=1571600&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/results/clientnegative/archive_partspec1.q.out (original)
+++ hive/branches/tez/ql/src/test/results/clientnegative/archive_partspec1.q.out Tue Feb 25 07:58:52 2014
@@ -3,11 +3,13 @@ PREHOOK: query: -- Tests trying to archi
 
 CREATE TABLE srcpart_archived LIKE srcpart
 PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
 POSTHOOK: query: -- Tests trying to archive a partition twice.
 -- EXCLUDE_HADOOP_MAJOR_VERSIONS(0.17, 0.18, 0.19)
 
 CREATE TABLE srcpart_archived LIKE srcpart
 POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@srcpart_archived
 PREHOOK: query: INSERT OVERWRITE TABLE srcpart_archived PARTITION (ds='2008-04-08', hr='12')
 SELECT key, value FROM srcpart WHERE ds='2008-04-08' AND hr='12'

Modified: hive/branches/tez/ql/src/test/results/clientnegative/archive_partspec2.q.out
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/results/clientnegative/archive_partspec2.q.out?rev=1571600&r1=1571599&r2=1571600&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/results/clientnegative/archive_partspec2.q.out (original)
+++ hive/branches/tez/ql/src/test/results/clientnegative/archive_partspec2.q.out Tue Feb 25 07:58:52 2014
@@ -3,11 +3,13 @@ PREHOOK: query: -- Tests trying to archi
 
 CREATE TABLE srcpart_archived LIKE srcpart
 PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
 POSTHOOK: query: -- Tests trying to archive a partition twice.
 -- EXCLUDE_HADOOP_MAJOR_VERSIONS(0.17, 0.18, 0.19)
 
 CREATE TABLE srcpart_archived LIKE srcpart
 POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@srcpart_archived
 PREHOOK: query: INSERT OVERWRITE TABLE srcpart_archived PARTITION (ds='2008-04-08', hr='12')
 SELECT key, value FROM srcpart WHERE ds='2008-04-08' AND hr='12'



Mime
View raw message