hadoop-common-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From a..@apache.org
Subject svn commit: r1148933 - in /hadoop/common/trunk/common: ./ src/test/core/org/apache/hadoop/io/file/tfile/
Date Wed, 20 Jul 2011 20:42:16 GMT
Author: atm
Date: Wed Jul 20 20:42:15 2011
New Revision: 1148933

URL: http://svn.apache.org/viewvc?rev=1148933&view=rev
Log:
HADOOP-7111. Several TFile tests failing when native libraries are present. (atm)

Modified:
    hadoop/common/trunk/common/CHANGES.txt
    hadoop/common/trunk/common/src/test/core/org/apache/hadoop/io/file/tfile/TestTFileByteArrays.java
    hadoop/common/trunk/common/src/test/core/org/apache/hadoop/io/file/tfile/TestTFileJClassComparatorByteArrays.java
    hadoop/common/trunk/common/src/test/core/org/apache/hadoop/io/file/tfile/TestTFileStreams.java

Modified: hadoop/common/trunk/common/CHANGES.txt
URL: http://svn.apache.org/viewvc/hadoop/common/trunk/common/CHANGES.txt?rev=1148933&r1=1148932&r2=1148933&view=diff
==============================================================================
--- hadoop/common/trunk/common/CHANGES.txt (original)
+++ hadoop/common/trunk/common/CHANGES.txt Wed Jul 20 20:42:15 2011
@@ -405,6 +405,9 @@ Trunk (unreleased changes)
     HADOOP-2081. Configuration getInt, getLong, and getFloat replace
     invalid numbers with the default value. (Harsh J via eli)
 
+    HADOOP-7111. Several TFile tests failing when native libraries are
+    present. (atm)
+
 Release 0.22.0 - Unreleased
 
   INCOMPATIBLE CHANGES

Modified: hadoop/common/trunk/common/src/test/core/org/apache/hadoop/io/file/tfile/TestTFileByteArrays.java
URL: http://svn.apache.org/viewvc/hadoop/common/trunk/common/src/test/core/org/apache/hadoop/io/file/tfile/TestTFileByteArrays.java?rev=1148933&r1=1148932&r2=1148933&view=diff
==============================================================================
--- hadoop/common/trunk/common/src/test/core/org/apache/hadoop/io/file/tfile/TestTFileByteArrays.java
(original)
+++ hadoop/common/trunk/common/src/test/core/org/apache/hadoop/io/file/tfile/TestTFileByteArrays.java
Wed Jul 20 20:42:15 2011
@@ -24,7 +24,6 @@ import java.io.IOException;
 import java.util.Random;
 
 import junit.framework.Assert;
-import junit.framework.TestCase;
 
 import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.fs.FSDataOutputStream;
@@ -35,6 +34,10 @@ import org.apache.hadoop.io.file.tfile.T
 import org.apache.hadoop.io.file.tfile.TFile.Writer;
 import org.apache.hadoop.io.file.tfile.TFile.Reader.Location;
 import org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner;
+import org.apache.hadoop.util.NativeCodeLoader;
+import org.junit.After;
+import org.junit.Before;
+import org.junit.Test;
 
 /**
  * 
@@ -42,7 +45,7 @@ import org.apache.hadoop.io.file.tfile.T
  * and LZO compression classes.
  * 
  */
-public class TestTFileByteArrays extends TestCase {
+public class TestTFileByteArrays {
   private static String ROOT =
       System.getProperty("test.build.data", "/tmp/tfile-test");
   private final static int BLOCK_SIZE = 512;
@@ -62,25 +65,29 @@ public class TestTFileByteArrays extends
   private String compression = Compression.Algorithm.GZ.getName();
   private String comparator = "memcmp";
   private String outputFile = "TFileTestByteArrays";
+
   /*
    * pre-sampled numbers of records in one block, based on the given the
-   * generated key and value strings
+   * generated key and value strings. This is slightly different based on
+   * whether or not the native libs are present.
    */
-  // private int records1stBlock = 4314;
-  // private int records2ndBlock = 4108;
-  private int records1stBlock = 4480;
-  private int records2ndBlock = 4263;
+  private int records1stBlock = NativeCodeLoader.isNativeCodeLoaded() ? 5674 : 4480;
+  private int records2ndBlock = NativeCodeLoader.isNativeCodeLoaded() ? 5574 : 4263;
 
   public void init(String compression, String comparator, String outputFile,
       int numRecords1stBlock, int numRecords2ndBlock) {
+    init(compression, comparator, outputFile);
+    this.records1stBlock = numRecords1stBlock;
+    this.records2ndBlock = numRecords2ndBlock;
+  }
+  
+  public void init(String compression, String comparator, String outputFile) {
     this.compression = compression;
     this.comparator = comparator;
     this.outputFile = outputFile;
-    this.records1stBlock = numRecords1stBlock;
-    this.records2ndBlock = numRecords2ndBlock;
   }
 
-  @Override
+  @Before
   public void setUp() throws IOException {
     conf = new Configuration();
     path = new Path(ROOT, outputFile);
@@ -89,12 +96,13 @@ public class TestTFileByteArrays extends
     writer = new Writer(out, BLOCK_SIZE, compression, comparator, conf);
   }
 
-  @Override
+  @After
   public void tearDown() throws IOException {
     if (!skip)
     fs.delete(path, true);
   }
 
+  @Test
   public void testNoDataEntry() throws IOException {
     if (skip) 
       return;
@@ -108,19 +116,21 @@ public class TestTFileByteArrays extends
     reader.close();
   }
 
+  @Test
   public void testOneDataEntry() throws IOException {
     if (skip)
       return;
     writeRecords(1);
     readRecords(1);
 
-    checkBlockIndex(1, 0, 0);
-    readValueBeforeKey(1, 0);
-    readKeyWithoutValue(1, 0);
-    readValueWithoutKey(1, 0);
-    readKeyManyTimes(1, 0);
+    checkBlockIndex(0, 0);
+    readValueBeforeKey(0);
+    readKeyWithoutValue(0);
+    readValueWithoutKey(0);
+    readKeyManyTimes(0);
   }
 
+  @Test
   public void testTwoDataEntries() throws IOException {
     if (skip)
       return;
@@ -133,6 +143,7 @@ public class TestTFileByteArrays extends
    * 
    * @throws IOException
    */
+  @Test
   public void testOneBlock() throws IOException {
     if (skip)
       return;
@@ -140,7 +151,7 @@ public class TestTFileByteArrays extends
     writeRecords(records1stBlock);
     readRecords(records1stBlock);
     // last key should be in the first block (block 0)
-    checkBlockIndex(records1stBlock, records1stBlock - 1, 0);
+    checkBlockIndex(records1stBlock - 1, 0);
   }
 
   /**
@@ -148,68 +159,70 @@ public class TestTFileByteArrays extends
    * 
    * @throws IOException
    */
+  @Test
   public void testOneBlockPlusOneEntry() throws IOException {
     if (skip)
       return;
     writeRecords(records1stBlock + 1);
     readRecords(records1stBlock + 1);
-    checkBlockIndex(records1stBlock + 1, records1stBlock - 1, 0);
-    checkBlockIndex(records1stBlock + 1, records1stBlock, 1);
+    checkBlockIndex(records1stBlock - 1, 0);
+    checkBlockIndex(records1stBlock, 1);
   }
 
+  @Test
   public void testTwoBlocks() throws IOException {
     if (skip)
       return;
     writeRecords(records1stBlock + 5);
     readRecords(records1stBlock + 5);
-    checkBlockIndex(records1stBlock + 5, records1stBlock + 4, 1);
+    checkBlockIndex(records1stBlock + 4, 1);
   }
 
+  @Test
   public void testThreeBlocks() throws IOException {
     if (skip) 
       return;
     writeRecords(2 * records1stBlock + 5);
     readRecords(2 * records1stBlock + 5);
 
-    checkBlockIndex(2 * records1stBlock + 5, 2 * records1stBlock + 4, 2);
+    checkBlockIndex(2 * records1stBlock + 4, 2);
     // 1st key in file
-    readValueBeforeKey(2 * records1stBlock + 5, 0);
-    readKeyWithoutValue(2 * records1stBlock + 5, 0);
-    readValueWithoutKey(2 * records1stBlock + 5, 0);
-    readKeyManyTimes(2 * records1stBlock + 5, 0);
+    readValueBeforeKey(0);
+    readKeyWithoutValue(0);
+    readValueWithoutKey(0);
+    readKeyManyTimes(0);
     // last key in file
-    readValueBeforeKey(2 * records1stBlock + 5, 2 * records1stBlock + 4);
-    readKeyWithoutValue(2 * records1stBlock + 5, 2 * records1stBlock + 4);
-    readValueWithoutKey(2 * records1stBlock + 5, 2 * records1stBlock + 4);
-    readKeyManyTimes(2 * records1stBlock + 5, 2 * records1stBlock + 4);
+    readValueBeforeKey(2 * records1stBlock + 4);
+    readKeyWithoutValue(2 * records1stBlock + 4);
+    readValueWithoutKey(2 * records1stBlock + 4);
+    readKeyManyTimes(2 * records1stBlock + 4);
 
     // 1st key in mid block, verify block indexes then read
-    checkBlockIndex(2 * records1stBlock + 5, records1stBlock - 1, 0);
-    checkBlockIndex(2 * records1stBlock + 5, records1stBlock, 1);
-    readValueBeforeKey(2 * records1stBlock + 5, records1stBlock);
-    readKeyWithoutValue(2 * records1stBlock + 5, records1stBlock);
-    readValueWithoutKey(2 * records1stBlock + 5, records1stBlock);
-    readKeyManyTimes(2 * records1stBlock + 5, records1stBlock);
+    checkBlockIndex(records1stBlock - 1, 0);
+    checkBlockIndex(records1stBlock, 1);
+    readValueBeforeKey(records1stBlock);
+    readKeyWithoutValue(records1stBlock);
+    readValueWithoutKey(records1stBlock);
+    readKeyManyTimes(records1stBlock);
 
     // last key in mid block, verify block indexes then read
-    checkBlockIndex(2 * records1stBlock + 5, records1stBlock + records2ndBlock
+    checkBlockIndex(records1stBlock + records2ndBlock
         - 1, 1);
-    checkBlockIndex(2 * records1stBlock + 5, records1stBlock + records2ndBlock,
-        2);
-    readValueBeforeKey(2 * records1stBlock + 5, records1stBlock
+    checkBlockIndex(records1stBlock + records2ndBlock, 2);
+    readValueBeforeKey(records1stBlock
         + records2ndBlock - 1);
-    readKeyWithoutValue(2 * records1stBlock + 5, records1stBlock
+    readKeyWithoutValue(records1stBlock
         + records2ndBlock - 1);
-    readValueWithoutKey(2 * records1stBlock + 5, records1stBlock
+    readValueWithoutKey(records1stBlock
         + records2ndBlock - 1);
-    readKeyManyTimes(2 * records1stBlock + 5, records1stBlock + records2ndBlock
+    readKeyManyTimes(records1stBlock + records2ndBlock
         - 1);
 
     // mid in mid block
-    readValueBeforeKey(2 * records1stBlock + 5, records1stBlock + 10);
-    readKeyWithoutValue(2 * records1stBlock + 5, records1stBlock + 10);
-    readValueWithoutKey(2 * records1stBlock + 5, records1stBlock + 10);
-    readKeyManyTimes(2 * records1stBlock + 5, records1stBlock + 10);
+    readValueBeforeKey(records1stBlock + 10);
+    readKeyWithoutValue(records1stBlock + 10);
+    readValueWithoutKey(records1stBlock + 10);
+    readKeyManyTimes(records1stBlock + 10);
   }
 
   Location locate(Scanner scanner, byte[] key) throws IOException {
@@ -219,27 +232,23 @@ public class TestTFileByteArrays extends
     return scanner.endLocation;
   }
   
+  @Test
   public void testLocate() throws IOException {
     if (skip)
       return;
     writeRecords(3 * records1stBlock);
     Reader reader = new Reader(fs.open(path), fs.getFileStatus(path).getLen(), conf);
     Scanner scanner = reader.createScanner();
-    Location loc2 =
-        locate(scanner, composeSortedKey(KEY, 3 * records1stBlock, 2)
-            .getBytes());
-    Location locLastIn1stBlock =
-        locate(scanner, composeSortedKey(KEY, 3 * records1stBlock,
-            records1stBlock - 1).getBytes());
-    Location locFirstIn2ndBlock =
-        locate(scanner, composeSortedKey(KEY, 3 * records1stBlock,
-            records1stBlock).getBytes());
+    locate(scanner, composeSortedKey(KEY, 2).getBytes());
+    locate(scanner, composeSortedKey(KEY, records1stBlock - 1).getBytes());
+    locate(scanner, composeSortedKey(KEY, records1stBlock).getBytes());
     Location locX = locate(scanner, "keyX".getBytes());
     Assert.assertEquals(scanner.endLocation, locX);
     scanner.close();
     reader.close();
   }
 
+  @Test
   public void testFailureWriterNotClosed() throws IOException {
     if (skip)
       return;
@@ -247,17 +256,16 @@ public class TestTFileByteArrays extends
     try {
       reader = new Reader(fs.open(path), fs.getFileStatus(path).getLen(), conf);
       Assert.fail("Cannot read before closing the writer.");
-    }
-    catch (IOException e) {
+    } catch (IOException e) {
       // noop, expecting exceptions
-    }
-    finally {
+    } finally {
       if (reader != null) {
         reader.close();
       }
     }
   }
 
+  @Test
   public void testFailureWriteMetaBlocksWithSameName() throws IOException {
     if (skip)
       return;
@@ -271,16 +279,15 @@ public class TestTFileByteArrays extends
     outMeta.close();
     // add the same metablock
     try {
-      DataOutputStream outMeta2 =
-          writer.prepareMetaBlock("testX", Compression.Algorithm.GZ.getName());
+      writer.prepareMetaBlock("testX", Compression.Algorithm.GZ.getName());
       Assert.fail("Cannot create metablocks with the same name.");
-    }
-    catch (Exception e) {
+    } catch (Exception e) {
       // noop, expecting exceptions
     }
     closeOutput();
   }
 
+  @Test
   public void testFailureGetNonExistentMetaBlock() throws IOException {
     if (skip)
       return;
@@ -300,15 +307,14 @@ public class TestTFileByteArrays extends
     mb.close();
     try {
       DataInputStream mbBad = reader.getMetaBlock("testY");
-      Assert.assertNull(mbBad);
       Assert.fail("Error on handling non-existent metablocks.");
-    }
-    catch (Exception e) {
+    } catch (Exception e) {
       // noop, expecting exceptions
     }
     reader.close();
   }
 
+  @Test
   public void testFailureWriteRecordAfterMetaBlock() throws IOException {
     if (skip)
       return;
@@ -324,13 +330,13 @@ public class TestTFileByteArrays extends
     try {
       writer.append("keyY".getBytes(), "valueY".getBytes());
       Assert.fail("Cannot add key/value after start adding meta blocks.");
-    }
-    catch (Exception e) {
+    } catch (Exception e) {
       // noop, expecting exceptions
     }
     closeOutput();
   }
 
+  @Test
   public void testFailureReadValueManyTimes() throws IOException {
     if (skip)
       return;
@@ -346,8 +352,7 @@ public class TestTFileByteArrays extends
     try {
       scanner.entry().getValue(vbuf);
       Assert.fail("Cannot get the value mlutiple times.");
-    }
-    catch (Exception e) {
+    } catch (Exception e) {
       // noop, expecting exceptions
     }
 
@@ -355,6 +360,7 @@ public class TestTFileByteArrays extends
     reader.close();
   }
 
+  @Test
   public void testFailureBadCompressionCodec() throws IOException {
     if (skip)
       return;
@@ -363,13 +369,13 @@ public class TestTFileByteArrays extends
     try {
       writer = new Writer(out, BLOCK_SIZE, "BAD", comparator, conf);
       Assert.fail("Error on handling invalid compression codecs.");
-    }
-    catch (Exception e) {
+    } catch (Exception e) {
       // noop, expecting exceptions
       // e.printStackTrace();
     }
   }
 
+  @Test
   public void testFailureOpenEmptyFile() throws IOException {
     if (skip)
       return;
@@ -379,15 +385,14 @@ public class TestTFileByteArrays extends
     out = fs.create(path);
     out.close();
     try {
-      Reader reader =
-          new Reader(fs.open(path), fs.getFileStatus(path).getLen(), conf);
+      new Reader(fs.open(path), fs.getFileStatus(path).getLen(), conf);
       Assert.fail("Error on handling empty files.");
-    }
-    catch (EOFException e) {
+    } catch (EOFException e) {
       // noop, expecting exceptions
     }
   }
 
+  @Test
   public void testFailureOpenRandomFile() throws IOException {
     if (skip)
       return;
@@ -404,15 +409,14 @@ public class TestTFileByteArrays extends
     }
     out.close();
     try {
-      Reader reader =
-          new Reader(fs.open(path), fs.getFileStatus(path).getLen(), conf);
+      new Reader(fs.open(path), fs.getFileStatus(path).getLen(), conf);
       Assert.fail("Error on handling random files.");
-    }
-    catch (IOException e) {
+    } catch (IOException e) {
       // noop, expecting exceptions
     }
   }
 
+  @Test
   public void testFailureKeyLongerThan64K() throws IOException {
     if (skip)
       return;
@@ -421,13 +425,13 @@ public class TestTFileByteArrays extends
     rand.nextBytes(buf);
     try {
       writer.append(buf, "valueX".getBytes());
-    }
-    catch (IndexOutOfBoundsException e) {
+    } catch (IndexOutOfBoundsException e) {
       // noop, expecting exceptions
     }
     closeOutput();
   }
 
+  @Test
   public void testFailureOutOfOrderKeys() throws IOException {
     if (skip)
       return;
@@ -435,8 +439,7 @@ public class TestTFileByteArrays extends
       writer.append("keyM".getBytes(), "valueM".getBytes());
       writer.append("keyA".getBytes(), "valueA".getBytes());
       Assert.fail("Error on handling out of order keys.");
-    }
-    catch (Exception e) {
+    } catch (Exception e) {
       // noop, expecting exceptions
       // e.printStackTrace();
     }
@@ -444,19 +447,20 @@ public class TestTFileByteArrays extends
     closeOutput();
   }
 
+  @Test
   public void testFailureNegativeOffset() throws IOException {
     if (skip)
       return;
     try {
       writer.append("keyX".getBytes(), -1, 4, "valueX".getBytes(), 0, 6);
       Assert.fail("Error on handling negative offset.");
-    }
-    catch (Exception e) {
+    } catch (Exception e) {
       // noop, expecting exceptions
     }
     closeOutput();
   }
 
+  @Test
   public void testFailureNegativeOffset_2() throws IOException {
     if (skip)
       return;
@@ -467,30 +471,29 @@ public class TestTFileByteArrays extends
     try {
       scanner.lowerBound("keyX".getBytes(), -1, 4);
       Assert.fail("Error on handling negative offset.");
-    }
-    catch (Exception e) {
+    } catch (Exception e) {
       // noop, expecting exceptions
-    }
-    finally {
+    } finally {
       reader.close();
       scanner.close();
     }
     closeOutput();
   }
 
+  @Test
   public void testFailureNegativeLength() throws IOException {
     if (skip)
       return;
     try {
       writer.append("keyX".getBytes(), 0, -1, "valueX".getBytes(), 0, 6);
       Assert.fail("Error on handling negative length.");
-    }
-    catch (Exception e) {
+    } catch (Exception e) {
       // noop, expecting exceptions
     }
     closeOutput();
   }
 
+  @Test
   public void testFailureNegativeLength_2() throws IOException {
     if (skip)
       return;
@@ -501,17 +504,16 @@ public class TestTFileByteArrays extends
     try {
       scanner.lowerBound("keyX".getBytes(), 0, -1);
       Assert.fail("Error on handling negative length.");
-    }
-    catch (Exception e) {
+    } catch (Exception e) {
       // noop, expecting exceptions
-    }
-    finally {
+    } finally {
       scanner.close();
       reader.close();
     }
     closeOutput();
   }
 
+  @Test
   public void testFailureNegativeLength_3() throws IOException {
     if (skip)
       return;
@@ -542,6 +544,7 @@ public class TestTFileByteArrays extends
     }
   }
 
+  @Test
   public void testFailureCompressionNotWorking() throws IOException {
     if (skip)
       return;
@@ -552,6 +555,7 @@ public class TestTFileByteArrays extends
     closeOutput();
   }
 
+  @Test
   public void testFailureFileWriteNotAt0Position() throws IOException {
     if (skip)
       return;
@@ -562,8 +566,7 @@ public class TestTFileByteArrays extends
     try {
       writer = new Writer(out, BLOCK_SIZE, compression, comparator, conf);
       Assert.fail("Failed to catch file write not at position 0.");
-    }
-    catch (Exception e) {
+    } catch (Exception e) {
       // noop, expecting exceptions
     }
     closeOutput();
@@ -585,7 +588,7 @@ public class TestTFileByteArrays extends
     long rawDataSize = 0;
     int nx;
     for (nx = 0; nx < count; nx++) {
-      byte[] key = composeSortedKey(KEY, count, nx).getBytes();
+      byte[] key = composeSortedKey(KEY, nx).getBytes();
       byte[] value = (VALUE + nx).getBytes();
       writer.append(key, value);
       rawDataSize +=
@@ -599,28 +602,13 @@ public class TestTFileByteArrays extends
    * Insert some leading 0's in front of the value, to make the keys sorted.
    * 
    * @param prefix
-   * @param total
    * @param value
    * @return
    */
-  static String composeSortedKey(String prefix, int total, int value) {
+  static String composeSortedKey(String prefix, int value) {
     return String.format("%s%010d", prefix, value);
   }
 
-  /**
-   * Calculate how many digits are in the 10-based integer.
-   * 
-   * @param value
-   * @return
-   */
-  private static int numberDigits(int value) {
-    int digits = 0;
-    while ((value = value / 10) > 0) {
-      digits++;
-    }
-    return digits;
-  }
-
   private void readRecords(int count) throws IOException {
     readRecords(fs, path, count, conf);
   }
@@ -640,7 +628,7 @@ public class TestTFileByteArrays extends
         int klen = scanner.entry().getKeyLength();
         scanner.entry().getKey(kbuf);
         Assert.assertEquals(new String(kbuf, 0, klen), composeSortedKey(KEY,
-            count, nx));
+            nx));
 
         byte[] vbuf = new byte[BUF_SIZE];
         int vlen = scanner.entry().getValueLength();
@@ -650,30 +638,28 @@ public class TestTFileByteArrays extends
 
       Assert.assertTrue(scanner.atEnd());
       Assert.assertFalse(scanner.advance());
-    }
-    finally {
+    } finally {
       scanner.close();
       reader.close();
     }
   }
 
-  private void checkBlockIndex(int count, int recordIndex,
-      int blockIndexExpected) throws IOException {
+  private void checkBlockIndex(int recordIndex, int blockIndexExpected) throws IOException
{
     Reader reader = new Reader(fs.open(path), fs.getFileStatus(path).getLen(), conf);
     Scanner scanner = reader.createScanner();
-    scanner.seekTo(composeSortedKey(KEY, count, recordIndex).getBytes());
+    scanner.seekTo(composeSortedKey(KEY, recordIndex).getBytes());
     Assert.assertEquals(blockIndexExpected, scanner.currentLocation
         .getBlockIndex());
     scanner.close();
     reader.close();
   }
 
-  private void readValueBeforeKey(int count, int recordIndex)
+  private void readValueBeforeKey(int recordIndex)
       throws IOException {
     Reader reader =
         new Reader(fs.open(path), fs.getFileStatus(path).getLen(), conf);
     Scanner scanner =
-        reader.createScannerByKey(composeSortedKey(KEY, count, recordIndex)
+        reader.createScannerByKey(composeSortedKey(KEY, recordIndex)
             .getBytes(), null);
 
     try {
@@ -686,19 +672,18 @@ public class TestTFileByteArrays extends
       int klen = scanner.entry().getKeyLength();
       scanner.entry().getKey(kbuf);
       Assert.assertEquals(new String(kbuf, 0, klen), composeSortedKey(KEY,
-          count, recordIndex));
-    }
-    finally {
+          recordIndex));
+    } finally {
       scanner.close();
       reader.close();
     }
   }
 
-  private void readKeyWithoutValue(int count, int recordIndex)
+  private void readKeyWithoutValue(int recordIndex)
       throws IOException {
     Reader reader = new Reader(fs.open(path), fs.getFileStatus(path).getLen(), conf);
     Scanner scanner =
-        reader.createScannerByKey(composeSortedKey(KEY, count, recordIndex)
+        reader.createScannerByKey(composeSortedKey(KEY, recordIndex)
             .getBytes(), null);
 
     try {
@@ -707,7 +692,7 @@ public class TestTFileByteArrays extends
       int klen1 = scanner.entry().getKeyLength();
       scanner.entry().getKey(kbuf1);
       Assert.assertEquals(new String(kbuf1, 0, klen1), composeSortedKey(KEY,
-          count, recordIndex));
+          recordIndex));
 
       if (scanner.advance() && !scanner.atEnd()) {
         // read the next key following the indexed
@@ -715,21 +700,20 @@ public class TestTFileByteArrays extends
         int klen2 = scanner.entry().getKeyLength();
         scanner.entry().getKey(kbuf2);
         Assert.assertEquals(new String(kbuf2, 0, klen2), composeSortedKey(KEY,
-            count, recordIndex + 1));
+            recordIndex + 1));
       }
-    }
-    finally {
+    } finally {
       scanner.close();
       reader.close();
     }
   }
 
-  private void readValueWithoutKey(int count, int recordIndex)
+  private void readValueWithoutKey(int recordIndex)
       throws IOException {
     Reader reader = new Reader(fs.open(path), fs.getFileStatus(path).getLen(), conf);
 
     Scanner scanner =
-        reader.createScannerByKey(composeSortedKey(KEY, count, recordIndex)
+        reader.createScannerByKey(composeSortedKey(KEY, recordIndex)
             .getBytes(), null);
 
     byte[] vbuf1 = new byte[BUF_SIZE];
@@ -749,11 +733,11 @@ public class TestTFileByteArrays extends
     reader.close();
   }
 
-  private void readKeyManyTimes(int count, int recordIndex) throws IOException {
+  private void readKeyManyTimes(int recordIndex) throws IOException {
     Reader reader = new Reader(fs.open(path), fs.getFileStatus(path).getLen(), conf);
 
     Scanner scanner =
-        reader.createScannerByKey(composeSortedKey(KEY, count, recordIndex)
+        reader.createScannerByKey(composeSortedKey(KEY, recordIndex)
             .getBytes(), null);
 
     // read the indexed key
@@ -761,17 +745,17 @@ public class TestTFileByteArrays extends
     int klen1 = scanner.entry().getKeyLength();
     scanner.entry().getKey(kbuf1);
     Assert.assertEquals(new String(kbuf1, 0, klen1), composeSortedKey(KEY,
-        count, recordIndex));
+        recordIndex));
 
     klen1 = scanner.entry().getKeyLength();
     scanner.entry().getKey(kbuf1);
     Assert.assertEquals(new String(kbuf1, 0, klen1), composeSortedKey(KEY,
-        count, recordIndex));
+        recordIndex));
 
     klen1 = scanner.entry().getKeyLength();
     scanner.entry().getKey(kbuf1);
     Assert.assertEquals(new String(kbuf1, 0, klen1), composeSortedKey(KEY,
-        count, recordIndex));
+        recordIndex));
 
     scanner.close();
     reader.close();

Modified: hadoop/common/trunk/common/src/test/core/org/apache/hadoop/io/file/tfile/TestTFileJClassComparatorByteArrays.java
URL: http://svn.apache.org/viewvc/hadoop/common/trunk/common/src/test/core/org/apache/hadoop/io/file/tfile/TestTFileJClassComparatorByteArrays.java?rev=1148933&r1=1148932&r2=1148933&view=diff
==============================================================================
--- hadoop/common/trunk/common/src/test/core/org/apache/hadoop/io/file/tfile/TestTFileJClassComparatorByteArrays.java
(original)
+++ hadoop/common/trunk/common/src/test/core/org/apache/hadoop/io/file/tfile/TestTFileJClassComparatorByteArrays.java
Wed Jul 20 20:42:15 2011
@@ -38,7 +38,7 @@ public class TestTFileJClassComparatorBy
   public void setUp() throws IOException {
     init(Compression.Algorithm.GZ.getName(),
         "jclass: org.apache.hadoop.io.file.tfile.MyComparator",
-        "TFileTestJClassComparator", 4480, 4263);
+        "TFileTestJClassComparator");
     super.setUp();
   }
 }

Modified: hadoop/common/trunk/common/src/test/core/org/apache/hadoop/io/file/tfile/TestTFileStreams.java
URL: http://svn.apache.org/viewvc/hadoop/common/trunk/common/src/test/core/org/apache/hadoop/io/file/tfile/TestTFileStreams.java?rev=1148933&r1=1148932&r2=1148933&view=diff
==============================================================================
--- hadoop/common/trunk/common/src/test/core/org/apache/hadoop/io/file/tfile/TestTFileStreams.java
(original)
+++ hadoop/common/trunk/common/src/test/core/org/apache/hadoop/io/file/tfile/TestTFileStreams.java
Wed Jul 20 20:42:15 2011
@@ -383,7 +383,7 @@ public class TestTFileStreams extends Te
       boolean knownValueLength, boolean close) throws IOException {
     long rawDataSize = 0;
     for (int nx = 0; nx < count; nx++) {
-      String key = TestTFileByteArrays.composeSortedKey("key", count, nx);
+      String key = TestTFileByteArrays.composeSortedKey("key", nx);
       DataOutputStream outKey =
           writer.prepareAppendKey(knownKeyLength ? key.length() : -1);
       outKey.write(key.getBytes());



Mime
View raw message