incubator-hcatalog-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From macy...@apache.org
Subject svn commit: r1104614 [2/2] - in /incubator/hcatalog/trunk: bin/ src/java/org/apache/hcatalog/cli/SemanticAnalysis/ src/java/org/apache/hcatalog/common/ src/java/org/apache/hcatalog/data/ src/java/org/apache/hcatalog/data/schema/ src/java/org/apache/hca...
Date Tue, 17 May 2011 22:50:51 GMT
Modified: incubator/hcatalog/trunk/src/java/org/apache/hcatalog/pig/HCatStorer.java
URL: http://svn.apache.org/viewvc/incubator/hcatalog/trunk/src/java/org/apache/hcatalog/pig/HCatStorer.java?rev=1104614&r1=1104613&r2=1104614&view=diff
==============================================================================
--- incubator/hcatalog/trunk/src/java/org/apache/hcatalog/pig/HCatStorer.java (original)
+++ incubator/hcatalog/trunk/src/java/org/apache/hcatalog/pig/HCatStorer.java Tue May 17 22:50:49 2011
@@ -40,7 +40,7 @@ import org.apache.pig.impl.util.ObjectSe
 import org.apache.pig.impl.util.UDFContext;
 
 /**
- * HowlStorer.
+ * HCatStorer.
  *
  */
 
@@ -75,11 +75,11 @@ public class HCatStorer extends HCatBase
     String[] userStr = location.split("\\.");
     HCatTableInfo tblInfo;
     if(userStr.length == 2) {
-      tblInfo = HCatTableInfo.getOutputTableInfo(PigHCatUtil.getHowlServerUri(job),
-          PigHCatUtil.getHowlServerPrincipal(job), userStr[0],userStr[1],partitions);
+      tblInfo = HCatTableInfo.getOutputTableInfo(PigHCatUtil.getHCatServerUri(job),
+          PigHCatUtil.getHCatServerPrincipal(job), userStr[0],userStr[1],partitions);
     } else {
-      tblInfo = HCatTableInfo.getOutputTableInfo(PigHCatUtil.getHowlServerUri(job),
-          PigHCatUtil.getHowlServerPrincipal(job), null,userStr[0],partitions);
+      tblInfo = HCatTableInfo.getOutputTableInfo(PigHCatUtil.getHCatServerUri(job),
+          PigHCatUtil.getHCatServerPrincipal(job), null,userStr[0],partitions);
     }
 
 
@@ -101,13 +101,13 @@ public class HCatStorer extends HCatBase
           // information passed to HCatOutputFormat was not right
           throw new PigException(he.getMessage(), PigHCatUtil.PIG_EXCEPTION_CODE, he);
       }
-      HCatSchema howlTblSchema = HCatOutputFormat.getTableSchema(job);
+      HCatSchema hcatTblSchema = HCatOutputFormat.getTableSchema(job);
       try{
-        doSchemaValidations(pigSchema, howlTblSchema);
+        doSchemaValidations(pigSchema, hcatTblSchema);
       } catch(HCatException he){
         throw new FrontendException(he.getMessage(), PigHCatUtil.PIG_EXCEPTION_CODE, he);
       }
-      computedSchema = convertPigSchemaToHCatSchema(pigSchema,howlTblSchema);
+      computedSchema = convertPigSchemaToHCatSchema(pigSchema,hcatTblSchema);
       HCatOutputFormat.setSchema(job, computedSchema);
       p.setProperty(HCatConstants.HCAT_KEY_OUTPUT_INFO, config.get(HCatConstants.HCAT_KEY_OUTPUT_INFO));
       if(config.get(HCatConstants.HCAT_KEY_HIVE_CONF) != null){
@@ -134,7 +134,7 @@ public class HCatStorer extends HCatBase
   @Override
   public void storeSchema(ResourceSchema schema, String arg1, Job job) throws IOException {
     if( job.getConfiguration().get("mapred.job.tracker", "").equalsIgnoreCase("local") ) {
-      //In local mode, mapreduce will not call HowlOutputCommitter.cleanupJob.
+      //In local mode, mapreduce will not call HCatOutputCommitter.cleanupJob.
       //Calling it from here so that the partition publish happens.
       //This call needs to be removed after MAPREDUCE-1447 is fixed.
       new HCatOutputCommitter(null).cleanupJob(job);

Modified: incubator/hcatalog/trunk/src/java/org/apache/hcatalog/pig/PigHCatUtil.java
URL: http://svn.apache.org/viewvc/incubator/hcatalog/trunk/src/java/org/apache/hcatalog/pig/PigHCatUtil.java?rev=1104614&r1=1104613&r2=1104614&view=diff
==============================================================================
--- incubator/hcatalog/trunk/src/java/org/apache/hcatalog/pig/PigHCatUtil.java (original)
+++ incubator/hcatalog/trunk/src/java/org/apache/hcatalog/pig/PigHCatUtil.java Tue May 17 22:50:49 2011
@@ -54,7 +54,7 @@ public class PigHCatUtil {
   static final int PIG_EXCEPTION_CODE = 1115; // http://wiki.apache.org/pig/PigErrorHandlingFunctionalSpecification#Error_codes
   private static final String DEFAULT_DB = MetaStoreUtils.DEFAULT_DATABASE_NAME;
 
-  private final  Map<Pair<String,String>, Table> howlTableCache =
+  private final  Map<Pair<String,String>, Table> hcatTableCache =
     new HashMap<Pair<String,String>, Table>();
 
   private static final TupleFactory tupFac = TupleFactory.getInstance();
@@ -62,7 +62,7 @@ public class PigHCatUtil {
   static public Pair<String, String> getDBTableNames(String location) throws IOException {
     // the location string will be of the form:
     // <database name>.<table name> - parse it and
-    // communicate the information to HowlInputFormat
+    // communicate the information to HCatInputFormat
 
     String[] dbTableNametokens = location.split("\\.");
     if(dbTableNametokens.length == 1) {
@@ -77,12 +77,12 @@ public class PigHCatUtil {
     }
   }
 
-  static public String getHowlServerUri(Job job) {
+  static public String getHCatServerUri(Job job) {
 
     return job.getConfiguration().get(HCatConstants.HCAT_METASTORE_URI);
   }
 
-  static public String getHowlServerPrincipal(Job job) {
+  static public String getHCatServerPrincipal(Job job) {
 
     return job.getConfiguration().get(HCatConstants.HCAT_METASTORE_PRINCIPAL);
   }
@@ -120,20 +120,20 @@ public class PigHCatUtil {
 
     Properties props = UDFContext.getUDFContext().getUDFProperties(
         classForUDFCLookup, new String[] {signature});
-    HCatSchema howlTableSchema = (HCatSchema) props.get(HCatConstants.HCAT_TABLE_SCHEMA);
+    HCatSchema hcatTableSchema = (HCatSchema) props.get(HCatConstants.HCAT_TABLE_SCHEMA);
 
     ArrayList<HCatFieldSchema> fcols = new ArrayList<HCatFieldSchema>();
     for(RequiredField rf: fields) {
-      fcols.add(howlTableSchema.getFields().get(rf.getIndex()));
+      fcols.add(hcatTableSchema.getFields().get(rf.getIndex()));
     }
     return new HCatSchema(fcols);
   }
 
-  public Table getTable(String location, String howlServerUri, String howlServerPrincipal) throws IOException{
-    Pair<String, String> loc_server = new Pair<String,String>(location, howlServerUri);
-    Table howlTable = howlTableCache.get(loc_server);
-    if(howlTable != null){
-      return howlTable;
+  public Table getTable(String location, String hcatServerUri, String hcatServerPrincipal) throws IOException{
+    Pair<String, String> loc_server = new Pair<String,String>(location, hcatServerUri);
+    Table hcatTable = hcatTableCache.get(loc_server);
+    if(hcatTable != null){
+      return hcatTable;
     }
 
     Pair<String, String> dbTablePair = PigHCatUtil.getDBTableNames(location);
@@ -141,21 +141,21 @@ public class PigHCatUtil {
     String tableName = dbTablePair.second;
     Table table = null;
     try {
-      client = createHiveMetaClient(howlServerUri, howlServerPrincipal, PigHCatUtil.class);
+      client = createHiveMetaClient(hcatServerUri, hcatServerPrincipal, PigHCatUtil.class);
       table = client.getTable(dbName, tableName);
     } catch (NoSuchObjectException nsoe){
       throw new PigException("Table not found : " + nsoe.getMessage(), PIG_EXCEPTION_CODE); // prettier error messages to frontend
     } catch (Exception e) {
       throw new IOException(e);
     }
-    howlTableCache.put(loc_server, table);
+    hcatTableCache.put(loc_server, table);
     return table;
   }
 
-  public static ResourceSchema getResourceSchema(HCatSchema howlSchema) throws IOException {
+  public static ResourceSchema getResourceSchema(HCatSchema hcatSchema) throws IOException {
 
     List<ResourceFieldSchema> rfSchemaList = new ArrayList<ResourceFieldSchema>();
-    for (HCatFieldSchema hfs : howlSchema.getFields()){
+    for (HCatFieldSchema hfs : hcatSchema.getFields()){
       ResourceFieldSchema rfSchema;
       rfSchema = getResourceSchemaFromFieldSchema(hfs);
       rfSchemaList.add(rfSchema);
@@ -229,7 +229,7 @@ public class PigHCatUtil {
   }
 
 /**
-   * @param type owl column type
+   * @param type hcat column type
    * @return corresponding pig type
    * @throws IOException
    */
@@ -351,8 +351,8 @@ public static Object extractPigObject(Ob
   }
 
 
-  public static void validateHowlTableSchemaFollowsPigRules(HCatSchema howlTableSchema) throws IOException {
-      for (HCatFieldSchema hfs : howlTableSchema.getFields()){
+  public static void validateHCatTableSchemaFollowsPigRules(HCatSchema hcatTableSchema) throws IOException {
+      for (HCatFieldSchema hfs : hcatTableSchema.getFields()){
           Type htype = hfs.getType();
           if (htype == Type.ARRAY){
               validateIsPigCompatibleArrayWithPrimitivesOrSimpleComplexTypes(hfs);

Modified: incubator/hcatalog/trunk/src/java/org/apache/hcatalog/pig/drivers/LoadFuncBasedInputDriver.java
URL: http://svn.apache.org/viewvc/incubator/hcatalog/trunk/src/java/org/apache/hcatalog/pig/drivers/LoadFuncBasedInputDriver.java?rev=1104614&r1=1104613&r2=1104614&view=diff
==============================================================================
--- incubator/hcatalog/trunk/src/java/org/apache/hcatalog/pig/drivers/LoadFuncBasedInputDriver.java (original)
+++ incubator/hcatalog/trunk/src/java/org/apache/hcatalog/pig/drivers/LoadFuncBasedInputDriver.java Tue May 17 22:50:49 2011
@@ -38,9 +38,9 @@ import org.apache.pig.data.Tuple;
 
 
 /**
- * This is a base class which wraps a Load func in HowlInputStorageDriver.
+ * This is a base class which wraps a Load func in HCatInputStorageDriver.
  * If you already have a LoadFunc, then this class along with LoadFuncBasedInputFormat
- * is doing all the heavy lifting. For a new Howl Input Storage Driver just extend it
+ * is doing all the heavy lifting. For a new HCat Input Storage Driver just extend it
  * and override the initialize(). {@link PigStorageInputDriver} illustrates
  * that well.
  */
@@ -57,7 +57,7 @@ public abstract class LoadFuncBasedInput
       throws IOException {
 
     List<Object> data = ((Tuple)baseValue).getAll();
-    List<Object> howlRecord = new ArrayList<Object>(desiredColNames.size());
+    List<Object> hcatRecord = new ArrayList<Object>(desiredColNames.size());
 
     /* Iterate through columns asked for in output schema, look them up in
      * original data schema. If found, put it. Else look up in partition columns
@@ -66,28 +66,28 @@ public abstract class LoadFuncBasedInput
      */
     for(String colName : desiredColNames){
       Integer idx = dataSchema.getPosition(colName);
-      howlRecord.add( idx != null ? data.get(idx) : partVals.get(colName));
+      hcatRecord.add( idx != null ? data.get(idx) : partVals.get(colName));
     }
-    return new DefaultHCatRecord(howlRecord);
+    return new DefaultHCatRecord(hcatRecord);
   }
 
   @Override
   public InputFormat<? extends WritableComparable, ? extends Writable> getInputFormat(
-      Properties howlProperties) {
+      Properties hcatProperties) {
 
     return inputFormat;
   }
 
   @Override
-  public void setOriginalSchema(JobContext jobContext, HCatSchema howlSchema) throws IOException {
+  public void setOriginalSchema(JobContext jobContext, HCatSchema hcatSchema) throws IOException {
 
-    dataSchema = howlSchema;
+    dataSchema = hcatSchema;
   }
 
   @Override
-  public void setOutputSchema(JobContext jobContext, HCatSchema howlSchema) throws IOException {
+  public void setOutputSchema(JobContext jobContext, HCatSchema hcatSchema) throws IOException {
 
-    desiredColNames = howlSchema.getFieldNames();
+    desiredColNames = hcatSchema.getFieldNames();
   }
 
   @Override

Modified: incubator/hcatalog/trunk/src/java/org/apache/hcatalog/rcfile/RCFileInputDriver.java
URL: http://svn.apache.org/viewvc/incubator/hcatalog/trunk/src/java/org/apache/hcatalog/rcfile/RCFileInputDriver.java?rev=1104614&r1=1104613&r2=1104614&view=diff
==============================================================================
--- incubator/hcatalog/trunk/src/java/org/apache/hcatalog/rcfile/RCFileInputDriver.java (original)
+++ incubator/hcatalog/trunk/src/java/org/apache/hcatalog/rcfile/RCFileInputDriver.java Tue May 17 22:50:49 2011
@@ -69,7 +69,7 @@ public class RCFileInputDriver extends H
   private Map<String,Integer> namePosMapping;
 
   @Override
-  public InputFormat<? extends WritableComparable, ? extends Writable> getInputFormat(Properties howlProperties) {
+  public InputFormat<? extends WritableComparable, ? extends Writable> getInputFormat(Properties hcatProperties) {
     return new RCFileMapReduceInputFormat<LongWritable, BytesRefArrayWritable>();
   }
 
@@ -117,7 +117,7 @@ public class RCFileInputDriver extends H
   public HCatRecord convertToHCatRecord(WritableComparable ignored, Writable bytesRefArray) throws IOException {
 
     // Deserialize bytesRefArray into struct and then convert that struct to
-    // HowlRecord.
+    // HCatRecord.
     ColumnarStruct struct;
     try {
       struct = (ColumnarStruct)serde.deserialize(bytesRefArray);
@@ -152,7 +152,7 @@ public class RCFileInputDriver extends H
   private Object getTypedObj(Object data, ObjectInspector oi) throws IOException{
 
     // The real work-horse method. We are gobbling up all the laziness benefits
-    // of Hive-RCFile by deserializing everything and creating crisp  HowlRecord
+    // of Hive-RCFile by deserializing everything and creating crisp  HCatRecord
     // with crisp Java objects inside it. We have to do it because higher layer
     // may not know how to do it.
 
@@ -200,26 +200,26 @@ public class RCFileInputDriver extends H
   }
 
   @Override
-  public void initialize(JobContext context,Properties howlProperties)
+  public void initialize(JobContext context,Properties hcatProperties)
   throws IOException {
 
-    super.initialize(context, howlProperties);
+    super.initialize(context, hcatProperties);
 
     // Columnar Serde needs to know names and types of columns it needs to read.
     List<FieldSchema> fields = HCatUtil.getFieldSchemaList(colsInData);
-    howlProperties.setProperty(Constants.LIST_COLUMNS,MetaStoreUtils.
+    hcatProperties.setProperty(Constants.LIST_COLUMNS,MetaStoreUtils.
         getColumnNamesFromFieldSchema(fields));
-    howlProperties.setProperty(Constants.LIST_COLUMN_TYPES, MetaStoreUtils.
+    hcatProperties.setProperty(Constants.LIST_COLUMN_TYPES, MetaStoreUtils.
         getColumnTypesFromFieldSchema(fields));
 
     // It seems RCFIle reads and writes nulls differently as compared to default hive.
     // setting these props to match LazySimpleSerde
-    howlProperties.setProperty(Constants.SERIALIZATION_NULL_FORMAT, "\\N");
-    howlProperties.setProperty(Constants.SERIALIZATION_FORMAT, "1");
+    hcatProperties.setProperty(Constants.SERIALIZATION_NULL_FORMAT, "\\N");
+    hcatProperties.setProperty(Constants.SERIALIZATION_FORMAT, "1");
 
     try {
       serde = new ColumnarSerDe();
-      serde.initialize(context.getConfiguration(), howlProperties);
+      serde.initialize(context.getConfiguration(), hcatProperties);
       oi = (StructObjectInspector) serde.getObjectInspector();
       structFields = oi.getAllStructFieldRefs();
 

Modified: incubator/hcatalog/trunk/src/test/org/apache/hcatalog/cli/TestPermsGrp.java
URL: http://svn.apache.org/viewvc/incubator/hcatalog/trunk/src/test/org/apache/hcatalog/cli/TestPermsGrp.java?rev=1104614&r1=1104613&r2=1104614&view=diff
==============================================================================
--- incubator/hcatalog/trunk/src/test/org/apache/hcatalog/cli/TestPermsGrp.java (original)
+++ incubator/hcatalog/trunk/src/test/org/apache/hcatalog/cli/TestPermsGrp.java Tue May 17 22:50:49 2011
@@ -52,7 +52,7 @@ public class TestPermsGrp extends TestCa
 
   private boolean isServerRunning = false;
   private static final String msPort = "20101";
-  private HiveConf howlConf;
+  private HiveConf hcatConf;
   private Warehouse clientWH;
   private Thread t;
   private HiveMetaStoreClient msc;
@@ -91,17 +91,17 @@ public class TestPermsGrp extends TestCa
     securityManager = System.getSecurityManager();
     System.setSecurityManager(new NoExitSecurityManager());
 
-    howlConf = new HiveConf(this.getClass());
-    howlConf.set("hive.metastore.local", "false");
-    howlConf.setVar(HiveConf.ConfVars.METASTOREURIS, "thrift://localhost:" + msPort);
-    howlConf.setIntVar(HiveConf.ConfVars.METASTORETHRIFTRETRIES, 3);
-
-    howlConf.set(HiveConf.ConfVars.SEMANTIC_ANALYZER_HOOK.varname, HCatSemanticAnalyzer.class.getName());
-    howlConf.set(HiveConf.ConfVars.PREEXECHOOKS.varname, "");
-    howlConf.set(HiveConf.ConfVars.POSTEXECHOOKS.varname, "");
-    howlConf.set(HiveConf.ConfVars.HIVE_SUPPORT_CONCURRENCY.varname, "false");
-    clientWH = new Warehouse(howlConf);
-    msc = new HiveMetaStoreClient(howlConf,null);
+    hcatConf = new HiveConf(this.getClass());
+    hcatConf.set("hive.metastore.local", "false");
+    hcatConf.setVar(HiveConf.ConfVars.METASTOREURIS, "thrift://localhost:" + msPort);
+    hcatConf.setIntVar(HiveConf.ConfVars.METASTORETHRIFTRETRIES, 3);
+
+    hcatConf.set(HiveConf.ConfVars.SEMANTIC_ANALYZER_HOOK.varname, HCatSemanticAnalyzer.class.getName());
+    hcatConf.set(HiveConf.ConfVars.PREEXECHOOKS.varname, "");
+    hcatConf.set(HiveConf.ConfVars.POSTEXECHOOKS.varname, "");
+    hcatConf.set(HiveConf.ConfVars.HIVE_SUPPORT_CONCURRENCY.varname, "false");
+    clientWH = new Warehouse(hcatConf);
+    msc = new HiveMetaStoreClient(hcatConf,null);
     System.setProperty(HiveConf.ConfVars.PREEXECHOOKS.varname, " ");
     System.setProperty(HiveConf.ConfVars.POSTEXECHOOKS.varname, " ");
   }
@@ -130,12 +130,12 @@ public class TestPermsGrp extends TestCa
         assertEquals(((ExitException)e).getStatus(), 0);
       }
       dfsPath = clientWH.getDefaultTablePath(dbName, tblName);
-      assertTrue(dfsPath.getFileSystem(howlConf).getFileStatus(dfsPath).getPermission().equals(FsPermission.valueOf("drwx-wx---")));
+      assertTrue(dfsPath.getFileSystem(hcatConf).getFileStatus(dfsPath).getPermission().equals(FsPermission.valueOf("drwx-wx---")));
 
       cleanupTbl(dbName, tblName, typeName);
 
       // User specified perms in invalid format.
-      howlConf.set(HCatConstants.HCAT_PERMS, "rwx");
+      hcatConf.set(HCatConstants.HCAT_PERMS, "rwx");
       // make sure create table fails.
       try{
         HCatCli.main(new String[]{"-e","create table simptbl (name string) stored as RCFILE", "-p","rwx"});
@@ -146,7 +146,7 @@ public class TestPermsGrp extends TestCa
       // No physical dir gets created.
       dfsPath = clientWH.getDefaultTablePath(MetaStoreUtils.DEFAULT_DATABASE_NAME,tblName);
       try{
-        dfsPath.getFileSystem(howlConf).getFileStatus(dfsPath);
+        dfsPath.getFileSystem(hcatConf).getFileStatus(dfsPath);
         assert false;
       } catch(Exception fnfe){
         assertTrue(fnfe instanceof FileNotFoundException);
@@ -162,8 +162,8 @@ public class TestPermsGrp extends TestCa
       }
 
       // test for invalid group name
-      howlConf.set(HCatConstants.HCAT_PERMS, "drw-rw-rw-");
-      howlConf.set(HCatConstants.HCAT_GROUP, "THIS_CANNOT_BE_A_VALID_GRP_NAME_EVER");
+      hcatConf.set(HCatConstants.HCAT_PERMS, "drw-rw-rw-");
+      hcatConf.set(HCatConstants.HCAT_GROUP, "THIS_CANNOT_BE_A_VALID_GRP_NAME_EVER");
 
       try{
         // create table must fail.
@@ -183,7 +183,7 @@ public class TestPermsGrp extends TestCa
       }
       try{
         // neither dir should get created.
-        dfsPath.getFileSystem(howlConf).getFileStatus(dfsPath);
+        dfsPath.getFileSystem(hcatConf).getFileStatus(dfsPath);
         assert false;
       } catch(Exception e){
         assertTrue(e instanceof FileNotFoundException);

Modified: incubator/hcatalog/trunk/src/test/org/apache/hcatalog/cli/TestSemanticAnalysis.java
URL: http://svn.apache.org/viewvc/incubator/hcatalog/trunk/src/test/org/apache/hcatalog/cli/TestSemanticAnalysis.java?rev=1104614&r1=1104613&r2=1104614&view=diff
==============================================================================
--- incubator/hcatalog/trunk/src/test/org/apache/hcatalog/cli/TestSemanticAnalysis.java (original)
+++ incubator/hcatalog/trunk/src/test/org/apache/hcatalog/cli/TestSemanticAnalysis.java Tue May 17 22:50:49 2011
@@ -50,26 +50,26 @@ import org.apache.thrift.TException;
 
 public class TestSemanticAnalysis extends TestCase{
 
-  private Driver howlDriver;
+  private Driver hcatDriver;
   private Driver hiveDriver;
   private HiveMetaStoreClient msc;
 
   @Override
   protected void setUp() throws Exception {
 
-    HiveConf howlConf = new HiveConf(this.getClass());
-    howlConf.set(ConfVars.PREEXECHOOKS.varname, "");
-    howlConf.set(ConfVars.POSTEXECHOOKS.varname, "");
-    howlConf.set(ConfVars.HIVE_SUPPORT_CONCURRENCY.varname, "false");
+    HiveConf hcatConf = new HiveConf(this.getClass());
+    hcatConf.set(ConfVars.PREEXECHOOKS.varname, "");
+    hcatConf.set(ConfVars.POSTEXECHOOKS.varname, "");
+    hcatConf.set(ConfVars.HIVE_SUPPORT_CONCURRENCY.varname, "false");
 
-    HiveConf hiveConf = new HiveConf(howlConf,this.getClass());
+    HiveConf hiveConf = new HiveConf(hcatConf,this.getClass());
     hiveDriver = new Driver(hiveConf);
 
-    howlConf.set(ConfVars.SEMANTIC_ANALYZER_HOOK.varname, HCatSemanticAnalyzer.class.getName());
-    howlDriver = new Driver(howlConf);
+    hcatConf.set(ConfVars.SEMANTIC_ANALYZER_HOOK.varname, HCatSemanticAnalyzer.class.getName());
+    hcatDriver = new Driver(hcatConf);
 
-    msc = new HiveMetaStoreClient(howlConf);
-    SessionState.start(new CliSessionState(howlConf));
+    msc = new HiveMetaStoreClient(hcatConf);
+    SessionState.start(new CliSessionState(hcatConf));
   }
 
   String query;
@@ -80,7 +80,7 @@ public class TestSemanticAnalysis extend
     hiveDriver.run("drop table junit_sem_analysis");
     hiveDriver.run("create table junit_sem_analysis (a int) partitioned by (b string) stored as TEXTFILE");
     hiveDriver.run("alter table junit_sem_analysis add partition (b='2010-10-10')");
-    howlDriver.run("alter table junit_sem_analysis partition (b='2010-10-10') set fileformat RCFILE");
+    hcatDriver.run("alter table junit_sem_analysis partition (b='2010-10-10') set fileformat RCFILE");
 
     Table tbl = msc.getTable(MetaStoreUtils.DEFAULT_DATABASE_NAME, tblName);
     assertEquals(TextInputFormat.class.getName(),tbl.getSd().getInputFormat());
@@ -100,7 +100,7 @@ public class TestSemanticAnalysis extend
     assertEquals(RCFileInputDriver.class.getName(), partParams.get(HCatConstants.HCAT_ISD_CLASS));
     assertEquals(RCFileOutputDriver.class.getName(), partParams.get(HCatConstants.HCAT_OSD_CLASS));
 
-    howlDriver.run("drop table junit_sem_analysis");
+    hcatDriver.run("drop table junit_sem_analysis");
   }
 
   public void testDatabaseOperations() throws MetaException, CommandNeedRetryException {
@@ -110,30 +110,30 @@ public class TestSemanticAnalysis extend
     String testDb2 = "testdatabaseoperatons2";
 
     if (dbs.contains(testDb1.toLowerCase())){
-      assertEquals(0,howlDriver.run("drop database "+testDb1).getResponseCode());
+      assertEquals(0,hcatDriver.run("drop database "+testDb1).getResponseCode());
     }
 
     if (dbs.contains(testDb2.toLowerCase())){
-      assertEquals(0,howlDriver.run("drop database "+testDb2).getResponseCode());
+      assertEquals(0,hcatDriver.run("drop database "+testDb2).getResponseCode());
     }
 
-    assertEquals(0,howlDriver.run("create database "+testDb1).getResponseCode());
+    assertEquals(0,hcatDriver.run("create database "+testDb1).getResponseCode());
     assertTrue(msc.getAllDatabases().contains(testDb1));
-    assertEquals(0,howlDriver.run("create database if not exists "+testDb1).getResponseCode());
+    assertEquals(0,hcatDriver.run("create database if not exists "+testDb1).getResponseCode());
     assertTrue(msc.getAllDatabases().contains(testDb1));
-    assertEquals(0,howlDriver.run("create database if not exists "+testDb2).getResponseCode());
+    assertEquals(0,hcatDriver.run("create database if not exists "+testDb2).getResponseCode());
     assertTrue(msc.getAllDatabases().contains(testDb2));
 
-    assertEquals(0,howlDriver.run("drop database "+testDb1).getResponseCode());
-    assertEquals(0,howlDriver.run("drop database "+testDb2).getResponseCode());
+    assertEquals(0,hcatDriver.run("drop database "+testDb1).getResponseCode());
+    assertEquals(0,hcatDriver.run("drop database "+testDb2).getResponseCode());
     assertFalse(msc.getAllDatabases().contains(testDb1));
     assertFalse(msc.getAllDatabases().contains(testDb2));
   }
 
   public void testCreateTableIfNotExists() throws MetaException, TException, NoSuchObjectException, CommandNeedRetryException{
 
-    howlDriver.run("drop table "+tblName);
-    howlDriver.run("create table junit_sem_analysis (a int) stored as RCFILE");
+    hcatDriver.run("drop table "+tblName);
+    hcatDriver.run("create table junit_sem_analysis (a int) stored as RCFILE");
     Table tbl = msc.getTable(MetaStoreUtils.DEFAULT_DATABASE_NAME, tblName);
     List<FieldSchema> cols = tbl.getSd().getCols();
     assertEquals(1, cols.size());
@@ -144,7 +144,7 @@ public class TestSemanticAnalysis extend
     assertEquals(RCFileInputDriver.class.getName(), tblParams.get(HCatConstants.HCAT_ISD_CLASS));
     assertEquals(RCFileOutputDriver.class.getName(), tblParams.get(HCatConstants.HCAT_OSD_CLASS));
 
-    CommandProcessorResponse resp = howlDriver.run("create table if not exists junit_sem_analysis (a int) stored as RCFILE");
+    CommandProcessorResponse resp = hcatDriver.run("create table if not exists junit_sem_analysis (a int) stored as RCFILE");
     assertEquals(0, resp.getResponseCode());
     assertNull(resp.getErrorMessage());
     tbl = msc.getTable(MetaStoreUtils.DEFAULT_DATABASE_NAME, tblName);
@@ -157,50 +157,50 @@ public class TestSemanticAnalysis extend
     tblParams = tbl.getParameters();
     assertEquals(RCFileInputDriver.class.getName(), tblParams.get(HCatConstants.HCAT_ISD_CLASS));
     assertEquals(RCFileOutputDriver.class.getName(), tblParams.get(HCatConstants.HCAT_OSD_CLASS));
-    howlDriver.run("drop table junit_sem_analysis");
+    hcatDriver.run("drop table junit_sem_analysis");
   }
 
   public void testAlterTblTouch() throws CommandNeedRetryException{
 
-    howlDriver.run("drop table junit_sem_analysis");
-    howlDriver.run("create table junit_sem_analysis (a int) partitioned by (b string) stored as RCFILE");
-    CommandProcessorResponse response = howlDriver.run("alter table junit_sem_analysis touch");
+    hcatDriver.run("drop table junit_sem_analysis");
+    hcatDriver.run("create table junit_sem_analysis (a int) partitioned by (b string) stored as RCFILE");
+    CommandProcessorResponse response = hcatDriver.run("alter table junit_sem_analysis touch");
     assertEquals(10, response.getResponseCode());
     assertTrue(response.getErrorMessage().contains("Operation not supported."));
 
-    howlDriver.run("alter table junit_sem_analysis touch partition (b='12')");
+    hcatDriver.run("alter table junit_sem_analysis touch partition (b='12')");
     assertEquals(10, response.getResponseCode());
     assertTrue(response.getErrorMessage().contains("Operation not supported."));
 
-    howlDriver.run("drop table junit_sem_analysis");
+    hcatDriver.run("drop table junit_sem_analysis");
   }
 
   public void testChangeColumns() throws CommandNeedRetryException{
-    howlDriver.run("drop table junit_sem_analysis");
-    howlDriver.run("create table junit_sem_analysis (a int, c string) partitioned by (b string) stored as RCFILE");
-    CommandProcessorResponse response = howlDriver.run("alter table junit_sem_analysis change a a1 int");
+    hcatDriver.run("drop table junit_sem_analysis");
+    hcatDriver.run("create table junit_sem_analysis (a int, c string) partitioned by (b string) stored as RCFILE");
+    CommandProcessorResponse response = hcatDriver.run("alter table junit_sem_analysis change a a1 int");
     assertEquals(10, response.getResponseCode());
     assertTrue(response.getErrorMessage().contains("Operation not supported."));
 
-    response = howlDriver.run("alter table junit_sem_analysis change a a string");
+    response = hcatDriver.run("alter table junit_sem_analysis change a a string");
     assertEquals(10, response.getResponseCode());
     assertTrue(response.getErrorMessage().contains("Operation not supported."));
 
-    response = howlDriver.run("alter table junit_sem_analysis change a a int after c");
+    response = hcatDriver.run("alter table junit_sem_analysis change a a int after c");
     assertEquals(10, response.getResponseCode());
     assertTrue(response.getErrorMessage().contains("Operation not supported."));
-    howlDriver.run("drop table junit_sem_analysis");
+    hcatDriver.run("drop table junit_sem_analysis");
   }
 
   public void testAddReplaceCols() throws IOException, MetaException, TException, NoSuchObjectException, CommandNeedRetryException{
 
-    howlDriver.run("drop table junit_sem_analysis");
-    howlDriver.run("create table junit_sem_analysis (a int, c string) partitioned by (b string) stored as RCFILE");
-    CommandProcessorResponse response = howlDriver.run("alter table junit_sem_analysis replace columns (a1 tinyint)");
+    hcatDriver.run("drop table junit_sem_analysis");
+    hcatDriver.run("create table junit_sem_analysis (a int, c string) partitioned by (b string) stored as RCFILE");
+    CommandProcessorResponse response = hcatDriver.run("alter table junit_sem_analysis replace columns (a1 tinyint)");
     assertEquals(10, response.getResponseCode());
     assertTrue(response.getErrorMessage().contains("Operation not supported."));
 
-    response = howlDriver.run("alter table junit_sem_analysis add columns (d tinyint)");
+    response = hcatDriver.run("alter table junit_sem_analysis add columns (d tinyint)");
     assertEquals(0, response.getResponseCode());
     assertNull(response.getErrorMessage());
     Table tbl = msc.getTable(MetaStoreUtils.DEFAULT_DATABASE_NAME, tblName);
@@ -209,23 +209,23 @@ public class TestSemanticAnalysis extend
     assertTrue(cols.get(0).equals(new FieldSchema("a", "int", "from deserializer")));
     assertTrue(cols.get(1).equals(new FieldSchema("c", "string", "from deserializer")));
     assertTrue(cols.get(2).equals(new FieldSchema("d", "tinyint", null)));
-    howlDriver.run("drop table junit_sem_analysis");
+    hcatDriver.run("drop table junit_sem_analysis");
   }
 
   public void testAlterTblClusteredBy() throws CommandNeedRetryException{
 
-    howlDriver.run("drop table junit_sem_analysis");
-    howlDriver.run("create table junit_sem_analysis (a int) partitioned by (b string) stored as RCFILE");
-    CommandProcessorResponse response = howlDriver.run("alter table junit_sem_analysis clustered by (a) into 7 buckets");
+    hcatDriver.run("drop table junit_sem_analysis");
+    hcatDriver.run("create table junit_sem_analysis (a int) partitioned by (b string) stored as RCFILE");
+    CommandProcessorResponse response = hcatDriver.run("alter table junit_sem_analysis clustered by (a) into 7 buckets");
     assertEquals(10, response.getResponseCode());
     assertTrue(response.getErrorMessage().contains("Operation not supported."));
-    howlDriver.run("drop table junit_sem_analysis");
+    hcatDriver.run("drop table junit_sem_analysis");
   }
 
   public void testAlterTableSetFF() throws IOException, MetaException, TException, NoSuchObjectException, CommandNeedRetryException{
 
-    howlDriver.run("drop table junit_sem_analysis");
-    howlDriver.run("create table junit_sem_analysis (a int) partitioned by (b string) stored as RCFILE");
+    hcatDriver.run("drop table junit_sem_analysis");
+    hcatDriver.run("create table junit_sem_analysis (a int) partitioned by (b string) stored as RCFILE");
 
     Table tbl = msc.getTable(MetaStoreUtils.DEFAULT_DATABASE_NAME, tblName);
     assertEquals(RCFileInputFormat.class.getName(),tbl.getSd().getInputFormat());
@@ -235,9 +235,9 @@ public class TestSemanticAnalysis extend
     assertEquals(RCFileInputDriver.class.getName(), tblParams.get(HCatConstants.HCAT_ISD_CLASS));
     assertEquals(RCFileOutputDriver.class.getName(), tblParams.get(HCatConstants.HCAT_OSD_CLASS));
 
-    howlDriver.run("alter table junit_sem_analysis set fileformat INPUTFORMAT 'org.apache.hadoop.hive.ql.io.RCFileInputFormat' OUTPUTFORMAT " +
+    hcatDriver.run("alter table junit_sem_analysis set fileformat INPUTFORMAT 'org.apache.hadoop.hive.ql.io.RCFileInputFormat' OUTPUTFORMAT " +
         "'org.apache.hadoop.hive.ql.io.RCFileOutputFormat' inputdriver 'mydriver' outputdriver 'yourdriver'");
-    howlDriver.run("desc extended junit_sem_analysis");
+    hcatDriver.run("desc extended junit_sem_analysis");
 
     tbl = msc.getTable(MetaStoreUtils.DEFAULT_DATABASE_NAME, tblName);
     assertEquals(RCFileInputFormat.class.getName(),tbl.getSd().getInputFormat());
@@ -246,14 +246,14 @@ public class TestSemanticAnalysis extend
     assertEquals("mydriver", tblParams.get(HCatConstants.HCAT_ISD_CLASS));
     assertEquals("yourdriver", tblParams.get(HCatConstants.HCAT_OSD_CLASS));
 
-    howlDriver.run("drop table junit_sem_analysis");
+    hcatDriver.run("drop table junit_sem_analysis");
   }
 
   public void testAddPartFail() throws CommandNeedRetryException{
 
     hiveDriver.run("drop table junit_sem_analysis");
     hiveDriver.run("create table junit_sem_analysis (a int) partitioned by (b string) stored as RCFILE");
-    CommandProcessorResponse response = howlDriver.run("alter table junit_sem_analysis add partition (b='2') location '/some/path'");
+    CommandProcessorResponse response = hcatDriver.run("alter table junit_sem_analysis add partition (b='2') location '/some/path'");
     assertEquals(10, response.getResponseCode());
     assertTrue(response.getErrorMessage().contains("FAILED: Error in semantic analysis: Operation not supported. Partitions can be added only in a table created through HCatalog. " +
     		"It seems table junit_sem_analysis was not created through HCatalog."));
@@ -262,39 +262,39 @@ public class TestSemanticAnalysis extend
 
   public void testAddPartPass() throws IOException, CommandNeedRetryException{
 
-    howlDriver.run("drop table junit_sem_analysis");
-    howlDriver.run("create table junit_sem_analysis (a int) partitioned by (b string) stored as RCFILE");
-    CommandProcessorResponse response = howlDriver.run("alter table junit_sem_analysis add partition (b='2') location '/tmp'");
+    hcatDriver.run("drop table junit_sem_analysis");
+    hcatDriver.run("create table junit_sem_analysis (a int) partitioned by (b string) stored as RCFILE");
+    CommandProcessorResponse response = hcatDriver.run("alter table junit_sem_analysis add partition (b='2') location '/tmp'");
     assertEquals(0, response.getResponseCode());
     assertNull(response.getErrorMessage());
-    howlDriver.run("drop table junit_sem_analysis");
+    hcatDriver.run("drop table junit_sem_analysis");
   }
 
   public void testCTAS() throws CommandNeedRetryException{
-    howlDriver.run("drop table junit_sem_analysis");
+    hcatDriver.run("drop table junit_sem_analysis");
     query = "create table junit_sem_analysis (a int) as select * from tbl2";
-    CommandProcessorResponse response = howlDriver.run(query);
+    CommandProcessorResponse response = hcatDriver.run(query);
     assertEquals(10, response.getResponseCode());
     assertTrue(response.getErrorMessage().contains("FAILED: Error in semantic analysis: Operation not supported. Create table as Select is not a valid operation."));
-    howlDriver.run("drop table junit_sem_analysis");
+    hcatDriver.run("drop table junit_sem_analysis");
   }
 
   public void testStoredAs() throws CommandNeedRetryException{
-    howlDriver.run("drop table junit_sem_analysis");
+    hcatDriver.run("drop table junit_sem_analysis");
     query = "create table junit_sem_analysis (a int)";
-    CommandProcessorResponse response = howlDriver.run(query);
+    CommandProcessorResponse response = hcatDriver.run(query);
     assertEquals(10, response.getResponseCode());
     assertTrue(response.getErrorMessage().contains("FAILED: Error in semantic analysis: STORED AS specification is either incomplete or incorrect."));
-    howlDriver.run("drop table junit_sem_analysis");
+    hcatDriver.run("drop table junit_sem_analysis");
   }
 
   public void testAddDriverInfo() throws IOException, MetaException, TException, NoSuchObjectException, CommandNeedRetryException{
 
-    howlDriver.run("drop table junit_sem_analysis");
+    hcatDriver.run("drop table junit_sem_analysis");
     query =  "create table junit_sem_analysis (a int) partitioned by (b string)  stored as " +
     		"INPUTFORMAT 'org.apache.hadoop.hive.ql.io.RCFileInputFormat' OUTPUTFORMAT " +
     		"'org.apache.hadoop.hive.ql.io.RCFileOutputFormat' inputdriver 'mydriver' outputdriver 'yourdriver' ";
-    assertEquals(0,howlDriver.run(query).getResponseCode());
+    assertEquals(0,hcatDriver.run(query).getResponseCode());
 
     Table tbl = msc.getTable(MetaStoreUtils.DEFAULT_DATABASE_NAME, tblName);
     assertEquals(RCFileInputFormat.class.getName(),tbl.getSd().getInputFormat());
@@ -303,15 +303,15 @@ public class TestSemanticAnalysis extend
     assertEquals("mydriver", tblParams.get(HCatConstants.HCAT_ISD_CLASS));
     assertEquals("yourdriver", tblParams.get(HCatConstants.HCAT_OSD_CLASS));
 
-    howlDriver.run("drop table junit_sem_analysis");
+    hcatDriver.run("drop table junit_sem_analysis");
   }
 
   public void testInvalidateNonStringPartition() throws IOException, CommandNeedRetryException{
 
-    howlDriver.run("drop table junit_sem_analysis");
+    hcatDriver.run("drop table junit_sem_analysis");
     query =  "create table junit_sem_analysis (a int) partitioned by (b int)  stored as RCFILE";
 
-    CommandProcessorResponse response = howlDriver.run(query);
+    CommandProcessorResponse response = hcatDriver.run(query);
     assertEquals(10,response.getResponseCode());
     assertEquals("FAILED: Error in semantic analysis: Operation not supported. HCatalog only supports partition columns of type string. For column: b Found type: int",
         response.getErrorMessage());
@@ -320,10 +320,10 @@ public class TestSemanticAnalysis extend
 
   public void testInvalidateSeqFileStoredAs() throws IOException, CommandNeedRetryException{
 
-    howlDriver.run("drop table junit_sem_analysis");
+    hcatDriver.run("drop table junit_sem_analysis");
     query =  "create table junit_sem_analysis (a int) partitioned by (b string)  stored as SEQUENCEFILE";
 
-    CommandProcessorResponse response = howlDriver.run(query);
+    CommandProcessorResponse response = hcatDriver.run(query);
     assertEquals(10,response.getResponseCode());
     assertEquals("FAILED: Error in semantic analysis: Operation not supported. HCatalog doesn't support Sequence File by default yet. You may specify it through INPUT/OUTPUT storage drivers.",
         response.getErrorMessage());
@@ -332,10 +332,10 @@ public class TestSemanticAnalysis extend
 
   public void testInvalidateTextFileStoredAs() throws IOException, CommandNeedRetryException{
 
-    howlDriver.run("drop table junit_sem_analysis");
+    hcatDriver.run("drop table junit_sem_analysis");
     query =  "create table junit_sem_analysis (a int) partitioned by (b string)  stored as TEXTFILE";
 
-    CommandProcessorResponse response = howlDriver.run(query);
+    CommandProcessorResponse response = hcatDriver.run(query);
     assertEquals(10,response.getResponseCode());
     assertEquals("FAILED: Error in semantic analysis: Operation not supported. HCatalog doesn't support Text File by default yet. You may specify it through INPUT/OUTPUT storage drivers.",
         response.getErrorMessage());
@@ -344,10 +344,10 @@ public class TestSemanticAnalysis extend
 
   public void testInvalidateClusteredBy() throws IOException, CommandNeedRetryException{
 
-    howlDriver.run("drop table junit_sem_analysis");
+    hcatDriver.run("drop table junit_sem_analysis");
     query =  "create table junit_sem_analysis (a int) partitioned by (b string) clustered by (a) into 10 buckets stored as TEXTFILE";
 
-    CommandProcessorResponse response = howlDriver.run(query);
+    CommandProcessorResponse response = hcatDriver.run(query);
     assertEquals(10,response.getResponseCode());
     assertEquals("FAILED: Error in semantic analysis: Operation not supported. HCatalog doesn't allow Clustered By in create table.",
         response.getErrorMessage());
@@ -360,7 +360,7 @@ public class TestSemanticAnalysis extend
 
     hiveDriver.run(query);
     query = "create table like_table like junit_sem_analysis";
-    CommandProcessorResponse response = howlDriver.run(query);
+    CommandProcessorResponse response = hcatDriver.run(query);
     assertEquals(10,response.getResponseCode());
     assertEquals("FAILED: Error in semantic analysis: Operation not supported. CREATE TABLE LIKE is not supported.", response.getErrorMessage());
   }
@@ -368,18 +368,18 @@ public class TestSemanticAnalysis extend
   public void testCTLPass() throws IOException, MetaException, TException, NoSuchObjectException, CommandNeedRetryException{
 
     try{
-      howlDriver.run("drop table junit_sem_analysis");
+      hcatDriver.run("drop table junit_sem_analysis");
     }
     catch( Exception e){
       System.err.println(e.getMessage());
     }
     query =  "create table junit_sem_analysis (a int) partitioned by (b string) stored as RCFILE";
 
-    howlDriver.run(query);
+    hcatDriver.run(query);
     String likeTbl = "like_table";
-    howlDriver.run("drop table "+likeTbl);
+    hcatDriver.run("drop table "+likeTbl);
     query = "create table like_table like junit_sem_analysis";
-    CommandProcessorResponse resp = howlDriver.run(query);
+    CommandProcessorResponse resp = hcatDriver.run(query);
     assertEquals(10, resp.getResponseCode());
     assertEquals("FAILED: Error in semantic analysis: Operation not supported. CREATE TABLE LIKE is not supported.", resp.getErrorMessage());
 //    Table tbl = msc.getTable(MetaStoreUtils.DEFAULT_DATABASE_NAME, likeTbl);
@@ -390,20 +390,20 @@ public class TestSemanticAnalysis extend
 //    assertEquals("org.apache.hadoop.hive.ql.io.RCFileInputFormat",tbl.getSd().getInputFormat());
 //    assertEquals("org.apache.hadoop.hive.ql.io.RCFileOutputFormat",tbl.getSd().getOutputFormat());
 //    Map<String, String> tblParams = tbl.getParameters();
-//    assertEquals("org.apache.hadoop.hive.howl.rcfile.RCFileInputStorageDriver", tblParams.get("howl.isd"));
-//    assertEquals("org.apache.hadoop.hive.howl.rcfile.RCFileOutputStorageDriver", tblParams.get("howl.osd"));
+//    assertEquals("org.apache.hadoop.hive.hcat.rcfile.RCFileInputStorageDriver", tblParams.get("hcat.isd"));
+//    assertEquals("org.apache.hadoop.hive.hcat.rcfile.RCFileOutputStorageDriver", tblParams.get("hcat.osd"));
 //
-//    howlDriver.run("drop table junit_sem_analysis");
-//    howlDriver.run("drop table "+likeTbl);
+//    hcatDriver.run("drop table junit_sem_analysis");
+//    hcatDriver.run("drop table "+likeTbl);
   }
 
 // This test case currently fails, since add partitions don't inherit anything from tables.
 
 //  public void testAddPartInheritDrivers() throws MetaException, TException, NoSuchObjectException{
 //
-//    howlDriver.run("drop table "+tblName);
-//    howlDriver.run("create table junit_sem_analysis (a int) partitioned by (b string) stored as RCFILE");
-//    howlDriver.run("alter table "+tblName+" add partition (b='2010-10-10')");
+//    hcatDriver.run("drop table "+tblName);
+//    hcatDriver.run("create table junit_sem_analysis (a int) partitioned by (b string) stored as RCFILE");
+//    hcatDriver.run("alter table "+tblName+" add partition (b='2010-10-10')");
 //
 //    List<String> partVals = new ArrayList<String>(1);
 //    partVals.add("2010-10-10");

Modified: incubator/hcatalog/trunk/src/test/org/apache/hcatalog/cli/TestUseDatabase.java
URL: http://svn.apache.org/viewvc/incubator/hcatalog/trunk/src/test/org/apache/hcatalog/cli/TestUseDatabase.java?rev=1104614&r1=1104613&r2=1104614&view=diff
==============================================================================
--- incubator/hcatalog/trunk/src/test/org/apache/hcatalog/cli/TestUseDatabase.java (original)
+++ incubator/hcatalog/trunk/src/test/org/apache/hcatalog/cli/TestUseDatabase.java Tue May 17 22:50:49 2011
@@ -33,19 +33,19 @@ import org.apache.hcatalog.cli.SemanticA
 /* Unit test for GitHub Howl issue #3 */
 public class TestUseDatabase extends TestCase{
 
-  private Driver howlDriver;
+  private Driver hcatDriver;
 
   @Override
   protected void setUp() throws Exception {
 
-    HiveConf howlConf = new HiveConf(this.getClass());
-    howlConf.set(ConfVars.PREEXECHOOKS.varname, "");
-    howlConf.set(ConfVars.POSTEXECHOOKS.varname, "");
-    howlConf.set(ConfVars.HIVE_SUPPORT_CONCURRENCY.varname, "false");
-
-    howlConf.set(ConfVars.SEMANTIC_ANALYZER_HOOK.varname, HCatSemanticAnalyzer.class.getName());
-    howlDriver = new Driver(howlConf);
-    SessionState.start(new CliSessionState(howlConf));
+    HiveConf hcatConf = new HiveConf(this.getClass());
+    hcatConf.set(ConfVars.PREEXECHOOKS.varname, "");
+    hcatConf.set(ConfVars.POSTEXECHOOKS.varname, "");
+    hcatConf.set(ConfVars.HIVE_SUPPORT_CONCURRENCY.varname, "false");
+
+    hcatConf.set(ConfVars.SEMANTIC_ANALYZER_HOOK.varname, HCatSemanticAnalyzer.class.getName());
+    hcatDriver = new Driver(hcatConf);
+    SessionState.start(new CliSessionState(hcatConf));
   }
 
   String query;
@@ -54,23 +54,23 @@ public class TestUseDatabase extends Tes
 
   public void testAlterTablePass() throws IOException, CommandNeedRetryException{
 
-    howlDriver.run("create database " + dbName);
-    howlDriver.run("use " + dbName);
-    howlDriver.run("create table " + tblName + " (a int) partitioned by (b string) stored as RCFILE");
+    hcatDriver.run("create database " + dbName);
+    hcatDriver.run("use " + dbName);
+    hcatDriver.run("create table " + tblName + " (a int) partitioned by (b string) stored as RCFILE");
 
     CommandProcessorResponse response;
 
-    response = howlDriver.run("alter table " + tblName + " add partition (b='2') location '/tmp'");
+    response = hcatDriver.run("alter table " + tblName + " add partition (b='2') location '/tmp'");
     assertEquals(0, response.getResponseCode());
     assertNull(response.getErrorMessage());
 
-    response = howlDriver.run("alter table " + tblName + " set fileformat INPUTFORMAT 'org.apache.hadoop.hive.ql.io.RCFileInputFormat' OUTPUTFORMAT " +
+    response = hcatDriver.run("alter table " + tblName + " set fileformat INPUTFORMAT 'org.apache.hadoop.hive.ql.io.RCFileInputFormat' OUTPUTFORMAT " +
         "'org.apache.hadoop.hive.ql.io.RCFileOutputFormat' inputdriver 'mydriver' outputdriver 'yourdriver'");
     assertEquals(0, response.getResponseCode());
     assertNull(response.getErrorMessage());
 
-    howlDriver.run("drop table " + tblName);
-    howlDriver.run("drop database " + dbName);
+    hcatDriver.run("drop table " + tblName);
+    hcatDriver.run("drop database " + dbName);
   }
 
 }

Modified: incubator/hcatalog/trunk/src/test/org/apache/hcatalog/data/TestDefaultHCatRecord.java
URL: http://svn.apache.org/viewvc/incubator/hcatalog/trunk/src/test/org/apache/hcatalog/data/TestDefaultHCatRecord.java?rev=1104614&r1=1104613&r2=1104614&view=diff
==============================================================================
--- incubator/hcatalog/trunk/src/test/org/apache/hcatalog/data/TestDefaultHCatRecord.java (original)
+++ incubator/hcatalog/trunk/src/test/org/apache/hcatalog/data/TestDefaultHCatRecord.java Tue May 17 22:50:49 2011
@@ -90,7 +90,7 @@ public class TestDefaultHCatRecord exten
     rec_1.add( new Integer(789));
     rec_1.add( new Long(1000L));
     rec_1.add( new Double(5.3D));
-    rec_1.add( new String("howl and hadoop"));
+    rec_1.add( new String("hcat and hadoop"));
     rec_1.add( null);
     rec_1.add( "null");
 
@@ -102,7 +102,7 @@ public class TestDefaultHCatRecord exten
     rec_2.add( new Integer(789));
     rec_2.add( new Long(1000L));
     rec_2.add( new Double(5.3D));
-    rec_2.add( new String("howl and hadoop"));
+    rec_2.add( new String("hcat and hadoop"));
     rec_2.add( null);
     rec_2.add( "null");
     HCatRecord tup_2 = new DefaultHCatRecord(rec_2);
@@ -113,14 +113,14 @@ public class TestDefaultHCatRecord exten
     rec_3.add( new Integer(789));
     rec_3.add( new Long(1000L));
     rec_3.add( new Double(5.3D));
-    rec_3.add( new String("howl and hadoop"));
+    rec_3.add( new String("hcat and hadoop"));
     rec_3.add( null);
     List<Integer> innerList = new ArrayList<Integer>();
     innerList.add(314);
     innerList.add(007);
     rec_3.add( innerList);
     Map<Short, String> map = new HashMap<Short, String>(3);
-    map.put(new Short("2"), "howl is cool");
+    map.put(new Short("2"), "hcat is cool");
     map.put(new Short("3"), "is it?");
     map.put(new Short("4"), "or is it not?");
     rec_3.add(map);
@@ -133,12 +133,12 @@ public class TestDefaultHCatRecord exten
     rec_4.add( new Integer(789));
     rec_4.add( new Long(1000L));
     rec_4.add( new Double(5.3D));
-    rec_4.add( new String("howl and hadoop"));
+    rec_4.add( new String("hcat and hadoop"));
     rec_4.add( null);
     rec_4.add( "null");
 
     Map<Short, String> map2 = new HashMap<Short, String>(3);
-    map2.put(new Short("2"), "howl is cool");
+    map2.put(new Short("2"), "hcat is cool");
     map2.put(new Short("3"), "is it?");
     map2.put(new Short("4"), "or is it not?");
     rec_4.add(map2);

Modified: incubator/hcatalog/trunk/src/test/org/apache/hcatalog/mapreduce/HCatMapReduceTest.java
URL: http://svn.apache.org/viewvc/incubator/hcatalog/trunk/src/test/org/apache/hcatalog/mapreduce/HCatMapReduceTest.java?rev=1104614&r1=1104613&r2=1104614&view=diff
==============================================================================
--- incubator/hcatalog/trunk/src/test/org/apache/hcatalog/mapreduce/HCatMapReduceTest.java (original)
+++ incubator/hcatalog/trunk/src/test/org/apache/hcatalog/mapreduce/HCatMapReduceTest.java Tue May 17 22:50:49 2011
@@ -66,13 +66,13 @@ import org.apache.hcatalog.rcfile.RCFile
 import org.apache.hcatalog.rcfile.RCFileOutputDriver;
 
 /**
- * Test for HowlOutputFormat. Writes a partition using HowlOutputFormat and reads
+ * Test for HCatOutputFormat. Writes a partition using HCatOutputFormat and reads
  * it back using HCatInputFormat, checks the column values and counts.
  */
 public abstract class HCatMapReduceTest extends TestCase {
 
   protected String dbName = "default";
-  protected String tableName = "testHowlMapReduceTable";
+  protected String tableName = "testHCatMapReduceTable";
 
   protected String inputFormat = RCFileInputFormat.class.getName();
   protected String outputFormat = RCFileOutputFormat.class.getName();
@@ -249,14 +249,14 @@ public abstract class HCatMapReduceTest 
     MapCreate.writeCount = 0;
 
     Configuration conf = new Configuration();
-    Job job = new Job(conf, "howl mapreduce write test");
+    Job job = new Job(conf, "hcat mapreduce write test");
     job.setJarByClass(this.getClass());
     job.setMapperClass(HCatMapReduceTest.MapCreate.class);
 
     // input/output settings
     job.setInputFormatClass(TextInputFormat.class);
 
-    Path path = new Path(fs.getWorkingDirectory(), "mapred/testHowlMapReduceInput");
+    Path path = new Path(fs.getWorkingDirectory(), "mapred/testHCatMapReduceInput");
     createInputFile(path, writeCount);
 
     TextInputFormat.setInputPaths(job, path);
@@ -273,7 +273,7 @@ public abstract class HCatMapReduceTest 
 
     HCatOutputFormat.setSchema(job, new HCatSchema(partitionColumns));
 
-    //new HowlOutputCommitter(null).setupJob(job);
+    //new HCatOutputCommitter(null).setupJob(job);
     job.waitForCompletion(true);
     new HCatOutputCommitter(null).cleanupJob(job);
     Assert.assertEquals(writeCount, MapCreate.writeCount);
@@ -289,7 +289,7 @@ public abstract class HCatMapReduceTest 
     readRecords.clear();
 
     Configuration conf = new Configuration();
-    Job job = new Job(conf, "howl mapreduce read test");
+    Job job = new Job(conf, "hcat mapreduce read test");
     job.setJarByClass(this.getClass());
     job.setMapperClass(HCatMapReduceTest.MapRead.class);
 
@@ -306,7 +306,7 @@ public abstract class HCatMapReduceTest 
 
     job.setNumReduceTasks(0);
 
-    Path path = new Path(fs.getWorkingDirectory(), "mapred/testHowlMapReduceOutput");
+    Path path = new Path(fs.getWorkingDirectory(), "mapred/testHCatMapReduceOutput");
     if( fs.exists(path) ) {
       fs.delete(path, true);
     }
@@ -323,7 +323,7 @@ public abstract class HCatMapReduceTest 
   protected HCatSchema getTableSchema() throws Exception {
 
     Configuration conf = new Configuration();
-    Job job = new Job(conf, "howl mapreduce read schema test");
+    Job job = new Job(conf, "hcat mapreduce read schema test");
     job.setJarByClass(this.getClass());
 
     // input/output settings

Modified: incubator/hcatalog/trunk/src/test/org/apache/hcatalog/mapreduce/TestHCatHiveCompatibility.java
URL: http://svn.apache.org/viewvc/incubator/hcatalog/trunk/src/test/org/apache/hcatalog/mapreduce/TestHCatHiveCompatibility.java?rev=1104614&r1=1104613&r2=1104614&view=diff
==============================================================================
--- incubator/hcatalog/trunk/src/test/org/apache/hcatalog/mapreduce/TestHCatHiveCompatibility.java (original)
+++ incubator/hcatalog/trunk/src/test/org/apache/hcatalog/mapreduce/TestHCatHiveCompatibility.java Tue May 17 22:50:49 2011
@@ -89,7 +89,7 @@ public class TestHCatHiveCompatibility e
       throw new IOException("Failed to create table.");
     }
 
-    // assert that the table created has no howl instrumentation, and that we're still able to read it.
+    // assert that the table created has no hcat instrumentation, and that we're still able to read it.
     Table table = client.getTable("default", "junit_unparted_noisd");
     assertFalse(table.getParameters().containsKey(HCatConstants.HCAT_ISD_CLASS));
     assertTrue(table.getSd().getInputFormat().equals(HCatConstants.HIVE_RCFILE_IF_CLASS));
@@ -113,7 +113,7 @@ public class TestHCatHiveCompatibility e
     assertFalse(itr.hasNext());
     assertEquals(11, i);
 
-    // assert that the table created still has no howl instrumentation
+    // assert that the table created still has no hcat instrumentation
     Table table2 = client.getTable("default", "junit_unparted_noisd");
     assertFalse(table2.getParameters().containsKey(HCatConstants.HCAT_ISD_CLASS));
     assertTrue(table2.getSd().getInputFormat().equals(HCatConstants.HIVE_RCFILE_IF_CLASS));
@@ -130,7 +130,7 @@ public class TestHCatHiveCompatibility e
       throw new IOException("Failed to create table.");
     }
 
-    // assert that the table created has no howl instrumentation, and that we're still able to read it.
+    // assert that the table created has no hcat instrumentation, and that we're still able to read it.
     Table table = client.getTable("default", "junit_parted_noisd");
 
     assertFalse(table.getParameters().containsKey(HCatConstants.HCAT_ISD_CLASS));
@@ -156,12 +156,12 @@ public class TestHCatHiveCompatibility e
     assertFalse(itr.hasNext());
     assertEquals(11, i);
 
-    // assert that the table created still has no howl instrumentation
+    // assert that the table created still has no hcat instrumentation
     Table table2 = client.getTable("default", "junit_parted_noisd");
     assertFalse(table2.getParameters().containsKey(HCatConstants.HCAT_ISD_CLASS));
     assertTrue(table2.getSd().getInputFormat().equals(HCatConstants.HIVE_RCFILE_IF_CLASS));
 
-    // assert that there is one partition present, and it had howl instrumentation inserted when it was created.
+    // assert that there is one partition present, and it had hcat instrumentation inserted when it was created.
     Partition ptn = client.getPartition("default", "junit_parted_noisd", Arrays.asList("42"));
 
     assertNotNull(ptn);

Modified: incubator/hcatalog/trunk/src/test/org/apache/hcatalog/mapreduce/TestHCatNonPartitioned.java
URL: http://svn.apache.org/viewvc/incubator/hcatalog/trunk/src/test/org/apache/hcatalog/mapreduce/TestHCatNonPartitioned.java?rev=1104614&r1=1104613&r2=1104614&view=diff
==============================================================================
--- incubator/hcatalog/trunk/src/test/org/apache/hcatalog/mapreduce/TestHCatNonPartitioned.java (original)
+++ incubator/hcatalog/trunk/src/test/org/apache/hcatalog/mapreduce/TestHCatNonPartitioned.java Tue May 17 22:50:49 2011
@@ -42,7 +42,7 @@ public class TestHCatNonPartitioned exte
   protected void initialize() throws HCatException {
 
     dbName = null; //test if null dbName works ("default" is used)
-    tableName = "testHowlNonPartitionedTable";
+    tableName = "testHCatNonPartitionedTable";
 
     writeRecords = new ArrayList<HCatRecord>();
 
@@ -75,7 +75,7 @@ public class TestHCatNonPartitioned exte
   }
 
 
-  public void testHowlNonPartitionedTable() throws Exception {
+  public void testHCatNonPartitionedTable() throws Exception {
 
     Map<String, String> partitionMap = new HashMap<String, String>();
     runMRCreate(null, partitionColumns, writeRecords, 10);
@@ -113,7 +113,7 @@ public class TestHCatNonPartitioned exte
     hiveReadTest();
   }
 
-  //Test that data inserted through howloutputformat is readable from hive
+  //Test that data inserted through hcatoutputformat is readable from hive
   private void hiveReadTest() throws Exception {
 
     String query = "select * from " + tableName;

Modified: incubator/hcatalog/trunk/src/test/org/apache/hcatalog/mapreduce/TestHCatOutputFormat.java
URL: http://svn.apache.org/viewvc/incubator/hcatalog/trunk/src/test/org/apache/hcatalog/mapreduce/TestHCatOutputFormat.java?rev=1104614&r1=1104613&r2=1104614&view=diff
==============================================================================
--- incubator/hcatalog/trunk/src/test/org/apache/hcatalog/mapreduce/TestHCatOutputFormat.java (original)
+++ incubator/hcatalog/trunk/src/test/org/apache/hcatalog/mapreduce/TestHCatOutputFormat.java Tue May 17 22:50:49 2011
@@ -53,8 +53,8 @@ public class TestHCatOutputFormat extend
   private HiveMetaStoreClient client;
   private HiveConf hiveConf;
 
-  private static final String dbName = "howlOutputFormatTestDB";
-  private static final String tblName = "howlOutputFormatTestTable";
+  private static final String dbName = "hcatOutputFormatTestDB";
+  private static final String tblName = "hcatOutputFormatTestTable";
 
   @Override
   protected void setUp() throws Exception {

Modified: incubator/hcatalog/trunk/src/test/org/apache/hcatalog/mapreduce/TestHCatPartitioned.java
URL: http://svn.apache.org/viewvc/incubator/hcatalog/trunk/src/test/org/apache/hcatalog/mapreduce/TestHCatPartitioned.java?rev=1104614&r1=1104613&r2=1104614&view=diff
==============================================================================
--- incubator/hcatalog/trunk/src/test/org/apache/hcatalog/mapreduce/TestHCatPartitioned.java (original)
+++ incubator/hcatalog/trunk/src/test/org/apache/hcatalog/mapreduce/TestHCatPartitioned.java Tue May 17 22:50:49 2011
@@ -42,7 +42,7 @@ public class TestHCatPartitioned extends
   @Override
   protected void initialize() throws Exception {
 
-    tableName = "testHowlPartitionedTable";
+    tableName = "testHCatPartitionedTable";
     writeRecords = new ArrayList<HCatRecord>();
 
     for(int i = 0;i < 20;i++) {
@@ -75,7 +75,7 @@ public class TestHCatPartitioned extends
   }
 
 
-  public void testHowlPartitionedTable() throws Exception {
+  public void testHCatPartitionedTable() throws Exception {
 
     Map<String, String> partitionMap = new HashMap<String, String>();
     partitionMap.put("part1", "p1value1");
@@ -297,7 +297,7 @@ public class TestHCatPartitioned extends
     runMRRead(70);
   }
 
-  //Test that data inserted through howloutputformat is readable from hive
+  //Test that data inserted through hcatoutputformat is readable from hive
   private void hiveReadTest() throws Exception {
 
     String query = "select * from " + tableName;

Modified: incubator/hcatalog/trunk/src/test/org/apache/hcatalog/pig/TestHCatStorer.java
URL: http://svn.apache.org/viewvc/incubator/hcatalog/trunk/src/test/org/apache/hcatalog/pig/TestHCatStorer.java?rev=1104614&r1=1104613&r2=1104614&view=diff
==============================================================================
--- incubator/hcatalog/trunk/src/test/org/apache/hcatalog/pig/TestHCatStorer.java (original)
+++ incubator/hcatalog/trunk/src/test/org/apache/hcatalog/pig/TestHCatStorer.java Tue May 17 22:50:49 2011
@@ -78,7 +78,7 @@ public class TestHCatStorer extends Test
 //    }
 //
 //    MiniCluster.deleteFile(cluster, fileName);
-//    MiniCluster.createInputFile(cluster, fileName, new String[]{"test\t{([a#haddop,b#pig])}","data\t{([b#hive,a#howl])}"});
+//    MiniCluster.createInputFile(cluster, fileName, new String[]{"test\t{([a#haddop,b#pig])}","data\t{([b#hive,a#hcat])}"});
 //
 //    PigServer server = new PigServer(ExecType.LOCAL, props);
 //    UDFContext.getUDFContext().setClientSystemProps();
@@ -469,8 +469,8 @@ public class TestHCatStorer extends Test
   }
 
   MiniCluster.deleteFile(cluster, fileName);
-  MiniCluster.createInputFile(cluster, fileName, new String[]{"zookeeper\t(2)\t{(pig)}\t{(pnuts,hdfs)}\t{(hadoop),(howl)}",
-      "chubby\t(2)\t{(sawzall)}\t{(bigtable,gfs)}\t{(mapreduce),(howl)}"});
+  MiniCluster.createInputFile(cluster, fileName, new String[]{"zookeeper\t(2)\t{(pig)}\t{(pnuts,hdfs)}\t{(hadoop),(hcat)}",
+      "chubby\t(2)\t{(sawzall)}\t{(bigtable,gfs)}\t{(mapreduce),(hcat)}"});
 
   PigServer server = new PigServer(ExecType.LOCAL, props);
   UDFContext.getUDFContext().setClientSystemProps();
@@ -489,8 +489,8 @@ public class TestHCatStorer extends Test
   driver.getResults(res);
   driver.run("drop table junit_unparted");
   Iterator<String> itr = res.iterator();
-  assertEquals("zookeeper\t{\"a1\":2}\t[\"pig\"]\t[{\"s1\":\"pnuts\",\"s2\":\"hdfs\"}]\t[{\"s3\":\"hadoop\"},{\"s3\":\"howl\"}]", itr.next());
-  assertEquals("chubby\t{\"a1\":2}\t[\"sawzall\"]\t[{\"s1\":\"bigtable\",\"s2\":\"gfs\"}]\t[{\"s3\":\"mapreduce\"},{\"s3\":\"howl\"}]",itr.next());
+  assertEquals("zookeeper\t{\"a1\":2}\t[\"pig\"]\t[{\"s1\":\"pnuts\",\"s2\":\"hdfs\"}]\t[{\"s3\":\"hadoop\"},{\"s3\":\"hcat\"}]", itr.next());
+  assertEquals("chubby\t{\"a1\":2}\t[\"sawzall\"]\t[{\"s1\":\"bigtable\",\"s2\":\"gfs\"}]\t[{\"s3\":\"mapreduce\"},{\"s3\":\"hcat\"}]",itr.next());
  assertFalse(itr.hasNext());
 
   }
@@ -509,7 +509,7 @@ public class TestHCatStorer extends Test
     int LOOP_SIZE = 3;
     String[] input = new String[LOOP_SIZE*LOOP_SIZE];
     for(int i = 0; i < LOOP_SIZE*LOOP_SIZE; i++) {
-      input[i] = i + "\t" + i * 2.1f +"\t"+ i*1.1d + "\t" + i * 2L +"\t"+"lets howl";
+      input[i] = i + "\t" + i * 2.1f +"\t"+ i*1.1d + "\t" + i * 2L +"\t"+"lets hcat";
     }
 
     MiniCluster.createInputFile(cluster, fileName, input);

Modified: incubator/hcatalog/trunk/src/test/org/apache/hcatalog/pig/TestPigStorageDriver.java
URL: http://svn.apache.org/viewvc/incubator/hcatalog/trunk/src/test/org/apache/hcatalog/pig/TestPigStorageDriver.java?rev=1104614&r1=1104613&r2=1104614&view=diff
==============================================================================
--- incubator/hcatalog/trunk/src/test/org/apache/hcatalog/pig/TestPigStorageDriver.java (original)
+++ incubator/hcatalog/trunk/src/test/org/apache/hcatalog/pig/TestPigStorageDriver.java Tue May 17 22:50:49 2011
@@ -56,21 +56,21 @@ import org.apache.thrift.TException;
 
 public class TestPigStorageDriver extends TestCase {
 
-  private HiveConf howlConf;
-  private Driver howlDriver;
+  private HiveConf hcatConf;
+  private Driver hcatDriver;
   private HiveMetaStoreClient msc;
 
   @Override
   protected void setUp() throws Exception {
 
-    howlConf = new HiveConf(this.getClass());
-    howlConf.set(ConfVars.PREEXECHOOKS.varname, "");
-    howlConf.set(ConfVars.POSTEXECHOOKS.varname, "");
-    howlConf.set(ConfVars.HIVE_SUPPORT_CONCURRENCY.varname, "false");
-    howlConf.set(ConfVars.SEMANTIC_ANALYZER_HOOK.varname, HCatSemanticAnalyzer.class.getName());
-    howlDriver = new Driver(howlConf);
-    msc = new HiveMetaStoreClient(howlConf);
-    SessionState.start(new CliSessionState(howlConf));
+    hcatConf = new HiveConf(this.getClass());
+    hcatConf.set(ConfVars.PREEXECHOOKS.varname, "");
+    hcatConf.set(ConfVars.POSTEXECHOOKS.varname, "");
+    hcatConf.set(ConfVars.HIVE_SUPPORT_CONCURRENCY.varname, "false");
+    hcatConf.set(ConfVars.SEMANTIC_ANALYZER_HOOK.varname, HCatSemanticAnalyzer.class.getName());
+    hcatDriver = new Driver(hcatConf);
+    msc = new HiveMetaStoreClient(hcatConf);
+    SessionState.start(new CliSessionState(hcatConf));
     super.setUp();
   }
 
@@ -82,33 +82,33 @@ public class TestPigStorageDriver extend
   public void testPigStorageDriver() throws IOException, CommandNeedRetryException{
 
 
-    String fsLoc = howlConf.get("fs.default.name");
+    String fsLoc = hcatConf.get("fs.default.name");
     Path tblPath = new Path(fsLoc, "/tmp/test_pig/data");
     String anyExistingFileInCurDir = "ivy.xml";
-    tblPath.getFileSystem(howlConf).copyFromLocalFile(new Path(anyExistingFileInCurDir),tblPath);
+    tblPath.getFileSystem(hcatConf).copyFromLocalFile(new Path(anyExistingFileInCurDir),tblPath);
 
-    howlDriver.run("drop table junit_pigstorage");
+    hcatDriver.run("drop table junit_pigstorage");
     CommandProcessorResponse resp;
     String createTable = "create table junit_pigstorage (a string) partitioned by (b string) stored as RCFILE";
 
-    resp = howlDriver.run(createTable);
+    resp = hcatDriver.run(createTable);
     assertEquals(0, resp.getResponseCode());
     assertNull(resp.getErrorMessage());
 
-    resp = howlDriver.run("alter table junit_pigstorage add partition (b='2010-10-10') location '"+new Path(fsLoc, "/tmp/test_pig")+"'");
+    resp = hcatDriver.run("alter table junit_pigstorage add partition (b='2010-10-10') location '"+new Path(fsLoc, "/tmp/test_pig")+"'");
     assertEquals(0, resp.getResponseCode());
     assertNull(resp.getErrorMessage());
 
-    resp = howlDriver.run("alter table junit_pigstorage partition (b='2010-10-10') set fileformat inputformat '" + RCFileInputFormat.class.getName()
+    resp = hcatDriver.run("alter table junit_pigstorage partition (b='2010-10-10') set fileformat inputformat '" + RCFileInputFormat.class.getName()
         +"' outputformat '"+RCFileOutputFormat.class.getName()+"' inputdriver '"+PigStorageInputDriver.class.getName()+"' outputdriver 'non-existent'");
     assertEquals(0, resp.getResponseCode());
     assertNull(resp.getErrorMessage());
 
-    resp =  howlDriver.run("desc extended junit_pigstorage partition (b='2010-10-10')");
+    resp =  hcatDriver.run("desc extended junit_pigstorage partition (b='2010-10-10')");
     assertEquals(0, resp.getResponseCode());
     assertNull(resp.getErrorMessage());
 
-    PigServer server = new PigServer(ExecType.LOCAL, howlConf.getAllProperties());
+    PigServer server = new PigServer(ExecType.LOCAL, hcatConf.getAllProperties());
     UDFContext.getUDFContext().setClientSystemProps();
     server.registerQuery(" a = load 'junit_pigstorage' using "+HCatLoader.class.getName()+";");
     Iterator<Tuple> itr = server.openIterator("a");
@@ -131,26 +131,26 @@ public class TestPigStorageDriver extend
     }
     assertEquals(0,stream.available());
     stream.close();
-    howlDriver.run("drop table junit_pigstorage");
+    hcatDriver.run("drop table junit_pigstorage");
   }
 
   public void testDelim() throws MetaException, TException, UnknownTableException, NoSuchObjectException, InvalidOperationException, IOException, CommandNeedRetryException{
 
-    howlDriver.run("drop table junit_pigstorage_delim");
+    hcatDriver.run("drop table junit_pigstorage_delim");
 
     CommandProcessorResponse resp;
     String createTable = "create table junit_pigstorage_delim (a string) partitioned by (b string) stored as RCFILE";
 
-    resp = howlDriver.run(createTable);
+    resp = hcatDriver.run(createTable);
 
     assertEquals(0, resp.getResponseCode());
     assertNull(resp.getErrorMessage());
 
-    resp = howlDriver.run("alter table junit_pigstorage_delim add partition (b='2010-10-10')");
+    resp = hcatDriver.run("alter table junit_pigstorage_delim add partition (b='2010-10-10')");
     assertEquals(0, resp.getResponseCode());
     assertNull(resp.getErrorMessage());
 
-    resp = howlDriver.run("alter table junit_pigstorage_delim partition (b='2010-10-10') set fileformat inputformat '" + RCFileInputFormat.class.getName()
+    resp = hcatDriver.run("alter table junit_pigstorage_delim partition (b='2010-10-10') set fileformat inputformat '" + RCFileInputFormat.class.getName()
         +"' outputformat '"+RCFileOutputFormat.class.getName()+"' inputdriver '"+MyPigStorageDriver.class.getName()+"' outputdriver 'non-existent'");
 
     Partition part = msc.getPartition(MetaStoreUtils.DEFAULT_DATABASE_NAME, "junit_pigstorage_delim", "b=2010-10-10");
@@ -159,7 +159,7 @@ public class TestPigStorageDriver extend
 
     msc.alter_partition(MetaStoreUtils.DEFAULT_DATABASE_NAME, "junit_pigstorage_delim", part);
 
-    PigServer server = new PigServer(ExecType.LOCAL, howlConf.getAllProperties());
+    PigServer server = new PigServer(ExecType.LOCAL, hcatConf.getAllProperties());
     UDFContext.getUDFContext().setClientSystemProps();
     server.registerQuery(" a = load 'junit_pigstorage_delim' using "+HCatLoader.class.getName()+";");
     try{

Modified: incubator/hcatalog/trunk/src/test/org/apache/hcatalog/rcfile/TestRCFileOutputStorageDriver.java
URL: http://svn.apache.org/viewvc/incubator/hcatalog/trunk/src/test/org/apache/hcatalog/rcfile/TestRCFileOutputStorageDriver.java?rev=1104614&r1=1104613&r2=1104614&view=diff
==============================================================================
--- incubator/hcatalog/trunk/src/test/org/apache/hcatalog/rcfile/TestRCFileOutputStorageDriver.java (original)
+++ incubator/hcatalog/trunk/src/test/org/apache/hcatalog/rcfile/TestRCFileOutputStorageDriver.java Tue May 17 22:50:49 2011
@@ -60,7 +60,7 @@ public class TestRCFileOutputStorageDriv
       bytesWritable.set(i, cu);
     }
 
-    //Convert byte array to HowlRecord using isd, convert howlrecord back to byte array
+    //Convert byte array to HCatRecord using isd, convert hcatrecord back to byte array
     //using osd, compare the two arrays
     HCatRecord record = isd.convertToHCatRecord(null, bytesWritable);
 
@@ -77,7 +77,7 @@ public class TestRCFileOutputStorageDriv
   private byte[][] buildBytesArray() throws UnsupportedEncodingException {
     byte[][] bytes = {"123".getBytes("UTF-8"), "456".getBytes("UTF-8"),
         "789".getBytes("UTF-8"), "1000".getBytes("UTF-8"),
-        "5.3".getBytes("UTF-8"), "howl and hadoop".getBytes("UTF-8"),
+        "5.3".getBytes("UTF-8"), "hcat and hadoop".getBytes("UTF-8"),
         new byte[0], "\\N".getBytes("UTF-8") };
     return bytes;
   }



Mime
View raw message