hive-dev mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Shyam Sarkar <shyam_sar...@yahoo.com>
Subject Re: Eclipse run fails !!
Date Wed, 04 Feb 2009 01:28:22 GMT
Dear Prasad,

I did a clean and then performed build all for project hive. I am getting 10 errors and 1706 warnings. All errors are about "must override a superclass method". It seems to be compiler problem. I have added jre1.6.0_11 in build JRE. Why is the following problem coming ?

UDFMethodResolver is an interface::

public interface UDFMethodResolver {
  
  public Method getEvalMethod(List<Class<?>> argClasses) 
    throws AmbiguousMethodException;
}

Following method should override above method ::

public Method getEvalMethod(List<Class<?>> argClasses)
      throws AmbiguousMethodException {
    assert(argClasses.size() == 2);

    List<Class<?>> pClasses = null;
    if (argClasses.get(0) == Void.class ||
        argClasses.get(1) == Void.class) {
      pClasses = new ArrayList<Class<?>>();
      pClasses.add(Double.class);
      pClasses.add(Double.class);
    }
    else if (argClasses.get(0) == argClasses.get(1)) {
      pClasses = argClasses;
    }
    else if (argClasses.get(0) == java.sql.Date.class ||
             argClasses.get(1) == java.sql.Date.class) {
      pClasses = new ArrayList<Class<?>>();
      pClasses.add(java.sql.Date.class);
      pClasses.add(java.sql.Date.class);
    }
    else {
      pClasses = new ArrayList<Class<?>>();
      pClasses.add(Double.class);
      pClasses.add(Double.class);
    }

    Method udfMethod = null;

    for(Method m: Arrays.asList(udfClass.getMethods())) {
      if (m.getName().equals("evaluate")) {

        Class<?>[] argumentTypeInfos = m.getParameterTypes();

        boolean match = (argumentTypeInfos.length == pClasses.size());

        for(int i=0; i<pClasses.size() && match; i++) {
          Class<?> accepted = ObjectInspectorUtils.generalizePrimitive(argumentTypeInfos[i]);
          if (accepted != pClasses.get(i)) {
            match = false;
          }
        }

        if (match) {
          if (udfMethod != null) {
            throw new AmbiguousMethodException(udfClass, argClasses);
          }
          else {
            udfMethod = m;
          }
        }
      }
    }
    return udfMethod;      
  }

}



=====================Errors and Warnings=======================
Description	Resource	Path	Location	Type
The method add_partition(Partition) of type MetaStoreClient must override a superclass method	MetaStoreClient.java	hive/metastore/src/java/org/apache/hadoop/hive/metastore	line 466	Java Problem
The method getEvalMethod(List<Class<?>>) of type ComparisonOpMethodResolver must override a superclass method	ComparisonOpMethodResolver.java	hive/ql/src/java/org/apache/hadoop/hive/ql/exec	line 54	Java Problem
The method getEvalMethod(List<Class<?>>) of type NumericOpMethodResolver must override a superclass method	NumericOpMethodResolver.java	hive/ql/src/java/org/apache/hadoop/hive/ql/exec	line 52	Java Problem
The method getEvalMethod(List<Class<?>>) of type UDFIf.UDFIfMethodResolver must override a superclass method	UDFIf.java	hive/ql/src/java/org/apache/hadoop/hive/ql/udf	line 81	Java Problem
The method process(Node, NodeProcessorCtx, Object...) of type TypeCheckProcFactory.BoolExprProcessor must override a superclass method	TypeCheckProcFactory.java	hive/ql/src/java/org/apache/hadoop/hive/ql/parse	line 205	Java Problem
The method process(Node, NodeProcessorCtx, Object...) of type TypeCheckProcFactory.ColumnExprProcessor must override a superclass method	TypeCheckProcFactory.java	hive/ql/src/java/org/apache/hadoop/hive/ql/parse	line 245	Java Problem
The method process(Node, NodeProcessorCtx, Object...) of type TypeCheckProcFactory.DefaultExprProcessor must override a superclass method	TypeCheckProcFactory.java	hive/ql/src/java/org/apache/hadoop/hive/ql/parse	line 584	Java Problem
The method process(Node, NodeProcessorCtx, Object...) of type TypeCheckProcFactory.NullExprProcessor must override a superclass method	TypeCheckProcFactory.java	hive/ql/src/java/org/apache/hadoop/hive/ql/parse	line 94	Java Problem
The method process(Node, NodeProcessorCtx, Object...) of type TypeCheckProcFactory.NumExprProcessor must override a superclass method	TypeCheckProcFactory.java	hive/ql/src/java/org/apache/hadoop/hive/ql/parse	line 121	Java Problem
The method process(Node, NodeProcessorCtx, Object...) of type TypeCheckProcFactory.StrExprProcessor must override a superclass method	TypeCheckProcFactory.java	hive/ql/src/java/org/apache/hadoop/hive/ql/parse	line 163	Java Problem
AbstractList is a raw type. References to generic type AbstractList<E> should be parameterized	ComplexSerDeField.java	hive/serde/src/java/org/apache/hadoop/hive/serde	line 361	Java Problem
ArrayList is a raw type. References to generic type ArrayList<E> should be parameterized	ComplexSerDeField.java	hive/serde/src/java/org/apache/hadoop/hive/serde	line 131	Java Problem
ArrayList is a raw type. References to generic type ArrayList<E> should be parameterized	ComplexSerDeField.java	hive/serde/src/java/org/apache/hadoop/hive/serde	line 135	Java Problem
ArrayList is a raw type. References to generic type ArrayList<E> should be parameterized	ComplexSerDeField.java	hive/serde/src/java/org/apache/hadoop/hive/serde	line 139	Java Problem
ArrayList is a raw type. References to generic type ArrayList<E> should be parameterized	ComplexSerDeField.java	hive/serde/src/java/org/apache/hadoop/hive/serde	line 143	Java Problem
ArrayList is a raw type. References to generic type ArrayList<E> should be parameterized	ComplexSerDeField.java	hive/serde/src/java/org/apache/hadoop/hive/serde	line 143	Java Problem
ArrayList is a raw type. References to generic type ArrayList<E> should be parameterized	ComplexSerDeField.java	hive/serde/src/java/org/apache/hadoop/hive/serde	line 234	Java Problem
ArrayList is a raw type. References to generic type ArrayList<E> should be parameterized	ComplexSerDeField.java	hive/serde/src/java/org/apache/hadoop/hive/serde	line 306	Java Problem
ArrayList is a raw type. References to generic type ArrayList<E> should be parameterized	ComplexSerDeField.java	hive/serde/src/java/org/apache/hadoop/hive/serde	line 307	Java Problem
ArrayList is a raw type. References to generic type ArrayList<E> should be parameterized	ComplexSerDeField.java	hive/serde/src/java/org/apache/hadoop/hive/serde	line 370	Java Problem
ArrayList is a raw type. References to generic type ArrayList<E> should be parameterized	JJTthrift_grammarState.java	hive/serde/src/java/org/apache/hadoop/hive/serde2/dynamic_type	line 13	Java Problem
ArrayList is a raw type. References to generic type ArrayList<E> should be parameterized	JJTthrift_grammarState.java	hive/serde/src/java/org/apache/hadoop/hive/serde2/dynamic_type	line 14	Java Problem
ArrayList is a raw type. References to generic type ArrayList<E> should be parameterized	OptionsProcessor.java	hive/cli/src/java/org/apache/hadoop/hive/cli	line 76	Java Problem
ArrayList is a raw type. References to generic type ArrayList<E> should be parameterized	OptionsProcessor.java	hive/cli/src/java/org/apache/hadoop/hive/cli	line 76	Java Problem
ArrayList is a raw type. References to generic type ArrayList<E> should be parameterized	ScriptOperator.java	hive/ql/src/java/org/apache/hadoop/hive/ql/exec	line 397	Java Problem
ArrayList is a raw type. References to generic type ArrayList<E> should be parameterized	ScriptOperator.java	hive/ql/src/java/org/apache/hadoop/hive/ql/exec	line 397	Java Problem
ArrayList is a raw type. References to generic type ArrayList<E> should be parameterized	Utilities.java	hive/ql/src/java/org/apache/hadoop/hive/ql/exec	line 249	Java Problem
ArrayList is a raw type. References to generic type ArrayList<E> should be parameterized	Utilities.java	hive/ql/src/java/org/apache/hadoop/hive/ql/exec	line 250	Java Problem
ArrayList is a raw type. References to generic type ArrayList<E> should be parameterized	Utilities.java	hive/ql/src/java/org/apache/hadoop/hive/ql/exec	line 250	Java Problem
ArrayList is a raw type. References to generic type ArrayList<E> should be parameterized	thrift_grammar.java	hive/serde/src/java/org/apache/hadoop/hive/serde2/dynamic_type	line 2283	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	ByteStreamTypedSerDe.java	hive/serde/src/java/org/apache/hadoop/hive/serde	line 70	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	ColumnInfo.java	hive/ql/src/java/org/apache/hadoop/hive/ql/exec	line 56	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	ComplexSerDeField.java	hive/serde/src/java/org/apache/hadoop/hive/serde	line 158	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	ComplexSerDeField.java	hive/serde/src/java/org/apache/hadoop/hive/serde	line 186	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	ComplexSerDeField.java	hive/serde/src/java/org/apache/hadoop/hive/serde	line 193	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	ComplexSerDeField.java	hive/serde/src/java/org/apache/hadoop/hive/serde	line 200	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	ComplexSerDeField.java	hive/serde/src/java/org/apache/hadoop/hive/serde	line 377	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	ComplexSerDeField.java	hive/serde/src/java/org/apache/hadoop/hive/serde	line 389	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	ComplexSerDeField.java	hive/serde/src/java/org/apache/hadoop/hive/serde	line 404	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	ComplexSerDeField.java	hive/serde/src/java/org/apache/hadoop/hive/serde	line 412	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	ConstantTypedSerDeField.java	hive/serde/src/java/org/apache/hadoop/hive/serde	line 65	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	ConstantTypedSerDeField.java	hive/serde/src/java/org/apache/hadoop/hive/serde	line 70	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	ConstantTypedSerDeField.java	hive/serde/src/java/org/apache/hadoop/hive/serde	line 74	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	ConstantTypedSerDeField.java	hive/serde/src/java/org/apache/hadoop/hive/serde	line 78	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	DynamicSerDeStructBase.java	hive/serde/src/java/org/apache/hadoop/hive/serde2/dynamic_type	line 50	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	DynamicSerDeTypeBase.java	hive/serde/src/java/org/apache/hadoop/hive/serde2/dynamic_type	line 44	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	DynamicSerDeTypeBool.java	hive/serde/src/java/org/apache/hadoop/hive/serde2/dynamic_type	line 70	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	DynamicSerDeTypeDouble.java	hive/serde/src/java/org/apache/hadoop/hive/serde2/dynamic_type	line 67	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	DynamicSerDeTypeList.java	hive/serde/src/java/org/apache/hadoop/hive/serde2/dynamic_type	line 40	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	DynamicSerDeTypeMap.java	hive/serde/src/java/org/apache/hadoop/hive/serde2/dynamic_type	line 46	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	DynamicSerDeTypeMap.java	hive/serde/src/java/org/apache/hadoop/hive/serde2/dynamic_type	line 48	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	DynamicSerDeTypeMap.java	hive/serde/src/java/org/apache/hadoop/hive/serde2/dynamic_type	line 49	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	DynamicSerDeTypeSet.java	hive/serde/src/java/org/apache/hadoop/hive/serde2/dynamic_type	line 49	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	DynamicSerDeTypeSet.java	hive/serde/src/java/org/apache/hadoop/hive/serde2/dynamic_type	line 51	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	DynamicSerDeTypeString.java	hive/serde/src/java/org/apache/hadoop/hive/serde2/dynamic_type	line 45	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	DynamicSerDeTypei16.java	hive/serde/src/java/org/apache/hadoop/hive/serde2/dynamic_type	line 37	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	DynamicSerDeTypei32.java	hive/serde/src/java/org/apache/hadoop/hive/serde2/dynamic_type	line 65	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	DynamicSerDeTypei64.java	hive/serde/src/java/org/apache/hadoop/hive/serde2/dynamic_type	line 37	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	FetchTask.java	hive/ql/src/java/org/apache/hadoop/hive/ql/exec	line 112	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	FetchTask.java	hive/ql/src/java/org/apache/hadoop/hive/ql/exec	line 113	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	FetchTask.java	hive/ql/src/java/org/apache/hadoop/hive/ql/exec	line 115	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	HiveInputFormat.java	hive/ql/src/java/org/apache/hadoop/hive/ql/io	line 148	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	HiveInputFormat.java	hive/ql/src/java/org/apache/hadoop/hive/ql/io	line 149	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	HiveInputFormat.java	hive/ql/src/java/org/apache/hadoop/hive/ql/io	line 151	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	HiveInputFormat.java	hive/ql/src/java/org/apache/hadoop/hive/ql/io	line 173	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	HiveInputFormat.java	hive/ql/src/java/org/apache/hadoop/hive/ql/io	line 209	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	JuteSerDe.java	hive/serde/src/java/org/apache/hadoop/hive/serde/jute	line 79	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	JuteSerDe.java	hive/serde/src/java/org/apache/hadoop/hive/serde/jute	line 96	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	MapOperator.java	hive/ql/src/java/org/apache/hadoop/hive/ql/exec	line 96	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	MetadataTypedSerDeField.java	hive/serde/src/java/org/apache/hadoop/hive/serde/simple_meta	line 44	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	MetadataTypedSerDeField.java	hive/serde/src/java/org/apache/hadoop/hive/serde/simple_meta	line 93	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	MetadataTypedSerDeField.java	hive/serde/src/java/org/apache/hadoop/hive/serde/simple_meta	line 98	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	MetadataTypedSerDeField.java	hive/serde/src/java/org/apache/hadoop/hive/serde/simple_meta	line 102	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	MetadataTypedSerDeField.java	hive/serde/src/java/org/apache/hadoop/hive/serde/simple_meta	line 106	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	PrimitiveTypeInfo.java	hive/ql/src/java/org/apache/hadoop/hive/ql/typeinfo	line 39	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	PrimitiveTypeInfo.java	hive/ql/src/java/org/apache/hadoop/hive/ql/typeinfo	line 52	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	PrimitiveTypeInfo.java	hive/ql/src/java/org/apache/hadoop/hive/ql/typeinfo	line 58	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	PrimitiveTypeInfo.java	hive/ql/src/java/org/apache/hadoop/hive/ql/typeinfo	line 66	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	RandomDimension.java	hive/ql/src/java/org/apache/hadoop/hive/ql/metadata	line 32	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	ReflectionSerDeField.java	hive/serde/src/java/org/apache/hadoop/hive/serde	line 30	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	ReflectionSerDeField.java	hive/serde/src/java/org/apache/hadoop/hive/serde	line 31	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	ReflectionSerDeField.java	hive/serde/src/java/org/apache/hadoop/hive/serde	line 36	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	ReflectionSerDeField.java	hive/serde/src/java/org/apache/hadoop/hive/serde	line 37	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	ReflectionSerDeField.java	hive/serde/src/java/org/apache/hadoop/hive/serde	line 41	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	ReflectionSerDeField.java	hive/serde/src/java/org/apache/hadoop/hive/serde	line 81	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	ReflectionSerDeField.java	hive/serde/src/java/org/apache/hadoop/hive/serde	line 83	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	ReflectionSerDeField.java	hive/serde/src/java/org/apache/hadoop/hive/serde	line 84	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	ReflectionSerDeField.java	hive/serde/src/java/org/apache/hadoop/hive/serde	line 119	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	ReflectionSerDeField.java	hive/serde/src/java/org/apache/hadoop/hive/serde	line 123	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	ReflectionSerDeField.java	hive/serde/src/java/org/apache/hadoop/hive/serde	line 131	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	ReflectionSerDeField.java	hive/serde/src/java/org/apache/hadoop/hive/serde	line 139	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	SerDeField.java	hive/serde/src/java/org/apache/hadoop/hive/serde	line 68	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	SerDeField.java	hive/serde/src/java/org/apache/hadoop/hive/serde	line 73	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	SerDeField.java	hive/serde/src/java/org/apache/hadoop/hive/serde	line 78	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	SerDeField.java	hive/serde/src/java/org/apache/hadoop/hive/serde	line 83	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	SerDeUtils.java	hive/serde/src/java/org/apache/hadoop/hive/serde	line 33	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	SerDeUtils.java	hive/serde/src/java/org/apache/hadoop/hive/serde	line 33	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	SerDeUtils.java	hive/serde/src/java/org/apache/hadoop/hive/serde	line 35	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	SerDeUtils.java	hive/serde/src/java/org/apache/hadoop/hive/serde	line 43	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	SerDeUtils.java	hive/serde/src/java/org/apache/hadoop/hive/serde	line 85	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	SerDeUtils.java	hive/serde/src/java/org/apache/hadoop/hive/serde	line 122	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	SerDeUtils.java	hive/serde/src/java/org/apache/hadoop/hive/serde	line 123	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	TReflectionUtils.java	hive/serde/src/java/org/apache/hadoop/hive/serde	line 27	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	TReflectionUtils.java	hive/serde/src/java/org/apache/hadoop/hive/serde	line 32	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	TReflectionUtils.java	hive/serde/src/java/org/apache/hadoop/hive/serde	line 45	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	ThriftByteStreamTypedSerDe.java	hive/serde/src/java/org/apache/hadoop/hive/serde/thrift	line 100	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	ThriftByteStreamTypedSerDe.java	hive/serde/src/java/org/apache/hadoop/hive/serde/thrift	line 124	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	ThriftSerDeField.java	hive/serde/src/java/org/apache/hadoop/hive/serde/thrift	line 33	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	UDFIf.java	hive/ql/src/java/org/apache/hadoop/hive/ql/udf	line 67	Java Problem
Class is a raw type. References to generic type Class<T> should be parameterized	Utilities.java	hive/ql/src/java/org/apache/hadoop/hive/ql/exec	line 493	Java Problem

=======================================================================

Do you have any suggestion ?

Thanks,
Shyam



--- On Tue, 2/3/09, Prasad Chakka <prasad@facebook.com> wrote:

> From: Prasad Chakka <prasad@facebook.com>
> Subject: Re: Eclipse run fails !!
> To: "shyam_sarkar@yahoo.com" <shyam_sarkar@yahoo.com>, "hive-dev@hadoop.apache.org" <hive-dev@hadoop.apache.org>
> Date: Tuesday, February 3, 2009, 4:57 PM
> There are compilation errors in Hive project so that is why
> running tests is causing issues. Could you send what are the
> compilation errors?
> One of the errors should be on following line. It most
> probably a Eclipse and java issue. You can most probably
> remove the @override annotation and get successful
> compilation. If there are any more errors send them to us.
> 
> The method getEvalMethod(List<Class<?>>) of
> type NumericOpMethodResolver must override a superclass
> method
>         at
> org.apache.hadoop.hive.ql.exec.NumericOpMethodResolver.getEvalMethod(NumericOpMethodResolver.java:52)
> 
> 
> ________________________________
> From: Shyam Sarkar <shyam_sarkar@yahoo.com>
> Reply-To: <shyam_sarkar@yahoo.com>
> Date: Tue, 3 Feb 2009 16:51:47 -0800
> To: <hive-dev@hadoop.apache.org>, Prasad Chakka
> <prasad@facebook.com>
> Subject: Re: Eclipse run fails !!
> 
> Dear Prasad,
> 
> I followed your instructions with 0.17.2.1 hadoop version
> and changed jre to version 1.6_11. When I ran JUnit test, I
> still got the following message :
> 
> "Errors exist in required Project(s):
> hive
> Proceed with Launch ?"
> 
> When I launched I got following errors ::
> =================================== It is long
> ======================
> at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown
> Source)
>         at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown
> Source)
>         at java.lang.reflect.Method.invoke(Unknown Source)
>         at
> junit.framework.TestCase.runTest(TestCase.java:154)
>         at
> junit.framework.TestCase.runBare(TestCase.java:127)
>         at
> junit.framework.TestResult$1.protect(TestResult.java:106)
>         at
> junit.framework.TestResult.runProtected(TestResult.java:124)
>         at
> junit.framework.TestResult.run(TestResult.java:109)
>         at junit.framework.TestCase.run(TestCase.java:118)
>         at
> junit.framework.TestSuite.runTest(TestSuite.java:208)
>         at
> junit.framework.TestSuite.run(TestSuite.java:203)
>         at
> org.eclipse.jdt.internal.junit.runner.junit3.JUnit3TestReference.run(JUnit3TestReference.java:130)
>         at
> org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:38)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:460)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:673)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:386)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:196)
> Hive history
> file=/home/ssarkar/hive/ql/../build/ql/tmp/hive_job_log_ssarkar_200902031616_-1598728140.txt
> Begin query: sample6.q
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=12}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=12}
> OK
> Loading data to table srcbucket
> OK
> Loading data to table srcbucket
> OK
> Loading data to table src
> OK
> diff
> /home/ssarkar/hive/build/ql/test/logs/positive/sample6.q.out
> /home/ssarkar/hive/ql/src/test/results/compiler/parse/sample6.q.out
> Exception: Unresolved compilation problem:
>         The method
> getEvalMethod(List<Class<?>>) of type
> NumericOpMethodResolver must override a superclass method
> 
> java.lang.Error: Unresolved compilation problem:
>         The method
> getEvalMethod(List<Class<?>>) of type
> NumericOpMethodResolver must override a superclass method
> 
>         at
> org.apache.hadoop.hive.ql.exec.NumericOpMethodResolver.getEvalMethod(NumericOpMethodResolver.java:52)
>         at
> org.apache.hadoop.hive.ql.exec.FunctionRegistry.getUDFMethod(FunctionRegistry.java:274)
>         at
> org.apache.hadoop.hive.ql.parse.TypeCheckProcFactory$DefaultExprProcessor.getFuncExprNodeDesc(TypeCheckProcFactory.java:423)
>         at
> org.apache.hadoop.hive.ql.parse.TypeCheckProcFactory$DefaultExprProcessor.getFuncExprNodeDesc(TypeCheckProcFactory.java:379)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genSamplePredicate(SemanticAnalyzer.java:2872)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genTablePlan(SemanticAnalyzer.java:2985)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:3027)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:3229)
>         at
> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:71)
>         at
> org.apache.hadoop.hive.ql.QTestUtil.analyzeAST(QTestUtil.java:691)
>         at
> org.apache.hadoop.hive.ql.parse.TestParse.testParse_sample6(TestParse.java:1044)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
>         at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown
> Source)
>         at java.lang.reflect.Method.invoke(Unknown Source)
>         at
> junit.framework.TestCase.runTest(TestCase.java:154)
>         at
> junit.framework.TestCase.runBare(TestCase.java:127)
>         at
> junit.framework.TestResult$1.protect(TestResult.java:106)
>         at
> junit.framework.TestResult.runProtected(TestResult.java:124)
>         at
> junit.framework.TestResult.run(TestResult.java:109)
>         at junit.framework.TestCase.run(TestCase.java:118)
>         at
> junit.framework.TestSuite.runTest(TestSuite.java:208)
>         at
> junit.framework.TestSuite.run(TestSuite.java:203)
>         at
> org.eclipse.jdt.internal.junit.runner.junit3.JUnit3TestReference.run(JUnit3TestReference.java:130)
>         at
> org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:38)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:460)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:673)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:386)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:196)
> Hive history
> file=/home/ssarkar/hive/ql/../build/ql/tmp/hive_job_log_ssarkar_200902031616_1450160017.txt
> Begin query: sample7.q
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=12}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=12}
> OK
> Loading data to table srcbucket
> OK
> Loading data to table srcbucket
> OK
> Loading data to table src
> OK
> diff
> /home/ssarkar/hive/build/ql/test/logs/positive/sample7.q.out
> /home/ssarkar/hive/ql/src/test/results/compiler/parse/sample7.q.out
> Exception: Unresolved compilation problem:
>         The method
> getEvalMethod(List<Class<?>>) of type
> NumericOpMethodResolver must override a superclass method
> 
> java.lang.Error: Unresolved compilation problem:
>         The method
> getEvalMethod(List<Class<?>>) of type
> NumericOpMethodResolver must override a superclass method
> 
>         at
> org.apache.hadoop.hive.ql.exec.NumericOpMethodResolver.getEvalMethod(NumericOpMethodResolver.java:52)
>         at
> org.apache.hadoop.hive.ql.exec.FunctionRegistry.getUDFMethod(FunctionRegistry.java:274)
>         at
> org.apache.hadoop.hive.ql.parse.TypeCheckProcFactory$DefaultExprProcessor.getFuncExprNodeDesc(TypeCheckProcFactory.java:423)
>         at
> org.apache.hadoop.hive.ql.parse.TypeCheckProcFactory$DefaultExprProcessor.getFuncExprNodeDesc(TypeCheckProcFactory.java:379)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genSamplePredicate(SemanticAnalyzer.java:2872)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genTablePlan(SemanticAnalyzer.java:2985)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:3027)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:3229)
>         at
> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:71)
>         at
> org.apache.hadoop.hive.ql.QTestUtil.analyzeAST(QTestUtil.java:691)
>         at
> org.apache.hadoop.hive.ql.parse.TestParse.testParse_sample7(TestParse.java:1070)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
>         at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown
> Source)
>         at java.lang.reflect.Method.invoke(Unknown Source)
>         at
> junit.framework.TestCase.runTest(TestCase.java:154)
>         at
> junit.framework.TestCase.runBare(TestCase.java:127)
>         at
> junit.framework.TestResult$1.protect(TestResult.java:106)
>         at
> junit.framework.TestResult.runProtected(TestResult.java:124)
>         at
> junit.framework.TestResult.run(TestResult.java:109)
>         at junit.framework.TestCase.run(TestCase.java:118)
>         at
> junit.framework.TestSuite.runTest(TestSuite.java:208)
>         at
> junit.framework.TestSuite.run(TestSuite.java:203)
>         at
> org.eclipse.jdt.internal.junit.runner.junit3.JUnit3TestReference.run(JUnit3TestReference.java:130)
>         at
> org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:38)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:460)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:673)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:386)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:196)
> Hive history
> file=/home/ssarkar/hive/ql/../build/ql/tmp/hive_job_log_ssarkar_200902031616_514371634.txt
> Begin query: subq.q
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=12}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=12}
> OK
> Loading data to table srcbucket
> OK
> Loading data to table srcbucket
> OK
> Loading data to table src
> OK
> diff
> /home/ssarkar/hive/build/ql/test/logs/positive/subq.q.out
> /home/ssarkar/hive/ql/src/test/results/compiler/parse/subq.q.out
> java.lang.Error: Unresolved compilation problem:
>         The method process(Node, NodeProcessorCtx,
> Object...) of type TypeCheckProcFactory.StrExprProcessor
> must override a superclass method
> 
> Exception: Unresolved compilation problem:
>         The method process(Node, NodeProcessorCtx,
> Object...) of type TypeCheckProcFactory.StrExprProcessor
> must override a superclass method
> 
>         at
> org.apache.hadoop.hive.ql.parse.TypeCheckProcFactory$StrExprProcessor.process(TypeCheckProcFactory.java:163)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultRuleDispatcher.dispatch(DefaultRuleDispatcher.java:80)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.dispatch(DefaultGraphWalker.java:83)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.walk(DefaultGraphWalker.java:113)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.startWalking(DefaultGraphWalker.java:95)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genExprNodeDesc(SemanticAnalyzer.java:3311)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genFilterPlan(SemanticAnalyzer.java:904)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genBodyPlan(SemanticAnalyzer.java:2712)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:3048)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:3000)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:3021)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:3229)
>         at
> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:71)
>         at
> org.apache.hadoop.hive.ql.QTestUtil.analyzeAST(QTestUtil.java:691)
>         at
> org.apache.hadoop.hive.ql.parse.TestParse.testParse_subq(TestParse.java:1096)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
>         at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown
> Source)
>         at java.lang.reflect.Method.invoke(Unknown Source)
>         at
> junit.framework.TestCase.runTest(TestCase.java:154)
>         at
> junit.framework.TestCase.runBare(TestCase.java:127)
>         at
> junit.framework.TestResult$1.protect(TestResult.java:106)
>         at
> junit.framework.TestResult.runProtected(TestResult.java:124)
>         at
> junit.framework.TestResult.run(TestResult.java:109)
>         at junit.framework.TestCase.run(TestCase.java:118)
>         at
> junit.framework.TestSuite.runTest(TestSuite.java:208)
>         at
> junit.framework.TestSuite.run(TestSuite.java:203)
>         at
> org.eclipse.jdt.internal.junit.runner.junit3.JUnit3TestReference.run(JUnit3TestReference.java:130)
>         at
> org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:38)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:460)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:673)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:386)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:196)
> Hive history
> file=/home/ssarkar/hive/ql/../build/ql/tmp/hive_job_log_ssarkar_200902031616_520907971.txt
> Begin query: udf1.q
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=12}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=12}
> OK
> Loading data to table srcbucket
> OK
> Loading data to table srcbucket
> OK
> Loading data to table src
> OK
> diff
> /home/ssarkar/hive/build/ql/test/logs/positive/udf1.q.out
> /home/ssarkar/hive/ql/src/test/results/compiler/parse/udf1.q.out
> Exception: Unresolved compilation problem:
>         The method process(Node, NodeProcessorCtx,
> Object...) of type TypeCheckProcFactory.StrExprProcessor
> must override a superclass method
> 
> java.lang.Error: Unresolved compilation problem:
>         The method process(Node, NodeProcessorCtx,
> Object...) of type TypeCheckProcFactory.StrExprProcessor
> must override a superclass method
> 
>         at
> org.apache.hadoop.hive.ql.parse.TypeCheckProcFactory$StrExprProcessor.process(TypeCheckProcFactory.java:163)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultRuleDispatcher.dispatch(DefaultRuleDispatcher.java:80)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.dispatch(DefaultGraphWalker.java:83)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.walk(DefaultGraphWalker.java:113)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.startWalking(DefaultGraphWalker.java:95)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genExprNodeDesc(SemanticAnalyzer.java:3311)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genFilterPlan(SemanticAnalyzer.java:904)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genBodyPlan(SemanticAnalyzer.java:2712)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:3048)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:3229)
>         at
> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:71)
>         at
> org.apache.hadoop.hive.ql.QTestUtil.analyzeAST(QTestUtil.java:691)
>         at
> org.apache.hadoop.hive.ql.parse.TestParse.testParse_udf1(TestParse.java:1122)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
>         at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown
> Source)
>         at java.lang.reflect.Method.invoke(Unknown Source)
>         at
> junit.framework.TestCase.runTest(TestCase.java:154)
>         at
> junit.framework.TestCase.runBare(TestCase.java:127)
>         at
> junit.framework.TestResult$1.protect(TestResult.java:106)
>         at
> junit.framework.TestResult.runProtected(TestResult.java:124)
>         at
> junit.framework.TestResult.run(TestResult.java:109)
>         at junit.framework.TestCase.run(TestCase.java:118)
>         at
> junit.framework.TestSuite.runTest(TestSuite.java:208)
>         at
> junit.framework.TestSuite.run(TestSuite.java:203)
>         at
> org.eclipse.jdt.internal.junit.runner.junit3.JUnit3TestReference.run(JUnit3TestReference.java:130)
>         at
> org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:38)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:460)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:673)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:386)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:196)
> Hive history
> file=/home/ssarkar/hive/ql/../build/ql/tmp/hive_job_log_ssarkar_200902031616_656206857.txt
> Begin query: udf4.q
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=12}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=12}
> OK
> Loading data to table srcbucket
> OK
> Loading data to table srcbucket
> OK
> Loading data to table src
> OK
> diff
> /home/ssarkar/hive/build/ql/test/logs/positive/udf4.q.out
> /home/ssarkar/hive/ql/src/test/results/compiler/parse/udf4.q.out
> Exception: Unresolved compilation problem:
>         The method process(Node, NodeProcessorCtx,
> Object...) of type TypeCheckProcFactory.StrExprProcessor
> must override a superclass method
> java.lang.Error: Unresolved compilation problem:
>         The method process(Node, NodeProcessorCtx,
> Object...) of type TypeCheckProcFactory.StrExprProcessor
> must override a superclass method
> 
>         at
> org.apache.hadoop.hive.ql.parse.TypeCheckProcFactory$StrExprProcessor.process(TypeCheckProcFactory.java:163)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultRuleDispatcher.dispatch(DefaultRuleDispatcher.java:80)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.dispatch(DefaultGraphWalker.java:83)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.walk(DefaultGraphWalker.java:113)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.startWalking(DefaultGraphWalker.java:95)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genExprNodeDesc(SemanticAnalyzer.java:3311)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genSelectPlan(SemanticAnalyzer.java:1167)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genBodyPlan(SemanticAnalyzer.java:2724)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:3048)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:3229)
>         at
> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:71)
>         at
> org.apache.hadoop.hive.ql.QTestUtil.analyzeAST(QTestUtil.java:691)
>         at
> org.apache.hadoop.hive.ql.parse.TestParse.testParse_udf4(TestParse.java:1148)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
>         at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown
> Source)
>         at java.lang.reflect.Method.invoke(Unknown Source)
>         at
> junit.framework.TestCase.runTest(TestCase.java:154)
>         at
> junit.framework.TestCase.runBare(TestCase.java:127)
>         at
> junit.framework.TestResult$1.protect(TestResult.java:106)
>         at
> junit.framework.TestResult.runProtected(TestResult.java:124)
>         at
> junit.framework.TestResult.run(TestResult.java:109)
>         at junit.framework.TestCase.run(TestCase.java:118)
>         at
> junit.framework.TestSuite.runTest(TestSuite.java:208)
>         at
> junit.framework.TestSuite.run(TestSuite.java:203)
>         at
> org.eclipse.jdt.internal.junit.runner.junit3.JUnit3TestReference.run(JUnit3TestReference.java:130)
>         at
> org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:38)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:460)
> 
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:673)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:386)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:196)
> Hive history
> file=/home/ssarkar/hive/ql/../build/ql/tmp/hive_job_log_ssarkar_200902031616_545867528.txt
> Begin query: udf6.q
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=12}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=12}
> OK
> Loading data to table srcbucket
> OK
> Loading data to table srcbucket
> OK
> Loading data to table src
> OK
> diff
> /home/ssarkar/hive/build/ql/test/logs/positive/udf6.q.out
> /home/ssarkar/hive/ql/src/test/results/compiler/parse/udf6.q.out
> Exception: Unresolved compilation problem:
>         The method process(Node, NodeProcessorCtx,
> Object...) of type TypeCheckProcFactory.StrExprProcessor
> must override a superclass method
> 
> java.lang.Error: Unresolved compilation problem:
>         The method process(Node, NodeProcessorCtx,
> Object...) of type TypeCheckProcFactory.StrExprProcessor
> must override a superclass method
> 
>         at
> org.apache.hadoop.hive.ql.parse.TypeCheckProcFactory$StrExprProcessor.process(TypeCheckProcFactory.java:163)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultRuleDispatcher.dispatch(DefaultRuleDispatcher.java:80)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.dispatch(DefaultGraphWalker.java:83)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.walk(DefaultGraphWalker.java:113)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.startWalking(DefaultGraphWalker.java:95)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genExprNodeDesc(SemanticAnalyzer.java:3311)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genSelectPlan(SemanticAnalyzer.java:1167)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genBodyPlan(SemanticAnalyzer.java:2724)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:3048)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:3229)
>         at
> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:71)
>         at
> org.apache.hadoop.hive.ql.QTestUtil.analyzeAST(QTestUtil.java:691)
>         at
> org.apache.hadoop.hive.ql.parse.TestParse.testParse_udf6(TestParse.java:1174)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
>         at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown
> Source)
>         at java.lang.reflect.Method.invoke(Unknown Source)
>         at
> junit.framework.TestCase.runTest(TestCase.java:154)
>         at
> junit.framework.TestCase.runBare(TestCase.java:127)
>         at
> junit.framework.TestResult$1.protect(TestResult.java:106)
>         at
> junit.framework.TestResult.runProtected(TestResult.java:124)
>         at
> junit.framework.TestResult.run(TestResult.java:109)
>         at junit.framework.TestCase.run(TestCase.java:118)
>         at
> junit.framework.TestSuite.runTest(TestSuite.java:208)
>         at
> junit.framework.TestSuite.run(TestSuite.java:203)
>         at
> org.eclipse.jdt.internal.junit.runner.junit3.JUnit3TestReference.run(JUnit3TestReference.java:130)
>         at
> org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:38)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:460)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:673)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:386)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:196)
> Hive history
> file=/home/ssarkar/hive/ql/../build/ql/tmp/hive_job_log_ssarkar_200902031616_-1947338661.txt
> Begin query: union.q
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=12}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=12}
> OK
> Loading data to table srcbucket
> OK
> Loading data to table srcbucket
> OK
> Loading data to table src
> OK
> diff
> /home/ssarkar/hive/build/ql/test/logs/positive/union.q.out
> /home/ssarkar/hive/ql/src/test/results/compiler/parse/union.q.out
> Exception: Unresolved compilation problem:
>         The method process(Node, NodeProcessorCtx,
> Object...) of type TypeCheckProcFactory.StrExprProcessor
> must override a superclass method
> 
> java.lang.Error: Unresolved compilation problem:
>         The method process(Node, NodeProcessorCtx,
> Object...) of type TypeCheckProcFactory.StrExprProcessor
> must override a superclass method
> 
>         at
> org.apache.hadoop.hive.ql.parse.TypeCheckProcFactory$StrExprProcessor.process(TypeCheckProcFactory.java:163)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultRuleDispatcher.dispatch(DefaultRuleDispatcher.java:80)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.dispatch(DefaultGraphWalker.java:83)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.walk(DefaultGraphWalker.java:113)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.startWalking(DefaultGraphWalker.java:95)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genExprNodeDesc(SemanticAnalyzer.java:3311)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genFilterPlan(SemanticAnalyzer.java:904)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genBodyPlan(SemanticAnalyzer.java:2712)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:3048)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:3000)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:3003)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:3021)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:3229)
>         at
> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:71)
>         at
> org.apache.hadoop.hive.ql.QTestUtil.analyzeAST(QTestUtil.java:691)
>         at
> org.apache.hadoop.hive.ql.parse.TestParse.testParse_union(TestParse.java:1200)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
>         at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown
> Source)
>         at java.lang.reflect.Method.invoke(Unknown Source)
>         at
> junit.framework.TestCase.runTest(TestCase.java:154)
>         at
> junit.framework.TestCase.runBare(TestCase.java:127)
>         at
> junit.framework.TestResult$1.protect(TestResult.java:106)
>         at
> junit.framework.TestResult.runProtected(TestResult.java:124)
>         at
> junit.framework.TestResult.run(TestResult.java:109)
>         at junit.framework.TestCase.run(TestCase.java:118)
>         at
> junit.framework.TestSuite.runTest(TestSuite.java:208)
>         at
> junit.framework.TestSuite.run(TestSuite.java:203)
>         at
> org.eclipse.jdt.internal.junit.runner.junit3.JUnit3TestReference.run(JUnit3TestReference.java:130)
>         at
> org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:38)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:460)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:673)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:386)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:196)
> Table doesnotexist does not exist
> Testing Filter Operator
> java.lang.Error: Unresolved compilation problem:
>         The method
> getEvalMethod(List<Class<?>>) of type
> ComparisonOpMethodResolver must override a superclass method
> 
>         at
> org.apache.hadoop.hive.ql.exec.ComparisonOpMethodResolver.getEvalMethod(ComparisonOpMethodResolver.java:54)
>         at
> org.apache.hadoop.hive.ql.exec.FunctionRegistry.getUDFMethod(FunctionRegistry.java:274)
>         at
> org.apache.hadoop.hive.ql.parse.TypeCheckProcFactory$DefaultExprProcessor.getFuncExprNodeDesc(TypeCheckProcFactory.java:423)
>         at
> org.apache.hadoop.hive.ql.parse.TypeCheckProcFactory$DefaultExprProcessor.getFuncExprNodeDesc(TypeCheckProcFactory.java:379)
>         at
> org.apache.hadoop.hive.ql.exec.TestOperators.testBaseFilterOperator(TestOperators.java:79)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
>         at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown
> Source)
>         at java.lang.reflect.Method.invoke(Unknown Source)
>         at
> junit.framework.TestCase.runTest(TestCase.java:154)
>         at
> junit.framework.TestCase.runBare(TestCase.java:127)
>         at
> junit.framework.TestResult$1.protect(TestResult.java:106)
>         at
> junit.framework.TestResult.runProtected(TestResult.java:124)
>         at
> junit.framework.TestResult.run(TestResult.java:109)
>         at junit.framework.TestCase.run(TestCase.java:118)
>         at
> junit.framework.TestSuite.runTest(TestSuite.java:208)
>         at
> junit.framework.TestSuite.run(TestSuite.java:203)
>         at
> org.eclipse.jdt.internal.junit.runner.junit3.JUnit3TestReference.run(JUnit3TestReference.java:130)
>         at
> org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:38)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:460)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:673)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:386)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:196)
> Testing FileSink Operator
> FileSink Operator ok
> Testing Script Operator
> [0] io.o=[1, 01]
> [0]
> io.oi=org.apache.hadoop.hive.serde2.objectinspector.StandardStructObjectInspector@acb988
> [1] io.o=[2, 11]
> [1]
> io.oi=org.apache.hadoop.hive.serde2.objectinspector.StandardStructObjectInspector@acb988
> [2] io.o=[3, 21]
> [2]
> io.oi=org.apache.hadoop.hive.serde2.objectinspector.StandardStructObjectInspector@acb988
> [3] io.o=[4, 31]
> [3]
> io.oi=org.apache.hadoop.hive.serde2.objectinspector.StandardStructObjectInspector@acb988
> [4] io.o=[5, 41]
> [4]
> io.oi=org.apache.hadoop.hive.serde2.objectinspector.StandardStructObjectInspector@acb988
> Script Operator ok
> Testing Map Operator
> io1.o.toString() = [[0, 1, 2]]
> io2.o.toString() = [[0, 1, 2]]
> answer.toString() = [[0, 1, 2]]
> io1.o.toString() = [[1, 2, 3]]
> io2.o.toString() = [[1, 2, 3]]
> answer.toString() = [[1, 2, 3]]
> io1.o.toString() = [[2, 3, 4]]
> io2.o.toString() = [[2, 3, 4]]
> answer.toString() = [[2, 3, 4]]
> io1.o.toString() = [[3, 4, 5]]
> io2.o.toString() = [[3, 4, 5]]
> answer.toString() = [[3, 4, 5]]
> io1.o.toString() = [[4, 5, 6]]
> io2.o.toString() = [[4, 5, 6]]
> answer.toString() = [[4, 5, 6]]
> Map Operator ok
> JEXL library test ok
> Evaluating 1 + 2 for 10000000 times
> Evaluation finished: 0.562 seconds, 0.056 seconds/million
> call.
> Evaluating __udf__concat.evaluate("1",
> "2") for 1000000 times
> Evaluation finished: 1.028 seconds, 1.028 seconds/million
> call.
> Hive history
> file=/home/ssarkar/hive/ql/../build/ql/tmp/hive_job_log_ssarkar_200902031616_1713747826.txt
> java.io.FileNotFoundException: join1.q (No such file or
> directory)
>         at java.io.FileInputStream.open(Native Method)
>         at java.io.FileInputStream.<init>(Unknown
> Source)
>         at
> org.apache.hadoop.hive.ql.QTestUtil.addFile(QTestUtil.java:188)
>         at
> org.apache.hadoop.hive.ql.QTestUtil.queryListRunner(QTestUtil.java:751)
>         at
> org.apache.hadoop.hive.ql.TestMTQueries.testMTQueries1(TestMTQueries.java:51)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
>         at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown
> Source)
>         at java.lang.reflect.Method.invoke(Unknown Source)
>         at
> junit.framework.TestCase.runTest(TestCase.java:154)
>         at
> junit.framework.TestCase.runBare(TestCase.java:127)
>         at
> junit.framework.TestResult$1.protect(TestResult.java:106)
>         at
> junit.framework.TestResult.runProtected(TestResult.java:124)
>         at
> junit.framework.TestResult.run(TestResult.java:109)
>         at junit.framework.TestCase.run(TestCase.java:118)
>         at
> junit.framework.TestSuite.runTest(TestSuite.java:208)
>         at
> junit.framework.TestSuite.run(TestSuite.java:203)
>         at
> org.eclipse.jdt.internal.junit.runner.junit3.JUnit3TestReference.run(JUnit3TestReference.java:130)
>         at
> org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:38)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:460)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:673)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:386)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:196)
> ExprNodeFuncEvaluator ok
> ExprNodeColumnEvaluator ok
> testExprNodeConversionEvaluator ok
> java.lang.Error: Unresolved compilation problem:
>         The method
> getEvalMethod(List<Class<?>>) of type
> NumericOpMethodResolver must override a superclass method
> 
>         at
> org.apache.hadoop.hive.ql.exec.NumericOpMethodResolver.getEvalMethod(NumericOpMethodResolver.java:52)
>         at
> org.apache.hadoop.hive.ql.exec.FunctionRegistry.getUDFMethod(FunctionRegistry.java:274)
>         at
> org.apache.hadoop.hive.ql.parse.TypeCheckProcFactory$DefaultExprProcessor.getFuncExprNodeDesc(TypeCheckProcFactory.java:423)
>         at
> org.apache.hadoop.hive.ql.parse.TypeCheckProcFactory$DefaultExprProcessor.getFuncExprNodeDesc(TypeCheckProcFactory.java:379)
>         at
> org.apache.hadoop.hive.ql.exec.TestExpressionEvaluator.testExprNodeSpeed(TestExpressionEvaluator.java:168)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
>         at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown
> Source)
>         at java.lang.reflect.Method.invoke(Unknown Source)
>         at
> junit.framework.TestCase.runTest(TestCase.java:154)
>         at
> junit.framework.TestCase.runBare(TestCase.java:127)
>         at
> junit.framework.TestResult$1.protect(TestResult.java:106)
>         at
> junit.framework.TestResult.runProtected(TestResult.java:124)
>         at
> junit.framework.TestResult.run(TestResult.java:109)
>         at junit.framework.TestCase.run(TestCase.java:118)
>         at
> junit.framework.TestSuite.runTest(TestSuite.java:208)
>         at
> junit.framework.TestSuite.run(TestSuite.java:203)
>         at
> org.eclipse.jdt.internal.junit.runner.junit3.JUnit3TestReference.run(JUnit3TestReference.java:130)
>         at
> org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:38)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:460)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:673)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:386)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:196)
> input struct = [234, [firstString, secondString],
> {firstKey=1, secondKey=2}, -234, 1.0, -2.5]
> Testing protocol:
> org.apache.hadoop.hive.serde2.thrift.TBinarySortableProtocol
> TypeName =
> struct{_hello:int,2bye:array<string>,another:map<string,int>,nhello:int,d:double,nd:double}
> bytes
> =x01x80x00x00xeax01x80x00x00x02x01x66x69x72x73x74x53x74x72x69x6ex67x00x01x73x65x63x6fx6ex64x53x74x72x69x6ex67x00x01x80x00x00x02x01x66x69x72x73x74x4bx65x79x00x01x80x00x00x01x01x73x65x63x6fx6ex64x4bx65x79x00x01x80x00x00x02x01x7fxffxffx16x01xbfxf0x00x00x00x00x00x00x01x3fxfbxffxffxffxffxffxff
> o class = class java.util.ArrayList
> o size = 6
> o[0] class = class java.lang.Integer
> o[1] class = class java.util.ArrayList
> o[2] class = class java.util.HashMap
> o = [234, [firstString, secondString], {firstKey=1,
> secondKey=2}, -234, 1.0, -2.5]
> Testing protocol:
> org.apache.hadoop.hive.serde2.thrift.TBinarySortableProtocol
> TypeName =
> struct{_hello:int,2bye:array<string>,another:map<string,int>,nhello:int,d:double,nd:double}
> bytes
> =xfex7fxffxffx15xfex7fxffxffxfdxfex99x96x8dx8cx8bxacx8bx8dx96x91x98xffxfex8cx9ax9cx90x91x9bxacx8bx8dx96x91x98xffxfex7fxffxffxfdxfex99x96x8dx8cx8bxb4x9ax86xffxfex7fxffxffxfexfex8cx9ax9cx90x91x9bxb4x9ax86xffxfex7fxffxffxfdxfex80x00x00xe9xfex40x0fxffxffxffxffxffxffxfexc0x04x00x00x00x00x00x00
> o class = class java.util.ArrayList
> o size = 6
> o[0] class = class java.lang.Integer
> o[1] class = class java.util.ArrayList
> o[2] class = class java.util.HashMap
> o = [234, [firstString, secondString], {firstKey=1,
> secondKey=2}, -234, 1.0, -2.5]
> Testing protocol:
> com.facebook.thrift.protocol.TBinaryProtocol
> TypeName =
> struct{_hello:int,2bye:array<string>,another:map<string,int>,nhello:int,d:double,nd:double}
> bytes
> =x08xffxffx00x00x00xeax0fxffxfex0bx00x00x00x02x00x00x00x0bx66x69x72x73x74x53x74x72x69x6ex67x00x00x00x0cx73x65x63x6fx6ex64x53x74x72x69x6ex67x0dxffxfdx0bx08x00x00x00x02x00x00x00x08x66x69x72x73x74x4bx65x79x00x00x00x01x00x00x00x09x73x65x63x6fx6ex64x4bx65x79x00x00x00x02x08xffxfcxffxffxffx16x04xffxfbx3fxf0x00x00x00x00x00x00x04xffxfaxc0x04x00x00x00x00x00x00x00
> o class = class java.util.ArrayList
> o size = 6
> o[0] class = class java.lang.Integer
> o[1] class = class java.util.ArrayList
> o[2] class = class java.util.HashMap
> o = [234, [firstString, secondString], {firstKey=1,
> secondKey=2}, -234, 1.0, -2.5]
> Testing protocol:
> com.facebook.thrift.protocol.TJSONProtocol
> TypeName =
> struct{_hello:int,2bye:array<string>,another:map<string,int>,nhello:int,d:double,nd:double}
> bytes
> =x7bx22x2dx31x22x3ax7bx22x69x33x32x22x3ax32x33x34x7dx2cx22x2dx32x22x3ax7bx22x6cx73x74x22x3ax5bx22x73x74x72x22x2cx32x2cx22x66x69x72x73x74x53x74x72x69x6ex67x22x2cx22x73x65x63x6fx6ex64x53x74x72x69x6ex67x22x5dx7dx2cx22x2dx33x22x3ax7bx22x6dx61x70x22x3ax5bx22x73x74x72x22x2cx22x69x33x32x22x2cx32x2cx7bx22x66x69x72x73x74x4bx65x79x22x3ax31x2cx22x73x65x63x6fx6ex64x4bx65x79x22x3ax32x7dx5dx7dx2cx22x2dx34x22x3ax7bx22x69x33x32x22x3ax2dx32x33x34x7dx2cx22x2dx35x22x3ax7bx22x64x62x6cx22x3ax31x2ex30x7dx2cx22x2dx36x22x3ax7bx22x64x62x6cx22x3ax2dx32x2ex35x7dx7d
> bytes in text
> ={"-1":{"i32":234},"-2":{"lst":["str",2,"firstString","secondString"]},"-3":{"map":["str","i32",2,{"firstKey":1,"secondKey":2}]},"-4":{"i32":-234},"-5":{"dbl":1.0},"-6":{"dbl":-2.5}}
> o class = class java.util.ArrayList
> o size = 6
> o[0] class = class java.lang.Integer
> o[1] class = class java.util.ArrayList
> o[2] class = class java.util.HashMap
> o = [234, [firstString, secondString], {firstKey=1,
> secondKey=2}, -234, 1.0, -2.5]
> Testing protocol:
> org.apache.hadoop.hive.serde2.thrift.TCTLSeparatedProtocol
> TypeName =
> struct{_hello:int,2bye:array<string>,another:map<string,int>,nhello:int,d:double,nd:double}
> bytes
> =x32x33x34x01x66x69x72x73x74x53x74x72x69x6ex67x02x73x65x63x6fx6ex64x53x74x72x69x6ex67x01x66x69x72x73x74x4bx65x79x03x31x02x73x65x63x6fx6ex64x4bx65x79x03x32x01x2dx32x33x34x01x31x2ex30x01x2dx32x2ex35
> bytes in text
> =234firstStringsecondStringfirstKey1secondKey2-2341.0-2.5
> o class = class java.util.ArrayList
> o size = 6
> o[0] class = class java.lang.Integer
> o[1] class = class java.util.ArrayList
> o[2] class = class java.util.HashMap
> o = [234, [firstString, secondString], {firstKey=1,
> secondKey=2}, -234, 1.0, -2.5]
> Beginning Test testTBinarySortableProtocol:
> Testing struct test { double hello}
> Testing struct test { i32 hello}
> Testing struct test { i64 hello}
> Testing struct test { string hello}
> Testing struct test { string hello, double another}
> Test testTBinarySortableProtocol passed!
> bytes in text =234      firstStringsecondString
> firstKey1secondKey2>
> compare to    =234      firstStringsecondString
> firstKey1secondKey2>
> o class = class java.util.ArrayList
> o size = 3
> o[0] class = class java.lang.Integer
> o[1] class = class java.util.ArrayList
> o[2] class = class java.util.HashMap
> o = [234, [firstString, secondString], {firstKey=1,
> secondKey=2}]
> bytes in text =234      firstStringsecondString
> firstKey1secondKey2>
> compare to    =234      firstStringsecondString
> firstKey1secondKey2>
> o class = class java.util.ArrayList
> o size = 3
> o = [234, null, {firstKey=1, secondKey=2}]
> Hive history
> file=/home/ssarkar/hive/ql/../build/ql/tmp/hive_job_log_ssarkar_200902031616_992344490.txt
> Hive history
> file=/home/ssarkar/hive/ql/../build/ql/tmp/hive_job_log_ssarkar_200902031616_1962723908.txt
> OK
> OK
> Copying data from
> file:/home/ssarkar/hive/data/files/kv1.txt
> Loading data to table testhivedrivertable
> OK
> OK
> OK
> Hive history
> file=/home/ssarkar/hive/ql/../build/ql/tmp/hive_job_log_ssarkar_200902031616_247426390.txt
> Begin query: altern1.q
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=12}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=12}
> OK
> Loading data to table srcbucket
> OK
> Loading data to table srcbucket
> OK
> Loading data to table src
> OK
> diff -I \(file:\)\|\(/tmp/.*\)
> /home/ssarkar/hive/build/ql/test/logs/clientnegative/altern1.q.out
> /home/ssarkar/hive/ql/src/test/results/clientnegative/altern1.q.out
> Done query: altern1.q
> Hive history
> file=/home/ssarkar/hive/ql/../build/ql/tmp/hive_job_log_ssarkar_200902031616_587924093.txt
> Begin query: bad_sample_clause.q
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=12}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=12}
> OK
> Loading data to table srcbucket
> OK
> Loading data to table srcbucket
> OK
> Loading data to table src
> OK
> diff -I \(file:\)\|\(/tmp/.*\)
> /home/ssarkar/hive/build/ql/test/logs/clientnegative/bad_sample_clause.q.out
> /home/ssarkar/hive/ql/src/test/results/clientnegative/bad_sample_clause.q.out
> Done query: bad_sample_clause.q
> Hive history
> file=/home/ssarkar/hive/ql/../build/ql/tmp/hive_job_log_ssarkar_200902031616_-1415770190.txt
> Begin query: clusterbydistributeby.q
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=12}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=12}
> OK
> Loading data to table srcbucket
> OK
> Loading data to table srcbucket
> OK
> Loading data to table src
> OK
> diff -I \(file:\)\|\(/tmp/.*\)
> /home/ssarkar/hive/build/ql/test/logs/clientnegative/clusterbydistributeby.q.out
> /home/ssarkar/hive/ql/src/test/results/clientnegative/clusterbydistributeby.q.out
> Done query: clusterbydistributeby.q
> Hive history
> file=/home/ssarkar/hive/ql/../build/ql/tmp/hive_job_log_ssarkar_200902031616_-1882308680.txt
> Begin query: clusterbysortby.q
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=12}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=12}
> OK
> Loading data to table srcbucket
> OK
> Loading data to table srcbucket
> OK
> Loading data to table src
> OK
> diff -I \(file:\)\|\(/tmp/.*\)
> /home/ssarkar/hive/build/ql/test/logs/clientnegative/clusterbysortby.q.out
> /home/ssarkar/hive/ql/src/test/results/clientnegative/clusterbysortby.q.out
> Done query: clusterbysortby.q
> Hive history
> file=/home/ssarkar/hive/ql/../build/ql/tmp/hive_job_log_ssarkar_200902031616_1579535074.txt
> Begin query: clustern1.q
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=12}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=12}
> OK
> Loading data to table srcbucket
> OK
> Loading data to table srcbucket
> OK
> Loading data to table src
> OK
> Exception: Unresolved compilation problem:
>         The method process(Node, NodeProcessorCtx,
> Object...) of type TypeCheckProcFactory.StrExprProcessor
> must override a superclass method
> 
> java.lang.Error: Unresolved compilation problem:
>         The method process(Node, NodeProcessorCtx,
> Object...) of type TypeCheckProcFactory.StrExprProcessor
> must override a superclass method
> 
>         at
> org.apache.hadoop.hive.ql.parse.TypeCheckProcFactory$StrExprProcessor.process(TypeCheckProcFactory.java:163)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultRuleDispatcher.dispatch(DefaultRuleDispatcher.java:80)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.dispatch(DefaultGraphWalker.java:83)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.walk(DefaultGraphWalker.java:113)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.startWalking(DefaultGraphWalker.java:95)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genExprNodeDesc(SemanticAnalyzer.java:3311)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genSelectPlan(SemanticAnalyzer.java:1167)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genBodyPlan(SemanticAnalyzer.java:2724)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:3048)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:3229)
>         at
> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:71)
>         at
> org.apache.hadoop.hive.ql.parse.ExplainSemanticAnalyzer.analyzeInternal(ExplainSemanticAnalyzer.java:43)
>         at
> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:71)
>         at
> org.apache.hadoop.hive.ql.Driver.run(Driver.java:193)
>         at
> org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:174)
>         at
> org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:207)
>         at
> org.apache.hadoop.hive.ql.QTestUtil.executeClient(QTestUtil.java:411)
>         at
> org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_clustern1(TestNegativeCliDriver.java:205)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
>         at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown
> Source)
>         at java.lang.reflect.Method.invoke(Unknown Source)
>         at
> junit.framework.TestCase.runTest(TestCase.java:154)
>         at
> junit.framework.TestCase.runBare(TestCase.java:127)
>         at
> junit.framework.TestResult$1.protect(TestResult.java:106)
>         at
> junit.framework.TestResult.runProtected(TestResult.java:124)
>         at
> junit.framework.TestResult.run(TestResult.java:109)
>         at junit.framework.TestCase.run(TestCase.java:118)
>         at
> junit.framework.TestSuite.runTest(TestSuite.java:208)
>         at
> junit.framework.TestSuite.run(TestSuite.java:203)
>         at
> org.eclipse.jdt.internal.junit.runner.junit3.JUnit3TestReference.run(JUnit3TestReference.java:130)
>         at
> org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:38)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:460)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:673)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:386)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:196)
> Hive history
> file=/home/ssarkar/hive/ql/../build/ql/tmp/hive_job_log_ssarkar_200902031616_-430224382.txt
> Begin query: clustern2.q
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=12}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=12}
> OK
> Loading data to table srcbucket
> OK
> Loading data to table srcbucket
> OK
> Loading data to table src
> OK
> java.lang.Error: Unresolved compilation problem:
>         The method process(Node, NodeProcessorCtx,
> Object...) of type TypeCheckProcFactory.StrExprProcessor
> must override a superclass method
> 
>         at
> org.apache.hadoop.hive.ql.parse.TypeCheckProcFactory$StrExprProcessor.process(TypeCheckProcFactory.java:163)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultRuleDispatcher.dispatch(DefaultRuleDispatcher.java:80)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.dispatch(DefaultGraphWalker.java:83)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.walk(DefaultGraphWalker.java:113)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.startWalking(DefaultGraphWalker.java:95)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genExprNodeDesc(SemanticAnalyzer.java:3311)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genJoinReduceSinkChild(SemanticAnalyzer.java:2332)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genJoinOperator(SemanticAnalyzer.java:2380)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genJoinPlan(SemanticAnalyzer.java:2444)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:3041)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:3229)
>         at
> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:71)
>         at
> org.apache.hadoop.hive.ql.parse.ExplainSemanticAnalyzer.analyzeInternal(ExplainSemanticAnalyzer.java:43)
>         at
> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:71)
>         at
> org.apache.hadoop.hive.ql.Driver.run(Driver.java:193)
>         at
> org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:174)
>         at
> org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:207)
>         at
> org.apache.hadoop.hive.ql.QTestUtil.executeClient(QTestUtil.java:411)
>         at
> org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_clustern2(TestNegativeCliDriver.java:230)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
>         at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown
> Source)
>         at java.lang.reflect.Method.invoke(Unknown Source)
>         at
> junit.framework.TestCase.runTest(TestCase.java:154)
>         at
> junit.framework.TestCase.runBare(TestCase.java:127)
>         at
> junit.framework.TestResult$1.protect(TestResult.java:106)
>         at
> junit.framework.TestResult.runProtected(TestResult.java:124)
>         at
> junit.framework.TestResult.run(TestResult.java:109)
>         at junit.framework.TestCase.run(TestCase.java:118)
>         at
> junit.framework.TestSuite.runTest(TestSuite.java:208)
>         at
> junit.framework.TestSuite.run(TestSuite.java:203)
>         at
> org.eclipse.jdt.internal.junit.runner.junit3.JUnit3TestReference.run(JUnit3TestReference.java:130)
>         at
> org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:38)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:460)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:673)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:386)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:196)
> Exception: Unresolved compilation problem:
>         The method process(Node, NodeProcessorCtx,
> Object...) of type TypeCheckProcFactory.StrExprProcessor
> must override a superclass method
> 
> Hive history
> file=/home/ssarkar/hive/ql/../build/ql/tmp/hive_job_log_ssarkar_200902031616_-431481701.txt
> Begin query: clustern3.q
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=12}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=12}
> OK
> Loading data to table srcbucket
> OK
> Loading data to table srcbucket
> OK
> Loading data to table src
> OK
> Exception: Unresolved compilation problem:
>         The method process(Node, NodeProcessorCtx,
> Object...) of type TypeCheckProcFactory.StrExprProcessor
> must override a superclass method
> 
> java.lang.Error: Unresolved compilation problem:
>         The method process(Node, NodeProcessorCtx,
> Object...) of type TypeCheckProcFactory.StrExprProcessor
> must override a superclass method
> 
>         at
> org.apache.hadoop.hive.ql.parse.TypeCheckProcFactory$StrExprProcessor.process(TypeCheckProcFactory.java:163)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultRuleDispatcher.dispatch(DefaultRuleDispatcher.java:80)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.dispatch(DefaultGraphWalker.java:83)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.walk(DefaultGraphWalker.java:113)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.startWalking(DefaultGraphWalker.java:95)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genExprNodeDesc(SemanticAnalyzer.java:3311)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genSelectPlan(SemanticAnalyzer.java:1167)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genBodyPlan(SemanticAnalyzer.java:2724)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:3048)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:3229)
>         at
> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:71)
>         at
> org.apache.hadoop.hive.ql.parse.ExplainSemanticAnalyzer.analyzeInternal(ExplainSemanticAnalyzer.java:43)
>         at
> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:71)
>         at
> org.apache.hadoop.hive.ql.Driver.run(Driver.java:193)
>         at
> org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:174)
>         at
> org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:207)
>         at
> org.apache.hadoop.hive.ql.QTestUtil.executeClient(QTestUtil.java:411)
>         at
> org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_clustern3(TestNegativeCliDriver.java:255)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
>         at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown
> Source)
>         at java.lang.reflect.Method.invoke(Unknown Source)
>         at
> junit.framework.TestCase.runTest(TestCase.java:154)
>         at
> junit.framework.TestCase.runBare(TestCase.java:127)
>         at
> junit.framework.TestResult$1.protect(TestResult.java:106)
>         at
> junit.framework.TestResult.runProtected(TestResult.java:124)
>         at
> junit.framework.TestResult.run(TestResult.java:109)
>         at junit.framework.TestCase.run(TestCase.java:118)
>         at
> junit.framework.TestSuite.runTest(TestSuite.java:208)
>         at
> junit.framework.TestSuite.run(TestSuite.java:203)
>         at
> org.eclipse.jdt.internal.junit.runner.junit3.JUnit3TestReference.run(JUnit3TestReference.java:130)
>         at
> org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:38)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:460)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:673)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:386)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:196)
> Hive history
> file=/home/ssarkar/hive/ql/../build/ql/tmp/hive_job_log_ssarkar_200902031616_1179496399.txt
> Begin query: clustern4.q
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=12}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=12}
> OK
> Loading data to table srcbucket
> OK
> Loading data to table srcbucket
> OK
> Loading data to table src
> OK
> java.lang.Error: Unresolved compilation problem:
>         The method process(Node, NodeProcessorCtx,
> Object...) of type TypeCheckProcFactory.StrExprProcessor
> must override a superclass method
> 
>         at
> org.apache.hadoop.hive.ql.parse.TypeCheckProcFactory$StrExprProcessor.process(TypeCheckProcFactory.java:163)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultRuleDispatcher.dispatch(DefaultRuleDispatcher.java:80)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.dispatch(DefaultGraphWalker.java:83)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.walk(DefaultGraphWalker.java:113)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.startWalking(DefaultGraphWalker.java:95)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genExprNodeDesc(SemanticAnalyzer.java:3311)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genSelectPlan(SemanticAnalyzer.java:1167)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genBodyPlan(SemanticAnalyzer.java:2724)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:3048)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:3229)
>         at
> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:71)
>         at
> org.apache.hadoop.hive.ql.parse.ExplainSemanticAnalyzer.analyzeInternal(ExplainSemanticAnalyzer.java:43)
>         at
> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:71)
>         at
> org.apache.hadoop.hive.ql.Driver.run(Driver.java:193)
>         at
> org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:174)
>         at
> org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:207)
>         at
> org.apache.hadoop.hive.ql.QTestUtil.executeClient(QTestUtil.java:411)
>         at
> org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_clustern4(TestNegativeCliDriver.java:280)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
>         at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown
> Source)
>         at java.lang.reflect.Method.invoke(Unknown Source)
>         at
> junit.framework.TestCase.runTest(TestCase.java:154)
>         at
> junit.framework.TestCase.runBare(TestCase.java:127)
>         at
> junit.framework.TestResult$1.protect(TestResult.java:106)
>         at
> junit.framework.TestResult.runProtected(TestResult.java:124)
>         at
> junit.framework.TestResult.run(TestResult.java:109)
>         at junit.framework.TestCase.run(TestCase.java:118)
>         at
> junit.framework.TestSuite.runTest(TestSuite.java:208)
>         at
> junit.framework.TestSuite.run(TestSuite.java:203)
>         at
> org.eclipse.jdt.internal.junit.runner.junit3.JUnit3TestReference.run(JUnit3TestReference.java:130)
>         at
> org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:38)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:460)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:673)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:386)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:196)
> Exception: Unresolved compilation problem:
>         The method process(Node, NodeProcessorCtx,
> Object...) of type TypeCheckProcFactory.StrExprProcessor
> must override a superclass method
> 
> Hive history
> file=/home/ssarkar/hive/ql/../build/ql/tmp/hive_job_log_ssarkar_200902031616_1998238474.txt
> Begin query: describe_xpath1.q
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=12}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=12}
> OK
> Loading data to table srcbucket
> OK
> Loading data to table srcbucket
> OK
> Loading data to table src
> OK
> diff -I \(file:\)\|\(/tmp/.*\)
> /home/ssarkar/hive/build/ql/test/logs/clientnegative/describe_xpath1.q.out
> /home/ssarkar/hive/ql/src/test/results/clientnegative/describe_xpath1.q.out
> Done query: describe_xpath1.q
> Hive history
> file=/home/ssarkar/hive/ql/../build/ql/tmp/hive_job_log_ssarkar_200902031616_-93672182.txt
> Begin query: describe_xpath2.q
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=12}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=12}
> OK
> Loading data to table srcbucket
> OK
> Loading data to table srcbucket
> OK
> Loading data to table src
> OK
> diff -I \(file:\)\|\(/tmp/.*\)
> /home/ssarkar/hive/build/ql/test/logs/clientnegative/describe_xpath2.q.out
> /home/ssarkar/hive/ql/src/test/results/clientnegative/describe_xpath2.q.out
> Done query: describe_xpath2.q
> Hive history
> file=/home/ssarkar/hive/ql/../build/ql/tmp/hive_job_log_ssarkar_200902031616_1401990633.txt
> Begin query: describe_xpath3.q
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=12}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=12}
> OK
> Loading data to table srcbucket
> OK
> Loading data to table srcbucket
> OK
> Loading data to table src
> OK
> diff -I \(file:\)\|\(/tmp/.*\)
> /home/ssarkar/hive/build/ql/test/logs/clientnegative/describe_xpath3.q.out
> /home/ssarkar/hive/ql/src/test/results/clientnegative/describe_xpath3.q.out
> Done query: describe_xpath3.q
> Hive history
> file=/home/ssarkar/hive/ql/../build/ql/tmp/hive_job_log_ssarkar_200902031616_659750364.txt
> Begin query: describe_xpath4.q
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=12}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=12}
> OK
> Loading data to table srcbucket
> OK
> Loading data to table srcbucket
> OK
> Loading data to table src
> OK
> diff -I \(file:\)\|\(/tmp/.*\)
> /home/ssarkar/hive/build/ql/test/logs/clientnegative/describe_xpath4.q.out
> /home/ssarkar/hive/ql/src/test/results/clientnegative/describe_xpath4.q.out
> Done query: describe_xpath4.q
> Hive history
> file=/home/ssarkar/hive/ql/../build/ql/tmp/hive_job_log_ssarkar_200902031616_-778063141.txt
> Begin query: fileformat_bad_class.q
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=12}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=12}
> OK
> Loading data to table srcbucket
> OK
> Loading data to table srcbucket
> OK
> Loading data to table src
> OK
> diff -I \(file:\)\|\(/tmp/.*\)
> /home/ssarkar/hive/build/ql/test/logs/clientnegative/fileformat_bad_class.q.out
> /home/ssarkar/hive/ql/src/test/results/clientnegative/fileformat_bad_class.q.out
> Done query: fileformat_bad_class.q
> Hive history
> file=/home/ssarkar/hive/ql/../build/ql/tmp/hive_job_log_ssarkar_200902031616_1389054449.txt
> Begin query: fileformat_void_input.q
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=12}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=12}
> OK
> Loading data to table srcbucket
> OK
> Loading data to table srcbucket
> OK
> Loading data to table src
> OK
> Exception: Unresolved compilation problem:
>         The method process(Node, NodeProcessorCtx,
> Object...) of type TypeCheckProcFactory.StrExprProcessor
> must override a superclass method
> 
> java.lang.Error: Unresolved compilation problem:
>         The method process(Node, NodeProcessorCtx,
> Object...) of type TypeCheckProcFactory.StrExprProcessor
> must override a superclass method
> 
>         at
> org.apache.hadoop.hive.ql.parse.TypeCheckProcFactory$StrExprProcessor.process(TypeCheckProcFactory.java:163)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultRuleDispatcher.dispatch(DefaultRuleDispatcher.java:80)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.dispatch(DefaultGraphWalker.java:83)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.walk(DefaultGraphWalker.java:113)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.startWalking(DefaultGraphWalker.java:95)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genExprNodeDesc(SemanticAnalyzer.java:3311)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genFilterPlan(SemanticAnalyzer.java:904)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genBodyPlan(SemanticAnalyzer.java:2712)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:3048)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:3229)
>         at
> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:71)
>         at
> org.apache.hadoop.hive.ql.Driver.run(Driver.java:193)
>         at
> org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:174)
>         at
> org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:207)
>         at
> org.apache.hadoop.hive.ql.QTestUtil.executeClient(QTestUtil.java:411)
>         at
> org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_fileformat_void_input(TestNegativeCliDriver.java:430)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
>         at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown
> Source)
>         at java.lang.reflect.Method.invoke(Unknown Source)
>         at
> junit.framework.TestCase.runTest(TestCase.java:154)
>         at
> junit.framework.TestCase.runBare(TestCase.java:127)
>         at
> junit.framework.TestResult$1.protect(TestResult.java:106)
>         at
> junit.framework.TestResult.runProtected(TestResult.java:124)
>         at
> junit.framework.TestResult.run(TestResult.java:109)
>         at junit.framework.TestCase.run(TestCase.java:118)
>         at
> junit.framework.TestSuite.runTest(TestSuite.java:208)
>         at
> junit.framework.TestSuite.run(TestSuite.java:203)
>         at
> org.eclipse.jdt.internal.junit.runner.junit3.JUnit3TestReference.run(JUnit3TestReference.java:130)
>         at
> org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:38)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:460)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:673)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:386)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:196)
> Hive history
> file=/home/ssarkar/hive/ql/../build/ql/tmp/hive_job_log_ssarkar_200902031616_893718016.txt
> Begin query: fileformat_void_output.q
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=12}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=12}
> OK
> Loading data to table srcbucket
> OK
> Loading data to table srcbucket
> OK
> Loading data to table src
> OK
> diff -I \(file:\)\|\(/tmp/.*\)
> /home/ssarkar/hive/build/ql/test/logs/clientnegative/fileformat_void_output.q.out
> /home/ssarkar/hive/ql/src/test/results/clientnegative/fileformat_void_output.q.out
> Done query: fileformat_void_output.q
> Hive history
> file=/home/ssarkar/hive/ql/../build/ql/tmp/hive_job_log_ssarkar_200902031616_-1795879737.txt
> Begin query: input1.q
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=12}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=12}
> OK
> Loading data to table srcbucket
> OK
> Loading data to table srcbucket
> OK
> Loading data to table src
> OK
> diff -I \(file:\)\|\(/tmp/.*\)
> /home/ssarkar/hive/build/ql/test/logs/clientnegative/input1.q.out
> /home/ssarkar/hive/ql/src/test/results/clientnegative/input1.q.out
> Done query: input1.q
> Hive history
> file=/home/ssarkar/hive/ql/../build/ql/tmp/hive_job_log_ssarkar_200902031616_1786217678.txt
> Begin query: input2.q
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=12}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=12}
> OK
> Loading data to table srcbucket
> OK
> Loading data to table srcbucket
> OK
> Loading data to table src
> OK
> Exception: Unresolved compilation problem:
>         The method process(Node, NodeProcessorCtx,
> Object...) of type TypeCheckProcFactory.StrExprProcessor
> must override a superclass method
> 
> java.lang.Error: Unresolved compilation problem:
>         The method process(Node, NodeProcessorCtx,
> Object...) of type TypeCheckProcFactory.StrExprProcessor
> must override a superclass method
> 
>         at
> org.apache.hadoop.hive.ql.parse.TypeCheckProcFactory$StrExprProcessor.process(TypeCheckProcFactory.java:163)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultRuleDispatcher.dispatch(DefaultRuleDispatcher.java:80)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.dispatch(DefaultGraphWalker.java:83)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.walk(DefaultGraphWalker.java:113)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.startWalking(DefaultGraphWalker.java:95)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genExprNodeDesc(SemanticAnalyzer.java:3311)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genSelectPlan(SemanticAnalyzer.java:1167)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genBodyPlan(SemanticAnalyzer.java:2724)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:3048)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:3229)
>         at
> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:71)
>         at
> org.apache.hadoop.hive.ql.Driver.run(Driver.java:193)
>         at
> org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:174)
>         at
> org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:207)
>         at
> org.apache.hadoop.hive.ql.QTestUtil.executeClient(QTestUtil.java:411)
>         at
> org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_input2(TestNegativeCliDriver.java:505)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
>         at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown
> Source)
>         at java.lang.reflect.Method.invoke(Unknown Source)
>         at
> junit.framework.TestCase.runTest(TestCase.java:154)
>         at
> junit.framework.TestCase.runBare(TestCase.java:127)
>         at
> junit.framework.TestResult$1.protect(TestResult.java:106)
>         at
> junit.framework.TestResult.runProtected(TestResult.java:124)
>         at
> junit.framework.TestResult.run(TestResult.java:109)
>         at junit.framework.TestCase.run(TestCase.java:118)
>         at
> junit.framework.TestSuite.runTest(TestSuite.java:208)
>         at
> junit.framework.TestSuite.run(TestSuite.java:203)
>         at
> org.eclipse.jdt.internal.junit.runner.junit3.JUnit3TestReference.run(JUnit3TestReference.java:130)
>         at
> org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:38)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:460)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:673)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:386)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:196)
> Hive history
> file=/home/ssarkar/hive/ql/../build/ql/tmp/hive_job_log_ssarkar_200902031616_-1429356131.txt
> Begin query: input_testxpath4.q
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=12}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=12}
> OK
> Loading data to table srcbucket
> OK
> Loading data to table srcbucket
> OK
> Loading data to table src
> OK
> java.lang.Error: Unresolved compilation problem:
>         The method process(Node, NodeProcessorCtx,
> Object...) of type TypeCheckProcFactory.StrExprProcessor
> must override a superclass method
> 
> Exception: Unresolved compilation problem:
>         The method process(Node, NodeProcessorCtx,
> Object...) of type TypeCheckProcFactory.StrExprProcessor
> must override a superclass method
> 
>         at
> org.apache.hadoop.hive.ql.parse.TypeCheckProcFactory$StrExprProcessor.process(TypeCheckProcFactory.java:163)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultRuleDispatcher.dispatch(DefaultRuleDispatcher.java:80)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.dispatch(DefaultGraphWalker.java:83)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.walk(DefaultGraphWalker.java:113)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.startWalking(DefaultGraphWalker.java:95)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genExprNodeDesc(SemanticAnalyzer.java:3311)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genSelectPlan(SemanticAnalyzer.java:1167)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genBodyPlan(SemanticAnalyzer.java:2724)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:3048)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:3229)
>         at
> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:71)
>         at
> org.apache.hadoop.hive.ql.parse.ExplainSemanticAnalyzer.analyzeInternal(ExplainSemanticAnalyzer.java:43)
>         at
> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:71)
>         at
> org.apache.hadoop.hive.ql.Driver.run(Driver.java:193)
>         at
> org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:174)
>         at
> org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:207)
>         at
> org.apache.hadoop.hive.ql.QTestUtil.executeClient(QTestUtil.java:411)
>         at
> org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_input_testxpath4(TestNegativeCliDriver.java:530)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
>         at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown
> Source)
>         at java.lang.reflect.Method.invoke(Unknown Source)
>         at
> junit.framework.TestCase.runTest(TestCase.java:154)
>         at
> junit.framework.TestCase.runBare(TestCase.java:127)
>         at
> junit.framework.TestResult$1.protect(TestResult.java:106)
>         at
> junit.framework.TestResult.runProtected(TestResult.java:124)
>         at
> junit.framework.TestResult.run(TestResult.java:109)
>         at junit.framework.TestCase.run(TestCase.java:118)
>         at
> junit.framework.TestSuite.runTest(TestSuite.java:208)
>         at
> junit.framework.TestSuite.run(TestSuite.java:203)
>         at
> org.eclipse.jdt.internal.junit.runner.junit3.JUnit3TestReference.run(JUnit3TestReference.java:130)
>         at
> org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:38)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:460)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:673)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:386)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:196)
> Hive history
> file=/home/ssarkar/hive/ql/../build/ql/tmp/hive_job_log_ssarkar_200902031616_-299734685.txt
> Begin query: invalid_create_tbl1.q
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=12}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=12}
> OK
> Loading data to table srcbucket
> OK
> Loading data to table srcbucket
> OK
> Loading data to table src
> OK
> diff -I \(file:\)\|\(/tmp/.*\)
> /home/ssarkar/hive/build/ql/test/logs/clientnegative/invalid_create_tbl1.q.out
> /home/ssarkar/hive/ql/src/test/results/clientnegative/invalid_create_tbl1.q.out
> Done query: invalid_create_tbl1.q
> Hive history
> file=/home/ssarkar/hive/ql/../build/ql/tmp/hive_job_log_ssarkar_200902031616_-3796110.txt
> Begin query: invalid_create_tbl2.q
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=12}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=12}
> OK
> Loading data to table srcbucket
> OK
> Loading data to table srcbucket
> OK
> Loading data to table src
> OK
> diff -I \(file:\)\|\(/tmp/.*\)
> /home/ssarkar/hive/build/ql/test/logs/clientnegative/invalid_create_tbl2.q.out
> /home/ssarkar/hive/ql/src/test/results/clientnegative/invalid_create_tbl2.q.out
> Done query: invalid_create_tbl2.q
> Hive history
> file=/home/ssarkar/hive/ql/../build/ql/tmp/hive_job_log_ssarkar_200902031616_732040395.txt
> Begin query: invalid_select_expression.q
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=12}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=12}
> OK
> Loading data to table srcbucket
> OK
> Loading data to table srcbucket
> OK
> Loading data to table src
> OK
> diff -I \(file:\)\|\(/tmp/.*\)
> /home/ssarkar/hive/build/ql/test/logs/clientnegative/invalid_select_expression.q.out
> /home/ssarkar/hive/ql/src/test/results/clientnegative/invalid_select_expression.q.out
> Done query: invalid_select_expression.q
> Hive history
> file=/home/ssarkar/hive/ql/../build/ql/tmp/hive_job_log_ssarkar_200902031616_764555300.txt
> Begin query: invalid_tbl_name.q
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=12}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=12}
> OK
> Loading data to table srcbucket
> OK
> Loading data to table srcbucket
> OK
> Loading data to table src
> OK
> diff -I \(file:\)\|\(/tmp/.*\)
> /home/ssarkar/hive/build/ql/test/logs/clientnegative/invalid_tbl_name.q.out
> /home/ssarkar/hive/ql/src/test/results/clientnegative/invalid_tbl_name.q.out
> Done query: invalid_tbl_name.q
> Hive history
> file=/home/ssarkar/hive/ql/../build/ql/tmp/hive_job_log_ssarkar_200902031616_-1388068500.txt
> Begin query: joinneg.q
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=12}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=12}
> OK
> Loading data to table srcbucket
> OK
> Loading data to table srcbucket
> OK
> Loading data to table src
> OK
> diff -I \(file:\)\|\(/tmp/.*\)
> /home/ssarkar/hive/build/ql/test/logs/clientnegative/joinneg.q.out
> /home/ssarkar/hive/ql/src/test/results/clientnegative/joinneg.q.out
> Done query: joinneg.q
> Hive history
> file=/home/ssarkar/hive/ql/../build/ql/tmp/hive_job_log_ssarkar_200902031616_1214860.txt
> Begin query: load_wrong_fileformat.q
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=12}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=12}
> OK
> Loading data to table srcbucket
> OK
> Loading data to table srcbucket
> OK
> Loading data to table src
> OK
> diff -I \(file:\)\|\(/tmp/.*\)
> /home/ssarkar/hive/build/ql/test/logs/clientnegative/load_wrong_fileformat.q.out
> /home/ssarkar/hive/ql/src/test/results/clientnegative/load_wrong_fileformat.q.out
> Done query: load_wrong_fileformat.q
> Hive history
> file=/home/ssarkar/hive/ql/../build/ql/tmp/hive_job_log_ssarkar_200902031616_-1542677940.txt
> Begin query: notable_alias3.q
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=12}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=12}
> OK
> Loading data to table srcbucket
> OK
> Loading data to table srcbucket
> OK
> Loading data to table src
> OK
> Exception: Unresolved compilation problem:
>         The method process(Node, NodeProcessorCtx,
> Object...) of type TypeCheckProcFactory.StrExprProcessor
> must override a superclass method
> 
> java.lang.Error: Unresolved compilation problem:
>         The method process(Node, NodeProcessorCtx,
> Object...) of type TypeCheckProcFactory.StrExprProcessor
> must override a superclass method
> 
>         at
> org.apache.hadoop.hive.ql.parse.TypeCheckProcFactory$StrExprProcessor.process(TypeCheckProcFactory.java:163)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultRuleDispatcher.dispatch(DefaultRuleDispatcher.java:80)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.dispatch(DefaultGraphWalker.java:83)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.walk(DefaultGraphWalker.java:113)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.startWalking(DefaultGraphWalker.java:95)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genExprNodeDesc(SemanticAnalyzer.java:3311)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genFilterPlan(SemanticAnalyzer.java:904)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genBodyPlan(SemanticAnalyzer.java:2712)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:3048)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:3229)
>         at
> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:71)
>         at
> org.apache.hadoop.hive.ql.Driver.run(Driver.java:193)
>         at
> org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:174)
>         at
> org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:207)
>         at
> org.apache.hadoop.hive.ql.QTestUtil.executeClient(QTestUtil.java:411)
>         at
> org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_notable_alias3(TestNegativeCliDriver.java:705)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
>         at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown
> Source)
>         at java.lang.reflect.Method.invoke(Unknown Source)
>         at
> junit.framework.TestCase.runTest(TestCase.java:154)
>         at
> junit.framework.TestCase.runBare(TestCase.java:127)
>         at
> junit.framework.TestResult$1.protect(TestResult.java:106)
>         at
> junit.framework.TestResult.runProtected(TestResult.java:124)
>         at
> junit.framework.TestResult.run(TestResult.java:109)
>         at junit.framework.TestCase.run(TestCase.java:118)
>         at
> junit.framework.TestSuite.runTest(TestSuite.java:208)
>         at
> junit.framework.TestSuite.run(TestSuite.java:203)
>         at
> org.eclipse.jdt.internal.junit.runner.junit3.JUnit3TestReference.run(JUnit3TestReference.java:130)
>         at
> org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:38)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:460)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:673)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:386)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:196)
> Hive history
> file=/home/ssarkar/hive/ql/../build/ql/tmp/hive_job_log_ssarkar_200902031616_-555682788.txt
> Begin query: notable_alias4.q
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=12}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=12}
> OK
> Loading data to table srcbucket
> OK
> Loading data to table srcbucket
> OK
> Loading data to table src
> OK
> Exception: Unresolved compilation problem:
>         The method process(Node, NodeProcessorCtx,
> Object...) of type TypeCheckProcFactory.StrExprProcessor
> must override a superclass method
> 
> java.lang.Error: Unresolved compilation problem:
>         The method process(Node, NodeProcessorCtx,
> Object...) of type TypeCheckProcFactory.StrExprProcessor
> must override a superclass method
> 
>         at
> org.apache.hadoop.hive.ql.parse.TypeCheckProcFactory$StrExprProcessor.process(TypeCheckProcFactory.java:163)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultRuleDispatcher.dispatch(DefaultRuleDispatcher.java:80)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.dispatch(DefaultGraphWalker.java:83)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.walk(DefaultGraphWalker.java:113)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.startWalking(DefaultGraphWalker.java:95)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genExprNodeDesc(SemanticAnalyzer.java:3311)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genJoinReduceSinkChild(SemanticAnalyzer.java:2332)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genJoinOperator(SemanticAnalyzer.java:2380)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genJoinPlan(SemanticAnalyzer.java:2444)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:3041)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:3229)
>         at
> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:71)
>         at
> org.apache.hadoop.hive.ql.parse.ExplainSemanticAnalyzer.analyzeInternal(ExplainSemanticAnalyzer.java:43)
>         at
> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:71)
>         at
> org.apache.hadoop.hive.ql.Driver.run(Driver.java:193)
>         at
> org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:174)
>         at
> org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:207)
>         at
> org.apache.hadoop.hive.ql.QTestUtil.executeClient(QTestUtil.java:411)
>         at
> org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_notable_alias4(TestNegativeCliDriver.java:730)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
>         at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown
> Source)
>         at java.lang.reflect.Method.invoke(Unknown Source)
>         at
> junit.framework.TestCase.runTest(TestCase.java:154)
>         at
> junit.framework.TestCase.runBare(TestCase.java:127)
>         at
> junit.framework.TestResult$1.protect(TestResult.java:106)
>         at
> junit.framework.TestResult.runProtected(TestResult.java:124)
>         at
> junit.framework.TestResult.run(TestResult.java:109)
>         at junit.framework.TestCase.run(TestCase.java:118)
>         at
> junit.framework.TestSuite.runTest(TestSuite.java:208)
>         at
> junit.framework.TestSuite.run(TestSuite.java:203)
>         at
> org.eclipse.jdt.internal.junit.runner.junit3.JUnit3TestReference.run(JUnit3TestReference.java:130)
>         at
> org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:38)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:460)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:673)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:386)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:196)
> Hive history
> file=/home/ssarkar/hive/ql/../build/ql/tmp/hive_job_log_ssarkar_200902031616_-1604113442.txt
> Begin query: strict_pruning.q
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=12}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=12}
> OK
> Loading data to table srcbucket
> OK
> Loading data to table srcbucket
> OK
> Loading data to table src
> OK
> Exception: Unresolved compilation problem:
>         The method process(Node, NodeProcessorCtx,
> Object...) of type TypeCheckProcFactory.NumExprProcessor
> must override a superclass method
> 
> java.lang.Error: Unresolved compilation problem:
>         The method process(Node, NodeProcessorCtx,
> Object...) of type TypeCheckProcFactory.NumExprProcessor
> must override a superclass method
> 
>         at
> org.apache.hadoop.hive.ql.parse.TypeCheckProcFactory$NumExprProcessor.process(TypeCheckProcFactory.java:121)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultRuleDispatcher.dispatch(DefaultRuleDispatcher.java:80)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.dispatch(DefaultGraphWalker.java:83)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.walk(DefaultGraphWalker.java:113)
>         at
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.startWalking(DefaultGraphWalker.java:95)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genExprNodeDesc(SemanticAnalyzer.java:3311)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genGroupByPlanReduceSinkOperator(SemanticAnalyzer.java:1688)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genGroupByPlan2MR(SemanticAnalyzer.java:1892)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genBodyPlan(SemanticAnalyzer.java:2721)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:3048)
>         at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:3229)
>         at
> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:71)
>         at
> org.apache.hadoop.hive.ql.parse.ExplainSemanticAnalyzer.analyzeInternal(ExplainSemanticAnalyzer.java:43)
>         at
> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:71)
>         at
> org.apache.hadoop.hive.ql.Driver.run(Driver.java:193)
>         at
> org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:174)
>         at
> org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:207)
>         at
> org.apache.hadoop.hive.ql.QTestUtil.executeClient(QTestUtil.java:411)
>         at
> org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_strict_pruning(TestNegativeCliDriver.java:755)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
>         at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown
> Source)
>         at java.lang.reflect.Method.invoke(Unknown Source)
>         at
> junit.framework.TestCase.runTest(TestCase.java:154)
>         at
> junit.framework.TestCase.runBare(TestCase.java:127)
>         at
> junit.framework.TestResult$1.protect(TestResult.java:106)
>         at
> junit.framework.TestResult.runProtected(TestResult.java:124)
>         at
> junit.framework.TestResult.run(TestResult.java:109)
>         at junit.framework.TestCase.run(TestCase.java:118)
>         at
> junit.framework.TestSuite.runTest(TestSuite.java:208)
>         at
> junit.framework.TestSuite.run(TestSuite.java:203)
>         at
> org.eclipse.jdt.internal.junit.runner.junit3.JUnit3TestReference.run(JUnit3TestReference.java:130)
>         at
> org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:38)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:460)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:673)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:386)
>         at
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:196)
> Hive history
> file=/home/ssarkar/hive/ql/../build/ql/tmp/hive_job_log_ssarkar_200902031616_-327058962.txt
> Begin query: subq_insert.q
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=12}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=12}
> OK
> Loading data to table srcbucket
> OK
> Loading data to table srcbucket
> OK
> Loading data to table src
> OK
> diff -I \(file:\)\|\(/tmp/.*\)
> /home/ssarkar/hive/build/ql/test/logs/clientnegative/subq_insert.q.out
> /home/ssarkar/hive/ql/src/test/results/clientnegative/subq_insert.q.out
> Done query: subq_insert.q
> Hive history
> file=/home/ssarkar/hive/ql/../build/ql/tmp/hive_job_log_ssarkar_200902031616_-196827093.txt
> Begin query: union.q
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-08,
> hr=12}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=11}
> OK
> Loading data to table srcpart partition {ds=2008-04-09,
> hr=12}
> OK
> Loading data to table srcbucket
> OK
> Loading data to table srcbucket
> OK
> Loading data to table src
> OK
> diff -I \(file:\)\|\(/tmp/.*\)
> /home/ssarkar/hive/build/ql/test/logs/clientnegative/union.q.out
> /home/ssarkar/hive/ql/src/test/results/clientnegative/union.q.out
> Done query: union.q
> =====================================================================
> 
> Thanks,
> Shyam
> 
> 
> 
> 
> --- On Tue, 2/3/09, Prasad Chakka
> <prasad@facebook.com> wrote:
> 
> > From: Prasad Chakka <prasad@facebook.com>
> > Subject: Re: Eclipse run fails !!
> > To: "hive-dev@hadoop.apache.org"
> <hive-dev@hadoop.apache.org>,
> "shyam_sarkar@yahoo.com"
> <shyam_sarkar@yahoo.com>
> > Date: Tuesday, February 3, 2009, 2:51 PM
> > I think there are multiple issues. Please do the
> following
> >
> >
> >  1.  'ant clean' in hive directory
> >  2.  delete project in eclipse
> >  3.  Don't change any config values in
> hive-site.xml
> > (revert your changes to fs.default.name etc) and
> don't
> > start HDFS cluster since in unit tests we are working
> on
> > local file system.
> >  4.  check what java version is 1.6
> >  5.  Follow the steps in the hive eclipse setup wiki
> with
> > -Dhadoop.version=0.17.2.1
> >  6.  Open Eclipse and import the project
> >  7.  Open project preferences and make sure that it is
> > using java 6. If it is not then change it to use java6
> (let
> > me know if you need help here). If you change it then
> make
> > sure that you rebuild the project by doing a clean
> >  8.  Make sure that there are no compilation problems
> for
> > the hive project (check 'problems' tab in the
> bottom
> > panel of Eclipse)
> >  9.  Run the Junit test case. It should run without
> any
> > warning dialogs
> >
> > Let me know which of these steps fail and what is the
> > error. You need not change any files run a junit
> testcase.
> > Once you are at this point, we can help you in setting
> up
> > command shell that talks to DFS.
> >
> > Prasad
> >
> > ________________________________
> > From: Ashish Thusoo <athusoo@facebook.com>
> > Reply-To: <hive-dev@hadoop.apache.org>
> > Date: Tue, 3 Feb 2009 14:41:12 -0800
> > To: <shyam_sarkar@yahoo.com>,
> > <hive-dev@hadoop.apache.org>
> > Subject: RE: Eclipse run fails !!
> >
> > Actually for running hive through eclipse you
> don't
> > need to download and start hadoop. Hive tests
> automatically
> > create a local instance of hdfs and map/reduce and are
> able
> > to run it.
> >
> > The errors that you are getting seem to indicate some
> jpox
> > plugins missing in eclipse. Prasad is an expert in
> that area
> > and can perhaps comment on that...
> >
> > Ashish
> >
> > -----Original Message-----
> > From: Shyam Sarkar [mailto:shyam_sarkar@yahoo.com]
> > Sent: Tuesday, February 03, 2009 2:30 PM
> > To: hive-dev@hadoop.apache.org; Ashish Thusoo
> > Subject: RE: Eclipse run fails !!
> >
> > Dear Ashish,
> >
> > I downloaded hadoop 0.17.0 and tried bin/start-all.sh
> > script. I got one error ::
> >
> ==============================================================
> > [ssarkar@ayush2 hadoop-0.17.0]$ bin/start-all.sh
> starting
> > namenode, logging to
> >
> /home/ssarkar/hadoop/hadoop-0.17.0/bin/../logs/hadoop-ssarkar-namenode-ayush2.out
> > ssarkar@localhost's password:
> > localhost: starting datanode, logging to
> >
> /home/ssarkar/hadoop/hadoop-0.17.0/bin/../logs/hadoop-ssarkar-datanode-ayush2.out
> > ssarkar@localhost's password:
> > localhost: starting secondarynamenode, logging to
> >
> /home/ssarkar/hadoop/hadoop-0.17.0/bin/../logs/hadoop-ssarkar-secondarynamenode-ayush2.out
> > localhost: Exception in thread "main"
> > java.lang.NullPointerException
> > localhost:      at
> >
> org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:119)
> > localhost:      at
> >
> org.apache.hadoop.dfs.SecondaryNameNode.<init>(SecondaryNameNode.java:118)
> > localhost:      at
> >
> org.apache.hadoop.dfs.SecondaryNameNode.main(SecondaryNameNode.java:495)
> > starting jobtracker, logging to
> >
> /home/ssarkar/hadoop/hadoop-0.17.0/bin/../logs/hadoop-ssarkar-jobtracker-ayush2.out
> > ssarkar@localhost's password:
> > localhost: starting tasktracker, logging to
> >
> /home/ssarkar/hadoop/hadoop-0.17.0/bin/../logs/hadoop-ssarkar-tasktracker-ayush2.out
> > [ssarkar@ayush2 hadoop-0.17.0]$
> >
> ===================================================================
> >
> > Next I loaded hive project into eclipse following
> steps in
> > hive wiki.
> > I tried Run->Run Configurations->JUnit and
> selected
> > TestTruncate to run but got the following error ::
> >
> > "Errors exist in required Projest(s):
> >
> > hive
> >
> > Proceed with Launch ?"
> >
> > When I launch I got following errors ::
> >
> >
> =================================================================
> > 09/02/03 14:01:33 INFO metastore.HiveMetaStore: 0:
> Opening
> > raw store with implemenation
> > class:org.apache.hadoop.hive.metastore.ObjectStore
> > 09/02/03 14:01:33 INFO metastore.ObjectStore:
> ObjectStore,
> > initialize called
> > 09/02/03 14:01:33 INFO metastore.ObjectStore: found
> > resource jpox.properties at
> > file:/home/ssarkar/hive/conf/jpox.properties
> > 09/02/03 14:01:33 WARN JPOX.Plugin: Extension Point
> > "org.eclipse.ui.views" not registered, but
> plugin
> > "org.eclipse.jdt.junit" defined in
> >
> file:/home/ssarkar/eclipse/eclipse32_3.4.1_working/configuration/org.eclipse.osgi/bundles/97/1/.cp/META-INF/MANIFEST.MF
> > refers to it.
> > 09/02/03 14:01:33 WARN JPOX.Plugin: Extension Point
> > "org.eclipse.ui.perspectiveExtensions" not
> > registered, but plugin
> "org.eclipse.jdt.junit"
> > defined in
> >
> file:/home/ssarkar/eclipse/eclipse32_3.4.1_working/configuration/org.eclipse.osgi/bundles/97/1/.cp/META-INF/MANIFEST.MF
> > refers to it.
> > 09/02/03 14:01:33 WARN JPOX.Plugin: Extension Point
> > "org.eclipse.ui.preferencePages" not
> registered,
> > but plugin "org.eclipse.jdt.junit" defined
> in
> >
> file:/home/ssarkar/eclipse/eclipse32_3.4.1_working/configuration/org.eclipse.osgi/bundles/97/1/.cp/META-INF/MANIFEST.MF
> > refers to it.
> > 09/02/03 14:01:33 WARN JPOX.Plugin: Extension Point
> > "org.eclipse.ui.keywords" not registered,
> but
> > plugin "org.eclipse.jdt.junit" defined in
> >
> file:/home/ssarkar/eclipse/eclipse32_3.4.1_working/configuration/org.eclipse.osgi/bundles/97/1/.cp/META-INF/MANIFEST.MF
> > refers to it.
> > 09/02/03 14:01:33 WARN JPOX.Plugin: Extension Point
> >
> "org.eclipse.debug.core.launchConfigurationTypes"
> > not registered, but plugin
> "org.eclipse.jdt.junit"
> > defined in
> >
> file:/home/ssarkar/eclipse/eclipse32_3.4.1_working/configuration/org.eclipse.osgi/bundles/97/1/.cp/META-INF/MANIFEST.MF
> > refers to it.
> > 09/02/03 14:01:33 WARN JPOX.Plugin: Extension Point
> >
> "org.eclipse.debug.core.launchConfigurationComparators"
> > not registered, but plugin
> "org.eclipse.jdt.junit"
> > defined in
> >
> file:/home/ssarkar/eclipse/eclipse32_3.4.1_working/configuration/org.eclipse.osgi/bundles/97/1/.cp/META-INF/MANIFEST.MF
> > refers to it.
> > 09/02/03 14:01:33 WARN JPOX.Plugin: Extension Point
> >
> "org.eclipse.debug.ui.launchConfigurationTypeImages"
> > not registered, but plugin
> "org.eclipse.jdt.junit"
> > defined in
> >
> file:/home/ssarkar/eclipse/eclipse32_3.4.1_working/configuration/org.eclipse.osgi/bundles/97/1/.cp/META-INF/MANIFEST.MF
> > refers to it.
> > 09/02/03 14:01:33 WARN JPOX.Plugin: Extension Point
> >
> "org.eclipse.debug.ui.launchConfigurationTabGroups"
> > not registered, but plugin
> "org.eclipse.jdt.junit"
> > defined in
> >
> file:/home/ssarkar/eclipse/eclipse32_3.4.1_working/configuration/org.eclipse.osgi/bundles/97/1/.cp/META-INF/MANIFEST.MF
> > refers to it.
> > 09/02/03 14:01:33 WARN JPOX.Plugin: Extension Point
> > "org.eclipse.ui.newWizards" not registered,
> but
> > plugin "org.eclipse.jdt.junit" defined in
> >
> file:/home/ssarkar/eclipse/eclipse32_3.4.1_working/configuration/org.eclipse.osgi/bundles/97/1/.cp/META-INF/MANIFEST.MF
> > refers to it.
> > 09/02/03 14:01:33 WARN JPOX.Plugin: Extension Point
> > "org.eclipse.ui.popupMenus" not registered,
> but
> > plugin "org.eclipse.jdt.junit" defined in
> >
> file:/home/ssarkar/eclipse/eclipse32_3.4.1_working/configuration/org.eclipse.osgi/bundles/97/1/.cp/META-INF/MANIFEST.MF
> > refers to it.
> > 09/02/03 14:01:33 WARN JPOX.Plugin: Extension Point
> > "org.eclipse.ui.actionSets" not registered,
> but
> > plugin "org.eclipse.jdt.junit" defined in
> >
> file:/home/ssarkar/eclipse/eclipse32_3.4.1_working/configuration/org.eclipse.osgi/bundles/97/1/.cp/META-INF/MANIFEST.MF
> > refers to it.
> > 09/02/03 14:01:33 WARN JPOX.Plugin: Extension Point
> > "org.eclipse.ui.actionSetPartAssociations"
> not
> > registered, but plugin
> "org.eclipse.jdt.junit"
> > defined in
> >
> file:/home/ssarkar/eclipse/eclipse32_3.4.1_working/configuration/org.eclipse.osgi/bundles/97/1/.cp/META-INF/MANIFEST.MF
> > refers to it.
> > 09/02/03 14:01:33 WARN JPOX.Plugin: Extension Point
> > "org.eclipse.debug.ui.launchShortcuts" not
> > registered, but plugin
> "org.eclipse.jdt.junit"
> > defined in
> >
> file:/home/ssarkar/eclipse/eclipse32_3.4.1_working/configuration/org.eclipse.osgi/bundles/97/1/.cp/META-INF/MANIFEST.MF
> > refers to it.
> > 09/02/03 14:01:33 WARN JPOX.Plugin: Extension Point
> >
> "org.eclipse.jdt.core.classpathVariableInitializer"
> > not registered, but plugin
> "org.eclipse.jdt.junit"
> > defined in
> >
> file:/home/ssarkar/eclipse/eclipse32_3.4.1_working/configuration/org.eclipse.osgi/bundles/97/1/.cp/META-INF/MANIFEST.MF
> > refers to it.
> > 09/02/03 14:01:33 WARN JPOX.Plugin: Extension Point
> > "org.eclipse.jdt.ui.quickFixProcessors" not
> > registered, but plugin
> "org.eclipse.jdt.junit"
> > defined in
> >
> file:/home/ssarkar/eclipse/eclipse32_3.4.1_working/configuration/org.eclipse.osgi/bundles/97/1/.cp/META-INF/MANIFEST.MF
> > refers to it.
> > 09/02/03 14:01:33 WARN JPOX.Plugin: Extension Point
> > "org.eclipse.jdt.ui.classpathFixProcessors"
> not
> > registered, but plugin
> "org.eclipse.jdt.junit"
> > defined in
> >
> file:/home/ssarkar/eclipse/eclipse32_3.4.1_working/configuration/org.eclipse.osgi/bundles/97/1/.cp/META-INF/MANIFEST.MF
> > refers to it.
> > 09/02/03 14:01:33 WARN JPOX.Plugin: Extension Point
> > "org.eclipse.ui.ide.markerResolution" not
> > registered, but plugin
> "org.eclipse.jdt.junit"
> > defined in
> >
> file:/home/ssarkar/eclipse/eclipse32_3.4.1_working/configuration/org.eclipse.osgi/bundles/97/1/.cp/META-INF/MANIFEST.MF
> > refers to it.
> > 09/02/03 14:01:33 WARN JPOX.Plugin: Extension Point
> >
> "org.eclipse.core.expressions.propertyTesters" not
> > registered, but plugin
> "org.eclipse.jdt.junit"
> > defined in
> >
> file:/home/ssarkar/eclipse/eclipse32_3.4.1_working/configuration/org.eclipse.osgi/bundles/97/1/.cp/META-INF/MANIFEST.MF
> > refers to it.
> > 09/02/03 14:01:33 WARN JPOX.Plugin: Extension Point
> >
> "org.eclipse.ltk.core.refactoring.renameParticipants"
> > not registered, but plugin
> "org.eclipse.jdt.junit"
> > defined in
> >
> file:/home/ssarkar/eclipse/eclipse32_3.4.1_working/configuration/org.eclipse.osgi/bundles/97/1/.cp/META-INF/MANIFEST.MF
> > refers to it.
> > 09/02/03 14:01:33 WARN JPOX.Plugin: Extension Point
> > "org.eclipse.ui.commands" not registered,
> but
> > plugin "org.eclipse.jdt.junit" defined in
> >
> file:/home/ssarkar/eclipse/eclipse32_3.4.1_working/configuration/org.eclipse.osgi/bundles/97/1/.cp/META-INF/MANIFEST.MF
> > refers to it.
> > 09/02/03 14:01:33 WARN JPOX.Plugin: Extension Point
> > "org.eclipse.ui.bindings" not registered,
> but
> > plugin "org.eclipse.jdt.junit" defined in
> >
> file:/home/ssarkar/eclipse/eclipse32_3.4.1_working/configuration/org.eclipse.osgi/bundles/97/1/.cp/META-INF/MANIFEST.MF
> > refers to it.
> > 09/02/03 14:01:33 WARN JPOX.Plugin: Extension Point
> > "org.eclipse.core.runtime.preferences" not
> > registered, but plugin
> "org.eclipse.jdt.junit"
> > defined in
> >
> file:/home/ssarkar/eclipse/eclipse32_3.4.1_working/configuration/org.eclipse.osgi/bundles/97/1/.cp/META-INF/MANIFEST.MF
> > refers to it.
> > 09/02/03 14:01:33 WARN JPOX.Plugin: Extension Point
> >
> "org.eclipse.jdt.core.classpathContainerInitializer"
> > not registered, but plugin
> "org.eclipse.jdt.junit"
> > defined in
> >
> file:/home/ssarkar/eclipse/eclipse32_3.4.1_working/configuration/org.eclipse.osgi/bundles/97/1/.cp/META-INF/MANIFEST.MF
> > refers to it.
> > 09/02/03 14:01:33 WARN JPOX.Plugin: Extension Point
> > "org.eclipse.jdt.ui.classpathContainerPage"
> not
> > registered, but plugin
> "org.eclipse.jdt.junit"
> > defined in
> >
> file:/home/ssarkar/eclipse/eclipse32_3.4.1_working/configuration/org.eclipse.osgi/bundles/97/1/.cp/META-INF/MANIFEST.MF
> > refers to it.
> > 09/02/03 14:01:33 ERROR JPOX.Plugin: Bundle
> > "org.eclipse.jdt.junit" requires
> > "org.eclipse.ui.ide" but it cannot be
> resolved.
> > 09/02/03 14:01:33 ERROR JPOX.Plugin: Bundle
> > "org.eclipse.jdt.junit" requires
> > "org.eclipse.ui.views" but it cannot be
> resolved.
> > 09/02/03 14:01:33 ERROR JPOX.Plugin: Bundle
> > "org.eclipse.jdt.junit" requires
> > "org.eclipse.jface.text" but it cannot be
> > resolved.
> > 09/02/03 14:01:33 ERROR JPOX.Plugin: Bundle
> > "org.eclipse.jdt.junit" requires
> > "org.eclipse.ui.workbench.texteditor" but it
> > cannot be resolved.
> > 09/02/03 14:01:33 ERROR JPOX.Plugin: Bundle
> > "org.eclipse.jdt.junit" requires
> > "org.eclipse.ui.editors" but it cannot be
> > resolved.
> > 09/02/03 14:01:33 ERROR JPOX.Plugin: Bundle
> > "org.eclipse.jdt.junit" requires
> > "org.eclipse.ui" but it cannot be resolved.
> > 09/02/03 14:01:33 ERROR JPOX.Plugin: Bundle
> > "org.eclipse.jdt.junit" requires
> > "org.eclipse.core.expressions" but it cannot
> be
> > resolved.
> > 09/02/03 14:01:33 ERROR JPOX.Plugin: Bundle
> > "org.eclipse.jdt.junit" requires
> > "org.eclipse.core.resources" but it cannot
> be
> > resolved.
> > 09/02/03 14:01:33 ERROR JPOX.Plugin: Bundle
> > "org.eclipse.jdt.junit" requires
> > "org.eclipse.debug.core" but it cannot be
> > resolved.
> > 09/02/03 14:01:33 ERROR JPOX.Plugin: Bundle
> > "org.eclipse.jdt.junit" requires
> > "org.eclipse.debug.ui" but it cannot be
> resolved.
> > 09/02/03 14:01:33 ERROR JPOX.Plugin: Bundle
> > "org.eclipse.jdt.junit" requires
> > "org.eclipse.jdt.core" but it cannot be
> resolved.
> > 09/02/03 14:01:33 ERROR JPOX.Plugin: Bundle
> > "org.eclipse.jdt.junit" requires
> > "org.eclipse.jdt.ui" but it cannot be
> resolved.
> > 09/02/03 14:01:33 ERROR JPOX.Plugin: Bundle
> > "org.eclipse.jdt.junit" requires
> > "org.eclipse.core.runtime" but it cannot be
> > resolved.
> > 09/02/03 14:01:33 ERROR JPOX.Plugin: Bundle
> > "org.eclipse.jdt.junit" requires
> > "org.eclipse.jdt.launching" but it cannot be
> > resolved.
> > 09/02/03 14:01:33 ERROR JPOX.Plugin: Bundle
> > "org.eclipse.jdt.junit" requires
> > "org.eclipse.jdt.debug.ui" but it cannot be
> > resolved.
> > 09/02/03 14:01:33 ERROR JPOX.Plugin: Bundle
> > "org.eclipse.jdt.junit" requires
> > "org.eclipse.compare" but it cannot be
> resolved.
> > 09/02/03 14:01:33 ERROR JPOX.Plugin: Bundle
> > "org.eclipse.jdt.junit" requires
> > "org.eclipse.ltk.core.refactoring" but it
> cannot
> > be resolved.
> > 09/02/03 14:01:33 ERROR JPOX.Plugin: Bundle
> > "org.eclipse.jdt.junit" requires
> > "org.eclipse.core.variables" but it cannot
> be
> > resolved.
> > 09/02/03 14:01:33 ERROR JPOX.Plugin: Bundle
> > "org.eclipse.jdt.junit" requires
> > "org.eclipse.ltk.ui.refactoring" but it
> cannot be
> > resolved.
> > 09/02/03 14:01:33 ERROR JPOX.Plugin: Bundle
> > "org.eclipse.jdt.junit.runtime" requires
> > "org.junit" but it cannot be resolved.
> > 09/02/03 14:01:33 WARN JPOX.Plugin: Bundle
> > "org.jpox" has an optional dependency to
> > "org.eclipse.equinox.registry" but it cannot
> be
> > resolved
> > 09/02/03 14:01:33 WARN JPOX.Plugin: Bundle
> > "org.jpox" has an optional dependency to
> > "org.eclipse.core.runtime" but it cannot be
> > resolved
> > 09/02/03 14:01:33 INFO JPOX.Persistence:
> =================
> > Persistence Configuration ===============
> > 09/02/03 14:01:33 INFO JPOX.Persistence: JPOX
> Persistence
> > Factory - Vendor: "JPOX"  Version:
> > "1.2.2"
> > 09/02/03 14:01:33 INFO JPOX.Persistence: JPOX
> Persistence
> > Factory initialised for datastore
> >
> URL="jdbc:derby:;databaseName=../build/test/junit_metastore_db;create=true"
> >
> driver="org.apache.derby.jdbc.EmbeddedDriver"
> > userName="APP"
> > 09/02/03 14:01:33 INFO JPOX.Persistence:
> >
> ===========================================================
> > 09/02/03 14:01:35 INFO Datastore.Schema: Initialising
> > Catalog "", Schema "APP" using
> > "SchemaTable" auto-start option
> > 09/02/03 14:01:36 INFO Datastore.Schema: Catalog
> > "", Schema "APP" initialised -
> managing
> > 0 classes
> > 09/02/03 14:01:36 INFO JPOX.JDO: >> Found
> > StoreManager org.jpox.store.rdbms.RDBMSManager
> > java.lang.UnsupportedClassVersionError: Bad version
> number
> > in .class file
> >         at java.lang.ClassLoader.defineClass1(Native
> > Method)
> >         at
> >
> java.lang.ClassLoader.defineClass(ClassLoader.java:620)
> >         at
> >
> java.security.SecureClassLoader.defineClass(SecureClassLoader.java:124)
> >         at
> >
> java.net.URLClassLoader.defineClass(URLClassLoader.java:260)
> >         at
> >
> java.net.URLClassLoader.access$100(URLClassLoader.java:56)
> >         at
> > java.net.URLClassLoader$1.run(URLClassLoader.java:195)
> >         at
> > java.security.AccessController.doPrivileged(Native
> Method)
> >         at
> >
> java.net.URLClassLoader.findClass(URLClassLoader.java:188)
> >         at
> > java.lang.ClassLoader.loadClass(ClassLoader.java:306)
> >         at
> >
> sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:268)
> >         at
> > java.lang.ClassLoader.loadClass(ClassLoader.java:251)
> >         at
> >
> java.lang.ClassLoader.loadClassInternal(ClassLoader.java:319)
> >         at
> >
> org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:180)
> >         at
> >
> org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:194)
> >         at
> >
> org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:124)
> >         at
> >
> org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:103)
> >         at
> >
> org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:54)
> >         at
> >
> org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:82)
> >         at
> >
> org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:127)
> >         at
> >
> org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:143)
> >         at
> >
> org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:115)
> >         at
> >
> org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.<init>(HiveMetaStore.java:100)
> >         at
> >
> org.apache.hadoop.hive.metastore.HiveMetaStoreClient.<init>(HiveMetaStoreClient.java:73)
> >         at
> >
> org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:785)
> >         at
> >
> org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:798)
> >         at
> >
> org.apache.hadoop.hive.ql.metadata.Hive.dropTable(Hive.java:316)
> >         at
> >
> org.apache.hadoop.hive.ql.metadata.Hive.dropTable(Hive.java:300)
> >         at
> >
> org.apache.hadoop.hive.ql.exec.TestExecDriver.<clinit>(TestExecDriver.java:105)
> >         at
> >
> sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
> > Method)
> >         at
> >
> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39)
> >         at
> >
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27)
> >         at
> >
> java.lang.reflect.Constructor.newInstance(Constructor.java:494)
> >         at
> >
> junit.framework.TestSuite.createTest(TestSuite.java:131)
> >         at
> >
> junit.framework.TestSuite.addTestMethod(TestSuite.java:114)
> >         at
> >
> junit.framework.TestSuite.<init>(TestSuite.java:75)
> >         at
> >
> org.eclipse.jdt.internal.junit.runner.junit3.JUnit3TestLoader.getTest(JUnit3TestLoader.java:102)
> >         at
> >
> org.eclipse.jdt.internal.junit.runner.junit3.JUnit3TestLoader.loadTests(JUnit3TestLoader.java:59)
> >         at
> >
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:445)
> >         at
> >
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:673)
> >         at
> >
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:386)
> >         at
> >
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:196)
> > java.lang.ExceptionInInitializerError
> >         at
> >
> sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
> > Method)
> >         at
> >
> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39)
> >         at
> >
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27)
> >         at
> >
> java.lang.reflect.Constructor.newInstance(Constructor.java:494)
> >         at
> >
> junit.framework.TestSuite.createTest(TestSuite.java:131)
> >         at
> >
> junit.framework.TestSuite.addTestMethod(TestSuite.java:114)
> >         at
> >
> junit.framework.TestSuite.<init>(TestSuite.java:75)
> >         at
> >
> org.eclipse.jdt.internal.junit.runner.junit3.JUnit3TestLoader.getTest(JUnit3TestLoader.java:102)
> >         at
> >
> org.eclipse.jdt.internal.junit.runner.junit3.JUnit3TestLoader.loadTests(JUnit3TestLoader.java:59)
> >         at
> >
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:445)
> >         at
> >
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:673)
> >         at
> >
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:386)
> >         at
> >
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:196)
> > Caused by: java.lang.RuntimeException: Encountered
> > throwable
> >         at
> >
> org.apache.hadoop.hive.ql.exec.TestExecDriver.<clinit>(TestExecDriver.java:113)
> >         ... 13 more
> >
> ======================================================================
> >
> > regards,
> > Shyam
> >
> >
> >
> >
> >
> > --- On Tue, 2/3/09, Ashish Thusoo
> > <athusoo@facebook.com> wrote:
> >
> > > From: Ashish Thusoo <athusoo@facebook.com>
> > > Subject: RE: Eclipse run fails !!
> > > To: "Shyam Sarkar"
> > <shyam_sarkar@yahoo.com>,
> > > "hive-dev@hadoop.apache.org"
> > <hive-dev@hadoop.apache.org>
> > > Date: Tuesday, February 3, 2009, 1:46 PM Hi
> Shyam,
> > >
> > > I can certainly say that 0.17.0 should work with
> > eclipse. I have been
> > > doing that for a while.
> > >
> > > Maybe we can concentrate on fixing why you are
> not
> > able to create a
> > > table in hdfs. I am not sure why you could not
> create
> > the
> > > /user/hive/warehouse directory in 0.17. Are you
> saying
> > that
> > >
> > > hadoop dfs -mkdir /user/facebook/hive
> > >
> > > does not work for you? Can you send out the
> output
> > when you run this
> > > command.
> > >
> > > Ashish
> > >
> > > PS: using -Dhadoop.versoion="0.17.0"
> for all
> > the commands that are
> > > given in the wiki should make things work in
> eclipse.
> > >
> > > -----Original Message-----
> > > From: Shyam Sarkar
> [mailto:shyam_sarkar@yahoo.com]
> > > Sent: Tuesday, February 03, 2009 12:00 PM
> > > To: hive-dev@hadoop.apache.org; Ashish Thusoo
> > > Subject: RE: Eclipse run fails !!
> > >
> > > Dear Ashish,
> > >
> > > For the last few days I tried eclipse 3.4.1 with
> > 0.17.2.1 version and
> > > got the same errors with run->run. Then I
> looked
> > into bin/hive command
> > > and found that it could not create table in HDFS.
> The
> > reason was that
> > > I could not create /user/hive/warehouse directory
> > inside HDFS. It was
> > > using Linux FS.
> > > This is why I switched to 0.19.0 where
> directories in
> > HDFS can be
> > > created.
> > >
> > > Could you please tell me which exact version of
> hadoop
> > will work fine
> > > with eclipse runs ?  I want to get rid of errors
> in
> > project itself
> > > (before any run).
> > >
> > > Regards,
> > > Shyam
> > >
> > > --- On Tue, 2/3/09, Ashish Thusoo
> > > <athusoo@facebook.com> wrote:
> > >
> > > > From: Ashish Thusoo
> <athusoo@facebook.com>
> > > > Subject: RE: Eclipse run fails !!
> > > > To: "hive-dev@hadoop.apache.org"
> > > <hive-dev@hadoop.apache.org>,
> > > > "shyam_sarkar@yahoo.com"
> > > <shyam_sarkar@yahoo.com>
> > > > Date: Tuesday, February 3, 2009, 11:38 AM Hi
> > Shyam,
> > > >
> > > > We have not really tried the eclipse stuff
> for
> > 0.19.0.
> > > Is it possible
> > > > for you to use 0.17.0 for now, while we
> figure
> > this
> > > out...
> > > >
> > > > Ashish
> > > >
> > > > -----Original Message-----
> > > > From: Shyam Sarkar
> > [mailto:shyam_sarkar@yahoo.com]
> > > > Sent: Tuesday, February 03, 2009 11:26 AM
> > > > To: hive-dev@hadoop.apache.org
> > > > Subject: Eclipse run fails !!
> > > >
> > > > Hello,
> > > >
> > > > I have hive project loaded inside eclipse
> 3.4.1
> > and
> > > hadoop 0.19.0 is
> > > > running in the background. I could create
> tables
> > from
> > > bin/hive
> > > > command.
> > > > But when I try to run->run inside eclipse
> it
> > says::
> > > >
> > > > "Errors exist with required project(s):
> > > >
> > > > hive
> > > >
> > > > Proceed with launch ?"
> > > >
> > > > and then it gives many errors.
> > > >
> > > > Can someone please tell me why there are
> errors
> > in
> > > project hive ?  I
> > > > followed all steps correctly from hive wiki.
> > > >
> > > > Regards,
> > > > shyam_sarkar@yahoo.com


      

Mime
View raw message