hadoop-hive-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From zs...@apache.org
Subject svn commit: r902921 [1/26] - in /hadoop/hive/trunk: ./ contrib/src/java/org/apache/hadoop/hive/contrib/genericudf/example/ contrib/src/java/org/apache/hadoop/hive/contrib/udtf/example/ ql/src/java/org/apache/hadoop/hive/ql/ ql/src/java/org/apache/hadoo...
Date Mon, 25 Jan 2010 18:49:05 GMT
Author: zshao
Date: Mon Jan 25 18:48:58 2010
New Revision: 902921

URL: http://svn.apache.org/viewvc?rev=902921&view=rev
Log:
HIVE-1084. Capitalize first letter of class names. (Carl Steinbach via zshao)

Added:
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/Description.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/description.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/JoinCond.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/joinCond.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/JoinType.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/joinType.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/AggregationDesc.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/aggregationDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/AlterTableDesc.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/alterTableDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/CollectDesc.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/collectDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/CopyWork.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/copyWork.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/CreateFunctionDesc.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/createFunctionDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/CreateTableDesc.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/createTableDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/CreateTableLikeDesc.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/createTableLikeDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/CreateViewDesc.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/createViewDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/DDLDesc.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/ddlDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/DescFunctionDesc.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/descFunctionDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/DescTableDesc.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/descTableDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/DropFunctionDesc.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/dropFunctionDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/DropTableDesc.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/dropTableDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/Explain.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/explain.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/ExplainWork.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/explainWork.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/ExplosionDesc.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/explosionDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/ExprNodeColumnDesc.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/exprNodeColumnDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/ExprNodeConstantDesc.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/exprNodeConstantDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/ExprNodeDesc.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/exprNodeDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/ExprNodeFieldDesc.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/exprNodeFieldDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/ExprNodeGenericFuncDesc.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/exprNodeGenericFuncDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/ExprNodeNullDesc.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/exprNodeNullDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/ExtractDesc.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/extractDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/FetchWork.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/fetchWork.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/FileSinkDesc.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/fileSinkDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/FilterDesc.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/filterDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/ForwardDesc.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/forwardDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/GroupByDesc.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/groupByDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/JoinCondDesc.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/joinCond.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/JoinDesc.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/joinDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/LateralViewJoinDesc.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/lateralViewJoinDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/LimitDesc.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/limitDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/LoadDesc.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/loadDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/LoadFileDesc.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/loadFileDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/LoadTableDesc.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/loadTableDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/MapJoinDesc.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/mapJoinDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/MapredLocalWork.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/mapredLocalWork.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/MapredWork.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/mapredWork.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/MoveWork.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/moveWork.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/PartitionDesc.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/partitionDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/ReduceSinkDesc.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/reduceSinkDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/SchemaDesc.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/schemaDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/ScriptDesc.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/scriptDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/SelectDesc.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/selectDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/ShowFunctionsDesc.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/showFunctionsDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/ShowPartitionsDesc.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/showPartitionsDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/ShowTableStatusDesc.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/showTableStatusDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/ShowTablesDesc.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/showTablesDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/TableDesc.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/tableDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/TableScanDesc.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/tableScanDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/UDTFDesc.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/udtfDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/UnionDesc.java   (contents, props changed)
      - copied, changed from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/unionDesc.java
Removed:
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/description.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/joinCond.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/joinType.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/aggregationDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/alterTableDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/collectDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/copyWork.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/createFunctionDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/createTableDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/createTableLikeDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/createViewDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/ddlDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/descFunctionDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/descTableDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/dropFunctionDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/dropTableDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/explain.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/explainWork.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/explosionDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/exprNodeColumnDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/exprNodeConstantDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/exprNodeDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/exprNodeFieldDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/exprNodeGenericFuncDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/exprNodeNullDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/extractDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/fetchWork.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/fileSinkDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/filterDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/forwardDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/groupByDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/joinCond.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/joinDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/lateralViewJoinDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/limitDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/loadDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/loadFileDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/loadTableDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/mapJoinDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/mapredLocalWork.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/mapredWork.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/moveWork.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/partitionDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/reduceSinkDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/schemaDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/scriptDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/selectDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/showFunctionsDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/showPartitionsDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/showTableStatusDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/showTablesDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/tableDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/tableScanDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/udtfDesc.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/unionDesc.java
Modified:
    hadoop/hive/trunk/CHANGES.txt
    hadoop/hive/trunk/contrib/src/java/org/apache/hadoop/hive/contrib/genericudf/example/GenericUDFDBOutput.java
    hadoop/hive/trunk/contrib/src/java/org/apache/hadoop/hive/contrib/udtf/example/GenericUDTFExplode2.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/Driver.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/CollectOperator.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/CommonJoinOperator.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/CopyTask.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/DDLTask.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExecDriver.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExecMapper.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExecReducer.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExplainTask.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExprNodeColumnEvaluator.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExprNodeConstantEvaluator.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExprNodeEvaluatorFactory.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExprNodeFieldEvaluator.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExprNodeGenericFuncEvaluator.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExprNodeNullEvaluator.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExtractOperator.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/FetchOperator.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/FetchTask.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/FileSinkOperator.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/FilterOperator.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ForwardOperator.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionTask.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/GroupByOperator.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/JoinOperator.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/LateralViewJoinOperator.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/LimitOperator.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/MapJoinOperator.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/MapOperator.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/MapRedTask.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/MoveTask.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/Operator.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/OperatorFactory.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ReduceSinkOperator.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ScriptOperator.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/SelectOperator.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/SkewJoinHandler.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/TableScanOperator.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/TaskFactory.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/UDTFOperator.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/UnionOperator.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/Utilities.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/persistence/RowContainer.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/io/CombineHiveInputFormat.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/io/HiveFileFormatUtils.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/io/HiveInputFormat.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/optimizer/ColumnPrunerProcCtx.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/optimizer/ColumnPrunerProcFactory.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/optimizer/GenMRFileSink1.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/optimizer/GenMRProcContext.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/optimizer/GenMRRedSink1.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/optimizer/GenMRRedSink3.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/optimizer/GenMRRedSink4.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/optimizer/GenMRUnion1.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/optimizer/GenMapRedUtils.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/optimizer/GroupByOptimizer.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/optimizer/MapJoinFactory.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/optimizer/MapJoinProcessor.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/optimizer/SamplePruner.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/GenMRSkewJoinProcessor.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/SkewJoinResolver.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/optimizer/ppr/ExprProcFactory.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/optimizer/ppr/OpProcFactory.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/optimizer/ppr/OpWalkerCtx.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/optimizer/ppr/PartitionPruner.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/DDLSemanticAnalyzer.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/ExplainSemanticAnalyzer.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/FunctionSemanticAnalyzer.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/LoadSemanticAnalyzer.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/ParseContext.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/QB.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/QBJoinTree.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/TypeCheckProcFactory.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/ConditionalResolverMergeFiles.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/ConditionalWork.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/DDLWork.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/FunctionWork.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/PlanUtils.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/ppd/ExprWalkerInfo.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/ppd/ExprWalkerProcFactory.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/ppd/OpProcFactory.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDAFMax.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDAFMin.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFAbs.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFAcos.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFAscii.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFAsin.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFBin.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFCeil.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFConcat.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFConv.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFCos.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFDate.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFDateAdd.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFDateDiff.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFDateSub.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFDayOfMonth.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFExp.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFFindInSet.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFFloor.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFFromUnixTime.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFHex.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFHour.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFJson.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFLTrim.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFLength.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFLike.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFLn.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFLog.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFLog10.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFLog2.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFLower.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFLpad.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFMinute.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFMonth.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPAnd.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPBitAnd.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPBitNot.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPBitOr.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPBitXor.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPDivide.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPEqual.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPEqualOrGreaterThan.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPEqualOrLessThan.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPGreaterThan.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPLessThan.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPLongDivide.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPMinus.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPMod.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPMultiply.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPNegative.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPNot.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPNotEqual.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPOr.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPPlus.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPPositive.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFParseUrl.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFPosMod.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFPower.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFRTrim.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFRand.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFRegExp.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFRegExpExtract.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFRegExpReplace.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFRepeat.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFReverse.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFRound.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFRpad.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFSecond.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFSin.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFSpace.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFSqrt.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFSubstr.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFTrim.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFUnhex.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFUnixTimeStamp.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFUpper.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFWeekOfYear.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFYear.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFAverage.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFCount.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFStd.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFStdSample.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFSum.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFVariance.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFVarianceSample.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFArray.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFCoalesce.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFConcatWS.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFElt.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFField.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFHash.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFIndex.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFInstr.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFLocate.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFMap.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFOPNotNull.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFOPNull.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFSize.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFSplit.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDTFExplode.java
    hadoop/hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/exec/TestExecDriver.java
    hadoop/hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/exec/TestExpressionEvaluator.java
    hadoop/hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/exec/TestOperators.java
    hadoop/hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/exec/TestPlan.java
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/case_sensitivity.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/cast1.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/groupby1.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/groupby2.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/groupby3.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/groupby4.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/groupby5.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/groupby6.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/input1.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/input2.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/input20.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/input3.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/input4.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/input5.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/input6.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/input7.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/input8.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/input9.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/input_part1.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/input_testsequencefile.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/input_testxpath.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/input_testxpath2.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/join1.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/join2.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/join3.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/join4.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/join5.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/join6.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/join7.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/join8.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/sample1.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/sample2.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/sample3.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/sample4.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/sample5.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/sample6.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/sample7.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/subq.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/udf1.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/udf4.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/udf6.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/udf_case.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/udf_when.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/union.q.xml

Modified: hadoop/hive/trunk/CHANGES.txt
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/CHANGES.txt?rev=902921&r1=902920&r2=902921&view=diff
==============================================================================
--- hadoop/hive/trunk/CHANGES.txt (original)
+++ hadoop/hive/trunk/CHANGES.txt Mon Jan 25 18:48:58 2010
@@ -28,6 +28,9 @@
     HIVE-1081. Automated source code cleanup - Part 3 - ql.
     (Carl Steinbach via zshao)
 
+    HIVE-1084. Capitalize first letter of class names.
+    (Carl Steinbach via zshao)
+
   OPTIMIZATIONS
 
   BUG FIXES

Modified: hadoop/hive/trunk/contrib/src/java/org/apache/hadoop/hive/contrib/genericudf/example/GenericUDFDBOutput.java
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/contrib/src/java/org/apache/hadoop/hive/contrib/genericudf/example/GenericUDFDBOutput.java?rev=902921&r1=902920&r2=902921&view=diff
==============================================================================
--- hadoop/hive/trunk/contrib/src/java/org/apache/hadoop/hive/contrib/genericudf/example/GenericUDFDBOutput.java (original)
+++ hadoop/hive/trunk/contrib/src/java/org/apache/hadoop/hive/contrib/genericudf/example/GenericUDFDBOutput.java Mon Jan 25 18:48:58 2010
@@ -25,7 +25,7 @@
 import org.apache.commons.logging.Log;
 import org.apache.commons.logging.LogFactory;
 import org.apache.hadoop.hive.ql.exec.UDFArgumentTypeException;
-import org.apache.hadoop.hive.ql.exec.description;
+import org.apache.hadoop.hive.ql.exec.Description;
 import org.apache.hadoop.hive.ql.metadata.HiveException;
 import org.apache.hadoop.hive.ql.udf.UDFType;
 import org.apache.hadoop.hive.ql.udf.generic.GenericUDF;
@@ -51,7 +51,7 @@
  * Use hive's ADD JAR feature to add your JDBC Driver to the distributed cache,
  * otherwise GenericUDFDBoutput will fail.
  */
-@description(name = "dboutput", value = "_FUNC_(jdbcstring,username,password,preparedstatement,[arguments]) - sends data to a jdbc driver", extended = "argument 0 is the JDBC connection string\n"
+@Description(name = "dboutput", value = "_FUNC_(jdbcstring,username,password,preparedstatement,[arguments]) - sends data to a jdbc driver", extended = "argument 0 is the JDBC connection string\n"
     + "argument 1 is the user name\n"
     + "argument 2 is the password\n"
     + "argument 3 is an SQL query to be used in the PreparedStatement\n"

Modified: hadoop/hive/trunk/contrib/src/java/org/apache/hadoop/hive/contrib/udtf/example/GenericUDTFExplode2.java
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/contrib/src/java/org/apache/hadoop/hive/contrib/udtf/example/GenericUDTFExplode2.java?rev=902921&r1=902920&r2=902921&view=diff
==============================================================================
--- hadoop/hive/trunk/contrib/src/java/org/apache/hadoop/hive/contrib/udtf/example/GenericUDTFExplode2.java (original)
+++ hadoop/hive/trunk/contrib/src/java/org/apache/hadoop/hive/contrib/udtf/example/GenericUDTFExplode2.java Mon Jan 25 18:48:58 2010
@@ -22,7 +22,7 @@
 import java.util.List;
 
 import org.apache.hadoop.hive.ql.exec.UDFArgumentException;
-import org.apache.hadoop.hive.ql.exec.description;
+import org.apache.hadoop.hive.ql.exec.Description;
 import org.apache.hadoop.hive.ql.metadata.HiveException;
 import org.apache.hadoop.hive.ql.udf.generic.GenericUDTF;
 import org.apache.hadoop.hive.serde2.objectinspector.ListObjectInspector;
@@ -30,7 +30,7 @@
 import org.apache.hadoop.hive.serde2.objectinspector.ObjectInspectorFactory;
 import org.apache.hadoop.hive.serde2.objectinspector.StructObjectInspector;
 
-@description(name = "explode2", value = "_FUNC_(a) - like explode, but outputs two identical columns (for "
+@Description(name = "explode2", value = "_FUNC_(a) - like explode, but outputs two identical columns (for "
     + "testing purposes)")
 public class GenericUDTFExplode2 extends GenericUDTF {
 

Modified: hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/Driver.java
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/Driver.java?rev=902921&r1=902920&r2=902921&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/Driver.java (original)
+++ hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/Driver.java Mon Jan 25 18:48:58 2010
@@ -59,7 +59,7 @@
 import org.apache.hadoop.hive.ql.parse.ParseUtils;
 import org.apache.hadoop.hive.ql.parse.SemanticAnalyzerFactory;
 import org.apache.hadoop.hive.ql.parse.SemanticException;
-import org.apache.hadoop.hive.ql.plan.tableDesc;
+import org.apache.hadoop.hive.ql.plan.TableDesc;
 import org.apache.hadoop.hive.ql.processors.CommandProcessor;
 import org.apache.hadoop.hive.ql.session.SessionState;
 import org.apache.hadoop.hive.ql.session.SessionState.LogHelper;
@@ -152,7 +152,7 @@
         }
         FetchTask ft = (FetchTask) sem.getFetchTask();
 
-        tableDesc td = ft.getTblDesc();
+        TableDesc td = ft.getTblDesc();
         // partitioned tables don't have tableDesc set on the FetchTask. Instead
         // they have a list of PartitionDesc objects, each with a table desc.
         // Let's

Modified: hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/CollectOperator.java
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/CollectOperator.java?rev=902921&r1=902920&r2=902921&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/CollectOperator.java (original)
+++ hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/CollectOperator.java Mon Jan 25 18:48:58 2010
@@ -23,7 +23,7 @@
 
 import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.hive.ql.metadata.HiveException;
-import org.apache.hadoop.hive.ql.plan.collectDesc;
+import org.apache.hadoop.hive.ql.plan.CollectDesc;
 import org.apache.hadoop.hive.serde2.objectinspector.InspectableObject;
 import org.apache.hadoop.hive.serde2.objectinspector.ObjectInspector;
 import org.apache.hadoop.hive.serde2.objectinspector.ObjectInspectorUtils;
@@ -31,7 +31,7 @@
 /**
  * Buffers rows emitted by other operators
  **/
-public class CollectOperator extends Operator<collectDesc> implements
+public class CollectOperator extends Operator<CollectDesc> implements
     Serializable {
 
   private static final long serialVersionUID = 1L;

Modified: hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/CommonJoinOperator.java
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/CommonJoinOperator.java?rev=902921&r1=902920&r2=902921&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/CommonJoinOperator.java (original)
+++ hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/CommonJoinOperator.java Mon Jan 25 18:48:58 2010
@@ -34,10 +34,10 @@
 import org.apache.hadoop.hive.ql.exec.persistence.RowContainer;
 import org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat;
 import org.apache.hadoop.hive.ql.metadata.HiveException;
-import org.apache.hadoop.hive.ql.plan.exprNodeDesc;
-import org.apache.hadoop.hive.ql.plan.joinCond;
-import org.apache.hadoop.hive.ql.plan.joinDesc;
-import org.apache.hadoop.hive.ql.plan.tableDesc;
+import org.apache.hadoop.hive.ql.plan.ExprNodeDesc;
+import org.apache.hadoop.hive.ql.plan.JoinCondDesc;
+import org.apache.hadoop.hive.ql.plan.JoinDesc;
+import org.apache.hadoop.hive.ql.plan.TableDesc;
 import org.apache.hadoop.hive.serde2.SerDe;
 import org.apache.hadoop.hive.serde2.SerDeException;
 import org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe;
@@ -52,7 +52,7 @@
 /**
  * Join operator implementation.
  */
-public abstract class CommonJoinOperator<T extends joinDesc> extends
+public abstract class CommonJoinOperator<T extends JoinDesc> extends
     Operator<T> implements Serializable {
   private static final long serialVersionUID = 1L;
   static final protected Log LOG = LogFactory.getLog(CommonJoinOperator.class
@@ -104,7 +104,7 @@
 
   transient static protected Byte[] order; // order in which the results should
                                            // be output
-  transient protected joinCond[] condn;
+  transient protected JoinCondDesc[] condn;
   transient protected boolean noOuterJoin;
   transient private Object[] dummyObj; // for outer joins, contains the
   // potential nulls for the concerned
@@ -122,7 +122,7 @@
   transient private Map<Integer, Set<String>> posToAliasMap;
 
   transient LazyBinarySerDe[] spillTableSerDe;
-  transient protected Map<Byte, tableDesc> spillTableDesc; // spill tables are
+  transient protected Map<Byte, TableDesc> spillTableDesc; // spill tables are
                                                            // used if the join
                                                            // input is too large
                                                            // to fit in memory
@@ -137,17 +137,17 @@
   transient boolean handleSkewJoin = false;
 
   protected int populateJoinKeyValue(Map<Byte, List<ExprNodeEvaluator>> outMap,
-      Map<Byte, List<exprNodeDesc>> inputMap) {
+      Map<Byte, List<ExprNodeDesc>> inputMap) {
 
     int total = 0;
 
-    Iterator<Map.Entry<Byte, List<exprNodeDesc>>> entryIter = inputMap
+    Iterator<Map.Entry<Byte, List<ExprNodeDesc>>> entryIter = inputMap
         .entrySet().iterator();
     while (entryIter.hasNext()) {
-      Map.Entry<Byte, List<exprNodeDesc>> e = entryIter.next();
+      Map.Entry<Byte, List<ExprNodeDesc>> e = entryIter.next();
       Byte key = order[e.getKey()];
 
-      List<exprNodeDesc> expr = e.getValue();
+      List<ExprNodeDesc> expr = e.getValue();
       int sz = expr.size();
       total += sz;
 
@@ -199,7 +199,7 @@
 
   }
 
-  protected static <T extends joinDesc> ObjectInspector getJoinOutputObjectInspector(
+  protected static <T extends JoinDesc> ObjectInspector getJoinOutputObjectInspector(
       Byte[] order, Map<Byte, List<ObjectInspector>> aliasToObjectInspectors,
       T conf) {
     ArrayList<ObjectInspector> structFieldObjectInspectors = new ArrayList<ObjectInspector>();
@@ -288,7 +288,7 @@
 
   RowContainer getRowContainer(Configuration hconf, byte pos, Byte alias,
       int containerSize) throws HiveException {
-    tableDesc tblDesc = getSpillTableDesc(alias);
+    TableDesc tblDesc = getSpillTableDesc(alias);
     SerDe serde = getSpillSerDe(alias);
 
     if (serde == null) {
@@ -311,7 +311,7 @@
   }
 
   private SerDe getSpillSerDe(byte alias) {
-    tableDesc desc = getSpillTableDesc(alias);
+    TableDesc desc = getSpillTableDesc(alias);
     if (desc == null) {
       return null;
     }
@@ -328,14 +328,14 @@
 
   transient boolean newGroupStarted = false;
 
-  public tableDesc getSpillTableDesc(Byte alias) {
+  public TableDesc getSpillTableDesc(Byte alias) {
     if (spillTableDesc == null || spillTableDesc.size() == 0) {
       initSpillTables();
     }
     return spillTableDesc.get(alias);
   }
 
-  public Map<Byte, tableDesc> getSpillTableDesc() {
+  public Map<Byte, TableDesc> getSpillTableDesc() {
     if (spillTableDesc == null) {
       initSpillTables();
     }
@@ -343,10 +343,10 @@
   }
 
   private void initSpillTables() {
-    Map<Byte, List<exprNodeDesc>> exprs = conf.getExprs();
-    spillTableDesc = new HashMap<Byte, tableDesc>(exprs.size());
+    Map<Byte, List<ExprNodeDesc>> exprs = conf.getExprs();
+    spillTableDesc = new HashMap<Byte, TableDesc>(exprs.size());
     for (int tag = 0; tag < exprs.size(); tag++) {
-      List<exprNodeDesc> valueCols = exprs.get((byte) tag);
+      List<ExprNodeDesc> valueCols = exprs.get((byte) tag);
       int columnSize = valueCols.size();
       StringBuffer colNames = new StringBuffer();
       StringBuffer colTypes = new StringBuffer();
@@ -364,7 +364,7 @@
       // remove the last ','
       colNames.setLength(colNames.length() - 1);
       colTypes.setLength(colTypes.length() - 1);
-      tableDesc tblDesc = new tableDesc(LazyBinarySerDe.class,
+      TableDesc tblDesc = new TableDesc(LazyBinarySerDe.class,
           SequenceFileInputFormat.class, HiveSequenceFileOutputFormat.class,
           Utilities.makeProperties(
               org.apache.hadoop.hive.serde.Constants.SERIALIZATION_FORMAT, ""
@@ -656,7 +656,7 @@
     int type = condn[joinPos - 1].getType();
 
     // process all nulls for RIGHT and FULL OUTER JOINS
-    if (((type == joinDesc.RIGHT_OUTER_JOIN) || (type == joinDesc.FULL_OUTER_JOIN))
+    if (((type == JoinDesc.RIGHT_OUTER_JOIN) || (type == JoinDesc.FULL_OUTER_JOIN))
         && !newObjNull && (inputNulls == null) && firstRow) {
       boolean[] newNulls = new boolean[intObj.getCurSize()];
       for (int i = 0; i < newNulls.length - 1; i++) {
@@ -671,21 +671,21 @@
       return null;
     }
 
-    if (type == joinDesc.INNER_JOIN) {
+    if (type == JoinDesc.INNER_JOIN) {
       return joinObjectsInnerJoin(resNulls, inputNulls, newObj, intObj, left,
           newObjNull);
-    } else if (type == joinDesc.LEFT_OUTER_JOIN) {
+    } else if (type == JoinDesc.LEFT_OUTER_JOIN) {
       return joinObjectsLeftOuterJoin(resNulls, inputNulls, newObj, intObj,
           left, newObjNull);
-    } else if (type == joinDesc.RIGHT_OUTER_JOIN) {
+    } else if (type == JoinDesc.RIGHT_OUTER_JOIN) {
       return joinObjectsRightOuterJoin(resNulls, inputNulls, newObj, intObj,
           left, newObjNull, firstRow);
-    } else if (type == joinDesc.LEFT_SEMI_JOIN) {
+    } else if (type == JoinDesc.LEFT_SEMI_JOIN) {
       return joinObjectsLeftSemiJoin(resNulls, inputNulls, newObj, intObj,
           left, newObjNull);
     }
 
-    assert (type == joinDesc.FULL_OUTER_JOIN);
+    assert (type == JoinDesc.FULL_OUTER_JOIN);
     return joinObjectsFullOuterJoin(resNulls, inputNulls, newObj, intObj, left,
         newObjNull, firstRow);
   }
@@ -712,7 +712,7 @@
 
         // check for skipping in case of left semi join
         if (aliasNum > 0
-            && condn[aliasNum - 1].getType() == joinDesc.LEFT_SEMI_JOIN
+            && condn[aliasNum - 1].getType() == JoinDesc.LEFT_SEMI_JOIN
             && newObj != dummyObj[aliasNum]) { // successful match
           skipping = true;
         }
@@ -784,7 +784,7 @@
   }
 
   protected void checkAndGenObject() throws HiveException {
-    if (condn[0].getType() == joinDesc.UNIQUE_JOIN) {
+    if (condn[0].getType() == JoinDesc.UNIQUE_JOIN) {
       new IntermediateObject(new ArrayList[numAliases], 0);
 
       // Check if results need to be emitted.

Modified: hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/CopyTask.java
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/CopyTask.java?rev=902921&r1=902920&r2=902921&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/CopyTask.java (original)
+++ hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/CopyTask.java Mon Jan 25 18:48:58 2010
@@ -25,14 +25,14 @@
 import org.apache.hadoop.fs.FileUtil;
 import org.apache.hadoop.fs.Path;
 import org.apache.hadoop.hive.ql.parse.LoadSemanticAnalyzer;
-import org.apache.hadoop.hive.ql.plan.copyWork;
+import org.apache.hadoop.hive.ql.plan.CopyWork;
 import org.apache.hadoop.hive.ql.plan.api.StageType;
 import org.apache.hadoop.util.StringUtils;
 
 /**
  * CopyTask implementation
  **/
-public class CopyTask extends Task<copyWork> implements Serializable {
+public class CopyTask extends Task<CopyWork> implements Serializable {
 
   private static final long serialVersionUID = 1L;
 

Modified: hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/DDLTask.java
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/DDLTask.java?rev=902921&r1=902920&r2=902921&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/DDLTask.java (original)
+++ hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/DDLTask.java Mon Jan 25 18:48:58 2010
@@ -66,17 +66,17 @@
 import org.apache.hadoop.hive.ql.plan.AddPartitionDesc;
 import org.apache.hadoop.hive.ql.plan.DDLWork;
 import org.apache.hadoop.hive.ql.plan.MsckDesc;
-import org.apache.hadoop.hive.ql.plan.alterTableDesc;
-import org.apache.hadoop.hive.ql.plan.createTableDesc;
-import org.apache.hadoop.hive.ql.plan.createTableLikeDesc;
-import org.apache.hadoop.hive.ql.plan.createViewDesc;
-import org.apache.hadoop.hive.ql.plan.descFunctionDesc;
-import org.apache.hadoop.hive.ql.plan.descTableDesc;
-import org.apache.hadoop.hive.ql.plan.dropTableDesc;
-import org.apache.hadoop.hive.ql.plan.showFunctionsDesc;
-import org.apache.hadoop.hive.ql.plan.showPartitionsDesc;
-import org.apache.hadoop.hive.ql.plan.showTableStatusDesc;
-import org.apache.hadoop.hive.ql.plan.showTablesDesc;
+import org.apache.hadoop.hive.ql.plan.AlterTableDesc;
+import org.apache.hadoop.hive.ql.plan.CreateTableDesc;
+import org.apache.hadoop.hive.ql.plan.CreateTableLikeDesc;
+import org.apache.hadoop.hive.ql.plan.CreateViewDesc;
+import org.apache.hadoop.hive.ql.plan.DescFunctionDesc;
+import org.apache.hadoop.hive.ql.plan.DescTableDesc;
+import org.apache.hadoop.hive.ql.plan.DropTableDesc;
+import org.apache.hadoop.hive.ql.plan.ShowFunctionsDesc;
+import org.apache.hadoop.hive.ql.plan.ShowPartitionsDesc;
+import org.apache.hadoop.hive.ql.plan.ShowTableStatusDesc;
+import org.apache.hadoop.hive.ql.plan.ShowTablesDesc;
 import org.apache.hadoop.hive.ql.plan.api.StageType;
 import org.apache.hadoop.hive.serde.Constants;
 import org.apache.hadoop.hive.serde2.Deserializer;
@@ -117,27 +117,27 @@
     try {
       db = Hive.get(conf);
 
-      createTableDesc crtTbl = work.getCreateTblDesc();
+      CreateTableDesc crtTbl = work.getCreateTblDesc();
       if (crtTbl != null) {
         return createTable(db, crtTbl);
       }
 
-      createTableLikeDesc crtTblLike = work.getCreateTblLikeDesc();
+      CreateTableLikeDesc crtTblLike = work.getCreateTblLikeDesc();
       if (crtTblLike != null) {
         return createTableLike(db, crtTblLike);
       }
 
-      dropTableDesc dropTbl = work.getDropTblDesc();
+      DropTableDesc dropTbl = work.getDropTblDesc();
       if (dropTbl != null) {
         return dropTable(db, dropTbl);
       }
 
-      alterTableDesc alterTbl = work.getAlterTblDesc();
+      AlterTableDesc alterTbl = work.getAlterTblDesc();
       if (alterTbl != null) {
         return alterTable(db, alterTbl);
       }
 
-      createViewDesc crtView = work.getCreateViewDesc();
+      CreateViewDesc crtView = work.getCreateViewDesc();
       if (crtView != null) {
         return createView(db, crtView);
       }
@@ -152,32 +152,32 @@
         return msck(db, msckDesc);
       }
 
-      descTableDesc descTbl = work.getDescTblDesc();
+      DescTableDesc descTbl = work.getDescTblDesc();
       if (descTbl != null) {
         return describeTable(db, descTbl);
       }
 
-      descFunctionDesc descFunc = work.getDescFunctionDesc();
+      DescFunctionDesc descFunc = work.getDescFunctionDesc();
       if (descFunc != null) {
         return describeFunction(descFunc);
       }
 
-      showTablesDesc showTbls = work.getShowTblsDesc();
+      ShowTablesDesc showTbls = work.getShowTblsDesc();
       if (showTbls != null) {
         return showTables(db, showTbls);
       }
 
-      showTableStatusDesc showTblStatus = work.getShowTblStatusDesc();
+      ShowTableStatusDesc showTblStatus = work.getShowTblStatusDesc();
       if (showTblStatus != null) {
         return showTableStatus(db, showTblStatus);
       }
 
-      showFunctionsDesc showFuncs = work.getShowFuncsDesc();
+      ShowFunctionsDesc showFuncs = work.getShowFuncsDesc();
       if (showFuncs != null) {
         return showFunctions(showFuncs);
       }
 
-      showPartitionsDesc showParts = work.getShowPartsDesc();
+      ShowPartitionsDesc showParts = work.getShowPartsDesc();
       if (showParts != null) {
         return showPartitions(db, showParts);
       }
@@ -360,7 +360,7 @@
    * @throws HiveException
    *           Throws this exception if an unexpected error occurs.
    */
-  private int showPartitions(Hive db, showPartitionsDesc showParts)
+  private int showPartitions(Hive db, ShowPartitionsDesc showParts)
       throws HiveException {
     // get the partitions for the table and populate the output
     String tabName = showParts.getTabName();
@@ -413,7 +413,7 @@
    * @throws HiveException
    *           Throws this exception if an unexpected error occurs.
    */
-  private int showTables(Hive db, showTablesDesc showTbls) throws HiveException {
+  private int showTables(Hive db, ShowTablesDesc showTbls) throws HiveException {
     // get the tables for the desired pattenn - populate the output stream
     List<String> tbls = null;
     if (showTbls.getPattern() != null) {
@@ -458,7 +458,7 @@
    * @throws HiveException
    *           Throws this exception if an unexpected error occurs.
    */
-  private int showFunctions(showFunctionsDesc showFuncs) throws HiveException {
+  private int showFunctions(ShowFunctionsDesc showFuncs) throws HiveException {
     // get the tables for the desired pattenn - populate the output stream
     Set<String> funcs = null;
     if (showFuncs.getPattern() != null) {
@@ -501,7 +501,7 @@
    *          is the function we are describing
    * @throws HiveException
    */
-  private int describeFunction(descFunctionDesc descFunc) throws HiveException {
+  private int describeFunction(DescFunctionDesc descFunc) throws HiveException {
     String funcName = descFunc.getName();
 
     // write the results in the file
@@ -510,14 +510,14 @@
       DataOutput outStream = (DataOutput) fs.create(descFunc.getResFile());
 
       // get the function documentation
-      description desc = null;
+      Description desc = null;
       Class<?> funcClass = null;
       FunctionInfo functionInfo = FunctionRegistry.getFunctionInfo(funcName);
       if (functionInfo != null) {
         funcClass = functionInfo.getFunctionClass();
       }
       if (funcClass != null) {
-        desc = funcClass.getAnnotation(description.class);
+        desc = funcClass.getAnnotation(Description.class);
       }
       if (desc != null) {
         outStream.writeBytes(desc.value().replace("_FUNC_", funcName));
@@ -564,7 +564,7 @@
    *          tables we are interested in
    * @return Return 0 when execution succeeds and above 0 if it fails.
    */
-  private int showTableStatus(Hive db, showTableStatusDesc showTblStatus)
+  private int showTableStatus(Hive db, ShowTableStatusDesc showTblStatus)
       throws HiveException {
     // get the tables for the desired pattenn - populate the output stream
     List<Table> tbls = new ArrayList<Table>();
@@ -688,7 +688,7 @@
    * @throws HiveException
    *           Throws this exception if an unexpected error occurs.
    */
-  private int describeTable(Hive db, descTableDesc descTbl)
+  private int describeTable(Hive db, DescTableDesc descTbl)
       throws HiveException {
     String colPath = descTbl.getTableName();
     String tableName = colPath.substring(0,
@@ -943,7 +943,7 @@
    * @throws HiveException
    *           Throws this exception if an unexpected error occurs.
    */
-  private int alterTable(Hive db, alterTableDesc alterTbl) throws HiveException {
+  private int alterTable(Hive db, AlterTableDesc alterTbl) throws HiveException {
     // alter the table
     Table tbl = db.getTable(MetaStoreUtils.DEFAULT_DATABASE_NAME, alterTbl
         .getOldName());
@@ -953,9 +953,9 @@
     }
     Table oldTbl = tbl.copy();
 
-    if (alterTbl.getOp() == alterTableDesc.alterTableTypes.RENAME) {
+    if (alterTbl.getOp() == AlterTableDesc.alterTableTypes.RENAME) {
       tbl.getTTable().setTableName(alterTbl.getNewName());
-    } else if (alterTbl.getOp() == alterTableDesc.alterTableTypes.ADDCOLS) {
+    } else if (alterTbl.getOp() == AlterTableDesc.alterTableTypes.ADDCOLS) {
       List<FieldSchema> newCols = alterTbl.getNewCols();
       List<FieldSchema> oldCols = tbl.getCols();
       if (tbl.getSerializationLib().equals(
@@ -982,7 +982,7 @@
         }
         tbl.getTTable().getSd().setCols(oldCols);
       }
-    } else if (alterTbl.getOp() == alterTableDesc.alterTableTypes.RENAMECOLUMN) {
+    } else if (alterTbl.getOp() == AlterTableDesc.alterTableTypes.RENAMECOLUMN) {
       List<FieldSchema> oldCols = tbl.getCols();
       List<FieldSchema> newCols = new ArrayList<FieldSchema>();
       Iterator<FieldSchema> iterOldCols = oldCols.iterator();
@@ -1048,7 +1048,7 @@
       }
 
       tbl.getTTable().getSd().setCols(newCols);
-    } else if (alterTbl.getOp() == alterTableDesc.alterTableTypes.REPLACECOLS) {
+    } else if (alterTbl.getOp() == AlterTableDesc.alterTableTypes.REPLACECOLS) {
       // change SerDe to LazySimpleSerDe if it is columnsetSerDe
       if (tbl.getSerializationLib().equals(
           "org.apache.hadoop.hive.serde.thrift.columnsetSerDe")) {
@@ -1065,12 +1065,12 @@
         return 1;
       }
       tbl.getTTable().getSd().setCols(alterTbl.getNewCols());
-    } else if (alterTbl.getOp() == alterTableDesc.alterTableTypes.ADDPROPS) {
+    } else if (alterTbl.getOp() == AlterTableDesc.alterTableTypes.ADDPROPS) {
       tbl.getTTable().getParameters().putAll(alterTbl.getProps());
-    } else if (alterTbl.getOp() == alterTableDesc.alterTableTypes.ADDSERDEPROPS) {
+    } else if (alterTbl.getOp() == AlterTableDesc.alterTableTypes.ADDSERDEPROPS) {
       tbl.getTTable().getSd().getSerdeInfo().getParameters().putAll(
           alterTbl.getProps());
-    } else if (alterTbl.getOp() == alterTableDesc.alterTableTypes.ADDSERDE) {
+    } else if (alterTbl.getOp() == AlterTableDesc.alterTableTypes.ADDSERDE) {
       tbl.setSerializationLib(alterTbl.getSerdeName());
       if ((alterTbl.getProps() != null) && (alterTbl.getProps().size() > 0)) {
         tbl.getTTable().getSd().getSerdeInfo().getParameters().putAll(
@@ -1081,13 +1081,13 @@
       tbl.reinitSerDe();
       tbl.setFields(Hive.getFieldsFromDeserializer(tbl.getName(), tbl
           .getDeserializer()));
-    } else if (alterTbl.getOp() == alterTableDesc.alterTableTypes.ADDFILEFORMAT) {
+    } else if (alterTbl.getOp() == AlterTableDesc.alterTableTypes.ADDFILEFORMAT) {
       tbl.getTTable().getSd().setInputFormat(alterTbl.getInputFormat());
       tbl.getTTable().getSd().setOutputFormat(alterTbl.getOutputFormat());
       if (alterTbl.getSerdeName() != null) {
         tbl.setSerializationLib(alterTbl.getSerdeName());
       }
-    } else if (alterTbl.getOp() == alterTableDesc.alterTableTypes.ADDCLUSTERSORTCOLUMN) {
+    } else if (alterTbl.getOp() == AlterTableDesc.alterTableTypes.ADDCLUSTERSORTCOLUMN) {
       // validate sort columns and bucket columns
       List<String> columns = Utilities.getColumnNamesFromFieldSchema(tbl
           .getCols());
@@ -1154,7 +1154,7 @@
    * @throws HiveException
    *           Throws this exception if an unexpected error occurs.
    */
-  private int dropTable(Hive db, dropTableDesc dropTbl) throws HiveException {
+  private int dropTable(Hive db, DropTableDesc dropTbl) throws HiveException {
     // We need to fetch the table before it is dropped so that it can be passed
     // to
     // post-execution hook
@@ -1259,7 +1259,7 @@
    * @throws HiveException
    *           Throws this exception if an unexpected error occurs.
    */
-  private int createTable(Hive db, createTableDesc crtTbl) throws HiveException {
+  private int createTable(Hive db, CreateTableDesc crtTbl) throws HiveException {
     // create the table
     Table tbl = new Table(crtTbl.getTableName());
     StorageDescriptor tblStorDesc = tbl.getTTable().getSd();
@@ -1398,7 +1398,7 @@
    * @throws HiveException
    *           Throws this exception if an unexpected error occurs.
    */
-  private int createTableLike(Hive db, createTableLikeDesc crtTbl)
+  private int createTableLike(Hive db, CreateTableLikeDesc crtTbl)
       throws HiveException {
     // Get the existing table
     Table tbl = db.getTable(MetaStoreUtils.DEFAULT_DATABASE_NAME, crtTbl
@@ -1437,7 +1437,7 @@
    * @throws HiveException
    *           Throws this exception if an unexpected error occurs.
    */
-  private int createView(Hive db, createViewDesc crtView) throws HiveException {
+  private int createView(Hive db, CreateViewDesc crtView) throws HiveException {
     Table tbl = new Table(crtView.getViewName());
     tbl.setViewOriginalText(crtView.getViewOriginalText());
     tbl.setViewExpandedText(crtView.getViewExpandedText());

Copied: hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/Description.java (from r902715, hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/description.java)
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/Description.java?p2=hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/Description.java&p1=hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/description.java&r1=902715&r2=902921&rev=902921&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/description.java (original)
+++ hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/Description.java Mon Jan 25 18:48:58 2010
@@ -22,7 +22,7 @@
 import java.lang.annotation.RetentionPolicy;
 
 @Retention(RetentionPolicy.RUNTIME)
-public @interface description {
+public @interface Description {
   String value() default "_FUNC_ is undocumented";
 
   String extended() default "";

Propchange: hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/Description.java
------------------------------------------------------------------------------
    svn:mergeinfo = 

Modified: hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExecDriver.java
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExecDriver.java?rev=902921&r1=902920&r2=902921&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExecDriver.java (original)
+++ hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExecDriver.java Mon Jan 25 18:48:58 2010
@@ -56,8 +56,8 @@
 import org.apache.hadoop.hive.ql.io.HiveKey;
 import org.apache.hadoop.hive.ql.io.HiveOutputFormat;
 import org.apache.hadoop.hive.ql.metadata.HiveException;
-import org.apache.hadoop.hive.ql.plan.mapredWork;
-import org.apache.hadoop.hive.ql.plan.partitionDesc;
+import org.apache.hadoop.hive.ql.plan.MapredWork;
+import org.apache.hadoop.hive.ql.plan.PartitionDesc;
 import org.apache.hadoop.hive.ql.plan.api.StageType;
 import org.apache.hadoop.hive.ql.session.SessionState;
 import org.apache.hadoop.hive.ql.session.SessionState.LogHelper;
@@ -75,7 +75,7 @@
 import org.apache.log4j.BasicConfigurator;
 import org.apache.log4j.varia.NullAppender;
 
-public class ExecDriver extends Task<mapredWork> implements Serializable {
+public class ExecDriver extends Task<MapredWork> implements Serializable {
 
   private static final long serialVersionUID = 1L;
 
@@ -154,7 +154,7 @@
   /**
    * Constructor/Initialization for invocation as independent utility
    */
-  public ExecDriver(mapredWork plan, JobConf job, boolean isSilent)
+  public ExecDriver(MapredWork plan, JobConf job, boolean isSilent)
       throws HiveException {
     setWork(plan);
     this.job = job;
@@ -375,7 +375,7 @@
    * @return the number of reducers.
    */
   public int estimateNumberOfReducers(HiveConf hive, JobConf job,
-      mapredWork work) throws IOException {
+      MapredWork work) throws IOException {
     if (hive == null) {
       hive = new HiveConf();
     }
@@ -442,7 +442,7 @@
    * @return the total size in bytes.
    * @throws IOException
    */
-  public long getTotalInputFileSize(JobConf job, mapredWork work)
+  public long getTotalInputFileSize(JobConf job, MapredWork work)
       throws IOException {
     long r = 0;
     // For each input path, calculate the total size.
@@ -930,7 +930,7 @@
       }
     }
 
-    mapredWork plan = Utilities.deserializeMapRedWork(pathData, conf);
+    MapredWork plan = Utilities.deserializeMapRedWork(pathData, conf);
     ExecDriver ed = new ExecDriver(plan, conf, isSilent);
 
     int ret = ed.execute();
@@ -1002,7 +1002,7 @@
 
   @Override
   public boolean hasReduce() {
-    mapredWork w = getWork();
+    MapredWork w = getWork();
     return w.getReducer() != null;
   }
 
@@ -1022,7 +1022,7 @@
   /**
    * Handle a empty/null path for a given alias
    */
-  private int addInputPath(String path, JobConf job, mapredWork work,
+  private int addInputPath(String path, JobConf job, MapredWork work,
       String hiveScratchDir, int numEmptyPaths, boolean isEmptyPath,
       String alias) throws Exception {
     // either the directory does not exist or it is empty
@@ -1065,14 +1065,14 @@
 
     work.setPathToAliases(pathToAliases);
 
-    LinkedHashMap<String, partitionDesc> pathToPartitionInfo = work
+    LinkedHashMap<String, PartitionDesc> pathToPartitionInfo = work
         .getPathToPartitionInfo();
     if (isEmptyPath) {
       pathToPartitionInfo.put(newPath.toUri().toString(), pathToPartitionInfo
           .get(path));
       pathToPartitionInfo.remove(path);
     } else {
-      partitionDesc pDesc = work.getAliasToPartnInfo().get(alias).clone();
+      PartitionDesc pDesc = work.getAliasToPartnInfo().get(alias).clone();
       pathToPartitionInfo.put(newPath.toUri().toString(), pDesc);
     }
     work.setPathToPartitionInfo(pathToPartitionInfo);
@@ -1085,7 +1085,7 @@
     return numEmptyPaths;
   }
 
-  private void addInputPaths(JobConf job, mapredWork work, String hiveScratchDir)
+  private void addInputPaths(JobConf job, MapredWork work, String hiveScratchDir)
       throws Exception {
     int numEmptyPaths = 0;
 

Modified: hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExecMapper.java
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExecMapper.java?rev=902921&r1=902920&r2=902921&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExecMapper.java (original)
+++ hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExecMapper.java Mon Jan 25 18:48:58 2010
@@ -29,9 +29,9 @@
 
 import org.apache.commons.logging.Log;
 import org.apache.commons.logging.LogFactory;
-import org.apache.hadoop.hive.ql.plan.fetchWork;
-import org.apache.hadoop.hive.ql.plan.mapredLocalWork;
-import org.apache.hadoop.hive.ql.plan.mapredWork;
+import org.apache.hadoop.hive.ql.plan.FetchWork;
+import org.apache.hadoop.hive.ql.plan.MapredLocalWork;
+import org.apache.hadoop.hive.ql.plan.MapredWork;
 import org.apache.hadoop.hive.serde2.objectinspector.InspectableObject;
 import org.apache.hadoop.hive.serde2.objectinspector.ObjectInspector;
 import org.apache.hadoop.io.Writable;
@@ -75,7 +75,7 @@
     try {
       jc = job;
       // create map and fetch operators
-      mapredWork mrwork = Utilities.getMapRedWork(job);
+      MapredWork mrwork = Utilities.getMapRedWork(job);
       mo = new MapOperator();
       mo.setConf(mrwork);
       // initialize map operator
@@ -84,13 +84,13 @@
       mo.initialize(jc, null);
 
       // initialize map local work
-      mapredLocalWork localWork = mrwork.getMapLocalWork();
+      MapredLocalWork localWork = mrwork.getMapLocalWork();
       if (localWork == null) {
         return;
       }
       fetchOperators = new HashMap<String, FetchOperator>();
       // create map local operators
-      for (Map.Entry<String, fetchWork> entry : localWork.getAliasToFetchWork()
+      for (Map.Entry<String, FetchWork> entry : localWork.getAliasToFetchWork()
           .entrySet()) {
         fetchOperators.put(entry.getKey(), new FetchOperator(entry.getValue(),
             job));
@@ -130,7 +130,7 @@
       // process map local operators
       if (fetchOperators != null) {
         try {
-          mapredLocalWork localWork = mo.getConf().getMapLocalWork();
+          MapredLocalWork localWork = mo.getConf().getMapLocalWork();
           int fetchOpNum = 0;
           for (Map.Entry<String, FetchOperator> entry : fetchOperators
               .entrySet()) {
@@ -227,7 +227,7 @@
     try {
       mo.close(abort);
       if (fetchOperators != null) {
-        mapredLocalWork localWork = mo.getConf().getMapLocalWork();
+        MapredLocalWork localWork = mo.getConf().getMapLocalWork();
         for (Map.Entry<String, FetchOperator> entry : fetchOperators.entrySet()) {
           Operator<? extends Serializable> forwardOp = localWork
               .getAliasToWork().get(entry.getKey());

Modified: hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExecReducer.java
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExecReducer.java?rev=902921&r1=902920&r2=902921&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExecReducer.java (original)
+++ hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExecReducer.java Mon Jan 25 18:48:58 2010
@@ -30,8 +30,8 @@
 import org.apache.commons.logging.LogFactory;
 import org.apache.hadoop.hive.ql.exec.ExecMapper.reportStats;
 import org.apache.hadoop.hive.ql.metadata.HiveException;
-import org.apache.hadoop.hive.ql.plan.mapredWork;
-import org.apache.hadoop.hive.ql.plan.tableDesc;
+import org.apache.hadoop.hive.ql.plan.MapredWork;
+import org.apache.hadoop.hive.ql.plan.TableDesc;
 import org.apache.hadoop.hive.serde2.Deserializer;
 import org.apache.hadoop.hive.serde2.SerDe;
 import org.apache.hadoop.hive.serde2.SerDeException;
@@ -77,8 +77,8 @@
     fieldNames = fieldNameArray.toArray(new String[0]);
   }
 
-  tableDesc keyTableDesc;
-  tableDesc[] valueTableDesc;
+  TableDesc keyTableDesc;
+  TableDesc[] valueTableDesc;
 
   @Override
   public void configure(JobConf job) {
@@ -100,7 +100,7 @@
       l4j.info("cannot get classpath: " + e.getMessage());
     }
     jc = job;
-    mapredWork gWork = Utilities.getMapRedWork(job);
+    MapredWork gWork = Utilities.getMapRedWork(job);
     reducer = gWork.getReducer();
     reducer.setParentOperators(null); // clear out any parents as reducer is the
                                       // root
@@ -111,7 +111,7 @@
           .getDeserializerClass(), null);
       inputKeyDeserializer.initialize(null, keyTableDesc.getProperties());
       keyObjectInspector = inputKeyDeserializer.getObjectInspector();
-      valueTableDesc = new tableDesc[gWork.getTagToValueDesc().size()];
+      valueTableDesc = new TableDesc[gWork.getTagToValueDesc().size()];
       for (int tag = 0; tag < gWork.getTagToValueDesc().size(); tag++) {
         // We should initialize the SerDe with the TypeInfo when available.
         valueTableDesc[tag] = gWork.getTagToValueDesc().get(tag);

Modified: hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExplainTask.java
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExplainTask.java?rev=902921&r1=902920&r2=902921&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExplainTask.java (original)
+++ hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExplainTask.java Mon Jan 25 18:48:58 2010
@@ -31,8 +31,8 @@
 import java.util.Set;
 import java.util.Map.Entry;
 
-import org.apache.hadoop.hive.ql.plan.explain;
-import org.apache.hadoop.hive.ql.plan.explainWork;
+import org.apache.hadoop.hive.ql.plan.Explain;
+import org.apache.hadoop.hive.ql.plan.ExplainWork;
 import org.apache.hadoop.hive.ql.plan.api.StageType;
 import org.apache.hadoop.util.StringUtils;
 
@@ -40,7 +40,7 @@
  * ExplainTask implementation
  * 
  **/
-public class ExplainTask extends Task<explainWork> implements Serializable {
+public class ExplainTask extends Task<ExplainWork> implements Serializable {
   private static final long serialVersionUID = 1L;
 
   public ExplainTask() {
@@ -161,10 +161,10 @@
   private void outputPlan(Serializable work, PrintStream out, boolean extended,
       int indent) throws Exception {
     // Check if work has an explain annotation
-    Annotation note = work.getClass().getAnnotation(explain.class);
+    Annotation note = work.getClass().getAnnotation(Explain.class);
 
-    if (note instanceof explain) {
-      explain xpl_note = (explain) note;
+    if (note instanceof Explain) {
+      Explain xpl_note = (Explain) note;
       if (extended || xpl_note.normalExplain()) {
         out.print(indentString(indent));
         out.println(xpl_note.displayName());
@@ -193,10 +193,10 @@
 
     for (Method m : methods) {
       int prop_indents = indent + 2;
-      note = m.getAnnotation(explain.class);
+      note = m.getAnnotation(Explain.class);
 
-      if (note instanceof explain) {
-        explain xpl_note = (explain) note;
+      if (note instanceof Explain) {
+        Explain xpl_note = (Explain) note;
 
         if (extended || xpl_note.normalExplain()) {
 

Modified: hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExprNodeColumnEvaluator.java
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExprNodeColumnEvaluator.java?rev=902921&r1=902920&r2=902921&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExprNodeColumnEvaluator.java (original)
+++ hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExprNodeColumnEvaluator.java Mon Jan 25 18:48:58 2010
@@ -19,7 +19,7 @@
 package org.apache.hadoop.hive.ql.exec;
 
 import org.apache.hadoop.hive.ql.metadata.HiveException;
-import org.apache.hadoop.hive.ql.plan.exprNodeColumnDesc;
+import org.apache.hadoop.hive.ql.plan.ExprNodeColumnDesc;
 import org.apache.hadoop.hive.serde2.objectinspector.ObjectInspector;
 import org.apache.hadoop.hive.serde2.objectinspector.StructField;
 import org.apache.hadoop.hive.serde2.objectinspector.StructObjectInspector;
@@ -29,12 +29,12 @@
  */
 public class ExprNodeColumnEvaluator extends ExprNodeEvaluator {
 
-  protected exprNodeColumnDesc expr;
+  protected ExprNodeColumnDesc expr;
 
   transient StructObjectInspector[] inspectors;
   transient StructField[] fields;
 
-  public ExprNodeColumnEvaluator(exprNodeColumnDesc expr) {
+  public ExprNodeColumnEvaluator(ExprNodeColumnDesc expr) {
     this.expr = expr;
   }
 

Modified: hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExprNodeConstantEvaluator.java
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExprNodeConstantEvaluator.java?rev=902921&r1=902920&r2=902921&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExprNodeConstantEvaluator.java (original)
+++ hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExprNodeConstantEvaluator.java Mon Jan 25 18:48:58 2010
@@ -19,7 +19,7 @@
 package org.apache.hadoop.hive.ql.exec;
 
 import org.apache.hadoop.hive.ql.metadata.HiveException;
-import org.apache.hadoop.hive.ql.plan.exprNodeConstantDesc;
+import org.apache.hadoop.hive.ql.plan.ExprNodeConstantDesc;
 import org.apache.hadoop.hive.serde2.objectinspector.ObjectInspector;
 import org.apache.hadoop.hive.serde2.objectinspector.PrimitiveObjectInspector.PrimitiveCategory;
 import org.apache.hadoop.hive.serde2.objectinspector.primitive.PrimitiveObjectInspectorFactory;
@@ -27,11 +27,11 @@
 
 public class ExprNodeConstantEvaluator extends ExprNodeEvaluator {
 
-  protected exprNodeConstantDesc expr;
+  protected ExprNodeConstantDesc expr;
   transient ObjectInspector writableObjectInspector;
   transient Object writableValue;
 
-  public ExprNodeConstantEvaluator(exprNodeConstantDesc expr) {
+  public ExprNodeConstantEvaluator(ExprNodeConstantDesc expr) {
     this.expr = expr;
     PrimitiveCategory pc = ((PrimitiveTypeInfo) expr.getTypeInfo())
         .getPrimitiveCategory();

Modified: hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExprNodeEvaluatorFactory.java
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExprNodeEvaluatorFactory.java?rev=902921&r1=902920&r2=902921&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExprNodeEvaluatorFactory.java (original)
+++ hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExprNodeEvaluatorFactory.java Mon Jan 25 18:48:58 2010
@@ -18,38 +18,38 @@
 
 package org.apache.hadoop.hive.ql.exec;
 
-import org.apache.hadoop.hive.ql.plan.exprNodeColumnDesc;
-import org.apache.hadoop.hive.ql.plan.exprNodeConstantDesc;
-import org.apache.hadoop.hive.ql.plan.exprNodeDesc;
-import org.apache.hadoop.hive.ql.plan.exprNodeFieldDesc;
-import org.apache.hadoop.hive.ql.plan.exprNodeGenericFuncDesc;
-import org.apache.hadoop.hive.ql.plan.exprNodeNullDesc;
+import org.apache.hadoop.hive.ql.plan.ExprNodeColumnDesc;
+import org.apache.hadoop.hive.ql.plan.ExprNodeConstantDesc;
+import org.apache.hadoop.hive.ql.plan.ExprNodeDesc;
+import org.apache.hadoop.hive.ql.plan.ExprNodeFieldDesc;
+import org.apache.hadoop.hive.ql.plan.ExprNodeGenericFuncDesc;
+import org.apache.hadoop.hive.ql.plan.ExprNodeNullDesc;
 
 public class ExprNodeEvaluatorFactory {
 
   public ExprNodeEvaluatorFactory() {
   }
 
-  public static ExprNodeEvaluator get(exprNodeDesc desc) {
+  public static ExprNodeEvaluator get(ExprNodeDesc desc) {
     // Constant node
-    if (desc instanceof exprNodeConstantDesc) {
-      return new ExprNodeConstantEvaluator((exprNodeConstantDesc) desc);
+    if (desc instanceof ExprNodeConstantDesc) {
+      return new ExprNodeConstantEvaluator((ExprNodeConstantDesc) desc);
     }
     // Column-reference node, e.g. a column in the input row
-    if (desc instanceof exprNodeColumnDesc) {
-      return new ExprNodeColumnEvaluator((exprNodeColumnDesc) desc);
+    if (desc instanceof ExprNodeColumnDesc) {
+      return new ExprNodeColumnEvaluator((ExprNodeColumnDesc) desc);
     }
     // Generic Function node, e.g. CASE, an operator or a UDF node
-    if (desc instanceof exprNodeGenericFuncDesc) {
-      return new ExprNodeGenericFuncEvaluator((exprNodeGenericFuncDesc) desc);
+    if (desc instanceof ExprNodeGenericFuncDesc) {
+      return new ExprNodeGenericFuncEvaluator((ExprNodeGenericFuncDesc) desc);
     }
     // Field node, e.g. get a.myfield1 from a
-    if (desc instanceof exprNodeFieldDesc) {
-      return new ExprNodeFieldEvaluator((exprNodeFieldDesc) desc);
+    if (desc instanceof ExprNodeFieldDesc) {
+      return new ExprNodeFieldEvaluator((ExprNodeFieldDesc) desc);
     }
     // Null node, a constant node with value NULL and no type information
-    if (desc instanceof exprNodeNullDesc) {
-      return new ExprNodeNullEvaluator((exprNodeNullDesc) desc);
+    if (desc instanceof ExprNodeNullDesc) {
+      return new ExprNodeNullEvaluator((ExprNodeNullDesc) desc);
     }
 
     throw new RuntimeException(

Modified: hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExprNodeFieldEvaluator.java
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExprNodeFieldEvaluator.java?rev=902921&r1=902920&r2=902921&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExprNodeFieldEvaluator.java (original)
+++ hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExprNodeFieldEvaluator.java Mon Jan 25 18:48:58 2010
@@ -22,7 +22,7 @@
 import java.util.List;
 
 import org.apache.hadoop.hive.ql.metadata.HiveException;
-import org.apache.hadoop.hive.ql.plan.exprNodeFieldDesc;
+import org.apache.hadoop.hive.ql.plan.ExprNodeFieldDesc;
 import org.apache.hadoop.hive.serde2.objectinspector.ListObjectInspector;
 import org.apache.hadoop.hive.serde2.objectinspector.ObjectInspector;
 import org.apache.hadoop.hive.serde2.objectinspector.ObjectInspectorFactory;
@@ -36,7 +36,7 @@
  */
 public class ExprNodeFieldEvaluator extends ExprNodeEvaluator {
 
-  protected exprNodeFieldDesc desc;
+  protected ExprNodeFieldDesc desc;
   transient ExprNodeEvaluator leftEvaluator;
   transient ObjectInspector leftInspector;
   transient StructObjectInspector structObjectInspector;
@@ -44,7 +44,7 @@
   transient ObjectInspector structFieldObjectInspector;
   transient ObjectInspector resultObjectInspector;
 
-  public ExprNodeFieldEvaluator(exprNodeFieldDesc desc) {
+  public ExprNodeFieldEvaluator(ExprNodeFieldDesc desc) {
     this.desc = desc;
     leftEvaluator = ExprNodeEvaluatorFactory.get(desc.getDesc());
   }

Modified: hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExprNodeGenericFuncEvaluator.java
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExprNodeGenericFuncEvaluator.java?rev=902921&r1=902920&r2=902921&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExprNodeGenericFuncEvaluator.java (original)
+++ hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExprNodeGenericFuncEvaluator.java Mon Jan 25 18:48:58 2010
@@ -21,7 +21,7 @@
 import org.apache.commons.logging.Log;
 import org.apache.commons.logging.LogFactory;
 import org.apache.hadoop.hive.ql.metadata.HiveException;
-import org.apache.hadoop.hive.ql.plan.exprNodeGenericFuncDesc;
+import org.apache.hadoop.hive.ql.plan.ExprNodeGenericFuncDesc;
 import org.apache.hadoop.hive.ql.udf.generic.GenericUDF;
 import org.apache.hadoop.hive.serde2.objectinspector.ObjectInspector;
 
@@ -30,7 +30,7 @@
   private static final Log LOG = LogFactory
       .getLog(ExprNodeGenericFuncEvaluator.class.getName());
 
-  protected exprNodeGenericFuncDesc expr;
+  protected ExprNodeGenericFuncDesc expr;
 
   transient GenericUDF genericUDF;
   transient Object rowObject;
@@ -53,7 +53,7 @@
     }
   };
 
-  public ExprNodeGenericFuncEvaluator(exprNodeGenericFuncDesc expr) {
+  public ExprNodeGenericFuncEvaluator(ExprNodeGenericFuncDesc expr) {
     this.expr = expr;
     children = new ExprNodeEvaluator[expr.getChildExprs().size()];
     for (int i = 0; i < children.length; i++) {

Modified: hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExprNodeNullEvaluator.java
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExprNodeNullEvaluator.java?rev=902921&r1=902920&r2=902921&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExprNodeNullEvaluator.java (original)
+++ hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExprNodeNullEvaluator.java Mon Jan 25 18:48:58 2010
@@ -19,7 +19,7 @@
 package org.apache.hadoop.hive.ql.exec;
 
 import org.apache.hadoop.hive.ql.metadata.HiveException;
-import org.apache.hadoop.hive.ql.plan.exprNodeNullDesc;
+import org.apache.hadoop.hive.ql.plan.ExprNodeNullDesc;
 import org.apache.hadoop.hive.serde2.objectinspector.ObjectInspector;
 import org.apache.hadoop.hive.serde2.objectinspector.primitive.PrimitiveObjectInspectorFactory;
 
@@ -27,9 +27,9 @@
 // change the void to the first matching argument
 public class ExprNodeNullEvaluator extends ExprNodeEvaluator {
 
-  protected exprNodeNullDesc expr;
+  protected ExprNodeNullDesc expr;
 
-  public ExprNodeNullEvaluator(exprNodeNullDesc expr) {
+  public ExprNodeNullEvaluator(ExprNodeNullDesc expr) {
     this.expr = expr;
   }
 



Mime
View raw message