hive-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Nicholas Murphy <>
Subject Re: basic, dumb getting started question (single-node)
Date Wed, 12 Nov 2014 22:40:33 GMT
Default installation only comes with a (lots of stuff, all of it commented

# Licensed to the Apache Software Foundation (ASF) under one
# or more contributor license agreements.  See the NOTICE file
# distributed with this work for additional information
# regarding copyright ownership.  The ASF licenses this file
# to you under the Apache License, Version 2.0 (the
# "License"); you may not use this file except in compliance
# with the License.  You may obtain a copy of the License at
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# See the License for the specific language governing permissions and
# limitations under the License.

# Set Hive and Hadoop environment variables here. These variables can be used
# to control the execution of Hive. It should be used by admins to configure
# the Hive installation (so that users do not have to set environment variables
# or set command line parameters to get correct behavior).
# The hive service being invoked (CLI/HWI etc.) is available via the environment
# variable SERVICE

# Hive Client memory usage can be an issue if a large number of clients
# are running at the same time. The flags below have been useful in 
# reducing memory usage:
# if [ "$SERVICE" = "cli" ]; then
#   if [ -z "$DEBUG" ]; then
#     export HADOOP_OPTS="$HADOOP_OPTS -XX:NewRatio=12 -Xms10m -XX:MaxHeapFreeRa
tio=40 -XX:MinHeapFreeRatio=15 -XX:+UseParNewGC -XX:-UseGCOverheadLimit"
#   else
#     export HADOOP_OPTS="$HADOOP_OPTS -XX:NewRatio=12 -Xms10m -XX:MaxHeapFreeRa
tio=40 -XX:MinHeapFreeRatio=15 -XX:-UseGCOverheadLimit"
#   fi
# fi

# The heap size of the jvm stared by hive shell script can be controlled via:
# export HADOOP_HEAPSIZE=1024
# Larger heap size may be required when running queries over large number of fil
es or partitions. 
# By default hive shell scripts use a heap size of 256 (MB).  Larger heap size w
ould also be 
# appropriate for hive server (hwi etc).

# Set HADOOP_HOME to point to a specific hadoop install directory
# HADOOP_HOME=${bin}/../../hadoop

# Hive Configuration Directory can be controlled by:
# export HIVE_CONF_DIR=

# Folder containing extra ibraries required for hive compilation/execution can b
e controlled by:


> On Nov 12, 2014, at 2:25 PM, Ritesh Kumar Singh <>
> If you have correctly set your HIVE_HOME variable in your .bash_rc file,
> What's the output of:
> >cat $HIVE_HOME/conf/
> If it doesn't work, run
> >cat $HIVE_HOME/conf/
> On Thu, Nov 13, 2014 at 2:57 AM, Nicholas Murphy <> wrote:
> Sorry for the (as per subject) basic, dumb getting started question but: I'm trying to
just set up Hive on a single node.  Hadoop is set up and running happily, but following the
directions on the Getting Started guide don't actually work (following the directions literally
results in it complaining about a lack of a hive-site.conf file; putting a dummy one in results
in, among other things, the HCatalog server not starting...looking at logs, it looks like
it's encountering configuration parsing problems, which suggests I just don't have the configuration
files set up correctly).  HADOOP_HOME is set, the correct directories are set up in HDFS,
> Hadoop version 2.5.1, Hive version 0.13.1, Oracle JDK (1.6, I believe), Debian 7.7.
> I notice the default conf/ directory has a bunch of template files, but only that.  Can
someone point me to a resource, or to an example of what configuration I need, to actually
run Hive correctly out of the box on a single machine?
> Thanks,
> Nick

View raw message