accumulo-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Ott, Charles H." <>
Subject RE: accumulo startup issue: Accumulo not initialized, there is no instance id at /accumulo/instance_id
Date Thu, 16 Jan 2014 20:02:24 GMT
Forgot this part of



Also in the example below, I make the folder to log in /var/log/accumulo… however, in the example I have it pointing to $accumulo_home/logs.  Just change that to wherever
your logs are stored.


Also you’ll need to run ssh-keygen on the master before pushing the ssh keys to your nodes.
(ssh-copy-id).  Also hdfs user will proabably have read-only access to the Hadoop home folder…
just make a folder called HDFS_USER_HOMEFOLDER/.ssh and give r/w/x permissions to hdfs on
that folder (.ssh) only.  Otherwise ssh-copy-id wont be able to update that folder.


In regard to optimization and write ahead logs, that stuff is usually environment/application


If anyone has questions or comments on this installation plan let me know, I would love to
know what you’re doing differently, and why.



From: []
On Behalf Of Ott, Charles H.
Sent: Thursday, January 16, 2014 2:50 PM
Subject: RE: accumulo startup issue: Accumulo not initialized, there is no instance id at




Not advocating this is the best approach, just what I’m currently doing, put this together
pretty quick, but it should be mostly complete for settting up accumulo on cdh hdfs/zk




I always do something like this first on CentOS:


$ yum install –y ntpd openssh-clients unzip


#setup ssh and ntpd as needed


$install the jdk RPM


# bash this to setup OS specifics


echo "Disabling SELINUX for Optimal CDH Compatability..."

sed -i 's/SELINUX=enforcing/SELINUX=disabled/g' /etc/selinux/config


echo "Increasing uLimit, aka File Descripter/Handlers for all Users..."

echo "# Adding Support for CDH" >> /etc/security/limits.conf

echo "*              -     nofile              65536" >> /etc/security/limits.conf


echo "Disabling IPv6..."

echo "# Disable ipv6" >> /etc/sysctl.conf

echo "net.ipv6.conf.all.disable_ipv6 = 1" >> /etc/sysctl.conf

echo "net.ipv6.conf.default.disable_ipv6 = 1" >> /etc/sysctl.conf


echo "Increasing Swapiness Factor to limit use of swap space."

echo "# swappiness for accumulo" >> /etc/sysctl.conf

echo "vm.swappiness = 10" >> /etc/sysctl.conf



reboot and test OS/services/jdk version…


then I usually extract Accumulo to /opt/Accumulo/accumulo-1.5.0


Make a sym link, /opt/accumulo/Accumulo-current -> ./accumulo-1.5.0


#make dirs. For Accumulo logs, where ever…

mkdir /var/log/accumulo


#let HDFS own all your Accumulo folders

chown –R hdfs:hdfs /opt/accumulo

chown –R hdfs:hdfs /var/log/accumulo


#update the hdfs password for the next step

user root: passwd hdfs


#setup passwordless ssh (test using hdfs afterwards, should be able to ssh <node> w/o
entering credentials)

su –hdfs

ssh-copy-id <for all tablet server nodes>


#update your iptables



#env vars

JAVA_HOME=/usr/java/default (jdk7 in my last install worked fine)


Settings for in /conf:

# cdh4

export HADOOP_HDFS_HOME=/opt/cloudera/parcels/CDH/lib/hadoop-hdfs

export HADOOP_MAPREDUCE_HOME=/opt/cloudera/parcels/CDH/lib/hadoop-mapreduce

test -z "$HADOOP_CONF_DIR"       && export HADOOP_CONF_DIR="$HADOOP_PREFIX/etc/hadoop"

test -z "$JAVA_HOME"             && export JAVA_HOME=/usr/java/default

test -z "$ZOOKEEPER_HOME"        && export ZOOKEEPER_HOME=/opt/cloudera/parcels/CDH/lib/zookeeper



#update all files as appropriate in /opt/Accumulo/Accumulo-current/conf/*

masters, monitor,slaves,tracers,gc,Accumulo-site.xml,


























    <description>Classpaths that accumulo checks for updates and class files.

      When using the Security Manager, please remove the ".../target/classes/" values.




then of course, always run your Accumulo binaries/scripts using the HDFS account.  I’m sure
I’m missing a few steps here and there… 


$ACCUMULO_HOME/bin/accumulo init






From: []
On Behalf Of Sean Busbey
Sent: Thursday, January 16, 2014 2:20 PM
To: Accumulo User List
Subject: Re: accumulo startup issue: Accumulo not initialized, there is no instance id at



On Thu, Jan 16, 2014 at 1:14 PM, Kesten Broughton <> wrote:

"You should make sure to correct the maximum number of open files for the user that is running

I have the following in all /etc/security/limits.conf in my accumulo cluster

hdfs soft nofile 65536
hdfs hard nofile 65536

However, i see this for all nodes.
WARN : Max files open on is 32768, recommend 65536

Should it be a different user or something?

'the user that is running Accumulo'
sudo hdfs
hdfs$ bin/accumulo -u root

so is hdfs or root the accumulo user?


The user in question here is the one who starts the Accumulo server processes. In production
environments this should be a user dedicated to Accumulo. FWIW, I usually name this user "accumulo".


How do you start up Accumulo? a service script? running $ACCUMULO_HOME/bin/ something

View raw message