Dennis Kubes wrote: > I had a somewhat difficult time figuring out how to get hbase started. > In the end, it was pretty simple. Here are the steps: > > 1. Download hadoop from svn, untar to directory say ~/hadooptrunk and > compile through ant. > 2. Move the build hadoop-xx directory to where you want to run it, > say ~/hadoop > 3. Set the hadoop tmp directory in hadoop-site.xml (as default all > other variables should be file) > 4. Copy scripts from ~/hadooptrunk/src/contrib/hbase/bin to > ~/hadoop/src/contrib/hbase/bin > 5. Format hadoop dfs through ~/hadoop/bin/hadoop namenode -format > 6. Start the dfs through ~/hadoop/bin/start-dfs.sh (logs are > viewable in ~/hadoop/logs by default, don't need mapreduce for hbase) > 7. Go to the hbase directory ~/hadoop/src/contrib/hbase > 8. Hbase default values are fine for now, start hbase with > ~/hadoop/src/contrib/hbase/bin/start-hbase.sh (logs are viewable > in ~/hadoop/logs by default) > 9. Enter hbase shell with ~/hadoop/src/contrib/hbase/bin/hbase shell > 10. Have fun with Hbase > 11. Stop the hbase servers with > ~/hadoop/src/contrib/hbase/bin/stop-hbase.sh. Wait until the > servers are finished stopping. > 12. Stop the hadoop dfs with ~/hadoop/bin/stop-dfs.sh > > Hope this helps. Did you try to run it with LocalFS / Cygwin, and if so, did you notice any peculiarities? I tried this once, and first the start-hbase.sh script wouldn't work (missing log files? it looked like some variables in paths were expanded in a wrong way), and then when I started the master and a regionserver by hand, it would complain about missing map files and all requests would time out ... I gave up after that and moved to HDFS. -- Best regards, Andrzej Bialecki <>< ___. ___ ___ ___ _ _ __________________________________ [__ || __|__/|__||\/| Information Retrieval, Semantic Web ___|||__|| \| || | Embedded Unix, System Integration http://www.sigram.com Contact: info at sigram dot com