hbase-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From st...@apache.org
Subject svn commit: r1034230 - in /hbase/trunk/src: docbkx/ main/java/org/apache/hadoop/hbase/ main/java/org/apache/hadoop/hbase/mapreduce/ main/java/org/apache/hadoop/hbase/regionserver/ main/java/org/apache/hadoop/hbase/replication/regionserver/ main/java/or...
Date Fri, 12 Nov 2010 01:23:19 GMT
Author: stack
Date: Fri Nov 12 01:23:19 2010
New Revision: 1034230

URL: http://svn.apache.org/viewvc?rev=1034230&view=rev
Log:
Fixes to site and moved getting started out of javadoc and into book

Removed:
    hbase/trunk/src/site/fml/
Modified:
    hbase/trunk/src/docbkx/book.xml
    hbase/trunk/src/main/java/org/apache/hadoop/hbase/KeyValue.java
    hbase/trunk/src/main/java/org/apache/hadoop/hbase/mapreduce/package-info.java
    hbase/trunk/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java
    hbase/trunk/src/main/java/org/apache/hadoop/hbase/replication/regionserver/Replication.java
    hbase/trunk/src/main/java/org/apache/hadoop/hbase/security/User.java
    hbase/trunk/src/main/javadoc/overview.html
    hbase/trunk/src/site/site.xml
    hbase/trunk/src/site/xdoc/index.xml
    hbase/trunk/src/site/xdoc/old_news.xml

Modified: hbase/trunk/src/docbkx/book.xml
URL: http://svn.apache.org/viewvc/hbase/trunk/src/docbkx/book.xml?rev=1034230&r1=1034229&r2=1034230&view=diff
==============================================================================
--- hbase/trunk/src/docbkx/book.xml (original)
+++ hbase/trunk/src/docbkx/book.xml Fri Nov 12 01:23:19 2010
@@ -83,16 +83,25 @@
 
   <chapter xml:id="getting_started">
     <title>Getting Started</title>
+    <section >
+      <title>Introduction</title>
+      <para>
+          <link linkend="quickstart">Quick Start</link> will get you up and running
+          on a single-node instance of HBase using the local filesystem.
+          The <link linkend="notsoquick">Not-so-quick Start Guide</link> 
+          describes setup of HBase in distributed mode running on top of HDFS.
+      </para>
+    </section>
 
     <section xml:id="quickstart">
       <title>Quick Start</title>
 
       <para><itemizedlist>
           <para>Here is a quick guide to starting up a standalone HBase
-          instance (an HBase instance that uses the local filesystem rather than
-          Hadoop HDFS), creating a table and inserting rows into a table via the
+              instance that uses the local filesystem.  It leads you
+              through creating a table, inserting rows via the
           <link linkend="shell">HBase Shell</link>, and then cleaning up and shutting
-          down your running instance. The below exercise should take no more than
+          down your instance. The below exercise should take no more than
           ten minutes (not including download time).
       </para>
           
@@ -101,7 +110,7 @@
 
             <para>Choose a download site from this list of <link
             xlink:href="http://www.apache.org/dyn/closer.cgi/hbase/">Apache
-            Download Mirrors</link>. Click on it. This will take you to a
+            Download Mirrors</link>. Click on suggested top link. This will take you to a
             mirror of <emphasis>HBase Releases</emphasis>. Click on
             the folder named <filename>stable</filename> and then download the
             file that ends in <filename>.tar.gz</filename> to your local filesystem;
@@ -146,7 +155,7 @@ starting master, logging to logs/hbase-u
 
             <note>
             <title>Is <application>java</application> installed?</title>
-            <para>The above presumes a 1.6 version of Oracle
+            <para>The above presumes a 1.6 version of SUN
             <application>java</application> is installed on your
             machine and available on your path; i.e. when you type
             <application>java</application>, you see output that describes the options
@@ -257,6 +266,7 @@ stopping hbase...............</programli
 
     <section xml:id="notsoquick">
       <title>Not-so-quick Start Guide</title>
+      
       <section xml:id="requirements"><title>Requirements</title>
       <para>HBase has the following requirements.  Please read the
       section below carefully and ensure that all requirements have been
@@ -271,7 +281,8 @@ Usually you'll want to use the latest ve
 </section>
   <section xml:id="hadoop"><title><link xlink:href="http://hadoop.apache.org">hadoop</link></title>
 <para>This version of HBase will only run on <link xlink:href="http://hadoop.apache.org/common/releases.html">Hadoop 0.20.x</link>.
- HBase will lose data unless it is running on an HDFS that has a durable sync.
+    It will not run on hadoop 0.21.x as of this writing.
+    HBase will lose data unless it is running on an HDFS that has a durable <code>sync</code>.
  Currently only the <link xlink:href="http://svn.apache.org/viewvc/hadoop/common/branches/branch-0.20-append/">branch-0.20-append</link>
  branch has this attribute.  No official releases have been made from this branch as of this writing
  so you will have to build your own Hadoop from the tip of this branch
@@ -297,6 +308,7 @@ Usually you'll want to use the latest ve
   </para>
 </section>
 
+
       <section xml:id="ulimit">
       <title><varname>ulimit</varname></title>
       <para>HBase is a database, it uses a lot of files at the same time.
@@ -330,27 +342,328 @@ Usually you'll want to use the latest ve
       </programlisting>
       </para>
       </section>
-      </section>
 
-      <section><title>HBase run modes: Standalone, Pseudo-distributed, and Distributed</title>
-      <para>HBase has three different run modes: standalone, this is what is described above in
-      <link linkend="quickstart">Quick Start,</link> pseudo-distributed mode where all
-      daemons run on a single server, and distributed, where each of the daemons runs
-      on different cluster node.</para>
-      <section><title>Standalone HBase</title>
-      <para>TODO</para>
+<section xml:id="windows">
+<title>Windows</title>
+<para>
+If you are running HBase on Windows, you must install
+<link xlink:href="http://cygwin.com/">Cygwin</link>
+to have a *nix-like environment for the shell scripts. The full details
+are explained in the <link xlink:href="../cygwin.html">Windows Installation</link>
+guide.
+</para>
+</section>
+
       </section>
-      <section><title>Pseudo-distributed</title>
-      <para>TODO</para>
+
+      <section><title>HBase run modes: Standalone and Distributed</title>
+          <para>HBase has two run modes: <link linkend="standalone">standalone</link>
+              and <link linkend="distributed">distributed</link>.</para>
+
+<para>Whatever your mode, define <code>${HBASE_HOME}</code> to be the location of the root of your HBase installation, e.g.
+<code>/user/local/hbase</code>. Edit <code>${HBASE_HOME}/conf/hbase-env.sh</code>. In this file you can
+set the heapsize for HBase, etc. At a minimum, set <code>JAVA_HOME</code> to point at the root of
+your Java installation.</para>
+
+      <section xml:id="standalone"><title>Standalone HBase</title>
+          <para>This mode is what <link linkend="quickstart">Quick Start</link> covered;
+          all daemons are run in the one JVM and HBase writes the local filesystem.</para>
       </section>
       <section><title>Distributed</title>
-      <para>TODO</para>
-      </section>
-      </section>
+          <para>Distributed mode can be subdivided into distributed but all daemons run on a
+          single node AND distibuted with daemons spread across all nodes in the cluster.</para>
+      <para>
+          Distributed modes require an instance of the <emphasis>Hadoop Distributed File System</emphasis> (HDFS).
+See the Hadoop <link xlink:href="http://hadoop.apache.org/common/docs/current/api/overview-summary.html#overview_description">
+requirements and instructions</link> for how to set up a HDFS.
+      </para>
+
+
+
+      <section xml:id="pseudo"><title>Pseudo-distributed</title>
+<para>A pseudo-distributed mode is simply a distributed mode run on a single host.
+Use this configuration testing and prototyping on hbase.  Do not use this configuration
+for production nor for evaluating HBase performance.
+</para>
+<para>Once you have confirmed your HDFS setup, configuring HBase for use on one host requires modification of
+<filename>./conf/hbase-site.xml</filename>, which needs to be pointed at the running Hadoop HDFS instance.
+Use <filename>hbase-site.xml</filename> to override the properties defined in
+<filename>conf/hbase-default.xml</filename> (<filename>hbase-default.xml</filename> itself
+should never be modified) and for HDFS client configurations.
+At a minimum, the <varname>hbase.rootdir</varname>,
+which points HBase at the Hadoop filesystem to use,
+should be redefined in <filename>hbase-site.xml</filename>. For example,
+adding the properties below to your <filename>hbase-site.xml</filename> says that HBase
+should use the <filename>/hbase</filename> 
+directory in the HDFS whose namenode is at port 9000 on your local machine, and that
+it should run with one replica only (recommended for pseudo-distributed mode):</para>
+<programlisting>
+&lt;configuration&gt;
+  ...
+  &lt;property&gt;
+    &lt;name&gt;hbase.rootdir&lt;/name&gt;
+    &lt;value&gt;hdfs://localhost:9000/hbase&lt;/value&gt;
+    &lt;description&gt;The directory shared by region servers.
+    &lt;/description&gt;
+  &lt;/property&gt;
+  &lt;property&gt;
+    &lt;name&gt;dfs.replication&lt;/name&gt;
+    &lt;value&gt;1&lt;/value&gt;
+    &lt;description&gt;The replication count for HLog &amp; HFile storage. Should not be greater than HDFS datanode count.
+    &lt;/description&gt;
+  &lt;/property&gt;
+  ...
+&lt;/configuration&gt;
+</programlisting>
+
+<note>
+<para>Let HBase create the directory. If you don't, you'll get warning saying HBase
+needs a migration run because the directory is missing files expected by HBase (it'll
+create them if you let it).</para>
+</note>
+
+<note>
+<para>Above we bind to localhost.  This means that a remote client cannot
+connect.  Amend accordingly, if you want to connect from a remote location.</para>
+</note>
+      </section>
+
+      <section><title>Distributed across multiple machines</title>
+
+
+<para>For running a fully-distributed operation on more than one host, the following
+configurations must be made <emphasis>in addition</emphasis> to those described in the
+<link linkend="pseudo">pseudo-distributed operation</link> section above.</para>
+
+<para>In <filename>hbase-site.xml</filename>, set <varname>hbase.cluster.distributed</varname> to <varname>true</varname>.</para>
+<programlisting>
+&lt;configuration&gt;
+  ...
+  &lt;property&gt;
+    &lt;name&gt;hbase.cluster.distributed&lt;/name&gt;
+    &lt;value&gt;true&lt;/value&gt;
+    &lt;description&gt;The mode the cluster will be in. Possible values are
+      false: standalone and pseudo-distributed setups with managed Zookeeper
+      true: fully-distributed with unmanaged Zookeeper Quorum (see hbase-env.sh)
+    &lt;/description&gt;
+  &lt;/property&gt;
+  ...
+&lt;/configuration&gt;
+</programlisting>
+
+<para>In fully-distributed mode, you probably want to change your <varname>hbase.rootdir</varname>
+from localhost to the name of the node running the HDFS NameNode and you should set
+the dfs.replication to be the number of datanodes you have in your cluster or 3, which
+ever is the smaller.
+</para>
+<para>In addition
+to <filename>hbase-site.xml</filename> changes, a fully-distributed mode requires that you
+modify <filename>${HBASE_HOME}/conf/regionservers</filename>.
+The <filename>regionserver</filename> file lists all hosts running <application>HRegionServer</application>s, one host per line
+(This file in HBase is like the Hadoop slaves file at <filename>${HADOOP_HOME}/conf/slaves</filename>).</para>
+
+<para>A distributed HBase depends on a running ZooKeeper cluster. All participating nodes and clients
+need to be able to get to the running ZooKeeper cluster.
+HBase by default manages a ZooKeeper cluster for you, or you can manage it on your own and point HBase to it.
+To toggle HBase management of ZooKeeper, use the <varname>HBASE_MANAGES_ZK</varname> variable in <filename>${HBASE_HOME}/conf/hbase-env.sh</filename>.
+This variable, which defaults to <varname>true</varname>, tells HBase whether to
+start/stop the ZooKeeper quorum servers alongside the rest of the servers.</para>
+
+<para>When HBase manages the ZooKeeper cluster, you can specify ZooKeeper configuration
+using its canonical <filename>zoo.cfg</filename> file (see below), or 
+just specify ZookKeeper options directly in the <filename>${HBASE_HOME}/conf/hbase-site.xml</filename>
+(If new to ZooKeeper, go the path of specifying your configuration in HBase's hbase-site.xml).
+Every ZooKeeper configuration option has a corresponding property in the HBase hbase-site.xml
+XML configuration file named <varname>hbase.zookeeper.property.OPTION</varname>.
+For example, the <varname>clientPort</varname> setting in ZooKeeper can be changed by
+setting the <varname>hbase.zookeeper.property.clientPort</varname> property.
+For the full list of available properties, see ZooKeeper's <filename>zoo.cfg</filename>.
+For the default values used by HBase, see <filename>${HBASE_HOME}/conf/hbase-default.xml</filename>.</para>
+
+<para>At minimum, you should set the list of servers that you want ZooKeeper to run
+on using the <varname>hbase.zookeeper.quorum</varname> property.
+This property defaults to <varname>localhost</varname> which is not suitable for a
+fully distributed HBase (it binds to the local machine only and remote clients
+will not be able to connect).
+It is recommended to run a ZooKeeper quorum of 3, 5 or 7 machines, and give each
+ZooKeeper server around 1GB of RAM, and if possible, its own dedicated disk.
+For very heavily loaded clusters, run ZooKeeper servers on separate machines from the
+Region Servers (DataNodes and TaskTrackers).</para>
+
+
+<para>To point HBase at an existing ZooKeeper cluster, add 
+a suitably configured <filename>zoo.cfg</filename> to the <filename>CLASSPATH</filename>.
+HBase will see this file and use it to figure out where ZooKeeper is.
+Additionally set <varname>HBASE_MANAGES_ZK</varname> in <filename>${HBASE_HOME}/conf/hbase-env.sh</filename>
+to <filename>false</filename> so that HBase doesn't mess with your ZooKeeper setup:</para>
+<programlisting>
+   ...
+  # Tell HBase whether it should manage it's own instance of Zookeeper or not.
+  export HBASE_MANAGES_ZK=false
+</programlisting>
+
+<para>As an example, to have HBase manage a ZooKeeper quorum on nodes
+<empahsis>rs{1,2,3,4,5}.example.com</empahsis>, bound to port 2222 (the default is 2181), use:</para>
+<programlisting>
+  ${HBASE_HOME}/conf/hbase-env.sh:
+
+       ...
+      # Tell HBase whether it should manage it's own instance of Zookeeper or not.
+      export HBASE_MANAGES_ZK=true
+
+  ${HBASE_HOME}/conf/hbase-site.xml:
+
+  &lt;configuration&gt;
+    ...
+    &lt;property&gt;
+      &lt;name&gt;hbase.zookeeper.property.clientPort&lt;/name&gt;
+      &lt;value&gt;2222&lt;/value&gt;
+      &lt;description&gt;Property from ZooKeeper's config zoo.cfg.
+      The port at which the clients will connect.
+      &lt;/description&gt;
+    &lt;/property&gt;
+    ...
+    &lt;property&gt;
+      &lt;name&gt;hbase.zookeeper.quorum&lt;/name&gt;
+      &lt;value&gt;rs1.example.com,rs2.example.com,rs3.example.com,rs4.example.com,rs5.example.com&lt;/value&gt;
+      &lt;description&gt;Comma separated list of servers in the ZooKeeper Quorum.
+      For example, "host1.mydomain.com,host2.mydomain.com,host3.mydomain.com".
+      By default this is set to localhost for local and pseudo-distributed modes
+      of operation. For a fully-distributed setup, this should be set to a full
+      list of ZooKeeper quorum servers. If HBASE_MANAGES_ZK is set in hbase-env.sh
+      this is the list of servers which we will start/stop ZooKeeper on.
+      &lt;/description&gt;
+    &lt;/property&gt;
+    ...
+  &lt;/configuration&gt;
+</programlisting>
+
+<para>When HBase manages ZooKeeper, it will start/stop the ZooKeeper servers as a part
+of the regular start/stop scripts. If you would like to run it yourself, you can
+do:</para>
+<programlisting>
+${HBASE_HOME}/bin/hbase-daemons.sh {start,stop} zookeeper
+</programlisting>
+
+<para>If you do let HBase manage ZooKeeper for you, make sure you configure
+where it's data is stored. By default, it will be stored in /tmp which is
+sometimes cleaned in live systems. Do modify this configuration:</para>
+<programlisting>
+    &lt;property&gt;
+      &lt;name&gt;hbase.zookeeper.property.dataDir&lt;/name&gt;
+      &lt;value&gt;${hbase.tmp.dir}/zookeeper&lt;/value&gt;
+      &lt;description>Property from ZooKeeper's config zoo.cfg.
+      The directory where the snapshot is stored.
+      &lt;/description&gt;
+    &lt;/property&gt;
+</programlisting>
+
+<para>Note that you can use HBase in this manner to spin up a ZooKeeper cluster,
+unrelated to HBase. Just make sure to set <varname>HBASE_MANAGES_ZK</varname> to
+<varname>false</varname> if you want it to stay up so that when HBase shuts down it
+doesn't take ZooKeeper with it.</para>
+
+<para>For more information about setting up a ZooKeeper cluster on your own, see
+the ZooKeeper <link xlink:href="http://hadoop.apache.org/zookeeper/docs/current/zookeeperStarted.html">Getting Started Guide</link>.
+HBase currently uses ZooKeeper version 3.3.2, so any cluster setup with a
+3.x.x version of ZooKeeper should work.</para>
+
+<para>Of note, if you have made <em>HDFS client configuration</em> on your Hadoop cluster, HBase will not
+see this configuration unless you do one of the following:</para>
+<orderedlist>
+  <listitem><para>Add a pointer to your <varname>HADOOP_CONF_DIR</varname> to <varname>CLASSPATH</varname> in <filename>hbase-env.sh</filename>.</para></listitem>
+  <listitem><para>Add a copy of <filename>hdfs-site.xml</filename> (or <filename>hadoop-site.xml</filename>) to <filename>${HBASE_HOME}/conf</filename>, or</para></listitem>
+  <listitem><para>if only a small set of HDFS client configurations, add them to <filename>hbase-site.xml</filename>.</para></listitem>
+</orderedlist>
+
+<para>An example of such an HDFS client configuration is <varname>dfs.replication</varname>. If for example,
+you want to run with a replication factor of 5, hbase will create files with the default of 3 unless
+you do the above to make the configuration available to HBase.</para>
+      </section>
+
+<section xml:id="confirm"><title>Running and Confirming Your Installation</title>
+<para>If you are running in standalone, non-distributed mode, HBase by default uses the local filesystem.</para>
+
+<para>If you are running a distributed cluster you will need to start the Hadoop DFS daemons and
+ZooKeeper Quorum before starting HBase and stop the daemons after HBase has shut down.</para>
+
+<para>Start and stop the Hadoop DFS daemons by running <filename>${HADOOP_HOME}/bin/start-dfs.sh</filename>.
+You can ensure it started properly by testing the put and get of files into the Hadoop filesystem.
+HBase does not normally use the mapreduce daemons.  These do not need to be started.</para>
+
+<para>Start up your ZooKeeper cluster.</para>
+
+<para>Start HBase with the following command:</para>
+<programlisting>
+${HBASE_HOME}/bin/start-hbase.sh
+</programlisting>
+
+<para>Once HBase has started, enter <filename>${HBASE_HOME}/bin/hbase shell</filename> to obtain a
+shell against HBase from which you can execute commands.
+Type 'help' at the shells' prompt to get a list of commands.
+Test your running install by creating tables, inserting content, viewing content, and then dropping your tables.
+For example:</para>
+<programlisting>
+hbase&gt; # Type "help" to see shell help screen
+hbase&gt; help
+hbase&gt; # To create a table named "mylittletable" with a column family of "mylittlecolumnfamily", type
+hbase&gt; create "mylittletable", "mylittlecolumnfamily"
+hbase&gt; # To see the schema for you just created "mylittletable" table and its single "mylittlecolumnfamily", type
+hbase&gt; describe "mylittletable"
+hbase&gt; # To add a row whose id is "myrow", to the column "mylittlecolumnfamily:x" with a value of 'v', do
+hbase&gt; put "mylittletable", "myrow", "mylittlecolumnfamily:x", "v"
+hbase&gt; # To get the cell just added, do
+hbase&gt; get "mylittletable", "myrow"
+hbase&gt; # To scan you new table, do
+hbase&gt; scan "mylittletable"
+</programlisting>
+
+<para>To stop HBase, exit the HBase shell and enter:</para>
+<programlisting>
+${HBASE_HOME}/bin/stop-hbase.sh
+</programlisting>
+
+<para>If you are running a distributed operation, be sure to wait until HBase has shut down completely
+before stopping the Hadoop daemons.</para>
+
+<para>The default location for logs is <filename>${HBASE_HOME}/logs</filename>.</para>
+
+<para>HBase also puts up a UI listing vital attributes. By default its deployed on the master host
+at port 60010 (HBase RegionServers listen on port 60020 by default and put up an informational
+http server at 60030).</para>
+</section>
+
+
+
+
+
+
+
+
+</section>
+</section>
+
+
+
       <section><title>Client configuration and dependencies connecting to an HBase cluster</title>
       <para>TODO</para>
       </section>
 
+
+      <section xml:id="upgrading">
+          <title>Upgrading your HBase Install</title>
+          <para>This version of 0.90.x HBase can be started on data written by
+              HBase 0.20.x or HBase 0.89.x.  There is no need of a migration step.
+              HBase 0.89.x and 0.90.x does write out the name of region directories
+              differently -- it names them with a md5 hash of the region name rather
+              than a jenkins hash -- so this means that once started, there is no
+              going back to HBase 0.20.x.
+          </para>
+      </section>
+
+
+
+
     <section><title>Example Configurations</title>
     <para>In this section we provide a few sample configurations.</para>
     <section><title>Basic Distributed HBase Install</title>
@@ -366,7 +679,7 @@ Below we show what the main configuratio
 <filename>hbase-env.sh</filename> -- found in the <filename>conf</filename> directory
 might look like.
 </para>
-    <section><title><filename>hbase-site.xml</filename></title>
+    <section xml:id="hbase_site"><title><filename>hbase-site.xml</filename></title>
     <programlisting>
 <![CDATA[
 <?xml version="1.0"?>
@@ -404,7 +717,7 @@ might look like.
     </programlisting>
     </section>
 
-    <section><title><filename>regionservers</filename></title>
+    <section xml:id="regionservers"><title><filename>regionservers</filename></title>
     <para>In this file you list the nodes that will run regionservers.  In
     our case we run regionservers on all but the head node example1 which is
     carrying the HBase master and the HDFS namenode</para>
@@ -420,7 +733,7 @@ might look like.
     </programlisting>
     </section>
 
-    <section><title><filename>hbase-env.sh</filename></title>
+    <section xml:id="hbase_env"><title><filename>hbase-env.sh</filename></title>
     <para>Below we use a <command>diff</command> to show the differences from 
     default in the <filename>hbase-env.sh</filename> file. Here we are setting
 the HBase heap to be 4G instead of the default 1G.
@@ -487,7 +800,7 @@ index e70ebc6..96f8c27 100644
       <para></para>
       </section>
 
-      <section>
+      <section xml:id="log4j">
       <title><filename>log4j.properties</filename></title>
       <para></para>
       </section>
@@ -569,7 +882,7 @@ index e70ebc6..96f8c27 100644
         </para>
     </section>
 
-    <section><title>Shell Tricks</title>
+    <section xml:id="shell_tricks"><title>Shell Tricks</title>
         <section><title><filename>irbrc</filename></title>
                 <para>Create an <filename>.irbrc</filename> file for yourself in your
                     home directory. Add HBase Shell customizations. A useful one is
@@ -639,13 +952,13 @@ index e70ebc6..96f8c27 100644
       via the table row key -- its primary key.
 </para>
 
-    <section>
+    <section xml:id="table">
       <title>Table</title>
 
       <para></para>
     </section>
 
-    <section>
+    <section xml:id="row">
       <title>Row</title>
 
       <para></para>
@@ -874,7 +1187,7 @@ index e70ebc6..96f8c27 100644
 
     <subtitle>How HBase is persisted on the Filesystem</subtitle>
 
-    <section>
+    <section xml:id="hfile">
       <title>HFile</title>
 
       <section xml:id="hfile_tool">
@@ -1524,8 +1837,8 @@ index e70ebc6..96f8c27 100644
     </section>
   </chapter>
 
-  <chapter>
-    <title xml:id="wal">The WAL</title>
+  <chapter xml:id="wal">
+    <title >The WAL</title>
 
     <subtitle>HBase's<link
     xlink:href="http://en.wikipedia.org/wiki/Write-ahead_logging"> Write-Ahead
@@ -1540,7 +1853,7 @@ index e70ebc6..96f8c27 100644
       <para>The HBase WAL is...</para>
     </section>
 
-    <section>
+    <section xml:id="wal_splitting">
       <title>WAL splitting</title>
 
       <subtitle>How edits are recovered from a crashed RegionServer</subtitle>
@@ -1584,8 +1897,8 @@ index e70ebc6..96f8c27 100644
 
   </chapter>
 
-  <chapter>
-    <title xml:id="blooms">Bloom Filters</title>
+  <chapter xml:id="blooms">
+    <title>Bloom Filters</title>
 
     <para>Bloom filters were developed over in <link
     xlink:href="https://issues.apache.org/jira/browse/HBASE-1200">HBase-1200
@@ -1658,7 +1971,7 @@ index e70ebc6..96f8c27 100644
       </section>
     </section>
 
-    <section>
+    <section xml:id="bloom_footprint">
       <title>Bloom StoreFile footprint</title>
 
       <para>Bloom filters add an entry to the <classname>StoreFile</classname>
@@ -1791,8 +2104,38 @@ index e70ebc6..96f8c27 100644
     </section>
   </appendix>
 
+  <appendix xml:id="faq">
+    <title >FAQ</title>
+    <qandaset defaultlabel='faq'>
+        <qandadiv><title>General</title>
+        <qandaentry>
+                <question><para>Are there other HBase FAQs?</para></question>
+            <answer>
+                <para>
+              See the FAQ that is up on the wiki, <link xlink:href="http://wiki.apache.org/hadoop/Hbase/FAQ">HBase Wiki FAQ</link>
+              as well as the <link xlink:href="http://wiki.apache.org/hadoop/Hbase/Troubleshooting">Troubleshooting</link> page and
+              the <link xlink:href="http://wiki.apache.org/hadoop/Hbase/FrequentlySeenErrors">Frequently Seen Errors</link> page.
+                </para>
+            </answer>
+        </qandaentry>
+    </qandadiv>
+    <qandadiv xml:id="ec2"><title>EC2</title>
+        <qandaentry>
+            <question><para>
+            Why doesn't my remote java connection into my ec2 cluster work?
+            </para></question>
+            <answer>
+                <para>
+          See Andrew's answer here, up on the user list: <link xlink:href="http://search-hadoop.com/m/sPdqNFAwyg2">Remote Java client connection into EC2 instance</link>.
+                </para>
+            </answer>
+        </qandaentry>
+    </qandadiv>
+    </qandaset>
+  </appendix>
+
 
-  <index>
+  <index xml:id="book_index">
   <title>Index</title>
   </index>
 </book>

Modified: hbase/trunk/src/main/java/org/apache/hadoop/hbase/KeyValue.java
URL: http://svn.apache.org/viewvc/hbase/trunk/src/main/java/org/apache/hadoop/hbase/KeyValue.java?rev=1034230&r1=1034229&r2=1034230&view=diff
==============================================================================
--- hbase/trunk/src/main/java/org/apache/hadoop/hbase/KeyValue.java (original)
+++ hbase/trunk/src/main/java/org/apache/hadoop/hbase/KeyValue.java Fri Nov 12 01:23:19 2010
@@ -1193,7 +1193,7 @@ public class KeyValue implements Writabl
    * changed to be null).  This method does a full copy of the backing byte
    * array and does not modify the original byte array of this KeyValue.
    * <p>
-   * This method is used by {@link KeyOnlyFilter} and is an advanced feature of
+   * This method is used by <code>KeyOnlyFilter</code> and is an advanced feature of
    * KeyValue, proceed with caution.
    */
   public void convertToKeyOnly() {

Modified: hbase/trunk/src/main/java/org/apache/hadoop/hbase/mapreduce/package-info.java
URL: http://svn.apache.org/viewvc/hbase/trunk/src/main/java/org/apache/hadoop/hbase/mapreduce/package-info.java?rev=1034230&r1=1034229&r2=1034230&view=diff
==============================================================================
--- hbase/trunk/src/main/java/org/apache/hadoop/hbase/mapreduce/package-info.java (original)
+++ hbase/trunk/src/main/java/org/apache/hadoop/hbase/mapreduce/package-info.java Fri Nov 12 01:23:19 2010
@@ -62,7 +62,7 @@ is set to the HBase <code>CLASSPATH</cod
 etc., dependencies on the passed
 </code>HADOOP_CLASSPATH</code> and adds the found jars to the mapreduce
 job configuration. See the source at
-{@link TableMapReduceUtil#addDependencyJars(org.apache.hadoop.mapreduce.Job)}
+<code>TableMapReduceUtil#addDependencyJars(org.apache.hadoop.mapreduce.Job)</code>
 for how this is done.
 </p>
 <p>The above may not work if you are running your HBase from its build directory;

Modified: hbase/trunk/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java
URL: http://svn.apache.org/viewvc/hbase/trunk/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java?rev=1034230&r1=1034229&r2=1034230&view=diff
==============================================================================
--- hbase/trunk/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java (original)
+++ hbase/trunk/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java Fri Nov 12 01:23:19 2010
@@ -2103,6 +2103,7 @@ public class HRegionServer implements HR
         list.add(e.getValue().getRegionInfo());
       }
     }
+    Collections.sort(list);
     return list;
   }
 

Modified: hbase/trunk/src/main/java/org/apache/hadoop/hbase/replication/regionserver/Replication.java
URL: http://svn.apache.org/viewvc/hbase/trunk/src/main/java/org/apache/hadoop/hbase/replication/regionserver/Replication.java?rev=1034230&r1=1034229&r2=1034230&view=diff
==============================================================================
--- hbase/trunk/src/main/java/org/apache/hadoop/hbase/replication/regionserver/Replication.java (original)
+++ hbase/trunk/src/main/java/org/apache/hadoop/hbase/replication/regionserver/Replication.java Fri Nov 12 01:23:19 2010
@@ -40,7 +40,7 @@ import org.apache.hadoop.hbase.util.Byte
 import org.apache.zookeeper.KeeperException;
 
 /**
- * Gateway to Replication.  Used by {@link HRegionServer}.
+ * Gateway to Replication.  Used by {@link org.apache.hadoop.hbase.regionserver.HRegionServer}.
  */
 public class Replication implements WALObserver {
   private final boolean replication;
@@ -159,4 +159,4 @@ public class Replication implements WALO
   public void logCloseRequested() {
     // not interested
   }
-}
\ No newline at end of file
+}

Modified: hbase/trunk/src/main/java/org/apache/hadoop/hbase/security/User.java
URL: http://svn.apache.org/viewvc/hbase/trunk/src/main/java/org/apache/hadoop/hbase/security/User.java?rev=1034230&r1=1034229&r2=1034230&view=diff
==============================================================================
--- hbase/trunk/src/main/java/org/apache/hadoop/hbase/security/User.java (original)
+++ hbase/trunk/src/main/java/org/apache/hadoop/hbase/security/User.java Fri Nov 12 01:23:19 2010
@@ -74,7 +74,7 @@ public abstract class User {
   /**
    * Returns the shortened version of the user name -- the portion that maps
    * to an operating system user name.
-   * @return
+   * @return Short name
    */
   public abstract String getShortName();
 

Modified: hbase/trunk/src/main/javadoc/overview.html
URL: http://svn.apache.org/viewvc/hbase/trunk/src/main/javadoc/overview.html?rev=1034230&r1=1034229&r2=1034230&view=diff
==============================================================================
--- hbase/trunk/src/main/javadoc/overview.html (original)
+++ hbase/trunk/src/main/javadoc/overview.html Fri Nov 12 01:23:19 2010
@@ -27,351 +27,29 @@
 <h2>Table of Contents</h2>
 <ul>
 <li>
-  <a href="#requirements">Requirements</a>
-  <ul>
-    <li><a href="#windows">Windows</a></li>
-  </ul>
-</li>
-<li>
   <a href="#getting_started" >Getting Started</a>
-  <ul>
-    <li><a href="#standalone">Standalone</a></li>
-    <li>
-      <a href="#distributed">Distributed Operation: Pseudo- and Fully-distributed modes</a>
-      <ul>
-        <li><a href="#pseudo-distrib">Pseudo-distributed</a></li>
-        <li><a href="#fully-distrib">Fully-distributed</a></li>
-      </ul>
-    </li>
-  </ul>
 </li>
-<li><a href="#runandconfirm">Running and Confirming Your Installation</a></li>
-<li><a href="#upgrading" >Upgrading</a></li>
 <li><a href="#client_example">Example API Usage</a></li>
 <li><a href="#related" >Related Documentation</a></li>
 </ul>
 
 
 <h2><a name="getting_started" >Getting Started</a></h2>
-<p>First review the <a href="http://hbase.apache.org/docs/current/book.html#requirements">requirements</a>
-section of the HBase Book.  A careful reading will save you grief down the road.</p>
-
-<p>What follows presumes you have obtained a copy of HBase,
-see <a href="http://hadoop.apache.org/hbase/releases.html">Releases</a>, and are installing
-for the first time. If upgrading your HBase instance, see <a href="#upgrading">Upgrading</a>.</p>
-
-<p>Three modes are described: <em>standalone</em>, <em>pseudo-distributed</em> (where all servers are run on
-a single host), and <em>fully-distributed</em>. If new to HBase start by following the standalone instructions.</p>
-
-<p>Begin by reading <a href="#requirements">Requirements</a>.</p>
-
-<p>Whatever your mode, define <code>${HBASE_HOME}</code> to be the location of the root of your HBase installation, e.g.
-<code>/user/local/hbase</code>. Edit <code>${HBASE_HOME}/conf/hbase-env.sh</code>. In this file you can
-set the heapsize for HBase, etc. At a minimum, set <code>JAVA_HOME</code> to point at the root of
-your Java installation.</p>
-
-<h3><a name="standalone">Standalone mode</a></h3>
-<p>If you are running a standalone operation, there should be nothing further to configure; proceed to
-<a href="#runandconfirm">Running and Confirming Your Installation</a>. If you are running a distributed
-operation, continue reading.</p>
-
-<h3><a name="distributed">Distributed Operation: Pseudo- and Fully-distributed modes</a></h3>
-<p>Distributed modes require an instance of the <em>Hadoop Distributed File System</em> (DFS).
-See the Hadoop <a href="http://hadoop.apache.org/common/docs/r0.20.1/api/overview-summary.html#overview_description">
-requirements and instructions</a> for how to set up a DFS.</p>
-
-<h4><a name="pseudo-distrib">Pseudo-distributed mode</a></h4>
-<p>A pseudo-distributed mode is simply a distributed mode run on a single host.
-Use this configuration testing and prototyping on hbase.  Do not use this configuration
-for production nor for evaluating HBase performance.
-</p>
-<p>Once you have confirmed your DFS setup, configuring HBase for use on one host requires modification of
-<code>${HBASE_HOME}/conf/hbase-site.xml</code>, which needs to be pointed at the running Hadoop DFS instance.
-Use <code>hbase-site.xml</code> to override the properties defined in
-<code>${HBASE_HOME}/conf/hbase-default.xml</code> (<code>hbase-default.xml</code> itself
-should never be modified) and for HDFS client configurations.
-At a minimum, the <code>hbase.rootdir</code>,
-which points HBase at the Hadoop filesystem to use,
-and the <code>dfs.replication</code>, an hdfs client-side
-configuration stipulating how many replicas to keep up,
-should be redefined in <code>hbase-site.xml</code>. For example,
-adding the properties below to your <code>hbase-site.xml</code> says that HBase
-should use the <code>/hbase</code> 
-directory in the HDFS whose namenode is at port 9000 on your local machine, and that
-it should run with one replica only (recommended for pseudo-distributed mode):</p>
-<blockquote>
-<pre>
-&lt;configuration&gt;
-  ...
-  &lt;property&gt;
-    &lt;name&gt;hbase.rootdir&lt;/name&gt;
-    &lt;value&gt;hdfs://localhost:9000/hbase&lt;/value&gt;
-    &lt;description&gt;The directory shared by region servers.
-    &lt;/description&gt;
-  &lt;/property&gt;
-  &lt;property&gt;
-    &lt;name&gt;dfs.replication&lt;/name&gt;
-    &lt;value&gt;1&lt;/value&gt;
-    &lt;description&gt;The replication count for HLog &amp; HFile storage. Should not be greater than HDFS datanode count.
-    &lt;/description&gt;
-  &lt;/property&gt;
-  ...
-&lt;/configuration&gt;
-</pre>
-</blockquote>
-
-<p>Note: Let HBase create the directory. If you don't, you'll get warning saying HBase
-needs a migration run because the directory is missing files expected by HBase (it'll
-create them if you let it).</p>
-<p>Also Note: Above we bind to localhost.  This means that a remote client cannot
-connect.  Amend accordingly, if you want to connect from a remote location.</p>
-
-<h4><a name="fully-distrib">Fully-Distributed Operation</a></h4>
-<p>For running a fully-distributed operation on more than one host, the following
-configurations must be made <em>in addition</em> to those described in the
-<a href="#pseudo-distrib">pseudo-distributed operation</a> section above.</p>
-
-<p>In <code>hbase-site.xml</code>, set <code>hbase.cluster.distributed</code> to <code>true</code>.</p>
-<blockquote>
-<pre>
-&lt;configuration&gt;
-  ...
-  &lt;property&gt;
-    &lt;name&gt;hbase.cluster.distributed&lt;/name&gt;
-    &lt;value&gt;true&lt;/value&gt;
-    &lt;description&gt;The mode the cluster will be in. Possible values are
-      false: standalone and pseudo-distributed setups with managed Zookeeper
-      true: fully-distributed with unmanaged Zookeeper Quorum (see hbase-env.sh)
-    &lt;/description&gt;
-  &lt;/property&gt;
-  ...
-&lt;/configuration&gt;
-</pre>
-</blockquote>
-
-<p>In fully-distributed mode, you probably want to change your <code>hbase.rootdir</code>
-from localhost to the name of the node running the HDFS NameNode and you should set
-the dfs.replication to be the number of datanodes you have in your cluster or 3, which
-ever is the smaller.
+<p>See the <a href="../book.html#getting_started">Getting Started</a>
+section of the <a href="../book.html">HBase Book</a>.
 </p>
-<p>In addition
-to <code>hbase-site.xml</code> changes, a fully-distributed mode requires that you
-modify <code>${HBASE_HOME}/conf/regionservers</code>.
-The <code>regionserver</code> file lists all hosts running <code>HRegionServer</code>s, one host per line
-(This file in HBase is like the Hadoop slaves file at <code>${HADOOP_HOME}/conf/slaves</code>).</p>
-
-<p>A distributed HBase depends on a running ZooKeeper cluster. All participating nodes and clients
-need to be able to get to the running ZooKeeper cluster.
-HBase by default manages a ZooKeeper cluster for you, or you can manage it on your own and point HBase to it.
-To toggle HBase management of ZooKeeper, use the <code>HBASE_MANAGES_ZK</code> variable in <code>${HBASE_HOME}/conf/hbase-env.sh</code>.
-This variable, which defaults to <code>true</code>, tells HBase whether to
-start/stop the ZooKeeper quorum servers alongside the rest of the servers.</p>
-
-<p>When HBase manages the ZooKeeper cluster, you can specify ZooKeeper configuration
-using its canonical <code>zoo.cfg</code> file (see below), or 
-just specify ZookKeeper options directly in the <code>${HBASE_HOME}/conf/hbase-site.xml</code>
-(If new to ZooKeeper, go the path of specifying your configuration in HBase's hbase-site.xml).
-Every ZooKeeper configuration option has a corresponding property in the HBase hbase-site.xml
-XML configuration file named <code>hbase.zookeeper.property.OPTION</code>.
-For example, the <code>clientPort</code> setting in ZooKeeper can be changed by
-setting the <code>hbase.zookeeper.property.clientPort</code> property.
-For the full list of available properties, see ZooKeeper's <code>zoo.cfg</code>.
-For the default values used by HBase, see <code>${HBASE_HOME}/conf/hbase-default.xml</code>.</p>
-
-<p>At minimum, you should set the list of servers that you want ZooKeeper to run
-on using the <code>hbase.zookeeper.quorum</code> property.
-This property defaults to <code>localhost</code> which is not suitable for a
-fully distributed HBase (it binds to the local machine only and remote clients
-will not be able to connect).
-It is recommended to run a ZooKeeper quorum of 3, 5 or 7 machines, and give each
-ZooKeeper server around 1GB of RAM, and if possible, its own dedicated disk.
-For very heavily loaded clusters, run ZooKeeper servers on separate machines from the
-Region Servers (DataNodes and TaskTrackers).</p>
-
-<p>To point HBase at an existing ZooKeeper cluster, add 
-a suitably configured <code>zoo.cfg</code> to the <code>CLASSPATH</code>.
-HBase will see this file and use it to figure out where ZooKeeper is.
-Additionally set <code>HBASE_MANAGES_ZK</code> in <code>${HBASE_HOME}/conf/hbase-env.sh</code>
-to <code>false</code> so that HBase doesn't mess with your ZooKeeper setup:</p>
-<blockquote>
-<pre>
-   ...
-  # Tell HBase whether it should manage it's own instance of Zookeeper or not.
-  export HBASE_MANAGES_ZK=false
-</pre>
-</blockquote>
-
-<p>As an example, to have HBase manage a ZooKeeper quorum on nodes
-<em>rs{1,2,3,4,5}.example.com</em>, bound to port 2222 (the default is 2181), use:</p>
-<blockquote>
-<pre>
-  ${HBASE_HOME}/conf/hbase-env.sh:
-
-       ...
-      # Tell HBase whether it should manage it's own instance of Zookeeper or not.
-      export HBASE_MANAGES_ZK=true
-
-  ${HBASE_HOME}/conf/hbase-site.xml:
-
-  &lt;configuration&gt;
-    ...
-    &lt;property&gt;
-      &lt;name&gt;hbase.zookeeper.property.clientPort&lt;/name&gt;
-      &lt;value&gt;2222&lt;/value&gt;
-      &lt;description&gt;Property from ZooKeeper's config zoo.cfg.
-      The port at which the clients will connect.
-      &lt;/description&gt;
-    &lt;/property&gt;
-    ...
-    &lt;property&gt;
-      &lt;name&gt;hbase.zookeeper.quorum&lt;/name&gt;
-      &lt;value&gt;rs1.example.com,rs2.example.com,rs3.example.com,rs4.example.com,rs5.example.com&lt;/value&gt;
-      &lt;description&gt;Comma separated list of servers in the ZooKeeper Quorum.
-      For example, "host1.mydomain.com,host2.mydomain.com,host3.mydomain.com".
-      By default this is set to localhost for local and pseudo-distributed modes
-      of operation. For a fully-distributed setup, this should be set to a full
-      list of ZooKeeper quorum servers. If HBASE_MANAGES_ZK is set in hbase-env.sh
-      this is the list of servers which we will start/stop ZooKeeper on.
-      &lt;/description&gt;
-    &lt;/property&gt;
-    ...
-  &lt;/configuration&gt;
-</pre>
-</blockquote>
-
-<p>When HBase manages ZooKeeper, it will start/stop the ZooKeeper servers as a part
-of the regular start/stop scripts. If you would like to run it yourself, you can
-do:</p>
-<blockquote>
-<pre>${HBASE_HOME}/bin/hbase-daemons.sh {start,stop} zookeeper</pre>
-</blockquote>
-
-<p>If you do let HBase manage ZooKeeper for you, make sure you configure
-where it's data is stored. By default, it will be stored in /tmp which is
-sometimes cleaned in live systems. Do modify this configuration:</p>
-<pre>
-    &lt;property&gt;
-      &lt;name&gt;hbase.zookeeper.property.dataDir&lt;/name&gt;
-      &lt;value&gt;${hbase.tmp.dir}/zookeeper&lt;/value&gt;
-      &lt;description>Property from ZooKeeper's config zoo.cfg.
-      The directory where the snapshot is stored.
-      &lt;/description&gt;
-    &lt;/property&gt;
-
-</pre>
-
-<p>Note that you can use HBase in this manner to spin up a ZooKeeper cluster,
-unrelated to HBase. Just make sure to set <code>HBASE_MANAGES_ZK</code> to
-<code>false</code> if you want it to stay up so that when HBase shuts down it
-doesn't take ZooKeeper with it.</p>
-
-<p>For more information about setting up a ZooKeeper cluster on your own, see
-the ZooKeeper <a href="http://hadoop.apache.org/zookeeper/docs/current/zookeeperStarted.html">Getting Started Guide</a>.
-HBase currently uses ZooKeeper version 3.3.1, so any cluster setup with a
-3.x.x version of ZooKeeper should work.</p>
-
-<p>Of note, if you have made <em>HDFS client configuration</em> on your Hadoop cluster, HBase will not
-see this configuration unless you do one of the following:</p>
-<ul>
-  <li>Add a pointer to your <code>HADOOP_CONF_DIR</code> to <code>CLASSPATH</code> in <code>hbase-env.sh</code>.</li>
-  <li>Add a copy of <code>hdfs-site.xml</code> (or <code>hadoop-site.xml</code>) to <code>${HBASE_HOME}/conf</code>, or</li>
-  <li>if only a small set of HDFS client configurations, add them to <code>hbase-site.xml</code>.</li>
-</ul>
-
-<p>An example of such an HDFS client configuration is <code>dfs.replication</code>. If for example,
-you want to run with a replication factor of 5, hbase will create files with the default of 3 unless
-you do the above to make the configuration available to HBase.</p>
-
-
-<h2><a name="runandconfirm">Running and Confirming Your Installation</a></h2>
-<p>If you are running in standalone, non-distributed mode, HBase by default uses the local filesystem.</p>
-
-<p>If you are running a distributed cluster you will need to start the Hadoop DFS daemons and
-ZooKeeper Quorum before starting HBase and stop the daemons after HBase has shut down.</p>
-
-<p>Start and stop the Hadoop DFS daemons by running <code>${HADOOP_HOME}/bin/start-dfs.sh</code>.
-You can ensure it started properly by testing the put and get of files into the Hadoop filesystem.
-HBase does not normally use the mapreduce daemons.  These do not need to be started.</p>
-
-<p>Start up your ZooKeeper cluster.</p>
-
-<p>Start HBase with the following command:</p>
-<blockquote>
-<pre>${HBASE_HOME}/bin/start-hbase.sh</pre>
-</blockquote>
-
-<p>Once HBase has started, enter <code>${HBASE_HOME}/bin/hbase shell</code> to obtain a
-shell against HBase from which you can execute commands.
-Type 'help' at the shells' prompt to get a list of commands.
-Test your running install by creating tables, inserting content, viewing content, and then dropping your tables.
-For example:</p>
-<blockquote>
-<pre>
-hbase&gt; # Type "help" to see shell help screen
-hbase&gt; help
-hbase&gt; # To create a table named "mylittletable" with a column family of "mylittlecolumnfamily", type
-hbase&gt; create "mylittletable", "mylittlecolumnfamily"
-hbase&gt; # To see the schema for you just created "mylittletable" table and its single "mylittlecolumnfamily", type
-hbase&gt; describe "mylittletable"
-hbase&gt; # To add a row whose id is "myrow", to the column "mylittlecolumnfamily:x" with a value of 'v', do
-hbase&gt; put "mylittletable", "myrow", "mylittlecolumnfamily:x", "v"
-hbase&gt; # To get the cell just added, do
-hbase&gt; get "mylittletable", "myrow"
-hbase&gt; # To scan you new table, do
-hbase&gt; scan "mylittletable"
-</pre>
-</blockquote>
-
-<p>To stop HBase, exit the HBase shell and enter:</p>
-<blockquote>
-<pre>${HBASE_HOME}/bin/stop-hbase.sh</pre>
-</blockquote>
-
-<p>If you are running a distributed operation, be sure to wait until HBase has shut down completely
-before stopping the Hadoop daemons.</p>
-
-<p>The default location for logs is <code>${HBASE_HOME}/logs</code>.</p>
-
-<p>HBase also puts up a UI listing vital attributes. By default its deployed on the master host
-at port 60010 (HBase RegionServers listen on port 60020 by default and put up an informational
-http server at 60030).</p>
-
-<h2><a name="upgrading" >Upgrading</a></h2>
-<p>After installing a new HBase on top of data written by a previous HBase version, before
-starting your cluster, run the <code>${HBASE_DIR}/bin/hbase migrate</code> migration script.
-It will make any adjustments to the filesystem data under <code>hbase.rootdir</code> necessary to run
-the HBase version. It does not change your install unless you explicitly ask it to.</p>
 
 <h2><a name="client_example">Example API Usage</a></h2>
 <p>For sample Java code, see <a href="org/apache/hadoop/hbase/client/package-summary.html#package_description">org.apache.hadoop.hbase.client</a> documentation.</p>
 
 <p>If your client is NOT Java, consider the Thrift or REST libraries.</p>
 
-<h2><a name="windows">Windows</a></h2>
-If you are running HBase on Windows, you must install
-<a href="http://cygwin.com/">Cygwin</a>
-to have a *nix-like environment for the shell scripts. The full details
-are explained in 
-the <a href="../cygwin.html">Windows Installation</a>
-guide.
-
 <h2><a name="related" >Related Documentation</a></h2>
 <ul>
-  <li><a href="http://hbase.org">HBase Home Page</a>
-  </li>
-  <li><a href="http://wiki.apache.org/hadoop/Hbase">HBase Wiki</a>
-  </li>
-  <li><a href="http://hadoop.apache.org/">Hadoop Home Page</a>
-  </li>
-  <li><a href="http://wiki.apache.org/hadoop/Hbase/MultipleMasters">Setting up Multiple HBase Masters</a>
-  </li>
-  <li><a href="http://wiki.apache.org/hadoop/Hbase/RollingRestart">Rolling Upgrades</a>
-  </li>
-  <li><a href="org/apache/hadoop/hbase/client/transactional/package-summary.html#package_description">Transactional HBase</a>
-  </li>
-  <li><a href="org/apache/hadoop/hbase/client/tableindexed/package-summary.html">Table Indexed HBase</a>
-  </li>
-  <li><a href="org/apache/hadoop/hbase/stargate/package-summary.html#package_description">Stargate</a> -- a RESTful Web service front end for HBase.
+  <li><a href="http://hbase.org">HBase Home Page</a> </li>
+  <li><a href="http://hbase.org/docs/current/book.html">HBase Book</a> </li>
+  <li><a href="http://wiki.apache.org/hadoop/Hbase">HBase Wiki</a> </li>
+  <li><a href="http://hadoop.apache.org/">Hadoop Home Page</a> </li>
   </li>
 </ul>
 

Modified: hbase/trunk/src/site/site.xml
URL: http://svn.apache.org/viewvc/hbase/trunk/src/site/site.xml?rev=1034230&r1=1034229&r2=1034230&view=diff
==============================================================================
--- hbase/trunk/src/site/site.xml (original)
+++ hbase/trunk/src/site/site.xml Fri Nov 12 01:23:19 2010
@@ -15,7 +15,7 @@
   <version position="right" />
   <publishDate position="right" />
   <body>
-    <menu name="HBase">
+    <menu name="HBase Project">
       <item name="Overview" href="index.html"/>
       <item name="License" href="license.html" />
       <item name="Downloads" href="http://www.apache.org/dyn/closer.cgi/hbase/" />
@@ -23,21 +23,22 @@
       <item name="Issue Tracking" href="issue-tracking.html" />
       <item name="Mailing Lists" href="mail-lists.html" />
       <item name="Source Repository" href="source-repository.html" />
-      <item name="FAQ" href="faq.html" />
-      <item name="Wiki" href="http://wiki.apache.org/hadoop/Hbase" />
       <item name="Team" href="team-list.html" />
     </menu>
     <menu name="Documentation">
-        <item name="Getting Started" href="apidocs/overview-summary.html#overview_description" />
+      <item name="Getting Started: Quick" href="quickstart.html" />
+      <item name="Getting Started: Detailed" href="notsoquick.html" />
       <item name="API" href="apidocs/index.html" />
       <item name="X-Ref" href="xref/index.html" />
+      <item name="Book"      href="book.html" />
+      <item name="FAQ" href="faq.html" />
+      <item name="Wiki" href="http://wiki.apache.org/hadoop/Hbase" />
       <item name="ACID Semantics" href="acid-semantics.html" />
       <item name="Bulk Loads" href="bulk-loads.html" />
       <item name="Metrics"      href="metrics.html" />
       <item name="HBase on Windows"      href="cygwin.html" />
       <item name="Cluster replication"      href="replication.html" />
-      <item name="Pseudo-Distributed HBase"      href="pseudo-distributed.html" />
-      <item name="HBase Book"      href="book.html" />
+      <item name="Pseudo-Dist. Extras"      href="pseudo-distributed.html" />
     </menu>
   </body>
     <skin>
@@ -45,4 +46,3 @@
       <artifactId>maven-stylus-skin</artifactId>
     </skin>
 </project>
-

Modified: hbase/trunk/src/site/xdoc/index.xml
URL: http://svn.apache.org/viewvc/hbase/trunk/src/site/xdoc/index.xml?rev=1034230&r1=1034229&r2=1034230&view=diff
==============================================================================
--- hbase/trunk/src/site/xdoc/index.xml (original)
+++ hbase/trunk/src/site/xdoc/index.xml Fri Nov 12 01:23:19 2010
@@ -53,9 +53,7 @@ HBase includes:
       <p>November 15-19th, <a href="http://www.devoxx.com/display/Devoxx2K10/Home">Devoxx</a> features HBase Training and multiple HBase presentations</p>
       <p>October 12th, HBase-related presentations by core contributors and users at <a href="http://www.cloudera.com/company/press-center/hadoop-world-nyc/">Hadoop World 2010</a></p>
       <p>October 11th, <a href="http://www.meetup.com/hbaseusergroup/calendar/14606174/">HUG-NYC: HBase User Group NYC Edition</a> (Night before Hadoop World)</p>
-      <p>June 30th, <a href="http://www.meetup.com/hbaseusergroup/calendar/13562846/">HBase Contributor Workshop</a> (Day after Hadoop Summit)</p>
-      <p>May 10th, 2010: HBase graduates from Hadoop sub-project to Apache Top Level Project </p>
-      <p><a href="old_news.html">...</a></p>
+      <p><small><a href="old_news.html">Old News</a></small></p>
     </section>
   </body>
   

Modified: hbase/trunk/src/site/xdoc/old_news.xml
URL: http://svn.apache.org/viewvc/hbase/trunk/src/site/xdoc/old_news.xml?rev=1034230&r1=1034229&r2=1034230&view=diff
==============================================================================
--- hbase/trunk/src/site/xdoc/old_news.xml (original)
+++ hbase/trunk/src/site/xdoc/old_news.xml Fri Nov 12 01:23:19 2010
@@ -28,6 +28,8 @@
   </properties>
   <body>
   <section name="Old News">
+      <p>June 30th, <a href="http://www.meetup.com/hbaseusergroup/calendar/13562846/">HBase Contributor Workshop</a> (Day after Hadoop Summit)</p>
+      <p>May 10th, 2010: HBase graduates from Hadoop sub-project to Apache Top Level Project </p>
       <p>Signup for <a href="http://www.meetup.com/hbaseusergroup/calendar/12689490/">HBase User Group Meeting, HUG10</a> hosted by Trend Micro, April 19th, 2010</p>
 
       <p><a href="http://www.meetup.com/hbaseusergroup/calendar/12689351/">HBase User Group Meeting, HUG9</a> hosted by Mozilla, March 10th, 2010</p>



Mime
View raw message