accumulo-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From bil...@apache.org
Subject svn commit: r1304563 [2/4] - in /incubator/accumulo/site/trunk: content/accumulo/ content/accumulo/1.4/ content/accumulo/1.4/examples/ content/accumulo/1.4/user_manual/ content/accumulo/user_manual_1.3-incubating/examples/ content/accumulo/user_manual_...
Date Fri, 23 Mar 2012 19:07:52 GMT
Added: incubator/accumulo/site/trunk/content/accumulo/1.4/user_manual/Accumulo_Design.mdtext
URL: http://svn.apache.org/viewvc/incubator/accumulo/site/trunk/content/accumulo/1.4/user_manual/Accumulo_Design.mdtext?rev=1304563&view=auto
==============================================================================
--- incubator/accumulo/site/trunk/content/accumulo/1.4/user_manual/Accumulo_Design.mdtext (added)
+++ incubator/accumulo/site/trunk/content/accumulo/1.4/user_manual/Accumulo_Design.mdtext Fri Mar 23 19:07:51 2012
@@ -0,0 +1,118 @@
+Title: Apache Accumulo User Manual: Accumulo Design
+Notice:    Licensed to the Apache Software Foundation (ASF) under one
+           or more contributor license agreements.  See the NOTICE file
+           distributed with this work for additional information
+           regarding copyright ownership.  The ASF licenses this file
+           to you under the Apache License, Version 2.0 (the
+           "License"); you may not use this file except in compliance
+           with the License.  You may obtain a copy of the License at
+           .
+             http://www.apache.org/licenses/LICENSE-2.0
+           .
+           Unless required by applicable law or agreed to in writing,
+           software distributed under the License is distributed on an
+           "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+           KIND, either express or implied.  See the License for the
+           specific language governing permissions and limitations
+           under the License.
+
+** Next:** [Accumulo Shell][2] ** Up:** [Apache Accumulo User Manual Version 1.4][4] ** Previous:** [Introduction][6]   ** [Contents][8]**   
+  
+<a id=CHILD_LINKS></a>**Subsections**
+
+* [Data Model][9]
+* [Architecture][10]
+* [Components][11]
+* [Data Management][12]
+* [Tablet Service][13]
+* [Compactions][14]
+* [Fault-Tolerance][15]
+
+* * *
+
+## <a id=Accumulo_Design></a> Accumulo Design
+
+## <a id=Data_Model></a> Data Model
+
+Accumulo provides a richer data model than simple key-value stores, but is not a fully relational database. Data is represented as key-value pairs, where the key and value are comprised of the following elements: 
+
+![converted table][16]
+
+All elements of the Key and the Value are represented as byte arrays except for Timestamp, which is a Long. Accumulo sorts keys by element and lexicographically in ascending order. Timestamps are sorted in descending order so that later versions of the same Key appear first in a sequential scan. Tables consist of a set of sorted key-value pairs. 
+
+## <a id=Architecture></a> Architecture
+
+Accumulo is a distributed data storage and retrieval system and as such consists of several architectural components, some of which run on many individual servers. Much of the work Accumulo does involves maintaining certain properties of the data, such as organization, availability, and integrity, across many commodity-class machines. 
+
+## <a id=Components></a> Components
+
+An instance of Accumulo includes many TabletServers, write-ahead Logger servers, one Garbage Collector process, one Master server and many Clients. 
+
+### <a id=Tablet_Server></a> Tablet Server
+
+The TabletServer manages some subset of all the tablets (partitions of tables). This includes receiving writes from clients, persisting writes to a write‐ahead log, sorting new key‐value pairs in memory, periodically flushing sorted key‐value pairs to new files in HDFS, and responding to reads from clients, forming a merge‐sorted view of all keys and values from all the files it has created and the sorted in‐memory store. 
+
+TabletServers also perform recovery of a tablet that was previously on a server that failed, reapplying any writes found in the write-ahead log to the tablet. 
+
+### <a id=Loggers></a> Loggers
+
+The Loggers accept updates to Tablet servers and write them to local on-disk storage. Each tablet server will write their updates to multiple loggers to preserve data in case of hardware failure. 
+
+### <a id=Garbage_Collector></a> Garbage Collector
+
+Accumulo processes will share files stored in HDFS. Periodically, the Garbage Collector will identify files that are no longer needed by any process, and delete them. 
+
+### <a id=Master></a> Master
+
+The Accumulo Master is responsible for detecting and responding to TabletServer failure. It tries to balance the load across TabletServer by assigning tablets carefully and instructing TabletServers to migrate tablets when necessary. The Master ensures all tablets are assigned to one TabletServer each, and handles table creation, alteration, and deletion requests from clients. The Master also coordinates startup, graceful shutdown and recovery of changes in write-ahead logs when Tablet servers fail. 
+
+### <a id=Client></a> Client
+
+Accumulo includes a client library that is linked to every application. The client library contains logic for finding servers managing a particular tablet, and communicating with TabletServers to write and retrieve key-value pairs. 
+
+## <a id=Data_Management></a> Data Management
+
+Accumulo stores data in tables, which are partitioned into tablets. Tablets are partitioned on row boundaries so that all of the columns and values for a particular row are found together within the same tablet. The Master assigns Tablets to one TabletServer at a time. This enables row-level transactions to take place without using distributed locking or some other complicated synchronization mechanism. As clients insert and query data, and as machines are added and removed from the cluster, the Master migrates tablets to ensure they remain available and that the ingest and query load is balanced across the cluster. 
+
+![Image data_distribution][17]
+
+## <a id=Tablet_Service></a> Tablet Service
+
+When a write arrives at a TabletServer it is written to a Write‐Ahead Log and then inserted into a sorted data structure in memory called a MemTable. When the MemTable reaches a certain size the TabletServer writes out the sorted key-value pairs to a file in HDFS called Indexed Sequential Access Method (ISAM) file. This process is called a minor compaction. A new MemTable is then created and the fact of the compaction is recorded in the Write‐Ahead Log. 
+
+When a request to read data arrives at a TabletServer, the TabletServer does a binary search across the MemTable as well as the in-memory indexes associated with each ISAM file to find the relevant values. If clients are performing a scan, several key‐value pairs are returned to the client in order from the MemTable and the set of ISAM files by performing a merge‐sort as they are read. 
+
+## <a id=Compactions></a> Compactions
+
+In order to manage the number of files per tablet, periodically the TabletServer performs Major Compactions of files within a tablet, in which some set of ISAM files are combined into one file. The previous files will eventually be removed by the Garbage Collector. This also provides an opportunity to permanently remove deleted key‐value pairs by omitting key‐value pairs suppressed by a delete entry when the new file is created. 
+
+## <a id=Fault-Tolerance></a> Fault-Tolerance
+
+If a TabletServer fails, the Master detects it and automatically reassigns the tablets assigned from the failed server to other servers. Any key-value pairs that were in memory at the time the TabletServer are automatically reapplied from the Write-Ahead Log to prevent any loss of data. 
+
+The Master will coordinate the copying of write-ahead logs to HDFS so the logs are available to all tablet servers. To make recovery efficient, the updates within a log are grouped by tablet. The sorting process can be performed by Hadoops MapReduce or the Logger server. TabletServers can quickly apply the mutations from the sorted logs that are destined for the tablets they have now been assigned. 
+
+TabletServer failures are noted on the Master's monitor page, accessible via   
+http://master-address:50095/monitor. 
+
+![Image failure_handling][18]
+
+* * *
+
+** Next:** [Accumulo Shell][2] ** Up:** [Apache Accumulo User Manual Version 1.4][4] ** Previous:** [Introduction][6]   ** [Contents][8]**
+
+   [2]: Accumulo_Shell.html
+   [4]: accumulo_user_manual.html
+   [6]: Introduction.html
+   [8]: Contents.html
+   [9]: Accumulo_Design.html#Data_Model
+   [10]: Accumulo_Design.html#Architecture
+   [11]: Accumulo_Design.html#Components
+   [12]: Accumulo_Design.html#Data_Management
+   [13]: Accumulo_Design.html#Tablet_Service
+   [14]: Accumulo_Design.html#Compactions
+   [15]: Accumulo_Design.html#Fault-Tolerance
+   [16]: img1.png
+   [17]: ./data_distribution.png
+   [18]: ./failure_handling.png
+

Propchange: incubator/accumulo/site/trunk/content/accumulo/1.4/user_manual/Accumulo_Design.mdtext
------------------------------------------------------------------------------
    svn:eol-style = native

Added: incubator/accumulo/site/trunk/content/accumulo/1.4/user_manual/Accumulo_Shell.mdtext
URL: http://svn.apache.org/viewvc/incubator/accumulo/site/trunk/content/accumulo/1.4/user_manual/Accumulo_Shell.mdtext?rev=1304563&view=auto
==============================================================================
--- incubator/accumulo/site/trunk/content/accumulo/1.4/user_manual/Accumulo_Shell.mdtext (added)
+++ incubator/accumulo/site/trunk/content/accumulo/1.4/user_manual/Accumulo_Shell.mdtext Fri Mar 23 19:07:51 2012
@@ -0,0 +1,152 @@
+Title: Apache Accumulo User Manual: Accumulo Shell
+Notice:    Licensed to the Apache Software Foundation (ASF) under one
+           or more contributor license agreements.  See the NOTICE file
+           distributed with this work for additional information
+           regarding copyright ownership.  The ASF licenses this file
+           to you under the Apache License, Version 2.0 (the
+           "License"); you may not use this file except in compliance
+           with the License.  You may obtain a copy of the License at
+           .
+             http://www.apache.org/licenses/LICENSE-2.0
+           .
+           Unless required by applicable law or agreed to in writing,
+           software distributed under the License is distributed on an
+           "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+           KIND, either express or implied.  See the License for the
+           specific language governing permissions and limitations
+           under the License.
+
+** Next:** [Writing Accumulo Clients][2] ** Up:** [Apache Accumulo User Manual Version 1.4][4] ** Previous:** [Accumulo Design][6]   ** [Contents][8]**   
+  
+<a id=CHILD_LINKS></a>**Subsections**
+
+* [Basic Administration][9]
+* [Table Maintenance][10]
+* [User Administration][11]
+
+* * *
+
+## <a id=Accumulo_Shell></a> Accumulo Shell
+
+Accumulo provides a simple shell that can be used to examine the contents and configuration settings of tables, insert/update/delete values, and change configuration settings. 
+
+The shell can be started by the following command: 
+    
+    
+    $ACCUMULO_HOME/bin/accumulo shell -u [username]
+    
+
+The shell will prompt for the corresponding password to the username specified and then display the following prompt: 
+    
+    
+    Shell - Apache Accumulo Interactive Shell
+    -
+    - version 1.3
+    - instance name: myinstance
+    - instance id: 00000000-0000-0000-0000-000000000000
+    -
+    - type 'help' for a list of available commands
+    -
+    
+
+## <a id=Basic_Administration></a> Basic Administration
+
+The Accumulo shell can be used to create and delete tables, as well as to configure table and instance specific options. 
+    
+    
+    root@myinstance> tables
+    !METADATA
+    
+    root@myinstance> createtable mytable
+    
+    root@myinstance mytable>
+    
+    root@myinstance mytable> tables
+    !METADATA
+    mytable
+    
+    root@myinstance mytable> createtable testtable
+    
+    root@myinstance testtable>
+    
+    root@myinstance junk> deletetable testtable
+    
+    root@myinstance>
+    
+
+The Shell can also be used to insert updates and scan tables. This is useful for inspecting tables. 
+    
+    
+    root@myinstance mytable> scan
+    
+    root@myinstance mytable> insert row1 colf colq value1
+    insert successful
+    
+    root@myinstance mytable> scan
+    row1 colf:colq [] value1
+    
+
+The value in brackets "[]" would be the visibility labels. Since none were used, this is empty for this row. You can use the "-t" option to scan to see the timestamp for the cell, too. 
+
+## <a id=Table_Maintenance></a> Table Maintenance
+
+The **compact** command instructs Accumulo to schedule a compaction of the table during which files are consolidated and deleted entries are removed. 
+    
+    
+    root@myinstance mytable> compact -t mytable
+    07 16:13:53,201 [shell.Shell] INFO : Compaction of table mytable
+    scheduled for 20100707161353EDT
+    
+
+The **flush** command instructs Accumulo to write all entries currently in memory for a given table to disk. 
+    
+    
+    root@myinstance mytable> flush -t mytable
+    07 16:14:19,351 [shell.Shell] INFO : Flush of table mytable
+    initiated...
+    
+
+## <a id=User_Administration></a> User Administration
+
+The Shell can be used to add, remove, and grant privileges to users. 
+    
+    
+    root@myinstance mytable> createuser bob
+    Enter new password for 'bob': *********
+    Please confirm new password for 'bob': *********
+    
+    root@myinstance mytable> authenticate bob
+    Enter current password for 'bob': *********
+    Valid
+    
+    root@myinstance mytable> grant System.CREATE_TABLE -s -u bob
+    
+    root@myinstance mytable> user bob
+    Enter current password for 'bob': *********
+    
+    bob@myinstance mytable> userpermissions
+    System permissions: System.CREATE_TABLE
+    Table permissions (!METADATA): Table.READ
+    Table permissions (mytable): NONE
+    
+    bob@myinstance mytable> createtable bobstable
+    bob@myinstance bobstable>
+    
+    bob@myinstance bobstable> user root
+    Enter current password for 'root': *********
+    
+    root@myinstance bobstable> revoke System.CREATE_TABLE -s -u bob
+    
+
+* * *
+
+** Next:** [Writing Accumulo Clients][2] ** Up:** [Apache Accumulo User Manual Version 1.4][4] ** Previous:** [Accumulo Design][6]   ** [Contents][8]**
+
+   [2]: Writing_Accumulo_Clients.html
+   [4]: accumulo_user_manual.html
+   [6]: Accumulo_Design.html
+   [8]: Contents.html
+   [9]: Accumulo_Shell.html#Basic_Administration
+   [10]: Accumulo_Shell.html#Table_Maintenance
+   [11]: Accumulo_Shell.html#User_Administration
+

Propchange: incubator/accumulo/site/trunk/content/accumulo/1.4/user_manual/Accumulo_Shell.mdtext
------------------------------------------------------------------------------
    svn:eol-style = native

Added: incubator/accumulo/site/trunk/content/accumulo/1.4/user_manual/Administration.mdtext
URL: http://svn.apache.org/viewvc/incubator/accumulo/site/trunk/content/accumulo/1.4/user_manual/Administration.mdtext?rev=1304563&view=auto
==============================================================================
--- incubator/accumulo/site/trunk/content/accumulo/1.4/user_manual/Administration.mdtext (added)
+++ incubator/accumulo/site/trunk/content/accumulo/1.4/user_manual/Administration.mdtext Fri Mar 23 19:07:51 2012
@@ -0,0 +1,183 @@
+Title: Apache Accumulo User Manual: Administration
+Notice:    Licensed to the Apache Software Foundation (ASF) under one
+           or more contributor license agreements.  See the NOTICE file
+           distributed with this work for additional information
+           regarding copyright ownership.  The ASF licenses this file
+           to you under the Apache License, Version 2.0 (the
+           "License"); you may not use this file except in compliance
+           with the License.  You may obtain a copy of the License at
+           .
+             http://www.apache.org/licenses/LICENSE-2.0
+           .
+           Unless required by applicable law or agreed to in writing,
+           software distributed under the License is distributed on an
+           "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+           KIND, either express or implied.  See the License for the
+           specific language governing permissions and limitations
+           under the License.
+
+** Next:** [Shell Commands][2] ** Up:** [Apache Accumulo User Manual Version 1.4][4] ** Previous:** [Security][6]   ** [Contents][8]**   
+  
+<a id=CHILD_LINKS></a>**Subsections**
+
+* [Hardware][9]
+* [Network][10]
+* [Installation][11]
+* [Dependencies][12]
+* [Configuration][13]
+* [Initialization][14]
+* [Running][15]
+* [Monitoring][16]
+* [Logging][17]
+* [Recovery][18]
+
+* * *
+
+## <a id=Administration></a> Administration
+
+## <a id=Hardware></a> Hardware
+
+Because we are running essentially two or three systems simultaneously layered across the cluster: HDFS, Accumulo and MapReduce, it is typical for hardware to consist of 4 to 8 cores, and 8 to 32 GB RAM. This is so each running process can have at least one core and 2 - 4 GB each. 
+
+One core running HDFS can typically keep 2 to 4 disks busy, so each machine may typically have as little as 2 x 300GB disks and as much as 4 x 1TB or 2TB disks. 
+
+It is possible to do with less than this, such as with 1u servers with 2 cores and 4GB each, but in this case it is recommended to only run up to two processes per machine - i.e. DataNode and TabletServer or DataNode and MapReduce worker but not all three. The constraint here is having enough available heap space for all the processes on a machine. 
+
+## <a id=Network></a> Network
+
+Accumulo communicates via remote procedure calls over TCP/IP for both passing data and control messages. In addition, Accumulo uses HDFS clients to communicate with HDFS. To achieve good ingest and query performance, sufficient network bandwidth must be available between any two machines. 
+
+## <a id=Installation></a> Installation
+
+Choose a directory for the Accumulo installation. This directory will be referenced by the environment variable $ACCUMULO_HOME. Run the following: 
+    
+    
+    $ tar xzf $ACCUMULO_HOME/accumulo.tar.gz
+    
+
+Repeat this step at each machine within the cluster. Usually all machines have the same $ACCUMULO_HOME. 
+
+## <a id=Dependencies></a> Dependencies
+
+Accumulo requires HDFS and ZooKeeper to be configured and running before starting. Password-less SSH should be configured between at least the Accumulo master and TabletServer machines. It is also a good idea to run Network Time Protocol (NTP) within the cluster to ensure nodes' clocks don't get too out of sync, which can cause problems with automatically timestamped data. Accumulo will remove from the set of TabletServers those machines whose times differ too much from the master's. 
+
+## <a id=Configuration></a> Configuration
+
+Accumulo is configured by editing several Shell and XML files found in $ACCUMULO_HOME/conf. The structure closely resembles Hadoop's configuration files. 
+
+### <a id=Edit_conf/accumulo-env.sh></a> Edit conf/accumulo-env.sh
+
+Accumulo needs to know where to find the software it depends on. Edit accumulo-env.sh and specify the following: 
+
+1. Enter the location of the installation directory of Accumulo for $ACCUMULO_HOME
+2. Enter your system's Java home for $JAVA_HOME
+3. Enter the location of Hadoop for $HADOOP_HOME
+4. Choose a location for Accumulo logs and enter it for $ACCUMULO_LOG_DIR
+5. Enter the location of ZooKeeper for $ZOOKEEPER_HOME
+
+By default Accumulo TabletServers are set to use 1GB of memory. You may change this by altering the value of $ACCUMULO_TSERVER_OPTS. Note the syntax is that of the Java JVM command line options. This value should be less than the physical memory of the machines running TabletServers. 
+
+There are similar options for the master's memory usage and the garbage collector process. Reduce these if they exceed the physical RAM of your hardware and increase them, within the bounds of the physical RAM, if a process fails because of insufficient memory. 
+
+Note that you will be specifying the Java heap space in accumulo-env.sh. You should make sure that the total heap space used for the Accumulo tserver and the Hadoop DataNode and TaskTracker is less than the available memory on each slave node in the cluster. On large clusters, it is recommended that the Accumulo master, Hadoop NameNode, secondary NameNode, and Hadoop JobTracker all be run on separate machines to allow them to use more heap space. If you are running these on the same machine on a small cluster, likewise make sure their heap space settings fit within the available memory. 
+
+### <a id=Cluster_Specification></a> Cluster Specification
+
+On the machine that will serve as the Accumulo master: 
+
+1. Write the IP address or domain name of the Accumulo Master to the   
+$ACCUMULO_HOME/conf/masters file. 
+2. Write the IP addresses or domain name of the machines that will be TabletServers in   
+$ACCUMULO_HOME/conf/slaves, one per line. 
+
+Note that if using domain names rather than IP addresses, DNS must be configured properly for all machines participating in the cluster. DNS can be a confusing source of errors. 
+
+### <a id=Accumulo_Settings></a> Accumulo Settings
+
+Specify appropriate values for the following settings in   
+$ACCUMULO_HOME/conf/accumulo-site.xml : 
+    
+    
+    <property>
+        <name>zookeeper</name>
+        <value>zooserver-one:2181,zooserver-two:2181</value>
+        <description>list of zookeeper servers</description>
+    </property>
+    <property>
+        <name>walog</name>
+        <value>/var/accumulo/walogs</value>
+        <description>local directory for write ahead logs</description>
+    </property>
+    
+
+This enables Accumulo to find ZooKeeper. Accumulo uses ZooKeeper to coordinate settings between processes and helps finalize TabletServer failure. 
+
+Accumulo records all changes to tables to a write-ahead log before committing them to the table. The `walog' setting specifies the local directory on each machine to which write-ahead logs are written. This directory should exist on all machines acting as TabletServers. 
+
+Some settings can be modified via the Accumulo shell and take effect immediately. However, any settings that should be persisted across system restarts must be recorded in the accumulo-site.xml file. 
+
+### <a id=Deploy_Configuration></a> Deploy Configuration
+
+Copy the masters, slaves, accumulo-env.sh, and if necessary, accumulo-site.xml from the   
+$ACCUMULO_HOME/conf/ directory on the master to all the machines specified in the slaves file. 
+
+## <a id=Initialization></a> Initialization
+
+Accumulo must be initialized to create the structures it uses internally to locate data across the cluster. HDFS is required to be configured and running before Accumulo can be initialized. 
+
+Once HDFS is started, initialization can be performed by executing   
+$ACCUMULO_HOME/bin/accumulo init . This script will prompt for a name for this instance of Accumulo. The instance name is used to identify a set of tables and instance-specific settings. The script will then write some information into HDFS so Accumulo can start properly. 
+
+The initialization script will prompt you to set a root password. Once Accumulo is initialized it can be started. 
+
+## <a id=Running></a> Running
+
+### <a id=Starting_Accumulo></a> Starting Accumulo
+
+Make sure Hadoop is configured on all of the machines in the cluster, including access to a shared HDFS instance. Make sure HDFS and ZooKeeper are running. Make sure ZooKeeper is configured and running on at least one machine in the cluster. Start Accumulo using the bin/start-all.sh script. 
+
+To verify that Accumulo is running, check the Status page as described under *Monitoring*. In addition, the Shell can provide some information about the status of tables via reading the !METADATA table. 
+
+### <a id=Stopping_Accumulo></a> Stopping Accumulo
+
+To shutdown cleanly, run bin/stop-all.sh and the master will orchestrate the shutdown of all the tablet servers. Shutdown waits for all minor compactions to finish, so it may take some time for particular configurations. 
+
+## <a id=Monitoring></a> Monitoring
+
+The Accumulo Master provides an interface for monitoring the status and health of Accumulo components. This interface can be accessed by pointing a web browser to   
+http://accumulomaster:50095/status
+
+## <a id=Logging></a> Logging
+
+Accumulo processes each write to a set of log files. By default these are found under   
+$ACCUMULO/logs/. 
+
+## <a id=Recovery></a> Recovery
+
+In the event of TabletServer failure or error on shutting Accumulo down, some mutations may not have been minor compacted to HDFS properly. In this case, Accumulo will automatically reapply such mutations from the write-ahead log either when the tablets from the failed server are reassigned by the Master, in the case of a single TabletServer failure or the next time Accumulo starts, in the event of failure during shutdown. 
+
+Recovery is performed by asking the loggers to copy their write-ahead logs into HDFS. As the logs are copied, they are also sorted, so that tablets can easily find their missing updates. The copy/sort status of each file is displayed on Accumulo monitor status page. Once the recovery is complete any tablets involved should return to an ``online" state. Until then those tablets will be unavailable to clients. 
+
+The Accumulo client library is configured to retry failed mutations and in many cases clients will be able to continue processing after the recovery process without throwing an exception. 
+
+Note that because Accumulo uses timestamps to order mutations, any mutations that are applied as part of the recovery process should appear to have been applied when they originally arrived at the TabletServer that failed. This makes the ordering of mutations consistent in the presence of failure. 
+
+* * *
+
+** Next:** [Shell Commands][2] ** Up:** [Apache Accumulo User Manual Version 1.4][4] ** Previous:** [Security][6]   ** [Contents][8]**
+
+   [2]: Shell_Commands.html
+   [4]: accumulo_user_manual.html
+   [6]: Security.html
+   [8]: Contents.html
+   [9]: Administration.html#Hardware
+   [10]: Administration.html#Network
+   [11]: Administration.html#Installation
+   [12]: Administration.html#Dependencies
+   [13]: Administration.html#Configuration
+   [14]: Administration.html#Initialization
+   [15]: Administration.html#Running
+   [16]: Administration.html#Monitoring
+   [17]: Administration.html#Logging
+   [18]: Administration.html#Recovery
+

Propchange: incubator/accumulo/site/trunk/content/accumulo/1.4/user_manual/Administration.mdtext
------------------------------------------------------------------------------
    svn:eol-style = native

Added: incubator/accumulo/site/trunk/content/accumulo/1.4/user_manual/Analytics.mdtext
URL: http://svn.apache.org/viewvc/incubator/accumulo/site/trunk/content/accumulo/1.4/user_manual/Analytics.mdtext?rev=1304563&view=auto
==============================================================================
--- incubator/accumulo/site/trunk/content/accumulo/1.4/user_manual/Analytics.mdtext (added)
+++ incubator/accumulo/site/trunk/content/accumulo/1.4/user_manual/Analytics.mdtext Fri Mar 23 19:07:51 2012
@@ -0,0 +1,164 @@
+Title: Apache Accumulo User Manual: Analytics
+Notice:    Licensed to the Apache Software Foundation (ASF) under one
+           or more contributor license agreements.  See the NOTICE file
+           distributed with this work for additional information
+           regarding copyright ownership.  The ASF licenses this file
+           to you under the Apache License, Version 2.0 (the
+           "License"); you may not use this file except in compliance
+           with the License.  You may obtain a copy of the License at
+           .
+             http://www.apache.org/licenses/LICENSE-2.0
+           .
+           Unless required by applicable law or agreed to in writing,
+           software distributed under the License is distributed on an
+           "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+           KIND, either express or implied.  See the License for the
+           specific language governing permissions and limitations
+           under the License.
+
+** Next:** [Security][2] ** Up:** [Apache Accumulo User Manual Version 1.4][4] ** Previous:** [High-Speed Ingest][6]   ** [Contents][8]**   
+  
+<a id=CHILD_LINKS></a>**Subsections**
+
+* [MapReduce][9]
+* [Combiners][10]
+* [Statistical Modeling][11]
+
+* * *
+
+## <a id=Analytics></a> Analytics
+
+Accumulo supports more advanced data processing than simply keeping keys sorted and performing efficient lookups. Analytics can be developed by using MapReduce and Iterators in conjunction with Accumulo tables. 
+
+## <a id=MapReduce></a> MapReduce
+
+Accumulo tables can be used as the source and destination of MapReduce jobs. To use an Accumulo table with a MapReduce job (specifically with the new Hadoop API as of version 0.20), configure the job parameters to use the AccumuloInputFormat and AccumuloOutputFormat. Accumulo specific parameters can be set via these two format classes to do the following: 
+
+* Authenticate and provide user credentials for the input 
+* Restrict the scan to a range of rows 
+* Restrict the input to a subset of available columns 
+
+### <a id=Mapper_and_Reducer_classes></a> Mapper and Reducer classes
+
+To read from an Accumulo table create a Mapper with the following class parameterization and be sure to configure the AccumuloInputFormat. 
+    
+    
+    class MyMapper extends Mapper<Key,Value,WritableComparable,Writable> {
+        public void map(Key k, Value v, Context c) {
+            // transform key and value data here
+        }
+    }
+    
+
+To write to an Accumulo table, create a Reducer with the following class parameterization and be sure to configure the AccumuloOutputFormat. The key emitted from the Reducer identifies the table to which the mutation is sent. This allows a single Reducer to write to more than one table if desired. A default table can be configured using the AccumuloOutputFormat, in which case the output table name does not have to be passed to the Context object within the Reducer. 
+    
+    
+    class MyReducer extends Reducer<WritableComparable, Writable, Text, Mutation> {
+    
+        public void reduce(WritableComparable key, Iterator<Text> values, Context c) {
+            
+            Mutation m;
+            
+            // create the mutation based on input key and value
+            
+            c.write(new Text("output-table"), m);
+        }
+    }
+    
+
+The Text object passed as the output should contain the name of the table to which this mutation should be applied. The Text can be null in which case the mutation will be applied to the default table name specified in the AccumuloOutputFormat options. 
+
+### <a id=AccumuloInputFormat_options></a> AccumuloInputFormat options
+    
+    
+    Job job = new Job(getConf());
+    AccumuloInputFormat.setInputInfo(job,
+            "user",
+            "passwd".getBytes(),
+            "table",
+            new Authorizations());
+    
+    AccumuloInputFormat.setZooKeeperInstance(job, "myinstance",
+            "zooserver-one,zooserver-two");
+    
+
+**Optional settings:**
+
+To restrict Accumulo to a set of row ranges: 
+    
+    
+    ArrayList<Range> ranges = new ArrayList<Range>();
+    // populate array list of row ranges ...
+    AccumuloInputFormat.setRanges(job, ranges);
+    
+
+To restrict accumulo to a list of columns: 
+    
+    
+    ArrayList<Pair<Text,Text>> columns = new ArrayList<Pair<Text,Text>>();
+    // populate list of columns
+    AccumuloInputFormat.fetchColumns(job, columns);
+    
+
+To use a regular expression to match row IDs: 
+    
+    
+    AccumuloInputFormat.setRegex(job, RegexType.ROW, "^.*");
+    
+
+### <a id=AccumuloOutputFormat_options></a> AccumuloOutputFormat options
+    
+    
+    boolean createTables = true;
+    String defaultTable = "mytable";
+    
+    AccumuloOutputFormat.setOutputInfo(job,
+            "user",
+            "passwd".getBytes(),
+            createTables,
+            defaultTable);
+    
+    AccumuloOutputFormat.setZooKeeperInstance(job, "myinstance",
+            "zooserver-one,zooserver-two");
+    
+
+**Optional Settings:**
+    
+    
+    AccumuloOutputFormat.setMaxLatency(job, 300); // milliseconds
+    AccumuloOutputFormat.setMaxMutationBufferSize(job, 5000000); // bytes
+    
+
+An example of using MapReduce with Accumulo can be found at   
+accumulo/docs/examples/README.mapred 
+
+## <a id=Combiners></a> Combiners
+
+Many applications can benefit from the ability to aggregate values across common keys. This can be done via Combiner iterators and is similar to the Reduce step in MapReduce. This provides the ability to define online, incrementally updated analytics without the overhead or latency associated with batch-oriented MapReduce jobs. 
+
+All that is needed to aggregate values of a table is to identify the fields over which values will be grouped, insert mutations with those fields as the key, and configure the table with a combining iterator that supports the summarizing operation desired. 
+
+The only restriction on an combining iterator is that the combiner developer should not assume that all values for a given key have been seen, since new mutations can be inserted at anytime. This precludes using the total number of values in the aggregation such as when calculating an average, for example. 
+
+### <a id=Feature_Vectors></a> Feature Vectors
+
+An interesting use of combining iterators within an Accumulo table is to store feature vectors for use in machine learning algorithms. For example, many algorithms such as k-means clustering, support vector machines, anomaly detection, etc. use the concept of a feature vector and the calculation of distance metrics to learn a particular model. The columns in an Accumulo table can be used to efficiently store sparse features and their weights to be incrementally updated via the use of an combining iterator. 
+
+## <a id=Statistical_Modeling></a> Statistical Modeling
+
+Statistical models that need to be updated by many machines in parallel could be similarly stored within an Accumulo table. For example, a MapReduce job that is iteratively updating a global statistical model could have each map or reduce worker reference the parts of the model to be read and updated through an embedded Accumulo client. 
+
+Using Accumulo this way enables efficient and fast lookups and updates of small pieces of information in a random access pattern, which is complementary to MapReduce's sequential access model. 
+
+* * *
+
+** Next:** [Security][2] ** Up:** [Apache Accumulo User Manual Version 1.4][4] ** Previous:** [High-Speed Ingest][6]   ** [Contents][8]**
+
+   [2]: Security.html
+   [4]: accumulo_user_manual.html
+   [6]: High_Speed_Ingest.html
+   [8]: Contents.html
+   [9]: Analytics.html#MapReduce
+   [10]: Analytics.html#Combiners
+   [11]: Analytics.html#Statistical_Modeling
+

Propchange: incubator/accumulo/site/trunk/content/accumulo/1.4/user_manual/Analytics.mdtext
------------------------------------------------------------------------------
    svn:eol-style = native

Added: incubator/accumulo/site/trunk/content/accumulo/1.4/user_manual/Contents.mdtext
URL: http://svn.apache.org/viewvc/incubator/accumulo/site/trunk/content/accumulo/1.4/user_manual/Contents.mdtext?rev=1304563&view=auto
==============================================================================
--- incubator/accumulo/site/trunk/content/accumulo/1.4/user_manual/Contents.mdtext (added)
+++ incubator/accumulo/site/trunk/content/accumulo/1.4/user_manual/Contents.mdtext Fri Mar 23 19:07:51 2012
@@ -0,0 +1,266 @@
+Title: Apache Accumulo User Manual: Contents
+Notice:    Licensed to the Apache Software Foundation (ASF) under one
+           or more contributor license agreements.  See the NOTICE file
+           distributed with this work for additional information
+           regarding copyright ownership.  The ASF licenses this file
+           to you under the Apache License, Version 2.0 (the
+           "License"); you may not use this file except in compliance
+           with the License.  You may obtain a copy of the License at
+           .
+             http://www.apache.org/licenses/LICENSE-2.0
+           .
+           Unless required by applicable law or agreed to in writing,
+           software distributed under the License is distributed on an
+           "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+           KIND, either express or implied.  See the License for the
+           specific language governing permissions and limitations
+           under the License.
+
+** Next:** [Introduction][2] ** Up:** [Apache Accumulo User Manual Version 1.4][4] ** Previous:** [Apache Accumulo User Manual Version 1.4][4]   
+  
+  
+
+
+### <a id=Contents></a> Contents
+
+* [Introduction][2]
+* [Accumulo Design][6]
+
+    * [Data Model][7]
+    * [Architecture][8]
+    * [Components][9]
+
+        * [Tablet Server][10]
+        * [Loggers][11]
+        * [Garbage Collector][12]
+        * [Master][13]
+        * [Client][14]
+
+    * [Data Management][15]
+    * [Tablet Service][16]
+    * [Compactions][17]
+    * [Fault-Tolerance][18]
+
+  
+
+* [Accumulo Shell][19]
+
+    * [Basic Administration][20]
+    * [Table Maintenance][21]
+    * [User Administration][22]
+
+  
+
+* [Writing Accumulo Clients][23]
+
+    * [Running Client Code][24]
+    * [Connecting][25]
+    * [Writing Data][26]
+
+        * [BatchWriter][27]
+
+    * [Reading Data][28]
+
+        * [Scanner][29]
+        * [Isolated Scanner][30]
+        * [BatchScanner][31]
+
+  
+
+* [Table Configuration][32]
+
+    * [Locality Groups][33]
+
+        * [Managing Locality Groups via the Shell][34]
+        * [Managing Locality Groups via the Client API][35]
+
+    * [Constraints][36]
+    * [Bloom Filters][37]
+    * [Iterators][38]
+
+        * [Setting Iterators via the Shell][39]
+        * [Setting Iterators Programmatically][40]
+        * [Versioning Iterators and Timestamps][41]
+        * [Filters][42]
+        * [Combiners][43]
+
+    * [Block Cache][44]
+    * [Compaction][45]
+    * [Pre-splitting tables][46]
+    * [Merging tablets][47]
+    * [Delete Range][48]
+    * [Cloning Tables][49]
+
+  
+
+* [Table Design][50]
+
+    * [Basic Table][51]
+    * [RowID Design][52]
+    * [Indexing][53]
+    * [Entity-Attribute and Graph Tables][54]
+    * [Document-Partitioned Indexing][55]
+
+  
+
+* [High-Speed Ingest][56]
+
+    * [Pre-Splitting New Tables][57]
+    * [Multiple Ingester Clients][58]
+    * [Bulk Ingest][59]
+    * [Logical Time for Bulk Ingest][60]
+    * [MapReduce Ingest][61]
+
+  
+
+* [Analytics][62]
+
+    * [MapReduce][63]
+
+        * [Mapper and Reducer classes][64]
+        * [AccumuloInputFormat options][65]
+        * [AccumuloOutputFormat options][66]
+
+    * [Combiners][67]
+
+        * [Feature Vectors][68]
+
+    * [Statistical Modeling][69]
+
+  
+
+* [Security][70]
+
+    * [Security Label Expressions][71]
+    * [Security Label Expression Syntax][72]
+    * [Authorization][73]
+    * [User Authorizations][74]
+    * [Secure Authorizations Handling][75]
+    * [Query Services Layer][76]
+
+  
+
+* [Administration][77]
+
+    * [Hardware][78]
+    * [Network][79]
+    * [Installation][80]
+    * [Dependencies][81]
+    * [Configuration][82]
+
+        * [Edit conf/accumulo-env.sh][83]
+        * [Cluster Specification][84]
+        * [Accumulo Settings][85]
+        * [Deploy Configuration][86]
+
+    * [Initialization][87]
+    * [Running][88]
+
+        * [Starting Accumulo][89]
+        * [Stopping Accumulo][90]
+
+    * [Monitoring][91]
+    * [Logging][92]
+    * [Recovery][93]
+
+  
+
+* [Shell Commands][94]
+
+  
+
+
+* * *
+
+   [2]: Introduction.html
+   [4]: accumulo_user_manual.html
+   [6]: Accumulo_Design.html
+   [7]: Accumulo_Design.html#Data_Model
+   [8]: Accumulo_Design.html#Architecture
+   [9]: Accumulo_Design.html#Components
+   [10]: Accumulo_Design.html#Tablet_Server
+   [11]: Accumulo_Design.html#Loggers
+   [12]: Accumulo_Design.html#Garbage_Collector
+   [13]: Accumulo_Design.html#Master
+   [14]: Accumulo_Design.html#Client
+   [15]: Accumulo_Design.html#Data_Management
+   [16]: Accumulo_Design.html#Tablet_Service
+   [17]: Accumulo_Design.html#Compactions
+   [18]: Accumulo_Design.html#Fault-Tolerance
+   [19]: Accumulo_Shell.html
+   [20]: Accumulo_Shell.html#Basic_Administration
+   [21]: Accumulo_Shell.html#Table_Maintenance
+   [22]: Accumulo_Shell.html#User_Administration
+   [23]: Writing_Accumulo_Clients.html
+   [24]: Writing_Accumulo_Clients.html#Running_Client_Code
+   [25]: Writing_Accumulo_Clients.html#Connecting
+   [26]: Writing_Accumulo_Clients.html#Writing_Data
+   [27]: Writing_Accumulo_Clients.html#BatchWriter
+   [28]: Writing_Accumulo_Clients.html#Reading_Data
+   [29]: Writing_Accumulo_Clients.html#Scanner
+   [30]: Writing_Accumulo_Clients.html#Isolated_Scanner
+   [31]: Writing_Accumulo_Clients.html#BatchScanner
+   [32]: Table_Configuration.html
+   [33]: Table_Configuration.html#Locality_Groups
+   [34]: Table_Configuration.html#Managing_Locality_Groups_via_the_Shell
+   [35]: Table_Configuration.html#Managing_Locality_Groups_via_the_Client_API
+   [36]: Table_Configuration.html#Constraints
+   [37]: Table_Configuration.html#Bloom_Filters
+   [38]: Table_Configuration.html#Iterators
+   [39]: Table_Configuration.html#Setting_Iterators_via_the_Shell
+   [40]: Table_Configuration.html#Setting_Iterators_Programmatically
+   [41]: Table_Configuration.html#Versioning_Iterators_and_Timestamps
+   [42]: Table_Configuration.html#Filters
+   [43]: Table_Configuration.html#Combiners
+   [44]: Table_Configuration.html#Block_Cache
+   [45]: Table_Configuration.html#Compaction
+   [46]: Table_Configuration.html#Pre-splitting_tables
+   [47]: Table_Configuration.html#Merging_tablets
+   [48]: Table_Configuration.html#Delete_Range
+   [49]: Table_Configuration.html#Cloning_Tables
+   [50]: Table_Design.html
+   [51]: Table_Design.html#Basic_Table
+   [52]: Table_Design.html#RowID_Design
+   [53]: Table_Design.html#Indexing
+   [54]: Table_Design.html#Entity-Attribute_and_Graph_Tables
+   [55]: Table_Design.html#Document-Partitioned_Indexing
+   [56]: High_Speed_Ingest.html
+   [57]: High_Speed_Ingest.html#Pre-Splitting_New_Tables
+   [58]: High_Speed_Ingest.html#Multiple_Ingester_Clients
+   [59]: High_Speed_Ingest.html#Bulk_Ingest
+   [60]: High_Speed_Ingest.html#Logical_Time_for_Bulk_Ingest
+   [61]: High_Speed_Ingest.html#MapReduce_Ingest
+   [62]: Analytics.html
+   [63]: Analytics.html#MapReduce
+   [64]: Analytics.html#Mapper_and_Reducer_classes
+   [65]: Analytics.html#AccumuloInputFormat_options
+   [66]: Analytics.html#AccumuloOutputFormat_options
+   [67]: Analytics.html#Combiners
+   [68]: Analytics.html#Feature_Vectors
+   [69]: Analytics.html#Statistical_Modeling
+   [70]: Security.html
+   [71]: Security.html#Security_Label_Expressions
+   [72]: Security.html#Security_Label_Expression_Syntax
+   [73]: Security.html#Authorization
+   [74]: Security.html#User_Authorizations
+   [75]: Security.html#Secure_Authorizations_Handling
+   [76]: Security.html#Query_Services_Layer
+   [77]: Administration.html
+   [78]: Administration.html#Hardware
+   [79]: Administration.html#Network
+   [80]: Administration.html#Installation
+   [81]: Administration.html#Dependencies
+   [82]: Administration.html#Configuration
+   [83]: Administration.html#Edit_conf/accumulo-env.sh
+   [84]: Administration.html#Cluster_Specification
+   [85]: Administration.html#Accumulo_Settings
+   [86]: Administration.html#Deploy_Configuration
+   [87]: Administration.html#Initialization
+   [88]: Administration.html#Running
+   [89]: Administration.html#Starting_Accumulo
+   [90]: Administration.html#Stopping_Accumulo
+   [91]: Administration.html#Monitoring
+   [92]: Administration.html#Logging
+   [93]: Administration.html#Recovery
+   [94]: Shell_Commands.html
+

Propchange: incubator/accumulo/site/trunk/content/accumulo/1.4/user_manual/Contents.mdtext
------------------------------------------------------------------------------
    svn:eol-style = native

Added: incubator/accumulo/site/trunk/content/accumulo/1.4/user_manual/High_Speed_Ingest.mdtext
URL: http://svn.apache.org/viewvc/incubator/accumulo/site/trunk/content/accumulo/1.4/user_manual/High_Speed_Ingest.mdtext?rev=1304563&view=auto
==============================================================================
--- incubator/accumulo/site/trunk/content/accumulo/1.4/user_manual/High_Speed_Ingest.mdtext (added)
+++ incubator/accumulo/site/trunk/content/accumulo/1.4/user_manual/High_Speed_Ingest.mdtext Fri Mar 23 19:07:51 2012
@@ -0,0 +1,107 @@
+Title: Apache Accumulo User Manual: High Speed Ingest
+Notice:    Licensed to the Apache Software Foundation (ASF) under one
+           or more contributor license agreements.  See the NOTICE file
+           distributed with this work for additional information
+           regarding copyright ownership.  The ASF licenses this file
+           to you under the Apache License, Version 2.0 (the
+           "License"); you may not use this file except in compliance
+           with the License.  You may obtain a copy of the License at
+           .
+             http://www.apache.org/licenses/LICENSE-2.0
+           .
+           Unless required by applicable law or agreed to in writing,
+           software distributed under the License is distributed on an
+           "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+           KIND, either express or implied.  See the License for the
+           specific language governing permissions and limitations
+           under the License.
+
+** Next:** [Analytics][2] ** Up:** [Apache Accumulo User Manual Version 1.4][4] ** Previous:** [Table Design][6]   ** [Contents][8]**   
+  
+<a id=CHILD_LINKS></a>**Subsections**
+
+* [Pre-Splitting New Tables][9]
+* [Multiple Ingester Clients][10]
+* [Bulk Ingest][11]
+* [Logical Time for Bulk Ingest][12]
+* [MapReduce Ingest][13]
+
+* * *
+
+## <a id=High-Speed_Ingest></a> High-Speed Ingest
+
+Accumulo is often used as part of a larger data processing and storage system. To maximize the performance of a parallel system involving Accumulo, the ingestion and query components should be designed to provide enough parallelism and concurrency to avoid creating bottlenecks for users and other systems writing to and reading from Accumulo. There are several ways to achieve high ingest performance. 
+
+## <a id=Pre-Splitting_New_Tables></a> Pre-Splitting New Tables
+
+New tables consist of a single tablet by default. As mutations are applied, the table grows and splits into multiple tablets which are balanced by the Master across TabletServers. This implies that the aggregate ingest rate will be limited to fewer servers than are available within the cluster until the table has reached the point where there are tablets on every TabletServer. 
+
+Pre-splitting a table ensures that there are as many tablets as desired available before ingest begins to take advantage of all the parallelism possible with the cluster hardware. Tables can be split anytime by using the shell: 
+    
+    
+    user@myinstance mytable> addsplits -sf /local_splitfile -t mytable
+    
+
+For the purposes of providing parallelism to ingest it is not necessary to create more tablets than there are physical machines within the cluster as the aggregate ingest rate is a function of the number of physical machines. Note that the aggregate ingest rate is still subject to the number of machines running ingest clients, and the distribution of rowIDs across the table. The aggregation ingest rate will be suboptimal if there are many inserts into a small number of rowIDs. 
+
+## <a id=Multiple_Ingester_Clients></a> Multiple Ingester Clients
+
+Accumulo is capable of scaling to very high rates of ingest, which is dependent upon not just the number of TabletServers in operation but also the number of ingest clients. This is because a single client, while capable of batching mutations and sending them to all TabletServers, is ultimately limited by the amount of data that can be processed on a single machine. The aggregate ingest rate will scale linearly with the number of clients up to the point at which either the aggregate I/O of TabletServers or total network bandwidth capacity is reached. 
+
+In operational settings where high rates of ingest are paramount, clusters are often configured to dedicate some number of machines solely to running Ingester Clients. The exact ratio of clients to TabletServers necessary for optimum ingestion rates will vary according to the distribution of resources per machine and by data type. 
+
+## <a id=Bulk_Ingest></a> Bulk Ingest
+
+Accumulo supports the ability to import files produced by an external process such as MapReduce into an existing table. In some cases it may be faster to load data this way rather than via ingesting through clients using BatchWriters. This allows a large number of machines to format data the way Accumulo expects. The new files can then simply be introduced to Accumulo via a shell command. 
+
+To configure MapReduce to format data in preparation for bulk loading, the job should be set to use a range partitioner instead of the default hash partitioner. The range partitioner uses the split points of the Accumulo table that will receive the data. The split points can be obtained from the shell and used by the MapReduce RangePartitioner. Note that this is only useful if the existing table is already split into multiple tablets. 
+    
+    
+    user@myinstance mytable> getsplits
+    aa
+    ab
+    ac
+    ...
+    zx
+    zy
+    zz
+    
+
+Run the MapReduce job, using the AccumuloFileOutputFormat to create the files to be introduced to Accumulo. Once this is complete, the files can be added to Accumulo via the shell: 
+    
+    
+    user@myinstance mytable> importdirectory /files_dir /failures
+    
+
+Note that the paths referenced are directories within the same HDFS instance over which Accumulo is running. Accumulo places any files that failed to be added to the second directory specified. 
+
+A complete example of using Bulk Ingest can be found at   
+accumulo/docs/examples/README.bulkIngest 
+
+## <a id=Logical_Time_for_Bulk_Ingest></a> Logical Time for Bulk Ingest
+
+Logical time is important for bulk imported data, for which the client code may be choosing a timestamp. At bulk import time, the user can choose to enable logical time for the set of files being imported. When its enabled, Accumulo uses a specialized system iterator to lazily set times in a bulk imported file. This mechanism guarantees that times set by unsynchronized multi-node applications (such as those running on MapReduce) will maintain some semblance of causal ordering. This mitigates the problem of the time being wrong on the system that created the file for bulk import. These times are not set when the file is imported, but whenever it is read by scans or compactions. At import, a time is obtained and always used by the specialized system iterator to set that time. 
+
+The timestamp assigned by accumulo will be the same for every key in the file. This could cause problems if the file contains multiple keys that are identical except for the timestamp. In this case, the sort order of the keys will be undefined. This could occur if an insert and an update were in the same bulk import file. 
+
+## <a id=MapReduce_Ingest></a> MapReduce Ingest
+
+It is possible to efficiently write many mutations to Accumulo in parallel via a MapReduce job. In this scenario the MapReduce is written to process data that lives in HDFS and write mutations to Accumulo using the AccumuloOutputFormat. See the MapReduce section under Analytics for details. 
+
+An example of using MapReduce can be found under   
+accumulo/docs/examples/README.mapred 
+
+* * *
+
+** Next:** [Analytics][2] ** Up:** [Apache Accumulo User Manual Version 1.4][4] ** Previous:** [Table Design][6]   ** [Contents][8]**
+
+   [2]: Analytics.html
+   [4]: accumulo_user_manual.html
+   [6]: Table_Design.html
+   [8]: Contents.html
+   [9]: High_Speed_Ingest.html#Pre-Splitting_New_Tables
+   [10]: High_Speed_Ingest.html#Multiple_Ingester_Clients
+   [11]: High_Speed_Ingest.html#Bulk_Ingest
+   [12]: High_Speed_Ingest.html#Logical_Time_for_Bulk_Ingest
+   [13]: High_Speed_Ingest.html#MapReduce_Ingest
+

Propchange: incubator/accumulo/site/trunk/content/accumulo/1.4/user_manual/High_Speed_Ingest.mdtext
------------------------------------------------------------------------------
    svn:eol-style = native

Added: incubator/accumulo/site/trunk/content/accumulo/1.4/user_manual/Introduction.mdtext
URL: http://svn.apache.org/viewvc/incubator/accumulo/site/trunk/content/accumulo/1.4/user_manual/Introduction.mdtext?rev=1304563&view=auto
==============================================================================
--- incubator/accumulo/site/trunk/content/accumulo/1.4/user_manual/Introduction.mdtext (added)
+++ incubator/accumulo/site/trunk/content/accumulo/1.4/user_manual/Introduction.mdtext Fri Mar 23 19:07:51 2012
@@ -0,0 +1,37 @@
+Title: Apache Accumulo User Manual: Introduction
+Notice:    Licensed to the Apache Software Foundation (ASF) under one
+           or more contributor license agreements.  See the NOTICE file
+           distributed with this work for additional information
+           regarding copyright ownership.  The ASF licenses this file
+           to you under the Apache License, Version 2.0 (the
+           "License"); you may not use this file except in compliance
+           with the License.  You may obtain a copy of the License at
+           .
+             http://www.apache.org/licenses/LICENSE-2.0
+           .
+           Unless required by applicable law or agreed to in writing,
+           software distributed under the License is distributed on an
+           "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+           KIND, either express or implied.  See the License for the
+           specific language governing permissions and limitations
+           under the License.
+
+** Next:** [Accumulo Design][2] ** Up:** [Apache Accumulo User Manual Version 1.4][4] ** Previous:** [Contents][6]   ** [Contents][6]**   
+  
+
+
+## <a id=Introduction></a> Introduction
+
+Apache Accumulo is a highly scalable structured store based on Google's BigTable. Accumulo is written in Java and operates over the Hadoop Distributed File System (HDFS), which is part of the popular Apache Hadoop project. Accumulo supports efficient storage and retrieval of structured data, including queries for ranges, and provides support for using Accumulo tables as input and output for MapReduce jobs. 
+
+Accumulo features automatic load-balancing and partitioning, data compression and fine-grained security labels. 
+
+  
+
+
+* * *
+
+   [2]: Accumulo_Design.html
+   [4]: accumulo_user_manual.html
+   [6]: Contents.html
+

Propchange: incubator/accumulo/site/trunk/content/accumulo/1.4/user_manual/Introduction.mdtext
------------------------------------------------------------------------------
    svn:eol-style = native

Added: incubator/accumulo/site/trunk/content/accumulo/1.4/user_manual/Security.mdtext
URL: http://svn.apache.org/viewvc/incubator/accumulo/site/trunk/content/accumulo/1.4/user_manual/Security.mdtext?rev=1304563&view=auto
==============================================================================
--- incubator/accumulo/site/trunk/content/accumulo/1.4/user_manual/Security.mdtext (added)
+++ incubator/accumulo/site/trunk/content/accumulo/1.4/user_manual/Security.mdtext Fri Mar 23 19:07:51 2012
@@ -0,0 +1,138 @@
+Title: Apache Accumulo User Manual: Security
+Notice:    Licensed to the Apache Software Foundation (ASF) under one
+           or more contributor license agreements.  See the NOTICE file
+           distributed with this work for additional information
+           regarding copyright ownership.  The ASF licenses this file
+           to you under the Apache License, Version 2.0 (the
+           "License"); you may not use this file except in compliance
+           with the License.  You may obtain a copy of the License at
+           .
+             http://www.apache.org/licenses/LICENSE-2.0
+           .
+           Unless required by applicable law or agreed to in writing,
+           software distributed under the License is distributed on an
+           "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+           KIND, either express or implied.  See the License for the
+           specific language governing permissions and limitations
+           under the License.
+
+** Next:** [Administration][2] ** Up:** [Apache Accumulo User Manual Version 1.4][4] ** Previous:** [Analytics][6]   ** [Contents][8]**   
+  
+<a id=CHILD_LINKS></a>**Subsections**
+
+* [Security Label Expressions][9]
+* [Security Label Expression Syntax][10]
+* [Authorization][11]
+* [User Authorizations][12]
+* [Secure Authorizations Handling][13]
+* [Query Services Layer][14]
+
+* * *
+
+## <a id=Security></a> Security
+
+Accumulo extends the BigTable data model to implement a security mechanism known as cell-level security. Every key-value pair has its own security label, stored under the column visibility element of the key, which is used to determine whether a given user meets the security requirements to read the value. This enables data of various security levels to be stored within the same row, and users of varying degrees of access to query the same table, while preserving data confidentiality. 
+
+## <a id=Security_Label_Expressions></a> Security Label Expressions
+
+When mutations are applied, users can specify a security label for each value. This is done as the Mutation is created by passing a ColumnVisibility object to the put() method: 
+    
+    
+    Text rowID = new Text("row1");
+    Text colFam = new Text("myColFam");
+    Text colQual = new Text("myColQual");
+    ColumnVisibility colVis = new ColumnVisibility("public");
+    long timestamp = System.currentTimeMillis();
+    
+    Value value = new Value("myValue");
+    
+    Mutation mutation = new Mutation(rowID);
+    mutation.put(colFam, colQual, colVis, timestamp, value);
+    
+
+## <a id=Security_Label_Expression_Syntax></a> Security Label Expression Syntax
+
+Security labels consist of a set of user-defined tokens that are required to read the value the label is associated with. The set of tokens required can be specified using syntax that supports logical AND and OR combinations of tokens, as well as nesting groups of tokens together. 
+
+For example, suppose within our organization we want to label our data values with security labels defined in terms of user roles. We might have tokens such as: 
+    
+    
+    admin
+    audit
+    system
+    
+
+These can be specified alone or combined using logical operators: 
+    
+    
+    // Users must have admin privileges:
+    admin
+    
+    // Users must have admin and audit privileges
+    admin&audit
+    
+    // Users with either admin or audit privileges
+    admin|audit
+    
+    // Users must have audit and one or both of admin or system
+    (admin|system)&audit
+    
+
+When both `|` and `&` operators are used, parentheses must be used to specify precedence of the operators. 
+
+## <a id=Authorization></a> Authorization
+
+When clients attempt to read data from Accumulo, any security labels present are examined against the set of authorizations passed by the client code when the Scanner or BatchScanner are created. If the authorizations are determined to be insufficient to satisfy the security label, the value is suppressed from the set of results sent back to the client. 
+
+Authorizations are specified as a comma-separated list of tokens the user possesses: 
+    
+    
+    // user possess both admin and system level access
+    Authorization auths = new Authorization("admin","system");
+    
+    Scanner s = connector.createScanner("table", auths);
+    
+
+## <a id=User_Authorizations></a> User Authorizations
+
+Each accumulo user has a set of associated security labels. To manipulate these in the shell use the setuaths and getauths commands. These may also be modified using the java security operations API. 
+
+When a user creates a scanner a set of Authorizations is passed. If the authorizations passed to the scanner are not a subset of the users authorizations, then an exception will be thrown. 
+
+To prevent users from writing data they can not read, add the visibility constraint to a table. Use the -evc option in the createtable shell command to enable this constraint. For existing tables use the following shell command to enable the visibility constraint. Ensure the constraint number does not conflict with any existing constraints. 
+    
+    
+    config -t table -s table.constraint.1=org.apache.accumulo.core.security.VisibilityConstraint
+    
+
+Any user with the alter table permission can add or remove this constraint. This constraint is not applied to bulk imported data, if this a concern then disable the bulk import permission. 
+
+## <a id=Secure_Authorizations_Handling></a> Secure Authorizations Handling
+
+For applications serving many users, it is not expected that an accumulo user will be created for each application user. In this case an accumulo user with all authorizations needed by any of the applications users must be created. To service queries, the application should create a scanner with the application users authorizations. These authorizations could be obtained from a trusted 3rd party. 
+
+Often production systems will integrate with Public-Key Infrastructure (PKI) and designate client code within the query layer to negotiate with PKI servers in order to authenticate users and retrieve their authorization tokens (credentials). This requires users to specify only the information necessary to authenticate themselves to the system. Once user identity is established, their credentials can be accessed by the client code and passed to Accumulo outside of the reach of the user. 
+
+## <a id=Query_Services_Layer></a> Query Services Layer
+
+Since the primary method of interaction with Accumulo is through the Java API, production environments often call for the implementation of a Query layer. This can be done using web services in containers such as Apache Tomcat, but is not a requirement. The Query Services Layer provides a mechanism for providing a platform on which user facing applications can be built. This allows the application designers to isolate potentially complex query logic, and enables a convenient point at which to perform essential security functions. 
+
+Several production environments choose to implement authentication at this layer, where users identifiers are used to retrieve their access credentials which are then cached within the query layer and presented to Accumulo through the Authorizations mechanism. 
+
+Typically, the query services layer sits between Accumulo and user workstations. 
+
+* * *
+
+** Next:** [Administration][2] ** Up:** [Apache Accumulo User Manual Version 1.4][4] ** Previous:** [Analytics][6]   ** [Contents][8]**
+
+   [2]: Administration.html
+   [4]: accumulo_user_manual.html
+   [6]: Analytics.html
+   [8]: Contents.html
+   [9]: Security.html#Security_Label_Expressions
+   [10]: Security.html#Security_Label_Expression_Syntax
+   [11]: Security.html#Authorization
+   [12]: Security.html#User_Authorizations
+   [13]: Security.html#Secure_Authorizations_Handling
+   [14]: Security.html#Query_Services_Layer
+

Propchange: incubator/accumulo/site/trunk/content/accumulo/1.4/user_manual/Security.mdtext
------------------------------------------------------------------------------
    svn:eol-style = native



Mime
View raw message