accumulo-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From mwa...@apache.org
Subject [6/7] accumulo git commit: ACCUMULO-4511 Removed Accumulo Examples
Date Fri, 09 Dec 2016 17:16:52 GMT
http://git-wip-us.apache.org/repos/asf/accumulo/blob/8e0f19a1/docs/src/main/resources/examples/filter.md
----------------------------------------------------------------------
diff --git a/docs/src/main/resources/examples/filter.md b/docs/src/main/resources/examples/filter.md
deleted file mode 100644
index 563e247..0000000
--- a/docs/src/main/resources/examples/filter.md
+++ /dev/null
@@ -1,112 +0,0 @@
-<!--
-Licensed to the Apache Software Foundation (ASF) under one or more
-contributor license agreements.  See the NOTICE file distributed with
-this work for additional information regarding copyright ownership.
-The ASF licenses this file to You under the Apache License, Version 2.0
-(the "License"); you may not use this file except in compliance with
-the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
--->
----
-title: Apache Accumulo Filter Example
----
-
-This is a simple filter example. It uses the AgeOffFilter that is provided as
-part of the core package org.apache.accumulo.core.iterators.user. Filters are
-iterators that select desired key/value pairs (or weed out undesired ones).
-Filters extend the org.apache.accumulo.core.iterators.Filter class
-and must implement a method accept(Key k, Value v). This method returns true
-if the key/value pair are to be delivered and false if they are to be ignored.
-Filter takes a "negate" parameter which defaults to false. If set to true, the
-return value of the accept method is negated, so that key/value pairs accepted
-by the method are omitted by the Filter.
-
-    username@instance> createtable filtertest
-    username@instance filtertest> setiter -t filtertest -scan -p 10 -n myfilter -ageoff
-    AgeOffFilter removes entries with timestamps more than <ttl> milliseconds old
-    ----------> set AgeOffFilter parameter negate, default false keeps k/v that pass accept method, true rejects k/v that pass accept method:
-    ----------> set AgeOffFilter parameter ttl, time to live (milliseconds): 30000
-    ----------> set AgeOffFilter parameter currentTime, if set, use the given value as the absolute time in milliseconds as the current time of day:
-    username@instance filtertest> scan
-    username@instance filtertest> insert foo a b c
-    username@instance filtertest> scan
-    foo a:b []    c
-    username@instance filtertest>
-
-... wait 30 seconds ...
-
-    username@instance filtertest> scan
-    username@instance filtertest>
-
-Note the absence of the entry inserted more than 30 seconds ago. Since the
-scope was set to "scan", this means the entry is still in Accumulo, but is
-being filtered out at query time. To delete entries from Accumulo based on
-the ages of their timestamps, AgeOffFilters should be set up for the "minc"
-and "majc" scopes, as well.
-
-To force an ageoff of the persisted data, after setting up the ageoff iterator
-on the "minc" and "majc" scopes you can flush and compact your table. This will
-happen automatically as a background operation on any table that is being
-actively written to, but can also be requested in the shell.
-
-The first setiter command used the special -ageoff flag to specify the
-AgeOffFilter, but any Filter can be configured by using the -class flag. The
-following commands show how to enable the AgeOffFilter for the minc and majc
-scopes using the -class flag, then flush and compact the table.
-
-    username@instance filtertest> setiter -t filtertest -minc -majc -p 10 -n myfilter -class org.apache.accumulo.core.iterators.user.AgeOffFilter
-    AgeOffFilter removes entries with timestamps more than <ttl> milliseconds old
-    ----------> set AgeOffFilter parameter negate, default false keeps k/v that pass accept method, true rejects k/v that pass accept method:
-    ----------> set AgeOffFilter parameter ttl, time to live (milliseconds): 30000
-    ----------> set AgeOffFilter parameter currentTime, if set, use the given value as the absolute time in milliseconds as the current time of day:
-    username@instance filtertest> flush
-    06 10:42:24,806 [shell.Shell] INFO : Flush of table filtertest initiated...
-    username@instance filtertest> compact
-    06 10:42:36,781 [shell.Shell] INFO : Compaction of table filtertest started for given range
-    username@instance filtertest> flush -t filtertest -w
-    06 10:42:52,881 [shell.Shell] INFO : Flush of table filtertest completed.
-    username@instance filtertest> compact -t filtertest -w
-    06 10:43:00,632 [shell.Shell] INFO : Compacting table ...
-    06 10:43:01,307 [shell.Shell] INFO : Compaction of table filtertest completed for given range
-    username@instance filtertest>
-
-By default, flush and compact execute in the background, but with the -w flag
-they will wait to return until the operation has completed. Both are
-demonstrated above, though only one call to each would be necessary. A
-specific table can be specified with -t.
-
-After the compaction runs, the newly created files will not contain any data
-that should have been aged off, and the Accumulo garbage collector will remove
-the old files.
-
-To see the iterator settings for a table, use config.
-
-    username@instance filtertest> config -t filtertest -f iterator
-    ---------+---------------------------------------------+---------------------------------------------------------------------------
-    SCOPE    | NAME                                        | VALUE
-    ---------+---------------------------------------------+---------------------------------------------------------------------------
-    table    | table.iterator.majc.myfilter .............. | 10,org.apache.accumulo.core.iterators.user.AgeOffFilter
-    table    | table.iterator.majc.myfilter.opt.ttl ...... | 30000
-    table    | table.iterator.majc.vers .................. | 20,org.apache.accumulo.core.iterators.user.VersioningIterator
-    table    | table.iterator.majc.vers.opt.maxVersions .. | 1
-    table    | table.iterator.minc.myfilter .............. | 10,org.apache.accumulo.core.iterators.user.AgeOffFilter
-    table    | table.iterator.minc.myfilter.opt.ttl ...... | 30000
-    table    | table.iterator.minc.vers .................. | 20,org.apache.accumulo.core.iterators.user.VersioningIterator
-    table    | table.iterator.minc.vers.opt.maxVersions .. | 1
-    table    | table.iterator.scan.myfilter .............. | 10,org.apache.accumulo.core.iterators.user.AgeOffFilter
-    table    | table.iterator.scan.myfilter.opt.ttl ...... | 30000
-    table    | table.iterator.scan.vers .................. | 20,org.apache.accumulo.core.iterators.user.VersioningIterator
-    table    | table.iterator.scan.vers.opt.maxVersions .. | 1
-    ---------+---------------------------------------------+---------------------------------------------------------------------------
-    username@instance filtertest>
-
-When setting new iterators, make sure to order their priority numbers
-(specified with -p) in the order you would like the iterators to be applied.
-Also, each iterator must have a unique name and priority within each scope.

http://git-wip-us.apache.org/repos/asf/accumulo/blob/8e0f19a1/docs/src/main/resources/examples/helloworld.md
----------------------------------------------------------------------
diff --git a/docs/src/main/resources/examples/helloworld.md b/docs/src/main/resources/examples/helloworld.md
deleted file mode 100644
index bc9f04b..0000000
--- a/docs/src/main/resources/examples/helloworld.md
+++ /dev/null
@@ -1,49 +0,0 @@
-<!--
-Licensed to the Apache Software Foundation (ASF) under one or more
-contributor license agreements.  See the NOTICE file distributed with
-this work for additional information regarding copyright ownership.
-The ASF licenses this file to You under the Apache License, Version 2.0
-(the "License"); you may not use this file except in compliance with
-the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
--->
----
-title: Apache Accumulo Hello World Example
----
-
-This tutorial uses the following Java classes, which can be found in org.apache.accumulo.examples.simple.helloworld in the examples-simple module:
-
- * InsertWithBatchWriter.java - Inserts 10K rows (50K entries) into accumulo with each row having 5 entries
- * ReadData.java - Reads all data between two rows
-
-Log into the accumulo shell:
-
-    $ ./bin/accumulo shell -u username -p password
-
-Create a table called 'hellotable':
-
-    username@instance> createtable hellotable
-
-Launch a Java program that inserts data with a BatchWriter:
-
-    $ ./bin/accumulo org.apache.accumulo.examples.simple.helloworld.InsertWithBatchWriter -i instance -z zookeepers -u username -p password -t hellotable
-
-On the accumulo status page at the URL below (where 'master' is replaced with the name or IP of your accumulo master), you should see 50K entries
-
-    http://master:9995/
-
-To view the entries, use the shell to scan the table:
-
-    username@instance> table hellotable
-    username@instance hellotable> scan
-
-You can also use a Java class to scan the table:
-
-    $ ./bin/accumulo org.apache.accumulo.examples.simple.helloworld.ReadData -i instance -z zookeepers -u username -p password -t hellotable --startKey row_0 --endKey row_1001

http://git-wip-us.apache.org/repos/asf/accumulo/blob/8e0f19a1/docs/src/main/resources/examples/index.md
----------------------------------------------------------------------
diff --git a/docs/src/main/resources/examples/index.md b/docs/src/main/resources/examples/index.md
deleted file mode 100644
index efb55f6..0000000
--- a/docs/src/main/resources/examples/index.md
+++ /dev/null
@@ -1,100 +0,0 @@
-<!--
-Licensed to the Apache Software Foundation (ASF) under one or more
-contributor license agreements.  See the NOTICE file distributed with
-this work for additional information regarding copyright ownership.
-The ASF licenses this file to You under the Apache License, Version 2.0
-(the "License"); you may not use this file except in compliance with
-the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
--->
----
-title: Apache Accumulo Examples
----
-
-## Setup instructions
-
-Before running any of the examples, the following steps must be performed.
-
-1. Install and run Accumulo via the instructions found in INSTALL.md.
-   Remember the instance name. It will be referred to as "instance" throughout
-   the examples. A comma-separated list of zookeeper servers will be referred
-   to as "zookeepers".
-
-2. Create an Accumulo user (for help see the 'User Administration' section of the 
-   [user manual][manual]), or use the root user. This user and their password
-   should replace any reference to "username" or "password" in the examples. This
-   user needs the ability to create tables.
-
-In all commands, you will need to replace "instance", "zookeepers",
-"username", and "password" with the values you set for your Accumulo instance.
-
-Commands intended to be run in bash are prefixed by '$'. These are always
-assumed to be run the from the root of your Accumulo installation.
-
-Commands intended to be run in the Accumulo shell are prefixed by '>'.
-
-## Accumulo Examples
-
-Each example below highlights a feature of Apache Accumulo.
-
-| Accumulo Example | Description |
-|------------------|-------------|
-| [batch] | Using the batch writer and batch scanner |
-| [bloom] | Creating a bloom filter enabled table to increase query performance |
-| [bulkIngest] | Ingesting bulk data using map/reduce jobs on Hadoop |
-| [classpath] | Using per-table classpaths |
-| [client] | Using table operations, reading and writing data in Java. |
-| [combiner] | Using example StatsCombiner to find min, max, sum, and count. |
-| [compactionStrategy] | Configuring a compaction strategy |
-| [constraints] | Using constraints with tables. |
-| [dirlist] | Storing filesystem information. |
-| [export] | Exporting and importing tables. |
-| [filedata] | Storing file data. |
-| [filter] | Using the AgeOffFilter to remove records more than 30 seconds old. |
-| [helloworld] | Inserting records both inside map/reduce jobs and outside. And reading records between two rows. |
-| [isolation] | Using the isolated scanner to ensure partial changes are not seen. |
-| [mapred] | Using MapReduce to read from and write to Accumulo tables. |
-| [maxmutation] | Limiting mutation size to avoid running out of memory. |
-| [regex] | Using MapReduce and Accumulo to find data using regular expressions. |
-| [reservations] | Using conditional mutations to implement simple reservation system. |
-| [rgbalancer] | Using a balancer to spread groups of tablets within a table evenly |
-| [rowhash] | Using MapReduce to read a table and write to a new column in the same table. |
-| [sample] | Building and using sample data in Accumulo. |
-| [shard] | Using the intersecting iterator with a term index partitioned by document. |
-| [tabletofile] | Using MapReduce to read a table and write one of its columns to a file in HDFS. |
-| [terasort] | Generating random data and sorting it using Accumulo. |
-| [visibility] | Using visibilities (or combinations of authorizations). Also shows user permissions. |
-
-[manual]: https://accumulo.apache.org/latest/accumulo_user_manual/
-[batch]: batch.md
-[bloom]: bloom.md
-[bulkIngest]: bulkIngest.md
-[classpath]: classpath.md
-[client]: client.md 
-[combiner]: combiner.md
-[compactionStrategy]: compactionStrategy.md
-[constraints]: constraints.md
-[dirlist]: dirlist.md
-[export]: export.md
-[filedata]: filedata.md
-[filter]: filter.md
-[helloworld]: helloworld.md
-[isolation]: isolation.md
-[mapred]: mapred.md
-[maxmutation]: maxmutation.md
-[regex]: regex.md
-[reservations]: reservations.md
-[rgbalancer]: rgbalancer.md
-[rowhash]: rowhash.md
-[sample]: sample.md
-[shard]: shard.md
-[tabletofile]: tabletofile.md
-[terasort]: terasort.md
-[visibility]: visibility.md

http://git-wip-us.apache.org/repos/asf/accumulo/blob/8e0f19a1/docs/src/main/resources/examples/isolation.md
----------------------------------------------------------------------
diff --git a/docs/src/main/resources/examples/isolation.md b/docs/src/main/resources/examples/isolation.md
deleted file mode 100644
index 9b4e0af..0000000
--- a/docs/src/main/resources/examples/isolation.md
+++ /dev/null
@@ -1,51 +0,0 @@
-<!--
-Licensed to the Apache Software Foundation (ASF) under one or more
-contributor license agreements.  See the NOTICE file distributed with
-this work for additional information regarding copyright ownership.
-The ASF licenses this file to You under the Apache License, Version 2.0
-(the "License"); you may not use this file except in compliance with
-the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
--->
----
-title: Apache Accumulo Isolation Example
----
-
-Accumulo has an isolated scanner that ensures partial changes to rows are not
-seen. Isolation is documented in ../docs/isolation.html and the user manual.
-
-InterferenceTest is a simple example that shows the effects of scanning with
-and without isolation. This program starts two threads. One threads
-continually upates all of the values in a row to be the same thing, but
-different from what it used to be. The other thread continually scans the
-table and checks that all values in a row are the same. Without isolation the
-scanning thread will sometimes see different values, which is the result of
-reading the row at the same time a mutation is changing the row.
-
-Below, Interference Test is run without isolation enabled for 5000 iterations
-and it reports problems.
-
-    $ ./bin/accumulo org.apache.accumulo.examples.simple.isolation.InterferenceTest -i instance -z zookeepers -u username -p password -t isotest --iterations 5000
-    ERROR Columns in row 053 had multiple values [53, 4553]
-    ERROR Columns in row 061 had multiple values [561, 61]
-    ERROR Columns in row 070 had multiple values [570, 1070]
-    ERROR Columns in row 079 had multiple values [1079, 1579]
-    ERROR Columns in row 088 had multiple values [2588, 1588]
-    ERROR Columns in row 106 had multiple values [2606, 3106]
-    ERROR Columns in row 115 had multiple values [4615, 3115]
-    finished
-
-Below, Interference Test is run with isolation enabled for 5000 iterations and
-it reports no problems.
-
-    $ ./bin/accumulo org.apache.accumulo.examples.simple.isolation.InterferenceTest -i instance -z zookeepers -u username -p password -t isotest --iterations 5000 --isolated
-    finished
-
-

http://git-wip-us.apache.org/repos/asf/accumulo/blob/8e0f19a1/docs/src/main/resources/examples/mapred.md
----------------------------------------------------------------------
diff --git a/docs/src/main/resources/examples/mapred.md b/docs/src/main/resources/examples/mapred.md
deleted file mode 100644
index e1a49eb..0000000
--- a/docs/src/main/resources/examples/mapred.md
+++ /dev/null
@@ -1,156 +0,0 @@
-<!--
-Licensed to the Apache Software Foundation (ASF) under one or more
-contributor license agreements.  See the NOTICE file distributed with
-this work for additional information regarding copyright ownership.
-The ASF licenses this file to You under the Apache License, Version 2.0
-(the "License"); you may not use this file except in compliance with
-the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
--->
----
-title: Apache Accumulo MapReduce Example
----
-
-This example uses mapreduce and accumulo to compute word counts for a set of
-documents. This is accomplished using a map-only mapreduce job and a
-accumulo table with combiners.
-
-To run this example you will need a directory in HDFS containing text files.
-The accumulo readme will be used to show how to run this example.
-
-    $ hadoop fs -copyFromLocal /path/to/accumulo/README.md /user/username/wc/Accumulo.README
-    $ hadoop fs -ls /user/username/wc
-    Found 1 items
-    -rw-r--r--   2 username supergroup       9359 2009-07-15 17:54 /user/username/wc/Accumulo.README
-
-The first part of running this example is to create a table with a combiner
-for the column family count.
-
-    $ ./bin/accumulo shell -u username -p password
-    Shell - Apache Accumulo Interactive Shell
-    - version: 1.5.0
-    - instance name: instance
-    - instance id: 00000000-0000-0000-0000-000000000000
-    -
-    - type 'help' for a list of available commands
-    -
-    username@instance> createtable wordCount
-    username@instance wordCount> setiter -class org.apache.accumulo.core.iterators.user.SummingCombiner -p 10 -t wordCount -majc -minc -scan
-    SummingCombiner interprets Values as Longs and adds them together. A variety of encodings (variable length, fixed length, or string) are available
-    ----------> set SummingCombiner parameter all, set to true to apply Combiner to every column, otherwise leave blank. if true, columns option will be ignored.: false
-    ----------> set SummingCombiner parameter columns, <col fam>[:<col qual>]{,<col fam>[:<col qual>]} escape non-alphanum chars using %<hex>.: count
-    ----------> set SummingCombiner parameter lossy, if true, failed decodes are ignored. Otherwise combiner will error on failed decodes (default false): <TRUE|FALSE>: false
-    ----------> set SummingCombiner parameter type, <VARLEN|FIXEDLEN|STRING|fullClassName>: STRING
-    username@instance wordCount> quit
-
-After creating the table, run the word count map reduce job.
-
-    $ ./contrib/tool.sh lib/accumulo-examples-simple.jar org.apache.accumulo.examples.simple.mapreduce.WordCount -i instance -z zookeepers  --input /user/username/wc -t wordCount -u username -p password
-
-    11/02/07 18:20:11 INFO input.FileInputFormat: Total input paths to process : 1
-    11/02/07 18:20:12 INFO mapred.JobClient: Running job: job_201102071740_0003
-    11/02/07 18:20:13 INFO mapred.JobClient:  map 0% reduce 0%
-    11/02/07 18:20:20 INFO mapred.JobClient:  map 100% reduce 0%
-    11/02/07 18:20:22 INFO mapred.JobClient: Job complete: job_201102071740_0003
-    11/02/07 18:20:22 INFO mapred.JobClient: Counters: 6
-    11/02/07 18:20:22 INFO mapred.JobClient:   Job Counters
-    11/02/07 18:20:22 INFO mapred.JobClient:     Launched map tasks=1
-    11/02/07 18:20:22 INFO mapred.JobClient:     Data-local map tasks=1
-    11/02/07 18:20:22 INFO mapred.JobClient:   FileSystemCounters
-    11/02/07 18:20:22 INFO mapred.JobClient:     HDFS_BYTES_READ=10487
-    11/02/07 18:20:22 INFO mapred.JobClient:   Map-Reduce Framework
-    11/02/07 18:20:22 INFO mapred.JobClient:     Map input records=255
-    11/02/07 18:20:22 INFO mapred.JobClient:     Spilled Records=0
-    11/02/07 18:20:22 INFO mapred.JobClient:     Map output records=1452
-
-After the map reduce job completes, query the accumulo table to see word
-counts.
-
-    $ ./bin/accumulo shell -u username -p password
-    username@instance> table wordCount
-    username@instance wordCount> scan -b the
-    the count:20080906 []    75
-    their count:20080906 []    2
-    them count:20080906 []    1
-    then count:20080906 []    1
-    there count:20080906 []    1
-    these count:20080906 []    3
-    this count:20080906 []    6
-    through count:20080906 []    1
-    time count:20080906 []    3
-    time. count:20080906 []    1
-    to count:20080906 []    27
-    total count:20080906 []    1
-    tserver, count:20080906 []    1
-    tserver.compaction.major.concurrent.max count:20080906 []    1
-    ...
-
-Another example to look at is
-org.apache.accumulo.examples.simple.mapreduce.UniqueColumns. This example
-computes the unique set of columns in a table and shows how a map reduce job
-can directly read a tables files from HDFS.
-
-One more example available is
-org.apache.accumulo.examples.simple.mapreduce.TokenFileWordCount.
-The TokenFileWordCount example works exactly the same as the WordCount example
-explained above except that it uses a token file rather than giving the
-password directly to the map-reduce job (this avoids having the password
-displayed in the job's configuration which is world-readable).
-
-To create a token file, use the create-token utility
-
-  $ ./bin/accumulo create-token
-
-It defaults to creating a PasswordToken, but you can specify the token class
-with -tc (requires the fully qualified class name). Based on the token class,
-it will prompt you for each property required to create the token.
-
-The last value it prompts for is a local filename to save to. If this file
-exists, it will append the new token to the end. Multiple tokens can exist in
-a file, but only the first one for each user will be recognized.
-
-Rather than waiting for the prompts, you can specify some options when calling
-create-token, for example
-
-  $ ./bin/accumulo create-token -u root -p secret -f root.pw
-
-would create a token file containing a PasswordToken for
-user 'root' with password 'secret' and saved to 'root.pw'
-
-This local file needs to be uploaded to hdfs to be used with the
-map-reduce job. For example, if the file were 'root.pw' in the local directory:
-
-  $ hadoop fs -put root.pw root.pw
-
-This would put 'root.pw' in the user's home directory in hdfs.
-
-Because the basic WordCount example uses Opts to parse its arguments
-(which extends ClientOnRequiredTable), you can use a token file with
-the basic WordCount example by calling the same command as explained above
-except replacing the password with the token file (rather than -p, use -tf).
-
-  $ ./contrib/tool.sh lib/accumulo-examples-simple.jar org.apache.accumulo.examples.simple.mapreduce.WordCount -i instance -z zookeepers  --input /user/username/wc -t wordCount -u username -tf tokenfile
-
-In the above examples, username was 'root' and tokenfile was 'root.pw'
-
-However, if you don't want to use the Opts class to parse arguments,
-the TokenFileWordCount is an example of using the token file manually.
-
-  $ ./contrib/tool.sh lib/accumulo-examples-simple.jar org.apache.accumulo.examples.simple.mapreduce.TokenFileWordCount instance zookeepers username tokenfile /user/username/wc wordCount
-
-The results should be the same as the WordCount example except that the
-authentication token was not stored in the configuration. It was instead
-stored in a file that the map-reduce job pulled into the distributed cache.
-(If you ran either of these on the same table right after the
-WordCount example, then the resulting counts should just double.)
-
-
-
-

http://git-wip-us.apache.org/repos/asf/accumulo/blob/8e0f19a1/docs/src/main/resources/examples/maxmutation.md
----------------------------------------------------------------------
diff --git a/docs/src/main/resources/examples/maxmutation.md b/docs/src/main/resources/examples/maxmutation.md
deleted file mode 100644
index 48c918a..0000000
--- a/docs/src/main/resources/examples/maxmutation.md
+++ /dev/null
@@ -1,51 +0,0 @@
-<!--
-Licensed to the Apache Software Foundation (ASF) under one or more
-contributor license agreements.  See the NOTICE file distributed with
-this work for additional information regarding copyright ownership.
-The ASF licenses this file to You under the Apache License, Version 2.0
-(the "License"); you may not use this file except in compliance with
-the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
--->
----
-title: Apache Accumulo MaxMutation Constraints Example
----
-
-This an example of how to limit the size of mutations that will be accepted into
-a table. Under the default configuration, accumulo does not provide a limitation
-on the size of mutations that can be ingested. Poorly behaved writers might
-inadvertently create mutations so large, that they cause the tablet servers to
-run out of memory. A simple contraint can be added to a table to reject very
-large mutations.
-
-    $ ./bin/accumulo shell -u username -p password
-
-    Shell - Apache Accumulo Interactive Shell
-    -
-    - version: 1.5.0
-    - instance name: instance
-    - instance id: 00000000-0000-0000-0000-000000000000
-    -
-    - type 'help' for a list of available commands
-    -
-    username@instance> createtable test_ingest
-    username@instance test_ingest> config -t test_ingest -s table.constraint.1=org.apache.accumulo.examples.simple.constraints.MaxMutationSize
-    username@instance test_ingest>
-
-
-Now the table will reject any mutation that is larger than 1/256th of the 
-working memory of the tablet server.  The following command attempts to ingest 
-a single row with 10000 columns, which exceeds the memory limit. Depending on the
-amount of Java heap your tserver(s) are given, you may have to increase the number
-of columns provided to see the failure.
-
-    $ ./bin/accumulo org.apache.accumulo.test.TestIngest -i instance -z zookeepers -u username -p password --rows 1 --cols 10000 
-    ERROR : Constraint violates : ConstraintViolationSummary(constrainClass:org.apache.accumulo.examples.simple.constraints.MaxMutationSize, violationCode:0, violationDescription:mutation exceeded maximum size of 188160, numberOfViolatingMutations:1)
-

http://git-wip-us.apache.org/repos/asf/accumulo/blob/8e0f19a1/docs/src/main/resources/examples/regex.md
----------------------------------------------------------------------
diff --git a/docs/src/main/resources/examples/regex.md b/docs/src/main/resources/examples/regex.md
deleted file mode 100644
index 29d47e1..0000000
--- a/docs/src/main/resources/examples/regex.md
+++ /dev/null
@@ -1,59 +0,0 @@
-<!--
-Licensed to the Apache Software Foundation (ASF) under one or more
-contributor license agreements.  See the NOTICE file distributed with
-this work for additional information regarding copyright ownership.
-The ASF licenses this file to You under the Apache License, Version 2.0
-(the "License"); you may not use this file except in compliance with
-the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
--->
----
-title: Apache Accumulo Regex Example
----
-
-This example uses mapreduce and accumulo to find items using regular expressions.
-This is accomplished using a map-only mapreduce job and a scan-time iterator.
-
-To run this example you will need some data in a table. The following will
-put a trivial amount of data into accumulo using the accumulo shell:
-
-    $ ./bin/accumulo shell -u username -p password
-    Shell - Apache Accumulo Interactive Shell
-    - version: 1.5.0
-    - instance name: instance
-    - instance id: 00000000-0000-0000-0000-000000000000
-    -
-    - type 'help' for a list of available commands
-    -
-    username@instance> createtable input
-    username@instance> insert dogrow dogcf dogcq dogvalue
-    username@instance> insert catrow catcf catcq catvalue
-    username@instance> quit
-
-The RegexExample class sets an iterator on the scanner. This does pattern matching
-against each key/value in accumulo, and only returns matching items. It will do this
-in parallel and will store the results in files in hdfs.
-
-The following will search for any rows in the input table that starts with "dog":
-
-    $ ./contrib/tool.sh lib/accumulo-examples-simple.jar org.apache.accumulo.examples.simple.mapreduce.RegexExample -u user -p passwd -i instance -t input --rowRegex 'dog.*' --output /tmp/output
-
-    $ hadoop fs -ls /tmp/output
-    Found 3 items
-    -rw-r--r--   1 username supergroup          0 2013-01-10 14:11 /tmp/output/_SUCCESS
-    drwxr-xr-x   - username supergroup          0 2013-01-10 14:10 /tmp/output/_logs
-    -rw-r--r--   1 username supergroup         51 2013-01-10 14:10 /tmp/output/part-m-00000
-
-We can see the output of our little map-reduce job:
-
-    $ hadoop fs -text /tmp/output/part-m-00000
-    dogrow dogcf:dogcq [] 1357844987994 false	dogvalue
-
-

http://git-wip-us.apache.org/repos/asf/accumulo/blob/8e0f19a1/docs/src/main/resources/examples/reservations.md
----------------------------------------------------------------------
diff --git a/docs/src/main/resources/examples/reservations.md b/docs/src/main/resources/examples/reservations.md
deleted file mode 100644
index 6b4886c..0000000
--- a/docs/src/main/resources/examples/reservations.md
+++ /dev/null
@@ -1,68 +0,0 @@
-<!--
-Licensed to the Apache Software Foundation (ASF) under one or more
-contributor license agreements.  See the NOTICE file distributed with
-this work for additional information regarding copyright ownership.
-The ASF licenses this file to You under the Apache License, Version 2.0
-(the "License"); you may not use this file except in compliance with
-the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
--->
----
-title: Apache Accumulo Reservations Example
----
-
-This example shows running a simple reservation system implemented using
-conditional mutations. This system guarantees that only one concurrent user can
-reserve a resource. The example's reserve command allows multiple users to be
-specified. When this is done, it creates a separate reservation thread for each
-user. In the example below threads are spun up for alice, bob, eve, mallory,
-and trent to reserve room06 on 20140101. Bob ends up getting the reservation
-and everyone else is put on a wait list. The example code will take any string
-for what, when and who.
-
-    $ ./bin/accumulo org.apache.accumulo.examples.simple.reservations.ARS
-    >connect test16 localhost root secret ars
-      connected
-    >
-      Commands :
-        reserve <what> <when> <who> {who}
-        cancel <what> <when> <who>
-        list <what> <when>
-    >reserve room06 20140101 alice bob eve mallory trent
-                       bob : RESERVED
-                   mallory : WAIT_LISTED
-                     alice : WAIT_LISTED
-                     trent : WAIT_LISTED
-                       eve : WAIT_LISTED
-    >list room06 20140101
-      Reservation holder : bob
-      Wait list : [mallory, alice, trent, eve]
-    >cancel room06 20140101 alice
-    >cancel room06 20140101 bob
-    >list room06 20140101
-      Reservation holder : mallory
-      Wait list : [trent, eve]
-    >quit
-
-Scanning the table in the Accumulo shell after running the example shows the
-following:
-
-    root@test16> table ars
-    root@test16 ars> scan
-    room06:20140101 res:0001 []    mallory
-    room06:20140101 res:0003 []    trent
-    room06:20140101 res:0004 []    eve
-    room06:20140101 tx:seq []    6
-
-The tx:seq column is incremented for each update to the row allowing for
-detection of concurrent changes. For an update to go through, the sequence
-number must not have changed since the data was read. If it does change,
-the conditional mutation will fail and the example code will retry.
-

http://git-wip-us.apache.org/repos/asf/accumulo/blob/8e0f19a1/docs/src/main/resources/examples/rgbalancer.md
----------------------------------------------------------------------
diff --git a/docs/src/main/resources/examples/rgbalancer.md b/docs/src/main/resources/examples/rgbalancer.md
deleted file mode 100644
index 3c80861..0000000
--- a/docs/src/main/resources/examples/rgbalancer.md
+++ /dev/null
@@ -1,161 +0,0 @@
-<!--
-Licensed to the Apache Software Foundation (ASF) under one or more
-contributor license agreements.  See the NOTICE file distributed with
-this work for additional information regarding copyright ownership.
-The ASF licenses this file to You under the Apache License, Version 2.0
-(the "License"); you may not use this file except in compliance with
-the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
--->
----
-title: Apache Accumulo Balancer Example
----
-
-For some data access patterns, its important to spread groups of tablets within
-a table out evenly.  Accumulo has a balancer that can do this using a regular
-expression to group tablets. This example shows how this balancer spreads 4
-groups of tablets within a table evenly across 17 tablet servers.
-
-Below shows creating a table and adding splits.  For this example we would like
-all of the tablets where the split point has the same two digits to be on
-different tservers.  This gives us four groups of tablets: 01, 02, 03, and 04.   
-
-    root@accumulo> createtable testRGB
-    root@accumulo testRGB> addsplits -t testRGB 01b 01m 01r 01z  02b 02m 02r 02z 03b 03m 03r 03z 04a 04b 04c 04d 04e 04f 04g 04h 04i 04j 04k 04l 04m 04n 04o 04p
-    root@accumulo testRGB> tables -l
-    accumulo.metadata    =>        !0
-    accumulo.replication =>      +rep
-    accumulo.root        =>        +r
-    testRGB              =>         2
-    trace                =>         1
-
-After adding the splits we look at the locations in the metadata table.
-
-    root@accumulo testRGB> scan -t accumulo.metadata -b 2; -e 2< -c loc
-    2;01b loc:34a5f6e086b000c []    ip-10-1-2-25:9997
-    2;01m loc:34a5f6e086b000c []    ip-10-1-2-25:9997
-    2;01r loc:14a5f6e079d0011 []    ip-10-1-2-15:9997
-    2;01z loc:14a5f6e079d000f []    ip-10-1-2-13:9997
-    2;02b loc:34a5f6e086b000b []    ip-10-1-2-26:9997
-    2;02m loc:14a5f6e079d000c []    ip-10-1-2-28:9997
-    2;02r loc:14a5f6e079d0012 []    ip-10-1-2-27:9997
-    2;02z loc:14a5f6e079d0012 []    ip-10-1-2-27:9997
-    2;03b loc:14a5f6e079d000d []    ip-10-1-2-21:9997
-    2;03m loc:14a5f6e079d000e []    ip-10-1-2-20:9997
-    2;03r loc:14a5f6e079d000d []    ip-10-1-2-21:9997
-    2;03z loc:14a5f6e079d000e []    ip-10-1-2-20:9997
-    2;04a loc:34a5f6e086b000b []    ip-10-1-2-26:9997
-    2;04b loc:14a5f6e079d0010 []    ip-10-1-2-17:9997
-    2;04c loc:14a5f6e079d0010 []    ip-10-1-2-17:9997
-    2;04d loc:24a5f6e07d3000c []    ip-10-1-2-16:9997
-    2;04e loc:24a5f6e07d3000d []    ip-10-1-2-29:9997
-    2;04f loc:24a5f6e07d3000c []    ip-10-1-2-16:9997
-    2;04g loc:24a5f6e07d3000a []    ip-10-1-2-14:9997
-    2;04h loc:14a5f6e079d000c []    ip-10-1-2-28:9997
-    2;04i loc:34a5f6e086b000d []    ip-10-1-2-19:9997
-    2;04j loc:34a5f6e086b000d []    ip-10-1-2-19:9997
-    2;04k loc:24a5f6e07d30009 []    ip-10-1-2-23:9997
-    2;04l loc:24a5f6e07d3000b []    ip-10-1-2-22:9997
-    2;04m loc:24a5f6e07d30009 []    ip-10-1-2-23:9997
-    2;04n loc:24a5f6e07d3000b []    ip-10-1-2-22:9997
-    2;04o loc:34a5f6e086b000a []    ip-10-1-2-18:9997
-    2;04p loc:24a5f6e07d30008 []    ip-10-1-2-24:9997
-    2< loc:24a5f6e07d30008 []    ip-10-1-2-24:9997
-
-Below the information above was massaged to show which tablet groups are on
-each tserver.  The four tablets in group 03 are on two tservers, ideally those
-tablets would be spread across 4 tservers.  Note the default tablet (2<) was
-categorized as group 04 below.
-
-    ip-10-1-2-13:9997 01
-    ip-10-1-2-14:9997 04
-    ip-10-1-2-15:9997 01
-    ip-10-1-2-16:9997 04 04
-    ip-10-1-2-17:9997 04 04
-    ip-10-1-2-18:9997 04
-    ip-10-1-2-19:9997 04 04
-    ip-10-1-2-20:9997 03 03
-    ip-10-1-2-21:9997 03 03
-    ip-10-1-2-22:9997 04 04
-    ip-10-1-2-23:9997 04 04
-    ip-10-1-2-24:9997 04 04
-    ip-10-1-2-25:9997 01 01
-    ip-10-1-2-26:9997 02 04
-    ip-10-1-2-27:9997 02 02
-    ip-10-1-2-28:9997 02 04
-    ip-10-1-2-29:9997 04
-
-To remedy this situation, the RegexGroupBalancer is configured with the
-commands below.  The configured regular expression selects the first two digits
-from a tablets end row as the group id.  Tablets that don't match and the
-default tablet are configured to be in group 04.
-
-    root@accumulo testRGB> config -t testRGB -s table.custom.balancer.group.regex.pattern=(\\d\\d).*
-    root@accumulo testRGB> config -t testRGB -s table.custom.balancer.group.regex.default=04
-    root@accumulo testRGB> config -t testRGB -s table.balancer=org.apache.accumulo.server.master.balancer.RegexGroupBalancer
-
-After waiting a little bit, look at the tablet locations again and all is good.
-
-    root@accumulo testRGB> scan -t accumulo.metadata -b 2; -e 2< -c loc
-    2;01b loc:34a5f6e086b000a []    ip-10-1-2-18:9997
-    2;01m loc:34a5f6e086b000c []    ip-10-1-2-25:9997
-    2;01r loc:14a5f6e079d0011 []    ip-10-1-2-15:9997
-    2;01z loc:14a5f6e079d000f []    ip-10-1-2-13:9997
-    2;02b loc:34a5f6e086b000b []    ip-10-1-2-26:9997
-    2;02m loc:14a5f6e079d000c []    ip-10-1-2-28:9997
-    2;02r loc:34a5f6e086b000d []    ip-10-1-2-19:9997
-    2;02z loc:14a5f6e079d0012 []    ip-10-1-2-27:9997
-    2;03b loc:24a5f6e07d3000d []    ip-10-1-2-29:9997
-    2;03m loc:24a5f6e07d30009 []    ip-10-1-2-23:9997
-    2;03r loc:14a5f6e079d000d []    ip-10-1-2-21:9997
-    2;03z loc:14a5f6e079d000e []    ip-10-1-2-20:9997
-    2;04a loc:34a5f6e086b000b []    ip-10-1-2-26:9997
-    2;04b loc:34a5f6e086b000c []    ip-10-1-2-25:9997
-    2;04c loc:14a5f6e079d0010 []    ip-10-1-2-17:9997
-    2;04d loc:14a5f6e079d000e []    ip-10-1-2-20:9997
-    2;04e loc:24a5f6e07d3000d []    ip-10-1-2-29:9997
-    2;04f loc:24a5f6e07d3000c []    ip-10-1-2-16:9997
-    2;04g loc:24a5f6e07d3000a []    ip-10-1-2-14:9997
-    2;04h loc:14a5f6e079d000c []    ip-10-1-2-28:9997
-    2;04i loc:14a5f6e079d0011 []    ip-10-1-2-15:9997
-    2;04j loc:34a5f6e086b000d []    ip-10-1-2-19:9997
-    2;04k loc:14a5f6e079d0012 []    ip-10-1-2-27:9997
-    2;04l loc:14a5f6e079d000f []    ip-10-1-2-13:9997
-    2;04m loc:24a5f6e07d30009 []    ip-10-1-2-23:9997
-    2;04n loc:24a5f6e07d3000b []    ip-10-1-2-22:9997
-    2;04o loc:34a5f6e086b000a []    ip-10-1-2-18:9997
-    2;04p loc:14a5f6e079d000d []    ip-10-1-2-21:9997
-    2< loc:24a5f6e07d30008 []    ip-10-1-2-24:9997
-
-Once again, the data above is transformed to make it easier to see which groups
-are on tservers.  The transformed data below shows that all groups are now
-evenly spread.
-
-    ip-10-1-2-13:9997 01 04
-    ip-10-1-2-14:9997    04
-    ip-10-1-2-15:9997 01 04
-    ip-10-1-2-16:9997    04
-    ip-10-1-2-17:9997    04
-    ip-10-1-2-18:9997 01 04
-    ip-10-1-2-19:9997 02 04
-    ip-10-1-2-20:9997 03 04
-    ip-10-1-2-21:9997 03 04
-    ip-10-1-2-22:9997    04
-    ip-10-1-2-23:9997 03 04
-    ip-10-1-2-24:9997    04
-    ip-10-1-2-25:9997 01 04
-    ip-10-1-2-26:9997 02 04
-    ip-10-1-2-27:9997 02 04
-    ip-10-1-2-28:9997 02 04
-    ip-10-1-2-29:9997 03 04
-
-If you need this functionality, but a regular expression does not meet your
-needs then extend GroupBalancer.  This allows you to specify a partitioning
-function in Java.  Use the RegexGroupBalancer source as an example.

http://git-wip-us.apache.org/repos/asf/accumulo/blob/8e0f19a1/docs/src/main/resources/examples/rowhash.md
----------------------------------------------------------------------
diff --git a/docs/src/main/resources/examples/rowhash.md b/docs/src/main/resources/examples/rowhash.md
deleted file mode 100644
index 9cd71a7..0000000
--- a/docs/src/main/resources/examples/rowhash.md
+++ /dev/null
@@ -1,61 +0,0 @@
-<!--
-Licensed to the Apache Software Foundation (ASF) under one or more
-contributor license agreements.  See the NOTICE file distributed with
-this work for additional information regarding copyright ownership.
-The ASF licenses this file to You under the Apache License, Version 2.0
-(the "License"); you may not use this file except in compliance with
-the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
--->
----
-title: Apache Accumulo RowHash Example
----
-
-This example shows a simple map/reduce job that reads from an accumulo table and
-writes back into that table.
-
-To run this example you will need some data in a table. The following will
-put a trivial amount of data into accumulo using the accumulo shell:
-
-    $ ./bin/accumulo shell -u username -p password
-    Shell - Apache Accumulo Interactive Shell
-    - version: 1.5.0
-    - instance name: instance
-    - instance id: 00000000-0000-0000-0000-000000000000
-    -
-    - type 'help' for a list of available commands
-    -
-    username@instance> createtable input
-    username@instance> insert a-row cf cq value
-    username@instance> insert b-row cf cq value
-    username@instance> quit
-
-The RowHash class will insert a hash for each row in the database if it contains a
-specified colum. Here's how you run the map/reduce job
-
-    $ ./contrib/tool.sh lib/accumulo-examples-simple.jar org.apache.accumulo.examples.simple.mapreduce.RowHash -u user -p passwd -i instance -t input --column cf:cq
-
-Now we can scan the table and see the hashes:
-
-    $ ./bin/accumulo shell -u username -p password
-    Shell - Apache Accumulo Interactive Shell
-    - version: 1.5.0
-    - instance name: instance
-    - instance id: 00000000-0000-0000-0000-000000000000
-    -
-    - type 'help' for a list of available commands
-    -
-    username@instance> scan -t input
-    a-row cf:cq []    value
-    a-row cf-HASHTYPE:cq-MD5BASE64 []    IGPBYI1uC6+AJJxC4r5YBA==
-    b-row cf:cq []    value
-    b-row cf-HASHTYPE:cq-MD5BASE64 []    IGPBYI1uC6+AJJxC4r5YBA==
-    username@instance>
-

http://git-wip-us.apache.org/repos/asf/accumulo/blob/8e0f19a1/docs/src/main/resources/examples/sample.md
----------------------------------------------------------------------
diff --git a/docs/src/main/resources/examples/sample.md b/docs/src/main/resources/examples/sample.md
deleted file mode 100644
index 432067e..0000000
--- a/docs/src/main/resources/examples/sample.md
+++ /dev/null
@@ -1,193 +0,0 @@
-<!--
-Licensed to the Apache Software Foundation (ASF) under one or more
-contributor license agreements.  See the NOTICE file distributed with
-this work for additional information regarding copyright ownership.
-The ASF licenses this file to You under the Apache License, Version 2.0
-(the "License"); you may not use this file except in compliance with
-the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
--->
----
-title: Apache Accumulo Sampling Example
----
-
-Basic Sampling Example
-----------------------
-
-Accumulo supports building a set of sample data that can be efficiently
-accessed by scanners.  What data is included in the sample set is configurable.
-Below, some data representing documents are inserted.  
-
-    root@instance sampex> createtable sampex
-    root@instance sampex> insert 9255 doc content 'abcde'
-    root@instance sampex> insert 9255 doc url file://foo.txt
-    root@instance sampex> insert 8934 doc content 'accumulo scales'
-    root@instance sampex> insert 8934 doc url file://accumulo_notes.txt
-    root@instance sampex> insert 2317 doc content 'milk, eggs, bread, parmigiano-reggiano'
-    root@instance sampex> insert 2317 doc url file://groceries/9.txt
-    root@instance sampex> insert 3900 doc content 'EC2 ate my homework'
-    root@instance sampex> insert 3900 doc uril file://final_project.txt
-
-Below the table sampex is configured to build a sample set.  The configuration
-causes Accumulo to include any row where `murmur3_32(row) % 3 ==0` in the
-tables sample data.
-
-    root@instance sampex> config -t sampex -s table.sampler.opt.hasher=murmur3_32
-    root@instance sampex> config -t sampex -s table.sampler.opt.modulus=3
-    root@instance sampex> config -t sampex -s table.sampler=org.apache.accumulo.core.client.sample.RowSampler
-
-Below, attempting to scan the sample returns an error.  This is because data
-was inserted before the sample set was configured.
-
-    root@instance sampex> scan --sample
-    2015-09-09 12:21:50,643 [shell.Shell] ERROR: org.apache.accumulo.core.client.SampleNotPresentException: Table sampex(ID:2) does not have sampling configured or built
-
-To remedy this problem, the following command will flush in memory data and
-compact any files that do not contain the correct sample data.   
-
-    root@instance sampex> compact -t sampex --sf-no-sample
-
-After the compaction, the sample scan works.  
-
-    root@instance sampex> scan --sample
-    2317 doc:content []    milk, eggs, bread, parmigiano-reggiano
-    2317 doc:url []    file://groceries/9.txt
-
-The commands below show that updates to data in the sample are seen when
-scanning the sample.
-
-    root@instance sampex> insert 2317 doc content 'milk, eggs, bread, parmigiano-reggiano, butter'
-    root@instance sampex> scan --sample
-    2317 doc:content []    milk, eggs, bread, parmigiano-reggiano, butter
-    2317 doc:url []    file://groceries/9.txt
-
-Inorder to make scanning the sample fast, sample data is partitioned as data is
-written to Accumulo.  This means if the sample configuration is changed, that
-data written previously is partitioned using a different criteria.  Accumulo
-will detect this situation and fail sample scans.  The commands below show this
-failure and fixiing the problem with a compaction.
-
-    root@instance sampex> config -t sampex -s table.sampler.opt.modulus=2
-    root@instance sampex> scan --sample
-    2015-09-09 12:22:51,058 [shell.Shell] ERROR: org.apache.accumulo.core.client.SampleNotPresentException: Table sampex(ID:2) does not have sampling configured or built
-    root@instance sampex> compact -t sampex --sf-no-sample
-    2015-09-09 12:23:07,242 [shell.Shell] INFO : Compaction of table sampex started for given range
-    root@instance sampex> scan --sample
-    2317 doc:content []    milk, eggs, bread, parmigiano-reggiano
-    2317 doc:url []    file://groceries/9.txt
-    3900 doc:content []    EC2 ate my homework
-    3900 doc:uril []    file://final_project.txt
-    9255 doc:content []    abcde
-    9255 doc:url []    file://foo.txt
-
-The example above is replicated in a java program using the Accumulo API.
-Below is the program name and the command to run it.
-
-    ./bin/accumulo org.apache.accumulo.examples.simple.sample.SampleExample -i instance -z localhost -u root -p secret
-
-The commands below look under the hood to give some insight into how this
-feature works.  The commands determine what files the sampex table is using.
-
-    root@instance sampex> tables -l
-    accumulo.metadata    =>        !0
-    accumulo.replication =>      +rep
-    accumulo.root        =>        +r
-    sampex               =>         2
-    trace                =>         1
-    root@instance sampex> scan -t accumulo.metadata -c file -b 2 -e 2<
-    2< file:hdfs://localhost:10000/accumulo/tables/2/default_tablet/A000000s.rf []    702,8
-
-Below shows running `accumulo rfile-info` on the file above.  This shows the
-rfile has a normal default locality group and a sample default locality group.
-The output also shows the configuration used to create the sample locality
-group.  The sample configuration within a rfile must match the tables sample
-configuration for sample scan to work.
-
-    $ ./bin/accumulo rfile-info hdfs://localhost:10000/accumulo/tables/2/default_tablet/A000000s.rf
-    Reading file: hdfs://localhost:10000/accumulo/tables/2/default_tablet/A000000s.rf
-    RFile Version            : 8
-    
-    Locality group           : <DEFAULT>
-    	Start block            : 0
-    	Num   blocks           : 1
-    	Index level 0          : 35 bytes  1 blocks
-    	First key              : 2317 doc:content [] 1437672014986 false
-    	Last key               : 9255 doc:url [] 1437672014875 false
-    	Num entries            : 8
-    	Column families        : [doc]
-    
-    Sample Configuration     :
-    	Sampler class          : org.apache.accumulo.core.client.sample.RowSampler
-    	Sampler options        : {hasher=murmur3_32, modulus=2}
-
-    Sample Locality group    : <DEFAULT>
-    	Start block            : 0
-    	Num   blocks           : 1
-    	Index level 0          : 36 bytes  1 blocks
-    	First key              : 2317 doc:content [] 1437672014986 false
-    	Last key               : 9255 doc:url [] 1437672014875 false
-    	Num entries            : 6
-    	Column families        : [doc]
-    
-    Meta block     : BCFile.index
-          Raw size             : 4 bytes
-          Compressed size      : 12 bytes
-          Compression type     : gz
-
-    Meta block     : RFile.index
-          Raw size             : 309 bytes
-          Compressed size      : 176 bytes
-          Compression type     : gz
-
-
-Shard Sampling Example
-----------------------
-
-The [shard example][shard] shows how to index and search files using Accumulo.  That
-example indexes documents into a table named `shard`.  The indexing scheme used
-in that example places the document name in the column qualifier.  A useful
-sample of this indexing scheme should contain all data for any document in the
-sample.   To accomplish this, the following commands build a sample for the
-shard table based on the column qualifier.
-
-    root@instance shard> config -t shard -s table.sampler.opt.hasher=murmur3_32
-    root@instance shard> config -t shard -s table.sampler.opt.modulus=101
-    root@instance shard> config -t shard -s table.sampler.opt.qualifier=true
-    root@instance shard> config -t shard -s table.sampler=org.apache.accumulo.core.client.sample.RowColumnSampler
-    root@instance shard> compact -t shard --sf-no-sample -w
-    2015-07-23 15:00:09,280 [shell.Shell] INFO : Compacting table ...
-    2015-07-23 15:00:10,134 [shell.Shell] INFO : Compaction of table shard completed for given range
-
-After enabling sampling, the command below counts the number of documents in
-the sample containing the words `import` and `int`.     
-
-    $ ./bin/accumulo org.apache.accumulo.examples.simple.shard.Query --sample -i instance16 -z localhost -t shard -u root -p secret import int | fgrep '.java' | wc
-         11      11    1246
-
-The command below counts the total number of documents containing the words
-`import` and `int`.
-
-    $ ./bin/accumulo org.apache.accumulo.examples.simple.shard.Query -i instance16 -z localhost -t shard -u root -p secret import int | fgrep '.java' | wc
-       1085    1085  118175
-
-The counts 11 out of 1085 total are around what would be expected for a modulus
-of 101.  Querying the sample first provides a quick way to estimate how much data
-the real query will bring back. 
-
-Another way sample data could be used with the shard example is with a
-specialized iterator.  In the examples source code there is an iterator named
-CutoffIntersectingIterator.  This iterator first checks how many documents are
-found in the sample data.  If too many documents are found in the sample data,
-then it returns nothing.   Otherwise it proceeds to query the full data set.
-To experiment with this iterator, use the following command.  The
-`--sampleCutoff` option below will cause the query to return nothing if based
-on the sample it appears a query would return more than 1000 documents.
-
-    $ ./bin/accumulo org.apache.accumulo.examples.simple.shard.Query --sampleCutoff 1000 -i instance16 -z localhost -t shard -u root -p secret import int | fgrep '.java' | wc

http://git-wip-us.apache.org/repos/asf/accumulo/blob/8e0f19a1/docs/src/main/resources/examples/shard.md
----------------------------------------------------------------------
diff --git a/docs/src/main/resources/examples/shard.md b/docs/src/main/resources/examples/shard.md
deleted file mode 100644
index 5e5789b..0000000
--- a/docs/src/main/resources/examples/shard.md
+++ /dev/null
@@ -1,68 +0,0 @@
-<!--
-Licensed to the Apache Software Foundation (ASF) under one or more
-contributor license agreements.  See the NOTICE file distributed with
-this work for additional information regarding copyright ownership.
-The ASF licenses this file to You under the Apache License, Version 2.0
-(the "License"); you may not use this file except in compliance with
-the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
--->
----
-title: Apache Accumulo Shard Example
----
-
-Accumulo has an iterator called the intersecting iterator which supports querying a term index that is partitioned by
-document, or "sharded". This example shows how to use the intersecting iterator through these four programs:
-
- * Index.java - Indexes a set of text files into an Accumulo table
- * Query.java - Finds documents containing a given set of terms.
- * Reverse.java - Reads the index table and writes a map of documents to terms into another table.
- * ContinuousQuery.java  Uses the table populated by Reverse.java to select N random terms per document. Then it continuously and randomly queries those terms.
-
-To run these example programs, create two tables like below.
-
-    username@instance> createtable shard
-    username@instance shard> createtable doc2term
-
-After creating the tables, index some files. The following command indexes all of the java files in the Accumulo source code.
-
-    $ cd /local/username/workspace/accumulo/
-    $ find core/src server/src -name "*.java" | xargs ./bin/accumulo org.apache.accumulo.examples.simple.shard.Index -i instance -z zookeepers -t shard -u username -p password --partitions 30
-
-The following command queries the index to find all files containing 'foo' and 'bar'.
-
-    $ ./bin/accumulo org.apache.accumulo.examples.simple.shard.Query -i instance -z zookeepers -t shard -u username -p password foo bar
-    /local/username/workspace/accumulo/src/core/src/test/java/accumulo/core/security/ColumnVisibilityTest.java
-    /local/username/workspace/accumulo/src/core/src/test/java/accumulo/core/client/mock/MockConnectorTest.java
-    /local/username/workspace/accumulo/src/core/src/test/java/accumulo/core/security/VisibilityEvaluatorTest.java
-    /local/username/workspace/accumulo/src/server/src/main/java/accumulo/test/functional/RowDeleteTest.java
-    /local/username/workspace/accumulo/src/server/src/test/java/accumulo/server/logger/TestLogWriter.java
-    /local/username/workspace/accumulo/src/server/src/main/java/accumulo/test/functional/DeleteEverythingTest.java
-    /local/username/workspace/accumulo/src/core/src/test/java/accumulo/core/data/KeyExtentTest.java
-    /local/username/workspace/accumulo/src/server/src/test/java/accumulo/server/constraints/MetadataConstraintsTest.java
-    /local/username/workspace/accumulo/src/core/src/test/java/accumulo/core/iterators/WholeRowIteratorTest.java
-    /local/username/workspace/accumulo/src/server/src/test/java/accumulo/server/util/DefaultMapTest.java
-    /local/username/workspace/accumulo/src/server/src/test/java/accumulo/server/tabletserver/InMemoryMapTest.java
-
-In order to run ContinuousQuery, we need to run Reverse.java to populate doc2term.
-
-    $ ./bin/accumulo org.apache.accumulo.examples.simple.shard.Reverse -i instance -z zookeepers --shardTable shard --doc2Term doc2term -u username -p password
-
-Below ContinuousQuery is run using 5 terms. So it selects 5 random terms from each document, then it continually
-randomly selects one set of 5 terms and queries. It prints the number of matching documents and the time in seconds.
-
-    $ ./bin/accumulo org.apache.accumulo.examples.simple.shard.ContinuousQuery -i instance -z zookeepers --shardTable shard --doc2Term doc2term -u username -p password --terms 5
-    [public, core, class, binarycomparable, b] 2  0.081
-    [wordtodelete, unindexdocument, doctablename, putdelete, insert] 1  0.041
-    [import, columnvisibilityinterpreterfactory, illegalstateexception, cv, columnvisibility] 1  0.049
-    [getpackage, testversion, util, version, 55] 1  0.048
-    [for, static, println, public, the] 55  0.211
-    [sleeptime, wrappingiterator, options, long, utilwaitthread] 1  0.057
-    [string, public, long, 0, wait] 12  0.132

http://git-wip-us.apache.org/repos/asf/accumulo/blob/8e0f19a1/docs/src/main/resources/examples/tabletofile.md
----------------------------------------------------------------------
diff --git a/docs/src/main/resources/examples/tabletofile.md b/docs/src/main/resources/examples/tabletofile.md
deleted file mode 100644
index 5316b51..0000000
--- a/docs/src/main/resources/examples/tabletofile.md
+++ /dev/null
@@ -1,61 +0,0 @@
-<!--
-Licensed to the Apache Software Foundation (ASF) under one or more
-contributor license agreements.  See the NOTICE file distributed with
-this work for additional information regarding copyright ownership.
-The ASF licenses this file to You under the Apache License, Version 2.0
-(the "License"); you may not use this file except in compliance with
-the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
--->
----
-Title: Apache Accumulo Table-to-File Example
----
-
-This example uses mapreduce to extract specified columns from an existing table.
-
-To run this example you will need some data in a table. The following will
-put a trivial amount of data into accumulo using the accumulo shell:
-
-    $ ./bin/accumulo shell -u username -p password
-    Shell - Apache Accumulo Interactive Shell
-    - version: 1.5.0
-    - instance name: instance
-    - instance id: 00000000-0000-0000-0000-000000000000
-    -
-    - type 'help' for a list of available commands
-    -
-    username@instance> createtable input
-    username@instance> insert dog cf cq dogvalue
-    username@instance> insert cat cf cq catvalue
-    username@instance> insert junk family qualifier junkvalue
-    username@instance> quit
-
-The TableToFile class configures a map-only job to read the specified columns and
-write the key/value pairs to a file in HDFS.
-
-The following will extract the rows containing the column "cf:cq":
-
-    $ ./contrib/tool.sh lib/accumulo-examples-simple.jar org.apache.accumulo.examples.simple.mapreduce.TableToFile -u user -p passwd -i instance -t input --columns cf:cq --output /tmp/output
-
-    $ hadoop fs -ls /tmp/output
-    -rw-r--r--   1 username supergroup          0 2013-01-10 14:44 /tmp/output/_SUCCESS
-    drwxr-xr-x   - username supergroup          0 2013-01-10 14:44 /tmp/output/_logs
-    drwxr-xr-x   - username supergroup          0 2013-01-10 14:44 /tmp/output/_logs/history
-    -rw-r--r--   1 username supergroup       9049 2013-01-10 14:44 /tmp/output/_logs/history/job_201301081658_0011_1357847072863_username_TableToFile%5F1357847071434
-    -rw-r--r--   1 username supergroup      26172 2013-01-10 14:44 /tmp/output/_logs/history/job_201301081658_0011_conf.xml
-    -rw-r--r--   1 username supergroup         50 2013-01-10 14:44 /tmp/output/part-m-00000
-
-We can see the output of our little map-reduce job:
-
-    $ hadoop fs -text /tmp/output/output/part-m-00000
-    catrow cf:cq []	catvalue
-    dogrow cf:cq []	dogvalue
-    $
-

http://git-wip-us.apache.org/repos/asf/accumulo/blob/8e0f19a1/docs/src/main/resources/examples/terasort.md
----------------------------------------------------------------------
diff --git a/docs/src/main/resources/examples/terasort.md b/docs/src/main/resources/examples/terasort.md
deleted file mode 100644
index 195bb4a..0000000
--- a/docs/src/main/resources/examples/terasort.md
+++ /dev/null
@@ -1,52 +0,0 @@
-<!--
-Licensed to the Apache Software Foundation (ASF) under one or more
-contributor license agreements.  See the NOTICE file distributed with
-this work for additional information regarding copyright ownership.
-The ASF licenses this file to You under the Apache License, Version 2.0
-(the "License"); you may not use this file except in compliance with
-the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
--->
----
-title: Apache Accumulo Terasort Example
----
-
-This example uses map/reduce to generate random input data that will
-be sorted by storing it into accumulo. It uses data very similar to the
-hadoop terasort benchmark.
-
-To run this example you run it with arguments describing the amount of data:
-
-    $ ./contrib/tool.sh lib/accumulo-examples-simple.jar org.apache.accumulo.examples.simple.mapreduce.TeraSortIngest \
-    -i instance -z zookeepers -u user -p password \
-    --count 10 \
-    --minKeySize 10 \
-    --maxKeySize 10 \
-    --minValueSize 78 \
-    --maxValueSize 78 \
-    --table sort \
-    --splits 10 \
-
-After the map reduce job completes, scan the data:
-
-    $ ./bin/accumulo shell -u username -p password
-    username@instance> scan -t sort
-    +l-$$OE/ZH c:         4 []    GGGGGGGGGGWWWWWWWWWWMMMMMMMMMMCCCCCCCCCCSSSSSSSSSSIIIIIIIIIIYYYYYYYYYYOOOOOOOO
-    ,C)wDw//u= c:        10 []    CCCCCCCCCCSSSSSSSSSSIIIIIIIIIIYYYYYYYYYYOOOOOOOOOOEEEEEEEEEEUUUUUUUUUUKKKKKKKK
-    75@~?'WdUF c:         1 []    IIIIIIIIIIYYYYYYYYYYOOOOOOOOOOEEEEEEEEEEUUUUUUUUUUKKKKKKKKKKAAAAAAAAAAQQQQQQQQ
-    ;L+!2rT~hd c:         8 []    MMMMMMMMMMCCCCCCCCCCSSSSSSSSSSIIIIIIIIIIYYYYYYYYYYOOOOOOOOOOEEEEEEEEEEUUUUUUUU
-    LsS8)|.ZLD c:         5 []    OOOOOOOOOOEEEEEEEEEEUUUUUUUUUUKKKKKKKKKKAAAAAAAAAAQQQQQQQQQQGGGGGGGGGGWWWWWWWW
-    M^*dDE;6^< c:         9 []    UUUUUUUUUUKKKKKKKKKKAAAAAAAAAAQQQQQQQQQQGGGGGGGGGGWWWWWWWWWWMMMMMMMMMMCCCCCCCC
-    ^Eu)<n#kdP c:         3 []    YYYYYYYYYYOOOOOOOOOOEEEEEEEEEEUUUUUUUUUUKKKKKKKKKKAAAAAAAAAAQQQQQQQQQQGGGGGGGG
-    le5awB.$sm c:         6 []    WWWWWWWWWWMMMMMMMMMMCCCCCCCCCCSSSSSSSSSSIIIIIIIIIIYYYYYYYYYYOOOOOOOOOOEEEEEEEE
-    q__[fwhKFg c:         7 []    EEEEEEEEEEUUUUUUUUUUKKKKKKKKKKAAAAAAAAAAQQQQQQQQQQGGGGGGGGGGWWWWWWWWWWMMMMMMMM
-    w[o||:N&H, c:         2 []    QQQQQQQQQQGGGGGGGGGGWWWWWWWWWWMMMMMMMMMMCCCCCCCCCCSSSSSSSSSSIIIIIIIIIIYYYYYYYY
-
-Of course, a real benchmark would ingest millions of entries.

http://git-wip-us.apache.org/repos/asf/accumulo/blob/8e0f19a1/docs/src/main/resources/examples/visibility.md
----------------------------------------------------------------------
diff --git a/docs/src/main/resources/examples/visibility.md b/docs/src/main/resources/examples/visibility.md
deleted file mode 100644
index 8345a9b..0000000
--- a/docs/src/main/resources/examples/visibility.md
+++ /dev/null
@@ -1,133 +0,0 @@
-<!--
-Licensed to the Apache Software Foundation (ASF) under one or more
-contributor license agreements.  See the NOTICE file distributed with
-this work for additional information regarding copyright ownership.
-The ASF licenses this file to You under the Apache License, Version 2.0
-(the "License"); you may not use this file except in compliance with
-the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
--->
----
-title: Apache Accumulo Visibility, Authorizations, and Permissions Example
----
-
-## Creating a new user
-
-    root@instance> createuser username
-    Enter new password for 'username': ********
-    Please confirm new password for 'username': ********
-    root@instance> user username
-    Enter password for user username: ********
-    username@instance> createtable vistest
-    06 10:48:47,931 [shell.Shell] ERROR: org.apache.accumulo.core.client.AccumuloSecurityException: Error PERMISSION_DENIED - User does not have permission to perform this action
-    username@instance> userpermissions
-    System permissions:
-
-    Table permissions (accumulo.metadata): Table.READ
-    username@instance>
-
-A user does not by default have permission to create a table.
-
-## Granting permissions to a user
-
-    username@instance> user root
-    Enter password for user root: ********
-    root@instance> grant -s System.CREATE_TABLE -u username
-    root@instance> user username
-    Enter password for user username: ********
-    username@instance> createtable vistest
-    username@instance> userpermissions
-    System permissions: System.CREATE_TABLE
-
-    Table permissions (accumulo.metadata): Table.READ
-    Table permissions (vistest): Table.READ, Table.WRITE, Table.BULK_IMPORT, Table.ALTER_TABLE, Table.GRANT, Table.DROP_TABLE
-    username@instance vistest>
-
-## Inserting data with visibilities
-
-Visibilities are boolean AND (&) and OR (|) combinations of authorization
-tokens. Authorization tokens are arbitrary strings taken from a restricted
-ASCII character set. Parentheses are required to specify order of operations
-in visibilities.
-
-    username@instance vistest> insert row f1 q1 v1 -l A
-    username@instance vistest> insert row f2 q2 v2 -l A&B
-    username@instance vistest> insert row f3 q3 v3 -l apple&carrot|broccoli|spinach
-    06 11:19:01,432 [shell.Shell] ERROR: org.apache.accumulo.core.util.BadArgumentException: cannot mix | and & near index 12
-    apple&carrot|broccoli|spinach
-                ^
-    username@instance vistest> insert row f3 q3 v3 -l (apple&carrot)|broccoli|spinach
-    username@instance vistest>
-
-## Scanning with authorizations
-
-Authorizations are sets of authorization tokens. Each Accumulo user has
-authorizations and each Accumulo scan has authorizations. Scan authorizations
-are only allowed to be a subset of the user's authorizations. By default, a
-user's authorizations set is empty.
-
-    username@instance vistest> scan
-    username@instance vistest> scan -s A
-    06 11:43:14,951 [shell.Shell] ERROR: java.lang.RuntimeException: org.apache.accumulo.core.client.AccumuloSecurityException: Error BAD_AUTHORIZATIONS - The user does not have the specified authorizations assigned
-    username@instance vistest>
-
-## Setting authorizations for a user
-
-    username@instance vistest> setauths -s A
-    06 11:53:42,056 [shell.Shell] ERROR: org.apache.accumulo.core.client.AccumuloSecurityException: Error PERMISSION_DENIED - User does not have permission to perform this action
-    username@instance vistest>
-
-A user cannot set authorizations unless the user has the System.ALTER_USER permission.
-The root user has this permission.
-
-    username@instance vistest> user root
-    Enter password for user root: ********
-    root@instance vistest> setauths -s A -u username
-    root@instance vistest> user username
-    Enter password for user username: ********
-    username@instance vistest> scan -s A
-    row f1:q1 [A]    v1
-    username@instance vistest> scan
-    row f1:q1 [A]    v1
-    username@instance vistest>
-
-The default authorizations for a scan are the user's entire set of authorizations.
-
-    username@instance vistest> user root
-    Enter password for user root: ********
-    root@instance vistest> setauths -s A,B,broccoli -u username
-    root@instance vistest> user username
-    Enter password for user username: ********
-    username@instance vistest> scan
-    row f1:q1 [A]    v1
-    row f2:q2 [A&B]    v2
-    row f3:q3 [(apple&carrot)|broccoli|spinach]    v3
-    username@instance vistest> scan -s B
-    username@instance vistest>
-
-If you want, you can limit a user to only be able to insert data which they can read themselves.
-It can be set with the following constraint.
-
-    username@instance vistest> user root
-    Enter password for user root: ******
-    root@instance vistest> config -t vistest -s table.constraint.1=org.apache.accumulo.core.security.VisibilityConstraint
-    root@instance vistest> user username
-    Enter password for user username: ********
-    username@instance vistest> insert row f4 q4 v4 -l spinach
-        Constraint Failures:
-            ConstraintViolationSummary(constrainClass:org.apache.accumulo.core.security.VisibilityConstraint, violationCode:2, violationDescription:User does not have authorization on column visibility, numberOfViolatingMutations:1)
-    username@instance vistest> insert row f4 q4 v4 -l spinach|broccoli
-    username@instance vistest> scan
-    row f1:q1 [A]    v1
-    row f2:q2 [A&B]    v2
-    row f3:q3 [(apple&carrot)|broccoli|spinach]    v3
-    row f4:q4 [spinach|broccoli]    v4
-    username@instance vistest>
-

http://git-wip-us.apache.org/repos/asf/accumulo/blob/8e0f19a1/examples/simple/.gitignore
----------------------------------------------------------------------
diff --git a/examples/simple/.gitignore b/examples/simple/.gitignore
deleted file mode 100644
index e77a822..0000000
--- a/examples/simple/.gitignore
+++ /dev/null
@@ -1,28 +0,0 @@
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-# Maven ignores
-/target/
-
-# IDE ignores
-/.settings/
-/.project
-/.classpath
-/.pydevproject
-/.idea
-/*.iml
-/nbproject/
-/nbactions.xml
-/nb-configuration.xml

http://git-wip-us.apache.org/repos/asf/accumulo/blob/8e0f19a1/examples/simple/pom.xml
----------------------------------------------------------------------
diff --git a/examples/simple/pom.xml b/examples/simple/pom.xml
deleted file mode 100644
index a9e6b7b..0000000
--- a/examples/simple/pom.xml
+++ /dev/null
@@ -1,104 +0,0 @@
-<?xml version="1.0" encoding="UTF-8"?>
-<!--
-  Licensed to the Apache Software Foundation (ASF) under one or more
-  contributor license agreements.  See the NOTICE file distributed with
-  this work for additional information regarding copyright ownership.
-  The ASF licenses this file to You under the Apache License, Version 2.0
-  (the "License"); you may not use this file except in compliance with
-  the License.  You may obtain a copy of the License at
-
-      http://www.apache.org/licenses/LICENSE-2.0
-
-  Unless required by applicable law or agreed to in writing, software
-  distributed under the License is distributed on an "AS IS" BASIS,
-  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-  See the License for the specific language governing permissions and
-  limitations under the License.
--->
-<project xmlns="http://maven.apache.org/POM/4.0.0" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 http://maven.apache.org/maven-v4_0_0.xsd">
-  <modelVersion>4.0.0</modelVersion>
-  <parent>
-    <groupId>org.apache.accumulo</groupId>
-    <artifactId>accumulo-project</artifactId>
-    <version>2.0.0-SNAPSHOT</version>
-    <relativePath>../../pom.xml</relativePath>
-  </parent>
-  <artifactId>accumulo-examples-simple</artifactId>
-  <name>Apache Accumulo Simple Examples</name>
-  <description>Simple Apache Accumulo examples.</description>
-  <dependencies>
-    <dependency>
-      <groupId>com.beust</groupId>
-      <artifactId>jcommander</artifactId>
-    </dependency>
-    <dependency>
-      <groupId>com.google.auto.service</groupId>
-      <artifactId>auto-service</artifactId>
-      <optional>true</optional>
-    </dependency>
-    <dependency>
-      <groupId>commons-cli</groupId>
-      <artifactId>commons-cli</artifactId>
-    </dependency>
-    <dependency>
-      <groupId>commons-configuration</groupId>
-      <artifactId>commons-configuration</artifactId>
-    </dependency>
-    <dependency>
-      <groupId>jline</groupId>
-      <artifactId>jline</artifactId>
-    </dependency>
-    <dependency>
-      <groupId>log4j</groupId>
-      <artifactId>log4j</artifactId>
-    </dependency>
-    <dependency>
-      <groupId>org.apache.accumulo</groupId>
-      <artifactId>accumulo-core</artifactId>
-    </dependency>
-    <dependency>
-      <groupId>org.apache.accumulo</groupId>
-      <artifactId>accumulo-fate</artifactId>
-    </dependency>
-    <dependency>
-      <groupId>org.apache.accumulo</groupId>
-      <artifactId>accumulo-shell</artifactId>
-    </dependency>
-    <dependency>
-      <groupId>org.apache.accumulo</groupId>
-      <artifactId>accumulo-tracer</artifactId>
-    </dependency>
-    <dependency>
-      <groupId>org.apache.hadoop</groupId>
-      <artifactId>hadoop-client</artifactId>
-    </dependency>
-    <dependency>
-      <groupId>org.apache.htrace</groupId>
-      <artifactId>htrace-core</artifactId>
-    </dependency>
-    <dependency>
-      <groupId>commons-httpclient</groupId>
-      <artifactId>commons-httpclient</artifactId>
-      <scope>test</scope>
-    </dependency>
-    <dependency>
-      <groupId>junit</groupId>
-      <artifactId>junit</artifactId>
-      <scope>test</scope>
-    </dependency>
-  </dependencies>
-  <build>
-    <pluginManagement>
-      <plugins>
-        <plugin>
-          <groupId>org.apache.maven.plugins</groupId>
-          <artifactId>maven-site-plugin</artifactId>
-          <configuration>
-            <!-- this is just an example. Leave it out. -->
-            <skip>true</skip>
-          </configuration>
-        </plugin>
-      </plugins>
-    </pluginManagement>
-  </build>
-</project>

http://git-wip-us.apache.org/repos/asf/accumulo/blob/8e0f19a1/examples/simple/src/main/findbugs/exclude-filter.xml
----------------------------------------------------------------------
diff --git a/examples/simple/src/main/findbugs/exclude-filter.xml b/examples/simple/src/main/findbugs/exclude-filter.xml
deleted file mode 100644
index 0ce1597..0000000
--- a/examples/simple/src/main/findbugs/exclude-filter.xml
+++ /dev/null
@@ -1,23 +0,0 @@
-<!--
-  Licensed to the Apache Software Foundation (ASF) under one or more
-  contributor license agreements.  See the NOTICE file distributed with
-  this work for additional information regarding copyright ownership.
-  The ASF licenses this file to You under the Apache License, Version 2.0
-  (the "License"); you may not use this file except in compliance with
-  the License.  You may obtain a copy of the License at
-
-      http://www.apache.org/licenses/LICENSE-2.0
-
-  Unless required by applicable law or agreed to in writing, software
-  distributed under the License is distributed on an "AS IS" BASIS,
-  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-  See the License for the specific language governing permissions and
-  limitations under the License.
--->
-<FindBugsFilter>
-  <Match>
-    <!-- example classes can call System.exit -->
-    <Class name="org.apache.accumulo.examples.simple.isolation.InterferenceTest$Writer" />
-    <Bug code="DM" pattern="DM_EXIT" />
-  </Match>
-</FindBugsFilter>

http://git-wip-us.apache.org/repos/asf/accumulo/blob/8e0f19a1/examples/simple/src/main/java/org/apache/accumulo/examples/simple/client/CountingVerifyingReceiver.java
----------------------------------------------------------------------
diff --git a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/client/CountingVerifyingReceiver.java b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/client/CountingVerifyingReceiver.java
deleted file mode 100644
index 4ac9db1..0000000
--- a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/client/CountingVerifyingReceiver.java
+++ /dev/null
@@ -1,64 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one or more
- * contributor license agreements.  See the NOTICE file distributed with
- * this work for additional information regarding copyright ownership.
- * The ASF licenses this file to You under the Apache License, Version 2.0
- * (the "License"); you may not use this file except in compliance with
- * the License.  You may obtain a copy of the License at
- *
- *     http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-package org.apache.accumulo.examples.simple.client;
-
-import static java.nio.charset.StandardCharsets.UTF_8;
-
-import java.util.Arrays;
-import java.util.HashMap;
-
-import org.apache.accumulo.core.data.Key;
-import org.apache.accumulo.core.data.Value;
-import org.apache.hadoop.io.Text;
-import org.slf4j.Logger;
-import org.slf4j.LoggerFactory;
-
-/**
- * Internal class used to verify validity of data read.
- */
-class CountingVerifyingReceiver {
-  private static final Logger log = LoggerFactory.getLogger(CountingVerifyingReceiver.class);
-
-  long count = 0;
-  int expectedValueSize = 0;
-  HashMap<Text,Boolean> expectedRows;
-
-  CountingVerifyingReceiver(HashMap<Text,Boolean> expectedRows, int expectedValueSize) {
-    this.expectedRows = expectedRows;
-    this.expectedValueSize = expectedValueSize;
-  }
-
-  public void receive(Key key, Value value) {
-
-    String row = key.getRow().toString();
-    long rowid = Integer.parseInt(row.split("_")[1]);
-
-    byte expectedValue[] = RandomBatchWriter.createValue(rowid, expectedValueSize);
-
-    if (!Arrays.equals(expectedValue, value.get())) {
-      log.error("Got unexpected value for " + key + " expected : " + new String(expectedValue, UTF_8) + " got : " + new String(value.get(), UTF_8));
-    }
-
-    if (!expectedRows.containsKey(key.getRow())) {
-      log.error("Got unexpected key " + key);
-    } else {
-      expectedRows.put(key.getRow(), true);
-    }
-
-    count++;
-  }
-}

http://git-wip-us.apache.org/repos/asf/accumulo/blob/8e0f19a1/examples/simple/src/main/java/org/apache/accumulo/examples/simple/client/Flush.java
----------------------------------------------------------------------
diff --git a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/client/Flush.java b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/client/Flush.java
deleted file mode 100644
index fb460f6..0000000
--- a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/client/Flush.java
+++ /dev/null
@@ -1,37 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one or more
- * contributor license agreements.  See the NOTICE file distributed with
- * this work for additional information regarding copyright ownership.
- * The ASF licenses this file to You under the Apache License, Version 2.0
- * (the "License"); you may not use this file except in compliance with
- * the License.  You may obtain a copy of the License at
- *
- *     http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-package org.apache.accumulo.examples.simple.client;
-
-import org.apache.accumulo.core.cli.ClientOnRequiredTable;
-import org.apache.accumulo.core.client.Connector;
-
-/**
- * Simple example for using tableOperations() (like create, delete, flush, etc).
- */
-public class Flush {
-
-  public static void main(String[] args) {
-    ClientOnRequiredTable opts = new ClientOnRequiredTable();
-    opts.parseArgs(Flush.class.getName(), args);
-    try {
-      Connector connector = opts.getConnector();
-      connector.tableOperations().flush(opts.getTableName(), null, null, true);
-    } catch (Exception e) {
-      throw new RuntimeException(e);
-    }
-  }
-}


Mime
View raw message