accumulo-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From mwa...@apache.org
Subject [accumulo-website] branch master updated: Improved links in docs
Date Wed, 16 Jan 2019 00:26:36 GMT
This is an automated email from the ASF dual-hosted git repository.

mwalch pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/accumulo-website.git


The following commit(s) were added to refs/heads/master by this push:
     new ab08d17  Improved links in docs
ab08d17 is described below

commit ab08d1718e49c359cf8b478bba072ed5fd22474b
Author: Mike Walch <mwalch@apache.org>
AuthorDate: Tue Jan 15 19:25:38 2019 -0500

    Improved links in docs
---
 _docs-2/administration/in-depth-install.md | 10 ++++------
 _docs-2/development/mapreduce.md           |  3 ++-
 2 files changed, 6 insertions(+), 7 deletions(-)

diff --git a/_docs-2/administration/in-depth-install.md b/_docs-2/administration/in-depth-install.md
index 530191c..c22ce2c 100644
--- a/_docs-2/administration/in-depth-install.md
+++ b/_docs-2/administration/in-depth-install.md
@@ -344,7 +344,7 @@ configuration is:
 general.vfs.context.classpath.app1.delegation=post
 ```
 
-To use contexts in your application you can set the `table.classpath.context` on your tables
or use the `setClassLoaderContext()` method on Scanner
+To use contexts in your application you can set the {% plink table.classpath.context %} on
your tables or use the `setClassLoaderContext()` method on Scanner
 and BatchScanner passing in the name of the context, app1 in the example above. Setting the
property on the table allows your minc, majc, and scan 
 iterators to load classes from the locations defined by the context. Passing the context
name to the scanners allows you to override the table setting
 to load only scan time iterators from a different location. 
@@ -445,12 +445,10 @@ to be able to scale to using 10's of GB of RAM and 10's of CPU cores.
 Accumulo TabletServers bind certain ports on the host to accommodate remote procedure calls
to/from
 other nodes. Running more than one TabletServer on a host requires that you set the environment
variable
 `ACCUMULO_SERVICE_INSTANCE` to an instance number (i.e 1, 2) for each instance that is started.
Also, set
-these properties in [accumulo.properties]:
+the these properties in [accumulo.properties]:
 
-```
-tserver.port.search=true
-replication.receipt.service.port=0
-```
+* {% plink tserver.port.search %} = `true`
+* {% plink replication.receipt.service.port %} = `0`
 
 ## Logging
 
diff --git a/_docs-2/development/mapreduce.md b/_docs-2/development/mapreduce.md
index adf0643..f5877e4 100644
--- a/_docs-2/development/mapreduce.md
+++ b/_docs-2/development/mapreduce.md
@@ -42,7 +42,7 @@ MapReduce jobs to run with both Accumulo's & Hadoop's dependencies on
the classp
 Since 2.0, Accumulo no longer has the same versions for dependencies as Hadoop. While this
allows
 Accumulo to update its dependencies more frequently, it can cause problems if both Accumulo's
&
 Hadoop's dependencies are on the classpath of the MapReduce job. When launching a MapReduce
job that
-use Accumulo, you should build a shaded jar with all of your dependencies and complete the
following
+use Accumulo, you should build a [shaded jar] with all of your dependencies and complete
the following
 steps so YARN only includes Hadoop code (and not all of Hadoop's dependencies) when running
your MapReduce job:
 
 1. Set `export HADOOP_USE_CLIENT_CLASSLOADER=true` in your environment before submitting
@@ -181,6 +181,7 @@ The [Accumulo Examples repo][examples-repo] has several MapReduce examples:
 * [tablettofile] - Uses MapReduce to read a table and write one of its columns to a file
in HDFS
 * [uniquecols] - Uses MapReduce to count unique columns in Accumulo
 
+[shaded jar]: https://maven.apache.org/plugins/maven-shade-plugin/index.html
 [AccumuloInputFormat]: {% jurl org.apache.accumulo.hadoop.mapreduce.AccumuloInputFormat %}
 [AccumuloOutputFormat]: {% jurl org.apache.accumulo.hadoop.mapreduce.AccumuloOutputFormat
%}
 [AccumuloFileOutputFormat]: {% jurl org.apache.accumulo.hadoop.mapreduce.AccumuloFileOutputFormat
%}


Mime
View raw message