hbase-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From bus...@apache.org
Subject [01/12] hbase git commit: HBASE-20344 Fix asciidoc warnings [Forced Update!]
Date Sun, 15 Apr 2018 02:45:50 GMT
Repository: hbase
Updated Branches:
  refs/heads/HBASE-20388 0deb8cd8a -> 3f5027795 (forced update)


HBASE-20344 Fix asciidoc warnings

Signed-off-by: Sean Busbey <busbey@apache.org>


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/826909a5
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/826909a5
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/826909a5

Branch: refs/heads/HBASE-20388
Commit: 826909a59cf27adb1578c627e743233b378922a5
Parents: d59a6c8
Author: Peter Somogyi <psomogyi@apache.org>
Authored: Wed Apr 4 13:36:48 2018 +0200
Committer: Peter Somogyi <psomogyi@apache.org>
Committed: Fri Apr 13 11:32:40 2018 +0200

----------------------------------------------------------------------
 src/main/asciidoc/_chapters/backup_restore.adoc | 106 +++++++++----------
 src/main/asciidoc/_chapters/compression.adoc    |  20 ----
 .../asciidoc/_chapters/getting_started.adoc     |   7 --
 src/main/asciidoc/_chapters/hbase_mob.adoc      |   4 -
 src/main/asciidoc/_chapters/ops_mgt.adoc        |   3 -
 src/main/asciidoc/_chapters/performance.adoc    |   2 -
 src/main/asciidoc/_chapters/schema_design.adoc  |   2 -
 src/main/asciidoc/_chapters/security.adoc       |  12 ---
 src/main/asciidoc/_chapters/shell.adoc          |   2 -
 9 files changed, 53 insertions(+), 105 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/hbase/blob/826909a5/src/main/asciidoc/_chapters/backup_restore.adoc
----------------------------------------------------------------------
diff --git a/src/main/asciidoc/_chapters/backup_restore.adoc b/src/main/asciidoc/_chapters/backup_restore.adoc
index b02af41..cb7fced 100644
--- a/src/main/asciidoc/_chapters/backup_restore.adoc
+++ b/src/main/asciidoc/_chapters/backup_restore.adoc
@@ -175,7 +175,7 @@ and its options. The below information is captured in this help message
for each
 // hbase backup create
 
 [[br.creating.complete.backup]]
-### Creating a Backup Image
+=== Creating a Backup Image
 
 [NOTE]
 ====
@@ -204,7 +204,7 @@ dataset with a restore operation, having the backup ID readily available
can sav
 ====
 
 [[br.create.positional.cli.arguments]]
-#### Positional Command-Line Arguments
+==== Positional Command-Line Arguments
 
 _type_::
   The type of backup to execute: _full_ or _incremental_. As a reminder, an _incremental_
backup requires a _full_ backup to
@@ -215,7 +215,7 @@ _backup_path_::
   are _hdfs:_, _webhdfs:_, _gpfs:_, and _s3fs:_.
 
 [[br.create.named.cli.arguments]]
-#### Named Command-Line Arguments
+==== Named Command-Line Arguments
 
 _-t <table_name[,table_name]>_::
   A comma-separated list of tables to back up. If no tables are specified, all tables are
backed up. No regular-expression or
@@ -242,7 +242,7 @@ _-q <name>_::
   is useful to prevent backup tasks from stealing resources away from other MapReduce jobs
of high importance.
 
 [[br.usage.examples]]
-#### Example usage
+==== Example usage
 
 [source]
 ----
@@ -255,7 +255,7 @@ in the path _/data/backup_. The _-w_ option specifies that no more than
three pa
 // hbase backup restore
 
 [[br.restoring.backup]]
-### Restoring a Backup Image
+=== Restoring a Backup Image
 
 Run the following command as an HBase superuser. You can only restore a backup on a running
HBase cluster because the data must be
 redistributed the RegionServers for the operation to complete successfully.
@@ -266,7 +266,7 @@ hbase restore <backup_path> <backup_id>
 ----
 
 [[br.restore.positional.args]]
-#### Positional Command-Line Arguments
+==== Positional Command-Line Arguments
 
 _backup_path_::
   The _backup_path_ argument specifies the full filesystem URI of where to store the backup
image. Valid prefixes are
@@ -277,7 +277,7 @@ _backup_id_::
 
 
 [[br.restore.named.args]]
-#### Named Command-Line Arguments
+==== Named Command-Line Arguments
 
 _-t <table_name[,table_name]>_::
   A comma-separated list of tables to restore. See <<br.using.backup.sets,Backup Sets>>
for more
@@ -304,7 +304,7 @@ _-o_::
 
 
 [[br.restore.usage]]
-#### Example of Usage
+==== Example of Usage
 
 [source]
 ----
@@ -319,7 +319,7 @@ This command restores two tables of an incremental backup image. In this
example
 // hbase backup merge
 
 [[br.merge.backup]]
-### Merging Incremental Backup Images
+=== Merging Incremental Backup Images
 
 This command can be used to merge two or more incremental backup images into a single incremental
 backup image. This can be used to consolidate multiple, small incremental backup images into
a single
@@ -332,18 +332,18 @@ $ hbase backup merge <backup_ids>
 ----
 
 [[br.merge.backup.positional.cli.arguments]]
-#### Positional Command-Line Arguments
+==== Positional Command-Line Arguments
 
 _backup_ids_::
   A comma-separated list of incremental backup image IDs that are to be combined into a single
image.
 
 [[br.merge.backup.named.cli.arguments]]
-#### Named Command-Line Arguments
+==== Named Command-Line Arguments
 
 None.
 
 [[br.merge.backup.example]]
-#### Example usage
+==== Example usage
 
 [source]
 ----
@@ -353,7 +353,7 @@ $ hbase backup merge backupId_1467823988425,backupId_1467827588425
 // hbase backup set
 
 [[br.using.backup.sets]]
-### Using Backup Sets
+=== Using Backup Sets
 
 Backup sets can ease the administration of HBase data backups and restores by reducing the
amount of repetitive input
 of table names. You can group tables into a named backup set with the `hbase backup set add`
command. You can then use
@@ -381,7 +381,7 @@ $ hbase backup set <subcommand> <backup_set_name> <tables>
 ----
 
 [[br.set.subcommands]]
-#### Backup Set Subcommands
+==== Backup Set Subcommands
 
 The following list details subcommands of the hbase backup set command.
 
@@ -406,7 +406,7 @@ _delete_::
   Deletes a backup set. Enter the value for the _backup_set_name_ option directly after the
`hbase backup set delete` command.
 
 [[br.set.positional.cli.arguments]]
-#### Positional Command-Line Arguments
+==== Positional Command-Line Arguments
 
 _backup_set_name_::
   Use to assign or invoke a backup set name. The backup set name must contain only printable
characters and cannot have any spaces.
@@ -419,7 +419,7 @@ TIP: Maintain a log or other record of the case-sensitive backup set names
and t
 or remote cluster, backup strategy. This information can help you in case of failure on the
primary cluster.
 
 [[br.set.usage]]
-#### Example of Usage
+==== Example of Usage
 
 [source]
 ----
@@ -432,7 +432,7 @@ Depending on the environment, this command results in _one_ of the following
act
 * If the `Q1Data` backup set exists already, the tables `TEAM_3` and `TEAM_4` are added to
the `Q1Data` backup set.
 
 [[br.administration]]
-## Administration of Backup Images
+== Administration of Backup Images
 
 The `hbase backup` command has several subcommands that help with administering backup images
as they accumulate. Most production
 environments require recurring backups, so it is necessary to have utilities to help manage
the data of the backup repository.
@@ -445,7 +445,7 @@ the HBase superuser.
 // hbase backup progress
 
 [[br.managing.backup.progress]]
-### Managing Backup Progress
+=== Managing Backup Progress
 
 You can monitor a running backup in another terminal session by running the _hbase backup
progress_ command and specifying the backup ID as an argument.
 
@@ -457,18 +457,18 @@ $ hbase backup progress <backup_id>
 ----
 
 [[br.progress.positional.cli.arguments]]
-#### Positional Command-Line Arguments
+==== Positional Command-Line Arguments
 
 _backup_id_::
   Specifies the backup that you want to monitor by seeing the progress information. The backupId
is case-sensitive.
 
 [[br.progress.named.cli.arguments]]
-#### Named Command-Line Arguments
+==== Named Command-Line Arguments
 
 None.
 
 [[br.progress.example]]
-#### Example usage
+==== Example usage
 
 [source]
 ----
@@ -478,7 +478,7 @@ hbase backup progress backupId_1467823988425
 // hbase backup history
 
 [[br.managing.backup.history]]
-### Managing Backup History
+=== Managing Backup History
 
 This command displays a log of backup sessions. The information for each session includes
backup ID, type (full or incremental), the tables
 in the backup, status, and start and end time. Specify the number of backup sessions to display
with the optional -n argument.
@@ -489,13 +489,13 @@ $ hbase backup history <backup_id>
 ----
 
 [[br.history.positional.cli.arguments]]
-#### Positional Command-Line Arguments
+==== Positional Command-Line Arguments
 
 _backup_id_::
   Specifies the backup that you want to monitor by seeing the progress information. The backupId
is case-sensitive.
 
 [[br.history.named.cli.arguments]]
-#### Named Command-Line Arguments
+==== Named Command-Line Arguments
 
 _-n <num_records>_::
   (Optional) The maximum number of backup records (Default: 10).
@@ -510,7 +510,7 @@ _-t_ <table_name>::
   The name of table to obtain history for. Mutually exclusive with the _-s_ option.
 
 [[br.history.backup.example]]
-#### Example usage
+==== Example usage
 
 [source]
 ----
@@ -522,7 +522,7 @@ $ hbase backup history -t WebIndexRecords
 // hbase backup describe
 
 [[br.describe.backup]]
-### Describing a Backup Image
+=== Describing a Backup Image
 
 This command can be used to obtain information about a specific backup image.
 
@@ -532,18 +532,18 @@ $ hbase backup describe <backup_id>
 ----
 
 [[br.describe.backup.positional.cli.arguments]]
-#### Positional Command-Line Arguments
+==== Positional Command-Line Arguments
 
 _backup_id_::
   The ID of the backup image to describe.
 
 [[br.describe.backup.named.cli.arguments]]
-#### Named Command-Line Arguments
+==== Named Command-Line Arguments
 
 None.
 
 [[br.describe.backup.example]]
-#### Example usage
+==== Example usage
 
 [source]
 ----
@@ -553,7 +553,7 @@ $ hbase backup describe backupId_1467823988425
 // hbase backup delete
 
 [[br.delete.backup]]
-### Deleting a Backup Image
+=== Deleting a Backup Image
 
 This command can be used to delete a backup image which is no longer needed.
 
@@ -563,18 +563,18 @@ $ hbase backup delete <backup_id>
 ----
 
 [[br.delete.backup.positional.cli.arguments]]
-#### Positional Command-Line Arguments
+==== Positional Command-Line Arguments
 
 _backup_id_::
   The ID to the backup image which should be deleted.
 
 [[br.delete.backup.named.cli.arguments]]
-#### Named Command-Line Arguments
+==== Named Command-Line Arguments
 
 None.
 
 [[br.delete.backup.example]]
-#### Example usage
+==== Example usage
 
 [source]
 ----
@@ -584,7 +584,7 @@ $ hbase backup delete backupId_1467823988425
 // hbase backup repair
 
 [[br.repair.backup]]
-### Backup Repair Command
+=== Backup Repair Command
 
 This command attempts to correct any inconsistencies in persisted backup metadata which exists
as
 the result of software errors or unhandled failure scenarios. While the backup implementation
tries
@@ -597,17 +597,17 @@ $ hbase backup repair
 ----
 
 [[br.repair.backup.positional.cli.arguments]]
-#### Positional Command-Line Arguments
+==== Positional Command-Line Arguments
 
 None.
 
 [[br.repair.backup.named.cli.arguments]]
-### Named Command-Line Arguments
+=== Named Command-Line Arguments
 
 None.
 
 [[br.repair.backup.example]]
-#### Example usage
+==== Example usage
 
 [source]
 ----
@@ -615,11 +615,11 @@ $ hbase backup repair
 ----
 
 [[br.backup.configuration]]
-## Configuration keys
+== Configuration keys
 
 The backup and restore feature includes both required and optional configuration keys.
 
-### Required properties
+=== Required properties
 
 _hbase.backup.enable_: Controls whether or not the feature is enabled (Default: `false`).
Set this value to `true`.
 
@@ -638,7 +638,7 @@ _hbase.coprocessor.region.classes_: A comma-separated list of RegionObservers
de
 _hbase.master.hfilecleaner.plugins_: A comma-separated list of HFileCleaners deployed on
the Master. Set this value
 to `org.apache.hadoop.hbase.backup.BackupHFileCleaner` or append it to the current value.
 
-### Optional properties
+=== Optional properties
 
 _hbase.backup.system.ttl_: The time-to-live in seconds of data in the `hbase:backup` tables
(default: forever). This property
 is only relevant prior to the creation of the `hbase:backup` table. Use the `alter` command
in the HBase shell to modify the TTL
@@ -653,9 +653,9 @@ _hbase.backup.logroll.timeout.millis_: The amount of time (in milliseconds)
to w
 in the Master's procedure framework (default: 30000).
 
 [[br.best.practices]]
-## Best Practices
+== Best Practices
 
-### Formulate a restore strategy and test it.
+=== Formulate a restore strategy and test it.
 
 Before you rely on a backup and restore strategy for your production environment, identify
how backups must be performed,
 and more importantly, how restores must be performed. Test the plan to ensure that it is
workable.
@@ -668,14 +668,14 @@ site renders locally stored backups useless. Consider storing the backup
data an
 and operator expertise) to restore the data at a site sufficiently remote from the production
site. In the case of a catastrophe
 at the whole primary site (fire, earthquake, etc.), the remote backup site can be very valuable.
 
-### Secure a full backup image first.
+=== Secure a full backup image first.
 
 As a baseline, you must complete a full backup of HBase data at least once before you can
rely on incremental backups. The full
 backup should be stored outside of the source cluster. To ensure complete dataset recovery,
you must run the restore utility
 with the option to restore baseline full backup. The full backup is the foundation of your
dataset. Incremental backup data
 is applied on top of the full backup during the restore operation to return you to the point
in time when backup was last taken.
 
-### Define and use backup sets for groups of tables that are logical subsets of the entire
dataset.
+=== Define and use backup sets for groups of tables that are logical subsets of the entire
dataset.
 
 You can group tables into an object called a backup set. A backup set can save time when
you have a particular group of tables
 that you expect to repeatedly back up or restore.
@@ -684,7 +684,7 @@ When you create a backup set, you type table names to include in the group.
The
 tables, but also retains the HBase backup metadata. Afterwards, you can invoke the backup
set name to indicate what tables apply
 to the command execution instead of entering all the table names individually.
 
-### Document the backup and restore strategy, and ideally log information about each backup.
+=== Document the backup and restore strategy, and ideally log information about each backup.
 
 Document the whole process so that the knowledge base can transfer to new administrators
after employee turnover. As an extra
 safety precaution, also log the calendar date, time, and other relevant details about the
data of each backup. This metadata
@@ -693,7 +693,7 @@ copies of all documentation: one copy at the production cluster site and
another
 accessed by an administrator remotely from the production cluster.
 
 [[br.s3.backup.scenario]]
-## Scenario: Safeguarding Application Datasets on Amazon S3
+== Scenario: Safeguarding Application Datasets on Amazon S3
 
 This scenario describes how a hypothetical retail business uses backups to safeguard application
data and then restore the dataset
 after failure.
@@ -760,7 +760,7 @@ existing data in the destination. In this case, the admin decides to overwrite
t
    s3a://$ACCESS_KEY:$SECRET_KEY@prodhbasebackups/backups backup_1467823988425 \ -overwrite
 
 [[br.data.security]]
-## Security of Backup Data
+== Security of Backup Data
 
 With this feature which makes copying data to remote locations, it's important to take a
moment to clearly state the procedural
 concerns that exist around data security. Like the HBase replication feature, backup and
restore provides the constructs to automatically
@@ -774,7 +774,7 @@ being accessed via HBase, and its authentication and authorization controls,
we
 providing a comparable level of security. This is a manual step which users *must* implement
on their own.
 
 [[br.technical.details]]
-## Technical Details of Incremental Backup and Restore
+== Technical Details of Incremental Backup and Restore
 
 HBase incremental backups enable more efficient capture of HBase table images than previous
attempts at serial backup and restore
 solutions, such as those that only used HBase Export and Import APIs. Incremental backups
use Write Ahead Logs (WALs) to capture
@@ -790,7 +790,7 @@ Bulk Load utility automatically imports as restored data in the table.
 You can only restore on a live HBase cluster because the data must be redistributed to complete
the restore operation successfully.
 
 [[br.filesystem.growth.warning]]
-## A Warning on File System Growth
+== A Warning on File System Growth
 
 As a reminder, incremental backups are implemented via retaining the write-ahead logs which
HBase primarily uses for data durability.
 Thus, to ensure that all data needing to be included in a backup is still available in the
system, the HBase backup and restore feature
@@ -806,14 +806,14 @@ more aggressive backup merges and deletions). As a reminder, the TTL
can be alte
 in the HBase shell. Modifying the configuration property `hbase.backup.system.ttl` in hbase-site.xml
after the system table exists has no effect.
 
 [[br.backup.capacity.planning]]
-## Capacity Planning
+== Capacity Planning
 
 When designing a distributed system deployment, it is critical that some basic mathmatical
rigor is executed to ensure sufficient computational
 capacity is available given the data and software requirements of the system. For this feature,
the availability of network capacity is the largest
 bottleneck when estimating the performance of some implementation of backup and restore.
The second most costly function is the speed at which
 data can be read/written.
 
-### Full Backups
+=== Full Backups
 
 To estimate the duration of a full backup, we have to understand the general actions which
are invoked:
 
@@ -840,7 +840,7 @@ queue which can limit the specific nodes where the workers will be spawned
-- th
 a set of non-critical nodes. Relating the `-b` and `-w` options to our earlier equations:
`-b` would be used to restrict each node from reading
 data at the full 80MB/s and `-w` is used to limit the job from spawning 16 worker tasks.
 
-### Incremental Backup
+=== Incremental Backup
 
 Like we did for full backups, we have to understand the incremental backup process to approximate
its runtime and cost.
 
@@ -854,7 +854,7 @@ this would require approximately 15 minutes to perform this step for 50GB
of dat
 DistCp MapReduce job would likely dominate the actual time taken to copy the data (50 / 1.25
= 40 seconds) and can be ignored.
 
 [[br.limitations]]
-## Limitations of the Backup and Restore Utility
+== Limitations of the Backup and Restore Utility
 
 *Serial backup operations*
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/826909a5/src/main/asciidoc/_chapters/compression.adoc
----------------------------------------------------------------------
diff --git a/src/main/asciidoc/_chapters/compression.adoc b/src/main/asciidoc/_chapters/compression.adoc
index 8fc1c55..67d10f5 100644
--- a/src/main/asciidoc/_chapters/compression.adoc
+++ b/src/main/asciidoc/_chapters/compression.adoc
@@ -335,25 +335,18 @@ You do not need to re-create the table or copy data.
 If you are changing codecs, be sure the old codec is still available until all the old StoreFiles
have been compacted.
 
 .Enabling Compression on a ColumnFamily of an Existing Table using HBaseShell
-====
 ----
-
 hbase> disable 'test'
 hbase> alter 'test', {NAME => 'cf', COMPRESSION => 'GZ'}
 hbase> enable 'test'
 ----
-====
 
 .Creating a New Table with Compression On a ColumnFamily
-====
 ----
-
 hbase> create 'test2', { NAME => 'cf2', COMPRESSION => 'SNAPPY' }
 ----
-====
 
 .Verifying a ColumnFamily's Compression Settings
-====
 ----
 
 hbase> describe 'test'
@@ -366,7 +359,6 @@ DESCRIPTION                                          ENABLED
  LOCKCACHE => 'true'}
 1 row(s) in 0.1070 seconds
 ----
-====
 
 ==== Testing Compression Performance
 
@@ -374,9 +366,7 @@ HBase includes a tool called LoadTestTool which provides mechanisms to
test your
 You must specify either `-write` or `-update-read` as your first parameter, and if you do
not specify another parameter, usage advice is printed for each option.
 
 .+LoadTestTool+ Usage
-====
 ----
-
 $ bin/hbase org.apache.hadoop.hbase.util.LoadTestTool -h
 usage: bin/hbase org.apache.hadoop.hbase.util.LoadTestTool <options>
 Options:
@@ -429,16 +419,12 @@ Options:
                               port numbers
  -zk_root <arg>               name of parent znode in zookeeper
 ----
-====
 
 .Example Usage of LoadTestTool
-====
 ----
-
 $ hbase org.apache.hadoop.hbase.util.LoadTestTool -write 1:10:100 -num_keys 1000000
           -read 100:30 -num_tables 1 -data_block_encoding NONE -tn load_test_tool_NONE
 ----
-====
 
 [[data.block.encoding.enable]]
 === Enable Data Block Encoding
@@ -449,9 +435,7 @@ Disable the table before altering its DATA_BLOCK_ENCODING setting.
 Following is an example using HBase Shell:
 
 .Enable Data Block Encoding On a Table
-====
 ----
-
 hbase>  disable 'test'
 hbase> alter 'test', { NAME => 'cf', DATA_BLOCK_ENCODING => 'FAST_DIFF' }
 Updating all regions with the new schema...
@@ -462,12 +446,9 @@ Done.
 hbase> enable 'test'
 0 row(s) in 0.1580 seconds
 ----
-====
 
 .Verifying a ColumnFamily's Data Block Encoding
-====
 ----
-
 hbase> describe 'test'
 DESCRIPTION                                          ENABLED
  'test', {NAME => 'cf', DATA_BLOCK_ENCODING => 'FAST true
@@ -478,7 +459,6 @@ DESCRIPTION                                          ENABLED
  e', BLOCKCACHE => 'true'}
 1 row(s) in 0.0650 seconds
 ----
-====
 
 :numbered:
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/826909a5/src/main/asciidoc/_chapters/getting_started.adoc
----------------------------------------------------------------------
diff --git a/src/main/asciidoc/_chapters/getting_started.adoc b/src/main/asciidoc/_chapters/getting_started.adoc
index 0edddfa..84ebcaa 100644
--- a/src/main/asciidoc/_chapters/getting_started.adoc
+++ b/src/main/asciidoc/_chapters/getting_started.adoc
@@ -604,18 +604,14 @@ On each node of the cluster, run the `jps` command and verify that the
correct p
 You may see additional Java processes running on your servers as well, if they are used for
other purposes.
 +
 .`node-a` `jps` Output
-====
 ----
-
 $ jps
 20355 Jps
 20071 HQuorumPeer
 20137 HMaster
 ----
-====
 +
 .`node-b` `jps` Output
-====
 ----
 $ jps
 15930 HRegionServer
@@ -623,17 +619,14 @@ $ jps
 15838 HQuorumPeer
 16010 HMaster
 ----
-====
 +
 .`node-c` `jps` Output
-====
 ----
 $ jps
 13901 Jps
 13639 HQuorumPeer
 13737 HRegionServer
 ----
-====
 +
 .ZooKeeper Process Name
 [NOTE]

http://git-wip-us.apache.org/repos/asf/hbase/blob/826909a5/src/main/asciidoc/_chapters/hbase_mob.adoc
----------------------------------------------------------------------
diff --git a/src/main/asciidoc/_chapters/hbase_mob.adoc b/src/main/asciidoc/_chapters/hbase_mob.adoc
index 9730529..8048772 100644
--- a/src/main/asciidoc/_chapters/hbase_mob.adoc
+++ b/src/main/asciidoc/_chapters/hbase_mob.adoc
@@ -61,12 +61,10 @@ an object is considered to be a MOB. Only `IS_MOB` is required. If you
do not
 specify the `MOB_THRESHOLD`, the default threshold value of 100 KB is used.
 
 .Configure a Column for MOB Using HBase Shell
-====
 ----
 hbase> create 't1', {NAME => 'f1', IS_MOB => true, MOB_THRESHOLD => 102400}
 hbase> alter 't1', {NAME => 'f1', IS_MOB => true, MOB_THRESHOLD => 102400}
 ----
-====
 
 .Configure a Column for MOB Using the Java API
 ====
@@ -91,7 +89,6 @@ weekly policy - compact MOB Files for one week into one large MOB file
 montly policy - compact MOB Files for one  month into one large MOB File
 
 .Configure MOB compaction policy Using HBase Shell
-====
 ----
 hbase> create 't1', {NAME => 'f1', IS_MOB => true, MOB_THRESHOLD => 102400, MOB_COMPACT_PARTITION_POLICY
=> 'daily'}
 hbase> create 't1', {NAME => 'f1', IS_MOB => true, MOB_THRESHOLD => 102400, MOB_COMPACT_PARTITION_POLICY
=> 'weekly'}
@@ -101,7 +98,6 @@ hbase> alter 't1', {NAME => 'f1', IS_MOB => true, MOB_THRESHOLD
=> 102400, MOB_C
 hbase> alter 't1', {NAME => 'f1', IS_MOB => true, MOB_THRESHOLD => 102400, MOB_COMPACT_PARTITION_POLICY
=> 'weekly'}
 hbase> alter 't1', {NAME => 'f1', IS_MOB => true, MOB_THRESHOLD => 102400, MOB_COMPACT_PARTITION_POLICY
=> 'monthly'}
 ----
-====
 
 === Configure MOB Compaction mergeable threshold
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/826909a5/src/main/asciidoc/_chapters/ops_mgt.adoc
----------------------------------------------------------------------
diff --git a/src/main/asciidoc/_chapters/ops_mgt.adoc b/src/main/asciidoc/_chapters/ops_mgt.adoc
index 38a7dff..0175ef7 100644
--- a/src/main/asciidoc/_chapters/ops_mgt.adoc
+++ b/src/main/asciidoc/_chapters/ops_mgt.adoc
@@ -1023,13 +1023,10 @@ The script requires you to set some environment variables before running
it.
 Examine the script and modify it to suit your needs.
 
 ._rolling-restart.sh_ General Usage
-====
 ----
-
 $ ./bin/rolling-restart.sh --help
 Usage: rolling-restart.sh [--config <hbase-confdir>] [--rs-only] [--master-only] [--graceful]
[--maxthreads xx]
 ----
-====
 
 Rolling Restart on RegionServers Only::
   To perform a rolling restart on the RegionServers only, use the `--rs-only` option.

http://git-wip-us.apache.org/repos/asf/hbase/blob/826909a5/src/main/asciidoc/_chapters/performance.adoc
----------------------------------------------------------------------
diff --git a/src/main/asciidoc/_chapters/performance.adoc b/src/main/asciidoc/_chapters/performance.adoc
index c917646..866779c 100644
--- a/src/main/asciidoc/_chapters/performance.adoc
+++ b/src/main/asciidoc/_chapters/performance.adoc
@@ -188,11 +188,9 @@ It is useful for tuning the IO impact of prefetching versus the time
before all
 To enable prefetching on a given column family, you can use HBase Shell or use the API.
 
 .Enable Prefetch Using HBase Shell
-====
 ----
 hbase> create 'MyTable', { NAME => 'myCF', PREFETCH_BLOCKS_ON_OPEN => 'true' }
 ----
-====
 
 .Enable Prefetch Using the API
 ====

http://git-wip-us.apache.org/repos/asf/hbase/blob/826909a5/src/main/asciidoc/_chapters/schema_design.adoc
----------------------------------------------------------------------
diff --git a/src/main/asciidoc/_chapters/schema_design.adoc b/src/main/asciidoc/_chapters/schema_design.adoc
index a25b85e..fdbd184 100644
--- a/src/main/asciidoc/_chapters/schema_design.adoc
+++ b/src/main/asciidoc/_chapters/schema_design.adoc
@@ -504,11 +504,9 @@ Deleted cells are still subject to TTL and there will never be more than
"maximu
 A new "raw" scan options returns all deleted rows and the delete markers.
 
 .Change the Value of `KEEP_DELETED_CELLS` Using HBase Shell
-====
 ----
 hbase> hbase> alter ‘t1′, NAME => ‘f1′, KEEP_DELETED_CELLS => true
 ----
-====
 
 .Change the Value of `KEEP_DELETED_CELLS` Using the API
 ====

http://git-wip-us.apache.org/repos/asf/hbase/blob/826909a5/src/main/asciidoc/_chapters/security.adoc
----------------------------------------------------------------------
diff --git a/src/main/asciidoc/_chapters/security.adoc b/src/main/asciidoc/_chapters/security.adoc
index ef7d6c4..c007053 100644
--- a/src/main/asciidoc/_chapters/security.adoc
+++ b/src/main/asciidoc/_chapters/security.adoc
@@ -1086,7 +1086,6 @@ public static void revokeFromTable(final HBaseTestingUtility util, final
String
 . Showing a User's Effective Permissions
 +
 .HBase Shell
-====
 ----
 hbase> user_permission 'user'
 
@@ -1094,7 +1093,6 @@ hbase> user_permission '.*'
 
 hbase> user_permission JAVA_REGEX
 ----
-====
 
 .API
 ====
@@ -1234,11 +1232,9 @@ Refer to the official API for usage instructions.
 . Define the List of Visibility Labels
 +
 .HBase Shell
-====
 ----
 hbase> add_labels [ 'admin', 'service', 'developer', 'test' ]
 ----
-====
 +
 .Java API
 ====
@@ -1265,7 +1261,6 @@ public static void addLabels() throws Exception {
 . Associate Labels with Users
 +
 .HBase Shell
-====
 ----
 hbase> set_auths 'service', [ 'service' ]
 ----
@@ -1281,7 +1276,6 @@ hbase> set_auths 'qa', [ 'test', 'developer' ]
 ----
 hbase> set_auths '@qagroup', [ 'test' ]
 ----
-====
 +
 .Java API
 ====
@@ -1305,7 +1299,6 @@ public void testSetAndGetUserAuths() throws Throwable {
 . Clear Labels From Users
 +
 .HBase Shell
-====
 ----
 hbase> clear_auths 'service', [ 'service' ]
 ----
@@ -1321,7 +1314,6 @@ hbase> clear_auths 'qa', [ 'test', 'developer' ]
 ----
 hbase> clear_auths '@qagroup', [ 'test', 'developer' ]
 ----
-====
 +
 .Java API
 ====
@@ -1345,7 +1337,6 @@ The label is only applied when data is written.
 The label is associated with a given version of the cell.
 +
 .HBase Shell
-====
 ----
 hbase> set_visibility 'user', 'admin|service|developer', { COLUMNS => 'i' }
 ----
@@ -1357,7 +1348,6 @@ hbase> set_visibility 'user', 'admin|service', { COLUMNS => 'pii'
}
 ----
 hbase> set_visibility 'user', 'test', { COLUMNS => [ 'i', 'pii' ], FILTER => "(PrefixFilter
('test'))" }
 ----
-====
 +
 NOTE: HBase Shell support for applying labels or permissions to cells is for testing and
verification support, and should not be employed for production use because it won't apply
the labels to cells that don't exist yet.
 The correct way to apply cell level labels is to do so in the application code when storing
the values.
@@ -1408,12 +1398,10 @@ set as an additional filter. It will further filter your results,
rather than
 giving you additional authorization.
 
 .HBase Shell
-====
 ----
 hbase> get_auths 'myUser'
 hbase> scan 'table1', AUTHORIZATIONS => ['private']
 ----
-====
 
 .Java API
 ====

http://git-wip-us.apache.org/repos/asf/hbase/blob/826909a5/src/main/asciidoc/_chapters/shell.adoc
----------------------------------------------------------------------
diff --git a/src/main/asciidoc/_chapters/shell.adoc b/src/main/asciidoc/_chapters/shell.adoc
index b246ab1..5612e1d 100644
--- a/src/main/asciidoc/_chapters/shell.adoc
+++ b/src/main/asciidoc/_chapters/shell.adoc
@@ -145,7 +145,6 @@ For instance, if your script creates a table, but returns a non-zero exit
value,
 You can enter HBase Shell commands into a text file, one command per line, and pass that
file to the HBase Shell.
 
 .Example Command File
-====
 ----
 create 'test', 'cf'
 list 'test'
@@ -158,7 +157,6 @@ get 'test', 'row1'
 disable 'test'
 enable 'test'
 ----
-====
 
 .Directing HBase Shell to Execute the Commands
 ====


Mime
View raw message