camel-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From davscl...@apache.org
Subject [06/20] camel git commit: Delete all the old karaf osgi tests that do not work. We should use the new camel-test-karaf module as base for testing.
Date Mon, 18 Apr 2016 11:40:12 GMT
http://git-wip-us.apache.org/repos/asf/camel/blob/6b77d012/tests/camel-itest-osgi/src/test/resources/hdfs-default.xml
----------------------------------------------------------------------
diff --git a/tests/camel-itest-osgi/src/test/resources/hdfs-default.xml b/tests/camel-itest-osgi/src/test/resources/hdfs-default.xml
deleted file mode 100644
index 24f0b03..0000000
--- a/tests/camel-itest-osgi/src/test/resources/hdfs-default.xml
+++ /dev/null
@@ -1,1607 +0,0 @@
-<?xml version="1.0"?>
-<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
-
-<!--
-   Licensed to the Apache Software Foundation (ASF) under one or more
-   contributor license agreements.  See the NOTICE file distributed with
-   this work for additional information regarding copyright ownership.
-   The ASF licenses this file to You under the Apache License, Version 2.0
-   (the "License"); you may not use this file except in compliance with
-   the License.  You may obtain a copy of the License at
-
-       http://www.apache.org/licenses/LICENSE-2.0
-
-   Unless required by applicable law or agreed to in writing, software
-   distributed under the License is distributed on an "AS IS" BASIS,
-   WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-   See the License for the specific language governing permissions and
-   limitations under the License.
--->
-
-<!-- Do not modify this file directly.  Instead, copy entries that you -->
-<!-- wish to modify from this file into hdfs-site.xml and change them -->
-<!-- there.  If hdfs-site.xml does not already exist, create it.      -->
-
-<configuration>
-
-<property>
-  <name>hadoop.hdfs.configuration.version</name>
-  <value>1</value>
-  <description>version of this configuration file</description>
-</property>
-
-<property>
-  <name>dfs.namenode.logging.level</name>
-  <value>info</value>
-  <description>
-    The logging level for dfs namenode. Other values are "dir" (trace
-    namespace mutations), "block" (trace block under/over replications
-    and block creations/deletions), or "all".
-  </description>
-</property>
-
-<property>
-  <name>dfs.namenode.rpc-address</name>
-  <value></value>
-  <description>
-    RPC address that handles all clients requests. In the case of HA/Federation where multiple namenodes exist,
-    the name service id is added to the name e.g. dfs.namenode.rpc-address.ns1
-    dfs.namenode.rpc-address.EXAMPLENAMESERVICE
-    The value of this property will take the form of nn-host1:rpc-port.
-  </description>
-</property>
-
-<property>
-  <name>dfs.namenode.rpc-bind-host</name>
-  <value></value>
-  <description>
-    The actual address the server will bind to. If this optional address is
-    set, the RPC server will bind to this address and the port specified in
-    dfs.namenode.rpc-address for the RPC server. It can also be specified
-    per name node or name service for HA/Federation. This is most useful for
-    making name node listen to all interfaces by setting to 0.0.0.0.
-  </description>
-</property>
-
-<property>
-  <name>dfs.namenode.servicerpc-address</name>
-  <value></value>
-  <description>
-    RPC address for HDFS Services communication. BackupNode, Datanodes and all other services should be
-    connecting to this address if it is configured. In the case of HA/Federation where multiple namenodes exist,
-    the name service id is added to the name e.g. dfs.namenode.servicerpc-address.ns1
-    dfs.namenode.rpc-address.EXAMPLENAMESERVICE
-    The value of this property will take the form of nn-host1:rpc-port.
-    If the value of this property is unset the value of dfs.namenode.rpc-address will be used as the default.
-  </description>
-</property>
-
-<property>
-  <name>dfs.namenode.servicerpc-bind-host</name>
-  <value></value>
-  <description>
-    The actual address the server will bind to. If this optional address is
-    set, the service RPC server will bind to this address and the port 
-    specified in dfs.namenode.servicerpc-address. It can also be specified
-    per name node or name service for HA/Federation. This is most useful for
-    making name node listen to all interfaces by setting to 0.0.0.0.
-  </description>
-</property>
-
-<property>
-  <name>dfs.namenode.secondary.http-address</name>
-  <value>0.0.0.0:50090</value>
-  <description>
-    The secondary namenode http server address and port.
-  </description>
-</property>
-
-<property>
-  <name>dfs.datanode.address</name>
-  <value>0.0.0.0:50010</value>
-  <description>
-    The datanode server address and port for data transfer.
-  </description>
-</property>
-
-<property>
-  <name>dfs.datanode.http.address</name>
-  <value>0.0.0.0:50075</value>
-  <description>
-    The datanode http server address and port.
-  </description>
-</property>
-
-<property>
-  <name>dfs.datanode.ipc.address</name>
-  <value>0.0.0.0:50020</value>
-  <description>
-    The datanode ipc server address and port.
-  </description>
-</property>
-
-<property>
-  <name>dfs.datanode.handler.count</name>
-  <value>10</value>
-  <description>The number of server threads for the datanode.</description>
-</property>
-
-<property>
-  <name>dfs.namenode.http-address</name>
-  <value>0.0.0.0:50070</value>
-  <description>
-    The address and the base port where the dfs namenode web ui will listen on.
-  </description>
-</property>
-
-<property>
-  <name>dfs.https.enable</name>
-  <value>false</value>
-  <description>
-    Deprecated. Use "dfs.http.policy" instead.
-  </description>
-</property>
-
-<property>
-  <name>dfs.http.policy</name>
-  <value>HTTP_ONLY</value>
-  <description>Decide if HTTPS(SSL) is supported on HDFS
-    This configures the HTTP endpoint for HDFS daemons:
-      The following values are supported:
-      - HTTP_ONLY : Service is provided only on http
-      - HTTPS_ONLY : Service is provided only on https
-      - HTTP_AND_HTTPS : Service is provided both on http and https
-  </description>
-</property>
-
-<property>
-  <name>dfs.client.https.need-auth</name>
-  <value>false</value>
-  <description>Whether SSL client certificate authentication is required
-  </description>
-</property>
-
-<property>
-  <name>dfs.https.server.keystore.resource</name>
-  <value>ssl-server.xml</value>
-  <description>Resource file from which ssl server keystore
-  information will be extracted
-  </description>
-</property>
-
-<property>
-  <name>dfs.client.https.keystore.resource</name>
-  <value>ssl-client.xml</value>
-  <description>Resource file from which ssl client keystore
-  information will be extracted
-  </description>
-</property>
-
-<property>
-  <name>dfs.datanode.https.address</name>
-  <value>0.0.0.0:50475</value>
-  <description>The datanode secure http server address and port.</description>
-</property>
-
-<property>
-  <name>dfs.namenode.https-address</name>
-  <value>0.0.0.0:50470</value>
-  <description>The namenode secure http server address and port.</description>
-</property>
-
- <property>
-  <name>dfs.datanode.dns.interface</name>
-  <value>default</value>
-  <description>The name of the Network Interface from which a data node should 
-  report its IP address.
-  </description>
- </property>
- 
-<property>
-  <name>dfs.datanode.dns.nameserver</name>
-  <value>default</value>
-  <description>The host name or IP address of the name server (DNS)
-  which a DataNode should use to determine the host name used by the
-  NameNode for communication and display purposes.
-  </description>
- </property>
- 
- <property>
-  <name>dfs.namenode.backup.address</name>
-  <value>0.0.0.0:50100</value>
-  <description>
-    The backup node server address and port.
-    If the port is 0 then the server will start on a free port.
-  </description>
-</property>
- 
- <property>
-  <name>dfs.namenode.backup.http-address</name>
-  <value>0.0.0.0:50105</value>
-  <description>
-    The backup node http server address and port.
-    If the port is 0 then the server will start on a free port.
-  </description>
-</property>
-
-<property>
-  <name>dfs.namenode.replication.considerLoad</name>
-  <value>true</value>
-  <description>Decide if chooseTarget considers the target's load or not
-  </description>
-</property>
-<property>
-  <name>dfs.default.chunk.view.size</name>
-  <value>32768</value>
-  <description>The number of bytes to view for a file on the browser.
-  </description>
-</property>
-
-<property>
-  <name>dfs.datanode.du.reserved</name>
-  <value>0</value>
-  <description>Reserved space in bytes per volume. Always leave this much space free for non dfs use.
-  </description>
-</property>
-
-<property>
-  <name>dfs.namenode.name.dir</name>
-  <value>file://${hadoop.tmp.dir}/dfs/name</value>
-  <description>Determines where on the local filesystem the DFS name node
-      should store the name table(fsimage).  If this is a comma-delimited list
-      of directories then the name table is replicated in all of the
-      directories, for redundancy. </description>
-</property>
-
-<property>
-  <name>dfs.namenode.name.dir.restore</name>
-  <value>false</value>
-  <description>Set to true to enable NameNode to attempt recovering a
-      previously failed dfs.namenode.name.dir. When enabled, a recovery of any
-      failed directory is attempted during checkpoint.</description>
-</property>
-
-<property>
-  <name>dfs.namenode.fs-limits.max-component-length</name>
-  <value>0</value>
-  <description>Defines the maximum number of characters in each component
-      of a path.  A value of 0 will disable the check.</description>
-</property>
-
-<property>
-  <name>dfs.namenode.fs-limits.max-directory-items</name>
-  <value>0</value>
-  <description>Defines the maximum number of items that a directory may
-      contain.  A value of 0 will disable the check.</description>
-</property>
-
-<property>
-  <name>dfs.namenode.fs-limits.min-block-size</name>
-  <value>1048576</value>
-  <description>Minimum block size in bytes, enforced by the Namenode at create
-      time. This prevents the accidental creation of files with tiny block
-      sizes (and thus many blocks), which can degrade
-      performance.</description>
-</property>
-
-<property>
-    <name>dfs.namenode.fs-limits.max-blocks-per-file</name>
-    <value>1048576</value>
-    <description>Maximum number of blocks per file, enforced by the Namenode on
-        write. This prevents the creation of extremely large files which can
-        degrade performance.</description>
-</property>
-
-<property>
-  <name>dfs.namenode.edits.dir</name>
-  <value>${dfs.namenode.name.dir}</value>
-  <description>Determines where on the local filesystem the DFS name node
-      should store the transaction (edits) file. If this is a comma-delimited list
-      of directories then the transaction file is replicated in all of the 
-      directories, for redundancy. Default value is same as dfs.namenode.name.dir
-  </description>
-</property>
-
-<property>
-  <name>dfs.namenode.shared.edits.dir</name>
-  <value></value>
-  <description>A directory on shared storage between the multiple namenodes
-  in an HA cluster. This directory will be written by the active and read
-  by the standby in order to keep the namespaces synchronized. This directory
-  does not need to be listed in dfs.namenode.edits.dir above. It should be
-  left empty in a non-HA cluster.
-  </description>
-</property>
-
-<property>
-  <name>dfs.namenode.edits.journal-plugin.qjournal</name>
-  <value>org.apache.hadoop.hdfs.qjournal.client.QuorumJournalManager</value>
-</property>
-
-<property>
-  <name>dfs.permissions.enabled</name>
-  <value>true</value>
-  <description>
-    If "true", enable permission checking in HDFS.
-    If "false", permission checking is turned off,
-    but all other behavior is unchanged.
-    Switching from one parameter value to the other does not change the mode,
-    owner or group of files or directories.
-  </description>
-</property>
-
-<property>
-  <name>dfs.permissions.superusergroup</name>
-  <value>supergroup</value>
-  <description>The name of the group of super-users.</description>
-</property>
-<!--
-<property>
-   <name>dfs.cluster.administrators</name>
-   <value>ACL for the admins</value>
-   <description>This configuration is used to control who can access the
-                default servlets in the namenode, etc.
-   </description>
-</property>
--->
-
-<property>
-  <name>dfs.block.access.token.enable</name>
-  <value>false</value>
-  <description>
-    If "true", access tokens are used as capabilities for accessing datanodes.
-    If "false", no access tokens are checked on accessing datanodes.
-  </description>
-</property>
-
-<property>
-  <name>dfs.block.access.key.update.interval</name>
-  <value>600</value>
-  <description>
-    Interval in minutes at which namenode updates its access keys.
-  </description>
-</property>
-
-<property>
-  <name>dfs.block.access.token.lifetime</name>
-  <value>600</value>
-  <description>The lifetime of access tokens in minutes.</description>
-</property>
-
-<property>
-  <name>dfs.datanode.data.dir</name>
-  <value>file://${hadoop.tmp.dir}/dfs/data</value>
-  <description>Determines where on the local filesystem an DFS data node
-  should store its blocks.  If this is a comma-delimited
-  list of directories, then data will be stored in all named
-  directories, typically on different devices.
-  Directories that do not exist are ignored.
-  </description>
-</property>
-
-<property>
-  <name>dfs.datanode.data.dir.perm</name>
-  <value>700</value>
-  <description>Permissions for the directories on on the local filesystem where
-  the DFS data node store its blocks. The permissions can either be octal or
-  symbolic.</description>
-</property>
-
-<property>
-  <name>dfs.replication</name>
-  <value>3</value>
-  <description>Default block replication. 
-  The actual number of replications can be specified when the file is created.
-  The default is used if replication is not specified in create time.
-  </description>
-</property>
-
-<property>
-  <name>dfs.replication.max</name>
-  <value>512</value>
-  <description>Maximal block replication. 
-  </description>
-</property>
-
-<property>
-  <name>dfs.namenode.replication.min</name>
-  <value>1</value>
-  <description>Minimal block replication. 
-  </description>
-</property>
-
-<property>
-  <name>dfs.blocksize</name>
-  <value>134217728</value>
-  <description>
-      The default block size for new files, in bytes.
-      You can use the following suffix (case insensitive):
-      k(kilo), m(mega), g(giga), t(tera), p(peta), e(exa) to specify the size (such as 128k, 512m, 1g, etc.),
-      Or provide complete size in bytes (such as 134217728 for 128 MB).
-  </description>
-</property>
-
-<property>
-  <name>dfs.client.block.write.retries</name>
-  <value>3</value>
-  <description>The number of retries for writing blocks to the data nodes, 
-  before we signal failure to the application.
-  </description>
-</property>
-
-<property>
-  <name>dfs.client.block.write.replace-datanode-on-failure.enable</name>
-  <value>true</value>
-  <description>
-    If there is a datanode/network failure in the write pipeline,
-    DFSClient will try to remove the failed datanode from the pipeline
-    and then continue writing with the remaining datanodes. As a result,
-    the number of datanodes in the pipeline is decreased.  The feature is
-    to add new datanodes to the pipeline.
-
-    This is a site-wide property to enable/disable the feature.
-
-    When the cluster size is extremely small, e.g. 3 nodes or less, cluster
-    administrators may want to set the policy to NEVER in the default
-    configuration file or disable this feature.  Otherwise, users may
-    experience an unusually high rate of pipeline failures since it is
-    impossible to find new datanodes for replacement.
-
-    See also dfs.client.block.write.replace-datanode-on-failure.policy
-  </description>
-</property>
-
-<property>
-  <name>dfs.client.block.write.replace-datanode-on-failure.policy</name>
-  <value>DEFAULT</value>
-  <description>
-    This property is used only if the value of
-    dfs.client.block.write.replace-datanode-on-failure.enable is true.
-
-    ALWAYS: always add a new datanode when an existing datanode is removed.
-    
-    NEVER: never add a new datanode.
-
-    DEFAULT: 
-      Let r be the replication number.
-      Let n be the number of existing datanodes.
-      Add a new datanode only if r is greater than or equal to 3 and either
-      (1) floor(r/2) is greater than or equal to n; or
-      (2) r is greater than n and the block is hflushed/appended.
-  </description>
-</property>
-
-<property>
-  <name>dfs.blockreport.intervalMsec</name>
-  <value>21600000</value>
-  <description>Determines block reporting interval in milliseconds.</description>
-</property>
-
-<property>
-  <name>dfs.blockreport.initialDelay</name>  <value>0</value>
-  <description>Delay for first block report in seconds.</description>
-</property>
-
-<property>
-  <name>dfs.datanode.directoryscan.interval</name>
-  <value>21600</value>
-  <description>Interval in seconds for Datanode to scan data directories and
-  reconcile the difference between blocks in memory and on the disk.
-  </description>
-</property>
-
-<property>
-  <name>dfs.datanode.directoryscan.threads</name>
-  <value>1</value>
-  <description>How many threads should the threadpool used to compile reports
-  for volumes in parallel have.
-  </description>
-</property>
-
-<property>
-  <name>dfs.heartbeat.interval</name>
-  <value>3</value>
-  <description>Determines datanode heartbeat interval in seconds.</description>
-</property>
-
-<property>
-  <name>dfs.namenode.handler.count</name>
-  <value>10</value>
-  <description>The number of server threads for the namenode.</description>
-</property>
-
-<property>
-  <name>dfs.namenode.safemode.threshold-pct</name>
-  <value>0.999f</value>
-  <description>
-    Specifies the percentage of blocks that should satisfy 
-    the minimal replication requirement defined by dfs.namenode.replication.min.
-    Values less than or equal to 0 mean not to wait for any particular
-    percentage of blocks before exiting safemode.
-    Values greater than 1 will make safe mode permanent.
-  </description>
-</property>
-
-<property>
-  <name>dfs.namenode.safemode.min.datanodes</name>
-  <value>0</value>
-  <description>
-    Specifies the number of datanodes that must be considered alive
-    before the name node exits safemode.
-    Values less than or equal to 0 mean not to take the number of live
-    datanodes into account when deciding whether to remain in safe mode
-    during startup.
-    Values greater than the number of datanodes in the cluster
-    will make safe mode permanent.
-  </description>
-</property>
-
-<property>
-  <name>dfs.namenode.safemode.extension</name>
-  <value>30000</value>
-  <description>
-    Determines extension of safe mode in milliseconds 
-    after the threshold level is reached.
-  </description>
-</property>
-
-<property>
-  <name>dfs.datanode.balance.bandwidthPerSec</name>
-  <value>1048576</value>
-  <description>
-        Specifies the maximum amount of bandwidth that each datanode
-        can utilize for the balancing purpose in term of
-        the number of bytes per second.
-  </description>
-</property>
-
-<property>
-  <name>dfs.hosts</name>
-  <value></value>
-  <description>Names a file that contains a list of hosts that are
-  permitted to connect to the namenode. The full pathname of the file
-  must be specified.  If the value is empty, all hosts are
-  permitted.</description>
-</property>
-
-<property>
-  <name>dfs.hosts.exclude</name>
-  <value></value>
-  <description>Names a file that contains a list of hosts that are
-  not permitted to connect to the namenode.  The full pathname of the
-  file must be specified.  If the value is empty, no hosts are
-  excluded.</description>
-</property> 
-
-<property>
-  <name>dfs.namenode.max.objects</name>
-  <value>0</value>
-  <description>The maximum number of files, directories and blocks
-  dfs supports. A value of zero indicates no limit to the number
-  of objects that dfs supports.
-  </description>
-</property>
-
-<property>
-  <name>dfs.namenode.datanode.registration.ip-hostname-check</name>
-  <value>true</value>
-  <description>
-    If true (the default), then the namenode requires that a connecting
-    datanode's address must be resolved to a hostname.  If necessary, a reverse
-    DNS lookup is performed.  All attempts to register a datanode from an
-    unresolvable address are rejected.
-
-    It is recommended that this setting be left on to prevent accidental
-    registration of datanodes listed by hostname in the excludes file during a
-    DNS outage.  Only set this to false in environments where there is no
-    infrastructure to support reverse DNS lookup.
-  </description>
-</property>
-
-<property>
-  <name>dfs.namenode.decommission.interval</name>
-  <value>30</value>
-  <description>Namenode periodicity in seconds to check if decommission is 
-  complete.</description>
-</property>
-
-<property>
-  <name>dfs.namenode.decommission.nodes.per.interval</name>
-  <value>5</value>
-  <description>The number of nodes namenode checks if decommission is complete
-  in each dfs.namenode.decommission.interval.</description>
-</property>
-
-<property>
-  <name>dfs.namenode.replication.interval</name>
-  <value>3</value>
-  <description>The periodicity in seconds with which the namenode computes 
-  repliaction work for datanodes. </description>
-</property>
-
-<property>
-  <name>dfs.namenode.accesstime.precision</name>
-  <value>3600000</value>
-  <description>The access time for HDFS file is precise upto this value. 
-               The default value is 1 hour. Setting a value of 0 disables
-               access times for HDFS.
-  </description>
-</property>
-
-<property>
-  <name>dfs.datanode.plugins</name>
-  <value></value>
-  <description>Comma-separated list of datanode plug-ins to be activated.
-  </description>
-</property>
-
-<property>
-  <name>dfs.namenode.plugins</name>
-  <value></value>
-  <description>Comma-separated list of namenode plug-ins to be activated.
-  </description>
-</property>
-
-<property>
-  <name>dfs.stream-buffer-size</name>
-  <value>4096</value>
-  <description>The size of buffer to stream files.
-  The size of this buffer should probably be a multiple of hardware
-  page size (4096 on Intel x86), and it determines how much data is
-  buffered during read and write operations.</description>
-</property>
-
-<property>
-  <name>dfs.bytes-per-checksum</name>
-  <value>512</value>
-  <description>The number of bytes per checksum.  Must not be larger than
-  dfs.stream-buffer-size</description>
-</property>
-
-<property>
-  <name>dfs.client-write-packet-size</name>
-  <value>65536</value>
-  <description>Packet size for clients to write</description>
-</property>
-
-<property>
-  <name>dfs.client.write.exclude.nodes.cache.expiry.interval.millis</name>
-  <value>600000</value>
-  <description>The maximum period to keep a DN in the excluded nodes list
-  at a client. After this period, in milliseconds, the previously excluded node(s) will
-  be removed automatically from the cache and will be considered good for block allocations
-  again. Useful to lower or raise in situations where you keep a file open for very long
-  periods (such as a Write-Ahead-Log (WAL) file) to make the writer tolerant to cluster maintenance
-  restarts. Defaults to 10 minutes.</description>
-</property>
-
-<property>
-  <name>dfs.namenode.checkpoint.dir</name>
-  <value>file://${hadoop.tmp.dir}/dfs/namesecondary</value>
-  <description>Determines where on the local filesystem the DFS secondary
-      name node should store the temporary images to merge.
-      If this is a comma-delimited list of directories then the image is
-      replicated in all of the directories for redundancy.
-  </description>
-</property>
-
-<property>
-  <name>dfs.namenode.checkpoint.edits.dir</name>
-  <value>${dfs.namenode.checkpoint.dir}</value>
-  <description>Determines where on the local filesystem the DFS secondary
-      name node should store the temporary edits to merge.
-      If this is a comma-delimited list of directoires then teh edits is
-      replicated in all of the directoires for redundancy.
-      Default value is same as dfs.namenode.checkpoint.dir
-  </description>
-</property>
-
-<property>
-  <name>dfs.namenode.checkpoint.period</name>
-  <value>3600</value>
-  <description>The number of seconds between two periodic checkpoints.
-  </description>
-</property>
-
-<property>
-  <name>dfs.namenode.checkpoint.txns</name>
-  <value>1000000</value>
-  <description>The Secondary NameNode or CheckpointNode will create a checkpoint
-  of the namespace every 'dfs.namenode.checkpoint.txns' transactions, regardless
-  of whether 'dfs.namenode.checkpoint.period' has expired.
-  </description>
-</property>
-
-<property>
-  <name>dfs.namenode.checkpoint.check.period</name>
-  <value>60</value>
-  <description>The SecondaryNameNode and CheckpointNode will poll the NameNode
-  every 'dfs.namenode.checkpoint.check.period' seconds to query the number
-  of uncheckpointed transactions.
-  </description>
-</property>
-
-<property>
-  <name>dfs.namenode.checkpoint.max-retries</name>
-  <value>3</value>
-  <description>The SecondaryNameNode retries failed checkpointing. If the 
-  failure occurs while loading fsimage or replaying edits, the number of
-  retries is limited by this variable. 
-  </description>
-</property>
-
-<property>
-  <name>dfs.namenode.num.checkpoints.retained</name>
-  <value>2</value>
-  <description>The number of image checkpoint files that will be retained by
-  the NameNode and Secondary NameNode in their storage directories. All edit
-  logs necessary to recover an up-to-date namespace from the oldest retained
-  checkpoint will also be retained.
-  </description>
-</property>
-
-<property>
-  <name>dfs.namenode.num.extra.edits.retained</name>
-  <value>1000000</value>
-  <description>The number of extra transactions which should be retained
-  beyond what is minimally necessary for a NN restart. This can be useful for
-  audit purposes or for an HA setup where a remote Standby Node may have
-  been offline for some time and need to have a longer backlog of retained
-  edits in order to start again.
-  Typically each edit is on the order of a few hundred bytes, so the default
-  of 1 million edits should be on the order of hundreds of MBs or low GBs.
-
-  NOTE: Fewer extra edits may be retained than value specified for this setting
-  if doing so would mean that more segments would be retained than the number
-  configured by dfs.namenode.max.extra.edits.segments.retained.
-  </description>
-</property>
-
-<property>
-  <name>dfs.namenode.max.extra.edits.segments.retained</name>
-  <value>10000</value>
-  <description>The maximum number of extra edit log segments which should be retained
-  beyond what is minimally necessary for a NN restart. When used in conjunction with
-  dfs.namenode.num.extra.edits.retained, this configuration property serves to cap
-  the number of extra edits files to a reasonable value.
-  </description>
-</property>
-
-<property>
-  <name>dfs.namenode.delegation.key.update-interval</name>
-  <value>86400000</value>
-  <description>The update interval for master key for delegation tokens 
-       in the namenode in milliseconds.
-  </description>
-</property>
-
-<property>
-  <name>dfs.namenode.delegation.token.max-lifetime</name>
-  <value>604800000</value>
-  <description>The maximum lifetime in milliseconds for which a delegation 
-      token is valid.
-  </description>
-</property>
-
-<property>
-  <name>dfs.namenode.delegation.token.renew-interval</name>
-  <value>86400000</value>
-  <description>The renewal interval for delegation token in milliseconds.
-  </description>
-</property>
-
-<property>
-  <name>dfs.datanode.failed.volumes.tolerated</name>
-  <value>0</value>
-  <description>The number of volumes that are allowed to
-  fail before a datanode stops offering service. By default
-  any volume failure will cause a datanode to shutdown.
-  </description>
-</property>
-
-<property>
-  <name>dfs.image.compress</name>
-  <value>false</value>
-  <description>Should the dfs image be compressed?
-  </description>
-</property>
-
-<property>
-  <name>dfs.image.compression.codec</name>
-  <value>org.apache.hadoop.io.compress.DefaultCodec</value>
-  <description>If the dfs image is compressed, how should they be compressed?
-               This has to be a codec defined in io.compression.codecs.
-  </description>
-</property>
-
-<property>
-  <name>dfs.image.transfer.timeout</name>
-  <value>600000</value>
-  <description>
-        Timeout for image transfer in milliseconds. This timeout and the related
-        dfs.image.transfer.bandwidthPerSec parameter should be configured such
-        that normal image transfer can complete within the timeout.
-        This timeout prevents client hangs when the sender fails during
-        image transfer, which is particularly important during checkpointing.
-        Note that this timeout applies to the entirety of image transfer, and
-        is not a socket timeout.
-  </description>
-</property>
-
-<property>
-  <name>dfs.image.transfer.bandwidthPerSec</name>
-  <value>0</value>
-  <description>
-        Maximum bandwidth used for image transfer in bytes per second.
-        This can help keep normal namenode operations responsive during
-        checkpointing. The maximum bandwidth and timeout in
-        dfs.image.transfer.timeout should be set such that normal image
-        transfers can complete successfully.
-        A default value of 0 indicates that throttling is disabled. 
-  </description>
-</property>
-
-<property>
-  <name>dfs.namenode.support.allow.format</name>
-  <value>true</value>
-  <description>Does HDFS namenode allow itself to be formatted?
-               You may consider setting this to false for any production
-               cluster, to avoid any possibility of formatting a running DFS.
-  </description>
-</property>
-
-<property>
-  <name>dfs.datanode.max.transfer.threads</name>
-  <value>4096</value>
-  <description>
-        Specifies the maximum number of threads to use for transferring data
-        in and out of the DN.
-  </description>
-</property>
-
-<property>
-  <name>dfs.datanode.readahead.bytes</name>
-  <value>4193404</value>
-  <description>
-        While reading block files, if the Hadoop native libraries are available,
-        the datanode can use the posix_fadvise system call to explicitly
-        page data into the operating system buffer cache ahead of the current
-        reader's position. This can improve performance especially when
-        disks are highly contended.
-
-        This configuration specifies the number of bytes ahead of the current
-        read position which the datanode will attempt to read ahead. This
-        feature may be disabled by configuring this property to 0.
-
-        If the native libraries are not available, this configuration has no
-        effect.
-  </description>
-</property>
-
-<property>
-  <name>dfs.datanode.drop.cache.behind.reads</name>
-  <value>false</value>
-  <description>
-        In some workloads, the data read from HDFS is known to be significantly
-        large enough that it is unlikely to be useful to cache it in the
-        operating system buffer cache. In this case, the DataNode may be
-        configured to automatically purge all data from the buffer cache
-        after it is delivered to the client. This behavior is automatically
-        disabled for workloads which read only short sections of a block
-        (e.g HBase random-IO workloads).
-
-        This may improve performance for some workloads by freeing buffer
-        cache spage usage for more cacheable data.
-
-        If the Hadoop native libraries are not available, this configuration
-        has no effect.
-  </description>
-</property>
-
-<property>
-  <name>dfs.datanode.drop.cache.behind.writes</name>
-  <value>false</value>
-  <description>
-        In some workloads, the data written to HDFS is known to be significantly
-        large enough that it is unlikely to be useful to cache it in the
-        operating system buffer cache. In this case, the DataNode may be
-        configured to automatically purge all data from the buffer cache
-        after it is written to disk.
-
-        This may improve performance for some workloads by freeing buffer
-        cache spage usage for more cacheable data.
-
-        If the Hadoop native libraries are not available, this configuration
-        has no effect.
-  </description>
-</property>
-
-<property>
-  <name>dfs.datanode.sync.behind.writes</name>
-  <value>false</value>
-  <description>
-        If this configuration is enabled, the datanode will instruct the
-        operating system to enqueue all written data to the disk immediately
-        after it is written. This differs from the usual OS policy which
-        may wait for up to 30 seconds before triggering writeback.
-
-        This may improve performance for some workloads by smoothing the
-        IO profile for data written to disk.
-
-        If the Hadoop native libraries are not available, this configuration
-        has no effect.
-  </description>
-</property>
-
-<property>
-  <name>dfs.client.failover.max.attempts</name>
-  <value>15</value>
-  <description>
-    Expert only. The number of client failover attempts that should be
-    made before the failover is considered failed.
-  </description>
-</property>
-
-<property>
-  <name>dfs.client.failover.sleep.base.millis</name>
-  <value>500</value>
-  <description>
-    Expert only. The time to wait, in milliseconds, between failover
-    attempts increases exponentially as a function of the number of
-    attempts made so far, with a random factor of +/- 50%. This option
-    specifies the base value used in the failover calculation. The
-    first failover will retry immediately. The 2nd failover attempt
-    will delay at least dfs.client.failover.sleep.base.millis
-    milliseconds. And so on.
-  </description>
-</property>
-
-<property>
-  <name>dfs.client.failover.sleep.max.millis</name>
-  <value>15000</value>
-  <description>
-    Expert only. The time to wait, in milliseconds, between failover
-    attempts increases exponentially as a function of the number of
-    attempts made so far, with a random factor of +/- 50%. This option
-    specifies the maximum value to wait between failovers. 
-    Specifically, the time between two failover attempts will not
-    exceed +/- 50% of dfs.client.failover.sleep.max.millis
-    milliseconds.
-  </description>
-</property>
-
-<property>
-  <name>dfs.client.failover.connection.retries</name>
-  <value>0</value>
-  <description>
-    Expert only. Indicates the number of retries a failover IPC client
-    will make to establish a server connection.
-  </description>
-</property>
-
-<property>
-  <name>dfs.client.failover.connection.retries.on.timeouts</name>
-  <value>0</value>
-  <description>
-    Expert only. The number of retry attempts a failover IPC client
-    will make on socket timeout when establishing a server connection.
-  </description>
-</property>
-
-<property>
-  <name>dfs.nameservices</name>
-  <value></value>
-  <description>
-    Comma-separated list of nameservices.
-  </description>
-</property>
-
-<property>
-  <name>dfs.nameservice.id</name>
-  <value></value>
-  <description>
-    The ID of this nameservice. If the nameservice ID is not
-    configured or more than one nameservice is configured for
-    dfs.nameservices it is determined automatically by
-    matching the local node's address with the configured address.
-  </description>
-</property>
-
-<property>
-  <name>dfs.ha.namenodes.EXAMPLENAMESERVICE</name>
-  <value></value>
-  <description>
-    The prefix for a given nameservice, contains a comma-separated
-    list of namenodes for a given nameservice (eg EXAMPLENAMESERVICE).
-  </description>
-</property>
-
-<property>
-  <name>dfs.ha.namenode.id</name>
-  <value></value>
-  <description>
-    The ID of this namenode. If the namenode ID is not configured it
-    is determined automatically by matching the local node's address
-    with the configured address.
-  </description>
-</property>
-
-<property>
-  <name>dfs.ha.log-roll.period</name>
-  <value>120</value>
-  <description>
-    How often, in seconds, the StandbyNode should ask the active to
-    roll edit logs. Since the StandbyNode only reads from finalized
-    log segments, the StandbyNode will only be as up-to-date as how
-    often the logs are rolled. Note that failover triggers a log roll
-    so the StandbyNode will be up to date before it becomes active.
-  </description>
-</property>
-
-<property>
-  <name>dfs.ha.tail-edits.period</name>
-  <value>60</value>
-  <description>
-    How often, in seconds, the StandbyNode should check for new
-    finalized log segments in the shared edits log.
-  </description>
-</property>
-
-<property>
-  <name>dfs.ha.automatic-failover.enabled</name>
-  <value>false</value>
-  <description>
-    Whether automatic failover is enabled. See the HDFS High
-    Availability documentation for details on automatic HA
-    configuration.
-  </description>
-</property>
-
-<property>
-  <name>dfs.support.append</name>
-  <value>true</value>
-  <description>
-    Does HDFS allow appends to files?
-  </description>
-</property>
-
-<property>
-  <name>dfs.client.use.datanode.hostname</name>
-  <value>false</value>
-  <description>Whether clients should use datanode hostnames when
-    connecting to datanodes.
-  </description>
-</property>
-
-<property>
-  <name>dfs.datanode.use.datanode.hostname</name>
-  <value>false</value>
-  <description>Whether datanodes should use datanode hostnames when
-    connecting to other datanodes for data transfer.
-  </description>
-</property>
-
-<property>
-  <name>dfs.client.local.interfaces</name>
-  <value></value>
-  <description>A comma separated list of network interface names to use
-    for data transfer between the client and datanodes. When creating
-    a connection to read from or write to a datanode, the client
-    chooses one of the specified interfaces at random and binds its
-    socket to the IP of that interface. Individual names may be
-    specified as either an interface name (eg "eth0"), a subinterface
-    name (eg "eth0:0"), or an IP address (which may be specified using
-    CIDR notation to match a range of IPs).
-  </description>
-</property>
-
-<property>
-  <name>dfs.namenode.kerberos.internal.spnego.principal</name>
-  <value>${dfs.web.authentication.kerberos.principal}</value>
-</property>
-
-<property>
-  <name>dfs.secondary.namenode.kerberos.internal.spnego.principal</name>
-  <value>${dfs.web.authentication.kerberos.principal}</value>
-</property>
-
-<property>
-  <name>dfs.namenode.avoid.read.stale.datanode</name>
-  <value>false</value>
-  <description>
-    Indicate whether or not to avoid reading from &quot;stale&quot; datanodes whose
-    heartbeat messages have not been received by the namenode 
-    for more than a specified time interval. Stale datanodes will be
-    moved to the end of the node list returned for reading. See
-    dfs.namenode.avoid.write.stale.datanode for a similar setting for writes.
-  </description>
-</property>
-
-<property>
-  <name>dfs.namenode.avoid.write.stale.datanode</name>
-  <value>false</value>
-  <description>
-    Indicate whether or not to avoid writing to &quot;stale&quot; datanodes whose 
-    heartbeat messages have not been received by the namenode 
-    for more than a specified time interval. Writes will avoid using 
-    stale datanodes unless more than a configured ratio 
-    (dfs.namenode.write.stale.datanode.ratio) of datanodes are marked as 
-    stale. See dfs.namenode.avoid.read.stale.datanode for a similar setting
-    for reads.
-  </description>
-</property>
-
-<property>
-  <name>dfs.namenode.stale.datanode.interval</name>
-  <value>30000</value>
-  <description>
-    Default time interval for marking a datanode as "stale", i.e., if 
-    the namenode has not received heartbeat msg from a datanode for 
-    more than this time interval, the datanode will be marked and treated 
-    as "stale" by default. The stale interval cannot be too small since 
-    otherwise this may cause too frequent change of stale states. 
-    We thus set a minimum stale interval value (the default value is 3 times 
-    of heartbeat interval) and guarantee that the stale interval cannot be less
-    than the minimum value. A stale data node is avoided during lease/block
-    recovery. It can be conditionally avoided for reads (see
-    dfs.namenode.avoid.read.stale.datanode) and for writes (see
-    dfs.namenode.avoid.write.stale.datanode).
-  </description>
-</property>
-
-<property>
-  <name>dfs.namenode.write.stale.datanode.ratio</name>
-  <value>0.5f</value>
-  <description>
-    When the ratio of number stale datanodes to total datanodes marked
-    is greater than this ratio, stop avoiding writing to stale nodes so
-    as to prevent causing hotspots.
-  </description>
-</property>
-
-<property>
-  <name>dfs.namenode.invalidate.work.pct.per.iteration</name>
-  <value>0.32f</value>
-  <description>
-    *Note*: Advanced property. Change with caution.
-    This determines the percentage amount of block
-    invalidations (deletes) to do over a single DN heartbeat
-    deletion command. The final deletion count is determined by applying this
-    percentage to the number of live nodes in the system.
-    The resultant number is the number of blocks from the deletion list
-    chosen for proper invalidation over a single heartbeat of a single DN.
-    Value should be a positive, non-zero percentage in float notation (X.Yf),
-    with 1.0f meaning 100%.
-  </description>
-</property>
-
-<property>
-  <name>dfs.namenode.replication.work.multiplier.per.iteration</name>
-  <value>2</value>
-  <description>
-    *Note*: Advanced property. Change with caution.
-    This determines the total amount of block transfers to begin in
-    parallel at a DN, for replication, when such a command list is being
-    sent over a DN heartbeat by the NN. The actual number is obtained by
-    multiplying this multiplier with the total number of live nodes in the
-    cluster. The result number is the number of blocks to begin transfers
-    immediately for, per DN heartbeat. This number can be any positive,
-    non-zero integer.
-  </description>
-</property>
-
-<property>
-  <name>dfs.webhdfs.enabled</name>
-  <value>true</value>
-  <description>
-    Enable WebHDFS (REST API) in Namenodes and Datanodes.
-  </description>
-</property>
-
-<property>
-  <name>hadoop.fuse.connection.timeout</name>
-  <value>300</value>
-  <description>
-    The minimum number of seconds that we'll cache libhdfs connection objects
-    in fuse_dfs. Lower values will result in lower memory consumption; higher
-    values may speed up access by avoiding the overhead of creating new
-    connection objects.
-  </description>
-</property>
-
-<property>
-  <name>hadoop.fuse.timer.period</name>
-  <value>5</value>
-  <description>
-    The number of seconds between cache expiry checks in fuse_dfs. Lower values
-    will result in fuse_dfs noticing changes to Kerberos ticket caches more
-    quickly.
-  </description>
-</property>
-
-<property>
-  <name>dfs.metrics.percentiles.intervals</name>
-  <value></value>
-  <description>
-    Comma-delimited set of integers denoting the desired rollover intervals 
-    (in seconds) for percentile latency metrics on the Namenode and Datanode.
-    By default, percentile latency metrics are disabled.
-  </description>
-</property>
-
-<property>
-  <name>dfs.encrypt.data.transfer</name>
-  <value>false</value>
-  <description>
-    Whether or not actual block data that is read/written from/to HDFS should
-    be encrypted on the wire. This only needs to be set on the NN and DNs,
-    clients will deduce this automatically.
-  </description>
-</property>
-
-<property>
-  <name>dfs.encrypt.data.transfer.algorithm</name>
-  <value></value>
-  <description>
-    This value may be set to either "3des" or "rc4". If nothing is set, then
-    the configured JCE default on the system is used (usually 3DES.) It is
-    widely believed that 3DES is more cryptographically secure, but RC4 is
-    substantially faster.
-  </description>
-</property>
-
-<property>
-  <name>dfs.datanode.hdfs-blocks-metadata.enabled</name>
-  <value>false</value>
-  <description>
-    Boolean which enables backend datanode-side support for the experimental DistributedFileSystem#getFileVBlockStorageLocations API.
-  </description>
-</property>
-
-<property>
-  <name>dfs.client.file-block-storage-locations.num-threads</name>
-  <value>10</value>
-  <description>
-    Number of threads used for making parallel RPCs in DistributedFileSystem#getFileBlockStorageLocations().
-  </description>
-</property>
-
-<property>
-  <name>dfs.client.file-block-storage-locations.timeout</name>
-  <value>60</value>
-  <description>
-    Timeout (in seconds) for the parallel RPCs made in DistributedFileSystem#getFileBlockStorageLocations().
-  </description>
-</property>
-
-<property>
-  <name>dfs.journalnode.rpc-address</name>
-  <value>0.0.0.0:8485</value>
-  <description>
-    The JournalNode RPC server address and port.
-  </description>
-</property>
-
-<property>
-  <name>dfs.journalnode.http-address</name>
-  <value>0.0.0.0:8480</value>
-  <description>
-    The address and port the JournalNode web UI listens on.
-    If the port is 0 then the server will start on a free port.
-  </description>
-</property>
-
-<property>
-  <name>dfs.namenode.audit.loggers</name>
-  <value>default</value>
-  <description>
-    List of classes implementing audit loggers that will receive audit events.
-    These should be implementations of org.apache.hadoop.hdfs.server.namenode.AuditLogger.
-    The special value "default" can be used to reference the default audit
-    logger, which uses the configured log system. Installing custom audit loggers
-    may affect the performance and stability of the NameNode. Refer to the custom
-    logger's documentation for more details.
-  </description>
-</property>
-
-<property>
-  <name>dfs.domain.socket.path</name>
-  <value></value>
-  <description>
-    Optional.  This is a path to a UNIX domain socket that will be used for
-    communication between the DataNode and local HDFS clients.
-    If the string "_PORT" is present in this path, it will be replaced by the
-    TCP port of the DataNode.
-  </description>
-</property>
-
-<property>
-  <name>dfs.datanode.available-space-volume-choosing-policy.balanced-space-threshold</name>
-  <value>10737418240</value> <!-- 10 GB -->
-  <description>
-    Only used when the dfs.datanode.fsdataset.volume.choosing.policy is set to
-    org.apache.hadoop.hdfs.server.datanode.fsdataset.AvailableSpaceVolumeChoosingPolicy.
-    This setting controls how much DN volumes are allowed to differ in terms of
-    bytes of free disk space before they are considered imbalanced. If the free
-    space of all the volumes are within this range of each other, the volumes
-    will be considered balanced and block assignments will be done on a pure
-    round robin basis.
-  </description>
-</property>
-
-<property>
-  <name>dfs.datanode.available-space-volume-choosing-policy.balanced-space-preference-fraction</name>
-  <value>0.75f</value>
-  <description>
-    Only used when the dfs.datanode.fsdataset.volume.choosing.policy is set to
-    org.apache.hadoop.hdfs.server.datanode.fsdataset.AvailableSpaceVolumeChoosingPolicy.
-    This setting controls what percentage of new block allocations will be sent
-    to volumes with more available disk space than others. This setting should
-    be in the range 0.0 - 1.0, though in practice 0.5 - 1.0, since there should
-    be no reason to prefer that volumes with less available disk space receive
-    more block allocations.
-  </description>
-</property>
-
-<property>
-  <name>dfs.namenode.edits.noeditlogchannelflush</name>
-  <value>false</value>
-  <description>
-    Specifies whether to flush edit log file channel. When set, expensive
-    FileChannel#force calls are skipped and synchronous disk writes are
-    enabled instead by opening the edit log file with RandomAccessFile("rws")
-    flags. This can significantly improve the performance of edit log writes
-    on the Windows platform.
-    Note that the behavior of the "rws" flags is platform and hardware specific
-    and might not provide the same level of guarantees as FileChannel#force.
-    For example, the write will skip the disk-cache on SAS and SCSI devices
-    while it might not on SATA devices. This is an expert level setting,
-    change with caution.
-  </description>
-</property>
-
-<property>
-  <name>dfs.client.cache.drop.behind.writes</name>
-  <value></value>
-  <description>
-    Just like dfs.datanode.drop.cache.behind.writes, this setting causes the
-    page cache to be dropped behind HDFS writes, potentially freeing up more
-    memory for other uses.  Unlike dfs.datanode.drop.cache.behind.writes, this
-    is a client-side setting rather than a setting for the entire datanode.
-    If present, this setting will override the DataNode default.
-
-    If the native libraries are not available to the DataNode, this
-    configuration has no effect.
-  </description>
-</property>
-
-<property>
-  <name>dfs.client.cache.drop.behind.reads</name>
-  <value></value>
-  <description>
-    Just like dfs.datanode.drop.cache.behind.reads, this setting causes the
-    page cache to be dropped behind HDFS reads, potentially freeing up more
-    memory for other uses.  Unlike dfs.datanode.drop.cache.behind.reads, this
-    is a client-side setting rather than a setting for the entire datanode.  If
-    present, this setting will override the DataNode default.
-
-    If the native libraries are not available to the DataNode, this
-    configuration has no effect.
-  </description>
-</property>
-
-<property>
-  <name>dfs.client.cache.readahead</name>
-  <value></value>
-  <description>
-    When using remote reads, this setting causes the datanode to
-    read ahead in the block file using posix_fadvise, potentially decreasing
-    I/O wait times.  Unlike dfs.datanode.readahead.bytes, this is a client-side
-    setting rather than a setting for the entire datanode.  If present, this
-    setting will override the DataNode default.
-
-    When using local reads, this setting determines how much readahead we do in
-    BlockReaderLocal.
-
-    If the native libraries are not available to the DataNode, this
-    configuration has no effect.
-  </description>
-</property>
-
-<property>
-  <name>dfs.namenode.enable.retrycache</name>
-  <value>true</value>
-  <description>
-    This enables the retry cache on the namenode. Namenode tracks for
-    non-idempotent requests the corresponding response. If a client retries the
-    request, the response from the retry cache is sent. Such operations
-    are tagged with annotation @AtMostOnce in namenode protocols. It is
-    recommended that this flag be set to true. Setting it to false, will result
-    in clients getting failure responses to retried request. This flag must 
-    be enabled in HA setup for transparent fail-overs.
-
-    The entries in the cache have expiration time configurable
-    using dfs.namenode.retrycache.expirytime.millis.
-  </description>
-</property>
-
-<property>
-  <name>dfs.namenode.retrycache.expirytime.millis</name>
-  <value>600000</value>
-  <description>
-    The time for which retry cache entries are retained.
-  </description>
-</property>
-
-<property>
-  <name>dfs.namenode.retrycache.heap.percent</name>
-  <value>0.03f</value>
-  <description>
-    This parameter configures the heap size allocated for retry cache
-    (excluding the response cached). This corresponds to approximately
-    4096 entries for every 64MB of namenode process java heap size.
-    Assuming retry cache entry expiration time (configured using
-    dfs.namenode.retrycache.expirytime.millis) of 10 minutes, this
-    enables retry cache to support 7 operations per second sustained
-    for 10 minutes. As the heap size is increased, the operation rate
-    linearly increases.
-  </description>
-</property>
-
-<property>
-  <name>dfs.client.mmap.cache.size</name>
-  <value>1024</value>
-  <description>
-    When zero-copy reads are used, the DFSClient keeps a cache of recently used
-    memory mapped regions.  This parameter controls the maximum number of
-    entries that we will keep in that cache.
-
-    If this is set to 0, we will not allow mmap.
-
-    The larger this number is, the more file descriptors we will potentially
-    use for memory-mapped files.  mmaped files also use virtual address space.
-    You may need to increase your ulimit virtual address space limits before
-    increasing the client mmap cache size.
-  </description>
-</property>
-
-<property>
-  <name>dfs.client.mmap.cache.timeout.ms</name>
-  <value>900000</value>
-  <description>
-    The minimum length of time that we will keep an mmap entry in the cache
-    between uses.  If an entry is in the cache longer than this, and nobody
-    uses it, it will be removed by a background thread.
-  </description>
-</property>
-
-<property>
-  <name>dfs.namenode.path.based.cache.block.map.allocation.percent</name>
-  <value>0.25</value>
-  <description>
-    The percentage of the Java heap which we will allocate to the cached blocks
-    map.  The cached blocks map is a hash map which uses chained hashing.
-    Smaller maps may be accessed more slowly if the number of cached blocks is
-    large; larger maps will consume more memory.
-  </description>
-</property>
-
-<property>
-  <name>dfs.datanode.max.locked.memory</name>
-  <value>0</value>
-  <description>
-    The amount of memory in bytes to use for caching of block replicas in
-    memory on the datanode. The datanode's maximum locked memory soft ulimit
-    (RLIMIT_MEMLOCK) must be set to at least this value, else the datanode
-    will abort on startup.
-
-    By default, this parameter is set to 0, which disables in-memory caching.
-
-    If the native libraries are not available to the DataNode, this
-    configuration has no effect.
-  </description>
-</property>
-
-<property>
-  <name>dfs.namenode.list.cache.directives.num.responses</name>
-  <value>100</value>
-  <description>
-    This value controls the number of cache directives that the NameNode will
-    send over the wire in response to a listDirectives RPC.
-  </description>
-</property>
-
-<property>
-  <name>dfs.namenode.list.cache.pools.num.responses</name>
-  <value>100</value>
-  <description>
-    This value controls the number of cache pools that the NameNode will
-    send over the wire in response to a listPools RPC.
-  </description>
-</property>
-
-<property>
-  <name>dfs.namenode.path.based.cache.refresh.interval.ms</name>
-  <value>300000</value>
-  <description>
-    The amount of milliseconds between subsequent path cache rescans.  Path
-    cache rescans are when we calculate which blocks should be cached, and on
-    what datanodes.
-
-    By default, this parameter is set to 300000, which is five minutes.
-  </description>
-</property>
-
-<property>
-  <name>dfs.namenode.path.based.cache.retry.interval.ms</name>
-  <value>60000</value>
-  <description>
-    When the NameNode needs to uncache something that is cached, or cache
-    something that is not cached, it must direct the DataNodes to do so by
-    sending a DNA_CACHE or DNA_UNCACHE command in response to a DataNode
-    heartbeat.  This parameter controls how frequently the NameNode will
-    resend these commands.
-  </description>
-</property>
-
-<property>
-  <name>dfs.datanode.fsdatasetcache.max.threads.per.volume</name>
-  <value>4</value>
-  <description>
-    The maximum number of threads per volume to use for caching new data
-    on the datanode. These threads consume both I/O and CPU. This can affect
-    normal datanode operations.
-  </description>
-</property>
-
-<property>
-  <name>dfs.cachereport.intervalMsec</name>
-  <value>10000</value>
-  <description>
-    Determines cache reporting interval in milliseconds.  After this amount of
-    time, the DataNode sends a full report of its cache state to the NameNode.
-    The NameNode uses the cache report to update its map of cached blocks to
-    DataNode locations.
-
-    This configuration has no effect if in-memory caching has been disabled by
-    setting dfs.datanode.max.locked.memory to 0 (which is the default).
-
-    If the native libraries are not available to the DataNode, this
-    configuration has no effect.
-  </description>
-</property>
-
-<property>
-  <name>dfs.namenode.edit.log.autoroll.multiplier.threshold</name>
-  <value>2.0</value>
-  <description>
-    Determines when an active namenode will roll its own edit log.
-    The actual threshold (in number of edits) is determined by multiplying
-    this value by dfs.namenode.checkpoint.txns.
-
-    This prevents extremely large edit files from accumulating on the active
-    namenode, which can cause timeouts during namenode startup and pose an
-    administrative hassle. This behavior is intended as a failsafe for when
-    the standby or secondary namenode fail to roll the edit log by the normal
-    checkpoint threshold.
-  </description>
-</property>
-
-<property>
-  <name>dfs.namenode.edit.log.autoroll.check.interval.ms</name>
-  <value>300000</value>
-  <description>
-    How often an active namenode will check if it needs to roll its edit log,
-    in milliseconds.
-  </description>
-</property>
-
-<property>
-  <name>dfs.webhdfs.user.provider.user.pattern</name>
-  <value>^[A-Za-z_][A-Za-z0-9._-]*[$]?$</value>
-  <description>
-    Valid pattern for user and group names for webhdfs, it must be a valid java regex.
-  </description>
-</property>
-
-</configuration>

http://git-wip-us.apache.org/repos/asf/camel/blob/6b77d012/tests/camel-itest-osgi/src/test/resources/org/apache/camel/itest/osgi/CamelContext.xml
----------------------------------------------------------------------
diff --git a/tests/camel-itest-osgi/src/test/resources/org/apache/camel/itest/osgi/CamelContext.xml b/tests/camel-itest-osgi/src/test/resources/org/apache/camel/itest/osgi/CamelContext.xml
deleted file mode 100644
index 6849e52..0000000
--- a/tests/camel-itest-osgi/src/test/resources/org/apache/camel/itest/osgi/CamelContext.xml
+++ /dev/null
@@ -1,33 +0,0 @@
-<?xml version="1.0" encoding="UTF-8"?>
-<!--
-    Licensed to the Apache Software Foundation (ASF) under one or more
-    contributor license agreements.  See the NOTICE file distributed with
-    this work for additional information regarding copyright ownership.
-    The ASF licenses this file to You under the Apache License, Version 2.0
-    (the "License"); you may not use this file except in compliance with
-    the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-    Unless required by applicable law or agreed to in writing, software
-    distributed under the License is distributed on an "AS IS" BASIS,
-    WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-    See the License for the specific language governing permissions and
-    limitations under the License.
--->
-
-<beans xmlns="http://www.springframework.org/schema/beans"
-       xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
-       xmlns:camel="http://camel.apache.org/schema/spring"
-       xsi:schemaLocation="
-       http://www.springframework.org/schema/beans http://www.springframework.org/schema/beans/spring-beans.xsd
-       http://camel.apache.org/schema/spring http://camel.apache.org/schema/spring/camel-spring.xsd">
-
-    <camelContext xmlns="http://camel.apache.org/schema/spring">
-        <camel:route>
-            <camel:from uri="direct:start"/>
-            <camel:to uri="mock:result"/>
-        </camel:route>
-    </camelContext>
-
-</beans>

http://git-wip-us.apache.org/repos/asf/camel/blob/6b77d012/tests/camel-itest-osgi/src/test/resources/org/apache/camel/itest/osgi/aws/CamelContext.xml
----------------------------------------------------------------------
diff --git a/tests/camel-itest-osgi/src/test/resources/org/apache/camel/itest/osgi/aws/CamelContext.xml b/tests/camel-itest-osgi/src/test/resources/org/apache/camel/itest/osgi/aws/CamelContext.xml
deleted file mode 100644
index 492313b..0000000
--- a/tests/camel-itest-osgi/src/test/resources/org/apache/camel/itest/osgi/aws/CamelContext.xml
+++ /dev/null
@@ -1,69 +0,0 @@
-<?xml version="1.0" encoding="UTF-8"?>
-<!--
-  Licensed to the Apache Software Foundation (ASF) under one or more
-  contributor license agreements. See the NOTICE file distributed with
-  this work for additional information regarding copyright ownership.
-  The ASF licenses this file to You under the Apache License, Version
-  2.0 (the "License"); you may not use this file except in compliance
-  with the License. You may obtain a copy of the License at
-
-  http://www.apache.org/licenses/LICENSE-2.0 Unless required by
-  applicable law or agreed to in writing, software distributed under
-  the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES
-  OR CONDITIONS OF ANY KIND, either express or implied. See the
-  License for the specific language governing permissions and
-  limitations under the License.
--->
-<beans xmlns="http://www.springframework.org/schema/beans"
-       xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
-       xsi:schemaLocation="
-    http://www.springframework.org/schema/beans http://www.springframework.org/schema/beans/spring-beans.xsd
-    http://camel.apache.org/schema/spring http://camel.apache.org/schema/spring/camel-spring.xsd">
-
-    <camelContext id="camel" xmlns="http://camel.apache.org/schema/spring">
-        <route>
-            <from uri="direct:start"/>
-            <to uri="aws-sqs://MyQueue?amazonSQSClient=#amazonSQSClient"/>
-        </route>
-
-        <route>
-            <from uri="aws-sqs://MyQueue?amazonSQSClient=#amazonSQSClient"/>
-            <to uri="mock:result"/>
-        </route>
-        
-        <route>
-            <from uri="direct:start-sns"/>
-            <to uri="aws-sns://MyTopic?amazonSNSClient=#amazonSNSClient"/>
-        </route>
-        
-        <route>
-            <from uri="direct:start-s3"/>
-            <to uri="aws-s3://mycamelbucket?amazonS3Client=#amazonS3Client"/>
-        </route>
-        
-        <route>
-            <from uri="aws-s3://mycamelbucket?amazonS3Client=#amazonS3Client&amp;maxMessagesPerPoll=5"/>
-            <to uri="mock:result-s3"/>
-        </route>
-        
-        <route>
-            <from uri="direct:start-ses"/>
-            <to uri="aws-ses://from@example.com?amazonSESClient=#amazonSESClient"/>
-        </route>
-        
-        <route>
-            <from uri="direct:start-sdb"/>
-            <to uri="aws-sdb://TestDomain?amazonSDBClient=#amazonSDBClient&amp;operation=GetAttributes"/>
-        </route>
-    </camelContext>
-
-    <bean id="amazonSQSClient" class="org.apache.camel.itest.osgi.aws.AmazonSQSClientMock"/>
-    
-    <bean id="amazonSNSClient" class="org.apache.camel.itest.osgi.aws.AmazonSNSClientMock"/>
-    
-    <bean id="amazonS3Client" class="org.apache.camel.itest.osgi.aws.AmazonS3ClientMock"/>
-    
-    <bean id="amazonSESClient" class="org.apache.camel.itest.osgi.aws.AmazonSESClientMock"/>
-    
-    <bean id="amazonSDBClient" class="org.apache.camel.itest.osgi.aws.AmazonSDBClientMock"/>
-</beans>
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/camel/blob/6b77d012/tests/camel-itest-osgi/src/test/resources/org/apache/camel/itest/osgi/aws/CamelIntegrationContext.xml
----------------------------------------------------------------------
diff --git a/tests/camel-itest-osgi/src/test/resources/org/apache/camel/itest/osgi/aws/CamelIntegrationContext.xml b/tests/camel-itest-osgi/src/test/resources/org/apache/camel/itest/osgi/aws/CamelIntegrationContext.xml
deleted file mode 100644
index 65422e7..0000000
--- a/tests/camel-itest-osgi/src/test/resources/org/apache/camel/itest/osgi/aws/CamelIntegrationContext.xml
+++ /dev/null
@@ -1,59 +0,0 @@
-<?xml version="1.0" encoding="UTF-8"?>
-<!--
-  Licensed to the Apache Software Foundation (ASF) under one or more
-  contributor license agreements. See the NOTICE file distributed with
-  this work for additional information regarding copyright ownership.
-  The ASF licenses this file to You under the Apache License, Version
-  2.0 (the "License"); you may not use this file except in compliance
-  with the License. You may obtain a copy of the License at
-
-  http://www.apache.org/licenses/LICENSE-2.0 Unless required by
-  applicable law or agreed to in writing, software distributed under
-  the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES
-  OR CONDITIONS OF ANY KIND, either express or implied. See the
-  License for the specific language governing permissions and
-  limitations under the License.
--->
-<beans xmlns="http://www.springframework.org/schema/beans"
-       xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
-       xsi:schemaLocation="
-    http://www.springframework.org/schema/beans http://www.springframework.org/schema/beans/spring-beans.xsd
-    http://camel.apache.org/schema/spring http://camel.apache.org/schema/spring/camel-spring.xsd">
-
-    <camelContext id="camel" xmlns="http://camel.apache.org/schema/spring">
-        <route>
-            <from uri="direct:start"/>
-            <to uri="aws-sqs://MyQueue?accessKey=xxx&amp;secretKey=yyy"/>
-        </route>
-
-        <route>
-            <from uri="aws-sqs://MyQueue?accessKey=xxx&amp;secretKey=yyy"/>
-            <to uri="mock:result"/>
-        </route>
-        
-        <route>
-            <from uri="direct:start-sns"/>
-            <to uri="aws-sns://MyTopic?accessKey=xxx&amp;secretKey=yyy"/>
-        </route>
-        
-        <route>
-            <from uri="direct:start-s3"/>
-            <to uri="aws-s3://mycamelbucket?accessKey=xxx&amp;secretKey=yyy"/>
-        </route>
-        
-        <route>
-            <from uri="aws-s3://mycamelbucket?accessKey=xxx&amp;secretKey=yyy&amp;maxMessagesPerPoll=5"/>
-            <to uri="mock:result-s3"/>
-        </route>
-        
-        <route>
-            <from uri="direct:start-ses"/>
-            <to uri="aws-ses://from@example.com?accessKey=xxx&amp;secretKey=yyy"/>
-        </route>
-        
-        <route>
-            <from uri="direct:start-sdb"/>
-            <to uri="aws-sdb://TestDomain?amazonSDBClient=#amazonSDBClient&amp;operation=GetAttributes"/>
-        </route>
-    </camelContext>
-</beans>
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/camel/blob/6b77d012/tests/camel-itest-osgi/src/test/resources/org/apache/camel/itest/osgi/beanio/mappings.xml
----------------------------------------------------------------------
diff --git a/tests/camel-itest-osgi/src/test/resources/org/apache/camel/itest/osgi/beanio/mappings.xml b/tests/camel-itest-osgi/src/test/resources/org/apache/camel/itest/osgi/beanio/mappings.xml
deleted file mode 100644
index 8874373..0000000
--- a/tests/camel-itest-osgi/src/test/resources/org/apache/camel/itest/osgi/beanio/mappings.xml
+++ /dev/null
@@ -1,30 +0,0 @@
-<!--
-  Licensed to the Apache Software Foundation (ASF) under one or more
-  contributor license agreements.  See the NOTICE file distributed with
-  this work for additional information regarding copyright ownership.
-  The ASF licenses this file to You under the Apache License, Version 2.0
-  (the "License"); you may not use this file except in compliance with
-  the License.  You may obtain a copy of the License at
-
-  http://www.apache.org/licenses/LICENSE-2.0
-
-  Unless required by applicable law or agreed to in writing, software
-  distributed under the License is distributed on an "AS IS" BASIS,
-  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-  See the License for the specific language governing permissions and
-  limitations under the License.
--->
-<beanio xmlns="http://www.beanio.org/2011/01" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
-        xsi:schemaLocation="http://www.beanio.org/2011/01 http://www.beanio.org/2011/01/mapping.xsd">
-
-  <stream name="employeeFile" format="csv">
-    <record name="employee" class="org.apache.camel.itest.osgi.beanio.Employee">
-      <field name="firstName"/>
-      <field name="lastName"/>
-      <field name="title"/>
-      <field name="salary"/>
-      <field name="hireDate" format="MMddyyyy"/>
-    </record>
-  </stream>
-
-</beanio>
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/camel/blob/6b77d012/tests/camel-itest-osgi/src/test/resources/org/apache/camel/itest/osgi/bindy/BindySpringDataFormatCsvJavaRouteTest.xml
----------------------------------------------------------------------
diff --git a/tests/camel-itest-osgi/src/test/resources/org/apache/camel/itest/osgi/bindy/BindySpringDataFormatCsvJavaRouteTest.xml b/tests/camel-itest-osgi/src/test/resources/org/apache/camel/itest/osgi/bindy/BindySpringDataFormatCsvJavaRouteTest.xml
deleted file mode 100644
index 59d8708..0000000
--- a/tests/camel-itest-osgi/src/test/resources/org/apache/camel/itest/osgi/bindy/BindySpringDataFormatCsvJavaRouteTest.xml
+++ /dev/null
@@ -1,38 +0,0 @@
-<?xml version="1.0" encoding="UTF-8"?>
-<!--
-    Licensed to the Apache Software Foundation (ASF) under one or more
-    contributor license agreements.  See the NOTICE file distributed with
-    this work for additional information regarding copyright ownership.
-    The ASF licenses this file to You under the Apache License, Version 2.0
-    (the "License"); you may not use this file except in compliance with
-    the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-    Unless required by applicable law or agreed to in writing, software
-    distributed under the License is distributed on an "AS IS" BASIS,
-    WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-    See the License for the specific language governing permissions and
-    limitations under the License.
--->
-
-<beans xmlns="http://www.springframework.org/schema/beans"
-       xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
-       xmlns:camel="http://camel.apache.org/schema/spring"
-       xsi:schemaLocation="
-       http://www.springframework.org/schema/beans http://www.springframework.org/schema/beans/spring-beans.xsd
-       http://camel.apache.org/schema/spring http://camel.apache.org/schema/spring/camel-spring.xsd">
-
-  <bean id="myRoutes" class="org.apache.camel.itest.osgi.bindy.MyRoutes"/>
-
-  <camelContext xmlns="http://camel.apache.org/schema/spring">
-
-    <routeBuilder ref="myRoutes"/>
-
-    <dataFormats>
-      <bindy id="myBindy" type="Csv" classType="org.apache.camel.itest.osgi.bindy.Employee"/>
-    </dataFormats>
-
-  </camelContext>
-
-</beans>

http://git-wip-us.apache.org/repos/asf/camel/blob/6b77d012/tests/camel-itest-osgi/src/test/resources/org/apache/camel/itest/osgi/bindy/BindySpringDataFormatCsvTest.xml
----------------------------------------------------------------------
diff --git a/tests/camel-itest-osgi/src/test/resources/org/apache/camel/itest/osgi/bindy/BindySpringDataFormatCsvTest.xml b/tests/camel-itest-osgi/src/test/resources/org/apache/camel/itest/osgi/bindy/BindySpringDataFormatCsvTest.xml
deleted file mode 100644
index fe4afa6..0000000
--- a/tests/camel-itest-osgi/src/test/resources/org/apache/camel/itest/osgi/bindy/BindySpringDataFormatCsvTest.xml
+++ /dev/null
@@ -1,48 +0,0 @@
-<?xml version="1.0" encoding="UTF-8"?>
-<!--
-    Licensed to the Apache Software Foundation (ASF) under one or more
-    contributor license agreements.  See the NOTICE file distributed with
-    this work for additional information regarding copyright ownership.
-    The ASF licenses this file to You under the Apache License, Version 2.0
-    (the "License"); you may not use this file except in compliance with
-    the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-    Unless required by applicable law or agreed to in writing, software
-    distributed under the License is distributed on an "AS IS" BASIS,
-    WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-    See the License for the specific language governing permissions and
-    limitations under the License.
--->
-
-<beans xmlns="http://www.springframework.org/schema/beans"
-       xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
-       xmlns:camel="http://camel.apache.org/schema/spring"
-       xsi:schemaLocation="
-       http://www.springframework.org/schema/beans http://www.springframework.org/schema/beans/spring-beans.xsd
-       http://camel.apache.org/schema/spring http://camel.apache.org/schema/spring/camel-spring.xsd">
-
-  <camelContext xmlns="http://camel.apache.org/schema/spring">
-
-    <dataFormats>
-      <bindy id="myBindy" type="Csv" classType="org.apache.camel.itest.osgi.bindy.Employee"/>
-    </dataFormats>
-
-    <route>
-      <from uri="direct:unmarshal"/>
-      <unmarshal ref="myBindy"/>
-      <split>
-        <simple>${body}</simple>
-        <to uri="mock:bindy-unmarshal"/>
-      </split>
-    </route>
-
-    <route>
-      <from uri="direct:marshal"/>
-      <marshal ref="myBindy"/>
-      <to uri="mock:bindy-marshal"/>
-    </route>
-  </camelContext>
-
-</beans>

http://git-wip-us.apache.org/repos/asf/camel/blob/6b77d012/tests/camel-itest-osgi/src/test/resources/org/apache/camel/itest/osgi/blueprint/blueprint-1.xml
----------------------------------------------------------------------
diff --git a/tests/camel-itest-osgi/src/test/resources/org/apache/camel/itest/osgi/blueprint/blueprint-1.xml b/tests/camel-itest-osgi/src/test/resources/org/apache/camel/itest/osgi/blueprint/blueprint-1.xml
deleted file mode 100644
index 3768e24..0000000
--- a/tests/camel-itest-osgi/src/test/resources/org/apache/camel/itest/osgi/blueprint/blueprint-1.xml
+++ /dev/null
@@ -1,27 +0,0 @@
-<?xml version="1.0" encoding="UTF-8"?>
-<!--
-    Licensed to the Apache Software Foundation (ASF) under one or more
-    contributor license agreements.  See the NOTICE file distributed with
-    this work for additional information regarding copyright ownership.
-    The ASF licenses this file to You under the Apache License, Version 2.0
-    (the "License"); you may not use this file except in compliance with
-    the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-    Unless required by applicable law or agreed to in writing, software
-    distributed under the License is distributed on an "AS IS" BASIS,
-    WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-    See the License for the specific language governing permissions and
-    limitations under the License.
--->
-<blueprint xmlns="http://www.osgi.org/xmlns/blueprint/v1.0.0">
-
-    <camelContext xmlns="http://camel.apache.org/schema/blueprint">
-        <route>
-            <from uri="direct:start"/>
-            <to uri="mock:result"/>
-        </route>
-    </camelContext>
-
-</blueprint>

http://git-wip-us.apache.org/repos/asf/camel/blob/6b77d012/tests/camel-itest-osgi/src/test/resources/org/apache/camel/itest/osgi/blueprint/blueprint-10.xml
----------------------------------------------------------------------
diff --git a/tests/camel-itest-osgi/src/test/resources/org/apache/camel/itest/osgi/blueprint/blueprint-10.xml b/tests/camel-itest-osgi/src/test/resources/org/apache/camel/itest/osgi/blueprint/blueprint-10.xml
deleted file mode 100644
index d1c9dfd..0000000
--- a/tests/camel-itest-osgi/src/test/resources/org/apache/camel/itest/osgi/blueprint/blueprint-10.xml
+++ /dev/null
@@ -1,26 +0,0 @@
-<?xml version="1.0" encoding="UTF-8"?>
-<!--
-    Licensed to the Apache Software Foundation (ASF) under one or more
-    contributor license agreements.  See the NOTICE file distributed with
-    this work for additional information regarding copyright ownership.
-    The ASF licenses this file to You under the Apache License, Version 2.0
-    (the "License"); you may not use this file except in compliance with
-    the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-    Unless required by applicable law or agreed to in writing, software
-    distributed under the License is distributed on an "AS IS" BASIS,
-    WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-    See the License for the specific language governing permissions and
-    limitations under the License.
--->
-<blueprint xmlns="http://www.osgi.org/xmlns/blueprint/v1.0.0">
-
-    <camelContext xmlns="http://camel.apache.org/schema/blueprint">
-        <endpoint id="testEndpoint" uri="seda:test.queue" />
-    </camelContext>
-
-    <bean id="producer" class="org.apache.camel.itest.osgi.blueprint.TestProducer"/>
-
-</blueprint>

http://git-wip-us.apache.org/repos/asf/camel/blob/6b77d012/tests/camel-itest-osgi/src/test/resources/org/apache/camel/itest/osgi/blueprint/blueprint-11.xml
----------------------------------------------------------------------
diff --git a/tests/camel-itest-osgi/src/test/resources/org/apache/camel/itest/osgi/blueprint/blueprint-11.xml b/tests/camel-itest-osgi/src/test/resources/org/apache/camel/itest/osgi/blueprint/blueprint-11.xml
deleted file mode 100644
index 407dfd9..0000000
--- a/tests/camel-itest-osgi/src/test/resources/org/apache/camel/itest/osgi/blueprint/blueprint-11.xml
+++ /dev/null
@@ -1,45 +0,0 @@
-<?xml version="1.0" encoding="UTF-8"?>
-<!--
-    Licensed to the Apache Software Foundation (ASF) under one or more
-    contributor license agreements.  See the NOTICE file distributed with
-    this work for additional information regarding copyright ownership.
-    The ASF licenses this file to You under the Apache License, Version 2.0
-    (the "License"); you may not use this file except in compliance with
-    the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-    Unless required by applicable law or agreed to in writing, software
-    distributed under the License is distributed on an "AS IS" BASIS,
-    WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-    See the License for the specific language governing permissions and
-    limitations under the License.
--->
-<blueprint xmlns="http://www.osgi.org/xmlns/blueprint/v1.0.0">
-
-    <routeContext id="myCoolRoutes" xmlns="http://camel.apache.org/schema/blueprint">
-        <!-- we can have a route -->
-        <route id="cool">
-            <from uri="direct:start"/>
-            <to uri="mock:result"/>
-        </route>
-        <!-- and another route, you can have as many your like -->
-        <route id="bar">
-            <from uri="direct:bar"/>
-            <to uri="mock:bar"/>
-        </route>
-    </routeContext>
-
-    <camelContext xmlns="http://camel.apache.org/schema/blueprint">
-
-        <!-- refer to a given route to be used -->
-        <routeContextRef ref="myCoolRoutes"/>
-
-        <!-- we can of course still use routes inside camelContext -->
-        <route id="inside">
-            <from uri="direct:inside"/>
-            <to uri="mock:inside"/>
-        </route>
-    </camelContext>
-
-</blueprint>

http://git-wip-us.apache.org/repos/asf/camel/blob/6b77d012/tests/camel-itest-osgi/src/test/resources/org/apache/camel/itest/osgi/blueprint/blueprint-12.xml
----------------------------------------------------------------------
diff --git a/tests/camel-itest-osgi/src/test/resources/org/apache/camel/itest/osgi/blueprint/blueprint-12.xml b/tests/camel-itest-osgi/src/test/resources/org/apache/camel/itest/osgi/blueprint/blueprint-12.xml
deleted file mode 100644
index 8649566..0000000
--- a/tests/camel-itest-osgi/src/test/resources/org/apache/camel/itest/osgi/blueprint/blueprint-12.xml
+++ /dev/null
@@ -1,38 +0,0 @@
-<?xml version="1.0" encoding="UTF-8"?>
-<!--
-    Licensed to the Apache Software Foundation (ASF) under one or more
-    contributor license agreements.  See the NOTICE file distributed with
-    this work for additional information regarding copyright ownership.
-    The ASF licenses this file to You under the Apache License, Version 2.0
-    (the "License"); you may not use this file except in compliance with
-    the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-    Unless required by applicable law or agreed to in writing, software
-    distributed under the License is distributed on an "AS IS" BASIS,
-    WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-    See the License for the specific language governing permissions and
-    limitations under the License.
--->
-<blueprint xmlns="http://www.osgi.org/xmlns/blueprint/v1.0.0">
-
-    <camelContext xmlns="http://camel.apache.org/schema/blueprint">
-
-        <!-- create a proxy that will route to the direct:start endpoint when invoked -->
-        <proxy id="myProxySender"
-               serviceInterface="org.apache.camel.itest.osgi.blueprint.TestProxySender"
-               serviceUrl="direct:start"/>
-
-        <!-- this is the route that our proxy will routed when invoked
-             and the output from this route is returned as reply on the proxy -->
-        <route>
-            <from uri="direct:start"/>
-            <transform>
-                <simple>Bye ${body}</simple>
-            </transform>
-        </route>
-
-    </camelContext>
-
-</blueprint>

http://git-wip-us.apache.org/repos/asf/camel/blob/6b77d012/tests/camel-itest-osgi/src/test/resources/org/apache/camel/itest/osgi/blueprint/blueprint-13.xml
----------------------------------------------------------------------
diff --git a/tests/camel-itest-osgi/src/test/resources/org/apache/camel/itest/osgi/blueprint/blueprint-13.xml b/tests/camel-itest-osgi/src/test/resources/org/apache/camel/itest/osgi/blueprint/blueprint-13.xml
deleted file mode 100644
index 01176da..0000000
--- a/tests/camel-itest-osgi/src/test/resources/org/apache/camel/itest/osgi/blueprint/blueprint-13.xml
+++ /dev/null
@@ -1,35 +0,0 @@
-<?xml version="1.0" encoding="UTF-8"?>
-<!--
-    Licensed to the Apache Software Foundation (ASF) under one or more
-    contributor license agreements.  See the NOTICE file distributed with
-    this work for additional information regarding copyright ownership.
-    The ASF licenses this file to You under the Apache License, Version 2.0
-    (the "License"); you may not use this file except in compliance with
-    the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-    Unless required by applicable law or agreed to in writing, software
-    distributed under the License is distributed on an "AS IS" BASIS,
-    WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-    See the License for the specific language governing permissions and
-    limitations under the License.
--->
-<blueprint xmlns="http://www.osgi.org/xmlns/blueprint/v1.0.0"
-           xmlns:camel="http://camel.apache.org/schema/blueprint">
-
-    <camel:camelContext id="camel1">
-        <camel:route>
-            <camel:from uri="file:target/foo"/>
-            <camel:to uri="mock:result"/>
-        </camel:route>
-    </camel:camelContext>
-
-    <camel:camelContext id="camel2">
-        <camel:route>
-            <camel:from uri="direct:start"/>
-            <camel:to uri="mock:end"/>
-        </camel:route>
-    </camel:camelContext>
-
-</blueprint>

http://git-wip-us.apache.org/repos/asf/camel/blob/6b77d012/tests/camel-itest-osgi/src/test/resources/org/apache/camel/itest/osgi/blueprint/blueprint-14.xml
----------------------------------------------------------------------
diff --git a/tests/camel-itest-osgi/src/test/resources/org/apache/camel/itest/osgi/blueprint/blueprint-14.xml b/tests/camel-itest-osgi/src/test/resources/org/apache/camel/itest/osgi/blueprint/blueprint-14.xml
deleted file mode 100644
index 972ebd2..0000000
--- a/tests/camel-itest-osgi/src/test/resources/org/apache/camel/itest/osgi/blueprint/blueprint-14.xml
+++ /dev/null
@@ -1,31 +0,0 @@
-<?xml version="1.0" encoding="UTF-8"?>
-<!--
-    Licensed to the Apache Software Foundation (ASF) under one or more
-    contributor license agreements.  See the NOTICE file distributed with
-    this work for additional information regarding copyright ownership.
-    The ASF licenses this file to You under the Apache License, Version 2.0
-    (the "License"); you may not use this file except in compliance with
-    the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-    Unless required by applicable law or agreed to in writing, software
-    distributed under the License is distributed on an "AS IS" BASIS,
-    WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-    See the License for the specific language governing permissions and
-    limitations under the License.
--->
-<blueprint xmlns="http://www.osgi.org/xmlns/blueprint/v1.0.0">
-
-    <camelContext xmlns="http://camel.apache.org/schema/blueprint">
-
-        <errorHandler id="dlc" deadLetterUri="mock:dead" type="DeadLetterChannel"/>
-
-        <route errorHandlerRef="dlc">
-            <from uri="direct:start"/>
-            <to uri="mock:result"/>
-        </route>
-
-    </camelContext>
-
-</blueprint>

http://git-wip-us.apache.org/repos/asf/camel/blob/6b77d012/tests/camel-itest-osgi/src/test/resources/org/apache/camel/itest/osgi/blueprint/blueprint-15.xml
----------------------------------------------------------------------
diff --git a/tests/camel-itest-osgi/src/test/resources/org/apache/camel/itest/osgi/blueprint/blueprint-15.xml b/tests/camel-itest-osgi/src/test/resources/org/apache/camel/itest/osgi/blueprint/blueprint-15.xml
deleted file mode 100644
index 17aa3a2..0000000
--- a/tests/camel-itest-osgi/src/test/resources/org/apache/camel/itest/osgi/blueprint/blueprint-15.xml
+++ /dev/null
@@ -1,29 +0,0 @@
-<?xml version="1.0" encoding="UTF-8"?>
-<!--
-    Licensed to the Apache Software Foundation (ASF) under one or more
-    contributor license agreements.  See the NOTICE file distributed with
-    this work for additional information regarding copyright ownership.
-    The ASF licenses this file to You under the Apache License, Version 2.0
-    (the "License"); you may not use this file except in compliance with
-    the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-    Unless required by applicable law or agreed to in writing, software
-    distributed under the License is distributed on an "AS IS" BASIS,
-    WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-    See the License for the specific language governing permissions and
-    limitations under the License.
--->
-<blueprint xmlns="http://www.osgi.org/xmlns/blueprint/v1.0.0">
-
-    <camelContext xmlns="http://camel.apache.org/schema/blueprint">
-        <route>
-            <from uri="mycomp:queue"/>
-            <to uri="mock:result"/>
-        </route>
-    </camelContext>
-
-    <bean id="mycomp" class="org.apache.camel.component.seda.SedaComponent"/>
-
-</blueprint>


Mime
View raw message