nifi-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From jperciv...@apache.org
Subject svn commit: r1752359 [4/9] - in /nifi/site/trunk/docs/nifi-docs/components: org.apache.nifi.amqp.processors.ConsumeAMQP/ org.apache.nifi.amqp.processors.PublishAMQP/ org.apache.nifi.controller.MonitorMemory/ org.apache.nifi.dbcp.hive.HiveConnectionPool...
Date Tue, 12 Jul 2016 22:38:05 GMT
Modified: nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.elasticsearch.PutElasticsearch/index.html
URL: http://svn.apache.org/viewvc/nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.elasticsearch.PutElasticsearch/index.html?rev=1752359&r1=1752358&r2=1752359&view=diff
==============================================================================
--- nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.elasticsearch.PutElasticsearch/index.html (original)
+++ nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.elasticsearch.PutElasticsearch/index.html Tue Jul 12 22:38:04 2016
@@ -1 +1 @@
-<!DOCTYPE html><html lang="en"><head><meta charset="utf-8"></meta><title>PutElasticsearch</title><link rel="stylesheet" href="../../css/component-usage.css" type="text/css"></link></head><body><h2>Description: </h2><p>Writes the contents of a FlowFile to Elasticsearch, using the specified parameters such as the index to insert into and the type of the document. If the cluster has been configured for authorization and/or secure transport (SSL/TLS) and the Shield plugin is available, secure connections can be made. This processor supports Elasticsearch 2.x clusters.</p><h3>Tags: </h3><p>elasticsearch, insert, update, write, put</p><h3>Properties: </h3><p>In the list below, the names of required properties appear in <strong>bold</strong>. Any other properties (not in bold) are considered optional. The table also indicates any default values, whether a property supports the <a href="../../html/expression-language-guide.html">NiFi Expression Language</a>, and whether a property is consid
 ered "sensitive", meaning that its value will be encrypted. Before entering a value in a sensitive property, ensure that the <strong>nifi.properties</strong> file has an entry for the property <strong>nifi.sensitive.props.key</strong>.</p><table id="properties"><tr><th>Name</th><th>Default Value</th><th>Allowable Values</th><th>Description</th></tr><tr><td id="name"><strong>Cluster Name</strong></td><td id="default-value">elasticsearch</td><td id="allowable-values"></td><td id="description">Name of the ES cluster (for example, elasticsearch_brew). Defaults to 'elasticsearch'</td></tr><tr><td id="name"><strong>ElasticSearch Hosts</strong></td><td id="default-value"></td><td id="allowable-values"></td><td id="description">ElasticSearch Hosts, which should be comma separated and colon for hostname/port host1:port,host2:port,....  For example testcluster:9300.</td></tr><tr><td id="name">SSL Context Service</td><td id="default-value"></td><td id="allowable-values"><strong>Controller Serv
 ice API: </strong><br/>SSLContextService<br/><strong>Implementation:</strong><br/><a href="../org.apache.nifi.ssl.StandardSSLContextService/index.html">StandardSSLContextService</a></td><td id="description">The SSL Context Service used to provide client certificate information for TLS/SSL connections. This service only applies if the Shield plugin is available.</td></tr><tr><td id="name">Shield Plugin Filename</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Specifies the path to the JAR for the Elasticsearch Shield plugin. If the Elasticsearch cluster has been secured with the Shield plugin, then the Shield plugin JAR must also be available to this processor. Note: Do NOT place the Shield JAR into NiFi's lib/ directory, doing so will prevent the Shield plugin from being loaded.</td></tr><tr><td id="name">Username</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Username to access the Elasticsearch cluster</td></t
 r><tr><td id="name">Password</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Password to access the Elasticsearch cluster<br/><strong>Sensitive Property: true</strong></td></tr><tr><td id="name"><strong>ElasticSearch Ping Timeout</strong></td><td id="default-value">5s</td><td id="allowable-values"></td><td id="description">The ping timeout used to determine when a node is unreachable. For example, 5s (5 seconds). If non-local recommended is 30s</td></tr><tr><td id="name"><strong>Sampler Interval</strong></td><td id="default-value">5s</td><td id="allowable-values"></td><td id="description">How often to sample / ping the nodes listed and connected. For example, 5s (5 seconds). If non-local recommended is 30s.</td></tr><tr><td id="name"><strong>Identifier Attribute</strong></td><td id="default-value"></td><td id="allowable-values"></td><td id="description">The name of the attribute containing the identifier for each FlowFile</td></tr><tr><td id="name
 "><strong>Index</strong></td><td id="default-value"></td><td id="allowable-values"></td><td id="description">The name of the index to insert into<br/><strong>Supports Expression Language: true</strong></td></tr><tr><td id="name"><strong>Type</strong></td><td id="default-value"></td><td id="allowable-values"></td><td id="description">The type of this document (used by Elasticsearch for indexing and searching)<br/><strong>Supports Expression Language: true</strong></td></tr><tr><td id="name"><strong>Character Set</strong></td><td id="default-value">UTF-8</td><td id="allowable-values"></td><td id="description">Specifies the character set of the document data.</td></tr><tr><td id="name"><strong>Batch Size</strong></td><td id="default-value">100</td><td id="allowable-values"></td><td id="description">The preferred number of FlowFiles to put to the database in a single transaction</td></tr></table><h3>Relationships: </h3><table id="relationships"><tr><th>Name</th><th>Description</th></tr>
 <tr><td>retry</td><td>A FlowFile is routed to this relationship if the database cannot be updated but attempting the operation again may succeed</td></tr><tr><td>success</td><td>All FlowFiles that are written to Elasticsearch are routed to this relationship</td></tr><tr><td>failure</td><td>All FlowFiles that cannot be written to Elasticsearch are routed to this relationship</td></tr></table><h3>Reads Attributes: </h3>None specified.<h3>Writes Attributes: </h3>None specified.</body></html>
\ No newline at end of file
+<!DOCTYPE html><html lang="en"><head><meta charset="utf-8"></meta><title>PutElasticsearch</title><link rel="stylesheet" href="../../css/component-usage.css" type="text/css"></link></head><body><h2>Description: </h2><p>Writes the contents of a FlowFile to Elasticsearch, using the specified parameters such as the index to insert into and the type of the document. If the cluster has been configured for authorization and/or secure transport (SSL/TLS) and the Shield plugin is available, secure connections can be made. This processor supports Elasticsearch 2.x clusters.</p><h3>Tags: </h3><p>elasticsearch, insert, update, write, put</p><h3>Properties: </h3><p>In the list below, the names of required properties appear in <strong>bold</strong>. Any other properties (not in bold) are considered optional. The table also indicates any default values, whether a property supports the <a href="../../html/expression-language-guide.html">NiFi Expression Language</a>, and whether a property is consid
 ered "sensitive", meaning that its value will be encrypted. Before entering a value in a sensitive property, ensure that the <strong>nifi.properties</strong> file has an entry for the property <strong>nifi.sensitive.props.key</strong>.</p><table id="properties"><tr><th>Name</th><th>Default Value</th><th>Allowable Values</th><th>Description</th></tr><tr><td id="name"><strong>Cluster Name</strong></td><td id="default-value">elasticsearch</td><td id="allowable-values"></td><td id="description">Name of the ES cluster (for example, elasticsearch_brew). Defaults to 'elasticsearch'</td></tr><tr><td id="name"><strong>ElasticSearch Hosts</strong></td><td id="default-value"></td><td id="allowable-values"></td><td id="description">ElasticSearch Hosts, which should be comma separated and colon for hostname/port host1:port,host2:port,....  For example testcluster:9300.</td></tr><tr><td id="name">SSL Context Service</td><td id="default-value"></td><td id="allowable-values"><strong>Controller Serv
 ice API: </strong><br/>SSLContextService<br/><strong>Implementation:</strong><br/><a href="../org.apache.nifi.ssl.StandardSSLContextService/index.html">StandardSSLContextService</a></td><td id="description">The SSL Context Service used to provide client certificate information for TLS/SSL connections. This service only applies if the Shield plugin is available.</td></tr><tr><td id="name">Shield Plugin Filename</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Specifies the path to the JAR for the Elasticsearch Shield plugin. If the Elasticsearch cluster has been secured with the Shield plugin, then the Shield plugin JAR must also be available to this processor. Note: Do NOT place the Shield JAR into NiFi's lib/ directory, doing so will prevent the Shield plugin from being loaded.</td></tr><tr><td id="name">Username</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Username to access the Elasticsearch cluster</td></t
 r><tr><td id="name">Password</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Password to access the Elasticsearch cluster<br/><strong>Sensitive Property: true</strong></td></tr><tr><td id="name"><strong>ElasticSearch Ping Timeout</strong></td><td id="default-value">5s</td><td id="allowable-values"></td><td id="description">The ping timeout used to determine when a node is unreachable. For example, 5s (5 seconds). If non-local recommended is 30s</td></tr><tr><td id="name"><strong>Sampler Interval</strong></td><td id="default-value">5s</td><td id="allowable-values"></td><td id="description">How often to sample / ping the nodes listed and connected. For example, 5s (5 seconds). If non-local recommended is 30s.</td></tr><tr><td id="name"><strong>Identifier Attribute</strong></td><td id="default-value"></td><td id="allowable-values"></td><td id="description">The name of the attribute containing the identifier for each FlowFile</td></tr><tr><td id="name
 "><strong>Index</strong></td><td id="default-value"></td><td id="allowable-values"></td><td id="description">The name of the index to insert into<br/><strong>Supports Expression Language: true</strong></td></tr><tr><td id="name"><strong>Type</strong></td><td id="default-value"></td><td id="allowable-values"></td><td id="description">The type of this document (used by Elasticsearch for indexing and searching)<br/><strong>Supports Expression Language: true</strong></td></tr><tr><td id="name"><strong>Character Set</strong></td><td id="default-value">UTF-8</td><td id="allowable-values"></td><td id="description">Specifies the character set of the document data.</td></tr><tr><td id="name"><strong>Batch Size</strong></td><td id="default-value">100</td><td id="allowable-values"></td><td id="description">The preferred number of FlowFiles to put to the database in a single transaction</td></tr><tr><td id="name"><strong>Index Operation</strong></td><td id="default-value">index</td><td id="allo
 wable-values"></td><td id="description">The type of the operation used to index (index, update, upsert)<br/><strong>Supports Expression Language: true</strong></td></tr></table><h3>Relationships: </h3><table id="relationships"><tr><th>Name</th><th>Description</th></tr><tr><td>retry</td><td>A FlowFile is routed to this relationship if the database cannot be updated but attempting the operation again may succeed</td></tr><tr><td>success</td><td>All FlowFiles that are written to Elasticsearch are routed to this relationship</td></tr><tr><td>failure</td><td>All FlowFiles that cannot be written to Elasticsearch are routed to this relationship</td></tr></table><h3>Reads Attributes: </h3>None specified.<h3>Writes Attributes: </h3>None specified.</body></html>
\ No newline at end of file

Modified: nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.CreateHadoopSequenceFile/index.html
URL: http://svn.apache.org/viewvc/nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.CreateHadoopSequenceFile/index.html?rev=1752359&r1=1752358&r2=1752359&view=diff
==============================================================================
--- nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.CreateHadoopSequenceFile/index.html (original)
+++ nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.CreateHadoopSequenceFile/index.html Tue Jul 12 22:38:04 2016
@@ -1 +1 @@
-<!DOCTYPE html><html lang="en"><head><meta charset="utf-8"></meta><title>CreateHadoopSequenceFile</title><link rel="stylesheet" href="../../css/component-usage.css" type="text/css"></link></head><body><h2>Description: </h2><p>Creates Hadoop Sequence Files from incoming flow files</p><p><a href="additionalDetails.html">Additional Details...</a></p><h3>Tags: </h3><p>hadoop, sequence file, create, sequencefile</p><h3>Properties: </h3><p>In the list below, the names of required properties appear in <strong>bold</strong>. Any other properties (not in bold) are considered optional. The table also indicates any default values.</p><table id="properties"><tr><th>Name</th><th>Default Value</th><th>Allowable Values</th><th>Description</th></tr><tr><td id="name">Hadoop Configuration Resources</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">A file or comma separated list of files which contains the Hadoop file system configuration. Without this, Hadoop will se
 arch the classpath for a 'core-site.xml' and 'hdfs-site.xml' file or will revert to a default configuration.</td></tr><tr><td id="name">Kerberos Principal</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Kerberos principal to authenticate as. Requires nifi.kerberos.krb5.file to be set in your nifi.properties</td></tr><tr><td id="name">Kerberos Keytab</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Kerberos keytab associated with the principal. Requires nifi.kerberos.krb5.file to be set in your nifi.properties</td></tr><tr><td id="name">Kerberos Relogin Period</td><td id="default-value">4 hours</td><td id="allowable-values"></td><td id="description">Period of time which should pass before attempting a kerberos relogin</td></tr><tr><td id="name">compression type</td><td id="default-value"></td><td id="allowable-values"><ul><li>NONE</li><li>DEFAULT</li><li>BZIP</li><li>GZIP</li><li>LZ4</li><li>SNAPPY</li><li>AUTOMAT
 IC</li></ul></td><td id="description">Type of compression to use when creating Sequence File</td></tr></table><h3>Relationships: </h3><table id="relationships"><tr><th>Name</th><th>Description</th></tr><tr><td>success</td><td>Generated Sequence Files are sent to this relationship</td></tr><tr><td>failure</td><td>Incoming files that failed to generate a Sequence File are sent to this relationship</td></tr></table><h3>Reads Attributes: </h3>None specified.<h3>Writes Attributes: </h3>None specified.<h3>See Also:</h3><p><a href="../org.apache.nifi.processors.hadoop.PutHDFS/index.html">PutHDFS</a></p></body></html>
\ No newline at end of file
+<!DOCTYPE html><html lang="en"><head><meta charset="utf-8"></meta><title>CreateHadoopSequenceFile</title><link rel="stylesheet" href="../../css/component-usage.css" type="text/css"></link></head><body><h2>Description: </h2><p>Creates Hadoop Sequence Files from incoming flow files</p><p><a href="additionalDetails.html">Additional Details...</a></p><h3>Tags: </h3><p>hadoop, sequence file, create, sequencefile</p><h3>Properties: </h3><p>In the list below, the names of required properties appear in <strong>bold</strong>. Any other properties (not in bold) are considered optional. The table also indicates any default values.</p><table id="properties"><tr><th>Name</th><th>Default Value</th><th>Allowable Values</th><th>Description</th></tr><tr><td id="name">Hadoop Configuration Resources</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">A file or comma separated list of files which contains the Hadoop file system configuration. Without this, Hadoop will se
 arch the classpath for a 'core-site.xml' and 'hdfs-site.xml' file or will revert to a default configuration.</td></tr><tr><td id="name">Kerberos Principal</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Kerberos principal to authenticate as. Requires nifi.kerberos.krb5.file to be set in your nifi.properties</td></tr><tr><td id="name">Kerberos Keytab</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Kerberos keytab associated with the principal. Requires nifi.kerberos.krb5.file to be set in your nifi.properties</td></tr><tr><td id="name">Kerberos Relogin Period</td><td id="default-value">4 hours</td><td id="allowable-values"></td><td id="description">Period of time which should pass before attempting a kerberos relogin</td></tr><tr><td id="name">compression type</td><td id="default-value"></td><td id="allowable-values"><ul><li>NONE</li><li>RECORD</li><li>BLOCK</li></ul></td><td id="description">Type of compression 
 to use when creating Sequence File</td></tr></table><h3>Relationships: </h3><table id="relationships"><tr><th>Name</th><th>Description</th></tr><tr><td>success</td><td>Generated Sequence Files are sent to this relationship</td></tr><tr><td>failure</td><td>Incoming files that failed to generate a Sequence File are sent to this relationship</td></tr></table><h3>Reads Attributes: </h3>None specified.<h3>Writes Attributes: </h3>None specified.<h3>See Also:</h3><p><a href="../org.apache.nifi.processors.hadoop.PutHDFS/index.html">PutHDFS</a></p></body></html>
\ No newline at end of file

Modified: nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.PutHDFS/index.html
URL: http://svn.apache.org/viewvc/nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.PutHDFS/index.html?rev=1752359&r1=1752358&r2=1752359&view=diff
==============================================================================
--- nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.PutHDFS/index.html (original)
+++ nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.PutHDFS/index.html Tue Jul 12 22:38:04 2016
@@ -1 +1 @@
-<!DOCTYPE html><html lang="en"><head><meta charset="utf-8"></meta><title>PutHDFS</title><link rel="stylesheet" href="../../css/component-usage.css" type="text/css"></link></head><body><h2>Description: </h2><p>Write FlowFile data to Hadoop Distributed File System (HDFS)</p><h3>Tags: </h3><p>hadoop, HDFS, put, copy, filesystem</p><h3>Properties: </h3><p>In the list below, the names of required properties appear in <strong>bold</strong>. Any other properties (not in bold) are considered optional. The table also indicates any default values, and whether a property supports the <a href="../../html/expression-language-guide.html">NiFi Expression Language</a>.</p><table id="properties"><tr><th>Name</th><th>Default Value</th><th>Allowable Values</th><th>Description</th></tr><tr><td id="name">Hadoop Configuration Resources</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">A file or comma separated list of files which contains the Hadoop file system configura
 tion. Without this, Hadoop will search the classpath for a 'core-site.xml' and 'hdfs-site.xml' file or will revert to a default configuration.</td></tr><tr><td id="name">Kerberos Principal</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Kerberos principal to authenticate as. Requires nifi.kerberos.krb5.file to be set in your nifi.properties</td></tr><tr><td id="name">Kerberos Keytab</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Kerberos keytab associated with the principal. Requires nifi.kerberos.krb5.file to be set in your nifi.properties</td></tr><tr><td id="name">Kerberos Relogin Period</td><td id="default-value">4 hours</td><td id="allowable-values"></td><td id="description">Period of time which should pass before attempting a kerberos relogin</td></tr><tr><td id="name"><strong>Directory</strong></td><td id="default-value"></td><td id="allowable-values"></td><td id="description">The parent HDFS directory t
 o which files should be written<br/><strong>Supports Expression Language: true</strong></td></tr><tr><td id="name"><strong>Conflict Resolution Strategy</strong></td><td id="default-value">fail</td><td id="allowable-values"><ul><li>replace</li><li>ignore</li><li>fail</li></ul></td><td id="description">Indicates what should happen when a file with the same name already exists in the output directory</td></tr><tr><td id="name">Block Size</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Size of each block as written to HDFS. This overrides the Hadoop Configuration</td></tr><tr><td id="name">IO Buffer Size</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Amount of memory to use to buffer file contents during IO. This overrides the Hadoop Configuration</td></tr><tr><td id="name">Replication</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Number of times that HDFS will replicate each f
 ile. This overrides the Hadoop Configuration</td></tr><tr><td id="name">Permissions umask</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">A umask represented as an octal number which determines the permissions of files written to HDFS. This overrides the Hadoop Configuration dfs.umaskmode</td></tr><tr><td id="name">Remote Owner</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Changes the owner of the HDFS file to this value after it is written. This only works if NiFi is running as a user that has HDFS super user privilege to change owner</td></tr><tr><td id="name">Remote Group</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Changes the group of the HDFS file to this value after it is written. This only works if NiFi is running as a user that has HDFS super user privilege to change group</td></tr><tr><td id="name"><strong>Compression codec</strong></td><td id="default-value">NO
 NE</td><td id="allowable-values"><ul><li>NONE</li><li>DEFAULT</li><li>BZIP</li><li>GZIP</li><li>LZ4</li><li>SNAPPY</li><li>AUTOMATIC</li></ul></td><td id="description">No Description Provided.</td></tr></table><h3>Relationships: </h3><table id="relationships"><tr><th>Name</th><th>Description</th></tr><tr><td>success</td><td>Files that have been successfully written to HDFS are transferred to this relationship</td></tr><tr><td>failure</td><td>Files that could not be written to HDFS for some reason are transferred to this relationship</td></tr></table><h3>Reads Attributes: </h3>None specified.<h3>Writes Attributes: </h3><table id="writes-attributes"><tr><th>Name</th><th>Description</th></tr><tr><td>filename</td><td>The name of the file written to HDFS comes from the value of this attribute.</td></tr></table><h3>See Also:</h3><p><a href="../org.apache.nifi.processors.hadoop.GetHDFS/index.html">GetHDFS</a></p></body></html>
\ No newline at end of file
+<!DOCTYPE html><html lang="en"><head><meta charset="utf-8"></meta><title>PutHDFS</title><link rel="stylesheet" href="../../css/component-usage.css" type="text/css"></link></head><body><h2>Description: </h2><p>Write FlowFile data to Hadoop Distributed File System (HDFS)</p><h3>Tags: </h3><p>hadoop, HDFS, put, copy, filesystem</p><h3>Properties: </h3><p>In the list below, the names of required properties appear in <strong>bold</strong>. Any other properties (not in bold) are considered optional. The table also indicates any default values, and whether a property supports the <a href="../../html/expression-language-guide.html">NiFi Expression Language</a>.</p><table id="properties"><tr><th>Name</th><th>Default Value</th><th>Allowable Values</th><th>Description</th></tr><tr><td id="name">Hadoop Configuration Resources</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">A file or comma separated list of files which contains the Hadoop file system configura
 tion. Without this, Hadoop will search the classpath for a 'core-site.xml' and 'hdfs-site.xml' file or will revert to a default configuration.</td></tr><tr><td id="name">Kerberos Principal</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Kerberos principal to authenticate as. Requires nifi.kerberos.krb5.file to be set in your nifi.properties</td></tr><tr><td id="name">Kerberos Keytab</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Kerberos keytab associated with the principal. Requires nifi.kerberos.krb5.file to be set in your nifi.properties</td></tr><tr><td id="name">Kerberos Relogin Period</td><td id="default-value">4 hours</td><td id="allowable-values"></td><td id="description">Period of time which should pass before attempting a kerberos relogin</td></tr><tr><td id="name"><strong>Directory</strong></td><td id="default-value"></td><td id="allowable-values"></td><td id="description">The parent HDFS directory t
 o which files should be written<br/><strong>Supports Expression Language: true</strong></td></tr><tr><td id="name"><strong>Conflict Resolution Strategy</strong></td><td id="default-value">fail</td><td id="allowable-values"><ul><li>replace</li><li>ignore</li><li>fail</li></ul></td><td id="description">Indicates what should happen when a file with the same name already exists in the output directory</td></tr><tr><td id="name">Block Size</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Size of each block as written to HDFS. This overrides the Hadoop Configuration</td></tr><tr><td id="name">IO Buffer Size</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Amount of memory to use to buffer file contents during IO. This overrides the Hadoop Configuration</td></tr><tr><td id="name">Replication</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Number of times that HDFS will replicate each f
 ile. This overrides the Hadoop Configuration</td></tr><tr><td id="name">Permissions umask</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">A umask represented as an octal number which determines the permissions of files written to HDFS. This overrides the Hadoop Configuration dfs.umaskmode</td></tr><tr><td id="name">Remote Owner</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Changes the owner of the HDFS file to this value after it is written. This only works if NiFi is running as a user that has HDFS super user privilege to change owner</td></tr><tr><td id="name">Remote Group</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Changes the group of the HDFS file to this value after it is written. This only works if NiFi is running as a user that has HDFS super user privilege to change group</td></tr><tr><td id="name"><strong>Compression codec</strong></td><td id="default-value">NO
 NE</td><td id="allowable-values"><ul><li>NONE</li><li>DEFAULT</li><li>BZIP</li><li>GZIP</li><li>LZ4</li><li>SNAPPY</li><li>AUTOMATIC</li></ul></td><td id="description">No Description Provided.</td></tr></table><h3>Relationships: </h3><table id="relationships"><tr><th>Name</th><th>Description</th></tr><tr><td>success</td><td>Files that have been successfully written to HDFS are transferred to this relationship</td></tr><tr><td>failure</td><td>Files that could not be written to HDFS for some reason are transferred to this relationship</td></tr></table><h3>Reads Attributes: </h3><table id="reads-attributes"><tr><th>Name</th><th>Description</th></tr><tr><td>filename</td><td>The name of the file written to HDFS comes from the value of this attribute.</td></tr></table><h3>Writes Attributes: </h3><table id="writes-attributes"><tr><th>Name</th><th>Description</th></tr><tr><td>filename</td><td>The name of the file written to HDFS is stored in this attribute.</td></tr><tr><td>absolute.hdfs.pa
 th</td><td>The absolute path to the file on HDFS is stored in this attribute.</td></tr></table><h3>See Also:</h3><p><a href="../org.apache.nifi.processors.hadoop.GetHDFS/index.html">GetHDFS</a></p></body></html>
\ No newline at end of file

Added: nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.inotify.GetHDFSEvents/index.html
URL: http://svn.apache.org/viewvc/nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.inotify.GetHDFSEvents/index.html?rev=1752359&view=auto
==============================================================================
--- nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.inotify.GetHDFSEvents/index.html (added)
+++ nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.inotify.GetHDFSEvents/index.html Tue Jul 12 22:38:04 2016
@@ -0,0 +1 @@
+<!DOCTYPE html><html lang="en"><head><meta charset="utf-8"></meta><title>GetHDFSEvents</title><link rel="stylesheet" href="../../css/component-usage.css" type="text/css"></link></head><body><h2>Description: </h2><p>This processor polls the notification events provided by the HdfsAdmin API. Since this uses the HdfsAdmin APIs it is required to run as an HDFS super user. Currently there are six types of events (append, close, create, metadata, rename, and unlink). Please see org.apache.hadoop.hdfs.inotify.Event documentation for full explanations of each event. This processor will poll for new events based on a defined duration. For each event received a new flow file will be created with the expected attributes and the event itself serialized to JSON and written to the flow file's content. For example, if event.type is APPEND then the content of the flow file will contain a JSON file containing the information about the append event. If successful the flow files are sent to the 'succe
 ss' relationship. Be careful of where the generated flow files are stored. If the flow files are stored in one of processor's watch directories there will be a never ending flow of events. It is also important to be aware that this processor must consume all events. The filtering must happen within the processor. This is because the HDFS admin's event notifications API does not have filtering.</p><h3>Tags: </h3><p>hadoop, events, inotify, notifications, filesystem</p><h3>Properties: </h3><p>In the list below, the names of required properties appear in <strong>bold</strong>. Any other properties (not in bold) are considered optional. The table also indicates any default values, and whether a property supports the <a href="../../html/expression-language-guide.html">NiFi Expression Language</a>.</p><table id="properties"><tr><th>Name</th><th>Default Value</th><th>Allowable Values</th><th>Description</th></tr><tr><td id="name">Hadoop Configuration Resources</td><td id="default-value"></
 td><td id="allowable-values"></td><td id="description">A file or comma separated list of files which contains the Hadoop file system configuration. Without this, Hadoop will search the classpath for a 'core-site.xml' and 'hdfs-site.xml' file or will revert to a default configuration.</td></tr><tr><td id="name">Kerberos Principal</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Kerberos principal to authenticate as. Requires nifi.kerberos.krb5.file to be set in your nifi.properties</td></tr><tr><td id="name">Kerberos Keytab</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Kerberos keytab associated with the principal. Requires nifi.kerberos.krb5.file to be set in your nifi.properties</td></tr><tr><td id="name">Kerberos Relogin Period</td><td id="default-value">4 hours</td><td id="allowable-values"></td><td id="description">Period of time which should pass before attempting a kerberos relogin</td></tr><tr><td id="na
 me"><strong>Poll Duration</strong></td><td id="default-value">1 second</td><td id="allowable-values"></td><td id="description">The time before the polling method returns with the next batch of events if they exist. It may exceed this amount of time by up to the time required for an RPC to the NameNode.</td></tr><tr><td id="name"><strong>HDFS Path to Watch</strong></td><td id="default-value"></td><td id="allowable-values"></td><td id="description">The HDFS path to get event notifications for. This property accepts both expression language and regular expressions. This will be evaluated during the OnScheduled phase.<br/><strong>Supports Expression Language: true</strong></td></tr><tr><td id="name"><strong>Ignore Hidden Files</strong></td><td id="default-value">false</td><td id="allowable-values"><ul><li>true</li><li>false</li></ul></td><td id="description">If true and the final component of the path associated with a given event starts with a '.' then that event will not be processed.
 </td></tr><tr><td id="name"><strong>Event Types to Filter On</strong></td><td id="default-value">append, close, create, metadata, rename, unlink</td><td id="allowable-values"></td><td id="description">A comma-separated list of event types to process. Valid event types are: append, close, create, metadata, rename, and unlink. Case does not matter.</td></tr><tr><td id="name"><strong>IOException Retries During Event Polling</strong></td><td id="default-value">3</td><td id="allowable-values"></td><td id="description">According to the HDFS admin API for event polling it is good to retry at least a few times. This number defines how many times the poll will be retried if it throws an IOException.</td></tr></table><h3>Relationships: </h3><table id="relationships"><tr><th>Name</th><th>Description</th></tr><tr><td>success</td><td>A flow file with updated information about a specific event will be sent to this relationship.</td></tr></table><h3>Reads Attributes: </h3>None specified.<h3>Writes
  Attributes: </h3><table id="writes-attributes"><tr><th>Name</th><th>Description</th></tr><tr><td>mime.type</td><td>This is always application/json.</td></tr><tr><td>hdfs.inotify.event.type</td><td>This will specify the specific HDFS notification event type. Currently there are six types of events (append, close, create, metadata, rename, and unlink).</td></tr><tr><td>hdfs.inotify.event.path</td><td>The specific path that the event is tied to.</td></tr></table><h3>See Also:</h3><p><a href="../org.apache.nifi.processors.hadoop.GetHDFS/index.html">GetHDFS</a>, <a href="../org.apache.nifi.processors.hadoop.FetchHDFS/index.html">FetchHDFS</a>, <a href="../org.apache.nifi.processors.hadoop.PutHDFS/index.html">PutHDFS</a>, <a href="../org.apache.nifi.processors.hadoop.ListHDFS/index.html">ListHDFS</a></p></body></html>
\ No newline at end of file

Added: nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hive.PutHiveQL/index.html
URL: http://svn.apache.org/viewvc/nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hive.PutHiveQL/index.html?rev=1752359&view=auto
==============================================================================
--- nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hive.PutHiveQL/index.html (added)
+++ nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hive.PutHiveQL/index.html Tue Jul 12 22:38:04 2016
@@ -0,0 +1 @@
+<!DOCTYPE html><html lang="en"><head><meta charset="utf-8"></meta><title>PutHiveQL</title><link rel="stylesheet" href="../../css/component-usage.css" type="text/css"></link></head><body><h2>Description: </h2><p>Executes a HiveQL DDL/DML command (UPDATE, INSERT, e.g.). The content of an incoming FlowFile is expected to be the HiveQL command to execute. The HiveQL command may use the ? to escape parameters. In this case, the parameters to use must exist as FlowFile attributes with the naming convention hiveql.args.N.type and hiveql.args.N.value, where N is a positive integer. The hiveql.args.N.type is expected to be a number indicating the JDBC Type. The content of the FlowFile is expected to be in UTF-8 format.</p><h3>Tags: </h3><p>sql, hive, put, database, update, insert</p><h3>Properties: </h3><p>In the list below, the names of required properties appear in <strong>bold</strong>. Any other properties (not in bold) are considered optional. The table also indicates any default values
 .</p><table id="properties"><tr><th>Name</th><th>Default Value</th><th>Allowable Values</th><th>Description</th></tr><tr><td id="name"><strong>Hive Database Connection Pooling Service</strong></td><td id="default-value"></td><td id="allowable-values"><strong>Controller Service API: </strong><br/>HiveDBCPService<br/><strong>Implementation:</strong><br/><a href="../org.apache.nifi.dbcp.hive.HiveConnectionPool/index.html">HiveConnectionPool</a></td><td id="description">The Hive Controller Service that is used to obtain connection(s) to the Hive database</td></tr><tr><td id="name"><strong>Batch Size</strong></td><td id="default-value">100</td><td id="allowable-values"></td><td id="description">The preferred number of FlowFiles to put to the database in a single transaction</td></tr><tr><td id="name"><strong>Character Set</strong></td><td id="default-value">UTF-8</td><td id="allowable-values"></td><td id="description">Specifies the character set of the record data.</td></tr></table><h3>R
 elationships: </h3><table id="relationships"><tr><th>Name</th><th>Description</th></tr><tr><td>retry</td><td>A FlowFile is routed to this relationship if the database cannot be updated but attempting the operation again may succeed</td></tr><tr><td>success</td><td>A FlowFile is routed to this relationship after the database is successfully updated</td></tr><tr><td>failure</td><td>A FlowFile is routed to this relationship if the database cannot be updated and retrying the operation will also fail, such as an invalid query or an integrity constraint violation</td></tr></table><h3>Reads Attributes: </h3><table id="reads-attributes"><tr><th>Name</th><th>Description</th></tr><tr><td>hiveql.args.N.type</td><td>Incoming FlowFiles are expected to be parameterized HiveQL statements. The type of each Parameter is specified as an integer that represents the JDBC Type of the parameter.</td></tr><tr><td>hiveql.args.N.value</td><td>Incoming FlowFiles are expected to be parameterized HiveQL statem
 ents. The value of the Parameters are specified as hiveql.args.1.value, hiveql.args.2.value, hiveql.args.3.value, and so on. The type of the hiveql.args.1.value Parameter is specified by the hiveql.args.1.type attribute.</td></tr></table><h3>Writes Attributes: </h3>None specified.<h3>See Also:</h3><p><a href="../org.apache.nifi.processors.hive.SelectHiveQL/index.html">SelectHiveQL</a></p></body></html>
\ No newline at end of file

Added: nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hive.SelectHiveQL/index.html
URL: http://svn.apache.org/viewvc/nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hive.SelectHiveQL/index.html?rev=1752359&view=auto
==============================================================================
--- nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hive.SelectHiveQL/index.html (added)
+++ nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hive.SelectHiveQL/index.html Tue Jul 12 22:38:04 2016
@@ -0,0 +1 @@
+<!DOCTYPE html><html lang="en"><head><meta charset="utf-8"></meta><title>SelectHiveQL</title><link rel="stylesheet" href="../../css/component-usage.css" type="text/css"></link></head><body><h2>Description: </h2><p>Execute provided HiveQL SELECT query against a Hive database connection. Query result will be converted to Avro or CSV format. Streaming is used so arbitrarily large result sets are supported. This processor can be scheduled to run on a timer, or cron expression, using the standard scheduling methods, or it can be triggered by an incoming FlowFile. If it is triggered by an incoming FlowFile, then attributes of that FlowFile will be available when evaluating the select query. FlowFile attribute 'selecthiveql.row.count' indicates how many rows were selected.</p><h3>Tags: </h3><p>hive, sql, select, jdbc, query, database</p><h3>Properties: </h3><p>In the list below, the names of required properties appear in <strong>bold</strong>. Any other properties (not in bold) are conside
 red optional. The table also indicates any default values, and whether a property supports the <a href="../../html/expression-language-guide.html">NiFi Expression Language</a>.</p><table id="properties"><tr><th>Name</th><th>Default Value</th><th>Allowable Values</th><th>Description</th></tr><tr><td id="name"><strong>Hive Database Connection Pooling Service</strong></td><td id="default-value"></td><td id="allowable-values"><strong>Controller Service API: </strong><br/>HiveDBCPService<br/><strong>Implementation:</strong><br/><a href="../org.apache.nifi.dbcp.hive.HiveConnectionPool/index.html">HiveConnectionPool</a></td><td id="description">The Hive Controller Service that is used to obtain connection(s) to the Hive database</td></tr><tr><td id="name"><strong>HiveQL Select Query</strong></td><td id="default-value"></td><td id="allowable-values"></td><td id="description">HiveQL SELECT query to execute<br/><strong>Supports Expression Language: true</strong></td></tr><tr><td id="name"><st
 rong>Output Format</strong></td><td id="default-value">Avro</td><td id="allowable-values"><ul><li>Avro</li><li>CSV</li></ul></td><td id="description">How to represent the records coming from Hive (Avro, CSV, e.g.)</td></tr></table><h3>Relationships: </h3><table id="relationships"><tr><th>Name</th><th>Description</th></tr><tr><td>success</td><td>Successfully created FlowFile from HiveQL query result set.</td></tr><tr><td>failure</td><td>HiveQL query execution failed. Incoming FlowFile will be penalized and routed to this relationship</td></tr></table><h3>Reads Attributes: </h3>None specified.<h3>Writes Attributes: </h3><table id="writes-attributes"><tr><th>Name</th><th>Description</th></tr><tr><td>mime.type</td><td>Sets the MIME type for the outgoing flowfile to application/avro-binary for Avro or text/csv for CSV.</td></tr><tr><td>filename</td><td>Adds .avro or .csv to the filename attribute depending on which output format is selected.</td></tr><tr><td>selecthiveql.row.count</td><t
 d>Indicates how many rows were selected/returned by the query.</td></tr></table></body></html>
\ No newline at end of file

Modified: nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.kafka.PutKafka/index.html
URL: http://svn.apache.org/viewvc/nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.kafka.PutKafka/index.html?rev=1752359&r1=1752358&r2=1752359&view=diff
==============================================================================
--- nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.kafka.PutKafka/index.html (original)
+++ nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.kafka.PutKafka/index.html Tue Jul 12 22:38:04 2016
@@ -1 +1 @@
-<!DOCTYPE html><html lang="en"><head><meta charset="utf-8"></meta><title>PutKafka</title><link rel="stylesheet" href="../../css/component-usage.css" type="text/css"></link></head><body><h2>Description: </h2><p>Sends the contents of a FlowFile as a message to Apache Kafka. The messages to send may be individual FlowFiles or may be delimited, using a user-specified delimiter, such as a new-line.</p><p><a href="additionalDetails.html">Additional Details...</a></p><h3>Tags: </h3><p>Apache, Kafka, Put, Send, Message, PubSub</p><h3>Properties: </h3><p>In the list below, the names of required properties appear in <strong>bold</strong>. Any other properties (not in bold) are considered optional. The table also indicates any default values, and whether a property supports the <a href="../../html/expression-language-guide.html">NiFi Expression Language</a>.</p><table id="properties"><tr><th>Name</th><th>Default Value</th><th>Allowable Values</th><th>Description</th></tr><tr><td id="name"><str
 ong>Known Brokers</strong></td><td id="default-value"></td><td id="allowable-values"></td><td id="description">A comma-separated list of known Kafka Brokers in the format &lt;host&gt;:&lt;port&gt;</td></tr><tr><td id="name"><strong>Topic Name</strong></td><td id="default-value"></td><td id="allowable-values"></td><td id="description">The Kafka Topic of interest<br/><strong>Supports Expression Language: true</strong></td></tr><tr><td id="name"><strong>Partition Strategy</strong></td><td id="default-value">Round Robin</td><td id="allowable-values"><ul><li>Round Robin <img src="../../html/images/iconInfo.png" alt="Messages will be assigned partitions in a round-robin fashion, sending the first message to Partition 1, the next Partition to Partition 2, and so on, wrapping as necessary." title="Messages will be assigned partitions in a round-robin fashion, sending the first message to Partition 1, the next Partition to Partition 2, and so on, wrapping as necessary."></img></li><li>Random
  <img src="../../html/images/iconInfo.png" alt="Messages will be assigned to random partitions." title="Messages will be assigned to random partitions."></img></li><li>User-Defined <img src="../../html/images/iconInfo.png" alt="The &lt;Partition&gt; property will be used to determine the partition. All messages within the same FlowFile will be assigned to the same partition." title="The &lt;Partition&gt; property will be used to determine the partition. All messages within the same FlowFile will be assigned to the same partition."></img></li></ul></td><td id="description">Specifies how messages should be partitioned when sent to Kafka</td></tr><tr><td id="name">Partition</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Specifies which Kafka Partition to add the message to. If using a message delimiter, all messages in the same FlowFile will be sent to the same partition. If a partition is specified but is not valid, then all messages within the sam
 e FlowFile will use the same partition but it remains undefined which partition is used.<br/><strong>Supports Expression Language: true</strong></td></tr><tr><td id="name">Kafka Key</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">The Key to use for the Message<br/><strong>Supports Expression Language: true</strong></td></tr><tr><td id="name"><strong>Delivery Guarantee</strong></td><td id="default-value">0</td><td id="allowable-values"><ul><li>Best Effort <img src="../../html/images/iconInfo.png" alt="FlowFile will be routed to success after successfully writing the content to a Kafka node, without waiting for a response. This provides the best performance but may result in data loss." title="FlowFile will be routed to success after successfully writing the content to a Kafka node, without waiting for a response. This provides the best performance but may result in data loss."></img></li><li>Guarantee Single Node Delivery <img src="../../html/image
 s/iconInfo.png" alt="FlowFile will be routed to success if the message is received by a single Kafka node, whether or not it is replicated. This is faster than &lt;Guarantee Replicated Delivery&gt; but can result in data loss if a Kafka node crashes" title="FlowFile will be routed to success if the message is received by a single Kafka node, whether or not it is replicated. This is faster than &lt;Guarantee Replicated Delivery&gt; but can result in data loss if a Kafka node crashes"></img></li><li>Guarantee Replicated Delivery <img src="../../html/images/iconInfo.png" alt="FlowFile will be routed to failure unless the message is replicated to the appropriate number of Kafka Nodes according to the Topic configuration" title="FlowFile will be routed to failure unless the message is replicated to the appropriate number of Kafka Nodes according to the Topic configuration"></img></li></ul></td><td id="description">Specifies the requirement for guaranteeing that a message is sent to Kafka
 </td></tr><tr><td id="name">Message Delimiter</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Specifies the delimiter (interpreted in its UTF-8 byte representation) to use for splitting apart multiple messages within a single FlowFile. If not specified, the entire content of the FlowFile will be used as a single message. If specified, the contents of the FlowFile will be split on this delimiter and each section sent as a separate Kafka message. Note that if messages are delimited and some messages for a given FlowFile are transferred successfully while others are not, the messages will be split into individual FlowFiles, such that those messages that were successfully sent are routed to the 'success' relationship while other messages are sent to the 'failure' relationship.<br/><strong>Supports Expression Language: true</strong></td></tr><tr><td id="name"><strong>Max Buffer Size</strong></td><td id="default-value">5 MB</td><td id="allowable-values"
 ></td><td id="description">The maximum amount of data to buffer in memory before sending to Kafka</td></tr><tr><td id="name"><strong>Max Record Size</strong></td><td id="default-value">1 MB</td><td id="allowable-values"></td><td id="description">The maximum size that any individual record can be.</td></tr><tr><td id="name"><strong>Communications Timeout</strong></td><td id="default-value">30 secs</td><td id="allowable-values"></td><td id="description">The amount of time to wait for a response from Kafka before determining that there is a communications error</td></tr><tr><td id="name"><strong>Batch Size</strong></td><td id="default-value">200</td><td id="allowable-values"></td><td id="description">The number of messages to send in one batch. The producer will wait until either this number of messages are ready to send or "Queue Buffering Max Time" is reached. NOTE: This property will be ignored unless the 'Message Delimiter' property is specified.</td></tr><tr><td id="name"><strong>
 Queue Buffering Max Time</strong></td><td id="default-value">5 secs</td><td id="allowable-values"></td><td id="description">Maximum time to buffer data before sending to Kafka. For example a setting of 100 ms will try to batch together 100 milliseconds' worth of messages to send at once. This will improve throughput but adds message delivery latency due to the buffering.</td></tr><tr><td id="name"><strong>Compression Codec</strong></td><td id="default-value">none</td><td id="allowable-values"><ul><li>None <img src="../../html/images/iconInfo.png" alt="Compression will not be used for any topic." title="Compression will not be used for any topic."></img></li><li>GZIP <img src="../../html/images/iconInfo.png" alt="Compress messages using GZIP" title="Compress messages using GZIP"></img></li><li>Snappy <img src="../../html/images/iconInfo.png" alt="Compress messages using Snappy" title="Compress messages using Snappy"></img></li></ul></td><td id="description">This parameter allows you 
 to specify the compression codec for all data generated by this producer.</td></tr><tr><td id="name"><strong>Client Name</strong></td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Client Name to use when communicating with Kafka</td></tr></table><h3>Dynamic Properties: </h3><p>Dynamic Properties allow the user to specify both the name and value of a property.<table id="dynamic-properties"><tr><th>Name</th><th>Value</th><th>Description</th></tr><tr><td id="name">The name of a Kafka configuration property.</td><td id="value">The value of a given Kafka configuration property.</td><td>These properties will be added on the Kafka configuration after loading any provided configuration properties. In the event a dynamic property represents a property that was already set as part of the static properties, its value wil be overriden with warning message describing the override. For the list of available Kafka properties please refer to: http://kafka.apache.or
 g/documentation.html#configuration.</td></tr></table></p><h3>Relationships: </h3><table id="relationships"><tr><th>Name</th><th>Description</th></tr><tr><td>success</td><td>Any FlowFile that is successfully sent to Kafka will be routed to this Relationship</td></tr><tr><td>failure</td><td>Any FlowFile that cannot be sent to Kafka will be routed to this Relationship</td></tr></table><h3>Reads Attributes: </h3>None specified.<h3>Writes Attributes: </h3>None specified.</body></html>
\ No newline at end of file
+<!DOCTYPE html><html lang="en"><head><meta charset="utf-8"></meta><title>PutKafka</title><link rel="stylesheet" href="../../css/component-usage.css" type="text/css"></link></head><body><h2>Description: </h2><p>Sends the contents of a FlowFile as a message to Apache Kafka. The messages to send may be individual FlowFiles or may be delimited, using a user-specified delimiter, such as a new-line.</p><p><a href="additionalDetails.html">Additional Details...</a></p><h3>Tags: </h3><p>Apache, Kafka, Put, Send, Message, PubSub</p><h3>Properties: </h3><p>In the list below, the names of required properties appear in <strong>bold</strong>. Any other properties (not in bold) are considered optional. The table also indicates any default values, and whether a property supports the <a href="../../html/expression-language-guide.html">NiFi Expression Language</a>.</p><table id="properties"><tr><th>Name</th><th>Default Value</th><th>Allowable Values</th><th>Description</th></tr><tr><td id="name"><str
 ong>Known Brokers</strong></td><td id="default-value"></td><td id="allowable-values"></td><td id="description">A comma-separated list of known Kafka Brokers in the format &lt;host&gt;:&lt;port&gt;</td></tr><tr><td id="name"><strong>Topic Name</strong></td><td id="default-value"></td><td id="allowable-values"></td><td id="description">The Kafka Topic of interest<br/><strong>Supports Expression Language: true</strong></td></tr><tr><td id="name"><strong>Partition Strategy</strong></td><td id="default-value">Round Robin</td><td id="allowable-values"><ul><li>Round Robin <img src="../../html/images/iconInfo.png" alt="Messages will be assigned partitions in a round-robin fashion, sending the first message to Partition 1, the next Partition to Partition 2, and so on, wrapping as necessary." title="Messages will be assigned partitions in a round-robin fashion, sending the first message to Partition 1, the next Partition to Partition 2, and so on, wrapping as necessary."></img></li><li>Random
  <img src="../../html/images/iconInfo.png" alt="Messages will be assigned to random partitions." title="Messages will be assigned to random partitions."></img></li><li>User-Defined <img src="../../html/images/iconInfo.png" alt="The &lt;Partition&gt; property will be used to determine the partition. All messages within the same FlowFile will be assigned to the same partition." title="The &lt;Partition&gt; property will be used to determine the partition. All messages within the same FlowFile will be assigned to the same partition."></img></li></ul></td><td id="description">Specifies how messages should be partitioned when sent to Kafka</td></tr><tr><td id="name">Partition</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Specifies which Kafka Partition to add the message to. If using a message delimiter, all messages in the same FlowFile will be sent to the same partition. If a partition is specified but is not valid, then all messages within the sam
 e FlowFile will use the same partition but it remains undefined which partition is used.<br/><strong>Supports Expression Language: true</strong></td></tr><tr><td id="name">Kafka Key</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">The Key to use for the Message<br/><strong>Supports Expression Language: true</strong></td></tr><tr><td id="name"><strong>Delivery Guarantee</strong></td><td id="default-value">0</td><td id="allowable-values"><ul><li>Best Effort <img src="../../html/images/iconInfo.png" alt="FlowFile will be routed to success after successfully writing the content to a Kafka node, without waiting for a response. This provides the best performance but may result in data loss." title="FlowFile will be routed to success after successfully writing the content to a Kafka node, without waiting for a response. This provides the best performance but may result in data loss."></img></li><li>Guarantee Single Node Delivery <img src="../../html/image
 s/iconInfo.png" alt="FlowFile will be routed to success if the message is received by a single Kafka node, whether or not it is replicated. This is faster than &lt;Guarantee Replicated Delivery&gt; but can result in data loss if a Kafka node crashes" title="FlowFile will be routed to success if the message is received by a single Kafka node, whether or not it is replicated. This is faster than &lt;Guarantee Replicated Delivery&gt; but can result in data loss if a Kafka node crashes"></img></li><li>Guarantee Replicated Delivery <img src="../../html/images/iconInfo.png" alt="FlowFile will be routed to failure unless the message is replicated to the appropriate number of Kafka Nodes according to the Topic configuration" title="FlowFile will be routed to failure unless the message is replicated to the appropriate number of Kafka Nodes according to the Topic configuration"></img></li></ul></td><td id="description">Specifies the requirement for guaranteeing that a message is sent to Kafka
 </td></tr><tr><td id="name">Message Delimiter</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Specifies the delimiter (interpreted in its UTF-8 byte representation) to use for splitting apart multiple messages within a single FlowFile. If not specified, the entire content of the FlowFile will be used as a single message. If specified, the contents of the FlowFile will be split on this delimiter and each section sent as a separate Kafka message. Note that if messages are delimited and some messages for a given FlowFile are transferred successfully while others are not, the messages will be split into individual FlowFiles, such that those messages that were successfully sent are routed to the 'success' relationship while other messages are sent to the 'failure' relationship.<br/><strong>Supports Expression Language: true</strong></td></tr><tr><td id="name"><strong>Max Buffer Size</strong></td><td id="default-value">5 MB</td><td id="allowable-values"
 ></td><td id="description">The maximum amount of data to buffer in memory before sending to Kafka</td></tr><tr><td id="name"><strong>Max Record Size</strong></td><td id="default-value">1 MB</td><td id="allowable-values"></td><td id="description">The maximum size that any individual record can be.</td></tr><tr><td id="name"><strong>Communications Timeout</strong></td><td id="default-value">30 secs</td><td id="allowable-values"></td><td id="description">The amount of time to wait for a response from Kafka before determining that there is a communications error</td></tr><tr><td id="name"><strong>Batch Size</strong></td><td id="default-value">16384</td><td id="allowable-values"></td><td id="description">This configuration controls the default batch size in bytes.The producer will attempt to batch records together into fewer requests whenever multiple records are being sent to the same partition. This helps performance on both the client and the server.</td></tr><tr><td id="name"><strong
 >Queue Buffering Max Time</strong></td><td id="default-value">5 secs</td><td id="allowable-values"></td><td id="description">Maximum time to buffer data before sending to Kafka. For example a setting of 100 ms will try to batch together 100 milliseconds' worth of messages to send at once. This will improve throughput but adds message delivery latency due to the buffering.</td></tr><tr><td id="name"><strong>Compression Codec</strong></td><td id="default-value">none</td><td id="allowable-values"><ul><li>None <img src="../../html/images/iconInfo.png" alt="Compression will not be used for any topic." title="Compression will not be used for any topic."></img></li><li>GZIP <img src="../../html/images/iconInfo.png" alt="Compress messages using GZIP" title="Compress messages using GZIP"></img></li><li>Snappy <img src="../../html/images/iconInfo.png" alt="Compress messages using Snappy" title="Compress messages using Snappy"></img></li></ul></td><td id="description">This parameter allows you
  to specify the compression codec for all data generated by this producer.</td></tr><tr><td id="name"><strong>Client Name</strong></td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Client Name to use when communicating with Kafka</td></tr></table><h3>Dynamic Properties: </h3><p>Dynamic Properties allow the user to specify both the name and value of a property.<table id="dynamic-properties"><tr><th>Name</th><th>Value</th><th>Description</th></tr><tr><td id="name">The name of a Kafka configuration property.</td><td id="value">The value of a given Kafka configuration property.</td><td>These properties will be added on the Kafka configuration after loading any provided configuration properties. In the event a dynamic property represents a property that was already set as part of the static properties, its value wil be overriden with warning message describing the override. For the list of available Kafka properties please refer to: http://kafka.apache.o
 rg/documentation.html#configuration.</td></tr></table></p><h3>Relationships: </h3><table id="relationships"><tr><th>Name</th><th>Description</th></tr><tr><td>success</td><td>Any FlowFile that is successfully sent to Kafka will be routed to this Relationship</td></tr><tr><td>failure</td><td>Any FlowFile that cannot be sent to Kafka will be routed to this Relationship</td></tr></table><h3>Reads Attributes: </h3>None specified.<h3>Writes Attributes: </h3>None specified.</body></html>
\ No newline at end of file

Added: nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.kafka.pubsub.ConsumeKafka/additionalDetails.html
URL: http://svn.apache.org/viewvc/nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.kafka.pubsub.ConsumeKafka/additionalDetails.html?rev=1752359&view=auto
==============================================================================
--- nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.kafka.pubsub.ConsumeKafka/additionalDetails.html (added)
+++ nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.kafka.pubsub.ConsumeKafka/additionalDetails.html Tue Jul 12 22:38:04 2016
@@ -0,0 +1,33 @@
+<!DOCTYPE html>
+<html lang="en">
+    <!--
+      Licensed to the Apache Software Foundation (ASF) under one or more
+      contributor license agreements.  See the NOTICE file distributed with
+      this work for additional information regarding copyright ownership.
+      The ASF licenses this file to You under the Apache License, Version 2.0
+      (the "License"); you may not use this file except in compliance with
+      the License.  You may obtain a copy of the License at
+          http://www.apache.org/licenses/LICENSE-2.0
+      Unless required by applicable law or agreed to in writing, software
+      distributed under the License is distributed on an "AS IS" BASIS,
+      WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+      See the License for the specific language governing permissions and
+      limitations under the License.
+    -->
+    <head>
+        <meta charset="utf-8" />
+        <title>ConsumeKafka</title>
+        <link rel="stylesheet" href="../../css/component-usage.css" type="text/css" />
+    </head>
+
+    <body>
+        <!-- Processor Documentation ================================================== -->
+        <h2>Description:</h2>
+        <p>
+            This Processors polls <a href="http://kafka.apache.org/">Apache Kafka</a>
+            for data using KafkaConsumer API available with Kafka 0.9+. When a message is received
+            from Kafka, this Processor emits a FlowFile where the content of the FlowFile is the value
+            of the Kafka message.
+        </p>
+    </body>
+</html>

Added: nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.kafka.pubsub.ConsumeKafka/index.html
URL: http://svn.apache.org/viewvc/nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.kafka.pubsub.ConsumeKafka/index.html?rev=1752359&view=auto
==============================================================================
--- nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.kafka.pubsub.ConsumeKafka/index.html (added)
+++ nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.kafka.pubsub.ConsumeKafka/index.html Tue Jul 12 22:38:04 2016
@@ -0,0 +1 @@
+<!DOCTYPE html><html lang="en"><head><meta charset="utf-8"></meta><title>ConsumeKafka</title><link rel="stylesheet" href="../../css/component-usage.css" type="text/css"></link></head><body><h2>Description: </h2><p>Consumes messages from Apache Kafka</p><p><a href="additionalDetails.html">Additional Details...</a></p><h3>Tags: </h3><p>Kafka, Get, Ingest, Ingress, Topic, PubSub, Consume</p><h3>Properties: </h3><p>In the list below, the names of required properties appear in <strong>bold</strong>. Any other properties (not in bold) are considered optional. The table also indicates any default values, and whether a property supports the <a href="../../html/expression-language-guide.html">NiFi Expression Language</a>.</p><table id="properties"><tr><th>Name</th><th>Default Value</th><th>Allowable Values</th><th>Description</th></tr><tr><td id="name"><strong>Kafka Brokers</strong></td><td id="default-value">localhost:9092</td><td id="allowable-values"></td><td id="description">A comma-sepa
 rated list of known Kafka Brokers in the format &lt;host&gt;:&lt;port&gt;<br/><strong>Supports Expression Language: true</strong></td></tr><tr><td id="name"><strong>Topic Name</strong></td><td id="default-value"></td><td id="allowable-values"></td><td id="description">The name of the Kafka Topic<br/><strong>Supports Expression Language: true</strong></td></tr><tr><td id="name"><strong>Client ID</strong></td><td id="default-value"></td><td id="allowable-values"></td><td id="description">String value uniquely identifying this client application. Corresponds to Kafka's 'client.id' property.<br/><strong>Supports Expression Language: true</strong></td></tr><tr><td id="name">Security Protocol</td><td id="default-value">PLAINTEXT</td><td id="allowable-values"><ul><li>PLAINTEXT <img src="../../html/images/iconInfo.png" alt="PLAINTEXT" title="PLAINTEXT"></img></li><li>SSL <img src="../../html/images/iconInfo.png" alt="SSL" title="SSL"></img></li><li>SASL_PLAINTEXT <img src="../../html/images
 /iconInfo.png" alt="SASL_PLAINTEXT" title="SASL_PLAINTEXT"></img></li><li>SASL_SSL <img src="../../html/images/iconInfo.png" alt="SASL_SSL" title="SASL_SSL"></img></li></ul></td><td id="description">Protocol used to communicate with brokers. Corresponds to Kafka's 'security.protocol' property.</td></tr><tr><td id="name">Kerberos Service Name</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">The Kerberos principal name that Kafka runs as. This can be defined either in Kafka's JAAS config or in Kafka's config. Corresponds to Kafka's 'security.protocol' property.It is ignored unless one of the SASL options of the &lt;Security Protocol&gt; are selected.</td></tr><tr><td id="name"><strong>Group ID</strong></td><td id="default-value"></td><td id="allowable-values"></td><td id="description">A Group ID is used to identify consumers that are within the same consumer group. Corresponds to Kafka's 'group.id' property.</td></tr><tr><td id="name"><strong>Offset 
 Reset</strong></td><td id="default-value">latest</td><td id="allowable-values"><ul><li>earliest <img src="../../html/images/iconInfo.png" alt="Automatically reset the offset to the earliest offset" title="Automatically reset the offset to the earliest offset"></img></li><li>latest <img src="../../html/images/iconInfo.png" alt="Automatically reset the offset to the latest offset" title="Automatically reset the offset to the latest offset"></img></li><li>none <img src="../../html/images/iconInfo.png" alt="Throw exception to the consumer if no previous offset is found for the consumer's group" title="Throw exception to the consumer if no previous offset is found for the consumer's group"></img></li></ul></td><td id="description">Allows you to manage the condition when there is no initial offset in Kafka or if the current offset does not exist any more on the server (e.g. because that data has been deleted). Corresponds to Kafka's 'auto.offset.reset' property.</td></tr><tr><td id="name"
 >Message Demarcator</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Since KafkaConsumer receives messages in batches, you have an option to output a single FlowFile which contains all Kafka messages in a single batch and this property allows you to provide a string (interpreted as UTF-8) to use for demarcating apart multiple Kafka messages. This is an optional property and if not provided each Kafka message received in a batch will result in a single FlowFile which essentially means that this processor may output multiple FlowFiles for each time it is triggered. To enter special character such as 'new line' use CTRL+Enter or Shift+Enter depending on the OS<br/><strong>Supports Expression Language: true</strong></td></tr></table><h3>Relationships: </h3><table id="relationships"><tr><th>Name</th><th>Description</th></tr><tr><td>success</td><td>All FlowFiles that are the are successfully sent to or received from Kafka are routed to this relationship<
 /td></tr></table><h3>Reads Attributes: </h3>None specified.<h3>Writes Attributes: </h3>None specified.</body></html>
\ No newline at end of file

Added: nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.kafka.pubsub.PublishKafka/additionalDetails.html
URL: http://svn.apache.org/viewvc/nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.kafka.pubsub.PublishKafka/additionalDetails.html?rev=1752359&view=auto
==============================================================================
--- nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.kafka.pubsub.PublishKafka/additionalDetails.html (added)
+++ nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.kafka.pubsub.PublishKafka/additionalDetails.html Tue Jul 12 22:38:04 2016
@@ -0,0 +1,47 @@
+<!DOCTYPE html>
+<html lang="en">
+    <!--
+      Licensed to the Apache Software Foundation (ASF) under one or more
+      contributor license agreements.  See the NOTICE file distributed with
+      this work for additional information regarding copyright ownership.
+      The ASF licenses this file to You under the Apache License, Version 2.0
+      (the "License"); you may not use this file except in compliance with
+      the License.  You may obtain a copy of the License at
+          http://www.apache.org/licenses/LICENSE-2.0
+      Unless required by applicable law or agreed to in writing, software
+      distributed under the License is distributed on an "AS IS" BASIS,
+      WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+      See the License for the specific language governing permissions and
+      limitations under the License.
+    -->
+    <head>
+        <meta charset="utf-8" />
+        <title>PublishKafka</title>
+        <link rel="stylesheet" href="../../css/component-usage.css" type="text/css" />
+    </head>
+
+    <body>
+        <!-- Processor Documentation ================================================== -->
+        <h2>Description:</h2>
+        <p>
+            This Processors puts the contents of a FlowFile to a Topic in
+            <a href="http://kafka.apache.org/">Apache Kafka</a> using KafkaProducer API available
+            with Kafka 0.9+ API. The content of a FlowFile becomes the contents of a Kafka message.
+            This message is optionally assigned a key by using the &lt;Kafka Key&gt; Property.
+        </p>
+
+        <p>
+            The Processor allows the user to configure an optional Message Demarcator that
+            can be used to send many messages per FlowFile. For example, a <i>\n</i> could be used
+            to indicate that the contents of the FlowFile should be used to send one message
+            per line of text. It also supports multi-char demarcators (e.g., 'my custom demarcator').
+            If the property is not set, the entire contents of the FlowFile
+            will be sent as a single message. When using the demarcator, if some messages are
+            successfully sent but other messages fail to send, the resulting FlowFile will be
+            considered a failed FlowFuile and will have additional attributes to that effect.
+            One of such attributes is 'failed.last.idx' which indicates the index of the last message
+            that was successfully ACKed by Kafka. (if no demarcator is used the value of this index will be -1).
+            This will allow PublishKafka to only re-send un-ACKed messages on the next re-try.
+        </p>
+    </body>
+</html>

Added: nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.kafka.pubsub.PublishKafka/index.html
URL: http://svn.apache.org/viewvc/nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.kafka.pubsub.PublishKafka/index.html?rev=1752359&view=auto
==============================================================================
--- nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.kafka.pubsub.PublishKafka/index.html (added)
+++ nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.kafka.pubsub.PublishKafka/index.html Tue Jul 12 22:38:04 2016
@@ -0,0 +1 @@
+<!DOCTYPE html><html lang="en"><head><meta charset="utf-8"></meta><title>PublishKafka</title><link rel="stylesheet" href="../../css/component-usage.css" type="text/css"></link></head><body><h2>Description: </h2><p>Sends the contents of a FlowFile as a message to Apache Kafka. The messages to send may be individual FlowFiles or may be delimited, using a user-specified delimiter, such as a new-line.</p><p><a href="additionalDetails.html">Additional Details...</a></p><h3>Tags: </h3><p>Apache, Kafka, Put, Send, Message, PubSub</p><h3>Properties: </h3><p>In the list below, the names of required properties appear in <strong>bold</strong>. Any other properties (not in bold) are considered optional. The table also indicates any default values, and whether a property supports the <a href="../../html/expression-language-guide.html">NiFi Expression Language</a>.</p><table id="properties"><tr><th>Name</th><th>Default Value</th><th>Allowable Values</th><th>Description</th></tr><tr><td id="name">
 <strong>Kafka Brokers</strong></td><td id="default-value">localhost:9092</td><td id="allowable-values"></td><td id="description">A comma-separated list of known Kafka Brokers in the format &lt;host&gt;:&lt;port&gt;<br/><strong>Supports Expression Language: true</strong></td></tr><tr><td id="name"><strong>Topic Name</strong></td><td id="default-value"></td><td id="allowable-values"></td><td id="description">The name of the Kafka Topic<br/><strong>Supports Expression Language: true</strong></td></tr><tr><td id="name"><strong>Client ID</strong></td><td id="default-value"></td><td id="allowable-values"></td><td id="description">String value uniquely identifying this client application. Corresponds to Kafka's 'client.id' property.<br/><strong>Supports Expression Language: true</strong></td></tr><tr><td id="name">Security Protocol</td><td id="default-value">PLAINTEXT</td><td id="allowable-values"><ul><li>PLAINTEXT <img src="../../html/images/iconInfo.png" alt="PLAINTEXT" title="PLAINTEXT"
 ></img></li><li>SSL <img src="../../html/images/iconInfo.png" alt="SSL" title="SSL"></img></li><li>SASL_PLAINTEXT <img src="../../html/images/iconInfo.png" alt="SASL_PLAINTEXT" title="SASL_PLAINTEXT"></img></li><li>SASL_SSL <img src="../../html/images/iconInfo.png" alt="SASL_SSL" title="SASL_SSL"></img></li></ul></td><td id="description">Protocol used to communicate with brokers. Corresponds to Kafka's 'security.protocol' property.</td></tr><tr><td id="name">Kerberos Service Name</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">The Kerberos principal name that Kafka runs as. This can be defined either in Kafka's JAAS config or in Kafka's config. Corresponds to Kafka's 'security.protocol' property.It is ignored unless one of the SASL options of the &lt;Security Protocol&gt; are selected.</td></tr><tr><td id="name"><strong>Delivery Guarantee</strong></td><td id="default-value">0</td><td id="allowable-values"><ul><li>Best Effort <img src="../../html/i
 mages/iconInfo.png" alt="FlowFile will be routed to success after successfully writing the content to a Kafka node, without waiting for a response. This provides the best performance but may result in data loss." title="FlowFile will be routed to success after successfully writing the content to a Kafka node, without waiting for a response. This provides the best performance but may result in data loss."></img></li><li>Guarantee Single Node Delivery <img src="../../html/images/iconInfo.png" alt="FlowFile will be routed to success if the message is received by a single Kafka node, whether or not it is replicated. This is faster than &lt;Guarantee Replicated Delivery&gt; but can result in data loss if a Kafka node crashes" title="FlowFile will be routed to success if the message is received by a single Kafka node, whether or not it is replicated. This is faster than &lt;Guarantee Replicated Delivery&gt; but can result in data loss if a Kafka node crashes"></img></li><li>Guarantee Repl
 icated Delivery <img src="../../html/images/iconInfo.png" alt="FlowFile will be routed to failure unless the message is replicated to the appropriate number of Kafka Nodes according to the Topic configuration" title="FlowFile will be routed to failure unless the message is replicated to the appropriate number of Kafka Nodes according to the Topic configuration"></img></li></ul></td><td id="description">Specifies the requirement for guaranteeing that a message is sent to Kafka. Corresponds to Kafka's 'acks' property.</td></tr><tr><td id="name">Kafka Key</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">The Key to use for the Message<br/><strong>Supports Expression Language: true</strong></td></tr><tr><td id="name">Message Demarcator</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Specifies the string (interpreted as UTF-8) to use for demarcating apart multiple messages within a single FlowFile. If not specified, th
 e entire content of the FlowFile will be used as a single message. If specified, the contents of the FlowFile will be split on this delimiter and each section sent as a separate Kafka message. To enter special character such as 'new line' use CTRL+Enter or Shift+Enter depending on your OS.<br/><strong>Supports Expression Language: true</strong></td></tr><tr><td id="name"><strong>Meta Data Wait Time</strong></td><td id="default-value">30 sec</td><td id="allowable-values"></td><td id="description">The amount of time KafkaConsumer will wait to obtain metadata during the 'send' call before failing the entire 'send' call. Corresponds to Kafka's 'max.block.ms' property<br/><strong>Supports Expression Language: true</strong></td></tr><tr><td id="name">Partitioner class</td><td id="default-value">org.apache.kafka.clients.producer.internals.DefaultPartitioner</td><td id="allowable-values"><ul><li>RoundRobinPartitioner <img src="../../html/images/iconInfo.png" alt="Messages will be assigned p
 artitions in a round-robin fashion, sending the first message to Partition 1, the next Partition to Partition 2, and so on, wrapping as necessary." title="Messages will be assigned partitions in a round-robin fashion, sending the first message to Partition 1, the next Partition to Partition 2, and so on, wrapping as necessary."></img></li><li>DefaultPartitioner <img src="../../html/images/iconInfo.png" alt="Messages will be assigned to random partitions." title="Messages will be assigned to random partitions."></img></li></ul></td><td id="description">Specifies which class to use to compute a partition id for a message. Corresponds to Kafka's 'partitioner.class' property.</td></tr><tr><td id="name"><strong>Compression Type</strong></td><td id="default-value">none</td><td id="allowable-values"><ul><li>none</li><li>gzip</li><li>snappy</li><li>lz4</li></ul></td><td id="description">This parameter allows you to specify the compression codec for all data generated by this producer.</td><
 /tr></table><h3>Dynamic Properties: </h3><p>Dynamic Properties allow the user to specify both the name and value of a property.<table id="dynamic-properties"><tr><th>Name</th><th>Value</th><th>Description</th></tr><tr><td id="name">The name of a Kafka configuration property.</td><td id="value">The value of a given Kafka configuration property.</td><td>These properties will be added on the Kafka configuration after loading any provided configuration properties. In the event a dynamic property represents a property that was already set, its value will be ignored and WARN message logged. For the list of available Kafka properties please refer to: http://kafka.apache.org/documentation.html#configuration.</td></tr></table></p><h3>Relationships: </h3><table id="relationships"><tr><th>Name</th><th>Description</th></tr><tr><td>success</td><td>All FlowFiles that are the are successfully sent to or received from Kafka are routed to this relationship</td></tr><tr><td>failure</td><td>Any FlowFi
 le that cannot be sent to Kafka will be routed to this Relationship</td></tr></table><h3>Reads Attributes: </h3>None specified.<h3>Writes Attributes: </h3>None specified.</body></html>
\ No newline at end of file

Modified: nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.kite.ConvertAvroSchema/index.html
URL: http://svn.apache.org/viewvc/nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.kite.ConvertAvroSchema/index.html?rev=1752359&r1=1752358&r2=1752359&view=diff
==============================================================================
--- nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.kite.ConvertAvroSchema/index.html (original)
+++ nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.kite.ConvertAvroSchema/index.html Tue Jul 12 22:38:04 2016
@@ -1 +1 @@
-<!DOCTYPE html><html lang="en"><head><meta charset="utf-8"></meta><title>ConvertAvroSchema</title><link rel="stylesheet" href="../../css/component-usage.css" type="text/css"></link></head><body><h2>Description: </h2><p>Convert records from one Avro schema to another, including support for flattening and simple type conversions</p><p><a href="additionalDetails.html">Additional Details...</a></p><h3>Tags: </h3><p>avro, convert, kite</p><h3>Properties: </h3><p>In the list below, the names of required properties appear in <strong>bold</strong>. Any other properties (not in bold) are considered optional. The table also indicates any default values, and whether a property supports the <a href="../../html/expression-language-guide.html">NiFi Expression Language</a>.</p><table id="properties"><tr><th>Name</th><th>Default Value</th><th>Allowable Values</th><th>Description</th></tr><tr><td id="name"><strong>Input Schema</strong></td><td id="default-value"></td><td id="allowable-values"></td><
 td id="description">Avro Schema of Input Flowfiles<br/><strong>Supports Expression Language: true</strong></td></tr><tr><td id="name"><strong>Output Schema</strong></td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Avro Schema of Output Flowfiles<br/><strong>Supports Expression Language: true</strong></td></tr><tr><td id="name">Locale</td><td id="default-value">default</td><td id="allowable-values"></td><td id="description">Locale to use for scanning data (see https://docs.oracle.com/javase/7/docs/api/java/util/Locale.html)or " default" for JVM default</td></tr></table><h3>Dynamic Properties: </h3><p>Dynamic Properties allow the user to specify both the name and value of a property.<table id="dynamic-properties"><tr><th>Name</th><th>Value</th><th>Description</th></tr><tr><td id="name">Field name from input schema</td><td id="value">Field name for output schema</td><td>Explicit mappings from input schema to output schema, which supports renaming fiel
 ds and stepping into nested records on the input schema using notation like parent.id</td></tr></table></p><h3>Relationships: </h3><table id="relationships"><tr><th>Name</th><th>Description</th></tr><tr><td>success</td><td>Avro content that converted successfully</td></tr><tr><td>failure</td><td>Avro content that failed to convert</td></tr></table><h3>Reads Attributes: </h3>None specified.<h3>Writes Attributes: </h3>None specified.</body></html>
\ No newline at end of file
+<!DOCTYPE html><html lang="en"><head><meta charset="utf-8"></meta><title>ConvertAvroSchema</title><link rel="stylesheet" href="../../css/component-usage.css" type="text/css"></link></head><body><h2>Description: </h2><p>Convert records from one Avro schema to another, including support for flattening and simple type conversions</p><p><a href="additionalDetails.html">Additional Details...</a></p><h3>Tags: </h3><p>avro, convert, kite</p><h3>Properties: </h3><p>In the list below, the names of required properties appear in <strong>bold</strong>. Any other properties (not in bold) are considered optional. The table also indicates any default values, and whether a property supports the <a href="../../html/expression-language-guide.html">NiFi Expression Language</a>.</p><table id="properties"><tr><th>Name</th><th>Default Value</th><th>Allowable Values</th><th>Description</th></tr><tr><td id="name"><strong>Input Schema</strong></td><td id="default-value"></td><td id="allowable-values"></td><
 td id="description">Avro Schema of Input Flowfiles.  This can be a URI (dataset, view, or resource) or literal JSON schema.<br/><strong>Supports Expression Language: true</strong></td></tr><tr><td id="name"><strong>Output Schema</strong></td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Avro Schema of Output Flowfiles.  This can be a URI (dataset, view, or resource) or literal JSON schema.<br/><strong>Supports Expression Language: true</strong></td></tr><tr><td id="name">Locale</td><td id="default-value">default</td><td id="allowable-values"></td><td id="description">Locale to use for scanning data (see https://docs.oracle.com/javase/7/docs/api/java/util/Locale.html)or " default" for JVM default</td></tr></table><h3>Dynamic Properties: </h3><p>Dynamic Properties allow the user to specify both the name and value of a property.<table id="dynamic-properties"><tr><th>Name</th><th>Value</th><th>Description</th></tr><tr><td id="name">Field name from input
  schema</td><td id="value">Field name for output schema</td><td>Explicit mappings from input schema to output schema, which supports renaming fields and stepping into nested records on the input schema using notation like parent.id</td></tr></table></p><h3>Relationships: </h3><table id="relationships"><tr><th>Name</th><th>Description</th></tr><tr><td>success</td><td>Avro content that converted successfully</td></tr><tr><td>failure</td><td>Avro content that failed to convert</td></tr></table><h3>Reads Attributes: </h3>None specified.<h3>Writes Attributes: </h3>None specified.</body></html>
\ No newline at end of file



Mime
View raw message