Modified: nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.aws.sqs.GetSQS/index.html URL: http://svn.apache.org/viewvc/nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.aws.sqs.GetSQS/index.html?rev=1719609&r1=1719608&r2=1719609&view=diff ============================================================================== --- nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.aws.sqs.GetSQS/index.html (original) +++ nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.aws.sqs.GetSQS/index.html Sat Dec 12 00:55:22 2015 @@ -1 +1 @@ -<!DOCTYPE html><html lang="en"><head><meta charset="utf-8"></meta><title>GetSQS</title><link rel="stylesheet" href="../../css/component-usage.css" type="text/css"></link></head><body><h2>Description: </h2><p>Fetches messages from an Amazon Simple Queuing Service Queue</p><h3>Tags: </h3><p>Amazon, AWS, SQS, Queue, Get, Fetch, Poll</p><h3>Properties: </h3><p>In the list below, the names of required properties appear in <strong>bold</strong>. Any other properties (not in bold) are considered optional. The table also indicates any default values, and whether a property is considered "sensitive", meaning that its value will be encrypted. Before entering a value in a sensitive property, ensure that the <strong>nifi.properties</strong> file has an entry for the property <strong>nifi.sensitive.props.key</strong>.</p><table id="properties"><tr><th>Name</th><th>Default Value</th><th>Allowable Values</th><th>Description</th></tr><tr><td id="name"><strong>Queue URL</strong></td><td id="default- value"></td><td id="allowable-values"></td><td id="description">The URL of the queue to act upon</td></tr><tr><td id="name"><strong>Auto Delete Messages</strong></td><td id="default-value">true</td><td id="allowable-values"><ul><li>true</li><li>false</li></ul></td><td id="description">Specifies whether the messages should be automatically deleted by the processors once they have been received.</td></tr><tr><td id="name">Access Key</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">No Description Provided.<br/><strong>Sensitive Property: true</strong></td></tr><tr><td id="name">Secret Key</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">No Description Provided.<br/><strong>Sensitive Property: true</strong></td></tr><tr><td id="name">Credentials File</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">No Description Provided.</td></tr><tr><td id="name"><strong>Region</strong></td><td id ="default-value">us-west-2</td><td id="allowable-values"><ul><li>us-gov-west-1 <img src="../../html/images/iconInfo.png" alt="us-gov-west-1" title="us-gov-west-1"></img></li><li>us-east-1 <img src="../../html/images/iconInfo.png" alt="us-east-1" title="us-east-1"></img></li><li>us-west-1 <img src="../../html/images/iconInfo.png" alt="us-west-1" title="us-west-1"></img></li><li>us-west-2 <img src="../../html/images/iconInfo.png" alt="us-west-2" title="us-west-2"></img></li><li>eu-west-1 <img src="../../html/images/iconInfo.png" alt="eu-west-1" title="eu-west-1"></img></li><li>eu-central-1 <img src="../../html/images/iconInfo.png" alt="eu-central-1" title="eu-central-1"></img></li><li>ap-southeast-1 <img src="../../html/images/iconInfo.png" alt="ap-southeast-1" title="ap-southeast-1"></img></li><li>ap-southeast-2 <img src="../../html/images/iconInfo.png" alt="ap-southeast-2" title="ap-southeast-2"></img></li><li>ap-northeast-1 <img src="../../html/images/iconInfo.png" alt="ap-northeas t-1" title="ap-northeast-1"></img></li><li>sa-east-1 <img src="../../html/images/iconInfo.png" alt="sa-east-1" title="sa-east-1"></img></li><li>cn-north-1 <img src="../../html/images/iconInfo.png" alt="cn-north-1" title="cn-north-1"></img></li></ul></td><td id="description">No Description Provided.</td></tr><tr><td id="name"><strong>Batch Size</strong></td><td id="default-value">10</td><td id="allowable-values"></td><td id="description">The maximum number of messages to send in a single network request</td></tr><tr><td id="name"><strong>Communications Timeout</strong></td><td id="default-value">30 secs</td><td id="allowable-values"></td><td id="description">No Description Provided.</td></tr><tr><td id="name"><strong>Character Set</strong></td><td id="default-value">UTF-8</td><td id="allowable-values"><ul><li>Big5</li><li>Big5-HKSCS</li><li>CESU-8</li><li>EUC-JP</li><li>EUC-KR</li><li>GB18030</li><li>GB2312</li><li>GBK</li><li>IBM-Thai</li><li>IBM00858</li><li>IBM01140</li><li>IBM011 41</li><li>IBM01142</li><li>IBM01143</li><li>IBM01144</li><li>IBM01145</li><li>IBM01146</li><li>IBM01147</li><li>IBM01148</li><li>IBM01149</li><li>IBM037</li><li>IBM1026</li><li>IBM1047</li><li>IBM273</li><li>IBM277</li><li>IBM278</li><li>IBM280</li><li>IBM284</li><li>IBM285</li><li>IBM290</li><li>IBM297</li><li>IBM420</li><li>IBM424</li><li>IBM437</li><li>IBM500</li><li>IBM775</li><li>IBM850</li><li>IBM852</li><li>IBM855</li><li>IBM857</li><li>IBM860</li><li>IBM861</li><li>IBM862</li><li>IBM863</li><li>IBM864</li><li>IBM865</li><li>IBM866</li><li>IBM868</li><li>IBM869</li><li>IBM870</li><li>IBM871</li><li>IBM918</li><li>ISO-2022-CN</li><li>ISO-2022-JP</li><li>ISO-2022-JP-2</li><li>ISO-2022-KR</li><li>ISO-8859-1</li><li>ISO-8859-13</li><li>ISO-8859-15</li><li>ISO-8859-2</li><li>ISO-8859-3</li><li>ISO-8859-4</li><li>ISO-8859-5</li><li>ISO-8859-6</li><li>ISO-8859-7</li><li>ISO-8859-8</li><li>ISO-8859-9</li><li>JIS_X0201</li><li>JIS_X0212-1990</li><li>KOI8-R</li><li>KOI8-U</li><li>Shif t_JIS</li><li>TIS-620</li><li>US-ASCII</li><li>UTF-16</li><li>UTF-16BE</li><li>UTF-16LE</li><li>UTF-32</li><li>UTF-32BE</li><li>UTF-32LE</li><li>UTF-8</li><li>windows-1250</li><li>windows-1251</li><li>windows-1252</li><li>windows-1253</li><li>windows-1254</li><li>windows-1255</li><li>windows-1256</li><li>windows-1257</li><li>windows-1258</li><li>windows-31j</li><li>x-Big5-HKSCS-2001</li><li>x-Big5-Solaris</li><li>x-COMPOUND_TEXT</li><li>x-euc-jp-linux</li><li>x-EUC-TW</li><li>x-eucJP-Open</li><li>x-IBM1006</li><li>x-IBM1025</li><li>x-IBM1046</li><li>x-IBM1097</li><li>x-IBM1098</li><li>x-IBM1112</li><li>x-IBM1122</li><li>x-IBM1123</li><li>x-IBM1124</li><li>x-IBM1166</li><li>x-IBM1364</li><li>x-IBM1381</li><li>x-IBM1383</li><li>x-IBM300</li><li>x-IBM33722</li><li>x-IBM737</li><li>x-IBM833</li><li>x-IBM834</li><li>x-IBM856</li><li>x-IBM874</li><li>x-IBM875</li><li>x-IBM921</li><li>x-IBM922</li><li>x-IBM930</li><li>x-IBM933</li><li>x-IBM935</li><li>x-IBM937</li><li>x-IBM939</li><li>x-IB M942</li><li>x-IBM942C</li><li>x-IBM943</li><li>x-IBM943C</li><li>x-IBM948</li><li>x-IBM949</li><li>x-IBM949C</li><li>x-IBM950</li><li>x-IBM964</li><li>x-IBM970</li><li>x-ISCII91</li><li>x-ISO-2022-CN-CNS</li><li>x-ISO-2022-CN-GB</li><li>x-iso-8859-11</li><li>x-JIS0208</li><li>x-JISAutoDetect</li><li>x-Johab</li><li>x-MacArabic</li><li>x-MacCentralEurope</li><li>x-MacCroatian</li><li>x-MacCyrillic</li><li>x-MacDingbat</li><li>x-MacGreek</li><li>x-MacHebrew</li><li>x-MacIceland</li><li>x-MacRoman</li><li>x-MacRomania</li><li>x-MacSymbol</li><li>x-MacThai</li><li>x-MacTurkish</li><li>x-MacUkraine</li><li>x-MS932_0213</li><li>x-MS950-HKSCS</li><li>x-MS950-HKSCS-XP</li><li>x-mswin-936</li><li>x-PCK</li><li>x-SJIS_0213</li><li>x-UTF-16LE-BOM</li><li>X-UTF-32BE-BOM</li><li>X-UTF-32LE-BOM</li><li>x-windows-50220</li><li>x-windows-50221</li><li>x-windows-874</li><li>x-windows-949</li><li>x-windows-950</li><li>x-windows-iso2022jp</li></ul></td><td id="description">The Character Set that shou ld be used to encode the textual content of the SQS message</td></tr><tr><td id="name"><strong>Visibility Timeout</strong></td><td id="default-value">15 mins</td><td id="allowable-values"></td><td id="description">The amount of time after a message is received but not deleted that the message is hidden from other consumers</td></tr></table><h3>Relationships: </h3><table id="relationships"><tr><th>Name</th><th>Description</th></tr><tr><td>success</td><td>FlowFiles are routed to success after being successfully copied to Amazon S3</td></tr></table><h3>Reads Attributes: </h3>None specified.<h3>Writes Attributes: </h3><table id="writes-attributes"><tr><th>Name</th><th>Description</th></tr><tr><td>hash.value</td><td>The MD5 sum of the message</td></tr><tr><td>hash.algorithm</td><td>MD5</td></tr><tr><td>sqs.message.id</td><td>The unique identifier of the SQS message</td></tr><tr><td>sqs.receipt.handle</td><td>The SQS Receipt Handle that is to be used to delete the message from the queue</ td></tr></table><h3>See Also:</h3><p><a href="../org.apache.nifi.processors.aws.sqs.PutSQS/index.html">PutSQS</a>, <a href="../org.apache.nifi.processors.aws.sqs.DeleteSQS/index.html">DeleteSQS</a></p></body></html> \ No newline at end of file +<!DOCTYPE html><html lang="en"><head><meta charset="utf-8"></meta><title>GetSQS</title><link rel="stylesheet" href="../../css/component-usage.css" type="text/css"></link></head><body><h2>Description: </h2><p>Fetches messages from an Amazon Simple Queuing Service Queue</p><h3>Tags: </h3><p>Amazon, AWS, SQS, Queue, Get, Fetch, Poll</p><h3>Properties: </h3><p>In the list below, the names of required properties appear in <strong>bold</strong>. Any other properties (not in bold) are considered optional. The table also indicates any default values, and whether a property is considered "sensitive", meaning that its value will be encrypted. Before entering a value in a sensitive property, ensure that the <strong>nifi.properties</strong> file has an entry for the property <strong>nifi.sensitive.props.key</strong>.</p><table id="properties"><tr><th>Name</th><th>Default Value</th><th>Allowable Values</th><th>Description</th></tr><tr><td id="name"><strong>Queue URL</strong></td><td id="default- value"></td><td id="allowable-values"></td><td id="description">The URL of the queue to act upon</td></tr><tr><td id="name"><strong>Auto Delete Messages</strong></td><td id="default-value">true</td><td id="allowable-values"><ul><li>true</li><li>false</li></ul></td><td id="description">Specifies whether the messages should be automatically deleted by the processors once they have been received.</td></tr><tr><td id="name">Access Key</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">No Description Provided.<br/><strong>Sensitive Property: true</strong></td></tr><tr><td id="name">Secret Key</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">No Description Provided.<br/><strong>Sensitive Property: true</strong></td></tr><tr><td id="name">Credentials File</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">No Description Provided.</td></tr><tr><td id="name"><strong>Region</strong></td><td id ="default-value">us-west-2</td><td id="allowable-values"><ul><li>us-gov-west-1 <img src="../../html/images/iconInfo.png" alt="us-gov-west-1" title="us-gov-west-1"></img></li><li>us-east-1 <img src="../../html/images/iconInfo.png" alt="us-east-1" title="us-east-1"></img></li><li>us-west-1 <img src="../../html/images/iconInfo.png" alt="us-west-1" title="us-west-1"></img></li><li>us-west-2 <img src="../../html/images/iconInfo.png" alt="us-west-2" title="us-west-2"></img></li><li>eu-west-1 <img src="../../html/images/iconInfo.png" alt="eu-west-1" title="eu-west-1"></img></li><li>eu-central-1 <img src="../../html/images/iconInfo.png" alt="eu-central-1" title="eu-central-1"></img></li><li>ap-southeast-1 <img src="../../html/images/iconInfo.png" alt="ap-southeast-1" title="ap-southeast-1"></img></li><li>ap-southeast-2 <img src="../../html/images/iconInfo.png" alt="ap-southeast-2" title="ap-southeast-2"></img></li><li>ap-northeast-1 <img src="../../html/images/iconInfo.png" alt="ap-northeas t-1" title="ap-northeast-1"></img></li><li>sa-east-1 <img src="../../html/images/iconInfo.png" alt="sa-east-1" title="sa-east-1"></img></li><li>cn-north-1 <img src="../../html/images/iconInfo.png" alt="cn-north-1" title="cn-north-1"></img></li></ul></td><td id="description">No Description Provided.</td></tr><tr><td id="name"><strong>Batch Size</strong></td><td id="default-value">10</td><td id="allowable-values"></td><td id="description">The maximum number of messages to send in a single network request</td></tr><tr><td id="name"><strong>Communications Timeout</strong></td><td id="default-value">30 secs</td><td id="allowable-values"></td><td id="description">No Description Provided.</td></tr><tr><td id="name"><strong>Character Set</strong></td><td id="default-value">UTF-8</td><td id="allowable-values"><ul><li>Big5</li><li>Big5-HKSCS</li><li>CESU-8</li><li>EUC-JP</li><li>EUC-KR</li><li>GB18030</li><li>GB2312</li><li>GBK</li><li>IBM-Thai</li><li>IBM00858</li><li>IBM01140</li><li>IBM011 41</li><li>IBM01142</li><li>IBM01143</li><li>IBM01144</li><li>IBM01145</li><li>IBM01146</li><li>IBM01147</li><li>IBM01148</li><li>IBM01149</li><li>IBM037</li><li>IBM1026</li><li>IBM1047</li><li>IBM273</li><li>IBM277</li><li>IBM278</li><li>IBM280</li><li>IBM284</li><li>IBM285</li><li>IBM290</li><li>IBM297</li><li>IBM420</li><li>IBM424</li><li>IBM437</li><li>IBM500</li><li>IBM775</li><li>IBM850</li><li>IBM852</li><li>IBM855</li><li>IBM857</li><li>IBM860</li><li>IBM861</li><li>IBM862</li><li>IBM863</li><li>IBM864</li><li>IBM865</li><li>IBM866</li><li>IBM868</li><li>IBM869</li><li>IBM870</li><li>IBM871</li><li>IBM918</li><li>ISO-2022-CN</li><li>ISO-2022-JP</li><li>ISO-2022-JP-2</li><li>ISO-2022-KR</li><li>ISO-8859-1</li><li>ISO-8859-13</li><li>ISO-8859-15</li><li>ISO-8859-2</li><li>ISO-8859-3</li><li>ISO-8859-4</li><li>ISO-8859-5</li><li>ISO-8859-6</li><li>ISO-8859-7</li><li>ISO-8859-8</li><li>ISO-8859-9</li><li>JIS_X0201</li><li>JIS_X0212-1990</li><li>KOI8-R</li><li>KOI8-U</li><li>Shif t_JIS</li><li>TIS-620</li><li>US-ASCII</li><li>UTF-16</li><li>UTF-16BE</li><li>UTF-16LE</li><li>UTF-32</li><li>UTF-32BE</li><li>UTF-32LE</li><li>UTF-8</li><li>windows-1250</li><li>windows-1251</li><li>windows-1252</li><li>windows-1253</li><li>windows-1254</li><li>windows-1255</li><li>windows-1256</li><li>windows-1257</li><li>windows-1258</li><li>windows-31j</li><li>x-Big5-HKSCS-2001</li><li>x-Big5-Solaris</li><li>x-COMPOUND_TEXT</li><li>x-euc-jp-linux</li><li>x-EUC-TW</li><li>x-eucJP-Open</li><li>x-IBM1006</li><li>x-IBM1025</li><li>x-IBM1046</li><li>x-IBM1097</li><li>x-IBM1098</li><li>x-IBM1112</li><li>x-IBM1122</li><li>x-IBM1123</li><li>x-IBM1124</li><li>x-IBM1166</li><li>x-IBM1364</li><li>x-IBM1381</li><li>x-IBM1383</li><li>x-IBM300</li><li>x-IBM33722</li><li>x-IBM737</li><li>x-IBM833</li><li>x-IBM834</li><li>x-IBM856</li><li>x-IBM874</li><li>x-IBM875</li><li>x-IBM921</li><li>x-IBM922</li><li>x-IBM930</li><li>x-IBM933</li><li>x-IBM935</li><li>x-IBM937</li><li>x-IBM939</li><li>x-IB M942</li><li>x-IBM942C</li><li>x-IBM943</li><li>x-IBM943C</li><li>x-IBM948</li><li>x-IBM949</li><li>x-IBM949C</li><li>x-IBM950</li><li>x-IBM964</li><li>x-IBM970</li><li>x-ISCII91</li><li>x-ISO-2022-CN-CNS</li><li>x-ISO-2022-CN-GB</li><li>x-iso-8859-11</li><li>x-JIS0208</li><li>x-JISAutoDetect</li><li>x-Johab</li><li>x-MacArabic</li><li>x-MacCentralEurope</li><li>x-MacCroatian</li><li>x-MacCyrillic</li><li>x-MacDingbat</li><li>x-MacGreek</li><li>x-MacHebrew</li><li>x-MacIceland</li><li>x-MacRoman</li><li>x-MacRomania</li><li>x-MacSymbol</li><li>x-MacThai</li><li>x-MacTurkish</li><li>x-MacUkraine</li><li>x-MS932_0213</li><li>x-MS950-HKSCS</li><li>x-MS950-HKSCS-XP</li><li>x-mswin-936</li><li>x-PCK</li><li>x-SJIS_0213</li><li>x-UTF-16LE-BOM</li><li>X-UTF-32BE-BOM</li><li>X-UTF-32LE-BOM</li><li>x-windows-50220</li><li>x-windows-50221</li><li>x-windows-874</li><li>x-windows-949</li><li>x-windows-950</li><li>x-windows-iso2022jp</li></ul></td><td id="description">The Character Set that shou ld be used to encode the textual content of the SQS message</td></tr><tr><td id="name"><strong>Visibility Timeout</strong></td><td id="default-value">15 mins</td><td id="allowable-values"></td><td id="description">The amount of time after a message is received but not deleted that the message is hidden from other consumers</td></tr><tr><td id="name"><strong>Receive Message Wait Time</strong></td><td id="default-value">0 sec</td><td id="allowable-values"></td><td id="description">The maximum amount of time to wait on a long polling receive call. Setting this to a value of 1 second or greater will reduce the number of SQS requests and decrease fetch latency at the cost of a constantly active thread.</td></tr></table><h3>Relationships: </h3><table id="relationships"><tr><th>Name</th><th>Description</th></tr><tr><td>success</td><td>FlowFiles are routed to success after being successfully copied to Amazon S3</td></tr></table><h3>Reads Attributes: </h3>None specified.<h3>Writes Attributes : </h3><table id="writes-attributes"><tr><th>Name</th><th>Description</th></tr><tr><td>hash.value</td><td>The MD5 sum of the message</td></tr><tr><td>hash.algorithm</td><td>MD5</td></tr><tr><td>sqs.message.id</td><td>The unique identifier of the SQS message</td></tr><tr><td>sqs.receipt.handle</td><td>The SQS Receipt Handle that is to be used to delete the message from the queue</td></tr></table><h3>See Also:</h3><p><a href="../org.apache.nifi.processors.aws.sqs.PutSQS/index.html">PutSQS</a>, <a href="../org.apache.nifi.processors.aws.sqs.DeleteSQS/index.html">DeleteSQS</a></p></body></html> \ No newline at end of file
Added: nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.azure.eventhub.GetAzureEventHub/index.html URL: http://svn.apache.org/viewvc/nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.azure.eventhub.GetAzureEventHub/index.html?rev=1719609&view=auto ============================================================================== --- nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.azure.eventhub.GetAzureEventHub/index.html (added) +++ nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.azure.eventhub.GetAzureEventHub/index.html Sat Dec 12 00:55:22 2015 @@ -0,0 +1 @@ +<!DOCTYPE html><html lang="en"><head><meta charset="utf-8"></meta><title>GetAzureEventHub</title><link rel="stylesheet" href="../../css/component-usage.css" type="text/css"></link></head><body><h2>Description: </h2><p>Receives messages from a Microsoft Azure Event Hub, writing the contents of the Azure message to the content of the FlowFile</p><h3>Tags: </h3><p>azure, microsoft, cloud, eventhub, events, streaming, streams</p><h3>Properties: </h3><p>In the list below, the names of required properties appear in <strong>bold</strong>. Any other properties (not in bold) are considered optional. The table also indicates any default values, and whether a property is considered "sensitive", meaning that its value will be encrypted. Before entering a value in a sensitive property, ensure that the <strong>nifi.properties</strong> file has an entry for the property <strong>nifi.sensitive.props.key</strong>.</p><table id="properties"><tr><th>Name</th><th>Default Value</th><th>Allowable Values< /th><th>Description</th></tr><tr><td id="name"><strong>Event Hub Name</strong></td><td id="default-value"></td><td id="allowable-values"></td><td id="description">The name of the Azure Event Hub to pull messages from</td></tr><tr><td id="name"><strong>Event Hub Namespace</strong></td><td id="default-value"></td><td id="allowable-values"></td><td id="description">The Azure Namespace that the Event Hub is assigned to. This is generally equal to <Event Hub Name>-ns</td></tr><tr><td id="name"><strong>Shared Access Policy Name</strong></td><td id="default-value"></td><td id="allowable-values"></td><td id="description">The name of the Event Hub Shared Access Policy. This Policy must have Listen permissions.</td></tr><tr><td id="name"><strong>Shared Access Policy Primary Key</strong></td><td id="default-value"></td><td id="allowable-values"></td><td id="description">The primary key of the Event Hub Shared Access Policy<br/><strong>Sensitive Property: true</strong></td></tr><tr><td id ="name"><strong>Number of Event Hub Partitions</strong></td><td id="default-value"></td><td id="allowable-values"></td><td id="description">The number of partitions that the Event Hub has. Only this number of partitions will be used, so it is important to ensure that if the number of partitions changes that this value be updated. Otherwise, some messages may not be consumed.</td></tr><tr><td id="name"><strong>Event Hub Consumer Group</strong></td><td id="default-value">$Default</td><td id="allowable-values"></td><td id="description">The name of the Event Hub Consumer Group to use when pulling events</td></tr></table><h3>Relationships: </h3><table id="relationships"><tr><th>Name</th><th>Description</th></tr><tr><td>success</td><td>Any FlowFile that is successfully received from the Azure Event Hub will be transferred to this Relationship.</td></tr></table><h3>Reads Attributes: </h3>None specified.<h3>Writes Attributes: </h3><table id="writes-attributes"><tr><th>Name</th><th>Descripti on</th></tr><tr><td>eventhub.enqueued.timestamp</td><td>The time (in milliseconds since epoch, UTC) at which the message was enqueued in the Azure Event Hub</td></tr><tr><td>eventhub.offset</td><td>The offset into the partition at which the message was stored</td></tr><tr><td>eventhub.sequence</td><td>The Azure Sequence number associated with the message</td></tr><tr><td>eventhub.name</td><td>The name of the Event Hub from which the message was pulled</td></tr><tr><td>eventhub.partition</td><td>The name of the Azure Partition from which the message was pulled</td></tr></table></body></html> \ No newline at end of file Added: nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.azure.eventhub.PutAzureEventHub/index.html URL: http://svn.apache.org/viewvc/nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.azure.eventhub.PutAzureEventHub/index.html?rev=1719609&view=auto ============================================================================== --- nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.azure.eventhub.PutAzureEventHub/index.html (added) +++ nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.azure.eventhub.PutAzureEventHub/index.html Sat Dec 12 00:55:22 2015 @@ -0,0 +1 @@ +<!DOCTYPE html><html lang="en"><head><meta charset="utf-8"></meta><title>PutAzureEventHub</title><link rel="stylesheet" href="../../css/component-usage.css" type="text/css"></link></head><body><h2>Description: </h2><p>Sends the contents of a FlowFile to a Windows Azure Event Hub. Note: the content of the FlowFile will be buffered into memory before being sent, so care should be taken to avoid sending FlowFiles to this Processor that exceed the amount of Java Heap Space available.</p><h3>Tags: </h3><p>microsoft, azure, cloud, eventhub, events, streams, streaming</p><h3>Properties: </h3><p>In the list below, the names of required properties appear in <strong>bold</strong>. Any other properties (not in bold) are considered optional. The table also indicates any default values, and whether a property is considered "sensitive", meaning that its value will be encrypted. Before entering a value in a sensitive property, ensure that the <strong>nifi.properties</strong> file has an entry for the property <strong>nifi.sensitive.props.key</strong>.</p><table id="properties"><tr><th>Name</th><th>Default Value</th><th>Allowable Values</th><th>Description</th></tr><tr><td id="name"><strong>Event Hub Name</strong></td><td id="default-value"></td><td id="allowable-values"></td><td id="description">The name of the Azure Event Hub to send to</td></tr><tr><td id="name"><strong>Event Hub Namespace</strong></td><td id="default-value"></td><td id="allowable-values"></td><td id="description">The Azure Namespace that the Event Hub is assigned to. This is generally equal to <Event Hub Name>-ns</td></tr><tr><td id="name"><strong>Shared Access Policy Name</strong></td><td id="default-value"></td><td id="allowable-values"></td><td id="description">The name of the Event Hub Shared Access Policy. This Policy must have Send permissions.</td></tr><tr><td id="name"><strong>Shared Access Policy Primary Key</strong></td><td id="default-value"></td><td id="allowable-values"></td><td id="des cription">The primary key of the Event Hub Shared Access Policy<br/><strong>Sensitive Property: true</strong></td></tr></table><h3>Relationships: </h3><table id="relationships"><tr><th>Name</th><th>Description</th></tr><tr><td>success</td><td>Any FlowFile that is successfully sent to the Azure Event Hub will be transferred to this Relationship.</td></tr><tr><td>failure</td><td>Any FlowFile that could not be sent to the Azure Event Hub will be transferred to this Relationship.</td></tr></table><h3>Reads Attributes: </h3>None specified.<h3>Writes Attributes: </h3>None specified.</body></html> \ No newline at end of file Added: nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.couchbase.GetCouchbaseKey/index.html URL: http://svn.apache.org/viewvc/nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.couchbase.GetCouchbaseKey/index.html?rev=1719609&view=auto ============================================================================== --- nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.couchbase.GetCouchbaseKey/index.html (added) +++ nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.couchbase.GetCouchbaseKey/index.html Sat Dec 12 00:55:22 2015 @@ -0,0 +1 @@ +<!DOCTYPE html><html lang="en"><head><meta charset="utf-8"></meta><title>GetCouchbaseKey</title><link rel="stylesheet" href="../../css/component-usage.css" type="text/css"></link></head><body><h2>Description: </h2><p>Get a document from Couchbase Server via Key/Value access. The ID of the document to fetch may be supplied by setting the <Document Id> property. NOTE: if the Document Id property is not set, the contents of the FlowFile will be read to determine the Document Id, which means that the contents of the entire FlowFile will be buffered in memory.</p><h3>Tags: </h3><p>nosql, couchbase, database, get</p><h3>Properties: </h3><p>In the list below, the names of required properties appear in <strong>bold</strong>. Any other properties (not in bold) are considered optional. The table also indicates any default values, and whether a property supports the <a href="../../html/expression-language-guide.html">NiFi Expression Language</a>.</p><table id="properties"><tr><th>Name</t h><th>Default Value</th><th>Allowable Values</th><th>Description</th></tr><tr><td id="name"><strong>Couchbase Cluster Controller Service</strong></td><td id="default-value"></td><td id="allowable-values"><strong>Controller Service API: </strong><br/>CouchbaseClusterControllerService<br/><strong>Implementation:</strong><br/><a href="../org.apache.nifi.couchbase.CouchbaseClusterService/index.html">CouchbaseClusterService</a></td><td id="description">A Couchbase Cluster Controller Service which manages connections to a Couchbase cluster.</td></tr><tr><td id="name"><strong>Bucket Name</strong></td><td id="default-value">default</td><td id="allowable-values"></td><td id="description">The name of bucket to access.</td></tr><tr><td id="name"><strong>Document Type</strong></td><td id="default-value">Json</td><td id="allowable-values"><ul><li>Json</li><li>Binary</li></ul></td><td id="description">The type of contents.</td></tr><tr><td id="name">Document Id</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">A static, fixed Couchbase document id, or an expression to construct the Couchbase document id.<br/><strong>Supports Expression Language: true</strong></td></tr></table><h3>Relationships: </h3><table id="relationships"><tr><th>Name</th><th>Description</th></tr><tr><td>retry</td><td>All FlowFiles that cannot written to Couchbase Server but can be retried are routed to this relationship.</td></tr><tr><td>success</td><td>All FlowFiles that are written to Couchbase Server are routed to this relationship.</td></tr><tr><td>failure</td><td>All FlowFiles that cannot written to Couchbase Server and can't be retried are routed to this relationship.</td></tr><tr><td>original</td><td>The original input file will be routed to this destination when it has been successfully processed.</td></tr></table><h3>Reads Attributes: </h3>None specified.<h3>Writes Attributes: </h3><table id="writes-attributes"><tr><th>Name</th><th>Description</th></tr><tr><td> couchbase.cluster</td><td>Cluster where the document was retrieved from.</td></tr><tr><td>couchbase.bucket</td><td>Bucket where the document was retrieved from.</td></tr><tr><td>couchbase.doc.id</td><td>Id of the document.</td></tr><tr><td>couchbase.doc.cas</td><td>CAS of the document.</td></tr><tr><td>couchbase.doc.expiry</td><td>Expiration of the document.</td></tr><tr><td>couchbase.exception</td><td>If Couchbase related error occurs the CouchbaseException class name will be captured here.</td></tr></table><h3>See Also:</h3><p><a href="../org.apache.nifi.couchbase.CouchbaseClusterControllerService/index.html">CouchbaseClusterControllerService</a></p></body></html> \ No newline at end of file Added: nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.couchbase.PutCouchbaseKey/index.html URL: http://svn.apache.org/viewvc/nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.couchbase.PutCouchbaseKey/index.html?rev=1719609&view=auto ============================================================================== --- nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.couchbase.PutCouchbaseKey/index.html (added) +++ nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.couchbase.PutCouchbaseKey/index.html Sat Dec 12 00:55:22 2015 @@ -0,0 +1 @@ +<!DOCTYPE html><html lang="en"><head><meta charset="utf-8"></meta><title>PutCouchbaseKey</title><link rel="stylesheet" href="../../css/component-usage.css" type="text/css"></link></head><body><h2>Description: </h2><p>Put a document to Couchbase Server via Key/Value access.</p><h3>Tags: </h3><p>nosql, couchbase, database, put</p><h3>Properties: </h3><p>In the list below, the names of required properties appear in <strong>bold</strong>. Any other properties (not in bold) are considered optional. The table also indicates any default values, and whether a property supports the <a href="../../html/expression-language-guide.html">NiFi Expression Language</a>.</p><table id="properties"><tr><th>Name</th><th>Default Value</th><th>Allowable Values</th><th>Description</th></tr><tr><td id="name"><strong>Couchbase Cluster Controller Service</strong></td><td id="default-value"></td><td id="allowable-values"><strong>Controller Service API: </strong><br/>CouchbaseClusterControllerService<br/><stron g>Implementation:</strong><br/><a href="../org.apache.nifi.couchbase.CouchbaseClusterService/index.html">CouchbaseClusterService</a></td><td id="description">A Couchbase Cluster Controller Service which manages connections to a Couchbase cluster.</td></tr><tr><td id="name"><strong>Bucket Name</strong></td><td id="default-value">default</td><td id="allowable-values"></td><td id="description">The name of bucket to access.</td></tr><tr><td id="name"><strong>Document Type</strong></td><td id="default-value">Json</td><td id="allowable-values"><ul><li>Json</li><li>Binary</li></ul></td><td id="description">The type of contents.</td></tr><tr><td id="name">Document Id</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">A static, fixed Couchbase document id, or an expression to construct the Couchbase document id.<br/><strong>Supports Expression Language: true</strong></td></tr><tr><td id="name"><strong>Persist To</strong></td><td id="default-value">NONE</td><t d id="allowable-values"><ul><li>MASTER</li><li>NONE</li><li>ONE</li><li>TWO</li><li>THREE</li><li>FOUR</li></ul></td><td id="description">Durability constraint about disk persistence.</td></tr><tr><td id="name"><strong>Replicate To</strong></td><td id="default-value">NONE</td><td id="allowable-values"><ul><li>NONE</li><li>ONE</li><li>TWO</li><li>THREE</li></ul></td><td id="description">Durability constraint about replication.</td></tr></table><h3>Relationships: </h3><table id="relationships"><tr><th>Name</th><th>Description</th></tr><tr><td>retry</td><td>All FlowFiles that cannot written to Couchbase Server but can be retried are routed to this relationship.</td></tr><tr><td>success</td><td>All FlowFiles that are written to Couchbase Server are routed to this relationship.</td></tr><tr><td>failure</td><td>All FlowFiles that cannot written to Couchbase Server and can't be retried are routed to this relationship.</td></tr></table><h3>Reads Attributes: </h3><table id="reads-attributes" ><tr><th>Name</th><th>Description</th></tr><tr><td>uuid</td><td>Used as a >document id if 'Document Id' is not specified</td></tr></table><h3>Writes >Attributes: </h3><table >id="writes-attributes"><tr><th>Name</th><th>Description</th></tr><tr><td>couchbase.cluster</td><td>Cluster > where the document was >stored.</td></tr><tr><td>couchbase.bucket</td><td>Bucket where the document >was stored.</td></tr><tr><td>couchbase.doc.id</td><td>Id of the >document.</td></tr><tr><td>couchbase.doc.cas</td><td>CAS of the >document.</td></tr><tr><td>couchbase.doc.expiry</td><td>Expiration of the >document.</td></tr><tr><td>couchbase.exception</td><td>If Couchbase related >error occurs the CouchbaseException class name will be captured >here.</td></tr></table><h3>See Also:</h3><p><a >href="../org.apache.nifi.couchbase.CouchbaseClusterControllerService/index.html">CouchbaseClusterControllerService</a></p></body></html> \ No newline at end of file Modified: nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.CreateHadoopSequenceFile/index.html URL: http://svn.apache.org/viewvc/nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.CreateHadoopSequenceFile/index.html?rev=1719609&r1=1719608&r2=1719609&view=diff ============================================================================== --- nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.CreateHadoopSequenceFile/index.html (original) +++ nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.CreateHadoopSequenceFile/index.html Sat Dec 12 00:55:22 2015 @@ -1 +1 @@ -<!DOCTYPE html><html lang="en"><head><meta charset="utf-8"></meta><title>CreateHadoopSequenceFile</title><link rel="stylesheet" href="../../css/component-usage.css" type="text/css"></link></head><body><h2>Description: </h2><p>Creates Hadoop Sequence Files from incoming flow files</p><p><a href="additionalDetails.html">Additional Details...</a></p><h3>Tags: </h3><p>hadoop, sequence file, create, sequencefile</p><h3>Properties: </h3><p>In the list below, the names of required properties appear in <strong>bold</strong>. Any other properties (not in bold) are considered optional. The table also indicates any default values.</p><table id="properties"><tr><th>Name</th><th>Default Value</th><th>Allowable Values</th><th>Description</th></tr><tr><td id="name">Hadoop Configuration Resources</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">A file or comma separated list of files which contains the Hadoop file system configuration. Without this, Hadoop will se arch the classpath for a 'core-site.xml' and 'hdfs-site.xml' file or will revert to a default configuration.</td></tr><tr><td id="name">Kerberos Principal</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Kerberos principal to authenticate as. Requires nifi.kerberos.krb5.file to be set in your nifi.properties</td></tr><tr><td id="name">Kerberos Keytab</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Kerberos keytab associated with the principal. Requires nifi.kerberos.krb5.file to be set in your nifi.properties</td></tr><tr><td id="name">compression type</td><td id="default-value"></td><td id="allowable-values"><ul><li>NONE</li><li>RECORD</li><li>BLOCK</li></ul></td><td id="description">Type of compression to use when creating Sequence File</td></tr></table><h3>Relationships: </h3><table id="relationships"><tr><th>Name</th><th>Description</th></tr><tr><td>success</td><td>Generated Sequence Files are sent to this re lationship</td></tr><tr><td>failure</td><td>Incoming files that failed to generate a Sequence File are sent to this relationship</td></tr></table><h3>Reads Attributes: </h3>None specified.<h3>Writes Attributes: </h3>None specified.<h3>See Also:</h3><p><a href="../org.apache.nifi.processors.hadoop.PutHDFS/index.html">PutHDFS</a></p></body></html> \ No newline at end of file +<!DOCTYPE html><html lang="en"><head><meta charset="utf-8"></meta><title>CreateHadoopSequenceFile</title><link rel="stylesheet" href="../../css/component-usage.css" type="text/css"></link></head><body><h2>Description: </h2><p>Creates Hadoop Sequence Files from incoming flow files</p><p><a href="additionalDetails.html">Additional Details...</a></p><h3>Tags: </h3><p>hadoop, sequence file, create, sequencefile</p><h3>Properties: </h3><p>In the list below, the names of required properties appear in <strong>bold</strong>. Any other properties (not in bold) are considered optional. The table also indicates any default values.</p><table id="properties"><tr><th>Name</th><th>Default Value</th><th>Allowable Values</th><th>Description</th></tr><tr><td id="name">Hadoop Configuration Resources</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">A file or comma separated list of files which contains the Hadoop file system configuration. Without this, Hadoop will se arch the classpath for a 'core-site.xml' and 'hdfs-site.xml' file or will revert to a default configuration.</td></tr><tr><td id="name">Kerberos Principal</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Kerberos principal to authenticate as. Requires nifi.kerberos.krb5.file to be set in your nifi.properties</td></tr><tr><td id="name">Kerberos Keytab</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Kerberos keytab associated with the principal. Requires nifi.kerberos.krb5.file to be set in your nifi.properties</td></tr><tr><td id="name">Kerberos Relogin Period</td><td id="default-value">4 hours</td><td id="allowable-values"></td><td id="description">Period of time which should pass before attempting a kerberos relogin</td></tr><tr><td id="name">compression type</td><td id="default-value"></td><td id="allowable-values"><ul><li>NONE</li><li>DEFAULT</li><li>BZIP</li><li>GZIP</li><li>LZ4</li><li>SNAPPY</li><li>AUTOMAT IC</li></ul></td><td id="description">Type of compression to use when creating Sequence File</td></tr></table><h3>Relationships: </h3><table id="relationships"><tr><th>Name</th><th>Description</th></tr><tr><td>success</td><td>Generated Sequence Files are sent to this relationship</td></tr><tr><td>failure</td><td>Incoming files that failed to generate a Sequence File are sent to this relationship</td></tr></table><h3>Reads Attributes: </h3>None specified.<h3>Writes Attributes: </h3>None specified.<h3>See Also:</h3><p><a href="../org.apache.nifi.processors.hadoop.PutHDFS/index.html">PutHDFS</a></p></body></html> \ No newline at end of file Modified: nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.FetchHDFS/index.html URL: http://svn.apache.org/viewvc/nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.FetchHDFS/index.html?rev=1719609&r1=1719608&r2=1719609&view=diff ============================================================================== --- nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.FetchHDFS/index.html (original) +++ nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.FetchHDFS/index.html Sat Dec 12 00:55:22 2015 @@ -1 +1 @@ -<!DOCTYPE html><html lang="en"><head><meta charset="utf-8"></meta><title>FetchHDFS</title><link rel="stylesheet" href="../../css/component-usage.css" type="text/css"></link></head><body><h2>Description: </h2><p>Retrieves a file from HDFS. The content of the incoming FlowFile is replaced by the content of the file in HDFS. The file in HDFS is left intact without any changes being made to it.</p><h3>Tags: </h3><p>hadoop, hdfs, get, ingest, fetch, source</p><h3>Properties: </h3><p>In the list below, the names of required properties appear in <strong>bold</strong>. Any other properties (not in bold) are considered optional. The table also indicates any default values, and whether a property supports the <a href="../../html/expression-language-guide.html">NiFi Expression Language</a>.</p><table id="properties"><tr><th>Name</th><th>Default Value</th><th>Allowable Values</th><th>Description</th></tr><tr><td id="name">Hadoop Configuration Resources</td><td id="default-value"></td><td id="al lowable-values"></td><td id="description">A file or comma separated list of files which contains the Hadoop file system configuration. Without this, Hadoop will search the classpath for a 'core-site.xml' and 'hdfs-site.xml' file or will revert to a default configuration.</td></tr><tr><td id="name"><strong>HDFS Filename</strong></td><td id="default-value">${path}/${filename}</td><td id="allowable-values"></td><td id="description">The name of the HDFS file to retrieve<br/><strong>Supports Expression Language: true</strong></td></tr></table><h3>Relationships: </h3><table id="relationships"><tr><th>Name</th><th>Description</th></tr><tr><td>success</td><td>FlowFiles will be routed to this relationship once they have been updated with the content of the HDFS file</td></tr><tr><td>comms.failure</td><td>FlowFiles will be routed to this relationship if the content of the HDFS file cannot be retrieve due to a communications failure. This generally indicates that the Fetch should be tried agai n.</td></tr><tr><td>failure</td><td>FlowFiles will be routed to this relationship if the content of the HDFS file cannot be retrieved and trying again will likely not be helpful. This would occur, for instance, if the file is not found or if there is a permissions issue</td></tr></table><h3>Reads Attributes: </h3>None specified.<h3>Writes Attributes: </h3><table id="writes-attributes"><tr><th>Name</th><th>Description</th></tr><tr><td>hdfs.failure.reason</td><td>When a FlowFile is routed to 'failure', this attribute is added indicating why the file could not be fetched from HDFS</td></tr></table><h3>See Also:</h3><p><a href="../org.apache.nifi.processors.hadoop.ListHDFS/index.html">ListHDFS</a>, <a href="../org.apache.nifi.processors.hadoop.GetHDFS/index.html">GetHDFS</a>, <a href="../org.apache.nifi.processors.hadoop.PutHDFS/index.html">PutHDFS</a></p></body></html> \ No newline at end of file +<!DOCTYPE html><html lang="en"><head><meta charset="utf-8"></meta><title>FetchHDFS</title><link rel="stylesheet" href="../../css/component-usage.css" type="text/css"></link></head><body><h2>Description: </h2><p>Retrieves a file from HDFS. The content of the incoming FlowFile is replaced by the content of the file in HDFS. The file in HDFS is left intact without any changes being made to it.</p><h3>Tags: </h3><p>hadoop, hdfs, get, ingest, fetch, source</p><h3>Properties: </h3><p>In the list below, the names of required properties appear in <strong>bold</strong>. Any other properties (not in bold) are considered optional. The table also indicates any default values, and whether a property supports the <a href="../../html/expression-language-guide.html">NiFi Expression Language</a>.</p><table id="properties"><tr><th>Name</th><th>Default Value</th><th>Allowable Values</th><th>Description</th></tr><tr><td id="name">Hadoop Configuration Resources</td><td id="default-value"></td><td id="al lowable-values"></td><td id="description">A file or comma separated list of files which contains the Hadoop file system configuration. Without this, Hadoop will search the classpath for a 'core-site.xml' and 'hdfs-site.xml' file or will revert to a default configuration.</td></tr><tr><td id="name"><strong>HDFS Filename</strong></td><td id="default-value">${path}/${filename}</td><td id="allowable-values"></td><td id="description">The name of the HDFS file to retrieve<br/><strong>Supports Expression Language: true</strong></td></tr><tr><td id="name">Kerberos Principal</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Kerberos principal to authenticate as. Requires nifi.kerberos.krb5.file to be set in your nifi.properties</td></tr><tr><td id="name">Kerberos Keytab</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Kerberos keytab associated with the principal. Requires nifi.kerberos.krb5.file to be set in your nifi.prop erties</td></tr><tr><td id="name">Kerberos Relogin Period</td><td id="default-value">4 hours</td><td id="allowable-values"></td><td id="description">Period of time which should pass before attempting a kerberos relogin</td></tr></table><h3>Relationships: </h3><table id="relationships"><tr><th>Name</th><th>Description</th></tr><tr><td>success</td><td>FlowFiles will be routed to this relationship once they have been updated with the content of the HDFS file</td></tr><tr><td>comms.failure</td><td>FlowFiles will be routed to this relationship if the content of the HDFS file cannot be retrieve due to a communications failure. This generally indicates that the Fetch should be tried again.</td></tr><tr><td>failure</td><td>FlowFiles will be routed to this relationship if the content of the HDFS file cannot be retrieved and trying again will likely not be helpful. This would occur, for instance, if the file is not found or if there is a permissions issue</td></tr></table><h3>Reads Attributes : </h3>None specified.<h3>Writes Attributes: </h3><table id="writes-attributes"><tr><th>Name</th><th>Description</th></tr><tr><td>hdfs.failure.reason</td><td>When a FlowFile is routed to 'failure', this attribute is added indicating why the file could not be fetched from HDFS</td></tr></table><h3>See Also:</h3><p><a href="../org.apache.nifi.processors.hadoop.ListHDFS/index.html">ListHDFS</a>, <a href="../org.apache.nifi.processors.hadoop.GetHDFS/index.html">GetHDFS</a>, <a href="../org.apache.nifi.processors.hadoop.PutHDFS/index.html">PutHDFS</a></p></body></html> \ No newline at end of file Modified: nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.GetHDFS/index.html URL: http://svn.apache.org/viewvc/nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.GetHDFS/index.html?rev=1719609&r1=1719608&r2=1719609&view=diff ============================================================================== --- nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.GetHDFS/index.html (original) +++ nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.GetHDFS/index.html Sat Dec 12 00:55:22 2015 @@ -1 +1 @@ -<!DOCTYPE html><html lang="en"><head><meta charset="utf-8"></meta><title>GetHDFS</title><link rel="stylesheet" href="../../css/component-usage.css" type="text/css"></link></head><body><h2>Description: </h2><p>Fetch files from Hadoop Distributed File System (HDFS) into FlowFiles. This Processor will delete the file from HDFS after fetching it.</p><h3>Tags: </h3><p>hadoop, HDFS, get, fetch, ingest, source, filesystem</p><h3>Properties: </h3><p>In the list below, the names of required properties appear in <strong>bold</strong>. Any other properties (not in bold) are considered optional. The table also indicates any default values.</p><table id="properties"><tr><th>Name</th><th>Default Value</th><th>Allowable Values</th><th>Description</th></tr><tr><td id="name">Hadoop Configuration Resources</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">A file or comma separated list of files which contains the Hadoop file system configuration. Without this, Hadoop will search the classpath for a 'core-site.xml' and 'hdfs-site.xml' file or will revert to a default configuration.</td></tr><tr><td id="name">Kerberos Principal</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Kerberos principal to authenticate as. Requires nifi.kerberos.krb5.file to be set in your nifi.properties</td></tr><tr><td id="name">Kerberos Keytab</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Kerberos keytab associated with the principal. Requires nifi.kerberos.krb5.file to be set in your nifi.properties</td></tr><tr><td id="name"><strong>Directory</strong></td><td id="default-value"></td><td id="allowable-values"></td><td id="description">The HDFS directory from which files should be read</td></tr><tr><td id="name"><strong>Recurse Subdirectories</strong></td><td id="default-value">true</td><td id="allowable-values"><ul><li>true</li><li>false</li></ul></td><td id="description">Indicates whether to pu ll files from subdirectories of the HDFS directory</td></tr><tr><td id="name"><strong>Keep Source File</strong></td><td id="default-value">false</td><td id="allowable-values"><ul><li>true</li><li>false</li></ul></td><td id="description">Determines whether to delete the file from HDFS after it has been successfully transferred. If true, the file will be fetched repeatedly. This is intended for testing only.</td></tr><tr><td id="name">File Filter Regex</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">A Java Regular Expression for filtering Filenames; if a filter is supplied then only files whose names match that Regular Expression will be fetched, otherwise all files will be fetched</td></tr><tr><td id="name"><strong>Filter Match Name Only</strong></td><td id="default-value">true</td><td id="allowable-values"><ul><li>true</li><li>false</li></ul></td><td id="description">If true then File Filter Regex will match on just the filename, otherwise subdire ctory names will be included with filename in the regex comparison</td></tr><tr><td id="name"><strong>Ignore Dotted Files</strong></td><td id="default-value">true</td><td id="allowable-values"><ul><li>true</li><li>false</li></ul></td><td id="description">If true, files whose names begin with a dot (".") will be ignored</td></tr><tr><td id="name"><strong>Minimum File Age</strong></td><td id="default-value">0 sec</td><td id="allowable-values"></td><td id="description">The minimum age that a file must be in order to be pulled; any file younger than this amount of time (based on last modification date) will be ignored</td></tr><tr><td id="name">Maximum File Age</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">The maximum age that a file must be in order to be pulled; any file older than this amount of time (based on last modification date) will be ignored</td></tr><tr><td id="name"><strong>Polling Interval</strong></td><td id="default-value">0 sec</td> <td id="allowable-values"></td><td id="description">Indicates how long to wait between performing directory listings</td></tr><tr><td id="name"><strong>Batch Size</strong></td><td id="default-value">100</td><td id="allowable-values"></td><td id="description">The maximum number of files to pull in each iteration, based on run schedule.</td></tr><tr><td id="name">IO Buffer Size</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Amount of memory to use to buffer file contents during IO. This overrides the Hadoop Configuration</td></tr><tr><td id="name">Compression codec</td><td id="default-value"></td><td id="allowable-values"><ul><li>org.apache.hadoop.io.compress.BZip2Codec</li><li>org.apache.hadoop.io.compress.DefaultCodec</li><li>org.apache.hadoop.io.compress.GzipCodec</li><li>org.apache.hadoop.io.compress.Lz4Codec</li><li>org.apache.hadoop.io.compress.SnappyCodec</li></ul></td><td id="description">No Description Provided.</td></tr></table><h3>Relati onships: </h3><table id="relationships"><tr><th>Name</th><th>Description</th></tr><tr><td>success</td><td>All files retrieved from HDFS are transferred to this relationship</td></tr><tr><td>passthrough</td><td>If this processor has an input queue for some reason, then FlowFiles arriving on that input are transferred to this relationship</td></tr></table><h3>Reads Attributes: </h3>None specified.<h3>Writes Attributes: </h3><table id="writes-attributes"><tr><th>Name</th><th>Description</th></tr><tr><td>filename</td><td>The name of the file that was read from HDFS.</td></tr><tr><td>path</td><td>The path is set to the relative path of the file's directory on HDFS. For example, if the Directory property is set to /tmp, then files picked up from /tmp will have the path attribute set to "./". If the Recurse Subdirectories property is set to true and a file is picked up from /tmp/abc/1/2/3, then the path attribute will be set to "abc/1/2/3".</td></tr></table><h3>See Also:</h3><p><a href=".. /org.apache.nifi.processors.hadoop.PutHDFS/index.html">PutHDFS</a>, <a href="../org.apache.nifi.processors.hadoop.ListHDFS/index.html">ListHDFS</a></p></body></html> \ No newline at end of file +<!DOCTYPE html><html lang="en"><head><meta charset="utf-8"></meta><title>GetHDFS</title><link rel="stylesheet" href="../../css/component-usage.css" type="text/css"></link></head><body><h2>Description: </h2><p>Fetch files from Hadoop Distributed File System (HDFS) into FlowFiles. This Processor will delete the file from HDFS after fetching it.</p><h3>Tags: </h3><p>hadoop, HDFS, get, fetch, ingest, source, filesystem</p><h3>Properties: </h3><p>In the list below, the names of required properties appear in <strong>bold</strong>. Any other properties (not in bold) are considered optional. The table also indicates any default values.</p><table id="properties"><tr><th>Name</th><th>Default Value</th><th>Allowable Values</th><th>Description</th></tr><tr><td id="name">Hadoop Configuration Resources</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">A file or comma separated list of files which contains the Hadoop file system configuration. Without this, Hadoop will search the classpath for a 'core-site.xml' and 'hdfs-site.xml' file or will revert to a default configuration.</td></tr><tr><td id="name">Kerberos Principal</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Kerberos principal to authenticate as. Requires nifi.kerberos.krb5.file to be set in your nifi.properties</td></tr><tr><td id="name">Kerberos Keytab</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Kerberos keytab associated with the principal. Requires nifi.kerberos.krb5.file to be set in your nifi.properties</td></tr><tr><td id="name">Kerberos Relogin Period</td><td id="default-value">4 hours</td><td id="allowable-values"></td><td id="description">Period of time which should pass before attempting a kerberos relogin</td></tr><tr><td id="name"><strong>Directory</strong></td><td id="default-value"></td><td id="allowable-values"></td><td id="description">The HDFS directory from which files should be read</t d></tr><tr><td id="name"><strong>Recurse Subdirectories</strong></td><td id="default-value">true</td><td id="allowable-values"><ul><li>true</li><li>false</li></ul></td><td id="description">Indicates whether to pull files from subdirectories of the HDFS directory</td></tr><tr><td id="name"><strong>Keep Source File</strong></td><td id="default-value">false</td><td id="allowable-values"><ul><li>true</li><li>false</li></ul></td><td id="description">Determines whether to delete the file from HDFS after it has been successfully transferred. If true, the file will be fetched repeatedly. This is intended for testing only.</td></tr><tr><td id="name">File Filter Regex</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">A Java Regular Expression for filtering Filenames; if a filter is supplied then only files whose names match that Regular Expression will be fetched, otherwise all files will be fetched</td></tr><tr><td id="name"><strong>Filter Match Name Only</s trong></td><td id="default-value">true</td><td id="allowable-values"><ul><li>true</li><li>false</li></ul></td><td id="description">If true then File Filter Regex will match on just the filename, otherwise subdirectory names will be included with filename in the regex comparison</td></tr><tr><td id="name"><strong>Ignore Dotted Files</strong></td><td id="default-value">true</td><td id="allowable-values"><ul><li>true</li><li>false</li></ul></td><td id="description">If true, files whose names begin with a dot (".") will be ignored</td></tr><tr><td id="name"><strong>Minimum File Age</strong></td><td id="default-value">0 sec</td><td id="allowable-values"></td><td id="description">The minimum age that a file must be in order to be pulled; any file younger than this amount of time (based on last modification date) will be ignored</td></tr><tr><td id="name">Maximum File Age</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">The maximum age that a file must be in order to be pulled; any file older than this amount of time (based on last modification date) will be ignored</td></tr><tr><td id="name"><strong>Polling Interval</strong></td><td id="default-value">0 sec</td><td id="allowable-values"></td><td id="description">Indicates how long to wait between performing directory listings</td></tr><tr><td id="name"><strong>Batch Size</strong></td><td id="default-value">100</td><td id="allowable-values"></td><td id="description">The maximum number of files to pull in each iteration, based on run schedule.</td></tr><tr><td id="name">IO Buffer Size</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Amount of memory to use to buffer file contents during IO. This overrides the Hadoop Configuration</td></tr><tr><td id="name"><strong>Compression codec</strong></td><td id="default-value">NONE</td><td id="allowable-values"><ul><li>NONE</li><li>DEFAULT</li><li>BZIP</li><li>GZIP</li><li>LZ4</li><li>SNAPPY</li><li>AUTOMATIC </li></ul></td><td id="description">No Description Provided.</td></tr></table><h3>Relationships: </h3><table id="relationships"><tr><th>Name</th><th>Description</th></tr><tr><td>success</td><td>All files retrieved from HDFS are transferred to this relationship</td></tr><tr><td>passthrough</td><td>If this processor has an input queue for some reason, then FlowFiles arriving on that input are transferred to this relationship</td></tr></table><h3>Reads Attributes: </h3>None specified.<h3>Writes Attributes: </h3><table id="writes-attributes"><tr><th>Name</th><th>Description</th></tr><tr><td>filename</td><td>The name of the file that was read from HDFS.</td></tr><tr><td>path</td><td>The path is set to the relative path of the file's directory on HDFS. For example, if the Directory property is set to /tmp, then files picked up from /tmp will have the path attribute set to "./". If the Recurse Subdirectories property is set to true and a file is picked up from /tmp/abc/1/2/3, then the path attribute will be set to "abc/1/2/3".</td></tr></table><h3>See Also:</h3><p><a href="../org.apache.nifi.processors.hadoop.PutHDFS/index.html">PutHDFS</a>, <a href="../org.apache.nifi.processors.hadoop.ListHDFS/index.html">ListHDFS</a></p></body></html> \ No newline at end of file Modified: nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.GetHDFSSequenceFile/index.html URL: http://svn.apache.org/viewvc/nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.GetHDFSSequenceFile/index.html?rev=1719609&r1=1719608&r2=1719609&view=diff ============================================================================== --- nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.GetHDFSSequenceFile/index.html (original) +++ nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.GetHDFSSequenceFile/index.html Sat Dec 12 00:55:22 2015 @@ -1 +1 @@ -<!DOCTYPE html><html lang="en"><head><meta charset="utf-8"></meta><title>GetHDFSSequenceFile</title><link rel="stylesheet" href="../../css/component-usage.css" type="text/css"></link></head><body><h2>Description: </h2><p>Fetch sequence files from Hadoop Distributed File System (HDFS) into FlowFiles</p><h3>Tags: </h3><p>hadoop, HDFS, get, fetch, ingest, source, sequence file</p><h3>Properties: </h3><p>In the list below, the names of required properties appear in <strong>bold</strong>. Any other properties (not in bold) are considered optional. The table also indicates any default values.</p><table id="properties"><tr><th>Name</th><th>Default Value</th><th>Allowable Values</th><th>Description</th></tr><tr><td id="name">Hadoop Configuration Resources</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">A file or comma separated list of files which contains the Hadoop file system configuration. Without this, Hadoop will search the classpath for a 'core-sit e.xml' and 'hdfs-site.xml' file or will revert to a default configuration.</td></tr><tr><td id="name">Kerberos Principal</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Kerberos principal to authenticate as. Requires nifi.kerberos.krb5.file to be set in your nifi.properties</td></tr><tr><td id="name">Kerberos Keytab</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Kerberos keytab associated with the principal. Requires nifi.kerberos.krb5.file to be set in your nifi.properties</td></tr><tr><td id="name"><strong>Directory</strong></td><td id="default-value"></td><td id="allowable-values"></td><td id="description">The HDFS directory from which files should be read</td></tr><tr><td id="name"><strong>Recurse Subdirectories</strong></td><td id="default-value">true</td><td id="allowable-values"><ul><li>true</li><li>false</li></ul></td><td id="description">Indicates whether to pull files from subdirectories of the HDFS d irectory</td></tr><tr><td id="name"><strong>Keep Source File</strong></td><td id="default-value">false</td><td id="allowable-values"><ul><li>true</li><li>false</li></ul></td><td id="description">Determines whether to delete the file from HDFS after it has been successfully transferred. If true, the file will be fetched repeatedly. This is intended for testing only.</td></tr><tr><td id="name">File Filter Regex</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">A Java Regular Expression for filtering Filenames; if a filter is supplied then only files whose names match that Regular Expression will be fetched, otherwise all files will be fetched</td></tr><tr><td id="name"><strong>Filter Match Name Only</strong></td><td id="default-value">true</td><td id="allowable-values"><ul><li>true</li><li>false</li></ul></td><td id="description">If true then File Filter Regex will match on just the filename, otherwise subdirectory names will be included with filename in the regex comparison</td></tr><tr><td id="name"><strong>Ignore Dotted Files</strong></td><td id="default-value">true</td><td id="allowable-values"><ul><li>true</li><li>false</li></ul></td><td id="description">If true, files whose names begin with a dot (".") will be ignored</td></tr><tr><td id="name"><strong>Minimum File Age</strong></td><td id="default-value">0 sec</td><td id="allowable-values"></td><td id="description">The minimum age that a file must be in order to be pulled; any file younger than this amount of time (based on last modification date) will be ignored</td></tr><tr><td id="name">Maximum File Age</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">The maximum age that a file must be in order to be pulled; any file older than this amount of time (based on last modification date) will be ignored</td></tr><tr><td id="name"><strong>Polling Interval</strong></td><td id="default-value">0 sec</td><td id="allowable-values"></td><td id="des cription">Indicates how long to wait between performing directory listings</td></tr><tr><td id="name"><strong>Batch Size</strong></td><td id="default-value">100</td><td id="allowable-values"></td><td id="description">The maximum number of files to pull in each iteration, based on run schedule.</td></tr><tr><td id="name">IO Buffer Size</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Amount of memory to use to buffer file contents during IO. This overrides the Hadoop Configuration</td></tr><tr><td id="name">Compression codec</td><td id="default-value"></td><td id="allowable-values"><ul><li>org.apache.hadoop.io.compress.BZip2Codec</li><li>org.apache.hadoop.io.compress.DefaultCodec</li><li>org.apache.hadoop.io.compress.GzipCodec</li><li>org.apache.hadoop.io.compress.Lz4Codec</li><li>org.apache.hadoop.io.compress.SnappyCodec</li></ul></td><td id="description">No Description Provided.</td></tr><tr><td id="name"><strong>FlowFile Content</strong></td><td id="default-value">VALUE ONLY</td><td id="allowable-values"><ul><li>VALUE ONLY</li><li>KEY VALUE PAIR</li></ul></td><td id="description">Indicate if the content is to be both the key and value of the Sequence File, or just the value.</td></tr></table><h3>Relationships: </h3><table id="relationships"><tr><th>Name</th><th>Description</th></tr><tr><td>success</td><td>All files retrieved from HDFS are transferred to this relationship</td></tr><tr><td>passthrough</td><td>If this processor has an input queue for some reason, then FlowFiles arriving on that input are transferred to this relationship</td></tr></table><h3>Reads Attributes: </h3>None specified.<h3>Writes Attributes: </h3><table id="writes-attributes"><tr><th>Name</th><th>Description</th></tr><tr><td>filename</td><td>The name of the file that was read from HDFS.</td></tr><tr><td>path</td><td>The path is set to the relative path of the file's directory on HDFS. For example, if the Directory property is set to /tmp, then files p icked up from /tmp will have the path attribute set to "./". If the Recurse Subdirectories property is set to true and a file is picked up from /tmp/abc/1/2/3, then the path attribute will be set to "abc/1/2/3".</td></tr></table><h3>See Also:</h3><p><a href="../org.apache.nifi.processors.hadoop.PutHDFS/index.html">PutHDFS</a></p></body></html> \ No newline at end of file +<!DOCTYPE html><html lang="en"><head><meta charset="utf-8"></meta><title>GetHDFSSequenceFile</title><link rel="stylesheet" href="../../css/component-usage.css" type="text/css"></link></head><body><h2>Description: </h2><p>Fetch sequence files from Hadoop Distributed File System (HDFS) into FlowFiles</p><h3>Tags: </h3><p>hadoop, HDFS, get, fetch, ingest, source, sequence file</p><h3>Properties: </h3><p>In the list below, the names of required properties appear in <strong>bold</strong>. Any other properties (not in bold) are considered optional. The table also indicates any default values.</p><table id="properties"><tr><th>Name</th><th>Default Value</th><th>Allowable Values</th><th>Description</th></tr><tr><td id="name">Hadoop Configuration Resources</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">A file or comma separated list of files which contains the Hadoop file system configuration. Without this, Hadoop will search the classpath for a 'core-sit e.xml' and 'hdfs-site.xml' file or will revert to a default configuration.</td></tr><tr><td id="name">Kerberos Principal</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Kerberos principal to authenticate as. Requires nifi.kerberos.krb5.file to be set in your nifi.properties</td></tr><tr><td id="name">Kerberos Keytab</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Kerberos keytab associated with the principal. Requires nifi.kerberos.krb5.file to be set in your nifi.properties</td></tr><tr><td id="name">Kerberos Relogin Period</td><td id="default-value">4 hours</td><td id="allowable-values"></td><td id="description">Period of time which should pass before attempting a kerberos relogin</td></tr><tr><td id="name"><strong>Directory</strong></td><td id="default-value"></td><td id="allowable-values"></td><td id="description">The HDFS directory from which files should be read</td></tr><tr><td id="name"><strong>Recurse S ubdirectories</strong></td><td id="default-value">true</td><td id="allowable-values"><ul><li>true</li><li>false</li></ul></td><td id="description">Indicates whether to pull files from subdirectories of the HDFS directory</td></tr><tr><td id="name"><strong>Keep Source File</strong></td><td id="default-value">false</td><td id="allowable-values"><ul><li>true</li><li>false</li></ul></td><td id="description">Determines whether to delete the file from HDFS after it has been successfully transferred. If true, the file will be fetched repeatedly. This is intended for testing only.</td></tr><tr><td id="name">File Filter Regex</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">A Java Regular Expression for filtering Filenames; if a filter is supplied then only files whose names match that Regular Expression will be fetched, otherwise all files will be fetched</td></tr><tr><td id="name"><strong>Filter Match Name Only</strong></td><td id="default-value">true</td ><td id="allowable-values"><ul><li>true</li><li>false</li></ul></td><td >id="description">If true then File Filter Regex will match on just the >filename, otherwise subdirectory names will be included with filename in the >regex comparison</td></tr><tr><td id="name"><strong>Ignore Dotted >Files</strong></td><td id="default-value">true</td><td >id="allowable-values"><ul><li>true</li><li>false</li></ul></td><td >id="description">If true, files whose names begin with a dot (".") will be >ignored</td></tr><tr><td id="name"><strong>Minimum File Age</strong></td><td >id="default-value">0 sec</td><td id="allowable-values"></td><td >id="description">The minimum age that a file must be in order to be pulled; >any file younger than this amount of time (based on last modification date) >will be ignored</td></tr><tr><td id="name">Maximum File Age</td><td >id="default-value"></td><td id="allowable-values"></td><td >id="description">The maximum age that a file must be in order to be pulled; >any file older tha n this amount of time (based on last modification date) will be ignored</td></tr><tr><td id="name"><strong>Polling Interval</strong></td><td id="default-value">0 sec</td><td id="allowable-values"></td><td id="description">Indicates how long to wait between performing directory listings</td></tr><tr><td id="name"><strong>Batch Size</strong></td><td id="default-value">100</td><td id="allowable-values"></td><td id="description">The maximum number of files to pull in each iteration, based on run schedule.</td></tr><tr><td id="name">IO Buffer Size</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Amount of memory to use to buffer file contents during IO. This overrides the Hadoop Configuration</td></tr><tr><td id="name"><strong>Compression codec</strong></td><td id="default-value">NONE</td><td id="allowable-values"><ul><li>NONE</li><li>DEFAULT</li><li>BZIP</li><li>GZIP</li><li>LZ4</li><li>SNAPPY</li><li>AUTOMATIC</li></ul></td><td id="description">No Des cription Provided.</td></tr><tr><td id="name"><strong>FlowFile Content</strong></td><td id="default-value">VALUE ONLY</td><td id="allowable-values"><ul><li>VALUE ONLY</li><li>KEY VALUE PAIR</li></ul></td><td id="description">Indicate if the content is to be both the key and value of the Sequence File, or just the value.</td></tr></table><h3>Relationships: </h3><table id="relationships"><tr><th>Name</th><th>Description</th></tr><tr><td>success</td><td>All files retrieved from HDFS are transferred to this relationship</td></tr><tr><td>passthrough</td><td>If this processor has an input queue for some reason, then FlowFiles arriving on that input are transferred to this relationship</td></tr></table><h3>Reads Attributes: </h3>None specified.<h3>Writes Attributes: </h3><table id="writes-attributes"><tr><th>Name</th><th>Description</th></tr><tr><td>filename</td><td>The name of the file that was read from HDFS.</td></tr><tr><td>path</td><td>The path is set to the relative path of the file' s directory on HDFS. For example, if the Directory property is set to /tmp, then files picked up from /tmp will have the path attribute set to "./". If the Recurse Subdirectories property is set to true and a file is picked up from /tmp/abc/1/2/3, then the path attribute will be set to "abc/1/2/3".</td></tr></table><h3>See Also:</h3><p><a href="../org.apache.nifi.processors.hadoop.PutHDFS/index.html">PutHDFS</a></p></body></html> \ No newline at end of file Modified: nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.ListHDFS/index.html URL: http://svn.apache.org/viewvc/nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.ListHDFS/index.html?rev=1719609&r1=1719608&r2=1719609&view=diff ============================================================================== --- nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.ListHDFS/index.html (original) +++ nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.ListHDFS/index.html Sat Dec 12 00:55:22 2015 @@ -1 +1 @@ -<!DOCTYPE html><html lang="en"><head><meta charset="utf-8"></meta><title>ListHDFS</title><link rel="stylesheet" href="../../css/component-usage.css" type="text/css"></link></head><body><h2>Description: </h2><p>Retrieves a listing of files from HDFS. For each file that is listed in HDFS, creates a FlowFile that represents the HDFS file so that it can be fetched in conjunction with ListHDFS. This Processor is designed to run on Primary Node only in a cluster. If the primary node changes, the new Primary Node will pick up where the previous node left off without duplicating all of the data. Unlike GetHDFS, this Processor does not delete any data from HDFS.</p><h3>Tags: </h3><p>hadoop, HDFS, get, list, ingest, source, filesystem</p><h3>Properties: </h3><p>In the list below, the names of required properties appear in <strong>bold</strong>. Any other properties (not in bold) are considered optional. The table also indicates any default values.</p><table id="properties"><tr><th>Name</th><t h>Default Value</th><th>Allowable Values</th><th>Description</th></tr><tr><td id="name">Hadoop Configuration Resources</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">A file or comma separated list of files which contains the Hadoop file system configuration. Without this, Hadoop will search the classpath for a 'core-site.xml' and 'hdfs-site.xml' file or will revert to a default configuration.</td></tr><tr><td id="name"><strong>Distributed Cache Service</strong></td><td id="default-value"></td><td id="allowable-values"><strong>Controller Service API: </strong><br/>DistributedMapCacheClient<br/><strong>Implementation:</strong><br/><a href="../org.apache.nifi.distributed.cache.client.DistributedMapCacheClientService/index.html">DistributedMapCacheClientService</a></td><td id="description">Specifies the Controller Service that should be used to maintain state about what has been pulled from HDFS so that if a new node begins pulling data, it won't dup licate all of the work that has been done.</td></tr><tr><td id="name"><strong>Directory</strong></td><td id="default-value"></td><td id="allowable-values"></td><td id="description">The HDFS directory from which files should be read</td></tr><tr><td id="name"><strong>Recurse Subdirectories</strong></td><td id="default-value">true</td><td id="allowable-values"><ul><li>true</li><li>false</li></ul></td><td id="description">Indicates whether to list files from subdirectories of the HDFS directory</td></tr></table><h3>Relationships: </h3><table id="relationships"><tr><th>Name</th><th>Description</th></tr><tr><td>success</td><td>All FlowFiles are transferred to this relationship</td></tr></table><h3>Reads Attributes: </h3>None specified.<h3>Writes Attributes: </h3><table id="writes-attributes"><tr><th>Name</th><th>Description</th></tr><tr><td>filename</td><td>The name of the file that was read from HDFS.</td></tr><tr><td>path</td><td>The path is set to the absolute path of the file's direc tory on HDFS. For example, if the Directory property is set to /tmp, then files picked up from /tmp will have the path attribute set to "./". If the Recurse Subdirectories property is set to true and a file is picked up from /tmp/abc/1/2/3, then the path attribute will be set to "/tmp/abc/1/2/3".</td></tr><tr><td>hdfs.owner</td><td>The user that owns the file in HDFS</td></tr><tr><td>hdfs.group</td><td>The group that owns the file in HDFS</td></tr><tr><td>hdfs.lastModified</td><td>The timestamp of when the file in HDFS was last modified, as milliseconds since midnight Jan 1, 1970 UTC</td></tr><tr><td>hdfs.length</td><td>The number of bytes in the file in HDFS</td></tr><tr><td>hdfs.replication</td><td>The number of HDFS replicas for hte file</td></tr><tr><td>hdfs.permissions</td><td>The permissions for the file in HDFS. This is formatted as 3 characters for the owner, 3 for the group, and 3 for other users. For example rw-rw-r--</td></tr></table><h3>See Also:</h3><p><a href="../org.a pache.nifi.processors.hadoop.GetHDFS/index.html">GetHDFS</a>, <a href="../org.apache.nifi.processors.hadoop.FetchHDFS/index.html">FetchHDFS</a>, <a href="../org.apache.nifi.processors.hadoop.PutHDFS/index.html">PutHDFS</a></p></body></html> \ No newline at end of file +<!DOCTYPE html><html lang="en"><head><meta charset="utf-8"></meta><title>ListHDFS</title><link rel="stylesheet" href="../../css/component-usage.css" type="text/css"></link></head><body><h2>Description: </h2><p>Retrieves a listing of files from HDFS. For each file that is listed in HDFS, creates a FlowFile that represents the HDFS file so that it can be fetched in conjunction with ListHDFS. This Processor is designed to run on Primary Node only in a cluster. If the primary node changes, the new Primary Node will pick up where the previous node left off without duplicating all of the data. Unlike GetHDFS, this Processor does not delete any data from HDFS.</p><h3>Tags: </h3><p>hadoop, HDFS, get, list, ingest, source, filesystem</p><h3>Properties: </h3><p>In the list below, the names of required properties appear in <strong>bold</strong>. Any other properties (not in bold) are considered optional. The table also indicates any default values.</p><table id="properties"><tr><th>Name</th><t h>Default Value</th><th>Allowable Values</th><th>Description</th></tr><tr><td id="name">Hadoop Configuration Resources</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">A file or comma separated list of files which contains the Hadoop file system configuration. Without this, Hadoop will search the classpath for a 'core-site.xml' and 'hdfs-site.xml' file or will revert to a default configuration.</td></tr><tr><td id="name"><strong>Distributed Cache Service</strong></td><td id="default-value"></td><td id="allowable-values"><strong>Controller Service API: </strong><br/>DistributedMapCacheClient<br/><strong>Implementation:</strong><br/><a href="../org.apache.nifi.distributed.cache.client.DistributedMapCacheClientService/index.html">DistributedMapCacheClientService</a></td><td id="description">Specifies the Controller Service that should be used to maintain state about what has been pulled from HDFS so that if a new node begins pulling data, it won't dup licate all of the work that has been done.</td></tr><tr><td id="name"><strong>Directory</strong></td><td id="default-value"></td><td id="allowable-values"></td><td id="description">The HDFS directory from which files should be read</td></tr><tr><td id="name"><strong>Recurse Subdirectories</strong></td><td id="default-value">true</td><td id="allowable-values"><ul><li>true</li><li>false</li></ul></td><td id="description">Indicates whether to list files from subdirectories of the HDFS directory</td></tr><tr><td id="name">Kerberos Principal</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Kerberos principal to authenticate as. Requires nifi.kerberos.krb5.file to be set in your nifi.properties</td></tr><tr><td id="name">Kerberos Keytab</td><td id="default-value"></td><td id="allowable-values"></td><td id="description">Kerberos keytab associated with the principal. Requires nifi.kerberos.krb5.file to be set in your nifi.properties</td></tr><tr><td id="na me">Kerberos Relogin Period</td><td id="default-value">4 hours</td><td id="allowable-values"></td><td id="description">Period of time which should pass before attempting a kerberos relogin</td></tr></table><h3>Relationships: </h3><table id="relationships"><tr><th>Name</th><th>Description</th></tr><tr><td>success</td><td>All FlowFiles are transferred to this relationship</td></tr></table><h3>Reads Attributes: </h3>None specified.<h3>Writes Attributes: </h3><table id="writes-attributes"><tr><th>Name</th><th>Description</th></tr><tr><td>filename</td><td>The name of the file that was read from HDFS.</td></tr><tr><td>path</td><td>The path is set to the absolute path of the file's directory on HDFS. For example, if the Directory property is set to /tmp, then files picked up from /tmp will have the path attribute set to "./". If the Recurse Subdirectories property is set to true and a file is picked up from /tmp/abc/1/2/3, then the path attribute will be set to "/tmp/abc/1/2/3".</td></tr>< tr><td>hdfs.owner</td><td>The user that owns the file in HDFS</td></tr><tr><td>hdfs.group</td><td>The group that owns the file in HDFS</td></tr><tr><td>hdfs.lastModified</td><td>The timestamp of when the file in HDFS was last modified, as milliseconds since midnight Jan 1, 1970 UTC</td></tr><tr><td>hdfs.length</td><td>The number of bytes in the file in HDFS</td></tr><tr><td>hdfs.replication</td><td>The number of HDFS replicas for hte file</td></tr><tr><td>hdfs.permissions</td><td>The permissions for the file in HDFS. This is formatted as 3 characters for the owner, 3 for the group, and 3 for other users. For example rw-rw-r--</td></tr></table><h3>See Also:</h3><p><a href="../org.apache.nifi.processors.hadoop.GetHDFS/index.html">GetHDFS</a>, <a href="../org.apache.nifi.processors.hadoop.FetchHDFS/index.html">FetchHDFS</a>, <a href="../org.apache.nifi.processors.hadoop.PutHDFS/index.html">PutHDFS</a></p></body></html> \ No newline at end of file
