Added: 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi/nifi-gcp-nar/1.9.0/org.apache.nifi.processors.gcp.storage.ListGCSBucket/index.html
URL: 
http://svn.apache.org/viewvc/nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi/nifi-gcp-nar/1.9.0/org.apache.nifi.processors.gcp.storage.ListGCSBucket/index.html?rev=1854109&view=auto
==============================================================================
--- 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi/nifi-gcp-nar/1.9.0/org.apache.nifi.processors.gcp.storage.ListGCSBucket/index.html
 (added)
+++ 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi/nifi-gcp-nar/1.9.0/org.apache.nifi.processors.gcp.storage.ListGCSBucket/index.html
 Fri Feb 22 01:03:44 2019
@@ -0,0 +1,4 @@
+<!DOCTYPE html><html lang="en"><head><meta 
charset="utf-8"></meta><title>ListGCSBucket</title><link rel="stylesheet" 
href="../../../../../css/component-usage.css" 
type="text/css"></link></head><script type="text/javascript">window.onload = 
function(){if(self==top) { document.getElementById('nameHeader').style.display 
= "inherit"; } }</script><body><h1 id="nameHeader" style="display: 
none;">ListGCSBucket</h1><h2>Description: </h2><p>Retrieves a listing of 
objects from an GCS bucket. For each object that is listed, creates a FlowFile 
that represents the object so that it can be fetched in conjunction with 
FetchGCSObject. This Processor is designed to run on Primary Node only in a 
cluster. If the primary node changes, the new Primary Node will pick up where 
the previous node left off without duplicating all of the data.</p><h3>Tags: 
</h3><p>google cloud, google, storage, gcs, list</p><h3>Properties: </h3><p>In 
the list below, the names of required properties appear in <strong>bold</str
 ong>. Any other properties (not in bold) are considered optional. The table 
also indicates any default values, whether a property supports the <a 
href="../../../../../html/expression-language-guide.html">NiFi Expression 
Language</a>, and whether a property is considered "sensitive", meaning that 
its value will be encrypted. Before entering a value in a sensitive property, 
ensure that the <strong>nifi.properties</strong> file has an entry for the 
property <strong>nifi.sensitive.props.key</strong>.</p><table 
id="properties"><tr><th>Name</th><th>Default Value</th><th>Allowable 
Values</th><th>Description</th></tr><tr><td id="name"><strong>GCP Credentials 
Provider Service</strong></td><td id="default-value"></td><td 
id="allowable-values"><strong>Controller Service API: 
</strong><br/>GCPCredentialsService<br/><strong>Implementation: </strong><a 
href="../org.apache.nifi.processors.gcp.credentials.service.GCPCredentialsControllerService/index.html">GCPCredentialsControllerService</a></td><t
 d id="description">The Controller Service used to obtain Google Cloud Platform 
credentials.</td></tr><tr><td id="name">Project ID</td><td 
id="default-value"></td><td id="allowable-values"></td><td 
id="description">Google Cloud Project ID<br/><strong>Supports Expression 
Language: true (will be evaluated using variable registry 
only)</strong></td></tr><tr><td id="name"><strong>Number of 
retries</strong></td><td id="default-value">6</td><td 
id="allowable-values"></td><td id="description">How many retry attempts should 
be made before routing to the failure relationship.</td></tr><tr><td 
id="name">Proxy host</td><td id="default-value"></td><td 
id="allowable-values"></td><td id="description">IP or hostname of the proxy to 
be used.
+ You might need to set the following properties in bootstrap for https proxy 
usage:
+-Djdk.http.auth.tunneling.disabledSchemes=
+-Djdk.http.auth.proxying.disabledSchemes=<br/><strong>Supports Expression 
Language: true (will be evaluated using variable registry 
only)</strong></td></tr><tr><td id="name">Proxy port</td><td 
id="default-value"></td><td id="allowable-values"></td><td 
id="description">Proxy port number<br/><strong>Supports Expression Language: 
true (will be evaluated using variable registry only)</strong></td></tr><tr><td 
id="name">HTTP Proxy Username</td><td id="default-value"></td><td 
id="allowable-values"></td><td id="description">HTTP Proxy 
Username<br/><strong>Supports Expression Language: true (will be evaluated 
using variable registry only)</strong></td></tr><tr><td id="name">HTTP Proxy 
Password</td><td id="default-value"></td><td id="allowable-values"></td><td 
id="description">HTTP Proxy Password<br/><strong>Sensitive Property: 
true</strong><br/><strong>Supports Expression Language: true (will be evaluated 
using variable registry only)</strong></td></tr><tr><td id="name">Proxy 
Configuration 
 Service</td><td id="default-value"></td><td 
id="allowable-values"><strong>Controller Service API: 
</strong><br/>ProxyConfigurationService<br/><strong>Implementation: </strong><a 
href="../../../nifi-proxy-configuration-nar/1.9.0/org.apache.nifi.proxy.StandardProxyConfigurationService/index.html">StandardProxyConfigurationService</a></td><td
 id="description">Specifies the Proxy Configuration Controller Service to proxy 
network requests. If set, it supersedes proxy settings configured per 
component. Supported proxies: HTTP + AuthN</td></tr><tr><td 
id="name"><strong>Bucket</strong></td><td id="default-value"></td><td 
id="allowable-values"></td><td id="description">Bucket of the 
object.<br/><strong>Supports Expression Language: true (will be evaluated using 
variable registry only)</strong></td></tr><tr><td id="name">Prefix</td><td 
id="default-value"></td><td id="allowable-values"></td><td id="description">The 
prefix used to filter the object list. In most cases, it should end with a forw
 ard slash ('/').<br/><strong>Supports Expression Language: true (will be 
evaluated using variable registry only)</strong></td></tr><tr><td 
id="name"><strong>Use Generations</strong></td><td 
id="default-value">false</td><td 
id="allowable-values"><ul><li>true</li><li>false</li></ul></td><td 
id="description">Specifies whether to use GCS Generations, if applicable.  If 
false, only the latest version of each object will be 
returned.</td></tr></table><h3>Relationships: </h3><table 
id="relationships"><tr><th>Name</th><th>Description</th></tr><tr><td>success</td><td>FlowFiles
 are routed to this relationship after a successful Google Cloud Storage 
operation.</td></tr></table><h3>Reads Attributes: </h3>None 
specified.<h3>Writes Attributes: </h3><table 
id="writes-attributes"><tr><th>Name</th><th>Description</th></tr><tr><td>filename</td><td>The
 name of the file</td></tr><tr><td>gcs.bucket</td><td>Bucket of the 
object.</td></tr><tr><td>gcs.key</td><td>Name of the 
object.</td></tr><tr><td>gcs.si
 ze</td><td>Size of the object.</td></tr><tr><td>gcs.cache.control</td><td>Data 
cache control of the object.</td></tr><tr><td>gcs.component.count</td><td>The 
number of components which make up the 
object.</td></tr><tr><td>gcs.content.disposition</td><td>The data content 
disposition of the object.</td></tr><tr><td>gcs.content.encoding</td><td>The 
content encoding of the 
object.</td></tr><tr><td>gcs.content.language</td><td>The content language of 
the object.</td></tr><tr><td>mime.type</td><td>The MIME/Content-Type of the 
object</td></tr><tr><td>gcs.crc32c</td><td>The CRC32C checksum of object's 
data, encoded in base64 in big-endian 
order.</td></tr><tr><td>gcs.create.time</td><td>The creation time of the object 
(milliseconds)</td></tr><tr><td>gcs.update.time</td><td>The last modification 
time of the object 
(milliseconds)</td></tr><tr><td>gcs.encryption.algorithm</td><td>The algorithm 
used to encrypt the object.</td></tr><tr><td>gcs.encryption.sha256</td><td>The 
SHA256 hash of the key u
 sed to encrypt the object</td></tr><tr><td>gcs.etag</td><td>The HTTP 1.1 
Entity tag for the object.</td></tr><tr><td>gcs.generated.id</td><td>The 
service-generated for the object</td></tr><tr><td>gcs.generation</td><td>The 
data generation of the object.</td></tr><tr><td>gcs.md5</td><td>The MD5 hash of 
the object's data encoded in 
base64.</td></tr><tr><td>gcs.media.link</td><td>The media download link to the 
object.</td></tr><tr><td>gcs.metageneration</td><td>The metageneration of the 
object.</td></tr><tr><td>gcs.owner</td><td>The owner (uploader) of the 
object.</td></tr><tr><td>gcs.owner.type</td><td>The ACL entity type of the 
uploader of the object.</td></tr><tr><td>gcs.uri</td><td>The URI of the object 
as a string.</td></tr></table><h3>State management: </h3><table 
id="stateful"><tr><th>Scope</th><th>Description</th></tr><tr><td>CLUSTER</td><td>After
 performing a listing of keys, the timestamp of the newest key is stored, along 
with the keys that share that same timestamp. This al
 lows the Processor to list only keys that have been added or modified after 
this date the next time that the Processor is run. State is stored across the 
cluster so that this Processor can be run on Primary Node only and if a new 
Primary Node is selected, the new node can pick up where the previous node left 
off, without duplicating the data.</td></tr></table><h3>Restricted: </h3>This 
component is not restricted.<h3>Input requirement: </h3>This component does not 
allow an incoming relationship.<h3>System Resource Considerations:</h3>None 
specified.<h3>See Also:</h3><p><a 
href="../org.apache.nifi.processors.gcp.storage.PutGCSObject/index.html">PutGCSObject</a>,
 <a 
href="../org.apache.nifi.processors.gcp.storage.DeleteGCSObject/index.html">DeleteGCSObject</a>,
 <a 
href="../org.apache.nifi.processors.gcp.storage.FetchGCSObject/index.html">FetchGCSObject</a></p></body></html>
\ No newline at end of file

Added: 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi/nifi-gcp-nar/1.9.0/org.apache.nifi.processors.gcp.storage.PutGCSObject/index.html
URL: 
http://svn.apache.org/viewvc/nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi/nifi-gcp-nar/1.9.0/org.apache.nifi.processors.gcp.storage.PutGCSObject/index.html?rev=1854109&view=auto
==============================================================================
--- 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi/nifi-gcp-nar/1.9.0/org.apache.nifi.processors.gcp.storage.PutGCSObject/index.html
 (added)
+++ 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi/nifi-gcp-nar/1.9.0/org.apache.nifi.processors.gcp.storage.PutGCSObject/index.html
 Fri Feb 22 01:03:44 2019
@@ -0,0 +1,4 @@
+<!DOCTYPE html><html lang="en"><head><meta 
charset="utf-8"></meta><title>PutGCSObject</title><link rel="stylesheet" 
href="../../../../../css/component-usage.css" 
type="text/css"></link></head><script type="text/javascript">window.onload = 
function(){if(self==top) { document.getElementById('nameHeader').style.display 
= "inherit"; } }</script><body><h1 id="nameHeader" style="display: 
none;">PutGCSObject</h1><h2>Description: </h2><p>Puts flow files to a Google 
Cloud Bucket.</p><h3>Tags: </h3><p>google, google cloud, gcs, archive, 
put</p><h3>Properties: </h3><p>In the list below, the names of required 
properties appear in <strong>bold</strong>. Any other properties (not in bold) 
are considered optional. The table also indicates any default values, whether a 
property supports the <a 
href="../../../../../html/expression-language-guide.html">NiFi Expression 
Language</a>, and whether a property is considered "sensitive", meaning that 
its value will be encrypted. Before entering a value in a
  sensitive property, ensure that the <strong>nifi.properties</strong> file has 
an entry for the property <strong>nifi.sensitive.props.key</strong>.</p><table 
id="properties"><tr><th>Name</th><th>Default Value</th><th>Allowable 
Values</th><th>Description</th></tr><tr><td id="name"><strong>GCP Credentials 
Provider Service</strong></td><td id="default-value"></td><td 
id="allowable-values"><strong>Controller Service API: 
</strong><br/>GCPCredentialsService<br/><strong>Implementation: </strong><a 
href="../org.apache.nifi.processors.gcp.credentials.service.GCPCredentialsControllerService/index.html">GCPCredentialsControllerService</a></td><td
 id="description">The Controller Service used to obtain Google Cloud Platform 
credentials.</td></tr><tr><td id="name">Project ID</td><td 
id="default-value"></td><td id="allowable-values"></td><td 
id="description">Google Cloud Project ID<br/><strong>Supports Expression 
Language: true (will be evaluated using variable registry 
only)</strong></td></tr><t
 r><td id="name"><strong>Number of retries</strong></td><td 
id="default-value">6</td><td id="allowable-values"></td><td 
id="description">How many retry attempts should be made before routing to the 
failure relationship.</td></tr><tr><td id="name">Proxy host</td><td 
id="default-value"></td><td id="allowable-values"></td><td id="description">IP 
or hostname of the proxy to be used.
+ You might need to set the following properties in bootstrap for https proxy 
usage:
+-Djdk.http.auth.tunneling.disabledSchemes=
+-Djdk.http.auth.proxying.disabledSchemes=<br/><strong>Supports Expression 
Language: true (will be evaluated using variable registry 
only)</strong></td></tr><tr><td id="name">Proxy port</td><td 
id="default-value"></td><td id="allowable-values"></td><td 
id="description">Proxy port number<br/><strong>Supports Expression Language: 
true (will be evaluated using variable registry only)</strong></td></tr><tr><td 
id="name">HTTP Proxy Username</td><td id="default-value"></td><td 
id="allowable-values"></td><td id="description">HTTP Proxy 
Username<br/><strong>Supports Expression Language: true (will be evaluated 
using variable registry only)</strong></td></tr><tr><td id="name">HTTP Proxy 
Password</td><td id="default-value"></td><td id="allowable-values"></td><td 
id="description">HTTP Proxy Password<br/><strong>Sensitive Property: 
true</strong><br/><strong>Supports Expression Language: true (will be evaluated 
using variable registry only)</strong></td></tr><tr><td id="name">Proxy 
Configuration 
 Service</td><td id="default-value"></td><td 
id="allowable-values"><strong>Controller Service API: 
</strong><br/>ProxyConfigurationService<br/><strong>Implementation: </strong><a 
href="../../../nifi-proxy-configuration-nar/1.9.0/org.apache.nifi.proxy.StandardProxyConfigurationService/index.html">StandardProxyConfigurationService</a></td><td
 id="description">Specifies the Proxy Configuration Controller Service to proxy 
network requests. If set, it supersedes proxy settings configured per 
component. Supported proxies: HTTP + AuthN</td></tr><tr><td 
id="name"><strong>Bucket</strong></td><td 
id="default-value">${gcs.bucket}</td><td id="allowable-values"></td><td 
id="description">Bucket of the object.<br/><strong>Supports Expression 
Language: true (will be evaluated using flow file attributes and variable 
registry)</strong></td></tr><tr><td id="name"><strong>Key</strong></td><td 
id="default-value">${filename}</td><td id="allowable-values"></td><td 
id="description">Name of the object.<br/><
 strong>Supports Expression Language: true (will be evaluated using flow file 
attributes and variable registry)</strong></td></tr><tr><td id="name">Content 
Type</td><td id="default-value">${mime.type}</td><td 
id="allowable-values"></td><td id="description">Content Type for the file, i.e. 
text/plain<br/><strong>Supports Expression Language: true (will be evaluated 
using flow file attributes and variable registry)</strong></td></tr><tr><td 
id="name">MD5 Hash</td><td id="default-value"></td><td 
id="allowable-values"></td><td id="description">MD5 Hash (encoded in Base64) of 
the file for server-side validation.<br/><strong>Supports Expression Language: 
true (will be evaluated using flow file attributes and variable 
registry)</strong></td></tr><tr><td id="name">CRC32C Checksum</td><td 
id="default-value"></td><td id="allowable-values"></td><td 
id="description">CRC32C Checksum (encoded in Base64, big-Endian order) of the 
file for server-side validation.<br/><strong>Supports Expression Langua
 ge: true (will be evaluated using flow file attributes and variable 
registry)</strong></td></tr><tr><td id="name">Object ACL</td><td 
id="default-value"></td><td id="allowable-values"><ul><li>All Authenticated 
Users <img src="../../../../../html/images/iconInfo.png" alt="Gives the bucket 
or object owner OWNER permission, and gives all authenticated Google account 
holders READER and WRITER permissions. All other permissions are removed." 
title="Gives the bucket or object owner OWNER permission, and gives all 
authenticated Google account holders READER and WRITER permissions. All other 
permissions are removed."></img></li><li>Authenticated Read <img 
src="../../../../../html/images/iconInfo.png" alt="Gives the bucket or object 
owner OWNER permission, and gives all authenticated Google account holders 
READER permission. All other permissions are removed." title="Gives the bucket 
or object owner OWNER permission, and gives all authenticated Google account 
holders READER permission. All ot
 her permissions are removed."></img></li><li>Bucket Owner Full Control <img 
src="../../../../../html/images/iconInfo.png" alt="Gives the object and bucket 
owners OWNER permission. All other permissions are removed." title="Gives the 
object and bucket owners OWNER permission. All other permissions are 
removed."></img></li><li>Bucket Owner Read Only <img 
src="../../../../../html/images/iconInfo.png" alt="Gives the object owner OWNER 
permission, and gives the bucket owner READER permission. All other permissions 
are removed." title="Gives the object owner OWNER permission, and gives the 
bucket owner READER permission. All other permissions are 
removed."></img></li><li>Private <img 
src="../../../../../html/images/iconInfo.png" alt="Gives the bucket or object 
owner OWNER permission for a bucket or object, and removes all other access 
permissions." title="Gives the bucket or object owner OWNER permission for a 
bucket or object, and removes all other access permissions."></img></li><li>Pro
 ject Private <img src="../../../../../html/images/iconInfo.png" alt="Gives 
permission to the project team based on their roles. Anyone who is part of the 
team has READER permission. Project owners and project editors have OWNER 
permission. This is the default ACL for newly created buckets. This is also the 
default ACL for newly created objects unless the default object ACL for that 
bucket has been changed." title="Gives permission to the project team based on 
their roles. Anyone who is part of the team has READER permission. Project 
owners and project editors have OWNER permission. This is the default ACL for 
newly created buckets. This is also the default ACL for newly created objects 
unless the default object ACL for that bucket has been 
changed."></img></li><li>Public Read Only <img 
src="../../../../../html/images/iconInfo.png" alt="Gives the bucket or object 
owner OWNER permission, and gives all users, both authenticated and anonymous, 
READER permission. When you apply this to a
 n object, anyone on the Internet can read the object without authenticating." 
title="Gives the bucket or object owner OWNER permission, and gives all users, 
both authenticated and anonymous, READER permission. When you apply this to an 
object, anyone on the Internet can read the object without 
authenticating."></img></li></ul></td><td id="description">Access Control to be 
attached to the object uploaded. Not providing this will revert to bucket 
defaults.</td></tr><tr><td id="name">Server Side Encryption Key</td><td 
id="default-value"></td><td id="allowable-values"></td><td id="description">An 
AES256 Encryption Key (encoded in base64) for server-side encryption of the 
object.<br/><strong>Sensitive Property: true</strong><br/><strong>Supports 
Expression Language: true (will be evaluated using flow file attributes and 
variable registry)</strong></td></tr><tr><td id="name"><strong>Overwrite 
Object</strong></td><td id="default-value">true</td><td 
id="allowable-values"><ul><li>true</li><l
 i>false</li></ul></td><td id="description">If false, the upload to GCS will 
succeed only if the object does not exist.</td></tr><tr><td id="name">Content 
Disposition Type</td><td id="default-value"></td><td 
id="allowable-values"><ul><li>Inline <img 
src="../../../../../html/images/iconInfo.png" alt="Indicates that the object 
should be loaded and rendered within the browser." title="Indicates that the 
object should be loaded and rendered within the 
browser."></img></li><li>Attachment <img 
src="../../../../../html/images/iconInfo.png" alt="Indicates that the object 
should be saved (using a Save As... dialog) rather than opened directly within 
the browser" title="Indicates that the object should be saved (using a Save 
As... dialog) rather than opened directly within the 
browser"></img></li></ul></td><td id="description">Type of RFC-6266 Content 
Disposition to be attached to the object</td></tr></table><h3>Dynamic 
Properties: </h3><p>Dynamic Properties allow the user to specify both the 
 name and value of a property.<table 
id="dynamic-properties"><tr><th>Name</th><th>Value</th><th>Description</th></tr><tr><td
 id="name">The name of a User-Defined Metadata field to add to the GCS 
Object</td><td id="value">The value of a User-Defined Metadata field to add to 
the GCS Object</td><td>Allows user-defined metadata to be added to the GCS 
object as key/value pairs<br/><strong>Supports Expression Language: true (will 
be evaluated using flow file attributes and variable 
registry)</strong></td></tr></table></p><h3>Relationships: </h3><table 
id="relationships"><tr><th>Name</th><th>Description</th></tr><tr><td>success</td><td>FlowFiles
 are routed to this relationship after a successful Google Cloud Storage 
operation.</td></tr><tr><td>failure</td><td>FlowFiles are routed to this 
relationship if the Google Cloud Storage operation 
fails.</td></tr></table><h3>Reads Attributes: </h3><table 
id="reads-attributes"><tr><th>Name</th><th>Description</th></tr><tr><td>filename</td><td>Uses
 the
  FlowFile's filename as the filename for the GCS 
object</td></tr><tr><td>mime.type</td><td>Uses the FlowFile's MIME type as the 
content-type for the GCS object</td></tr></table><h3>Writes Attributes: 
</h3><table 
id="writes-attributes"><tr><th>Name</th><th>Description</th></tr><tr><td>gcs.bucket</td><td>Bucket
 of the object.</td></tr><tr><td>gcs.key</td><td>Name of the 
object.</td></tr><tr><td>gcs.size</td><td>Size of the 
object.</td></tr><tr><td>gcs.cache.control</td><td>Data cache control of the 
object.</td></tr><tr><td>gcs.component.count</td><td>The number of components 
which make up the object.</td></tr><tr><td>gcs.content.disposition</td><td>The 
data content disposition of the 
object.</td></tr><tr><td>gcs.content.encoding</td><td>The content encoding of 
the object.</td></tr><tr><td>gcs.content.language</td><td>The content language 
of the object.</td></tr><tr><td>mime.type</td><td>The MIME/Content-Type of the 
object</td></tr><tr><td>gcs.crc32c</td><td>The CRC32C checksum of obje
 ct's data, encoded in base64 in big-endian 
order.</td></tr><tr><td>gcs.create.time</td><td>The creation time of the object 
(milliseconds)</td></tr><tr><td>gcs.update.time</td><td>The last modification 
time of the object 
(milliseconds)</td></tr><tr><td>gcs.encryption.algorithm</td><td>The algorithm 
used to encrypt the object.</td></tr><tr><td>gcs.encryption.sha256</td><td>The 
SHA256 hash of the key used to encrypt the 
object</td></tr><tr><td>gcs.etag</td><td>The HTTP 1.1 Entity tag for the 
object.</td></tr><tr><td>gcs.generated.id</td><td>The service-generated for the 
object</td></tr><tr><td>gcs.generation</td><td>The data generation of the 
object.</td></tr><tr><td>gcs.md5</td><td>The MD5 hash of the object's data 
encoded in base64.</td></tr><tr><td>gcs.media.link</td><td>The media download 
link to the object.</td></tr><tr><td>gcs.metageneration</td><td>The 
metageneration of the object.</td></tr><tr><td>gcs.owner</td><td>The owner 
(uploader) of the object.</td></tr><tr><td>gcs.owner.
 type</td><td>The ACL entity type of the uploader of the 
object.</td></tr><tr><td>gcs.uri</td><td>The URI of the object as a 
string.</td></tr></table><h3>State management: </h3>This component does not 
store state.<h3>Restricted: </h3>This component is not restricted.<h3>Input 
requirement: </h3>This component requires an incoming relationship.<h3>System 
Resource Considerations:</h3>None specified.<h3>See Also:</h3><p><a 
href="../org.apache.nifi.processors.gcp.storage.FetchGCSObject/index.html">FetchGCSObject</a>,
 <a 
href="../org.apache.nifi.processors.gcp.storage.DeleteGCSObject/index.html">DeleteGCSObject</a>,
 <a 
href="../org.apache.nifi.processors.gcp.storage.ListGCSBucket/index.html">ListGCSBucket</a></p></body></html>
\ No newline at end of file

Added: 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi/nifi-groovyx-nar/1.9.0/org.apache.nifi.processors.groovyx.ExecuteGroovyScript/SQL.gif
URL: 
http://svn.apache.org/viewvc/nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi/nifi-groovyx-nar/1.9.0/org.apache.nifi.processors.groovyx.ExecuteGroovyScript/SQL.gif?rev=1854109&view=auto
==============================================================================
Binary file - no diff available.

Propchange: 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi/nifi-groovyx-nar/1.9.0/org.apache.nifi.processors.groovyx.ExecuteGroovyScript/SQL.gif
------------------------------------------------------------------------------
    svn:mime-type = application/octet-stream

Added: 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi/nifi-groovyx-nar/1.9.0/org.apache.nifi.processors.groovyx.ExecuteGroovyScript/SQL2.gif
URL: 
http://svn.apache.org/viewvc/nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi/nifi-groovyx-nar/1.9.0/org.apache.nifi.processors.groovyx.ExecuteGroovyScript/SQL2.gif?rev=1854109&view=auto
==============================================================================
Binary file - no diff available.

Propchange: 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi/nifi-groovyx-nar/1.9.0/org.apache.nifi.processors.groovyx.ExecuteGroovyScript/SQL2.gif
------------------------------------------------------------------------------
    svn:mime-type = application/octet-stream

Added: 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi/nifi-groovyx-nar/1.9.0/org.apache.nifi.processors.groovyx.ExecuteGroovyScript/additionalDetails.html
URL: 
http://svn.apache.org/viewvc/nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi/nifi-groovyx-nar/1.9.0/org.apache.nifi.processors.groovyx.ExecuteGroovyScript/additionalDetails.html?rev=1854109&view=auto
==============================================================================
--- 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi/nifi-groovyx-nar/1.9.0/org.apache.nifi.processors.groovyx.ExecuteGroovyScript/additionalDetails.html
 (added)
+++ 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi/nifi-groovyx-nar/1.9.0/org.apache.nifi.processors.groovyx.ExecuteGroovyScript/additionalDetails.html
 Fri Feb 22 01:03:44 2019
@@ -0,0 +1,203 @@
+<!DOCTYPE html>
+<html lang="en">
+<!--
+  Licensed to the Apache Software Foundation (ASF) under one or more
+  contributor license agreements.  See the NOTICE file distributed with
+  this work for additional information regarding copyright ownership.
+  The ASF licenses this file to You under the Apache License, Version 2.0
+  (the "License"); you may not use this file except in compliance with
+  the License.  You may obtain a copy of the License at
+      http://www.apache.org/licenses/LICENSE-2.0
+  Unless required by applicable law or agreed to in writing, software
+  distributed under the License is distributed on an "AS IS" BASIS,
+  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+  See the License for the specific language governing permissions and
+  limitations under the License.
+-->
+<head>
+    <meta charset="utf-8" />
+    <title>Groovy</title>
+    <!--link rel="stylesheet" href="../../css/component-usage.css" 
type="text/css" /-->
+    <link rel="stylesheet" href="../../../../../css/component-usage.css" 
type="text/css" />
+</head>
+
+<body>
+<h2>Summary</h2>
+<p>This is a grooviest groovy script :)</p>
+<h2>Script Bindings:</h2>
+<table>
+<tr><th>variable</th><th>type</th><th>description</th></tr>
+<tr>
+       <td>session</td>
+       <td>org.apache.nifi.processor.ProcessSession</td>
+       <td>the session that is used to get, change, and transfer input 
files</td>
+</tr>
+<tr>
+       <td>context</td>
+       <td>org.apache.nifi.processor.ProcessContext</td>
+       <td>the context (almost unuseful)</td>
+</tr>
+<tr>
+       <td>log</td>
+       <td>org.apache.nifi.logging.ComponentLog</td>
+       <td>the logger for this processor instance</td>
+</tr>
+<tr>
+       <td>REL_SUCCESS</td>
+       <td>org.apache.nifi.processor.Relationship</td>
+       <td>the success relationship</td>
+</tr>
+<tr>
+       <td>REL_FAILURE</td>
+       <td>org.apache.nifi.processor.Relationship</td>
+       <td>the failure relationship</td>
+</tr>
+<tr>
+       <td>CTL</td>
+       <td>java.util.HashMap&lt;String,<a 
href="https://github.com/apache/nifi/blob/master/nifi-api/src/main/java/org/apache/nifi/controller/ControllerService.java";>ControllerService</a>&gt;</td>
+       <td>Map populated with controller services defined with `CTL.*` 
processor properties.
+       <br/>The `CTL.` prefixed properties could be linked to controller 
service and provides access to this service from a script without additional 
code.</td>
+</tr>
+<tr>
+       <td>SQL</td>
+       <td>java.util.HashMap&lt;String, <a 
href="http://docs.groovy-lang.org/latest/html/api/groovy/sql/Sql.html";>groovy.sql.Sql</a>&gt;</td>
+       <td>Map populated with `groovy.sql.Sql` objects connected to 
corresponding database defined with `SQL.*` processor properties. 
+       <br/>The `SQL.` prefixed properties could be linked only to 
DBCPSercice.</td>
+</tr>
+<tr>
+       <td>Dynamic processor properties</td>
+       <td>org.apache.nifi.components.PropertyDescriptor</td>
+       <td>All processor properties not started with `CTL.` or `SQL.` are 
bound to script variables</td>
+</tr>
+</table>
+
+<h2>SQL map details</h2>
+<p>
+<b>Example:</b> if you defined property <code>`SQL.mydb`</code> and linked it 
to any DBCPService, 
+then you can access it from code <code>SQL.mydb.rows('select * from 
mytable')</code><br/>
+
+<br/>The processor automatically takes connection from dbcp service before 
executing script and tries to handle transaction:
+<br/> database transactions automatically rolled back on script exception and 
committed on success.
+<br/>Or you can manage transaction manually.
+<br/>NOTE: Script must not disconnect connection.
+
+<br/><img src="SQL.gif"/>
+<br/><img src="SQL2.gif"/>
+</p>
+
+<h2>SessionFile - flow file extension</h2>
+<p>
+  The (org.apache.nifi.processors.groovyx.flow.SessionFile) is an actual 
object returned by session in Extended Groovy processor.<br/>
+  This flow file is a container that references session and the real flow 
file.<br/>
+  This allows to use simplified syntax to work with file attributes and 
content:
+</p>
+<p><i>set new attribute value</i></p>
+<pre>
+  flowFile.ATTRIBUTE_NAME = ATTRIBUTE_VALUE 
+  flowFile.'mime.type' = 'text/xml' 
+  flowFile.putAttribute("ATTRIBUTE_NAME", ATTRIBUTE_VALUE)
+  //the same as
+  flowFile = session.putAttribute(flowFile, "ATTRIBUTE_NAME", ATTRIBUTE_VALUE)
+</pre>
+  
+<p><i>remove attribute</i></p>
+<pre>
+  flowFile.ATTRIBUTE_NAME = null
+  //equals to
+  flowFile = session.removeAttribute(flowFile, "ATTRIBUTE_NAME")
+</pre>
+
+<p><i>get attribute value</i></p>
+<pre>
+  String a = flowFile.ATTRIBUTE_NAME
+</pre>
+  
+<p><i>write content</i></p>
+<pre>
+  flowFile.write("UTF-8", "THE CharSequence to write into flow file replacing 
current content")
+  flowFile.write("UTF-8"){writer-> 
+    do something with java.io.Writer...
+  }
+  flowFile.write{outStream-> 
+    do something with output stream...
+  }
+  flowFile.write{inStream, outStream-> 
+    do something with input and output streams... 
+  }
+</pre>
+
+<p><i>get content</i></p>
+<pre>
+  InputStream i = flowFile.read()
+  def json = new groovy.json.JsonSlurper().parse( flowFile.read() )
+  String text = flowFile.read().getText("UTF-8")
+</pre>
+  
+<p><i>transfer flow file to success relation</i></p>
+<pre>
+  REL_SUCCESS << flowFile 
+  flowFile.transfer(REL_SUCCESS)
+  //the same as:
+  session.transfer(flowFile, REL_SUCCESS)
+</pre>
+
+<p><i>work with dbcp</i></p>
+<pre>
+  import groovy.sql.Sql
+
+  //define property named `SQL.db` connected to a DBCPConnectionPool 
controller service
+  //for this case it's an H2 database example
+  
+  //read value from the database with prepared statement 
+  //and assign into flowfile attribute `db.yesterday`
+  def daysAdd = -1
+  def row = SQL.db.firstRow("select dateadd('DAY', ${daysAdd}, sysdate) as 
DB_DATE from dual")
+  flowFile.'db.yesterday' = row.DB_DATE
+
+  //to work with BLOBs and CLOBs in the database 
+  //use parameter casting using groovy.sql.Sql.BLOB(Stream) and 
groovy.sql.Sql.CLOB(Reader)
+  
+  //write content of the flow file into database blob
+  flowFile.read{ rawIn->
+    def parms = [
+      p_id   : flowFile.ID as Long, //get flow file attribute named `ID`
+      p_data : Sql.BLOB( rawIn ),   //use input stream as BLOB sql parameter
+    ]
+    SQL.db.executeUpdate(parms, "update mytable set data = :p_data where id = 
:p_id")
+  }
+</pre>
+
+
+<h2>Handling processor start &amp; stop</h2>
+
+<p>In the extended groovy processor you can catch `start` and `stop` events by 
providing corresponding static methods:</p>
+<pre>
+  import org.apache.nifi.processor.ProcessContext
+  import java.util.concurrent.atomic.AtomicLong
+
+  class Const{
+    static Date startTime = null;
+    static AtomicLong triggerCount = null;
+  }
+
+  static onStart(ProcessContext context){
+    Const.startTime = new Date()
+    Const.triggerCount = new AtomicLong(0)
+    println "onStart $context ${Const.startTime}"
+  }
+
+  static onStop(ProcessContext context){
+    def alive = (System.currentTimeMillis() - Const.startTime.getTime()) / 1000
+    println "onStop $context executed ${ Const.triggerCount } times during ${ 
alive } seconds"
+  }
+
+
+  flowFile.'trigger.count' = Const.triggerCount.incrementAndGet()
+  REL_SUCCESS << flowFile
+</pre>
+<br/>
+<br/>
+<br/>
+<br/>
+</body>
+</html>
\ No newline at end of file

Added: 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi/nifi-groovyx-nar/1.9.0/org.apache.nifi.processors.groovyx.ExecuteGroovyScript/index.html
URL: 
http://svn.apache.org/viewvc/nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi/nifi-groovyx-nar/1.9.0/org.apache.nifi.processors.groovyx.ExecuteGroovyScript/index.html?rev=1854109&view=auto
==============================================================================
--- 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi/nifi-groovyx-nar/1.9.0/org.apache.nifi.processors.groovyx.ExecuteGroovyScript/index.html
 (added)
+++ 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi/nifi-groovyx-nar/1.9.0/org.apache.nifi.processors.groovyx.ExecuteGroovyScript/index.html
 Fri Feb 22 01:03:44 2019
@@ -0,0 +1 @@
+<!DOCTYPE html><html lang="en"><head><meta 
charset="utf-8"></meta><title>ExecuteGroovyScript</title><link rel="stylesheet" 
href="../../../../../css/component-usage.css" 
type="text/css"></link></head><script type="text/javascript">window.onload = 
function(){if(self==top) { document.getElementById('nameHeader').style.display 
= "inherit"; } }</script><body><h1 id="nameHeader" style="display: 
none;">ExecuteGroovyScript</h1><h2>Description: </h2><p>Experimental Extended 
Groovy script processor. The script is responsible for handling the incoming 
flow file (transfer to SUCCESS or remove, e.g.) as well as any flow files 
created by the script. If the handling is incomplete or incorrect, the session 
will be rolled back.</p><p><a href="additionalDetails.html">Additional 
Details...</a></p><h3>Tags: </h3><p>script, groovy, groovyx</p><h3>Properties: 
</h3><p>In the list below, the names of required properties appear in 
<strong>bold</strong>. Any other properties (not in bold) are considered opti
 onal. The table also indicates any default values, and whether a property 
supports the <a href="../../../../../html/expression-language-guide.html">NiFi 
Expression Language</a>.</p><table id="properties"><tr><th>Name</th><th>Default 
Value</th><th>Allowable Values</th><th>Description</th></tr><tr><td 
id="name">Script File</td><td id="default-value"></td><td 
id="allowable-values"></td><td id="description">Path to script file to execute. 
Only one of Script File or Script Body may be used<br/><strong>Supports 
Expression Language: true (will be evaluated using variable registry 
only)</strong></td></tr><tr><td id="name">Script Body</td><td 
id="default-value"></td><td id="allowable-values"></td><td 
id="description">Body of script to execute. Only one of Script File or Script 
Body may be used</td></tr><tr><td id="name"><strong>Failure 
strategy</strong></td><td id="default-value">rollback</td><td 
id="allowable-values"><ul><li>rollback</li><li>transfer to 
failure</li></ul></td><td id="descrip
 tion">What to do with unhandled exceptions. If you want to manage exception by 
code then keep the default value `rollback`. If `transfer to failure` selected 
and unhandled exception occurred then all flowFiles received from incoming 
queues in this session will be transferred to `failure` relationship with 
additional attributes set: ERROR_MESSAGE and ERROR_STACKTRACE. If `rollback` 
selected and unhandled exception occurred then all flowFiles received from 
incoming queues will be penalized and returned. If the processor has no 
incoming connections then this parameter has no effect.</td></tr><tr><td 
id="name">Additional classpath</td><td id="default-value"></td><td 
id="allowable-values"></td><td id="description">Classpath list separated by 
semicolon. You can use masks like `*`, `*.jar` in file 
name.<br/><strong>Supports Expression Language: true (will be evaluated using 
variable registry only)</strong></td></tr></table><h3>Dynamic Properties: 
</h3><p>Dynamic Properties allow the user t
 o specify both the name and value of a property.<table 
id="dynamic-properties"><tr><th>Name</th><th>Value</th><th>Description</th></tr><tr><td
 id="name">A script engine property to update</td><td id="value">The value to 
set it to</td><td>Updates a script engine property specified by the Dynamic 
Property's key with the value specified by the Dynamic Property's value. Use 
`CTL.` to access any controller services.<br/><strong>Supports Expression 
Language: true (will be evaluated using flow file attributes and variable 
registry)</strong></td></tr></table></p><h3>Relationships: </h3><table 
id="relationships"><tr><th>Name</th><th>Description</th></tr><tr><td>success</td><td>FlowFiles
 that were successfully processed</td></tr><tr><td>failure</td><td>FlowFiles 
that failed to be processed</td></tr></table><h3>Reads Attributes: </h3>None 
specified.<h3>Writes Attributes: </h3>None specified.<h3>State management: 
</h3>This component does not store state.<h3>Restricted: </h3><table 
id="restricti
 ons"><tr><th>Required Permission</th><th>Explanation</th></tr><tr><td>execute 
code</td><td>Provides operator the ability to execute arbitrary code assuming 
all permissions that NiFi has.</td></tr></table><h3>Input requirement: 
</h3>This component allows an incoming relationship.<h3>System Resource 
Considerations:</h3>None specified.<h3>See Also:</h3><p><a 
href="../../../nifi-scripting-nar/1.9.0/org.apache.nifi.processors.script.ExecuteScript/index.html">ExecuteScript</a></p></body></html>
\ No newline at end of file

Added: 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi/nifi-hadoop-nar/1.9.0/org.apache.nifi.processors.hadoop.CreateHadoopSequenceFile/additionalDetails.html
URL: 
http://svn.apache.org/viewvc/nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi/nifi-hadoop-nar/1.9.0/org.apache.nifi.processors.hadoop.CreateHadoopSequenceFile/additionalDetails.html?rev=1854109&view=auto
==============================================================================
--- 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi/nifi-hadoop-nar/1.9.0/org.apache.nifi.processors.hadoop.CreateHadoopSequenceFile/additionalDetails.html
 (added)
+++ 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi/nifi-hadoop-nar/1.9.0/org.apache.nifi.processors.hadoop.CreateHadoopSequenceFile/additionalDetails.html
 Fri Feb 22 01:03:44 2019
@@ -0,0 +1,46 @@
+<!DOCTYPE html>
+<html lang="en">
+    <!--
+      Licensed to the Apache Software Foundation (ASF) under one or more
+      contributor license agreements.  See the NOTICE file distributed with
+      this work for additional information regarding copyright ownership.
+      The ASF licenses this file to You under the Apache License, Version 2.0
+      (the "License"); you may not use this file except in compliance with
+      the License.  You may obtain a copy of the License at
+          http://www.apache.org/licenses/LICENSE-2.0
+      Unless required by applicable law or agreed to in writing, software
+      distributed under the License is distributed on an "AS IS" BASIS,
+      WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+      See the License for the specific language governing permissions and
+      limitations under the License.
+    -->
+    <head>
+        <meta charset="utf-8" />
+        <title>CreateHadoopSequenceFile</title>
+
+        <link rel="stylesheet" href="../../../../../css/component-usage.css" 
type="text/css" />
+    </head>
+
+    <body>
+        <!-- Processor Documentation 
================================================== -->
+        <h2>Description:</h2>
+        <p>This processor is used to create a Hadoop Sequence File, which 
essentially is a file of key/value pairs. The key 
+            will be a file name and the value will be the flow file content. 
The processor will take either a merged (a.k.a. packaged) flow 
+            file or a singular flow file. Historically, this processor handled 
the merging by type and size or time prior to creating a 
+            SequenceFile output; it no longer does this. If creating a 
SequenceFile that contains multiple files of the same type is desired,
+            precede this processor with a <code>RouteOnAttribute</code> 
processor to segregate files of the same type and follow that with a
+            <code>MergeContent</code> processor to bundle up files. If the 
type of files is not important, just use the 
+            <code>MergeContent</code> processor. When using the 
<code>MergeContent</code> processor, the following Merge Formats are 
+            supported by this processor:
+        <ul>
+            <li>TAR</li>
+            <li>ZIP</li>
+            <li>FlowFileStream v3</li>
+        </ul>
+        The created SequenceFile is named the same as the incoming FlowFile 
with the suffix '.sf'. For incoming FlowFiles that are 
+        bundled, the keys in the SequenceFile are the individual file names, 
the values are the contents of each file.
+    </p>
+    NOTE: The value portion of a key/value pair is loaded into memory. While 
there is a max size limit of 2GB, this could cause memory
+    issues if there are too many concurrent tasks and the flow file sizes are 
large.
+</body>
+</html>

Added: 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi/nifi-hadoop-nar/1.9.0/org.apache.nifi.processors.hadoop.CreateHadoopSequenceFile/index.html
URL: 
http://svn.apache.org/viewvc/nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi/nifi-hadoop-nar/1.9.0/org.apache.nifi.processors.hadoop.CreateHadoopSequenceFile/index.html?rev=1854109&view=auto
==============================================================================
--- 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi/nifi-hadoop-nar/1.9.0/org.apache.nifi.processors.hadoop.CreateHadoopSequenceFile/index.html
 (added)
+++ 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi/nifi-hadoop-nar/1.9.0/org.apache.nifi.processors.hadoop.CreateHadoopSequenceFile/index.html
 Fri Feb 22 01:03:44 2019
@@ -0,0 +1,3 @@
+<!DOCTYPE html><html lang="en"><head><meta 
charset="utf-8"></meta><title>CreateHadoopSequenceFile</title><link 
rel="stylesheet" href="../../../../../css/component-usage.css" 
type="text/css"></link></head><script type="text/javascript">window.onload = 
function(){if(self==top) { document.getElementById('nameHeader').style.display 
= "inherit"; } }</script><body><h1 id="nameHeader" style="display: 
none;">CreateHadoopSequenceFile</h1><h2>Description: </h2><p>Creates Hadoop 
Sequence Files from incoming flow files</p><p><a 
href="additionalDetails.html">Additional Details...</a></p><h3>Tags: 
</h3><p>hadoop, sequence file, create, sequencefile</p><h3>Properties: 
</h3><p>In the list below, the names of required properties appear in 
<strong>bold</strong>. Any other properties (not in bold) are considered 
optional. The table also indicates any default values, and whether a property 
supports the <a href="../../../../../html/expression-language-guide.html">NiFi 
Expression Language</a>.</p><table 
 id="properties"><tr><th>Name</th><th>Default Value</th><th>Allowable 
Values</th><th>Description</th></tr><tr><td id="name">Hadoop Configuration 
Resources</td><td id="default-value"></td><td id="allowable-values"></td><td 
id="description">A file or comma separated list of files which contains the 
Hadoop file system configuration. Without this, Hadoop will search the 
classpath for a 'core-site.xml' and 'hdfs-site.xml' file or will revert to a 
default configuration. To use swebhdfs, see 'Additional Details' section of 
PutHDFS's documentation.<br/><strong>Supports Expression Language: true (will 
be evaluated using variable registry only)</strong></td></tr><tr><td 
id="name">Kerberos Credentials Service</td><td id="default-value"></td><td 
id="allowable-values"><strong>Controller Service API: 
</strong><br/>KerberosCredentialsService<br/><strong>Implementation: 
</strong><a 
href="../../../nifi-kerberos-credentials-service-nar/1.9.0/org.apache.nifi.kerberos.KeytabCredentialsService/index.html
 ">KeytabCredentialsService</a></td><td id="description">Specifies the Kerberos 
Credentials Controller Service that should be used for authenticating with 
Kerberos</td></tr><tr><td id="name">Kerberos Principal</td><td 
id="default-value"></td><td id="allowable-values"></td><td 
id="description">Kerberos principal to authenticate as. Requires 
nifi.kerberos.krb5.file to be set in your nifi.properties<br/><strong>Supports 
Expression Language: true (will be evaluated using variable registry 
only)</strong></td></tr><tr><td id="name">Kerberos Keytab</td><td 
id="default-value"></td><td id="allowable-values"></td><td 
id="description">Kerberos keytab associated with the principal. Requires 
nifi.kerberos.krb5.file to be set in your nifi.properties<br/><strong>Supports 
Expression Language: true (will be evaluated using variable registry 
only)</strong></td></tr><tr><td id="name">Kerberos Relogin Period</td><td 
id="default-value">4 hours</td><td id="allowable-values"></td><td 
id="description">Perio
 d of time which should pass before attempting a kerberos relogin.
+
+This property has been deprecated, and has no effect on processing. Relogins 
now occur automatically.<br/><strong>Supports Expression Language: true (will 
be evaluated using variable registry only)</strong></td></tr><tr><td 
id="name">Additional Classpath Resources</td><td id="default-value"></td><td 
id="allowable-values"></td><td id="description">A comma-separated list of paths 
to files and/or directories that will be added to the classpath. When 
specifying a directory, all files with in the directory will be added to the 
classpath, but further sub-directories will not be included.</td></tr><tr><td 
id="name">Compression type</td><td id="default-value"></td><td 
id="allowable-values"><ul><li>NONE</li><li>RECORD</li><li>BLOCK</li></ul></td><td
 id="description">Type of compression to use when creating Sequence 
File</td></tr><tr><td id="name"><strong>Compression codec</strong></td><td 
id="default-value">NONE</td><td id="allowable-values"><ul><li>NONE <img 
src="../../../../../html/images/
 iconInfo.png" alt="No compression" title="No 
compression"></img></li><li>DEFAULT <img 
src="../../../../../html/images/iconInfo.png" alt="Default ZLIB compression" 
title="Default ZLIB compression"></img></li><li>BZIP <img 
src="../../../../../html/images/iconInfo.png" alt="BZIP compression" 
title="BZIP compression"></img></li><li>GZIP <img 
src="../../../../../html/images/iconInfo.png" alt="GZIP compression" 
title="GZIP compression"></img></li><li>LZ4 <img 
src="../../../../../html/images/iconInfo.png" alt="LZ4 compression" title="LZ4 
compression"></img></li><li>LZO <img 
src="../../../../../html/images/iconInfo.png" alt="LZO compression - it assumes 
LD_LIBRARY_PATH has been set and jar is available" title="LZO compression - it 
assumes LD_LIBRARY_PATH has been set and jar is 
available"></img></li><li>SNAPPY <img 
src="../../../../../html/images/iconInfo.png" alt="Snappy compression" 
title="Snappy compression"></img></li><li>AUTOMATIC <img 
src="../../../../../html/images/iconInfo.png" alt=
 "Will attempt to automatically detect the compression codec." title="Will 
attempt to automatically detect the compression 
codec."></img></li></ul></td><td id="description">No Description 
Provided.</td></tr></table><h3>Relationships: </h3><table 
id="relationships"><tr><th>Name</th><th>Description</th></tr><tr><td>success</td><td>Generated
 Sequence Files are sent to this 
relationship</td></tr><tr><td>failure</td><td>Incoming files that failed to 
generate a Sequence File are sent to this 
relationship</td></tr></table><h3>Reads Attributes: </h3>None 
specified.<h3>Writes Attributes: </h3>None specified.<h3>State management: 
</h3>This component does not store state.<h3>Restricted: </h3>This component is 
not restricted.<h3>Input requirement: </h3>This component requires an incoming 
relationship.<h3>System Resource Considerations:</h3>None specified.<h3>See 
Also:</h3><p><a 
href="../org.apache.nifi.processors.hadoop.PutHDFS/index.html">PutHDFS</a></p></body></html>
\ No newline at end of file

Added: 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi/nifi-hadoop-nar/1.9.0/org.apache.nifi.processors.hadoop.DeleteHDFS/index.html
URL: 
http://svn.apache.org/viewvc/nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi/nifi-hadoop-nar/1.9.0/org.apache.nifi.processors.hadoop.DeleteHDFS/index.html?rev=1854109&view=auto
==============================================================================
--- 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi/nifi-hadoop-nar/1.9.0/org.apache.nifi.processors.hadoop.DeleteHDFS/index.html
 (added)
+++ 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi/nifi-hadoop-nar/1.9.0/org.apache.nifi.processors.hadoop.DeleteHDFS/index.html
 Fri Feb 22 01:03:44 2019
@@ -0,0 +1,3 @@
+<!DOCTYPE html><html lang="en"><head><meta 
charset="utf-8"></meta><title>DeleteHDFS</title><link rel="stylesheet" 
href="../../../../../css/component-usage.css" 
type="text/css"></link></head><script type="text/javascript">window.onload = 
function(){if(self==top) { document.getElementById('nameHeader').style.display 
= "inherit"; } }</script><body><h1 id="nameHeader" style="display: 
none;">DeleteHDFS</h1><h2>Description: </h2><p>Deletes one or more files or 
directories from HDFS. The path can be provided as an attribute from an 
incoming FlowFile, or a statically set path that is periodically removed. If 
this processor has an incoming connection, itwill ignore running on a periodic 
basis and instead rely on incoming FlowFiles to trigger a delete. Note that you 
may use a wildcard character to match multiple files or directories. If there 
are no incoming connections no flowfiles will be transfered to any output 
relationships.  If there is an incoming flowfile then provided there are no de
 tected failures it will be transferred to success otherwise it will be sent to 
false. If knowledge of globbed files deleted is necessary use ListHDFS first to 
produce a specific list of files to delete. </p><h3>Tags: </h3><p>hadoop, HDFS, 
delete, remove, filesystem</p><h3>Properties: </h3><p>In the list below, the 
names of required properties appear in <strong>bold</strong>. Any other 
properties (not in bold) are considered optional. The table also indicates any 
default values, and whether a property supports the <a 
href="../../../../../html/expression-language-guide.html">NiFi Expression 
Language</a>.</p><table id="properties"><tr><th>Name</th><th>Default 
Value</th><th>Allowable Values</th><th>Description</th></tr><tr><td 
id="name">Hadoop Configuration Resources</td><td id="default-value"></td><td 
id="allowable-values"></td><td id="description">A file or comma separated list 
of files which contains the Hadoop file system configuration. Without this, 
Hadoop will search the classpath
  for a 'core-site.xml' and 'hdfs-site.xml' file or will revert to a default 
configuration. To use swebhdfs, see 'Additional Details' section of PutHDFS's 
documentation.<br/><strong>Supports Expression Language: true (will be 
evaluated using variable registry only)</strong></td></tr><tr><td 
id="name">Kerberos Credentials Service</td><td id="default-value"></td><td 
id="allowable-values"><strong>Controller Service API: 
</strong><br/>KerberosCredentialsService<br/><strong>Implementation: 
</strong><a 
href="../../../nifi-kerberos-credentials-service-nar/1.9.0/org.apache.nifi.kerberos.KeytabCredentialsService/index.html">KeytabCredentialsService</a></td><td
 id="description">Specifies the Kerberos Credentials Controller Service that 
should be used for authenticating with Kerberos</td></tr><tr><td 
id="name">Kerberos Principal</td><td id="default-value"></td><td 
id="allowable-values"></td><td id="description">Kerberos principal to 
authenticate as. Requires nifi.kerberos.krb5.file to be set in
  your nifi.properties<br/><strong>Supports Expression Language: true (will be 
evaluated using variable registry only)</strong></td></tr><tr><td 
id="name">Kerberos Keytab</td><td id="default-value"></td><td 
id="allowable-values"></td><td id="description">Kerberos keytab associated with 
the principal. Requires nifi.kerberos.krb5.file to be set in your 
nifi.properties<br/><strong>Supports Expression Language: true (will be 
evaluated using variable registry only)</strong></td></tr><tr><td 
id="name">Kerberos Relogin Period</td><td id="default-value">4 hours</td><td 
id="allowable-values"></td><td id="description">Period of time which should 
pass before attempting a kerberos relogin.
+
+This property has been deprecated, and has no effect on processing. Relogins 
now occur automatically.<br/><strong>Supports Expression Language: true (will 
be evaluated using variable registry only)</strong></td></tr><tr><td 
id="name">Additional Classpath Resources</td><td id="default-value"></td><td 
id="allowable-values"></td><td id="description">A comma-separated list of paths 
to files and/or directories that will be added to the classpath. When 
specifying a directory, all files with in the directory will be added to the 
classpath, but further sub-directories will not be included.</td></tr><tr><td 
id="name"><strong>Path</strong></td><td id="default-value"></td><td 
id="allowable-values"></td><td id="description">The HDFS file or directory to 
delete. A wildcard expression may be used to only delete certain 
files<br/><strong>Supports Expression Language: true (will be evaluated using 
flow file attributes and variable registry)</strong></td></tr><tr><td 
id="name"><strong>Recursive</str
 ong></td><td id="default-value">true</td><td 
id="allowable-values"><ul><li>true</li><li>false</li></ul></td><td 
id="description">Remove contents of a non-empty directory 
recursively</td></tr></table><h3>Relationships: </h3><table 
id="relationships"><tr><th>Name</th><th>Description</th></tr><tr><td>success</td><td>When
 an incoming flowfile is used then if there are no errors invoking delete the 
flowfile will route here.</td></tr><tr><td>failure</td><td>When an incoming 
flowfile is used and there is a failure while deleting then the flowfile will 
route here.</td></tr></table><h3>Reads Attributes: </h3>None 
specified.<h3>Writes Attributes: </h3><table 
id="writes-attributes"><tr><th>Name</th><th>Description</th></tr><tr><td>hdfs.filename</td><td>HDFS
 file to be deleted. If multiple files are deleted, then only the last filename 
is set.</td></tr><tr><td>hdfs.path</td><td>HDFS Path specified in the delete 
request. If multiple paths are deleted, then only the last path is 
set.</td></tr><tr
 ><td>hdfs.error.message</td><td>HDFS error message related to the 
 >hdfs.error.code</td></tr></table><h3>State management: </h3>This component 
 >does not store state.<h3>Restricted: </h3><table 
 >id="restrictions"><tr><th>Required 
 >Permission</th><th>Explanation</th></tr><tr><td>write 
 >filesystem</td><td>Provides operator the ability to delete any file that NiFi 
 >has access to in HDFS or the local filesystem.</td></tr></table><h3>Input 
 >requirement: </h3>This component allows an incoming relationship.<h3>System 
 >Resource Considerations:</h3>None specified.<h3>See Also:</h3><p><a 
 >href="../org.apache.nifi.processors.hadoop.ListHDFS/index.html">ListHDFS</a>, 
 ><a 
 >href="../org.apache.nifi.processors.hadoop.PutHDFS/index.html">PutHDFS</a></p></body></html>
\ No newline at end of file

Added: 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi/nifi-hadoop-nar/1.9.0/org.apache.nifi.processors.hadoop.FetchHDFS/index.html
URL: 
http://svn.apache.org/viewvc/nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi/nifi-hadoop-nar/1.9.0/org.apache.nifi.processors.hadoop.FetchHDFS/index.html?rev=1854109&view=auto
==============================================================================
--- 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi/nifi-hadoop-nar/1.9.0/org.apache.nifi.processors.hadoop.FetchHDFS/index.html
 (added)
+++ 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi/nifi-hadoop-nar/1.9.0/org.apache.nifi.processors.hadoop.FetchHDFS/index.html
 Fri Feb 22 01:03:44 2019
@@ -0,0 +1,3 @@
+<!DOCTYPE html><html lang="en"><head><meta 
charset="utf-8"></meta><title>FetchHDFS</title><link rel="stylesheet" 
href="../../../../../css/component-usage.css" 
type="text/css"></link></head><script type="text/javascript">window.onload = 
function(){if(self==top) { document.getElementById('nameHeader').style.display 
= "inherit"; } }</script><body><h1 id="nameHeader" style="display: 
none;">FetchHDFS</h1><h2>Description: </h2><p>Retrieves a file from HDFS. The 
content of the incoming FlowFile is replaced by the content of the file in 
HDFS. The file in HDFS is left intact without any changes being made to 
it.</p><h3>Tags: </h3><p>hadoop, hdfs, get, ingest, fetch, 
source</p><h3>Properties: </h3><p>In the list below, the names of required 
properties appear in <strong>bold</strong>. Any other properties (not in bold) 
are considered optional. The table also indicates any default values, and 
whether a property supports the <a 
href="../../../../../html/expression-language-guide.html">NiFi Expre
 ssion Language</a>.</p><table id="properties"><tr><th>Name</th><th>Default 
Value</th><th>Allowable Values</th><th>Description</th></tr><tr><td 
id="name">Hadoop Configuration Resources</td><td id="default-value"></td><td 
id="allowable-values"></td><td id="description">A file or comma separated list 
of files which contains the Hadoop file system configuration. Without this, 
Hadoop will search the classpath for a 'core-site.xml' and 'hdfs-site.xml' file 
or will revert to a default configuration. To use swebhdfs, see 'Additional 
Details' section of PutHDFS's documentation.<br/><strong>Supports Expression 
Language: true (will be evaluated using variable registry 
only)</strong></td></tr><tr><td id="name">Kerberos Credentials Service</td><td 
id="default-value"></td><td id="allowable-values"><strong>Controller Service 
API: </strong><br/>KerberosCredentialsService<br/><strong>Implementation: 
</strong><a 
href="../../../nifi-kerberos-credentials-service-nar/1.9.0/org.apache.nifi.kerberos.Keyta
 bCredentialsService/index.html">KeytabCredentialsService</a></td><td 
id="description">Specifies the Kerberos Credentials Controller Service that 
should be used for authenticating with Kerberos</td></tr><tr><td 
id="name">Kerberos Principal</td><td id="default-value"></td><td 
id="allowable-values"></td><td id="description">Kerberos principal to 
authenticate as. Requires nifi.kerberos.krb5.file to be set in your 
nifi.properties<br/><strong>Supports Expression Language: true (will be 
evaluated using variable registry only)</strong></td></tr><tr><td 
id="name">Kerberos Keytab</td><td id="default-value"></td><td 
id="allowable-values"></td><td id="description">Kerberos keytab associated with 
the principal. Requires nifi.kerberos.krb5.file to be set in your 
nifi.properties<br/><strong>Supports Expression Language: true (will be 
evaluated using variable registry only)</strong></td></tr><tr><td 
id="name">Kerberos Relogin Period</td><td id="default-value">4 hours</td><td 
id="allowable-values"><
 /td><td id="description">Period of time which should pass before attempting a 
kerberos relogin.
+
+This property has been deprecated, and has no effect on processing. Relogins 
now occur automatically.<br/><strong>Supports Expression Language: true (will 
be evaluated using variable registry only)</strong></td></tr><tr><td 
id="name">Additional Classpath Resources</td><td id="default-value"></td><td 
id="allowable-values"></td><td id="description">A comma-separated list of paths 
to files and/or directories that will be added to the classpath. When 
specifying a directory, all files with in the directory will be added to the 
classpath, but further sub-directories will not be included.</td></tr><tr><td 
id="name"><strong>HDFS Filename</strong></td><td 
id="default-value">${path}/${filename}</td><td id="allowable-values"></td><td 
id="description">The name of the HDFS file to retrieve<br/><strong>Supports 
Expression Language: true (will be evaluated using flow file attributes and 
variable registry)</strong></td></tr><tr><td id="name"><strong>Compression 
codec</strong></td><td id="default-va
 lue">NONE</td><td id="allowable-values"><ul><li>NONE <img 
src="../../../../../html/images/iconInfo.png" alt="No compression" title="No 
compression"></img></li><li>DEFAULT <img 
src="../../../../../html/images/iconInfo.png" alt="Default ZLIB compression" 
title="Default ZLIB compression"></img></li><li>BZIP <img 
src="../../../../../html/images/iconInfo.png" alt="BZIP compression" 
title="BZIP compression"></img></li><li>GZIP <img 
src="../../../../../html/images/iconInfo.png" alt="GZIP compression" 
title="GZIP compression"></img></li><li>LZ4 <img 
src="../../../../../html/images/iconInfo.png" alt="LZ4 compression" title="LZ4 
compression"></img></li><li>LZO <img 
src="../../../../../html/images/iconInfo.png" alt="LZO compression - it assumes 
LD_LIBRARY_PATH has been set and jar is available" title="LZO compression - it 
assumes LD_LIBRARY_PATH has been set and jar is 
available"></img></li><li>SNAPPY <img 
src="../../../../../html/images/iconInfo.png" alt="Snappy compression" 
title="Snappy com
 pression"></img></li><li>AUTOMATIC <img 
src="../../../../../html/images/iconInfo.png" alt="Will attempt to 
automatically detect the compression codec." title="Will attempt to 
automatically detect the compression codec."></img></li></ul></td><td 
id="description">No Description Provided.</td></tr></table><h3>Relationships: 
</h3><table 
id="relationships"><tr><th>Name</th><th>Description</th></tr><tr><td>success</td><td>FlowFiles
 will be routed to this relationship once they have been updated with the 
content of the HDFS file</td></tr><tr><td>comms.failure</td><td>FlowFiles will 
be routed to this relationship if the content of the HDFS file cannot be 
retrieve due to a communications failure. This generally indicates that the 
Fetch should be tried again.</td></tr><tr><td>failure</td><td>FlowFiles will be 
routed to this relationship if the content of the HDFS file cannot be retrieved 
and trying again will likely not be helpful. This would occur, for instance, if 
the file is not found or i
 f there is a permissions issue</td></tr></table><h3>Reads Attributes: 
</h3>None specified.<h3>Writes Attributes: </h3><table 
id="writes-attributes"><tr><th>Name</th><th>Description</th></tr><tr><td>hdfs.failure.reason</td><td>When
 a FlowFile is routed to 'failure', this attribute is added indicating why the 
file could not be fetched from HDFS</td></tr></table><h3>State management: 
</h3>This component does not store state.<h3>Restricted: </h3><table 
id="restrictions"><tr><th>Required 
Permission</th><th>Explanation</th></tr><tr><td>read 
filesystem</td><td>Provides operator the ability to retrieve any file that NiFi 
has access to in HDFS or the local filesystem.</td></tr></table><h3>Input 
requirement: </h3>This component requires an incoming relationship.<h3>System 
Resource Considerations:</h3>None specified.<h3>See Also:</h3><p><a 
href="../org.apache.nifi.processors.hadoop.ListHDFS/index.html">ListHDFS</a>, 
<a href="../org.apache.nifi.processors.hadoop.GetHDFS/index.html">GetHDFS</a>,
  <a 
href="../org.apache.nifi.processors.hadoop.PutHDFS/index.html">PutHDFS</a></p></body></html>
\ No newline at end of file

Added: 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi/nifi-hadoop-nar/1.9.0/org.apache.nifi.processors.hadoop.GetHDFS/index.html
URL: 
http://svn.apache.org/viewvc/nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi/nifi-hadoop-nar/1.9.0/org.apache.nifi.processors.hadoop.GetHDFS/index.html?rev=1854109&view=auto
==============================================================================
--- 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi/nifi-hadoop-nar/1.9.0/org.apache.nifi.processors.hadoop.GetHDFS/index.html
 (added)
+++ 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi/nifi-hadoop-nar/1.9.0/org.apache.nifi.processors.hadoop.GetHDFS/index.html
 Fri Feb 22 01:03:44 2019
@@ -0,0 +1,3 @@
+<!DOCTYPE html><html lang="en"><head><meta 
charset="utf-8"></meta><title>GetHDFS</title><link rel="stylesheet" 
href="../../../../../css/component-usage.css" 
type="text/css"></link></head><script type="text/javascript">window.onload = 
function(){if(self==top) { document.getElementById('nameHeader').style.display 
= "inherit"; } }</script><body><h1 id="nameHeader" style="display: 
none;">GetHDFS</h1><h2>Description: </h2><p>Fetch files from Hadoop Distributed 
File System (HDFS) into FlowFiles. This Processor will delete the file from 
HDFS after fetching it.</p><h3>Tags: </h3><p>hadoop, HDFS, get, fetch, ingest, 
source, filesystem</p><h3>Properties: </h3><p>In the list below, the names of 
required properties appear in <strong>bold</strong>. Any other properties (not 
in bold) are considered optional. The table also indicates any default values, 
and whether a property supports the <a 
href="../../../../../html/expression-language-guide.html">NiFi Expression 
Language</a>.</p><table id="prope
 rties"><tr><th>Name</th><th>Default Value</th><th>Allowable 
Values</th><th>Description</th></tr><tr><td id="name">Hadoop Configuration 
Resources</td><td id="default-value"></td><td id="allowable-values"></td><td 
id="description">A file or comma separated list of files which contains the 
Hadoop file system configuration. Without this, Hadoop will search the 
classpath for a 'core-site.xml' and 'hdfs-site.xml' file or will revert to a 
default configuration. To use swebhdfs, see 'Additional Details' section of 
PutHDFS's documentation.<br/><strong>Supports Expression Language: true (will 
be evaluated using variable registry only)</strong></td></tr><tr><td 
id="name">Kerberos Credentials Service</td><td id="default-value"></td><td 
id="allowable-values"><strong>Controller Service API: 
</strong><br/>KerberosCredentialsService<br/><strong>Implementation: 
</strong><a 
href="../../../nifi-kerberos-credentials-service-nar/1.9.0/org.apache.nifi.kerberos.KeytabCredentialsService/index.html">KeytabC
 redentialsService</a></td><td id="description">Specifies the Kerberos 
Credentials Controller Service that should be used for authenticating with 
Kerberos</td></tr><tr><td id="name">Kerberos Principal</td><td 
id="default-value"></td><td id="allowable-values"></td><td 
id="description">Kerberos principal to authenticate as. Requires 
nifi.kerberos.krb5.file to be set in your nifi.properties<br/><strong>Supports 
Expression Language: true (will be evaluated using variable registry 
only)</strong></td></tr><tr><td id="name">Kerberos Keytab</td><td 
id="default-value"></td><td id="allowable-values"></td><td 
id="description">Kerberos keytab associated with the principal. Requires 
nifi.kerberos.krb5.file to be set in your nifi.properties<br/><strong>Supports 
Expression Language: true (will be evaluated using variable registry 
only)</strong></td></tr><tr><td id="name">Kerberos Relogin Period</td><td 
id="default-value">4 hours</td><td id="allowable-values"></td><td 
id="description">Period of time
  which should pass before attempting a kerberos relogin.
+
+This property has been deprecated, and has no effect on processing. Relogins 
now occur automatically.<br/><strong>Supports Expression Language: true (will 
be evaluated using variable registry only)</strong></td></tr><tr><td 
id="name">Additional Classpath Resources</td><td id="default-value"></td><td 
id="allowable-values"></td><td id="description">A comma-separated list of paths 
to files and/or directories that will be added to the classpath. When 
specifying a directory, all files with in the directory will be added to the 
classpath, but further sub-directories will not be included.</td></tr><tr><td 
id="name"><strong>Directory</strong></td><td id="default-value"></td><td 
id="allowable-values"></td><td id="description">The HDFS directory from which 
files should be read<br/><strong>Supports Expression Language: true (will be 
evaluated using variable registry only)</strong></td></tr><tr><td 
id="name"><strong>Recurse Subdirectories</strong></td><td 
id="default-value">true</td><td id="all
 owable-values"><ul><li>true</li><li>false</li></ul></td><td 
id="description">Indicates whether to pull files from subdirectories of the 
HDFS directory</td></tr><tr><td id="name"><strong>Keep Source 
File</strong></td><td id="default-value">false</td><td 
id="allowable-values"><ul><li>true</li><li>false</li></ul></td><td 
id="description">Determines whether to delete the file from HDFS after it has 
been successfully transferred. If true, the file will be fetched repeatedly. 
This is intended for testing only.</td></tr><tr><td id="name">File Filter 
Regex</td><td id="default-value"></td><td id="allowable-values"></td><td 
id="description">A Java Regular Expression for filtering Filenames; if a filter 
is supplied then only files whose names match that Regular Expression will be 
fetched, otherwise all files will be fetched</td></tr><tr><td 
id="name"><strong>Filter Match Name Only</strong></td><td 
id="default-value">true</td><td 
id="allowable-values"><ul><li>true</li><li>false</li></ul></td><t
 d id="description">If true then File Filter Regex will match on just the 
filename, otherwise subdirectory names will be included with filename in the 
regex comparison</td></tr><tr><td id="name"><strong>Ignore Dotted 
Files</strong></td><td id="default-value">true</td><td 
id="allowable-values"><ul><li>true</li><li>false</li></ul></td><td 
id="description">If true, files whose names begin with a dot (".") will be 
ignored</td></tr><tr><td id="name"><strong>Minimum File Age</strong></td><td 
id="default-value">0 sec</td><td id="allowable-values"></td><td 
id="description">The minimum age that a file must be in order to be pulled; any 
file younger than this amount of time (based on last modification date) will be 
ignored</td></tr><tr><td id="name">Maximum File Age</td><td 
id="default-value"></td><td id="allowable-values"></td><td id="description">The 
maximum age that a file must be in order to be pulled; any file older than this 
amount of time (based on last modification date) will be ignore
 d</td></tr><tr><td id="name"><strong>Polling Interval</strong></td><td 
id="default-value">0 sec</td><td id="allowable-values"></td><td 
id="description">Indicates how long to wait between performing directory 
listings</td></tr><tr><td id="name"><strong>Batch Size</strong></td><td 
id="default-value">100</td><td id="allowable-values"></td><td 
id="description">The maximum number of files to pull in each iteration, based 
on run schedule.</td></tr><tr><td id="name">IO Buffer Size</td><td 
id="default-value"></td><td id="allowable-values"></td><td 
id="description">Amount of memory to use to buffer file contents during IO. 
This overrides the Hadoop Configuration</td></tr><tr><td 
id="name"><strong>Compression codec</strong></td><td 
id="default-value">NONE</td><td id="allowable-values"><ul><li>NONE <img 
src="../../../../../html/images/iconInfo.png" alt="No compression" title="No 
compression"></img></li><li>DEFAULT <img 
src="../../../../../html/images/iconInfo.png" alt="Default ZLIB compression
 " title="Default ZLIB compression"></img></li><li>BZIP <img 
src="../../../../../html/images/iconInfo.png" alt="BZIP compression" 
title="BZIP compression"></img></li><li>GZIP <img 
src="../../../../../html/images/iconInfo.png" alt="GZIP compression" 
title="GZIP compression"></img></li><li>LZ4 <img 
src="../../../../../html/images/iconInfo.png" alt="LZ4 compression" title="LZ4 
compression"></img></li><li>LZO <img 
src="../../../../../html/images/iconInfo.png" alt="LZO compression - it assumes 
LD_LIBRARY_PATH has been set and jar is available" title="LZO compression - it 
assumes LD_LIBRARY_PATH has been set and jar is 
available"></img></li><li>SNAPPY <img 
src="../../../../../html/images/iconInfo.png" alt="Snappy compression" 
title="Snappy compression"></img></li><li>AUTOMATIC <img 
src="../../../../../html/images/iconInfo.png" alt="Will attempt to 
automatically detect the compression codec." title="Will attempt to 
automatically detect the compression codec."></img></li></ul></td><td id="de
 scription">No Description Provided.</td></tr></table><h3>Relationships: 
</h3><table 
id="relationships"><tr><th>Name</th><th>Description</th></tr><tr><td>success</td><td>All
 files retrieved from HDFS are transferred to this 
relationship</td></tr></table><h3>Reads Attributes: </h3>None 
specified.<h3>Writes Attributes: </h3><table 
id="writes-attributes"><tr><th>Name</th><th>Description</th></tr><tr><td>filename</td><td>The
 name of the file that was read from HDFS.</td></tr><tr><td>path</td><td>The 
path is set to the relative path of the file's directory on HDFS. For example, 
if the Directory property is set to /tmp, then files picked up from /tmp will 
have the path attribute set to "./". If the Recurse Subdirectories property is 
set to true and a file is picked up from /tmp/abc/1/2/3, then the path 
attribute will be set to "abc/1/2/3".</td></tr></table><h3>State management: 
</h3>This component does not store state.<h3>Restricted: </h3><table 
id="restrictions"><tr><th>Required Permissio
 n</th><th>Explanation</th></tr><tr><td>read filesystem</td><td>Provides 
operator the ability to retrieve any file that NiFi has access to in HDFS or 
the local filesystem.</td></tr><tr><td>write filesystem</td><td>Provides 
operator the ability to delete any file that NiFi has access to in HDFS or the 
local filesystem.</td></tr></table><h3>Input requirement: </h3>This component 
does not allow an incoming relationship.<h3>System Resource 
Considerations:</h3>None specified.<h3>See Also:</h3><p><a 
href="../org.apache.nifi.processors.hadoop.PutHDFS/index.html">PutHDFS</a>, <a 
href="../org.apache.nifi.processors.hadoop.ListHDFS/index.html">ListHDFS</a></p></body></html>
\ No newline at end of file

Added: 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi/nifi-hadoop-nar/1.9.0/org.apache.nifi.processors.hadoop.GetHDFSFileInfo/index.html
URL: 
http://svn.apache.org/viewvc/nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi/nifi-hadoop-nar/1.9.0/org.apache.nifi.processors.hadoop.GetHDFSFileInfo/index.html?rev=1854109&view=auto
==============================================================================
--- 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi/nifi-hadoop-nar/1.9.0/org.apache.nifi.processors.hadoop.GetHDFSFileInfo/index.html
 (added)
+++ 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi/nifi-hadoop-nar/1.9.0/org.apache.nifi.processors.hadoop.GetHDFSFileInfo/index.html
 Fri Feb 22 01:03:44 2019
@@ -0,0 +1,3 @@
+<!DOCTYPE html><html lang="en"><head><meta 
charset="utf-8"></meta><title>GetHDFSFileInfo</title><link rel="stylesheet" 
href="../../../../../css/component-usage.css" 
type="text/css"></link></head><script type="text/javascript">window.onload = 
function(){if(self==top) { document.getElementById('nameHeader').style.display 
= "inherit"; } }</script><body><h1 id="nameHeader" style="display: 
none;">GetHDFSFileInfo</h1><h2>Description: </h2><p>Retrieves a listing of 
files and directories from HDFS. This processor creates a FlowFile(s) that 
represents the HDFS file/dir with relevant information. Main purpose of this 
processor to provide functionality similar to HDFS Client, i.e. count, du, ls, 
test, etc. Unlike ListHDFS, this processor is stateless, supports incoming 
connections and provides information on a dir level. </p><h3>Tags: 
</h3><p>hadoop, HDFS, get, list, ingest, source, filesystem</p><h3>Properties: 
</h3><p>In the list below, the names of required properties appear in 
<strong>bold
 </strong>. Any other properties (not in bold) are considered optional. The 
table also indicates any default values, and whether a property supports the <a 
href="../../../../../html/expression-language-guide.html">NiFi Expression 
Language</a>.</p><table id="properties"><tr><th>Name</th><th>Default 
Value</th><th>Allowable Values</th><th>Description</th></tr><tr><td 
id="name">Hadoop Configuration Resources</td><td id="default-value"></td><td 
id="allowable-values"></td><td id="description">A file or comma separated list 
of files which contains the Hadoop file system configuration. Without this, 
Hadoop will search the classpath for a 'core-site.xml' and 'hdfs-site.xml' file 
or will revert to a default configuration. To use swebhdfs, see 'Additional 
Details' section of PutHDFS's documentation.<br/><strong>Supports Expression 
Language: true (will be evaluated using variable registry 
only)</strong></td></tr><tr><td id="name">Kerberos Credentials Service</td><td 
id="default-value"></td><td i
 d="allowable-values"><strong>Controller Service API: 
</strong><br/>KerberosCredentialsService<br/><strong>Implementation: 
</strong><a 
href="../../../nifi-kerberos-credentials-service-nar/1.9.0/org.apache.nifi.kerberos.KeytabCredentialsService/index.html">KeytabCredentialsService</a></td><td
 id="description">Specifies the Kerberos Credentials Controller Service that 
should be used for authenticating with Kerberos</td></tr><tr><td 
id="name">Kerberos Principal</td><td id="default-value"></td><td 
id="allowable-values"></td><td id="description">Kerberos principal to 
authenticate as. Requires nifi.kerberos.krb5.file to be set in your 
nifi.properties<br/><strong>Supports Expression Language: true (will be 
evaluated using variable registry only)</strong></td></tr><tr><td 
id="name">Kerberos Keytab</td><td id="default-value"></td><td 
id="allowable-values"></td><td id="description">Kerberos keytab associated with 
the principal. Requires nifi.kerberos.krb5.file to be set in your 
nifi.properties
 <br/><strong>Supports Expression Language: true (will be evaluated using 
variable registry only)</strong></td></tr><tr><td id="name">Kerberos Relogin 
Period</td><td id="default-value">4 hours</td><td 
id="allowable-values"></td><td id="description">Period of time which should 
pass before attempting a kerberos relogin.
+
+This property has been deprecated, and has no effect on processing. Relogins 
now occur automatically.<br/><strong>Supports Expression Language: true (will 
be evaluated using variable registry only)</strong></td></tr><tr><td 
id="name">Additional Classpath Resources</td><td id="default-value"></td><td 
id="allowable-values"></td><td id="description">A comma-separated list of paths 
to files and/or directories that will be added to the classpath. When 
specifying a directory, all files with in the directory will be added to the 
classpath, but further sub-directories will not be included.</td></tr><tr><td 
id="name"><strong>Full path</strong></td><td id="default-value"></td><td 
id="allowable-values"></td><td id="description">A directory to start listing 
from, or a file's full path.<br/><strong>Supports Expression Language: true 
(will be evaluated using flow file attributes and variable 
registry)</strong></td></tr><tr><td id="name"><strong>Recurse 
Subdirectories</strong></td><td id="default-
 value">true</td><td 
id="allowable-values"><ul><li>true</li><li>false</li></ul></td><td 
id="description">Indicates whether to list files from subdirectories of the 
HDFS directory</td></tr><tr><td id="name">Directory Filter</td><td 
id="default-value"></td><td id="allowable-values"></td><td 
id="description">Regex. Only directories whose names match the given regular 
expression will be picked up. If not provided, any filter would be apply 
(performance considerations).<br/><strong>Supports Expression Language: true 
(will be evaluated using flow file attributes and variable 
registry)</strong></td></tr><tr><td id="name">File Filter</td><td 
id="default-value"></td><td id="allowable-values"></td><td 
id="description">Regex. Only files whose names match the given regular 
expression will be picked up. If not provided, any filter would be apply 
(performance considerations).<br/><strong>Supports Expression Language: true 
(will be evaluated using flow file attributes and variable registry)</strong
 ></td></tr><tr><td id="name">Exclude Files</td><td id="default-value"></td><td 
 >id="allowable-values"></td><td id="description">Regex. Files whose names 
 >match the given regular expression will not be picked up. If not provided, 
 >any filter won't be apply (performance considerations).<br/><strong>Supports 
 >Expression Language: true (will be evaluated using flow file attributes and 
 >variable registry)</strong></td></tr><tr><td id="name"><strong>Ignore Dotted 
 >Directories</strong></td><td id="default-value">true</td><td 
 >id="allowable-values"><ul><li>true</li><li>false</li></ul></td><td 
 >id="description">If true, directories whose names begin with a dot (".") will 
 >be ignored</td></tr><tr><td id="name"><strong>Ignore Dotted 
 >Files</strong></td><td id="default-value">true</td><td 
 >id="allowable-values"><ul><li>true</li><li>false</li></ul></td><td 
 >id="description">If true, files whose names begin with a dot (".") will be 
 >ignored</td></tr><tr><td id="name"><strong>Group Results</strong></td><td id=
 "default-value">gethdfsfileinfo-group-all</td><td 
id="allowable-values"><ul><li>All <img 
src="../../../../../html/images/iconInfo.png" alt="Group all results into a 
single flowfile." title="Group all results into a single 
flowfile."></img></li><li>Parent Directory <img 
src="../../../../../html/images/iconInfo.png" alt="Group HDFS objects by their 
parent directories only. Processor will generate flowfile for each directory 
(if recursive). If 'Recurse Subdirectories' property set to 'false', then will 
have the same effect as 'All'" title="Group HDFS objects by their parent 
directories only. Processor will generate flowfile for each directory (if 
recursive). If 'Recurse Subdirectories' property set to 'false', then will have 
the same effect as 'All'"></img></li><li>None <img 
src="../../../../../html/images/iconInfo.png" alt="Don't group results. 
Generate flowfile per each HDFS object." title="Don't group results. Generate 
flowfile per each HDFS object."></img></li></ul></td><td id="des
 cription">Groups HDFS objects</td></tr><tr><td 
id="name"><strong>Destination</strong></td><td 
id="default-value">gethdfsfileinfo-dest-content</td><td 
id="allowable-values"><ul><li>Attributes <img 
src="../../../../../html/images/iconInfo.png" alt="Details of given HDFS object 
will be stored in attributes of flowfile. WARNING: In case when scan finds 
thousands or millions of objects, having huge values in attribute could impact 
flow file repo and GC/heap usage. Use content destination for such cases." 
title="Details of given HDFS object will be stored in attributes of flowfile. 
WARNING: In case when scan finds thousands or millions of objects, having huge 
values in attribute could impact flow file repo and GC/heap usage. Use content 
destination for such cases."></img></li><li>Content <img 
src="../../../../../html/images/iconInfo.png" alt="Details of given HDFS object 
will be stored in a content in JSON format" title="Details of given HDFS object 
will be stored in a content in JSON for
 mat"></img></li></ul></td><td id="description">Sets the destination for the 
resutls. When set to 'Content', attributes of flowfile won't be used for 
storing results. </td></tr></table><h3>Relationships: </h3><table 
id="relationships"><tr><th>Name</th><th>Description</th></tr><tr><td>success</td><td>All
 successfully generated FlowFiles are transferred to this 
relationship</td></tr><tr><td>not found</td><td>If no objects are found, 
original FlowFile are transferred to this 
relationship</td></tr><tr><td>failure</td><td>All failed attempts to access 
HDFS will be routed to this 
relationship</td></tr><tr><td>original</td><td>Original FlowFiles are 
transferred to this relationship</td></tr></table><h3>Reads Attributes: 
</h3>None specified.<h3>Writes Attributes: </h3><table 
id="writes-attributes"><tr><th>Name</th><th>Description</th></tr><tr><td>hdfs.objectName</td><td>The
 name of the file/dir found on HDFS.</td></tr><tr><td>hdfs.path</td><td>The 
path is set to the absolute path of the obje
 ct's parent directory on HDFS. For example, if an object is a directory 'foo', 
under directory '/bar' then 'hdfs.objectName' will have value 'foo', and 
'hdfs.path' will be '/bar'</td></tr><tr><td>hdfs.type</td><td>The type of an 
object. Possible values: directory, file, 
link</td></tr><tr><td>hdfs.owner</td><td>The user that owns the object in 
HDFS</td></tr><tr><td>hdfs.group</td><td>The group that owns the object in 
HDFS</td></tr><tr><td>hdfs.lastModified</td><td>The timestamp of when the 
object in HDFS was last modified, as milliseconds since midnight Jan 1, 1970 
UTC</td></tr><tr><td>hdfs.length</td><td>In case of files: The number of bytes 
in the file in HDFS.  In case of dirs: Retuns storage space consumed by 
directory. </td></tr><tr><td>hdfs.count.files</td><td>In case of 
type='directory' will represent total count of files under this dir. Won't be 
populated to other types of HDFS objects. 
</td></tr><tr><td>hdfs.count.dirs</td><td>In case of type='directory' will 
represent total
  count of directories under this dir (including itself). Won't be populated to 
other types of HDFS objects. </td></tr><tr><td>hdfs.replication</td><td>The 
number of HDFS replicas for the 
file</td></tr><tr><td>hdfs.permissions</td><td>The permissions for the object 
in HDFS. This is formatted as 3 characters for the owner, 3 for the group, and 
3 for other users. For example 
rw-rw-r--</td></tr><tr><td>hdfs.status</td><td>The status contains comma 
separated list of file/dir paths, which couldn't be listed/accessed. Status 
won't be set if no errors occured.</td></tr><tr><td>hdfs.full.tree</td><td>When 
destination is 'attribute', will be populated with full tree of HDFS directory 
in JSON format.WARNING: In case when scan finds thousands or millions of 
objects, having huge values in attribute could impact flow file repo and 
GC/heap usage. Use content destination for such 
cases</td></tr></table><h3>State management: </h3>This component does not store 
state.<h3>Restricted: </h3>This componen
 t is not restricted.<h3>Input requirement: </h3>This component allows an 
incoming relationship.<h3>System Resource Considerations:</h3>None 
specified.<h3>See Also:</h3><p><a 
href="../org.apache.nifi.processors.hadoop.ListHDFS/index.html">ListHDFS</a>, 
<a href="../org.apache.nifi.processors.hadoop.GetHDFS/index.html">GetHDFS</a>, 
<a 
href="../org.apache.nifi.processors.hadoop.FetchHDFS/index.html">FetchHDFS</a>, 
<a 
href="../org.apache.nifi.processors.hadoop.PutHDFS/index.html">PutHDFS</a></p></body></html>
\ No newline at end of file


Reply via email to