Modified: 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.GetHDFS/index.html
URL: 
http://svn.apache.org/viewvc/nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.GetHDFS/index.html?rev=1771896&r1=1771895&r2=1771896&view=diff
==============================================================================
--- 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.GetHDFS/index.html
 (original)
+++ 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.GetHDFS/index.html
 Tue Nov 29 12:33:21 2016
@@ -1 +1 @@
-<!DOCTYPE html><html lang="en"><head><meta 
charset="utf-8"></meta><title>GetHDFS</title><link rel="stylesheet" 
href="../../css/component-usage.css" 
type="text/css"></link></head><body><h2>Description: </h2><p>Fetch files from 
Hadoop Distributed File System (HDFS) into FlowFiles. This Processor will 
delete the file from HDFS after fetching it.</p><h3>Tags: </h3><p>hadoop, HDFS, 
get, fetch, ingest, source, filesystem</p><h3>Properties: </h3><p>In the list 
below, the names of required properties appear in <strong>bold</strong>. Any 
other properties (not in bold) are considered optional. The table also 
indicates any default values, and whether a property supports the <a 
href="../../html/expression-language-guide.html">NiFi Expression 
Language</a>.</p><table id="properties"><tr><th>Name</th><th>Default 
Value</th><th>Allowable Values</th><th>Description</th></tr><tr><td 
id="name">Hadoop Configuration Resources</td><td id="default-value"></td><td 
id="allowable-values"></td><td id="descript
 ion">A file or comma separated list of files which contains the Hadoop file 
system configuration. Without this, Hadoop will search the classpath for a 
'core-site.xml' and 'hdfs-site.xml' file or will revert to a default 
configuration.</td></tr><tr><td id="name">Kerberos Principal</td><td 
id="default-value"></td><td id="allowable-values"></td><td 
id="description">Kerberos principal to authenticate as. Requires 
nifi.kerberos.krb5.file to be set in your nifi.properties</td></tr><tr><td 
id="name">Kerberos Keytab</td><td id="default-value"></td><td 
id="allowable-values"></td><td id="description">Kerberos keytab associated with 
the principal. Requires nifi.kerberos.krb5.file to be set in your 
nifi.properties</td></tr><tr><td id="name">Kerberos Relogin Period</td><td 
id="default-value">4 hours</td><td id="allowable-values"></td><td 
id="description">Period of time which should pass before attempting a kerberos 
relogin</td></tr><tr><td id="name"><strong>Directory</strong></td><td 
id="default
 -value"></td><td id="allowable-values"></td><td id="description">The HDFS 
directory from which files should be read<br/><strong>Supports Expression 
Language: true</strong></td></tr><tr><td id="name"><strong>Recurse 
Subdirectories</strong></td><td id="default-value">true</td><td 
id="allowable-values"><ul><li>true</li><li>false</li></ul></td><td 
id="description">Indicates whether to pull files from subdirectories of the 
HDFS directory</td></tr><tr><td id="name"><strong>Keep Source 
File</strong></td><td id="default-value">false</td><td 
id="allowable-values"><ul><li>true</li><li>false</li></ul></td><td 
id="description">Determines whether to delete the file from HDFS after it has 
been successfully transferred. If true, the file will be fetched repeatedly. 
This is intended for testing only.</td></tr><tr><td id="name">File Filter 
Regex</td><td id="default-value"></td><td id="allowable-values"></td><td 
id="description">A Java Regular Expression for filtering Filenames; if a filter 
is suppli
 ed then only files whose names match that Regular Expression will be fetched, 
otherwise all files will be fetched</td></tr><tr><td id="name"><strong>Filter 
Match Name Only</strong></td><td id="default-value">true</td><td 
id="allowable-values"><ul><li>true</li><li>false</li></ul></td><td 
id="description">If true then File Filter Regex will match on just the 
filename, otherwise subdirectory names will be included with filename in the 
regex comparison</td></tr><tr><td id="name"><strong>Ignore Dotted 
Files</strong></td><td id="default-value">true</td><td 
id="allowable-values"><ul><li>true</li><li>false</li></ul></td><td 
id="description">If true, files whose names begin with a dot (".") will be 
ignored</td></tr><tr><td id="name"><strong>Minimum File Age</strong></td><td 
id="default-value">0 sec</td><td id="allowable-values"></td><td 
id="description">The minimum age that a file must be in order to be pulled; any 
file younger than this amount of time (based on last modification date) will 
 be ignored</td></tr><tr><td id="name">Maximum File Age</td><td 
id="default-value"></td><td id="allowable-values"></td><td id="description">The 
maximum age that a file must be in order to be pulled; any file older than this 
amount of time (based on last modification date) will be 
ignored</td></tr><tr><td id="name"><strong>Polling Interval</strong></td><td 
id="default-value">0 sec</td><td id="allowable-values"></td><td 
id="description">Indicates how long to wait between performing directory 
listings</td></tr><tr><td id="name"><strong>Batch Size</strong></td><td 
id="default-value">100</td><td id="allowable-values"></td><td 
id="description">The maximum number of files to pull in each iteration, based 
on run schedule.</td></tr><tr><td id="name">IO Buffer Size</td><td 
id="default-value"></td><td id="allowable-values"></td><td 
id="description">Amount of memory to use to buffer file contents during IO. 
This overrides the Hadoop Configuration</td></tr><tr><td 
id="name"><strong>Compression co
 dec</strong></td><td id="default-value">NONE</td><td 
id="allowable-values"><ul><li>NONE</li><li>DEFAULT</li><li>BZIP</li><li>GZIP</li><li>LZ4</li><li>SNAPPY</li><li>AUTOMATIC</li></ul></td><td
 id="description">No Description Provided.</td></tr></table><h3>Relationships: 
</h3><table 
id="relationships"><tr><th>Name</th><th>Description</th></tr><tr><td>success</td><td>All
 files retrieved from HDFS are transferred to this 
relationship</td></tr></table><h3>Reads Attributes: </h3>None 
specified.<h3>Writes Attributes: </h3><table 
id="writes-attributes"><tr><th>Name</th><th>Description</th></tr><tr><td>filename</td><td>The
 name of the file that was read from HDFS.</td></tr><tr><td>path</td><td>The 
path is set to the relative path of the file's directory on HDFS. For example, 
if the Directory property is set to /tmp, then files picked up from /tmp will 
have the path attribute set to "./". If the Recurse Subdirectories property is 
set to true and a file is picked up from /tmp/abc/1/2/3, then 
 the path attribute will be set to "abc/1/2/3".</td></tr></table><h3>See 
Also:</h3><p><a 
href="../org.apache.nifi.processors.hadoop.PutHDFS/index.html">PutHDFS</a>, <a 
href="../org.apache.nifi.processors.hadoop.ListHDFS/index.html">ListHDFS</a></p></body></html>
\ No newline at end of file
+<!DOCTYPE html><html lang="en"><head><meta 
charset="utf-8"></meta><title>GetHDFS</title><link rel="stylesheet" 
href="../../css/component-usage.css" 
type="text/css"></link></head><body><h2>Description: </h2><p>Fetch files from 
Hadoop Distributed File System (HDFS) into FlowFiles. This Processor will 
delete the file from HDFS after fetching it.</p><h3>Tags: </h3><p>hadoop, HDFS, 
get, fetch, ingest, source, filesystem, restricted</p><h3>Properties: 
</h3><p>In the list below, the names of required properties appear in 
<strong>bold</strong>. Any other properties (not in bold) are considered 
optional. The table also indicates any default values, and whether a property 
supports the <a href="../../html/expression-language-guide.html">NiFi 
Expression Language</a>.</p><table id="properties"><tr><th>Name</th><th>Default 
Value</th><th>Allowable Values</th><th>Description</th></tr><tr><td 
id="name">Hadoop Configuration Resources</td><td id="default-value"></td><td 
id="allowable-values"></td><td 
 id="description">A file or comma separated list of files which contains the 
Hadoop file system configuration. Without this, Hadoop will search the 
classpath for a 'core-site.xml' and 'hdfs-site.xml' file or will revert to a 
default configuration.</td></tr><tr><td id="name">Kerberos Principal</td><td 
id="default-value"></td><td id="allowable-values"></td><td 
id="description">Kerberos principal to authenticate as. Requires 
nifi.kerberos.krb5.file to be set in your nifi.properties</td></tr><tr><td 
id="name">Kerberos Keytab</td><td id="default-value"></td><td 
id="allowable-values"></td><td id="description">Kerberos keytab associated with 
the principal. Requires nifi.kerberos.krb5.file to be set in your 
nifi.properties</td></tr><tr><td id="name">Kerberos Relogin Period</td><td 
id="default-value">4 hours</td><td id="allowable-values"></td><td 
id="description">Period of time which should pass before attempting a kerberos 
relogin</td></tr><tr><td id="name">Additional Classpath Resources</td
 ><td id="default-value"></td><td id="allowable-values"></td><td 
 >id="description">A comma-separated list of paths to files and/or directories 
 >that will be added to the classpath. When specifying a directory, all files 
 >with in the directory will be added to the classpath, but further 
 >sub-directories will not be included.</td></tr><tr><td 
 >id="name"><strong>Directory</strong></td><td id="default-value"></td><td 
 >id="allowable-values"></td><td id="description">The HDFS directory from which 
 >files should be read<br/><strong>Supports Expression Language: 
 >true</strong></td></tr><tr><td id="name"><strong>Recurse 
 >Subdirectories</strong></td><td id="default-value">true</td><td 
 >id="allowable-values"><ul><li>true</li><li>false</li></ul></td><td 
 >id="description">Indicates whether to pull files from subdirectories of the 
 >HDFS directory</td></tr><tr><td id="name"><strong>Keep Source 
 >File</strong></td><td id="default-value">false</td><td 
 >id="allowable-values"><ul><li>true</li><li>false</li></ul></td><
 td id="description">Determines whether to delete the file from HDFS after it 
has been successfully transferred. If true, the file will be fetched 
repeatedly. This is intended for testing only.</td></tr><tr><td id="name">File 
Filter Regex</td><td id="default-value"></td><td id="allowable-values"></td><td 
id="description">A Java Regular Expression for filtering Filenames; if a filter 
is supplied then only files whose names match that Regular Expression will be 
fetched, otherwise all files will be fetched</td></tr><tr><td 
id="name"><strong>Filter Match Name Only</strong></td><td 
id="default-value">true</td><td 
id="allowable-values"><ul><li>true</li><li>false</li></ul></td><td 
id="description">If true then File Filter Regex will match on just the 
filename, otherwise subdirectory names will be included with filename in the 
regex comparison</td></tr><tr><td id="name"><strong>Ignore Dotted 
Files</strong></td><td id="default-value">true</td><td 
id="allowable-values"><ul><li>true</li><li>fal
 se</li></ul></td><td id="description">If true, files whose names begin with a 
dot (".") will be ignored</td></tr><tr><td id="name"><strong>Minimum File 
Age</strong></td><td id="default-value">0 sec</td><td 
id="allowable-values"></td><td id="description">The minimum age that a file 
must be in order to be pulled; any file younger than this amount of time (based 
on last modification date) will be ignored</td></tr><tr><td id="name">Maximum 
File Age</td><td id="default-value"></td><td id="allowable-values"></td><td 
id="description">The maximum age that a file must be in order to be pulled; any 
file older than this amount of time (based on last modification date) will be 
ignored</td></tr><tr><td id="name"><strong>Polling Interval</strong></td><td 
id="default-value">0 sec</td><td id="allowable-values"></td><td 
id="description">Indicates how long to wait between performing directory 
listings</td></tr><tr><td id="name"><strong>Batch Size</strong></td><td 
id="default-value">100</td><td id="al
 lowable-values"></td><td id="description">The maximum number of files to pull 
in each iteration, based on run schedule.</td></tr><tr><td id="name">IO Buffer 
Size</td><td id="default-value"></td><td id="allowable-values"></td><td 
id="description">Amount of memory to use to buffer file contents during IO. 
This overrides the Hadoop Configuration</td></tr><tr><td 
id="name"><strong>Compression codec</strong></td><td 
id="default-value">NONE</td><td 
id="allowable-values"><ul><li>NONE</li><li>DEFAULT</li><li>BZIP</li><li>GZIP</li><li>LZ4</li><li>SNAPPY</li><li>AUTOMATIC</li></ul></td><td
 id="description">No Description Provided.</td></tr></table><h3>Relationships: 
</h3><table 
id="relationships"><tr><th>Name</th><th>Description</th></tr><tr><td>success</td><td>All
 files retrieved from HDFS are transferred to this 
relationship</td></tr></table><h3>Reads Attributes: </h3>None 
specified.<h3>Writes Attributes: </h3><table 
id="writes-attributes"><tr><th>Name</th><th>Description</th></tr><tr><td>f
 ilename</td><td>The name of the file that was read from 
HDFS.</td></tr><tr><td>path</td><td>The path is set to the relative path of the 
file's directory on HDFS. For example, if the Directory property is set to 
/tmp, then files picked up from /tmp will have the path attribute set to "./". 
If the Recurse Subdirectories property is set to true and a file is picked up 
from /tmp/abc/1/2/3, then the path attribute will be set to 
"abc/1/2/3".</td></tr></table><h3>State management: </h3>This processor has no 
state management.<h3>Restricted: </h3><td>Provides operator the ability to 
retrieve and delete any file that NiFi has access to in HDFS or the local 
filesystem.</td><h3>See Also:</h3><p><a 
href="../org.apache.nifi.processors.hadoop.PutHDFS/index.html">PutHDFS</a>, <a 
href="../org.apache.nifi.processors.hadoop.ListHDFS/index.html">ListHDFS</a></p></body></html>
\ No newline at end of file

Modified: 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.GetHDFSSequenceFile/index.html
URL: 
http://svn.apache.org/viewvc/nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.GetHDFSSequenceFile/index.html?rev=1771896&r1=1771895&r2=1771896&view=diff
==============================================================================
--- 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.GetHDFSSequenceFile/index.html
 (original)
+++ 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.GetHDFSSequenceFile/index.html
 Tue Nov 29 12:33:21 2016
@@ -1 +1 @@
-<!DOCTYPE html><html lang="en"><head><meta 
charset="utf-8"></meta><title>GetHDFSSequenceFile</title><link rel="stylesheet" 
href="../../css/component-usage.css" 
type="text/css"></link></head><body><h2>Description: </h2><p>Fetch sequence 
files from Hadoop Distributed File System (HDFS) into FlowFiles</p><h3>Tags: 
</h3><p>hadoop, HDFS, get, fetch, ingest, source, sequence 
file</p><h3>Properties: </h3><p>In the list below, the names of required 
properties appear in <strong>bold</strong>. Any other properties (not in bold) 
are considered optional. The table also indicates any default values, and 
whether a property supports the <a 
href="../../html/expression-language-guide.html">NiFi Expression 
Language</a>.</p><table id="properties"><tr><th>Name</th><th>Default 
Value</th><th>Allowable Values</th><th>Description</th></tr><tr><td 
id="name">Hadoop Configuration Resources</td><td id="default-value"></td><td 
id="allowable-values"></td><td id="description">A file or comma separated list 
of fil
 es which contains the Hadoop file system configuration. Without this, Hadoop 
will search the classpath for a 'core-site.xml' and 'hdfs-site.xml' file or 
will revert to a default configuration.</td></tr><tr><td id="name">Kerberos 
Principal</td><td id="default-value"></td><td id="allowable-values"></td><td 
id="description">Kerberos principal to authenticate as. Requires 
nifi.kerberos.krb5.file to be set in your nifi.properties</td></tr><tr><td 
id="name">Kerberos Keytab</td><td id="default-value"></td><td 
id="allowable-values"></td><td id="description">Kerberos keytab associated with 
the principal. Requires nifi.kerberos.krb5.file to be set in your 
nifi.properties</td></tr><tr><td id="name">Kerberos Relogin Period</td><td 
id="default-value">4 hours</td><td id="allowable-values"></td><td 
id="description">Period of time which should pass before attempting a kerberos 
relogin</td></tr><tr><td id="name"><strong>Directory</strong></td><td 
id="default-value"></td><td id="allowable-values"></t
 d><td id="description">The HDFS directory from which files should be 
read<br/><strong>Supports Expression Language: true</strong></td></tr><tr><td 
id="name"><strong>Recurse Subdirectories</strong></td><td 
id="default-value">true</td><td 
id="allowable-values"><ul><li>true</li><li>false</li></ul></td><td 
id="description">Indicates whether to pull files from subdirectories of the 
HDFS directory</td></tr><tr><td id="name"><strong>Keep Source 
File</strong></td><td id="default-value">false</td><td 
id="allowable-values"><ul><li>true</li><li>false</li></ul></td><td 
id="description">Determines whether to delete the file from HDFS after it has 
been successfully transferred. If true, the file will be fetched repeatedly. 
This is intended for testing only.</td></tr><tr><td id="name">File Filter 
Regex</td><td id="default-value"></td><td id="allowable-values"></td><td 
id="description">A Java Regular Expression for filtering Filenames; if a filter 
is supplied then only files whose names match that 
 Regular Expression will be fetched, otherwise all files will be 
fetched</td></tr><tr><td id="name"><strong>Filter Match Name 
Only</strong></td><td id="default-value">true</td><td 
id="allowable-values"><ul><li>true</li><li>false</li></ul></td><td 
id="description">If true then File Filter Regex will match on just the 
filename, otherwise subdirectory names will be included with filename in the 
regex comparison</td></tr><tr><td id="name"><strong>Ignore Dotted 
Files</strong></td><td id="default-value">true</td><td 
id="allowable-values"><ul><li>true</li><li>false</li></ul></td><td 
id="description">If true, files whose names begin with a dot (".") will be 
ignored</td></tr><tr><td id="name"><strong>Minimum File Age</strong></td><td 
id="default-value">0 sec</td><td id="allowable-values"></td><td 
id="description">The minimum age that a file must be in order to be pulled; any 
file younger than this amount of time (based on last modification date) will be 
ignored</td></tr><tr><td id="name">Maxi
 mum File Age</td><td id="default-value"></td><td 
id="allowable-values"></td><td id="description">The maximum age that a file 
must be in order to be pulled; any file older than this amount of time (based 
on last modification date) will be ignored</td></tr><tr><td 
id="name"><strong>Polling Interval</strong></td><td id="default-value">0 
sec</td><td id="allowable-values"></td><td id="description">Indicates how long 
to wait between performing directory listings</td></tr><tr><td 
id="name"><strong>Batch Size</strong></td><td id="default-value">100</td><td 
id="allowable-values"></td><td id="description">The maximum number of files to 
pull in each iteration, based on run schedule.</td></tr><tr><td id="name">IO 
Buffer Size</td><td id="default-value"></td><td id="allowable-values"></td><td 
id="description">Amount of memory to use to buffer file contents during IO. 
This overrides the Hadoop Configuration</td></tr><tr><td 
id="name"><strong>Compression codec</strong></td><td id="default-value">NO
 NE</td><td 
id="allowable-values"><ul><li>NONE</li><li>DEFAULT</li><li>BZIP</li><li>GZIP</li><li>LZ4</li><li>SNAPPY</li><li>AUTOMATIC</li></ul></td><td
 id="description">No Description Provided.</td></tr><tr><td 
id="name"><strong>FlowFile Content</strong></td><td id="default-value">VALUE 
ONLY</td><td id="allowable-values"><ul><li>VALUE ONLY</li><li>KEY VALUE 
PAIR</li></ul></td><td id="description">Indicate if the content is to be both 
the key and value of the Sequence File, or just the 
value.</td></tr></table><h3>Relationships: </h3><table 
id="relationships"><tr><th>Name</th><th>Description</th></tr><tr><td>success</td><td>All
 files retrieved from HDFS are transferred to this 
relationship</td></tr></table><h3>Reads Attributes: </h3>None 
specified.<h3>Writes Attributes: </h3><table 
id="writes-attributes"><tr><th>Name</th><th>Description</th></tr><tr><td>filename</td><td>The
 name of the file that was read from HDFS.</td></tr><tr><td>path</td><td>The 
path is set to the relative path of t
 he file's directory on HDFS. For example, if the Directory property is set to 
/tmp, then files picked up from /tmp will have the path attribute set to "./". 
If the Recurse Subdirectories property is set to true and a file is picked up 
from /tmp/abc/1/2/3, then the path attribute will be set to 
"abc/1/2/3".</td></tr></table><h3>See Also:</h3><p><a 
href="../org.apache.nifi.processors.hadoop.PutHDFS/index.html">PutHDFS</a></p></body></html>
\ No newline at end of file
+<!DOCTYPE html><html lang="en"><head><meta 
charset="utf-8"></meta><title>GetHDFSSequenceFile</title><link rel="stylesheet" 
href="../../css/component-usage.css" 
type="text/css"></link></head><body><h2>Description: </h2><p>Fetch sequence 
files from Hadoop Distributed File System (HDFS) into FlowFiles</p><h3>Tags: 
</h3><p>hadoop, HDFS, get, fetch, ingest, source, sequence 
file</p><h3>Properties: </h3><p>In the list below, the names of required 
properties appear in <strong>bold</strong>. Any other properties (not in bold) 
are considered optional. The table also indicates any default values, and 
whether a property supports the <a 
href="../../html/expression-language-guide.html">NiFi Expression 
Language</a>.</p><table id="properties"><tr><th>Name</th><th>Default 
Value</th><th>Allowable Values</th><th>Description</th></tr><tr><td 
id="name">Hadoop Configuration Resources</td><td id="default-value"></td><td 
id="allowable-values"></td><td id="description">A file or comma separated list 
of fil
 es which contains the Hadoop file system configuration. Without this, Hadoop 
will search the classpath for a 'core-site.xml' and 'hdfs-site.xml' file or 
will revert to a default configuration.</td></tr><tr><td id="name">Kerberos 
Principal</td><td id="default-value"></td><td id="allowable-values"></td><td 
id="description">Kerberos principal to authenticate as. Requires 
nifi.kerberos.krb5.file to be set in your nifi.properties</td></tr><tr><td 
id="name">Kerberos Keytab</td><td id="default-value"></td><td 
id="allowable-values"></td><td id="description">Kerberos keytab associated with 
the principal. Requires nifi.kerberos.krb5.file to be set in your 
nifi.properties</td></tr><tr><td id="name">Kerberos Relogin Period</td><td 
id="default-value">4 hours</td><td id="allowable-values"></td><td 
id="description">Period of time which should pass before attempting a kerberos 
relogin</td></tr><tr><td id="name">Additional Classpath Resources</td><td 
id="default-value"></td><td id="allowable-values"
 ></td><td id="description">A comma-separated list of paths to files and/or 
 >directories that will be added to the classpath. When specifying a directory, 
 >all files with in the directory will be added to the classpath, but further 
 >sub-directories will not be included.</td></tr><tr><td 
 >id="name"><strong>Directory</strong></td><td id="default-value"></td><td 
 >id="allowable-values"></td><td id="description">The HDFS directory from which 
 >files should be read<br/><strong>Supports Expression Language: 
 >true</strong></td></tr><tr><td id="name"><strong>Recurse 
 >Subdirectories</strong></td><td id="default-value">true</td><td 
 >id="allowable-values"><ul><li>true</li><li>false</li></ul></td><td 
 >id="description">Indicates whether to pull files from subdirectories of the 
 >HDFS directory</td></tr><tr><td id="name"><strong>Keep Source 
 >File</strong></td><td id="default-value">false</td><td 
 >id="allowable-values"><ul><li>true</li><li>false</li></ul></td><td 
 >id="description">Determines whether to delete the f
 ile from HDFS after it has been successfully transferred. If true, the file 
will be fetched repeatedly. This is intended for testing only.</td></tr><tr><td 
id="name">File Filter Regex</td><td id="default-value"></td><td 
id="allowable-values"></td><td id="description">A Java Regular Expression for 
filtering Filenames; if a filter is supplied then only files whose names match 
that Regular Expression will be fetched, otherwise all files will be 
fetched</td></tr><tr><td id="name"><strong>Filter Match Name 
Only</strong></td><td id="default-value">true</td><td 
id="allowable-values"><ul><li>true</li><li>false</li></ul></td><td 
id="description">If true then File Filter Regex will match on just the 
filename, otherwise subdirectory names will be included with filename in the 
regex comparison</td></tr><tr><td id="name"><strong>Ignore Dotted 
Files</strong></td><td id="default-value">true</td><td 
id="allowable-values"><ul><li>true</li><li>false</li></ul></td><td 
id="description">If true, files w
 hose names begin with a dot (".") will be ignored</td></tr><tr><td 
id="name"><strong>Minimum File Age</strong></td><td id="default-value">0 
sec</td><td id="allowable-values"></td><td id="description">The minimum age 
that a file must be in order to be pulled; any file younger than this amount of 
time (based on last modification date) will be ignored</td></tr><tr><td 
id="name">Maximum File Age</td><td id="default-value"></td><td 
id="allowable-values"></td><td id="description">The maximum age that a file 
must be in order to be pulled; any file older than this amount of time (based 
on last modification date) will be ignored</td></tr><tr><td 
id="name"><strong>Polling Interval</strong></td><td id="default-value">0 
sec</td><td id="allowable-values"></td><td id="description">Indicates how long 
to wait between performing directory listings</td></tr><tr><td 
id="name"><strong>Batch Size</strong></td><td id="default-value">100</td><td 
id="allowable-values"></td><td id="description">The maximum 
 number of files to pull in each iteration, based on run 
schedule.</td></tr><tr><td id="name">IO Buffer Size</td><td 
id="default-value"></td><td id="allowable-values"></td><td 
id="description">Amount of memory to use to buffer file contents during IO. 
This overrides the Hadoop Configuration</td></tr><tr><td 
id="name"><strong>Compression codec</strong></td><td 
id="default-value">NONE</td><td 
id="allowable-values"><ul><li>NONE</li><li>DEFAULT</li><li>BZIP</li><li>GZIP</li><li>LZ4</li><li>SNAPPY</li><li>AUTOMATIC</li></ul></td><td
 id="description">No Description Provided.</td></tr><tr><td 
id="name"><strong>FlowFile Content</strong></td><td id="default-value">VALUE 
ONLY</td><td id="allowable-values"><ul><li>VALUE ONLY</li><li>KEY VALUE 
PAIR</li></ul></td><td id="description">Indicate if the content is to be both 
the key and value of the Sequence File, or just the 
value.</td></tr></table><h3>Relationships: </h3><table 
id="relationships"><tr><th>Name</th><th>Description</th></tr><tr><td>su
 ccess</td><td>All files retrieved from HDFS are transferred to this 
relationship</td></tr></table><h3>Reads Attributes: </h3>None 
specified.<h3>Writes Attributes: </h3><table 
id="writes-attributes"><tr><th>Name</th><th>Description</th></tr><tr><td>filename</td><td>The
 name of the file that was read from HDFS.</td></tr><tr><td>path</td><td>The 
path is set to the relative path of the file's directory on HDFS. For example, 
if the Directory property is set to /tmp, then files picked up from /tmp will 
have the path attribute set to "./". If the Recurse Subdirectories property is 
set to true and a file is picked up from /tmp/abc/1/2/3, then the path 
attribute will be set to "abc/1/2/3".</td></tr></table><h3>State management: 
</h3>This processor has no state management.<h3>Restricted: </h3><td>Provides 
operator the ability to retrieve and delete any file that NiFi has access to in 
HDFS or the local filesystem.</td><h3>See Also:</h3><p><a 
href="../org.apache.nifi.processors.hadoop.PutHDFS/i
 ndex.html">PutHDFS</a></p></body></html>
\ No newline at end of file

Modified: 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.ListHDFS/index.html
URL: 
http://svn.apache.org/viewvc/nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.ListHDFS/index.html?rev=1771896&r1=1771895&r2=1771896&view=diff
==============================================================================
--- 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.ListHDFS/index.html
 (original)
+++ 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.ListHDFS/index.html
 Tue Nov 29 12:33:21 2016
@@ -1 +1 @@
-<!DOCTYPE html><html lang="en"><head><meta 
charset="utf-8"></meta><title>ListHDFS</title><link rel="stylesheet" 
href="../../css/component-usage.css" 
type="text/css"></link></head><body><h2>Description: </h2><p>Retrieves a 
listing of files from HDFS. For each file that is listed in HDFS, creates a 
FlowFile that represents the HDFS file so that it can be fetched in conjunction 
with ListHDFS. This Processor is designed to run on Primary Node only in a 
cluster. If the primary node changes, the new Primary Node will pick up where 
the previous node left off without duplicating all of the data. Unlike GetHDFS, 
this Processor does not delete any data from HDFS.</p><h3>Tags: </h3><p>hadoop, 
HDFS, get, list, ingest, source, filesystem</p><h3>Properties: </h3><p>In the 
list below, the names of required properties appear in <strong>bold</strong>. 
Any other properties (not in bold) are considered optional. The table also 
indicates any default values, and whether a property supports the <a href="
 ../../html/expression-language-guide.html">NiFi Expression 
Language</a>.</p><table id="properties"><tr><th>Name</th><th>Default 
Value</th><th>Allowable Values</th><th>Description</th></tr><tr><td 
id="name">Hadoop Configuration Resources</td><td id="default-value"></td><td 
id="allowable-values"></td><td id="description">A file or comma separated list 
of files which contains the Hadoop file system configuration. Without this, 
Hadoop will search the classpath for a 'core-site.xml' and 'hdfs-site.xml' file 
or will revert to a default configuration.</td></tr><tr><td id="name">Kerberos 
Principal</td><td id="default-value"></td><td id="allowable-values"></td><td 
id="description">Kerberos principal to authenticate as. Requires 
nifi.kerberos.krb5.file to be set in your nifi.properties</td></tr><tr><td 
id="name">Kerberos Keytab</td><td id="default-value"></td><td 
id="allowable-values"></td><td id="description">Kerberos keytab associated with 
the principal. Requires nifi.kerberos.krb5.file to 
 be set in your nifi.properties</td></tr><tr><td id="name">Kerberos Relogin 
Period</td><td id="default-value">4 hours</td><td 
id="allowable-values"></td><td id="description">Period of time which should 
pass before attempting a kerberos relogin</td></tr><tr><td 
id="name">Distributed Cache Service</td><td id="default-value"></td><td 
id="allowable-values"><strong>Controller Service API: 
</strong><br/>DistributedMapCacheClient<br/><strong>Implementation:</strong><br/><a
 
href="../org.apache.nifi.distributed.cache.client.DistributedMapCacheClientService/index.html">DistributedMapCacheClientService</a></td><td
 id="description">Specifies the Controller Service that should be used to 
maintain state about what has been pulled from HDFS so that if a new node 
begins pulling data, it won't duplicate all of the work that has been 
done.</td></tr><tr><td id="name"><strong>Directory</strong></td><td 
id="default-value"></td><td id="allowable-values"></td><td id="description">The 
HDFS directory from wh
 ich files should be read<br/><strong>Supports Expression Language: 
true</strong></td></tr><tr><td id="name"><strong>Recurse 
Subdirectories</strong></td><td id="default-value">true</td><td 
id="allowable-values"><ul><li>true</li><li>false</li></ul></td><td 
id="description">Indicates whether to list files from subdirectories of the 
HDFS directory</td></tr></table><h3>Relationships: </h3><table 
id="relationships"><tr><th>Name</th><th>Description</th></tr><tr><td>success</td><td>All
 FlowFiles are transferred to this relationship</td></tr></table><h3>Reads 
Attributes: </h3>None specified.<h3>Writes Attributes: </h3><table 
id="writes-attributes"><tr><th>Name</th><th>Description</th></tr><tr><td>filename</td><td>The
 name of the file that was read from HDFS.</td></tr><tr><td>path</td><td>The 
path is set to the absolute path of the file's directory on HDFS. For example, 
if the Directory property is set to /tmp, then files picked up from /tmp will 
have the path attribute set to "./". If the Re
 curse Subdirectories property is set to true and a file is picked up from 
/tmp/abc/1/2/3, then the path attribute will be set to 
"/tmp/abc/1/2/3".</td></tr><tr><td>hdfs.owner</td><td>The user that owns the 
file in HDFS</td></tr><tr><td>hdfs.group</td><td>The group that owns the file 
in HDFS</td></tr><tr><td>hdfs.lastModified</td><td>The timestamp of when the 
file in HDFS was last modified, as milliseconds since midnight Jan 1, 1970 
UTC</td></tr><tr><td>hdfs.length</td><td>The number of bytes in the file in 
HDFS</td></tr><tr><td>hdfs.replication</td><td>The number of HDFS replicas for 
hte file</td></tr><tr><td>hdfs.permissions</td><td>The permissions for the file 
in HDFS. This is formatted as 3 characters for the owner, 3 for the group, and 
3 for other users. For example rw-rw-r--</td></tr></table><h3>See 
Also:</h3><p><a 
href="../org.apache.nifi.processors.hadoop.GetHDFS/index.html">GetHDFS</a>, <a 
href="../org.apache.nifi.processors.hadoop.FetchHDFS/index.html">FetchHDFS</a>, 
<a hre
 
f="../org.apache.nifi.processors.hadoop.PutHDFS/index.html">PutHDFS</a></p></body></html>
\ No newline at end of file
+<!DOCTYPE html><html lang="en"><head><meta 
charset="utf-8"></meta><title>ListHDFS</title><link rel="stylesheet" 
href="../../css/component-usage.css" 
type="text/css"></link></head><body><h2>Description: </h2><p>Retrieves a 
listing of files from HDFS. For each file that is listed in HDFS, creates a 
FlowFile that represents the HDFS file so that it can be fetched in conjunction 
with ListHDFS. This Processor is designed to run on Primary Node only in a 
cluster. If the primary node changes, the new Primary Node will pick up where 
the previous node left off without duplicating all of the data. Unlike GetHDFS, 
this Processor does not delete any data from HDFS.</p><h3>Tags: </h3><p>hadoop, 
HDFS, get, list, ingest, source, filesystem</p><h3>Properties: </h3><p>In the 
list below, the names of required properties appear in <strong>bold</strong>. 
Any other properties (not in bold) are considered optional. The table also 
indicates any default values, and whether a property supports the <a href="
 ../../html/expression-language-guide.html">NiFi Expression 
Language</a>.</p><table id="properties"><tr><th>Name</th><th>Default 
Value</th><th>Allowable Values</th><th>Description</th></tr><tr><td 
id="name">Hadoop Configuration Resources</td><td id="default-value"></td><td 
id="allowable-values"></td><td id="description">A file or comma separated list 
of files which contains the Hadoop file system configuration. Without this, 
Hadoop will search the classpath for a 'core-site.xml' and 'hdfs-site.xml' file 
or will revert to a default configuration.</td></tr><tr><td id="name">Kerberos 
Principal</td><td id="default-value"></td><td id="allowable-values"></td><td 
id="description">Kerberos principal to authenticate as. Requires 
nifi.kerberos.krb5.file to be set in your nifi.properties</td></tr><tr><td 
id="name">Kerberos Keytab</td><td id="default-value"></td><td 
id="allowable-values"></td><td id="description">Kerberos keytab associated with 
the principal. Requires nifi.kerberos.krb5.file to 
 be set in your nifi.properties</td></tr><tr><td id="name">Kerberos Relogin 
Period</td><td id="default-value">4 hours</td><td 
id="allowable-values"></td><td id="description">Period of time which should 
pass before attempting a kerberos relogin</td></tr><tr><td id="name">Additional 
Classpath Resources</td><td id="default-value"></td><td 
id="allowable-values"></td><td id="description">A comma-separated list of paths 
to files and/or directories that will be added to the classpath. When 
specifying a directory, all files with in the directory will be added to the 
classpath, but further sub-directories will not be included.</td></tr><tr><td 
id="name">Distributed Cache Service</td><td id="default-value"></td><td 
id="allowable-values"><strong>Controller Service API: 
</strong><br/>DistributedMapCacheClient<br/><strong>Implementation:</strong><br/><a
 
href="../org.apache.nifi.distributed.cache.client.DistributedMapCacheClientService/index.html">DistributedMapCacheClientService</a></td><td
 id="d
 escription">Specifies the Controller Service that should be used to maintain 
state about what has been pulled from HDFS so that if a new node begins pulling 
data, it won't duplicate all of the work that has been done.</td></tr><tr><td 
id="name"><strong>Directory</strong></td><td id="default-value"></td><td 
id="allowable-values"></td><td id="description">The HDFS directory from which 
files should be read<br/><strong>Supports Expression Language: 
true</strong></td></tr><tr><td id="name"><strong>Recurse 
Subdirectories</strong></td><td id="default-value">true</td><td 
id="allowable-values"><ul><li>true</li><li>false</li></ul></td><td 
id="description">Indicates whether to list files from subdirectories of the 
HDFS directory</td></tr></table><h3>Relationships: </h3><table 
id="relationships"><tr><th>Name</th><th>Description</th></tr><tr><td>success</td><td>All
 FlowFiles are transferred to this relationship</td></tr></table><h3>Reads 
Attributes: </h3>None specified.<h3>Writes Attributes: </h
 3><table 
id="writes-attributes"><tr><th>Name</th><th>Description</th></tr><tr><td>filename</td><td>The
 name of the file that was read from HDFS.</td></tr><tr><td>path</td><td>The 
path is set to the absolute path of the file's directory on HDFS. For example, 
if the Directory property is set to /tmp, then files picked up from /tmp will 
have the path attribute set to "./". If the Recurse Subdirectories property is 
set to true and a file is picked up from /tmp/abc/1/2/3, then the path 
attribute will be set to 
"/tmp/abc/1/2/3".</td></tr><tr><td>hdfs.owner</td><td>The user that owns the 
file in HDFS</td></tr><tr><td>hdfs.group</td><td>The group that owns the file 
in HDFS</td></tr><tr><td>hdfs.lastModified</td><td>The timestamp of when the 
file in HDFS was last modified, as milliseconds since midnight Jan 1, 1970 
UTC</td></tr><tr><td>hdfs.length</td><td>The number of bytes in the file in 
HDFS</td></tr><tr><td>hdfs.replication</td><td>The number of HDFS replicas for 
hte file</td></tr><tr><t
 d>hdfs.permissions</td><td>The permissions for the file in HDFS. This is 
formatted as 3 characters for the owner, 3 for the group, and 3 for other 
users. For example rw-rw-r--</td></tr></table><h3>State management: </h3><table 
id="stateful"><tr><th>Scope</th><th>Description</th></tr><tr><td>CLUSTER</td><td>After
 performing a listing of HDFS files, the timestamp of the newest file is 
stored, along with the filenames of all files that share that same timestamp. 
This allows the Processor to list only files that have been added or modified 
after this date the next time that the Processor is run. State is stored across 
the cluster so that this Processor can be run on Primary Node only and if a new 
Primary Node is selected, the new node can pick up where the previous node left 
off, without duplicating the data.</td></tr></table><h3>Restricted: 
</h3><h3>See Also:</h3><p><a 
href="../org.apache.nifi.processors.hadoop.GetHDFS/index.html">GetHDFS</a>, <a 
href="../org.apache.nifi.processors.had
 oop.FetchHDFS/index.html">FetchHDFS</a>, <a 
href="../org.apache.nifi.processors.hadoop.PutHDFS/index.html">PutHDFS</a></p></body></html>
\ No newline at end of file

Modified: 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.PutHDFS/index.html
URL: 
http://svn.apache.org/viewvc/nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.PutHDFS/index.html?rev=1771896&r1=1771895&r2=1771896&view=diff
==============================================================================
--- 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.PutHDFS/index.html
 (original)
+++ 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.PutHDFS/index.html
 Tue Nov 29 12:33:21 2016
@@ -1 +1 @@
-<!DOCTYPE html><html lang="en"><head><meta 
charset="utf-8"></meta><title>PutHDFS</title><link rel="stylesheet" 
href="../../css/component-usage.css" 
type="text/css"></link></head><body><h2>Description: </h2><p>Write FlowFile 
data to Hadoop Distributed File System (HDFS)</p><h3>Tags: </h3><p>hadoop, 
HDFS, put, copy, filesystem</p><h3>Properties: </h3><p>In the list below, the 
names of required properties appear in <strong>bold</strong>. Any other 
properties (not in bold) are considered optional. The table also indicates any 
default values, and whether a property supports the <a 
href="../../html/expression-language-guide.html">NiFi Expression 
Language</a>.</p><table id="properties"><tr><th>Name</th><th>Default 
Value</th><th>Allowable Values</th><th>Description</th></tr><tr><td 
id="name">Hadoop Configuration Resources</td><td id="default-value"></td><td 
id="allowable-values"></td><td id="description">A file or comma separated list 
of files which contains the Hadoop file system configura
 tion. Without this, Hadoop will search the classpath for a 'core-site.xml' and 
'hdfs-site.xml' file or will revert to a default 
configuration.</td></tr><tr><td id="name">Kerberos Principal</td><td 
id="default-value"></td><td id="allowable-values"></td><td 
id="description">Kerberos principal to authenticate as. Requires 
nifi.kerberos.krb5.file to be set in your nifi.properties</td></tr><tr><td 
id="name">Kerberos Keytab</td><td id="default-value"></td><td 
id="allowable-values"></td><td id="description">Kerberos keytab associated with 
the principal. Requires nifi.kerberos.krb5.file to be set in your 
nifi.properties</td></tr><tr><td id="name">Kerberos Relogin Period</td><td 
id="default-value">4 hours</td><td id="allowable-values"></td><td 
id="description">Period of time which should pass before attempting a kerberos 
relogin</td></tr><tr><td id="name"><strong>Directory</strong></td><td 
id="default-value"></td><td id="allowable-values"></td><td id="description">The 
parent HDFS directory t
 o which files should be written<br/><strong>Supports Expression Language: 
true</strong></td></tr><tr><td id="name"><strong>Conflict Resolution 
Strategy</strong></td><td id="default-value">fail</td><td 
id="allowable-values"><ul><li>replace</li><li>ignore</li><li>fail</li></ul></td><td
 id="description">Indicates what should happen when a file with the same name 
already exists in the output directory</td></tr><tr><td id="name">Block 
Size</td><td id="default-value"></td><td id="allowable-values"></td><td 
id="description">Size of each block as written to HDFS. This overrides the 
Hadoop Configuration</td></tr><tr><td id="name">IO Buffer Size</td><td 
id="default-value"></td><td id="allowable-values"></td><td 
id="description">Amount of memory to use to buffer file contents during IO. 
This overrides the Hadoop Configuration</td></tr><tr><td 
id="name">Replication</td><td id="default-value"></td><td 
id="allowable-values"></td><td id="description">Number of times that HDFS will 
replicate each f
 ile. This overrides the Hadoop Configuration</td></tr><tr><td 
id="name">Permissions umask</td><td id="default-value"></td><td 
id="allowable-values"></td><td id="description">A umask represented as an octal 
number which determines the permissions of files written to HDFS. This 
overrides the Hadoop Configuration dfs.umaskmode</td></tr><tr><td 
id="name">Remote Owner</td><td id="default-value"></td><td 
id="allowable-values"></td><td id="description">Changes the owner of the HDFS 
file to this value after it is written. This only works if NiFi is running as a 
user that has HDFS super user privilege to change owner</td></tr><tr><td 
id="name">Remote Group</td><td id="default-value"></td><td 
id="allowable-values"></td><td id="description">Changes the group of the HDFS 
file to this value after it is written. This only works if NiFi is running as a 
user that has HDFS super user privilege to change group</td></tr><tr><td 
id="name"><strong>Compression codec</strong></td><td id="default-value">NO
 NE</td><td 
id="allowable-values"><ul><li>NONE</li><li>DEFAULT</li><li>BZIP</li><li>GZIP</li><li>LZ4</li><li>SNAPPY</li><li>AUTOMATIC</li></ul></td><td
 id="description">No Description Provided.</td></tr></table><h3>Relationships: 
</h3><table 
id="relationships"><tr><th>Name</th><th>Description</th></tr><tr><td>success</td><td>Files
 that have been successfully written to HDFS are transferred to this 
relationship</td></tr><tr><td>failure</td><td>Files that could not be written 
to HDFS for some reason are transferred to this 
relationship</td></tr></table><h3>Reads Attributes: </h3><table 
id="reads-attributes"><tr><th>Name</th><th>Description</th></tr><tr><td>filename</td><td>The
 name of the file written to HDFS comes from the value of this 
attribute.</td></tr></table><h3>Writes Attributes: </h3><table 
id="writes-attributes"><tr><th>Name</th><th>Description</th></tr><tr><td>filename</td><td>The
 name of the file written to HDFS is stored in this 
attribute.</td></tr><tr><td>absolute.hdfs.pa
 th</td><td>The absolute path to the file on HDFS is stored in this 
attribute.</td></tr></table><h3>See Also:</h3><p><a 
href="../org.apache.nifi.processors.hadoop.GetHDFS/index.html">GetHDFS</a></p></body></html>
\ No newline at end of file
+<!DOCTYPE html><html lang="en"><head><meta 
charset="utf-8"></meta><title>PutHDFS</title><link rel="stylesheet" 
href="../../css/component-usage.css" 
type="text/css"></link></head><body><h2>Description: </h2><p>Write FlowFile 
data to Hadoop Distributed File System (HDFS)</p><h3>Tags: </h3><p>hadoop, 
HDFS, put, copy, filesystem, restricted</p><h3>Properties: </h3><p>In the list 
below, the names of required properties appear in <strong>bold</strong>. Any 
other properties (not in bold) are considered optional. The table also 
indicates any default values, and whether a property supports the <a 
href="../../html/expression-language-guide.html">NiFi Expression 
Language</a>.</p><table id="properties"><tr><th>Name</th><th>Default 
Value</th><th>Allowable Values</th><th>Description</th></tr><tr><td 
id="name">Hadoop Configuration Resources</td><td id="default-value"></td><td 
id="allowable-values"></td><td id="description">A file or comma separated list 
of files which contains the Hadoop file syst
 em configuration. Without this, Hadoop will search the classpath for a 
'core-site.xml' and 'hdfs-site.xml' file or will revert to a default 
configuration.</td></tr><tr><td id="name">Kerberos Principal</td><td 
id="default-value"></td><td id="allowable-values"></td><td 
id="description">Kerberos principal to authenticate as. Requires 
nifi.kerberos.krb5.file to be set in your nifi.properties</td></tr><tr><td 
id="name">Kerberos Keytab</td><td id="default-value"></td><td 
id="allowable-values"></td><td id="description">Kerberos keytab associated with 
the principal. Requires nifi.kerberos.krb5.file to be set in your 
nifi.properties</td></tr><tr><td id="name">Kerberos Relogin Period</td><td 
id="default-value">4 hours</td><td id="allowable-values"></td><td 
id="description">Period of time which should pass before attempting a kerberos 
relogin</td></tr><tr><td id="name">Additional Classpath Resources</td><td 
id="default-value"></td><td id="allowable-values"></td><td id="description">A 
comma-sep
 arated list of paths to files and/or directories that will be added to the 
classpath. When specifying a directory, all files with in the directory will be 
added to the classpath, but further sub-directories will not be 
included.</td></tr><tr><td id="name"><strong>Directory</strong></td><td 
id="default-value"></td><td id="allowable-values"></td><td id="description">The 
parent HDFS directory to which files should be written<br/><strong>Supports 
Expression Language: true</strong></td></tr><tr><td id="name"><strong>Conflict 
Resolution Strategy</strong></td><td id="default-value">fail</td><td 
id="allowable-values"><ul><li>replace <img src="../../html/images/iconInfo.png" 
alt="Replaces the existing file if any." title="Replaces the existing file if 
any."></img></li><li>ignore <img src="../../html/images/iconInfo.png" 
alt="Ignores the flow file and routes it to success." title="Ignores the flow 
file and routes it to success."></img></li><li>fail <img 
src="../../html/images/iconInfo.png" al
 t="Penalizes the flow file and routes it to failure." title="Penalizes the 
flow file and routes it to failure."></img></li><li>append <img 
src="../../html/images/iconInfo.png" alt="Appends to the existing file if any, 
creates a new file otherwise." title="Appends to the existing file if any, 
creates a new file otherwise."></img></li></ul></td><td 
id="description">Indicates what should happen when a file with the same name 
already exists in the output directory</td></tr><tr><td id="name">Block 
Size</td><td id="default-value"></td><td id="allowable-values"></td><td 
id="description">Size of each block as written to HDFS. This overrides the 
Hadoop Configuration</td></tr><tr><td id="name">IO Buffer Size</td><td 
id="default-value"></td><td id="allowable-values"></td><td 
id="description">Amount of memory to use to buffer file contents during IO. 
This overrides the Hadoop Configuration</td></tr><tr><td 
id="name">Replication</td><td id="default-value"></td><td 
id="allowable-values"></td><td 
 id="description">Number of times that HDFS will replicate each file. This 
overrides the Hadoop Configuration</td></tr><tr><td id="name">Permissions 
umask</td><td id="default-value"></td><td id="allowable-values"></td><td 
id="description">A umask represented as an octal number which determines the 
permissions of files written to HDFS. This overrides the Hadoop Configuration 
dfs.umaskmode</td></tr><tr><td id="name">Remote Owner</td><td 
id="default-value"></td><td id="allowable-values"></td><td 
id="description">Changes the owner of the HDFS file to this value after it is 
written. This only works if NiFi is running as a user that has HDFS super user 
privilege to change owner</td></tr><tr><td id="name">Remote Group</td><td 
id="default-value"></td><td id="allowable-values"></td><td 
id="description">Changes the group of the HDFS file to this value after it is 
written. This only works if NiFi is running as a user that has HDFS super user 
privilege to change group</td></tr><tr><td id="name">
 <strong>Compression codec</strong></td><td id="default-value">NONE</td><td 
id="allowable-values"><ul><li>NONE</li><li>DEFAULT</li><li>BZIP</li><li>GZIP</li><li>LZ4</li><li>SNAPPY</li><li>AUTOMATIC</li></ul></td><td
 id="description">No Description Provided.</td></tr></table><h3>Relationships: 
</h3><table 
id="relationships"><tr><th>Name</th><th>Description</th></tr><tr><td>success</td><td>Files
 that have been successfully written to HDFS are transferred to this 
relationship</td></tr><tr><td>failure</td><td>Files that could not be written 
to HDFS for some reason are transferred to this 
relationship</td></tr></table><h3>Reads Attributes: </h3><table 
id="reads-attributes"><tr><th>Name</th><th>Description</th></tr><tr><td>filename</td><td>The
 name of the file written to HDFS comes from the value of this 
attribute.</td></tr></table><h3>Writes Attributes: </h3><table 
id="writes-attributes"><tr><th>Name</th><th>Description</th></tr><tr><td>filename</td><td>The
 name of the file written to HDF
 S is stored in this attribute.</td></tr><tr><td>absolute.hdfs.path</td><td>The 
absolute path to the file on HDFS is stored in this 
attribute.</td></tr></table><h3>State management: </h3>This processor has no 
state management.<h3>Restricted: </h3><td>Provides operator the ability to 
write to any file that NiFi has access to in HDFS or the local 
filesystem.</td><h3>See Also:</h3><p><a 
href="../org.apache.nifi.processors.hadoop.GetHDFS/index.html">GetHDFS</a></p></body></html>
\ No newline at end of file

Modified: 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.inotify.GetHDFSEvents/index.html
URL: 
http://svn.apache.org/viewvc/nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.inotify.GetHDFSEvents/index.html?rev=1771896&r1=1771895&r2=1771896&view=diff
==============================================================================
--- 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.inotify.GetHDFSEvents/index.html
 (original)
+++ 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.inotify.GetHDFSEvents/index.html
 Tue Nov 29 12:33:21 2016
@@ -1 +1 @@
-<!DOCTYPE html><html lang="en"><head><meta 
charset="utf-8"></meta><title>GetHDFSEvents</title><link rel="stylesheet" 
href="../../css/component-usage.css" 
type="text/css"></link></head><body><h2>Description: </h2><p>This processor 
polls the notification events provided by the HdfsAdmin API. Since this uses 
the HdfsAdmin APIs it is required to run as an HDFS super user. Currently there 
are six types of events (append, close, create, metadata, rename, and unlink). 
Please see org.apache.hadoop.hdfs.inotify.Event documentation for full 
explanations of each event. This processor will poll for new events based on a 
defined duration. For each event received a new flow file will be created with 
the expected attributes and the event itself serialized to JSON and written to 
the flow file's content. For example, if event.type is APPEND then the content 
of the flow file will contain a JSON file containing the information about the 
append event. If successful the flow files are sent to the 'succe
 ss' relationship. Be careful of where the generated flow files are stored. If 
the flow files are stored in one of processor's watch directories there will be 
a never ending flow of events. It is also important to be aware that this 
processor must consume all events. The filtering must happen within the 
processor. This is because the HDFS admin's event notifications API does not 
have filtering.</p><h3>Tags: </h3><p>hadoop, events, inotify, notifications, 
filesystem</p><h3>Properties: </h3><p>In the list below, the names of required 
properties appear in <strong>bold</strong>. Any other properties (not in bold) 
are considered optional. The table also indicates any default values, and 
whether a property supports the <a 
href="../../html/expression-language-guide.html">NiFi Expression 
Language</a>.</p><table id="properties"><tr><th>Name</th><th>Default 
Value</th><th>Allowable Values</th><th>Description</th></tr><tr><td 
id="name">Hadoop Configuration Resources</td><td id="default-value"></
 td><td id="allowable-values"></td><td id="description">A file or comma 
separated list of files which contains the Hadoop file system configuration. 
Without this, Hadoop will search the classpath for a 'core-site.xml' and 
'hdfs-site.xml' file or will revert to a default 
configuration.</td></tr><tr><td id="name">Kerberos Principal</td><td 
id="default-value"></td><td id="allowable-values"></td><td 
id="description">Kerberos principal to authenticate as. Requires 
nifi.kerberos.krb5.file to be set in your nifi.properties</td></tr><tr><td 
id="name">Kerberos Keytab</td><td id="default-value"></td><td 
id="allowable-values"></td><td id="description">Kerberos keytab associated with 
the principal. Requires nifi.kerberos.krb5.file to be set in your 
nifi.properties</td></tr><tr><td id="name">Kerberos Relogin Period</td><td 
id="default-value">4 hours</td><td id="allowable-values"></td><td 
id="description">Period of time which should pass before attempting a kerberos 
relogin</td></tr><tr><td id="na
 me"><strong>Poll Duration</strong></td><td id="default-value">1 second</td><td 
id="allowable-values"></td><td id="description">The time before the polling 
method returns with the next batch of events if they exist. It may exceed this 
amount of time by up to the time required for an RPC to the 
NameNode.</td></tr><tr><td id="name"><strong>HDFS Path to 
Watch</strong></td><td id="default-value"></td><td 
id="allowable-values"></td><td id="description">The HDFS path to get event 
notifications for. This property accepts both expression language and regular 
expressions. This will be evaluated during the OnScheduled 
phase.<br/><strong>Supports Expression Language: true</strong></td></tr><tr><td 
id="name"><strong>Ignore Hidden Files</strong></td><td 
id="default-value">false</td><td 
id="allowable-values"><ul><li>true</li><li>false</li></ul></td><td 
id="description">If true and the final component of the path associated with a 
given event starts with a '.' then that event will not be processed.
 </td></tr><tr><td id="name"><strong>Event Types to Filter On</strong></td><td 
id="default-value">append, close, create, metadata, rename, unlink</td><td 
id="allowable-values"></td><td id="description">A comma-separated list of event 
types to process. Valid event types are: append, close, create, metadata, 
rename, and unlink. Case does not matter.</td></tr><tr><td 
id="name"><strong>IOException Retries During Event Polling</strong></td><td 
id="default-value">3</td><td id="allowable-values"></td><td 
id="description">According to the HDFS admin API for event polling it is good 
to retry at least a few times. This number defines how many times the poll will 
be retried if it throws an IOException.</td></tr></table><h3>Relationships: 
</h3><table 
id="relationships"><tr><th>Name</th><th>Description</th></tr><tr><td>success</td><td>A
 flow file with updated information about a specific event will be sent to this 
relationship.</td></tr></table><h3>Reads Attributes: </h3>None 
specified.<h3>Writes
  Attributes: </h3><table 
id="writes-attributes"><tr><th>Name</th><th>Description</th></tr><tr><td>mime.type</td><td>This
 is always 
application/json.</td></tr><tr><td>hdfs.inotify.event.type</td><td>This will 
specify the specific HDFS notification event type. Currently there are six 
types of events (append, close, create, metadata, rename, and 
unlink).</td></tr><tr><td>hdfs.inotify.event.path</td><td>The specific path 
that the event is tied to.</td></tr></table><h3>See Also:</h3><p><a 
href="../org.apache.nifi.processors.hadoop.GetHDFS/index.html">GetHDFS</a>, <a 
href="../org.apache.nifi.processors.hadoop.FetchHDFS/index.html">FetchHDFS</a>, 
<a href="../org.apache.nifi.processors.hadoop.PutHDFS/index.html">PutHDFS</a>, 
<a 
href="../org.apache.nifi.processors.hadoop.ListHDFS/index.html">ListHDFS</a></p></body></html>
\ No newline at end of file
+<!DOCTYPE html><html lang="en"><head><meta 
charset="utf-8"></meta><title>GetHDFSEvents</title><link rel="stylesheet" 
href="../../css/component-usage.css" 
type="text/css"></link></head><body><h2>Description: </h2><p>This processor 
polls the notification events provided by the HdfsAdmin API. Since this uses 
the HdfsAdmin APIs it is required to run as an HDFS super user. Currently there 
are six types of events (append, close, create, metadata, rename, and unlink). 
Please see org.apache.hadoop.hdfs.inotify.Event documentation for full 
explanations of each event. This processor will poll for new events based on a 
defined duration. For each event received a new flow file will be created with 
the expected attributes and the event itself serialized to JSON and written to 
the flow file's content. For example, if event.type is APPEND then the content 
of the flow file will contain a JSON file containing the information about the 
append event. If successful the flow files are sent to the 'succe
 ss' relationship. Be careful of where the generated flow files are stored. If 
the flow files are stored in one of processor's watch directories there will be 
a never ending flow of events. It is also important to be aware that this 
processor must consume all events. The filtering must happen within the 
processor. This is because the HDFS admin's event notifications API does not 
have filtering.</p><h3>Tags: </h3><p>hadoop, events, inotify, notifications, 
filesystem</p><h3>Properties: </h3><p>In the list below, the names of required 
properties appear in <strong>bold</strong>. Any other properties (not in bold) 
are considered optional. The table also indicates any default values, and 
whether a property supports the <a 
href="../../html/expression-language-guide.html">NiFi Expression 
Language</a>.</p><table id="properties"><tr><th>Name</th><th>Default 
Value</th><th>Allowable Values</th><th>Description</th></tr><tr><td 
id="name">Hadoop Configuration Resources</td><td id="default-value"></
 td><td id="allowable-values"></td><td id="description">A file or comma 
separated list of files which contains the Hadoop file system configuration. 
Without this, Hadoop will search the classpath for a 'core-site.xml' and 
'hdfs-site.xml' file or will revert to a default 
configuration.</td></tr><tr><td id="name">Kerberos Principal</td><td 
id="default-value"></td><td id="allowable-values"></td><td 
id="description">Kerberos principal to authenticate as. Requires 
nifi.kerberos.krb5.file to be set in your nifi.properties</td></tr><tr><td 
id="name">Kerberos Keytab</td><td id="default-value"></td><td 
id="allowable-values"></td><td id="description">Kerberos keytab associated with 
the principal. Requires nifi.kerberos.krb5.file to be set in your 
nifi.properties</td></tr><tr><td id="name">Kerberos Relogin Period</td><td 
id="default-value">4 hours</td><td id="allowable-values"></td><td 
id="description">Period of time which should pass before attempting a kerberos 
relogin</td></tr><tr><td id="na
 me">Additional Classpath Resources</td><td id="default-value"></td><td 
id="allowable-values"></td><td id="description">A comma-separated list of paths 
to files and/or directories that will be added to the classpath. When 
specifying a directory, all files with in the directory will be added to the 
classpath, but further sub-directories will not be included.</td></tr><tr><td 
id="name"><strong>Poll Duration</strong></td><td id="default-value">1 
second</td><td id="allowable-values"></td><td id="description">The time before 
the polling method returns with the next batch of events if they exist. It may 
exceed this amount of time by up to the time required for an RPC to the 
NameNode.</td></tr><tr><td id="name"><strong>HDFS Path to 
Watch</strong></td><td id="default-value"></td><td 
id="allowable-values"></td><td id="description">The HDFS path to get event 
notifications for. This property accepts both expression language and regular 
expressions. This will be evaluated during the OnScheduled 
 phase.<br/><strong>Supports Expression Language: 
true</strong></td></tr><tr><td id="name"><strong>Ignore Hidden 
Files</strong></td><td id="default-value">false</td><td 
id="allowable-values"><ul><li>true</li><li>false</li></ul></td><td 
id="description">If true and the final component of the path associated with a 
given event starts with a '.' then that event will not be 
processed.</td></tr><tr><td id="name"><strong>Event Types to Filter 
On</strong></td><td id="default-value">append, close, create, metadata, rename, 
unlink</td><td id="allowable-values"></td><td id="description">A 
comma-separated list of event types to process. Valid event types are: append, 
close, create, metadata, rename, and unlink. Case does not 
matter.</td></tr><tr><td id="name"><strong>IOException Retries During Event 
Polling</strong></td><td id="default-value">3</td><td 
id="allowable-values"></td><td id="description">According to the HDFS admin API 
for event polling it is good to retry at least a few times. This
  number defines how many times the poll will be retried if it throws an 
IOException.</td></tr></table><h3>Relationships: </h3><table 
id="relationships"><tr><th>Name</th><th>Description</th></tr><tr><td>success</td><td>A
 flow file with updated information about a specific event will be sent to this 
relationship.</td></tr></table><h3>Reads Attributes: </h3>None 
specified.<h3>Writes Attributes: </h3><table 
id="writes-attributes"><tr><th>Name</th><th>Description</th></tr><tr><td>mime.type</td><td>This
 is always 
application/json.</td></tr><tr><td>hdfs.inotify.event.type</td><td>This will 
specify the specific HDFS notification event type. Currently there are six 
types of events (append, close, create, metadata, rename, and 
unlink).</td></tr><tr><td>hdfs.inotify.event.path</td><td>The specific path 
that the event is tied to.</td></tr></table><h3>State management: </h3><table 
id="stateful"><tr><th>Scope</th><th>Description</th></tr><tr><td>CLUSTER</td><td>The
 last used transaction id is sto
 red. This is used </td></tr></table><h3>Restricted: </h3><h3>See 
Also:</h3><p><a 
href="../org.apache.nifi.processors.hadoop.GetHDFS/index.html">GetHDFS</a>, <a 
href="../org.apache.nifi.processors.hadoop.FetchHDFS/index.html">FetchHDFS</a>, 
<a href="../org.apache.nifi.processors.hadoop.PutHDFS/index.html">PutHDFS</a>, 
<a 
href="../org.apache.nifi.processors.hadoop.ListHDFS/index.html">ListHDFS</a></p></body></html>
\ No newline at end of file

Modified: 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hive.ConvertAvroToORC/index.html
URL: 
http://svn.apache.org/viewvc/nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hive.ConvertAvroToORC/index.html?rev=1771896&r1=1771895&r2=1771896&view=diff
==============================================================================
--- 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hive.ConvertAvroToORC/index.html
 (original)
+++ 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hive.ConvertAvroToORC/index.html
 Tue Nov 29 12:33:21 2016
@@ -1 +1 @@
-<!DOCTYPE html><html lang="en"><head><meta 
charset="utf-8"></meta><title>ConvertAvroToORC</title><link rel="stylesheet" 
href="../../css/component-usage.css" 
type="text/css"></link></head><body><h2>Description: </h2><p>Converts an Avro 
record into ORC file format. This processor provides a direct mapping of an 
Avro record to an ORC record, such that the resulting ORC file will have the 
same hierarchical structure as the Avro document. If an incoming FlowFile 
contains a stream of multiple Avro records, the resultant FlowFile will contain 
a ORC file containing all of the Avro records.  If an incoming FlowFile does 
not contain any records, an empty ORC file is the output.</p><h3>Tags: 
</h3><p>avro, orc, hive, convert</p><h3>Properties: </h3><p>In the list below, 
the names of required properties appear in <strong>bold</strong>. Any other 
properties (not in bold) are considered optional. The table also indicates any 
default values, and whether a property supports the <a href="../../html/e
 xpression-language-guide.html">NiFi Expression Language</a>.</p><table 
id="properties"><tr><th>Name</th><th>Default Value</th><th>Allowable 
Values</th><th>Description</th></tr><tr><td id="name">ORC Configuration 
Resources</td><td id="default-value"></td><td id="allowable-values"></td><td 
id="description">A file or comma separated list of files which contains the ORC 
configuration (hive-site.xml, e.g.). Without this, Hadoop will search the 
classpath for a 'hive-site.xml' file or will revert to a default configuration. 
Please see the ORC documentation for more details.</td></tr><tr><td 
id="name"><strong>Stripe Size</strong></td><td id="default-value">64 MB</td><td 
id="allowable-values"></td><td id="description">The size of the memory buffer 
(in bytes) for writing stripes to an ORC file</td></tr><tr><td 
id="name"><strong>Buffer Size</strong></td><td id="default-value">10 KB</td><td 
id="allowable-values"></td><td id="description">The maximum size of the memory 
buffers (in bytes) used fo
 r compressing and storing a stripe in memory. This is a hint to the ORC 
writer, which may choose to use a smaller buffer size based on stripe size and 
number of columns for efficient stripe writing and memory 
utilization.</td></tr><tr><td id="name"><strong>Compression 
Type</strong></td><td id="default-value">NONE</td><td 
id="allowable-values"><ul><li>NONE</li><li>ZLIB</li><li>SNAPPY</li><li>LZO</li></ul></td><td
 id="description">No Description Provided.</td></tr><tr><td id="name">Hive 
Table Name</td><td id="default-value"></td><td id="allowable-values"></td><td 
id="description">An optional table name to insert into the hive.ddl attribute. 
The generated DDL can be used by a PutHiveQL processor (presumably after a 
PutHDFS processor) to create a table backed by the converted ORC file. If this 
property is not provided, the full name (including namespace) of the incoming 
Avro record will be normalized and used as the table name.<br/><strong>Supports 
Expression Language: true</strong></td
 ></tr></table><h3>Relationships: </h3><table 
 >id="relationships"><tr><th>Name</th><th>Description</th></tr><tr><td>success</td><td>A
 > FlowFile is routed to this relationship after it has been converted to ORC 
 >format.</td></tr><tr><td>failure</td><td>A FlowFile is routed to this 
 >relationship if it cannot be parsed as Avro or cannot be converted to ORC for 
 >any reason</td></tr></table><h3>Reads Attributes: </h3>None 
 >specified.<h3>Writes Attributes: </h3><table 
 >id="writes-attributes"><tr><th>Name</th><th>Description</th></tr><tr><td>mime.type</td><td>Sets
 > the mime type to 
 >application/octet-stream</td></tr><tr><td>filename</td><td>Sets the filename 
 >to the existing filename with the extension replaced by / added to by 
 >.orc</td></tr><tr><td>record.count</td><td>Sets the number of records in the 
 >ORC file.</td></tr><tr><td>hive.ddl</td><td>Creates a partial Hive DDL 
 >statement for creating a table in Hive from this ORC file. This can be used 
 >in ReplaceText for setting the content to the DDL. To
  make it valid DDL, add "LOCATION '&lt;path_to_orc_file_in_hdfs&gt;'", where 
the path is the directory that contains this ORC file on HDFS. For example, 
ConvertAvroToORC can send flow files to a PutHDFS processor to send the file to 
HDFS, then to a ReplaceText to set the content to this DDL (plus the LOCATION 
clause as described), then to PutHiveQL processor to create the table if it 
doesn't exist.</td></tr></table></body></html>
\ No newline at end of file
+<!DOCTYPE html><html lang="en"><head><meta 
charset="utf-8"></meta><title>ConvertAvroToORC</title><link rel="stylesheet" 
href="../../css/component-usage.css" 
type="text/css"></link></head><body><h2>Description: </h2><p>Converts an Avro 
record into ORC file format. This processor provides a direct mapping of an 
Avro record to an ORC record, such that the resulting ORC file will have the 
same hierarchical structure as the Avro document. If an incoming FlowFile 
contains a stream of multiple Avro records, the resultant FlowFile will contain 
a ORC file containing all of the Avro records.  If an incoming FlowFile does 
not contain any records, an empty ORC file is the output.</p><h3>Tags: 
</h3><p>avro, orc, hive, convert</p><h3>Properties: </h3><p>In the list below, 
the names of required properties appear in <strong>bold</strong>. Any other 
properties (not in bold) are considered optional. The table also indicates any 
default values, and whether a property supports the <a href="../../html/e
 xpression-language-guide.html">NiFi Expression Language</a>.</p><table 
id="properties"><tr><th>Name</th><th>Default Value</th><th>Allowable 
Values</th><th>Description</th></tr><tr><td id="name">ORC Configuration 
Resources</td><td id="default-value"></td><td id="allowable-values"></td><td 
id="description">A file or comma separated list of files which contains the ORC 
configuration (hive-site.xml, e.g.). Without this, Hadoop will search the 
classpath for a 'hive-site.xml' file or will revert to a default configuration. 
Please see the ORC documentation for more details.</td></tr><tr><td 
id="name"><strong>Stripe Size</strong></td><td id="default-value">64 MB</td><td 
id="allowable-values"></td><td id="description">The size of the memory buffer 
(in bytes) for writing stripes to an ORC file</td></tr><tr><td 
id="name"><strong>Buffer Size</strong></td><td id="default-value">10 KB</td><td 
id="allowable-values"></td><td id="description">The maximum size of the memory 
buffers (in bytes) used fo
 r compressing and storing a stripe in memory. This is a hint to the ORC 
writer, which may choose to use a smaller buffer size based on stripe size and 
number of columns for efficient stripe writing and memory 
utilization.</td></tr><tr><td id="name"><strong>Compression 
Type</strong></td><td id="default-value">NONE</td><td 
id="allowable-values"><ul><li>NONE</li><li>ZLIB</li><li>SNAPPY</li><li>LZO</li></ul></td><td
 id="description">No Description Provided.</td></tr><tr><td id="name">Hive 
Table Name</td><td id="default-value"></td><td id="allowable-values"></td><td 
id="description">An optional table name to insert into the hive.ddl attribute. 
The generated DDL can be used by a PutHiveQL processor (presumably after a 
PutHDFS processor) to create a table backed by the converted ORC file. If this 
property is not provided, the full name (including namespace) of the incoming 
Avro record will be normalized and used as the table name.<br/><strong>Supports 
Expression Language: true</strong></td
 ></tr></table><h3>Relationships: </h3><table 
 >id="relationships"><tr><th>Name</th><th>Description</th></tr><tr><td>success</td><td>A
 > FlowFile is routed to this relationship after it has been converted to ORC 
 >format.</td></tr><tr><td>failure</td><td>A FlowFile is routed to this 
 >relationship if it cannot be parsed as Avro or cannot be converted to ORC for 
 >any reason</td></tr></table><h3>Reads Attributes: </h3>None 
 >specified.<h3>Writes Attributes: </h3><table 
 >id="writes-attributes"><tr><th>Name</th><th>Description</th></tr><tr><td>mime.type</td><td>Sets
 > the mime type to 
 >application/octet-stream</td></tr><tr><td>filename</td><td>Sets the filename 
 >to the existing filename with the extension replaced by / added to by 
 >.orc</td></tr><tr><td>record.count</td><td>Sets the number of records in the 
 >ORC file.</td></tr><tr><td>hive.ddl</td><td>Creates a partial Hive DDL 
 >statement for creating a table in Hive from this ORC file. This can be used 
 >in ReplaceText for setting the content to the DDL. To
  make it valid DDL, add "LOCATION '&lt;path_to_orc_file_in_hdfs&gt;'", where 
the path is the directory that contains this ORC file on HDFS. For example, 
ConvertAvroToORC can send flow files to a PutHDFS processor to send the file to 
HDFS, then to a ReplaceText to set the content to this DDL (plus the LOCATION 
clause as described), then to PutHiveQL processor to create the table if it 
doesn't exist.</td></tr></table><h3>State management: </h3>This processor has 
no state management.<h3>Restricted: </h3></body></html>
\ No newline at end of file

Modified: 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hive.PutHiveQL/index.html
URL: 
http://svn.apache.org/viewvc/nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hive.PutHiveQL/index.html?rev=1771896&r1=1771895&r2=1771896&view=diff
==============================================================================
--- 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hive.PutHiveQL/index.html
 (original)
+++ 
nifi/site/trunk/docs/nifi-docs/components/org.apache.nifi.processors.hive.PutHiveQL/index.html
 Tue Nov 29 12:33:21 2016
@@ -1 +1 @@
-<!DOCTYPE html><html lang="en"><head><meta 
charset="utf-8"></meta><title>PutHiveQL</title><link rel="stylesheet" 
href="../../css/component-usage.css" 
type="text/css"></link></head><body><h2>Description: </h2><p>Executes a HiveQL 
DDL/DML command (UPDATE, INSERT, e.g.). The content of an incoming FlowFile is 
expected to be the HiveQL command to execute. The HiveQL command may use the ? 
to escape parameters. In this case, the parameters to use must exist as 
FlowFile attributes with the naming convention hiveql.args.N.type and 
hiveql.args.N.value, where N is a positive integer. The hiveql.args.N.type is 
expected to be a number indicating the JDBC Type. The content of the FlowFile 
is expected to be in UTF-8 format.</p><h3>Tags: </h3><p>sql, hive, put, 
database, update, insert</p><h3>Properties: </h3><p>In the list below, the 
names of required properties appear in <strong>bold</strong>. Any other 
properties (not in bold) are considered optional. The table also indicates any 
default values
 .</p><table id="properties"><tr><th>Name</th><th>Default 
Value</th><th>Allowable Values</th><th>Description</th></tr><tr><td 
id="name"><strong>Hive Database Connection Pooling Service</strong></td><td 
id="default-value"></td><td id="allowable-values"><strong>Controller Service 
API: </strong><br/>HiveDBCPService<br/><strong>Implementation:</strong><br/><a 
href="../org.apache.nifi.dbcp.hive.HiveConnectionPool/index.html">HiveConnectionPool</a></td><td
 id="description">The Hive Controller Service that is used to obtain 
connection(s) to the Hive database</td></tr><tr><td id="name"><strong>Batch 
Size</strong></td><td id="default-value">100</td><td 
id="allowable-values"></td><td id="description">The preferred number of 
FlowFiles to put to the database in a single transaction</td></tr><tr><td 
id="name"><strong>Character Set</strong></td><td 
id="default-value">UTF-8</td><td id="allowable-values"></td><td 
id="description">Specifies the character set of the record 
data.</td></tr></table><h3>R
 elationships: </h3><table 
id="relationships"><tr><th>Name</th><th>Description</th></tr><tr><td>retry</td><td>A
 FlowFile is routed to this relationship if the database cannot be updated but 
attempting the operation again may succeed</td></tr><tr><td>success</td><td>A 
FlowFile is routed to this relationship after the database is successfully 
updated</td></tr><tr><td>failure</td><td>A FlowFile is routed to this 
relationship if the database cannot be updated and retrying the operation will 
also fail, such as an invalid query or an integrity constraint 
violation</td></tr></table><h3>Reads Attributes: </h3><table 
id="reads-attributes"><tr><th>Name</th><th>Description</th></tr><tr><td>hiveql.args.N.type</td><td>Incoming
 FlowFiles are expected to be parameterized HiveQL statements. The type of each 
Parameter is specified as an integer that represents the JDBC Type of the 
parameter.</td></tr><tr><td>hiveql.args.N.value</td><td>Incoming FlowFiles are 
expected to be parameterized HiveQL statem
 ents. The value of the Parameters are specified as hiveql.args.1.value, 
hiveql.args.2.value, hiveql.args.3.value, and so on. The type of the 
hiveql.args.1.value Parameter is specified by the hiveql.args.1.type 
attribute.</td></tr></table><h3>Writes Attributes: </h3>None specified.<h3>See 
Also:</h3><p><a 
href="../org.apache.nifi.processors.hive.SelectHiveQL/index.html">SelectHiveQL</a></p></body></html>
\ No newline at end of file
+<!DOCTYPE html><html lang="en"><head><meta 
charset="utf-8"></meta><title>PutHiveQL</title><link rel="stylesheet" 
href="../../css/component-usage.css" 
type="text/css"></link></head><body><h2>Description: </h2><p>Executes a HiveQL 
DDL/DML command (UPDATE, INSERT, e.g.). The content of an incoming FlowFile is 
expected to be the HiveQL command to execute. The HiveQL command may use the ? 
to escape parameters. In this case, the parameters to use must exist as 
FlowFile attributes with the naming convention hiveql.args.N.type and 
hiveql.args.N.value, where N is a positive integer. The hiveql.args.N.type is 
expected to be a number indicating the JDBC Type. The content of the FlowFile 
is expected to be in UTF-8 format.</p><h3>Tags: </h3><p>sql, hive, put, 
database, update, insert</p><h3>Properties: </h3><p>In the list below, the 
names of required properties appear in <strong>bold</strong>. Any other 
properties (not in bold) are considered optional. The table also indicates any 
default values
 .</p><table id="properties"><tr><th>Name</th><th>Default 
Value</th><th>Allowable Values</th><th>Description</th></tr><tr><td 
id="name"><strong>Hive Database Connection Pooling Service</strong></td><td 
id="default-value"></td><td id="allowable-values"><strong>Controller Service 
API: </strong><br/>HiveDBCPService<br/><strong>Implementation:</strong><br/><a 
href="../org.apache.nifi.dbcp.hive.HiveConnectionPool/index.html">HiveConnectionPool</a></td><td
 id="description">The Hive Controller Service that is used to obtain 
connection(s) to the Hive database</td></tr><tr><td id="name"><strong>Batch 
Size</strong></td><td id="default-value">100</td><td 
id="allowable-values"></td><td id="description">The preferred number of 
FlowFiles to put to the database in a single transaction</td></tr><tr><td 
id="name"><strong>Character Set</strong></td><td 
id="default-value">UTF-8</td><td id="allowable-values"></td><td 
id="description">Specifies the character set of the record 
data.</td></tr></table><h3>R
 elationships: </h3><table 
id="relationships"><tr><th>Name</th><th>Description</th></tr><tr><td>retry</td><td>A
 FlowFile is routed to this relationship if the database cannot be updated but 
attempting the operation again may succeed</td></tr><tr><td>success</td><td>A 
FlowFile is routed to this relationship after the database is successfully 
updated</td></tr><tr><td>failure</td><td>A FlowFile is routed to this 
relationship if the database cannot be updated and retrying the operation will 
also fail, such as an invalid query or an integrity constraint 
violation</td></tr></table><h3>Reads Attributes: </h3><table 
id="reads-attributes"><tr><th>Name</th><th>Description</th></tr><tr><td>hiveql.args.N.type</td><td>Incoming
 FlowFiles are expected to be parametrized HiveQL statements. The type of each 
Parameter is specified as an integer that represents the JDBC Type of the 
parameter.</td></tr><tr><td>hiveql.args.N.value</td><td>Incoming FlowFiles are 
expected to be parametrized HiveQL statemen
 ts. The value of the Parameters are specified as hiveql.args.1.value, 
hiveql.args.2.value, hiveql.args.3.value, and so on. The type of the 
hiveql.args.1.value Parameter is specified by the hiveql.args.1.type 
attribute.</td></tr></table><h3>Writes Attributes: </h3>None 
specified.<h3>State management: </h3>This processor has no state 
management.<h3>Restricted: </h3><h3>See Also:</h3><p><a 
href="../org.apache.nifi.processors.hive.SelectHiveQL/index.html">SelectHiveQL</a></p></body></html>
\ No newline at end of file


Reply via email to