Author: navis
Date: Wed Dec 18 07:44:22 2013
New Revision: 1551876

URL: http://svn.apache.org/r1551876
Log:
HIVE-5879 : Fix spelling errors in hive-default.xml.template (Lefty Leverenz 
via Navis)

Modified:
    hive/trunk/conf/hive-default.xml.template

Modified: hive/trunk/conf/hive-default.xml.template
URL: 
http://svn.apache.org/viewvc/hive/trunk/conf/hive-default.xml.template?rev=1551876&r1=1551875&r2=1551876&view=diff
==============================================================================
--- hive/trunk/conf/hive-default.xml.template (original)
+++ hive/trunk/conf/hive-default.xml.template Wed Dec 18 07:44:22 2013
@@ -30,7 +30,7 @@
   <value>-1</value>
     <description>The default number of reduce tasks per job.  Typically set
   to a prime close to the number of available hosts.  Ignored when
-  mapred.job.tracker is "local". Hadoop set this to 1 by default, whereas hive 
uses -1 as its default value.
+  mapred.job.tracker is "local". Hadoop set this to 1 by default, whereas Hive 
uses -1 as its default value.
   By setting this property to -1, Hive will automatically figure out what 
should be the number of reducers.
   </description>
 </property>
@@ -46,7 +46,7 @@
   <value>999</value>
   <description>max number of reducers will be used. If the one
        specified in the configuration parameter mapred.reduce.tasks is
-       negative, hive will use this one as the max number of reducers when
+       negative, Hive will use this one as the max number of reducers when
        automatically determine number of reducers.</description>
 </property>
 
@@ -59,15 +59,15 @@
 <property>
   <name>hive.cli.print.current.db</name>
   <value>false</value>
-  <description>Whether to include the current database in the hive 
prompt.</description>
+  <description>Whether to include the current database in the Hive 
prompt.</description>
 </property>
 
 <property>
   <name>hive.cli.prompt</name>
   <value>hive</value>
   <description>Command line prompt configuration value. Other hiveconf can be 
used in
-        this configuration value. Variable substitution will only be invoked 
at the hive
-        cli startup.</description>
+        this configuration value. Variable substitution will only be invoked 
at the Hive
+        CLI startup.</description>
 </property>
 
 <property>
@@ -75,7 +75,7 @@
   <value>-1</value>
   <description>The number of columns to use when formatting output generated
         by the DESCRIBE PRETTY table_name command.  If the value of this 
property
-        is -1, then hive will use the auto-detected terminal 
width.</description>
+        is -1, then Hive will use the auto-detected terminal 
width.</description>
 </property>
 
 <property>
@@ -93,16 +93,16 @@
 <property>
   <name>hive.test.mode</name>
   <value>false</value>
-  <description>whether hive is running in test mode. If yes, it turns on 
sampling and prefixes the output tablename</description>
+  <description>Whether Hive is running in test mode. If yes, it turns on 
sampling and prefixes the output tablename.</description>
 </property>
 
 <property>
   <name>hive.test.mode.prefix</name>
   <value>test_</value>
-  <description>if hive is running in test mode, prefixes the output table by 
this string</description>
+  <description>if Hive is running in test mode, prefixes the output table by 
this string</description>
 </property>
 
-<!-- If the input table is not bucketed, the denominator of the tablesample is 
determinied by the parameter below   -->
+<!-- If the input table is not bucketed, the denominator of the tablesample is 
determined by the parameter below   -->
 <!-- For example, the following query:                                         
                                     -->
 <!--   INSERT OVERWRITE TABLE dest                                             
                                     -->
 <!--   SELECT col1 from src                                                    
                                     -->
@@ -112,19 +112,19 @@
 <property>
   <name>hive.test.mode.samplefreq</name>
   <value>32</value>
-  <description>if hive is running in test mode and table is not bucketed, 
sampling frequency</description>
+  <description>if Hive is running in test mode and table is not bucketed, 
sampling frequency</description>
 </property>
 
 <property>
   <name>hive.test.mode.nosamplelist</name>
   <value></value>
-  <description>if hive is running in test mode, dont sample the above comma 
seperated list of tables</description>
+  <description>if Hive is running in test mode, don't sample the above comma 
separated list of tables</description>
 </property>
 
 <property>
   <name>hive.metastore.uris</name>
   <value></value>
-  <description>Thrift uri for the remote metastore. Used by metastore client 
to connect to remote metastore.</description>
+  <description>Thrift URI for the remote metastore. Used by metastore client 
to connect to remote metastore.</description>
 </property>
 
 <property>
@@ -226,7 +226,7 @@
 <property>
   <name>datanucleus.cache.level2</name>
   <value>false</value>
-  <description>Use a level 2 cache. Turn this off if metadata is changed 
independently of hive metastore server</description>
+  <description>Use a level 2 cache. Turn this off if metadata is changed 
independently of Hive metastore server</description>
 </property>
 
 <property>
@@ -262,13 +262,13 @@
 <property>
   <name>hive.metastore.event.listeners</name>
   <value></value>
-  <description>list of comma seperated listeners for metastore 
events.</description>
+  <description>list of comma separated listeners for metastore 
events.</description>
 </property>
 
 <property>
   <name>hive.metastore.partition.inherit.table.properties</name>
   <value></value>
-  <description>list of comma seperated keys occurring in table properties 
which will get inherited to newly created partitions. * implies all the keys 
will get inherited.</description>
+  <description>list of comma separated keys occurring in table properties 
which will get inherited to newly created partitions. * implies all the keys 
will get inherited.</description>
 </property>
 
 <property>
@@ -294,7 +294,7 @@
   <value></value>
   <description>If true (default is false), ALTER TABLE operations which change 
the type of 
     a column (say STRING) to an incompatible type (say MAP&lt;STRING, 
STRING&gt;) are disallowed.  
-    RCFile default serde (ColumnarSerde) serializes the values in such a way 
that the
+    RCFile default SerDe (ColumnarSerDe) serializes the values in such a way 
that the
     datatypes can be converted from string to any type. The map is also 
serialized as
     a string, which can be read as a string as well. However, with any binary 
     serialization, this is not true. Blocking the ALTER TABLE prevents 
ClassCastExceptions
@@ -376,7 +376,7 @@
 <property>
   <name>hive.default.rcfile.serde</name>
   <value>org.apache.hadoop.hive.serde2.columnar.LazyBinaryColumnarSerDe</value>
-  <description>The default SerDe hive will use for the rcfile 
format</description>
+  <description>The default SerDe Hive will use for the RCFile 
format</description>
 </property>
 
 <property>
@@ -411,7 +411,7 @@
     perform the 2 groups bys. This makes sense if map-side aggregation is 
turned off. However,
     with maps-side aggregation, it might be useful in some cases to treat the 
2 inserts independently, 
     thereby performing the query above in 2MR jobs instead of 3 (due to 
spraying by distinct key first).
-    If this parameter is turned off, we dont consider the fact that the 
distinct key is the same across
+    If this parameter is turned off, we don't consider the fact that the 
distinct key is the same across
     different MR jobs.
   </description>
 </property>
@@ -449,7 +449,7 @@
 <property>
   <name>hive.session.history.enabled</name>
   <value>false</value>
-  <description>Whether to log hive query, query plan, runtime statistics etc 
</description>
+  <description>Whether to log Hive query, query plan, runtime statistics 
etc.</description>
 </property>
 
 <property>
@@ -505,7 +505,7 @@
     a union is performed for the 2 joins generated above. So unless the same 
skewed key is present
     in both the joined tables, the join for the skewed key will be performed 
as a map-side join.
 
-    The main difference between this paramater and hive.optimize.skewjoin is 
that this parameter
+    The main difference between this parameter and hive.optimize.skewjoin is 
that this parameter
     uses the skew information stored in the metastore to optimize the plan at 
compile time itself.
     If there is no skew information in the metadata, this parameter will not 
have any affect.
     Both hive.optimize.skewjoin.compiletime and hive.optimize.skewjoin should 
be set to true.
@@ -529,19 +529,19 @@
     The merge is triggered if either of hive.merge.mapfiles or 
hive.merge.mapredfiles is set to true.
     If the user has set hive.merge.mapfiles to true and hive.merge.mapredfiles 
to false, the idea was the
     number of reducers are few, so the number of files anyway are small. 
However, with this optimization,
-    we are increasing the number of files possibly by a big margin. So, we 
merge aggresively.</description>
+    we are increasing the number of files possibly by a big margin. So, we 
merge aggressively.</description>
 </property>
 
 <property>
   <name>hive.mapred.supports.subdirectories</name>
   <value>false</value>
-  <description>Whether the version of hadoop which is running supports 
sub-directories for tables/partitions.
-    Many hive optimizations can be applied if the hadoop version supports 
sub-directories for
+  <description>Whether the version of Hadoop which is running supports 
sub-directories for tables/partitions.
+    Many Hive optimizations can be applied if the Hadoop version supports 
sub-directories for
     tables/partitions. It was added by MAPREDUCE-1501</description>
 </property>
 
 <property>
-  <name>hive.multigroupby.singlemr</name>
+  <name>hive.multigroupby.singlereducer</name>
   <value>false</value>
   <description>Whether to optimize multi group by query to generate single M/R
   job plan. If the multi group by query has common group by keys, it will be
@@ -576,16 +576,16 @@
     This can lead to explosion across map-reduce boundary if the cardinality 
of T is very high,
     and map-side aggregation does not do a very good job. 
 
-    This parameter decides if hive should add an additional map-reduce job. If 
the grouping set
+    This parameter decides if Hive should add an additional map-reduce job. If 
the grouping set
     cardinality (4 in the example above), is more than this value, a new MR 
job is added under the
-    assumption that the orginal group by will reduce the data size.
+    assumption that the original group by will reduce the data size.
   </description>
 </property>
 
 <property>
   <name>hive.join.emit.interval</name>
   <value>1000</value>
-  <description>How many rows in the right-most join operand Hive should buffer 
before emitting the join result. </description>
+  <description>How many rows in the right-most join operand Hive should buffer 
before emitting the join result.</description>
 </property>
 
 <property>
@@ -605,7 +605,7 @@
   <value>false</value>
   <description>Whether to enable skew join optimization.
     The algorithm is as follows: At runtime, detect the keys with a large 
skew. Instead of
-    processing those keys, store them temporarily in a hdfs directory. In a 
follow-up map-reduce
+    processing those keys, store them temporarily in an HDFS directory. In a 
follow-up map-reduce
     job, process those skewed keys. The same key need not be skewed for all 
the tables, and so,
     the follow-up map-reduce job (for the skewed keys) would be much faster, 
since it would be a
     map-join.
@@ -639,7 +639,7 @@
 <property>
   <name>hive.mapred.mode</name>
   <value>nonstrict</value>
-  <description>The mode in which the hive operations are being performed.
+  <description>The mode in which the Hive operations are being performed.
      In strict mode, some risky queries are not allowed to run. They include:
        Cartesian Product.
        No partition being picked up for a query.
@@ -653,7 +653,7 @@
   <name>hive.enforce.bucketmapjoin</name>
   <value>false</value>
   <description>If the user asked for bucketed map-side join, and it cannot be 
performed,
-    should the query fail or not ? For eg, if the buckets in the tables being 
joined are
+    should the query fail or not ? For example, if the buckets in the tables 
being joined are
     not a multiple of each other, bucketed map-side join cannot be performed, 
and the
     query will fail if hive.enforce.bucketmapjoin is set to true.
   </description>
@@ -688,13 +688,13 @@
 <property>
   <name>hive.exec.compress.output</name>
   <value>false</value>
-  <description> This controls whether the final outputs of a query (to a 
local/hdfs file or a hive table) is compressed. The compression codec and other 
options are determined from hadoop config variables mapred.output.compress* 
</description>
+  <description> This controls whether the final outputs of a query (to a 
local/HDFS file or a Hive table) is compressed. The compression codec and other 
options are determined from Hadoop config variables mapred.output.compress* 
</description>
 </property>
 
 <property>
   <name>hive.exec.compress.intermediate</name>
   <value>false</value>
-  <description> This controls whether intermediate files produced by hive 
between multiple map-reduce jobs are compressed. The compression codec and 
other options are determined from hadoop config variables 
mapred.output.compress* </description>
+  <description> This controls whether intermediate files produced by Hive 
between multiple map-reduce jobs are compressed. The compression codec and 
other options are determined from Hadoop config variables 
mapred.output.compress* </description>
 </property>
 
 <property>
@@ -718,7 +718,7 @@
 <property>
   <name>hive.counters.group.name</name>
   <value>HIVE</value>
-  <description>The name of counter group for internal hive 
variables(CREATED_FILE, FATAL_ERROR, etc.)</description>
+  <description>The name of counter group for internal Hive variables 
(CREATED_FILE, FATAL_ERROR, etc.)</description>
 </property>
 
 <property>
@@ -760,7 +760,7 @@
 <property>
   <name>hive.metastore.init.hooks</name>
   <value></value>
-  <description>A comma separated list of hooks to be invoked at the beginning 
of HMSHandler initialization. Aninit hook is specified as the name of Java 
class which extends 
org.apache.hadoop.hive.metastore.MetaStoreInitListener.</description>
+  <description>A comma separated list of hooks to be invoked at the beginning 
of HMSHandler initialization. An init hook is specified as the name of Java 
class which extends 
org.apache.hadoop.hive.metastore.MetaStoreInitListener.</description>
 </property>
 
 <property>
@@ -820,13 +820,13 @@
 <property>
   <name>hive.mapjoin.localtask.max.memory.usage</name>
   <value>0.90</value>
-  <description>This number means how much memory the local task can take to 
hold the key/value into in-memory hash table; If the local task's memory usage 
is more than this number, the local task will be abort by themself. It means 
the data of small table is too large to be hold in the memory.</description>
+  <description>This number means how much memory the local task can take to 
hold the key/value into an in-memory hash table. If the local task's memory 
usage is more than this number, the local task will abort by itself. It means 
the data of the small table is too large to be held in memory.</description>
 </property>
 
 <property>
   <name>hive.mapjoin.followby.gby.localtask.max.memory.usage</name>
   <value>0.55</value>
-  <description>This number means how much memory the local task can take to 
hold the key/value into in-memory hash table when this map join followed by a 
group by; If the local task's memory usage is more than this number, the local 
task will be abort by themself. It means the data of small table is too large 
to be hold in the memory.</description>
+  <description>This number means how much memory the local task can take to 
hold the key/value into an in-memory hash table when this map join is followed 
by a group by. If the local task's memory usage is more than this number, the 
local task will abort by itself. It means the data of the small table is too 
large to be held in memory.</description>
 </property>
 
 <property>
@@ -838,14 +838,14 @@
 <property>
   <name>hive.auto.convert.join</name>
   <value>false</value>
-  <description>Whether Hive enable the optimization about converting common 
join into mapjoin based on the input file size</description>
+  <description>Whether Hive enables the optimization about converting common 
join into mapjoin based on the input file size</description>
 </property>
 
 <property>
   <name>hive.auto.convert.join.noconditionaltask</name>
   <value>true</value>
-  <description>Whether Hive enable the optimization about converting common 
join into mapjoin based on the input file 
-    size. If this paramater is on, and the sum of size for n-1 of the 
tables/partitions for a n-way join is smaller than the
+  <description>Whether Hive enables the optimization about converting common 
join into mapjoin based on the input file 
+    size. If this parameter is on, and the sum of size for n-1 of the 
tables/partitions for a n-way join is smaller than the
     specified size, the join is directly converted to a mapjoin (there is no 
conditional task).
   </description>
 </property>
@@ -862,13 +862,13 @@
 <property>
   <name>hive.script.auto.progress</name>
   <value>false</value>
-  <description>Whether Hive Tranform/Map/Reduce Clause should automatically 
send progress information to TaskTracker to avoid the task getting killed 
because of inactivity.  Hive sends progress information when the script is 
outputting to stderr.  This option removes the need of periodically producing 
stderr messages, but users should be cautious because this may prevent infinite 
loops in the scripts to be killed by TaskTracker.  </description>
+  <description>Whether Hive Transform/Map/Reduce Clause should automatically 
send progress information to TaskTracker to avoid the task getting killed 
because of inactivity.  Hive sends progress information when the script is 
outputting to stderr.  This option removes the need of periodically producing 
stderr messages, but users should be cautious because this may prevent infinite 
loops in the scripts to be killed by TaskTracker. </description>
 </property>
 
 <property>
   <name>hive.script.serde</name>
   <value>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</value>
-  <description>The default serde for trasmitting input data to and reading 
output data from the user scripts. </description>
+  <description>The default SerDe for transmitting input data to and reading 
output data from the user scripts. </description>
 </property>
 
 <property>
@@ -917,7 +917,7 @@
 <property>
   <name>stream.stderr.reporter.prefix</name>
   <value>reporter:</value>
-  <description>Streaming jobs that log to stardard error with this prefix can 
log counter or status information.</description>
+  <description>Streaming jobs that log to standard error with this prefix can 
log counter or status information.</description>
 </property>
 
 <property>
@@ -941,7 +941,7 @@
 <property>
   <name>hive.udtf.auto.progress</name>
   <value>false</value>
-  <description>Whether Hive should automatically send progress information to 
TaskTracker when using UDTF's to prevent the task getting killed because of 
inactivity.  Users should be cautious because this may prevent TaskTracker from 
killing tasks with infinte loops.  </description>
+  <description>Whether Hive should automatically send progress information to 
TaskTracker when using UDTF's to prevent the task getting killed because of 
inactivity.  Users should be cautious because this may prevent TaskTracker from 
killing tasks with infinite loops.  </description>
 </property>
 
 <property>
@@ -1003,7 +1003,7 @@
 <property>
   <name>hive.optimize.bucketingsorting</name>
   <value>true</value>
-  <description>If hive.enforce.bucketing or hive.enforce.sorting is true, dont 
create a reducer for enforcing
+  <description>If hive.enforce.bucketing or hive.enforce.sorting is true, 
don't create a reducer for enforcing
     bucketing/sorting for queries of the form: 
     insert overwrite table T2 select * from T1;
     where T1 and T2 are bucketed/sorted by the same keys into the same number 
of buckets.
@@ -1045,10 +1045,10 @@
   <name>hive.auto.convert.sortmerge.join.to.mapjoin</name>
   <value>false</value>
   <description>If hive.auto.convert.sortmerge.join is set to true, and a join 
was converted to a sort-merge join,
-    this parameter decides whether each table should be tried as a big table, 
and effectviely a map-join should be
+    this parameter decides whether each table should be tried as a big table, 
and effectively a map-join should be
     tried. That would create a conditional task with n+1 children for a n-way 
join (1 child for each table as the
-    big table), and the backup task will be the sort-merge join. In some 
casess, a map-join would be faster than a
-    sort-merge join, if there is no advantage of having the output bucketed 
and sorted. For eg. if a very big sorted
+    big table), and the backup task will be the sort-merge join. In some 
cases, a map-join would be faster than a
+    sort-merge join, if there is no advantage of having the output bucketed 
and sorted. For example, if a very big sorted
     and bucketed table with few files (say 10 files) are being joined with a 
very small sorter and bucketed table
     with few files (10 files), the sort-merge join will only use 10 mappers, 
and a simple map-only join might be faster
     if the complete small table can fit in memory, and a map-join can be 
performed.
@@ -1058,7 +1058,7 @@
 <property>
   <name>hive.metastore.ds.connection.url.hook</name>
   <value></value>
-  <description>Name of the hook to use for retriving the JDO connection URL. 
If empty, the value in javax.jdo.option.ConnectionURL is used </description>
+  <description>Name of the hook to use for retrieving the JDO connection URL. 
If empty, the value in javax.jdo.option.ConnectionURL is used </description>
 </property>
 
 <property>
@@ -1070,7 +1070,7 @@
 <property>
    <name>hive.metastore.ds.retry.interval</name>
    <value>1000</value>
-   <description>The number of miliseconds between metastore retry 
attempts</description>
+   <description>The number of milliseconds between metastore retry 
attempts</description>
 </property>
 
 <property>
@@ -1094,25 +1094,25 @@
 <property>
   <name>hive.metastore.sasl.enabled</name>
   <value>false</value>
-  <description>If true, the metastore thrift interface will be secured with 
SASL. Clients must authenticate with Kerberos.</description>
+  <description>If true, the metastore Thrift interface will be secured with 
SASL. Clients must authenticate with Kerberos.</description>
 </property>
 
 <property>
   <name>hive.metastore.thrift.framed.transport.enabled</name>
   <value>false</value>
-  <description>If true, the metastore thrift interface will use 
TFramedTransport. When false (default) a standard TTransport is 
used.</description>
+  <description>If true, the metastore Thrift interface will use 
TFramedTransport. When false (default) a standard TTransport is 
used.</description>
 </property>
 
 <property>
   <name>hive.metastore.kerberos.keytab.file</name>
   <value></value>
-  <description>The path to the Kerberos Keytab file containing the metastore 
thrift server's service principal.</description>
+  <description>The path to the Kerberos Keytab file containing the metastore 
Thrift server's service principal.</description>
 </property>
 
 <property>
   <name>hive.metastore.kerberos.principal</name>
   <value>hive-metastore/[email protected]</value>
-  <description>The service principal for the metastore thrift server. The 
special string _HOST will be replaced automatically with the correct host 
name.</description>
+  <description>The service principal for the metastore Thrift server. The 
special string _HOST will be replaced automatically with the correct host 
name.</description>
 </property>
 
 <property>
@@ -1198,13 +1198,13 @@
 <property>
   <name>hive.exec.default.partition.name</name>
   <value>__HIVE_DEFAULT_PARTITION__</value>
-  <description>The default partition name in case the dynamic partition column 
value is null/empty string or anyother values that cannot be escaped. This 
value must not contain any special character used in HDFS URI (e.g., ':', '%', 
'/' etc). The user has to be aware that the dynamic partition value should not 
contain this value to avoid confusions.</description>
+  <description>The default partition name in case the dynamic partition column 
value is null/empty string or any other values that cannot be escaped. This 
value must not contain any special character used in HDFS URI (e.g., ':', '%', 
'/' etc). The user has to be aware that the dynamic partition value should not 
contain this value to avoid confusions.</description>
 </property>
 
 <property>
   <name>hive.stats.dbclass</name>
   <value>counter</value>
-  <description>The storage that stores temporary hive statistics. Currently, 
jdbc, hbase, counter and custom type is supported</description>
+  <description>The storage that stores temporary Hive statistics. Currently, 
jdbc, hbase, counter and custom type are supported.</description>
 </property>
 
 <property>
@@ -1216,13 +1216,13 @@
 <property>
   <name>hive.stats.jdbcdriver</name>
   <value>org.apache.derby.jdbc.EmbeddedDriver</value>
-  <description>The JDBC driver for the database that stores temporary hive 
statistics.</description>
+  <description>The JDBC driver for the database that stores temporary Hive 
statistics.</description>
 </property>
 
 <property>
   <name>hive.stats.dbconnectionstring</name>
   <value>jdbc:derby:;databaseName=TempStatsStore;create=true</value>
-  <description>The default connection string for the database that stores 
temporary hive statistics.</description>
+  <description>The default connection string for the database that stores 
temporary Hive statistics.</description>
 </property>
 
 <property>
@@ -1252,14 +1252,14 @@
 <property>
   <name>hive.stats.retries.wait</name>
   <value>3000</value>
-  <description>The base waiting window (in milliseconds) before the next 
retry. The actual wait time is calculated by baseWindow * failues  baseWindow * 
(failure  1) * (random number between [0.0,1.0]).</description>
+  <description>The base waiting window (in milliseconds) before the next 
retry. The actual wait time is calculated by baseWindow * failures  baseWindow 
* (failure  1) * (random number between [0.0,1.0]).</description>
 </property>
 
 <property>
   <name>hive.stats.reliable</name>
   <value>false</value>
   <description>Whether queries will fail because stats cannot be collected 
completely accurately.
-    If this is set to true, reading/writing from/into a partition may fail 
becuase the stats
+    If this is set to true, reading/writing from/into a partition may fail 
because the stats
     could not be computed accurately.
   </description>
 </property>
@@ -1301,7 +1301,7 @@
 <property>
   <name>hive.support.concurrency</name>
   <value>false</value>
-  <description>Whether hive supports concurrency or not. A zookeeper instance 
must be up and running for the default hive lock manager to support read-write 
locks.</description>
+  <description>Whether Hive supports concurrency or not. A ZooKeeper instance 
must be up and running for the default Hive lock manager to support read-write 
locks.</description>
 </property>
 
 <property>
@@ -1325,25 +1325,25 @@
 <property>
   <name>hive.zookeeper.quorum</name>
   <value></value>
-  <description>The list of zookeeper servers to talk to. This is only needed 
for read/write locks.</description>
+  <description>The list of ZooKeeper servers to talk to. This is only needed 
for read/write locks.</description>
 </property>
 
 <property>
   <name>hive.zookeeper.client.port</name>
   <value>2181</value>
-  <description>The port of zookeeper servers to talk to. This is only needed 
for read/write locks.</description>
+  <description>The port of ZooKeeper servers to talk to. This is only needed 
for read/write locks.</description>
 </property>
 
 <property>
   <name>hive.zookeeper.session.timeout</name>
   <value>600000</value>
-  <description>Zookeeper client's session timeout. The client is disconnected, 
and as a result, all locks released, if a heartbeat is not sent in the 
timeout.</description>
+  <description>ZooKeeper client's session timeout. The client is disconnected, 
and as a result, all locks released, if a heartbeat is not sent in the 
timeout.</description>
 </property>
 
 <property>
   <name>hive.zookeeper.namespace</name>
   <value>hive_zookeeper_namespace</value>
-  <description>The parent node under which all zookeeper nodes are 
created.</description>
+  <description>The parent node under which all ZooKeeper nodes are 
created.</description>
 </property>
 
 <property>
@@ -1355,7 +1355,7 @@
 <property>
   <name>fs.har.impl</name>
   <value>org.apache.hadoop.hive.shims.HiveHarFileSystem</value>
-  <description>The implementation for accessing Hadoop Archives. Note that 
this won't be applicable to Hadoop vers less than 0.20</description>
+  <description>The implementation for accessing Hadoop Archives. Note that 
this won't be applicable to Hadoop versions less than 0.20</description>
 </property>
 
 <property>
@@ -1367,13 +1367,13 @@
 <property>
   <name>hive.fetch.output.serde</name>
   <value>org.apache.hadoop.hive.serde2.DelimitedJSONSerDe</value>
-  <description>The serde used by FetchTask to serialize the fetch 
output.</description>
+  <description>The SerDe used by FetchTask to serialize the fetch 
output.</description>
 </property>
 
 <property>
   <name>hive.exec.mode.local.auto</name>
   <value>false</value>
-  <description> Let hive determine whether to run in local mode automatically 
</description>
+  <description> Let Hive determine whether to run in local mode automatically 
</description>
 </property>
 
 <property>
@@ -1443,19 +1443,19 @@
 <property>
   <name>hive.conf.validation</name>
   <value>true</value>
-  <description>Eables type checking for registered hive 
configurations</description>
+  <description>Enables type checking for registered Hive 
configurations</description>
 </property>
 
 <property>
   <name>hive.security.authorization.enabled</name>
   <value>false</value>
-  <description>enable or disable the hive client authorization</description>
+  <description>enable or disable the Hive client authorization</description>
 </property>
 
 <property>
   <name>hive.security.authorization.manager</name>
   
<value>org.apache.hadoop.hive.ql.security.authorization.DefaultHiveAuthorizationProvider</value>
-  <description>the hive client authorization manager class name.
+  <description>The Hive client authorization manager class name.
   The user defined authorization class should implement interface 
org.apache.hadoop.hive.ql.security.authorization.HiveAuthorizationProvider.
   </description>
 </property>
@@ -1516,13 +1516,13 @@
 <property>
   <name>hive.security.command.whitelist</name>
   <value>set,reset,dfs,add,delete</value>
-  <description>Comma seperated list of non-SQL Hive commands users are 
authorized to execute</description>
+  <description>Comma separated list of non-SQL Hive commands users are 
authorized to execute</description>
 </property>
 
 <property>
   <name>hive.conf.restricted.list</name>
   <value></value>
-  <description>Comma seperated list of configuration options which are 
immutable at runtime</description>
+  <description>Comma separated list of configuration options which are 
immutable at runtime</description>
 </property>
 
 <property>
@@ -1537,13 +1537,13 @@
 <property>
   <name>hive.error.on.empty.partition</name>
   <value>false</value>
-  <description>Whether to throw an excpetion if dynamic partition insert 
generates empty results.</description>
+  <description>Whether to throw an exception if dynamic partition insert 
generates empty results.</description>
 </property>
 
 <property>
   <name>hive.index.compact.file.ignore.hdfs</name>
   <value>false</value>
-  <description>True the hdfs location stored in the index file will be igbored 
at runtime.
+  <description>When true the HDFS location stored in the index file will be 
ignored at runtime.
   If the data got moved or the name of the cluster got changed, the index data 
should still be usable.</description>
 </property>
 
@@ -1634,7 +1634,7 @@
 <property>
   <name>hive.exec.concatenate.check.index</name>
   <value>true</value>
-  <description>If this sets to true, hive will throw error when doing
+  <description>If this is set to true, Hive will throw error when doing
    'alter table tbl_name [partSpec] concatenate' on a table/partition
     that has indexes on it. The reason the user want to set this to true
     is because it can help user to avoid handling all index drop, recreation,
@@ -1666,7 +1666,7 @@
 <property>
   <name>hive.autogen.columnalias.prefix.includefuncname</name>
   <value>false</value>
-  <description>Whether to include function name in the column alias auto 
generated by hive.</description>
+  <description>Whether to include function name in the column alias auto 
generated by Hive.</description>
 </property>
 
 <property>
@@ -1678,7 +1678,7 @@
 <property>
   <name>hive.start.cleanup.scratchdir</name>
   <value>false</value>
-  <description>To cleanup the hive scratchdir while starting the hive 
server</description>
+  <description>To cleanup the Hive scratchdir while starting the Hive 
Server</description>
 </property>
 
 <property>
@@ -1713,8 +1713,9 @@
 <property>
   <name>hive.exec.driver.run.hooks</name>
   <value></value>
-  <description>A comma separated list of hooks which implement 
HiveDriverRunHook and will be run at the
-                          beginning and end of Driver.run, these will be run 
in the order specified
+  <description>A comma separated list of hooks which implement 
HiveDriverRunHook
+    and will be run at the beginning and end of Driver.run, these will be run 
in
+    the order specified.
   </description>
 </property>
 
@@ -1732,7 +1733,7 @@
   <value>false</value>
   <description>
     This adds an option to escape special chars (newlines, carriage returns and
-    tabs) when they are passed to the user script. This is useful if the hive 
tables
+    tabs) when they are passed to the user script. This is useful if the Hive 
tables
     can contain data that contains special characters.
   </description>
 </property>
@@ -1741,9 +1742,9 @@
   <name>hive.exec.rcfile.use.explicit.header</name>
   <value>true</value>
   <description>
-    If this is set the header for RC Files will simply be RCF.  If this is not
+    If this is set the header for RCFiles will simply be RCF.  If this is not
     set the header will be that borrowed from sequence files, e.g. SEQ- 
followed
-    by the input and output RC File formats.
+    by the input and output RCFile formats.
   </description>
 </property>
 
@@ -1779,7 +1780,7 @@
   <description>
     Some select queries can be converted to single FETCH task minimizing 
latency.
     Currently the query should be single sourced not having any subquery and 
should not have
-    any aggregations or distincts (which incurrs RS), lateral views and joins.
+    any aggregations or distincts (which incurs RS), lateral views and joins.
     1. minimal : SELECT STAR, FILTER on partition columns, LIMIT only
     2. more    : SELECT, FILTER, LIMIT only (TABLESAMPLE, virtual columns)
   </description>
@@ -1799,8 +1800,8 @@
   <name>hive.fetch.task.aggr</name>
   <value>false</value>
   <description>
-    Aggregation queries with no group-by clause (for example, select count(*) 
from src) executes
-    final aggregations in single reduce task. If this is set true, hive 
delegates final aggregation
+    Aggregation queries with no group-by clause (for example, select count(*) 
from src) execute
+    final aggregations in single reduce task. If this is set true, Hive 
delegates final aggregation
     stage to fetch task, possibly decreasing the query time.
   </description>
 </property>
@@ -1826,7 +1827,7 @@
 <property>
    <name>hive.hmshandler.retry.interval</name>
    <value>1000</value>
-   <description>The number of miliseconds between HMSHandler retry 
attempts</description>
+   <description>The number of milliseconds between HMSHandler retry 
attempts</description>
 </property>
 
 <property>
@@ -1838,7 +1839,7 @@
 <property>
    <name>hive.server.tcp.keepalive</name>
    <value>true</value>
-   <description>Whether to enable TCP keepalive for the Hive server. Keepalive 
will prevent accumulation of half-open connections.</description>
+   <description>Whether to enable TCP keepalive for the Hive Server. Keepalive 
will prevent accumulation of half-open connections.</description>
 </property>
 
 <property>
@@ -1977,13 +1978,13 @@
     must be a proper implementation of the interface
     org.apache.hive.service.auth.PasswdAuthenticationProvider. HiveServer2
     will call its Authenticate(user, passed) method to authenticate requests.
-    The implementation may optionally extend the Hadoop's
+    The implementation may optionally extend Hadoop's
     org.apache.hadoop.conf.Configured class to grab Hive's Configuration 
object.
   </description>
 </property>
 
 <property>
-  <name>>hive.server2.authentication.kerberos.principal</name>
+  <name>hive.server2.authentication.kerberos.principal</name>
   <value></value>
   <description>
     Kerberos server principal
@@ -1991,7 +1992,7 @@
 </property>
 
 <property>
-  <name>>hive.server2.authentication.kerberos.keytab</name>
+  <name>hive.server2.authentication.kerberos.keytab</name>
   <value></value>
   <description>
     Kerberos keytab file for server principal
@@ -2018,8 +2019,8 @@
   <name>hive.server2.enable.doAs</name>
   <value>true</value>
   <description>
-   Setting this property to true will have hive server2 execute
-    hive operations as the user making the calls to it.
+   Setting this property to true will have HiveServer2 execute
+    Hive operations as the user making the calls to it.
   </description>
 </property>
 
@@ -2027,9 +2028,9 @@
   <name>hive.server2.table.type.mapping</name>
   <value>CLASSIC</value>
   <description>
-   This setting reflects how HiveServer will report the table types for JDBC 
and other
-   client implementations that retrieves the available tables and supported 
table types
-     HIVE : Exposes the hive's native table tyes like MANAGED_TABLE, 
EXTERNAL_TABLE, VIRTUAL_VIEW
+   This setting reflects how HiveServer2 will report the table types for JDBC 
and other
+   client implementations that retrieve the available tables and supported 
table types
+     HIVE : Exposes Hive's native table types like MANAGED_TABLE, 
EXTERNAL_TABLE, VIRTUAL_VIEW
      CLASSIC : More generic types like TABLE and VIEW
   </description>
 </property>
@@ -2038,11 +2039,11 @@
   <name>hive.server2.thrift.sasl.qop</name>
   <value>auth</value>
   <description>Sasl QOP value; Set it to one of following values to enable 
higher levels of
-     protection for hive server2 communication with clients.
+     protection for HiveServer2 communication with clients.
       "auth" - authentication only (default)
       "auth-int" - authentication plus integrity protection
       "auth-conf" - authentication plus integrity and confidentiality 
protection
-     This is applicable only hive server2 is configured to use kerberos 
authentication.
+     This is applicable only if HiveServer2 is configured to use Kerberos 
authentication.
  </description>
 </property>
 
@@ -2086,7 +2087,7 @@
   <name>hive.compute.query.using.stats</name>
   <value>false</value>
   <description>
-  When set to true hive will answer few queries like count(1) purely using 
stats
+  When set to true Hive will answer a few queries like count(1) purely using 
stats
   stored in metastore. For basic stats collection turn on the config 
hive.stats.autogather to true.
   For more advanced stats collection need to run analyze table queries.
   </description>
@@ -2098,7 +2099,7 @@
    <description>
    Enforce metastore schema version consistency.
    True: Verify that version information stored in metastore matches with one 
from Hive jars.  Also disable automatic
-         schema migration attempt. Users are required to manully migrate 
schema after Hive upgrade which ensures
+         schema migration attempt. Users are required to manually migrate 
schema after Hive upgrade which ensures
          proper metastore schema migration. (Default)
    False: Warn if the version information stored in metastore doesn't match 
with one from in Hive jars.
    </description>


Reply via email to