<configuration>
  <property>
      <name>tez.version</name>
          <value>0.4.0</value>
            </property>

  <property>
      <name>tez.lib.uris</name>
          <value>${fs.default.name}/apps/tez-0.4.0-incubating-full,${
fs.default.name}/apps/tez-0.4.0-incubating-full/lib/</value>
            </property>
                                      <!-- Client Submission timeout value
when submitting DAGs to a session -->
  <property>
      <name>tez.session.client.timeout.secs</name>
          <value>-1</value>
            </property>
              <!-- prewarm stuff -->
</configuration>



On Thu, May 15, 2014 at 12:11 PM, oracle cool <[email protected]> wrote:

> Hi
>
> We are running  Hadoop 2.2 , Hive 0.13 - Hive Server1.
>
>
> When we enabled tez 0.4 ( tables are in rcfile, no vectorization) &
> executed the query, we see major improvement in query but when we run
> concurrent 10 session & execute same query  ( select count(*) from tab ),
> One batch of 10 query executed but next batch hang after 3 query.
>
> -------------
> via hive server 1 jdbc connection - 10 connection which execute same query
> after 1 min.
>
> query execution time in first batch is 15 sec.
>
> set hive.execution.engine=tez;
> select count(*) from table1;
>
> ---------------
>
>
> yarn-site.xml
> -------------------
>
> <configuration>
>         <!-- Site specific YARN configuration properties -->
>         <property>
>                 <name>yarn.nodemanager.aux-services</name>
>                 <value>mapreduce_shuffle</value>
>         </property>
>         <property>
>
> <name>yarn.nodemanager.aux-services.mapreduce.shuffle.class</name>
>                 <value>org.apache.hadoop.mapred.ShuffleHandler</value>
>         </property>
>         <property>
>                 <name>yarn.resourcemanager.resource-tracker.address</name>
>                 <value>proj233:8025</value>
>         </property>
>         <property>
>                 <name>yarn.resourcemanager.scheduler.address</name>
>                 <value>proj233:8030</value>
>         </property>
>         <property>
>                 <name>yarn.resourcemanager.address</name>
>                 <value>proj233:8040</value>
>         </property>
>   <property>
>     <name>yarn.scheduler.minimum-allocation-mb</name>
>     <value>1024</value>
>       </property>
>   <property>
>     <name>yarn.scheduler.maximum-allocation-mb</name>
>     <value>40960</value>
>       </property>
>   <property>
>     <name>yarn.scheduler.minimum-allocation-vcores</name>
>     <value>1</value>
>       </property>
>   <property>
>     <name>yarn.scheduler.maximum-allocation-vcores</name>
>     <value>24</value>
>       </property>
>   <property>
>     <name>yarn.nodemanager.resource.memory-mb</name>
>     <value>40960</value>
>       </property>
>   <property>
>     <name>yarn.nodemanager.resource.cpu-vcores</name>
>     <value>1</value>
>       </property>
> <property>
>   <name>yarn.resourcemanager.scheduler.class</name>
>
> <value>org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacityScheduler</value>
> </property>
> </configuration>
>
>
> Mapred-site.xml
> ---------------------
> <configuration>
>
> <property>
>    <name>mapreduce.framework.name</name>
>       <value>yarn-tez</value>
> </property>
>
>  <property>
>     <name>yarn.nodemanager.vmem-pmem-ratio</name>
>     <value>2.1</value>
>       </property>
>   <property>
>     <name>mapreduce.task.io.sort.mb</name>
>     <value>32</value>
>       </property>
>   <property>
>         <name>yarn.app.mapreduce.am.resource.mb</name>
>         <value>1024</value>
>     </property>
>     <property>
>         <name>yarn.app.mapreduce.am.command-opts</name>
>         <value>-Xmx768m</value>
>     </property>
>     <property>
>         <name>mapreduce.map.cpu.vcores</name>
>         <value>1</value>
>             </property>
>     <property>
>         <name>mapreduce.reduce.cpu.vcores</name>
>         <value>1</value>
>             </property>
>     <property>
>         <name>mapreduce.map.memory.mb</name>
>         <value>1024</value>
>             </property>
>     <property>
>         <name>mapreduce.map.java.opts</name>
>         <value>-Xmx768m</value>
>             </property>
>     <property>
>         <name>mapreduce.reduce.memory.mb</name>
>         <value>1024</value>
>             </property>
>     <property>
>         <name>mapreduce.reduce.java.opts</name>
>         <value>-Xmx768m</value>
>             </property>
> </configuration>
>
>
> hdfs-site.xml
> --------------------
> <configuration>
>   <property>
>     <name>dfs.replication</name>
>     <value>3</value>
>   </property>
>   <property>
>     <name>dfs.namenode.name.dir</name>
>     <value>file:/Data0/yarn_data/hdfs/namenode</value>
>   </property>
>   <property>
>     <name>dfs.datanode.data.dir</name>
>
> <value>/Data0/yarn_data/hdfs/datanode,/Data1/yarn_data/hdfs/datanode,/Data3/yarn_data/hdfs/datanode,/Data4/yarn_data/hdfs/datanode</value>
>   </property>
>   <property>
>     <name>dfs.heartbeat.interval</name>
>     <value>3</value>
>         </description>
>   </property>
>
>   <property>
>     <name>dfs.safemode.threshold.pct</name>
>     <value>1.0f</value>
>   </property>
> <property>
>    <name>dfs.block.size</name>
>    <value>134217728</value>
> </property>
> <property>
>   <name>dfs.datanode.max.xcievers</name>
>   <value>16384</value>
> </property>
>
> <property>
>   <name>dfs.client.block.write.replace-datanode-on-failure.enable</name>
>   <value>false</value>
> </property>
> <property>
>   <name>dfs.datanode.handler.count</name>
>   <value>10</value>
> </property>
> <property>
>   <name>dfs.datanode.socket.write.timeout</name>
>   <value>0</value>
> </configuration>
>
>
>
> Thanks
> SS
>
>

Reply via email to