Re: Fail to load table via Tez
Hi Rajesh, Thanks for your quick answer. Seems like we also have a problem with the logs, as even with aggreagation enabled we get the following : /app-logs/yarn/logs/application_1499426430661_0113 does not exist. Log aggregation has not completed or is not enabled. Still, I tried what you suggested and still got the same problem: INFO : Map 1: 255(+85,-31)/340 INFO : Map 1: 256(+84,-31)/340 INFO : Map 1: 257(+77,-33)/340 INFO : Map 1: 257(+0,-33)/340 ERROR : Status: Failed ERROR : Vertex failed, vertexName=Map 1, vertexId=vertex_1499426430661_0119_1_00, diagnostics=[Task failed, taskId=task_1499426430661_0119_1_00_000273, diagnostics=[TaskAttempt 0 failed, info=[Container container_e17_1499426430661_0119_01_000170 finished with diagnostics set to [Container failed, exitCode=-104. Container [pid=37464,containerID=container_e17_1499426430661_0119_01_000170] is running beyond physical memory limits. Current usage: 2.5 GB of 2.5 GB physical memory used; 4.4 GB of 5.3 GB virtual memory used. Killing container. Do you have another idea ? Regards, Loïc Loïc CHANEL System Big Data engineer MS - Worldline Analytics Platform - Worldline (Villeurbanne, France) 2017-07-07 16:51 GMT+02:00 Rajesh Balamohan: > You can run *"yarn logs -applicationId application_1499426430661_0113 > > application_1499426430661_**0113.log"* to get the app logs. > > Would suggest you to try with *"hive --hiveconf > tez.grouping.max-size=134217728 --hiveconf tez.grouping.min-size=** 134217728" > *for running your hive query. You may want to adjust this parameter (to > say 256 MB or so) in case too may mappers are created. > > ~Rajesh.B > > > On Fri, Jul 7, 2017 at 8:02 PM, Loïc Chanel > wrote: > >> Hi guys, >> >> I'm having some troubles with Tez when I try to load some data stored in >> small JSON files in HDFS into a Hive table. >> >> At first I got some Out of memory exceptions, so I tried increasing the >> amount of memory allocated to Tez, until the problem turned to a GC >> Overhead limit exceeded after 10 GB of RAM was allocated to Tez containers. >> >> So I upgraded my common sense and put back memory limits to a normal >> level, and now the problem I hit is the following : >> >> INFO : Map 1: 276(+63,-84)/339 >> INFO : Map 1: 276(+63,-85)/339 >> INFO : Map 1: 276(+63,-85)/339 >> INFO : Map 1: 276(+0,-86)/339 >> INFO : Map 1: 276(+0,-86)/339 >> ERROR : Status: Failed >> ERROR : Status: Failed >> ERROR : Vertex failed, vertexName=Map 1, >> vertexId=vertex_1499426430661_0113_1_00, >> diagnostics=[Task failed, taskId=task_1499426430661_0113_1_00_000241, >> diagnostics=[TaskAttempt 0 failed, info=[Container >> container_e17_1499426430661_0113_01_000170 finished with diagnostics set >> to [Container failed, exitCode=-104. Container >> [pid=59528,containerID=container_e17_1499426430661_0113_01_000170] is >> running beyond physical memory limits. Current usage: 2.7 GB of 2.5 GB >> physical memory used; 4.4 GB of 5.3 GB virtual memory used. Killing >> container. >> >> The problem is I can't see how the container could be allocated so much >> memory, and why can't Tez split the jobs into smaller ones when it fails >> for memory reasons. >> >> FYI, in YARN, Max container memory is 92160 MB, in MR2 Map can have 4GB >> and Reduce 5GB, Tez container size is set to 2560 MB and tez.grouping. >> max-size is set to 1073741824. >> >> If you need more information feel free to ask. >> >> I am currently running out of ideas on how to debug this as I have a >> limited access to Tez container logs, so any inputs will be highly >> appreciated. >> >> Thanks ! >> >> >> Loïc >> >> Loïc CHANEL >> System Big Data engineer >> MS - Worldline Analytics Platform - Worldline (Villeurbanne, France) >> > >
Re: Fail to load table via Tez
You can run *"yarn logs -applicationId application_1499426430661_0113 > application_1499426430661_**0113.log"* to get the app logs. Would suggest you to try with *"hive --hiveconf tez.grouping.max-size=134217728 --hiveconf tez.grouping.min-size=** 134217728" *for running your hive query. You may want to adjust this parameter (to say 256 MB or so) in case too may mappers are created. ~Rajesh.B On Fri, Jul 7, 2017 at 8:02 PM, Loïc Chanelwrote: > Hi guys, > > I'm having some troubles with Tez when I try to load some data stored in > small JSON files in HDFS into a Hive table. > > At first I got some Out of memory exceptions, so I tried increasing the > amount of memory allocated to Tez, until the problem turned to a GC > Overhead limit exceeded after 10 GB of RAM was allocated to Tez containers. > > So I upgraded my common sense and put back memory limits to a normal > level, and now the problem I hit is the following : > > INFO : Map 1: 276(+63,-84)/339 > INFO : Map 1: 276(+63,-85)/339 > INFO : Map 1: 276(+63,-85)/339 > INFO : Map 1: 276(+0,-86)/339 > INFO : Map 1: 276(+0,-86)/339 > ERROR : Status: Failed > ERROR : Status: Failed > ERROR : Vertex failed, vertexName=Map 1, > vertexId=vertex_1499426430661_0113_1_00, > diagnostics=[Task failed, taskId=task_1499426430661_0113_1_00_000241, > diagnostics=[TaskAttempt 0 failed, info=[Container > container_e17_1499426430661_0113_01_000170 finished with diagnostics set > to [Container failed, exitCode=-104. Container [pid=59528,containerID= > container_e17_1499426430661_0113_01_000170] is running beyond physical > memory limits. Current usage: 2.7 GB of 2.5 GB physical memory used; 4.4 GB > of 5.3 GB virtual memory used. Killing container. > > The problem is I can't see how the container could be allocated so much > memory, and why can't Tez split the jobs into smaller ones when it fails > for memory reasons. > > FYI, in YARN, Max container memory is 92160 MB, in MR2 Map can have 4GB > and Reduce 5GB, Tez container size is set to 2560 MB and tez.grouping. > max-size is set to 1073741824. > > If you need more information feel free to ask. > > I am currently running out of ideas on how to debug this as I have a > limited access to Tez container logs, so any inputs will be highly > appreciated. > > Thanks ! > > > Loïc > > Loïc CHANEL > System Big Data engineer > MS - Worldline Analytics Platform - Worldline (Villeurbanne, France) >
Fail to load table via Tez
Hi guys, I'm having some troubles with Tez when I try to load some data stored in small JSON files in HDFS into a Hive table. At first I got some Out of memory exceptions, so I tried increasing the amount of memory allocated to Tez, until the problem turned to a GC Overhead limit exceeded after 10 GB of RAM was allocated to Tez containers. So I upgraded my common sense and put back memory limits to a normal level, and now the problem I hit is the following : INFO : Map 1: 276(+63,-84)/339 INFO : Map 1: 276(+63,-85)/339 INFO : Map 1: 276(+63,-85)/339 INFO : Map 1: 276(+0,-86)/339 INFO : Map 1: 276(+0,-86)/339 ERROR : Status: Failed ERROR : Status: Failed ERROR : Vertex failed, vertexName=Map 1, vertexId=vertex_1499426430661_0113_1_00, diagnostics=[Task failed, taskId=task_1499426430661_0113_1_00_000241, diagnostics=[TaskAttempt 0 failed, info=[Container container_e17_1499426430661_0113_01_000170 finished with diagnostics set to [Container failed, exitCode=-104. Container [pid=59528,containerID=container_e17_1499426430661_0113_01_000170] is running beyond physical memory limits. Current usage: 2.7 GB of 2.5 GB physical memory used; 4.4 GB of 5.3 GB virtual memory used. Killing container. The problem is I can't see how the container could be allocated so much memory, and why can't Tez split the jobs into smaller ones when it fails for memory reasons. FYI, in YARN, Max container memory is 92160 MB, in MR2 Map can have 4GB and Reduce 5GB, Tez container size is set to 2560 MB and tez.grouping.max-size is set to 1073741824. If you need more information feel free to ask. I am currently running out of ideas on how to debug this as I have a limited access to Tez container logs, so any inputs will be highly appreciated. Thanks ! Loïc Loïc CHANEL System Big Data engineer MS - Worldline Analytics Platform - Worldline (Villeurbanne, France)