[ https://issues.apache.org/jira/browse/HIVE-457?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12792744#action_12792744 ]
Zheng Shao commented on HIVE-457: --------------------------------- This also means that we are limiting the row length for the data. Can we add a conf variable into TextRecord and set the default to maybe 10MB? In this way, nobody will notice the difference unless the row is bigger than 10MB, which is rare enough but is still smaller than typical memory size. {code} set hive.text.record.reader.max.length=10485760; {code} We can treat this one as an internal variable (by not adding it to hive-default.xml), following hadoop internal variable convention. > ScriptOperator should NOT cache all data in stderr > -------------------------------------------------- > > Key: HIVE-457 > URL: https://issues.apache.org/jira/browse/HIVE-457 > Project: Hadoop Hive > Issue Type: Bug > Reporter: Zheng Shao > Assignee: Paul Yang > Priority: Blocker > Fix For: 0.5.0 > > Attachments: err.sh, HIVE-457.1.patch > > > Sometimes user scripts output a lot of data to stderr without a new line, and > this causes Hive to go out-of-memory. > We should directly output the data from stderr without caching it. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.