[
https://issues.apache.org/jira/browse/HADOOP-11984?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14549616#comment-14549616
]
Allen Wittenauer commented on HADOOP-11984:
-------------------------------------------
bq. Is it worthwhile to trade the additional complexity and possibly reduced
robustness for the optimized execution time?
I don't find the current code complex at all compared to some of the gymnastics
happening in hadoop-functions.sh. I might have a different opinion if the
docker bits ever get committed.
bq. The UNIX world supports bash pretty well. However, Hadoop officially
supports Windows as well. How does the compatibility story work for Windows
folks? Unfortunately things like path separators are quite annoying to handle.
We've already done it for the client code in trunk. It's likely better than
the DOS batch code. My hunch is that it won't be that bad to get test-patch.sh
running under cygwin. More complex than the client code though.
bq. What I'm trying to explore here from a maintainability / productivity
prospective whether we should allow writing supplement scripts in languages
that have a wide range of libraries to take care of things like text processing
/ OS manipulation, regardless it's Python, Perl or other languages.
Until the portability issue gets dealt with (which will happen as more time
passes), the answer here is likely no without potentially significantly
increasing the burden of the installed software set.
Of course, unless someone is actually going to do the work, this whole
conversation is sort of moot. Traditionally, the non-java parts of Hadoop are
regularly ignored until it impacts someone's corporate goals. I don't expect
that to change any time soon.
bq.I can't repro the OutOfMemoryError problems after setting reuseForks to
false. I don't know if I'm just overloading these hosts by running multiple
processes with a maximum 1 GB heap. I don't have much visibility into these
hosts. If anyone else out there knows how to find host-level details, please
let me know. Otherwise, I'll probably need to coordinate with infra.
Hmm. I wonder if the fact that MAVEN_OPTS is set to "-Xmx3072m
-XX:MaxPermSize=768m" prior to test-patch.sh invocation in the HADOOP precommit
is part of the problem here.
This is where it'd be super useful to run precommit in the Docker container so
we could see if it was actually a maven version problem. :(
> Enable parallel JUnit tests in pre-commit.
> ------------------------------------------
>
> Key: HADOOP-11984
> URL: https://issues.apache.org/jira/browse/HADOOP-11984
> Project: Hadoop Common
> Issue Type: Improvement
> Components: scripts
> Reporter: Chris Nauroth
> Assignee: Chris Nauroth
> Attachments: HADOOP-11984.001.patch, HADOOP-11984.002.patch,
> HADOOP-11984.003.patch, HADOOP-11984.004.patch, HADOOP-11984.005.patch
>
>
> HADOOP-9287 and related issues implemented the parallel-tests Maven profile
> for running JUnit tests in multiple concurrent processes. This issue
> proposes to activate that profile during pre-commit to speed up execution.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)