Re: Results of tests

2015-01-09 Thread Ted Yu
k.deploy<
>> >
>> https://amplab.cs.berkeley.edu/jenkins/view/Spark/job/Spark-1.2-Maven-with-YARN/lastSuccessfulBuild/HADOOP_PROFILE=hadoop-2.4,label=centos/testReport/org.apache.spark.deploy/
>> >
>> > 16 s0
>> > > 0
>> > > 29
>> > > 29
>> > >
>> > > org.apache.spark.deploy.worker<
>> >
>> https://amplab.cs.berkeley.edu/jenkins/view/Spark/job/Spark-1.2-Maven-with-YARN/lastSuccessfulBuild/HADOOP_PROFILE=hadoop-2.4,label=centos/testReport/org.apache.spark.deploy.worker/
>> >
>> >  0.55 s  0
>> > > 0
>> > > 12
>> > > 12
>> > >
>> > > 
>> > >
>> > >
>> > > Moreover, in my Ubuntu/x86_64 environment, I do not find 3745 tests
>> and
>> > 0 failures, but 3485 tests and 4 failures (when using Oracle JVM 1.7 ).
>> > When using IBM JVM, there are only 2566 tests and 5 failures (in same
>> > component: Streaming).
>> > >
>> > > On my PPC64BE (BE = Big-Endian)environment, the tests block after 2
>> > hundreds of tests.
>> > > Is Spark independent of Little/Big-Endian stuff ?
>> > >
>> > > On my PPC64LE (LE = Little-Endian) environment, I have 3485 tests only
>> > (like on Ubuntu/x86_64 with IBM JVM), with 6 or 285 failures...
>> > >
>> > > So, I need to learn more about how your Jenkins environment extracts
>> > details about the results.
>> > > Moreover, which JVM is used ?
>> > >
>> > > Do you plan to use IBM JVM in order to check that Spark and IBM JVM
>> are
>> > compatible ? (they already do not look to be compatible 100% ...).
>> > >
>> > > Thanks
>> > >
>> > > Tony
>> > >
>> > > IBM Coop Architect & Technical Leader
>> > > Office : +33 (0) 4 76 29 72 67
>> > > 1 rue de Provence - 38432 Échirolles - France
>> > > www.atos.net<http://www.atos.net/>
>> > > 
>> > > De : Ted Yu [yuzhih...@gmail.com]
>> > > Envoyé : jeudi 8 janvier 2015 17:43
>> > > À : Tony Reix
>> > > Cc : dev@spark.apache.org
>> > > Objet : Re: Results of tests
>> > >
>> > > Here it is:
>> > >
>> > > [centos] $
>> >
>> /home/jenkins/tools/hudson.tasks.Maven_MavenInstallation/Maven_3.0.5/bin/mvn
>> > -DHADOOP_PROFILE=hadoop-2.4 -Dlabel=centos -DskipTests -Phadoop-2.4
>> -Pyarn
>> > -Phive clean package
>> > >
>> > >
>> > > You can find the above in
>> >
>> https://amplab.cs.berkeley.edu/jenkins/view/Spark/job/Spark-1.2-Maven-with-YARN/lastSuccessfulBuild/HADOOP_PROFILE=hadoop-2.4,label=centos/consoleFull
>> > >
>> > >
>> > > Cheers
>> > >
>> > > On Thu, Jan 8, 2015 at 8:05 AM, Tony Reix > > tony.r...@bull.net>> wrote:
>> > > Thanks !
>> > >
>> > > I've been able to see that there are 3745 tests for version 1.2.0 with
>> > profile Hadoop 2.4  .
>> > > However, on my side, the maximum tests I've seen are 3485... About 300
>> > tests are missing on my side.
>> > > Which Maven option has been used for producing the report file used
>> for
>> > building the page:
>> > >
>> >
>> https://amplab.cs.berkeley.edu/jenkins/view/Spark/job/Spark-1.2-Maven-with-YARN/lastSuccessfulBuild/HADOOP_PROFILE=hadoop-2.4,label=centos/testReport/
>> > >   ? (I'm not authorized to look at the "configuration" part)
>> > >
>> > > Thx !
>> > >
>> > > Tony
>> > >
>> > > 
>> > > De : Ted Yu [yuzhih...@gmail.com<mailto:yuzhih...@gmail.com>]
>> > > Envoyé : jeudi 8 janvier 2015 16:11
>> > > À : Tony Reix
>> > > Cc : dev@spark.apache.org<mailto:dev@spark.apache.org>
>> > > Objet : Re: Results of tests
>> > >
>> > > Please take a look at
>> https://amplab.cs.berkeley.edu/jenkins/view/Spark/
>> > >
>> > > On Thu, Jan 8, 2015 at 5:40 AM, Tony Reix > > tony.r...@bull.net>> wrote:
>> > > Hi,
>> > > I'm checking that Spark works fine on a new environment (PPC64
>> hardware).
>> > > I've found some issues, with versions 1.1.0, 1.1.1, and 1.2.0, even
>> when
>> > running on Ubuntu on x86_64 with Oracle JVM. I'd like to know where I
>> can
>> > find the results of the tests of Spark, for each version and for the
>> > different versions, in order to have a reference to compare my results
>> > with. I cannot find them on Spark web-site.
>> > > Thx
>> > > Tony
>> > >
>> > >
>> > >
>> >
>> > -
>> > To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org
>> > For additional commands, e-mail: dev-h...@spark.apache.org
>> >
>> >
>>
>
>


Re: Results of tests

2015-01-09 Thread Nicholas Chammas
; > 0
> > > > 17
> > > > 17
> > > >
> > > > org.apache.spark.deploy<
> > >
> > https://amplab.cs.berkeley.edu/jenkins/view/Spark/job/Spark-
> 1.2-Maven-with-YARN/lastSuccessfulBuild/HADOOP_PROFILE=hadoop-2.4,label=
> centos/testReport/org.apache.spark.deploy/
> > >
> > > 16 s0
> > > > 0
> > > > 29
> > > > 29
> > > >
> > > > org.apache.spark.deploy.worker<
> > >
> > https://amplab.cs.berkeley.edu/jenkins/view/Spark/job/Spark-
> 1.2-Maven-with-YARN/lastSuccessfulBuild/HADOOP_PROFILE=hadoop-2.4,label=
> centos/testReport/org.apache.spark.deploy.worker/
> > >
> > >  0.55 s  0
> > > > 0
> > > > 12
> > > > 12
> > > >
> > > > 
> > > >
> > > >
> > > > Moreover, in my Ubuntu/x86_64 environment, I do not find 3745 tests
> and
> > > 0 failures, but 3485 tests and 4 failures (when using Oracle JVM 1.7 ).
> > > When using IBM JVM, there are only 2566 tests and 5 failures (in same
> > > component: Streaming).
> > > >
> > > > On my PPC64BE (BE = Big-Endian)environment, the tests block after 2
> > > hundreds of tests.
> > > > Is Spark independent of Little/Big-Endian stuff ?
> > > >
> > > > On my PPC64LE (LE = Little-Endian) environment, I have 3485 tests
> only
> > > (like on Ubuntu/x86_64 with IBM JVM), with 6 or 285 failures...
> > > >
> > > > So, I need to learn more about how your Jenkins environment extracts
> > > details about the results.
> > > > Moreover, which JVM is used ?
> > > >
> > > > Do you plan to use IBM JVM in order to check that Spark and IBM JVM
> are
> > > compatible ? (they already do not look to be compatible 100% ...).
> > > >
> > > > Thanks
> > > >
> > > > Tony
> > > >
> > > > IBM Coop Architect & Technical Leader
> > > > Office : +33 (0) 4 76 29 72 67
> > > > 1 rue de Provence - 38432 Échirolles - France
> > > > www.atos.net<http://www.atos.net/>
> > > > 
> > > > De : Ted Yu [yuzhih...@gmail.com]
> > > > Envoyé : jeudi 8 janvier 2015 17:43
> > > > À : Tony Reix
> > > > Cc : dev@spark.apache.org
> > > > Objet : Re: Results of tests
> > > >
> > > > Here it is:
> > > >
> > > > [centos] $
> > >
> > /home/jenkins/tools/hudson.tasks.Maven_MavenInstallation/Mav
> en_3.0.5/bin/mvn
> > > -DHADOOP_PROFILE=hadoop-2.4 -Dlabel=centos -DskipTests -Phadoop-2.4
> > -Pyarn
> > > -Phive clean package
> > > >
> > > >
> > > > You can find the above in
> > >
> > https://amplab.cs.berkeley.edu/jenkins/view/Spark/job/Spark-
> 1.2-Maven-with-YARN/lastSuccessfulBuild/HADOOP_PROFILE=hadoop-2.4,label=
> centos/consoleFull
> > > >
> > > >
> > > > Cheers
> > > >
> > > > On Thu, Jan 8, 2015 at 8:05 AM, Tony Reix   > > tony.r...@bull.net>> wrote:
> > > > Thanks !
> > > >
> > > > I've been able to see that there are 3745 tests for version 1.2.0
> with
> > > profile Hadoop 2.4  .
> > > > However, on my side, the maximum tests I've seen are 3485... About
> 300
> > > tests are missing on my side.
> > > > Which Maven option has been used for producing the report file used
> for
> > > building the page:
> > > >
> > >
> > https://amplab.cs.berkeley.edu/jenkins/view/Spark/job/Spark-
> 1.2-Maven-with-YARN/lastSuccessfulBuild/HADOOP_PROFILE=hadoop-2.4,label=
> centos/testReport/
> > > >   ? (I'm not authorized to look at the "configuration" part)
> > > >
> > > > Thx !
> > > >
> > > > Tony
> > > >
> > > > 
> > > > De : Ted Yu [yuzhih...@gmail.com<mailto:yuzhih...@gmail.com>]
> > > > Envoyé : jeudi 8 janvier 2015 16:11
> > > > À : Tony Reix
> > > > Cc : dev@spark.apache.org<mailto:dev@spark.apache.org>
> > > > Objet : Re: Results of tests
> > > >
> > > > Please take a look at
> > https://amplab.cs.berkeley.edu/jenkins/view/Spark/
> > > >
> > > > On Thu, Jan 8, 2015 at 5:40 AM, Tony Reix   > > tony.r...@bull.net>> wrote:
> > > > Hi,
> > > > I'm checking that Spark works fine on a new environment (PPC64
> > hardware).
> > > > I've found some issues, with versions 1.1.0, 1.1.1, and 1.2.0, even
> > when
> > > running on Ubuntu on x86_64 with Oracle JVM. I'd like to know where I
> can
> > > find the results of the tests of Spark, for each version and for the
> > > different versions, in order to have a reference to compare my results
> > > with. I cannot find them on Spark web-site.
> > > > Thx
> > > > Tony
> > > >
> > > >
> > > >
> > >
> > > -
> > > To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org
> > > For additional commands, e-mail: dev-h...@spark.apache.org
> > >
> > >
> >
>


Re: Results of tests

2015-01-09 Thread Josh Rosen
> > > 12
> > >
> > > 
> > >
> > >
> > > Moreover, in my Ubuntu/x86_64 environment, I do not find 3745 tests and
> > 0 failures, but 3485 tests and 4 failures (when using Oracle JVM 1.7 ).
> > When using IBM JVM, there are only 2566 tests and 5 failures (in same
> > component: Streaming).
> > >
> > > On my PPC64BE (BE = Big-Endian)environment, the tests block after 2
> > hundreds of tests.
> > > Is Spark independent of Little/Big-Endian stuff ?
> > >
> > > On my PPC64LE (LE = Little-Endian) environment, I have 3485 tests only
> > (like on Ubuntu/x86_64 with IBM JVM), with 6 or 285 failures...
> > >
> > > So, I need to learn more about how your Jenkins environment extracts
> > details about the results.
> > > Moreover, which JVM is used ?
> > >
> > > Do you plan to use IBM JVM in order to check that Spark and IBM JVM are
> > compatible ? (they already do not look to be compatible 100% ...).
> > >
> > > Thanks
> > >
> > > Tony
> > >
> > > IBM Coop Architect & Technical Leader
> > > Office : +33 (0) 4 76 29 72 67
> > > 1 rue de Provence - 38432 Échirolles - France
> > > www.atos.net<http://www.atos.net/>
> > > 
> > > De : Ted Yu [yuzhih...@gmail.com]
> > > Envoyé : jeudi 8 janvier 2015 17:43
> > > À : Tony Reix
> > > Cc : dev@spark.apache.org
> > > Objet : Re: Results of tests
> > >
> > > Here it is:
> > >
> > > [centos] $
> >
> /home/jenkins/tools/hudson.tasks.Maven_MavenInstallation/Maven_3.0.5/bin/mvn
> > -DHADOOP_PROFILE=hadoop-2.4 -Dlabel=centos -DskipTests -Phadoop-2.4
> -Pyarn
> > -Phive clean package
> > >
> > >
> > > You can find the above in
> >
> https://amplab.cs.berkeley.edu/jenkins/view/Spark/job/Spark-1.2-Maven-with-YARN/lastSuccessfulBuild/HADOOP_PROFILE=hadoop-2.4,label=centos/consoleFull
> > >
> > >
> > > Cheers
> > >
> > > On Thu, Jan 8, 2015 at 8:05 AM, Tony Reix  > tony.r...@bull.net>> wrote:
> > > Thanks !
> > >
> > > I've been able to see that there are 3745 tests for version 1.2.0 with
> > profile Hadoop 2.4  .
> > > However, on my side, the maximum tests I've seen are 3485... About 300
> > tests are missing on my side.
> > > Which Maven option has been used for producing the report file used for
> > building the page:
> > >
> >
> https://amplab.cs.berkeley.edu/jenkins/view/Spark/job/Spark-1.2-Maven-with-YARN/lastSuccessfulBuild/HADOOP_PROFILE=hadoop-2.4,label=centos/testReport/
> > >   ? (I'm not authorized to look at the "configuration" part)
> > >
> > > Thx !
> > >
> > > Tony
> > >
> > > 
> > > De : Ted Yu [yuzhih...@gmail.com<mailto:yuzhih...@gmail.com>]
> > > Envoyé : jeudi 8 janvier 2015 16:11
> > > À : Tony Reix
> > > Cc : dev@spark.apache.org<mailto:dev@spark.apache.org>
> > > Objet : Re: Results of tests
> > >
> > > Please take a look at
> https://amplab.cs.berkeley.edu/jenkins/view/Spark/
> > >
> > > On Thu, Jan 8, 2015 at 5:40 AM, Tony Reix  > tony.r...@bull.net>> wrote:
> > > Hi,
> > > I'm checking that Spark works fine on a new environment (PPC64
> hardware).
> > > I've found some issues, with versions 1.1.0, 1.1.1, and 1.2.0, even
> when
> > running on Ubuntu on x86_64 with Oracle JVM. I'd like to know where I can
> > find the results of the tests of Spark, for each version and for the
> > different versions, in order to have a reference to compare my results
> > with. I cannot find them on Spark web-site.
> > > Thx
> > > Tony
> > >
> > >
> > >
> >
> > -
> > To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org
> > For additional commands, e-mail: dev-h...@spark.apache.org
> >
> >
>


Re: Results of tests

2015-01-09 Thread Ted Yu
For a build which uses JUnit, we would see a summary such as the following (
https://builds.apache.org/job/HBase-TRUNK/6007/console):

Tests run: 2199, Failures: 0, Errors: 0, Skipped: 25


In 
https://amplab.cs.berkeley.edu/jenkins/view/Spark/job/Spark-1.2-Maven-with-YARN/lastSuccessfulBuild/HADOOP_PROFILE=hadoop-2.4,label=centos/consoleFull
, I don't see such statistics.


Looks like scalatest-maven-plugin can be enhanced :-)


On Fri, Jan 9, 2015 at 3:52 AM, Sean Owen  wrote:

> Hey Tony, the number of tests run could vary depending on how the
> build is configured. For example, YARN-related tests would only run
> when the yarn profile is turned on. Java 8 tests would only run under
> Java 8.
>
> Although I don't know that there's any reason to believe the IBM JVM
> has a problem with Spark, I see this issue that is potentially related
> to endian-ness : https://issues.apache.org/jira/browse/SPARK-2018 I
> don't know if that was a Spark issue. Certainly, would be good for you
> to investigate if you are interested in resolving it.
>
> The Jenkins output shows you exactly what tests were run and how --
> have a look at the logs.
>
>
> https://amplab.cs.berkeley.edu/jenkins/view/Spark/job/Spark-1.2-Maven-with-YARN/lastSuccessfulBuild/HADOOP_PROFILE=hadoop-2.4,label=centos/consoleFull
>
> On Fri, Jan 9, 2015 at 9:15 AM, Tony Reix  wrote:
> > Hi Ted
> >
> > Thanks for the info.
> > However, I'm still unable to understand how the page:
> >
> https://amplab.cs.berkeley.edu/jenkins/view/Spark/job/Spark-1.2-Maven-with-YARN/lastSuccessfulBuild/HADOOP_PROFILE=hadoop-2.4,label=centos/testReport/
> > has been built.
> > This page contains details I do not find in the page you indicated to me:
> >
> https://amplab.cs.berkeley.edu/jenkins/view/Spark/job/Spark-1.2-Maven-with-YARN/lastSuccessfulBuild/HADOOP_PROFILE=hadoop-2.4,label=centos/consoleFull
> >
> > As an example, I'm still unable to find these details:
> > org.apache.spark<
> https://amplab.cs.berkeley.edu/jenkins/view/Spark/job/Spark-1.2-Maven-with-YARN/lastSuccessfulBuild/HADOOP_PROFILE=hadoop-2.4,label=centos/testReport/org.apache.spark/>
>  12 mn   0
> > 1
> > 247
> > 248
> >
> > org.apache.spark.api.python<
> https://amplab.cs.berkeley.edu/jenkins/view/Spark/job/Spark-1.2-Maven-with-YARN/lastSuccessfulBuild/HADOOP_PROFILE=hadoop-2.4,label=centos/testReport/org.apache.spark.api.python/>
> 20 ms   0
> > 0
> > 2
> > 2
> >
> > org.apache.spark.bagel<
> https://amplab.cs.berkeley.edu/jenkins/view/Spark/job/Spark-1.2-Maven-with-YARN/lastSuccessfulBuild/HADOOP_PROFILE=hadoop-2.4,label=centos/testReport/org.apache.spark.bagel/>
>  7.7 s   0
> > 0
> > 4
> > 4
> >
> > org.apache.spark.broadcast<
> https://amplab.cs.berkeley.edu/jenkins/view/Spark/job/Spark-1.2-Maven-with-YARN/lastSuccessfulBuild/HADOOP_PROFILE=hadoop-2.4,label=centos/testReport/org.apache.spark.broadcast/>
>  43 s0
> > 0
> > 17
> > 17
> >
> > org.apache.spark.deploy<
> https://amplab.cs.berkeley.edu/jenkins/view/Spark/job/Spark-1.2-Maven-with-YARN/lastSuccessfulBuild/HADOOP_PROFILE=hadoop-2.4,label=centos/testReport/org.apache.spark.deploy/>
> 16 s0
> > 0
> > 29
> > 29
> >
> > org.apache.spark.deploy.worker<
> https://amplab.cs.berkeley.edu/jenkins/view/Spark/job/Spark-1.2-Maven-with-YARN/lastSuccessfulBuild/HADOOP_PROFILE=hadoop-2.4,label=centos/testReport/org.apache.spark.deploy.worker/>
>  0.55 s  0
> > 0
> > 12
> > 12
> >
> > 
> >
> >
> > Moreover, in my Ubuntu/x86_64 environment, I do not find 3745 tests and
> 0 failures, but 3485 tests and 4 failures (when using Oracle JVM 1.7 ).
> When using IBM JVM, there are only 2566 tests and 5 failures (in same
> component: Streaming).
> >
> > On my PPC64BE (BE = Big-Endian)environment, the tests block after 2
> hundreds of tests.
> > Is Spark independent of Little/Big-Endian stuff ?
> >
> > On my PPC64LE (LE = Little-Endian) environment, I have 3485 tests only
> (like on Ubuntu/x86_64 with IBM JVM), with 6 or 285 failures...
> >
> > So, I need to learn more about how your Jenkins environment extracts
> details about the results.
> > Moreover, which JVM is used ?
> >
> > Do you plan to use IBM JVM in order to check that Spark and IBM JVM are
> compatible ? (they already do not look to be compatible 100% ...).
> >
> > Thanks
> >
> > Tony
> >
> > IBM Coop 

Re: Results of tests

2015-01-09 Thread Sean Owen
Hey Tony, the number of tests run could vary depending on how the
build is configured. For example, YARN-related tests would only run
when the yarn profile is turned on. Java 8 tests would only run under
Java 8.

Although I don't know that there's any reason to believe the IBM JVM
has a problem with Spark, I see this issue that is potentially related
to endian-ness : https://issues.apache.org/jira/browse/SPARK-2018 I
don't know if that was a Spark issue. Certainly, would be good for you
to investigate if you are interested in resolving it.

The Jenkins output shows you exactly what tests were run and how --
have a look at the logs.

https://amplab.cs.berkeley.edu/jenkins/view/Spark/job/Spark-1.2-Maven-with-YARN/lastSuccessfulBuild/HADOOP_PROFILE=hadoop-2.4,label=centos/consoleFull

On Fri, Jan 9, 2015 at 9:15 AM, Tony Reix  wrote:
> Hi Ted
>
> Thanks for the info.
> However, I'm still unable to understand how the page:
>
> https://amplab.cs.berkeley.edu/jenkins/view/Spark/job/Spark-1.2-Maven-with-YARN/lastSuccessfulBuild/HADOOP_PROFILE=hadoop-2.4,label=centos/testReport/
> has been built.
> This page contains details I do not find in the page you indicated to me:
>
> https://amplab.cs.berkeley.edu/jenkins/view/Spark/job/Spark-1.2-Maven-with-YARN/lastSuccessfulBuild/HADOOP_PROFILE=hadoop-2.4,label=centos/consoleFull
>
> As an example, I'm still unable to find these details:
> org.apache.spark<https://amplab.cs.berkeley.edu/jenkins/view/Spark/job/Spark-1.2-Maven-with-YARN/lastSuccessfulBuild/HADOOP_PROFILE=hadoop-2.4,label=centos/testReport/org.apache.spark/>
>12 mn   0
> 1
> 247
> 248
>
> org.apache.spark.api.python<https://amplab.cs.berkeley.edu/jenkins/view/Spark/job/Spark-1.2-Maven-with-YARN/lastSuccessfulBuild/HADOOP_PROFILE=hadoop-2.4,label=centos/testReport/org.apache.spark.api.python/>
>  20 ms   0
> 0
> 2
> 2
>
> org.apache.spark.bagel<https://amplab.cs.berkeley.edu/jenkins/view/Spark/job/Spark-1.2-Maven-with-YARN/lastSuccessfulBuild/HADOOP_PROFILE=hadoop-2.4,label=centos/testReport/org.apache.spark.bagel/>
>7.7 s   0
> 0
> 4
> 4
>
> org.apache.spark.broadcast<https://amplab.cs.berkeley.edu/jenkins/view/Spark/job/Spark-1.2-Maven-with-YARN/lastSuccessfulBuild/HADOOP_PROFILE=hadoop-2.4,label=centos/testReport/org.apache.spark.broadcast/>
>43 s0
> 0
> 17
> 17
>
> org.apache.spark.deploy<https://amplab.cs.berkeley.edu/jenkins/view/Spark/job/Spark-1.2-Maven-with-YARN/lastSuccessfulBuild/HADOOP_PROFILE=hadoop-2.4,label=centos/testReport/org.apache.spark.deploy/>
>  16 s0
> 0
> 29
> 29
>
> org.apache.spark.deploy.worker<https://amplab.cs.berkeley.edu/jenkins/view/Spark/job/Spark-1.2-Maven-with-YARN/lastSuccessfulBuild/HADOOP_PROFILE=hadoop-2.4,label=centos/testReport/org.apache.spark.deploy.worker/>
>0.55 s  0
> 0
> 12
> 12
>
> 
>
>
> Moreover, in my Ubuntu/x86_64 environment, I do not find 3745 tests and 0 
> failures, but 3485 tests and 4 failures (when using Oracle JVM 1.7 ). When 
> using IBM JVM, there are only 2566 tests and 5 failures (in same component: 
> Streaming).
>
> On my PPC64BE (BE = Big-Endian)environment, the tests block after 2 hundreds 
> of tests.
> Is Spark independent of Little/Big-Endian stuff ?
>
> On my PPC64LE (LE = Little-Endian) environment, I have 3485 tests only (like 
> on Ubuntu/x86_64 with IBM JVM), with 6 or 285 failures...
>
> So, I need to learn more about how your Jenkins environment extracts details 
> about the results.
> Moreover, which JVM is used ?
>
> Do you plan to use IBM JVM in order to check that Spark and IBM JVM are 
> compatible ? (they already do not look to be compatible 100% ...).
>
> Thanks
>
> Tony
>
> IBM Coop Architect & Technical Leader
> Office : +33 (0) 4 76 29 72 67
> 1 rue de Provence - 38432 Échirolles - France
> www.atos.net<http://www.atos.net/>
> 
> De : Ted Yu [yuzhih...@gmail.com]
> Envoyé : jeudi 8 janvier 2015 17:43
> À : Tony Reix
> Cc : dev@spark.apache.org
> Objet : Re: Results of tests
>
> Here it is:
>
> [centos] $ 
> /home/jenkins/tools/hudson.tasks.Maven_MavenInstallation/Maven_3.0.5/bin/mvn 
> -DHADOOP_PROFILE=hadoop-2.4 -Dlabel=centos -DskipTests -Phadoop-2.4 -Pyarn 
> -Phive clean package
>
>
> You can find the above in 
> https://amplab.cs.berkeley.edu/jenkins/view/Spark/job/Spark-1.2-Maven-with-YARN/lastSuccessfulBuild/HADOOP_PROFILE=hadoop-2.4,label=centos/consoleFull
>
>
> Cheers
>
> On Thu, Jan 8, 2015 at 8:05 AM, Tony 

Re: Results of tests

2015-01-08 Thread Ted Yu
Here it is:

[centos] $ 
/home/jenkins/tools/hudson.tasks.Maven_MavenInstallation/Maven_3.0.5/bin/mvn
-DHADOOP_PROFILE=hadoop-2.4 -Dlabel=centos -DskipTests -Phadoop-2.4
-Pyarn -Phive clean package


You can find the above in
https://amplab.cs.berkeley.edu/jenkins/view/Spark/job/Spark-1.2-Maven-with-YARN/lastSuccessfulBuild/HADOOP_PROFILE=hadoop-2.4,label=centos/consoleFull


Cheers


On Thu, Jan 8, 2015 at 8:05 AM, Tony Reix  wrote:

>  Thanks !
>
> I've been able to see that there are 3745 tests for version 1.2.0 with
> profile Hadoop 2.4  .
> However, on my side, the maximum tests I've seen are 3485... About 300
> tests are missing on my side.
> Which Maven option has been used for producing the report file used for
> building the page:
>
> https://amplab.cs.berkeley.edu/jenkins/view/Spark/job/Spark-1.2-Maven-with-YARN/lastSuccessfulBuild/HADOOP_PROFILE=hadoop-2.4,label=centos/testReport/
>   ? (I'm not authorized to look at the "configuration" part)
>
> Thx !
>
> Tony
>
>  --
> *De :* Ted Yu [yuzhih...@gmail.com]
> *Envoyé :* jeudi 8 janvier 2015 16:11
> *À :* Tony Reix
> *Cc :* dev@spark.apache.org
> *Objet :* Re: Results of tests
>
>   Please take a look at https://amplab.cs.berkeley.edu/jenkins/view/Spark/
>
> On Thu, Jan 8, 2015 at 5:40 AM, Tony Reix  wrote:
>
>> Hi,
>> I'm checking that Spark works fine on a new environment (PPC64 hardware).
>> I've found some issues, with versions 1.1.0, 1.1.1, and 1.2.0, even when
>> running on Ubuntu on x86_64 with Oracle JVM. I'd like to know where I can
>> find the results of the tests of Spark, for each version and for the
>> different versions, in order to have a reference to compare my results
>> with. I cannot find them on Spark web-site.
>> Thx
>> Tony
>>
>>
>


Re: Results of tests

2015-01-08 Thread Ted Yu
Please take a look at https://amplab.cs.berkeley.edu/jenkins/view/Spark/

On Thu, Jan 8, 2015 at 5:40 AM, Tony Reix  wrote:

> Hi,
> I'm checking that Spark works fine on a new environment (PPC64 hardware).
> I've found some issues, with versions 1.1.0, 1.1.1, and 1.2.0, even when
> running on Ubuntu on x86_64 with Oracle JVM. I'd like to know where I can
> find the results of the tests of Spark, for each version and for the
> different versions, in order to have a reference to compare my results
> with. I cannot find them on Spark web-site.
> Thx
> Tony
>
>