Re: Understanding the build params for spark with sbt.

2015-04-21 Thread Sree V
Hi Shiyao,
>From the same page you referred:Maven is the official recommendation for 
>packaging Spark, and is the “build of reference”. But SBT is supported for 
>day-to-day development since it can provide much faster iterative compilation. 
>More advanced developers may wish to use SBT.

For maven, pom.xml is the main and important file.
-P stands for Profilesearch for 'profile' in spark/pom.xmlMore on it: 
http://maven.apache.org/guides/introduction/introduction-to-profiles.html
-D stands for Definemaven takes it from Java or earlier languages.It is a way 
to pass system.properties and/or override existing properties from build file.
Core build:spark/core/pom.xml is your build file for building only Spark-Core.

Thanking you.

With Regards
Sree 


 On Tuesday, April 21, 2015 12:12 AM, Akhil Das 
 wrote:
   

 With maven you could like:

mvn -Dhadoop.version=2.3.0 -DskipTests clean package -pl core

ThanksBest Regards
On Mon, Apr 20, 2015 at 8:10 PM, Shiyao Ma  wrote:

Hi.

My usage is only about the spark core and hdfs, so no spark sql or
mlib or other components invovled.


I saw the hint on the
http://spark.apache.org/docs/latest/building-spark.html, with a sample
like:
build/sbt -Pyarn -Phadoop-2.3 assembly. (what's the -P for?)


Fundamentally, I'd like to let sbt only compile and package the core
and the hadoop.

Meanwhile, it would be appreciated if you could inform me what's the
scala file that controls the logic of "-Pyarn", so that I can dig into
the build source and have a finer control.



Thanks.

--

吾輩は猫である。ホームーページはhttp://introo.me。

-
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org





  

Re: Spark on Windows

2015-04-17 Thread Sree V
spark 'master' branch (i.e. v1.4.0) builds successfully on windows 8.1 intel i7 
64-bit with oracle jdk8_45.with maven opts without the flag 
"-XX:ReservedCodeCacheSize=1g".
takes about 33 minutes.
Thanking you.

With Regards
Sree  


 On Thursday, April 16, 2015 9:07 PM, Arun Lists  
wrote:
   

 Here is what I got from the engineer who worked on building Spark and using it 
on Windows:
1)  Hadoop winutils.exe is needed on Windows, even for local files – and you 
have to set the Hadoop.home.dir in the spark-class2.cmd (for the two lines with 
$RUNNER near the end, by adding “-Dhadoop.home.dir=” file after 
downloading Hadoop binaries + winutils. 2)  Java/Spark cannot delete the spark 
temporary files and it throws an exception (program still works though).  
Manual clean-up works just fine, and it is not a permissions issue as it has 
rights to create the file (I have also tried using my own directory rather than 
the default, same error).3)  tried building Spark again, and have attached the 
log – I don’t get any errors, just warnings.  However when I try to use that 
JAR I just get the error message “Error: Could not find or load main class 
org.apache.spark.deploy.SparkSubmit”.
On Thu, Apr 16, 2015 at 12:19 PM, Arun Lists  wrote:

Thanks, Matei! We'll try that and let you know if it works. You are correct in 
inferring that some of the problems we had were with dependencies.
We also had problems with the spark-submit scripts. I will get the details from 
the engineer who worked on the Windows builds and provide them to you.
arun

On Thu, Apr 16, 2015 at 10:44 AM, Matei Zaharia  wrote:

You could build Spark with Scala 2.11 on Mac / Linux and transfer it over to 
Windows. AFAIK it should build on Windows too, the only problem is that Maven 
might take a long time to download dependencies. What errors are you seeing?

Matei

> On Apr 16, 2015, at 9:23 AM, Arun Lists  wrote:
>
> We run Spark on Mac and Linux but also need to run it on Windows 8.1 and  
> Windows Server. We ran into problems with the Scala 2.10 binary bundle for 
> Spark 1.3.0 but managed to get it working. However, on Mac/Linux, we are on 
> Scala 2.11.6 (we built Spark from the sources). On Windows, however despite 
> our best efforts we cannot get Spark 1.3.0 as built from sources working for 
> Scala 2.11.6. Spark has too many moving parts and dependencies!
>
> When can we expect to see a binary bundle for Spark 1.3.0 that is built for 
> Scala 2.11.6?  I read somewhere that the only reason that Spark 1.3.0 is 
> still built for Scala 2.10 is because Kafka is still on Scala 2.10. For those 
> of us who don't use Kafka, can we have a Scala 2.10 bundle.
>
> If there isn't an official bundle arriving any time soon, can someone who has 
> built it for Windows 8.1 successfully please share with the group?
>
> Thanks,
> arun
>







-
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org