Re: Spark - configuration setting doesn't work

2019-10-29 Thread Chetan Khatri
Oct 2019 at 11:02 AM, Chetan Khatri < > chetan.opensou...@gmail.com> wrote: > >> Could someone please help me. >> >> On Thu, Oct 17, 2019 at 7:29 PM Chetan Khatri < >> chetan.opensou...@gmail.com> wrote: >> >>> Hi Users, >>> >

Re: Spark - configuration setting doesn't work

2019-10-27 Thread hemant singh
an Khatri > wrote: > >> Hi Users, >> >> I am setting spark configuration in below way; >> >> val spark = SparkSession.builder().appName(APP_NAME).getOrCreate() >> >> spark.conf.set("spark.speculation", "false") >> spark.conf.se

Re: Spark - configuration setting doesn't work

2019-10-26 Thread Chetan Khatri
Could someone please help me. On Thu, Oct 17, 2019 at 7:29 PM Chetan Khatri wrote: > Hi Users, > > I am setting spark configuration in below way; > > val spark = SparkSession.builder().appName(APP_NAME).getOrCreate() > > spark.conf.set("spark.speculation&q

Spark - configuration setting doesn't work

2019-10-17 Thread Chetan Khatri
Hi Users, I am setting spark configuration in below way; val spark = SparkSession.builder().appName(APP_NAME).getOrCreate() spark.conf.set("spark.speculation", "false") spark.conf.set("spark.broadcast.compress", "true") spark.conf.set("spark.s

Re: Apache Hive with Spark Configuration

2017-01-04 Thread Chetan Khatri
Ryan, I agree that Hive 1.2.1 work reliably with Spark 2.x , but i went through with current stable version of Hive which is 2.0.1 and I am working with that. seems good but i want to make sure the which version of Hive is more reliable with Spark 2.x and i think @Ryan you replied the same which

Re: Apache Hive with Spark Configuration

2017-01-03 Thread Ryan Blue
Chetan, Spark is currently using Hive 1.2.1 to interact with the Metastore. Using that version for Hive is going to be the most reliable, but the metastore API doesn't change very often and we've found (from having different versions as well) that older versions are mostly compatible. Some things

Re: Apache Hive with Spark Configuration

2016-12-28 Thread Gourav Sengupta
Hi, I think that you can configure the hive metastore versions in SPARK. Regards, Gourav On Wed, Dec 28, 2016 at 12:22 PM, Chetan Khatri wrote: > Hello Users / Developers, > > I am using Hive 2.0.1 with MySql as a Metastore, can you tell me which > version is

Apache Hive with Spark Configuration

2016-12-28 Thread Chetan Khatri
Hello Users / Developers, I am using Hive 2.0.1 with MySql as a Metastore, can you tell me which version is more compatible with Spark 2.0.2 ? THanks

Re: Changing Spark configuration midway through application.

2016-08-10 Thread Andrew Ehrlich
If you're changing properties for the SparkContext, then I believe you will have to start a new SparkContext with the new properties. On Wed, Aug 10, 2016 at 8:47 AM, Jestin Ma wrote: > If I run an application, for example with 3 joins: > > [join 1] > [join 2] > [join

Changing Spark configuration midway through application.

2016-08-10 Thread Jestin Ma
If I run an application, for example with 3 joins: [join 1] [join 2] [join 3] [final join and save to disk] Could I change Spark properties in between each join? [join 1] [change properties] [join 2] [change properties] ... Or would I have to create a separate application with different

Re: Spark configuration with 5 nodes

2016-03-19 Thread Mich Talebzadeh
Thanks Steve, For NN it all depends how fast you want a start-up. 1GB of NameNode memory accommodates around 42T so if you are talking about 100GB of NN memory then SSD may make sense to speed up the start-up. Raid 10 is the best one that one can get assuming all internal disks. In general it

Re: Spark configuration with 5 nodes

2016-03-19 Thread Steve Loughran
> On 17 Mar 2016, at 12:28, Mich Talebzadeh wrote: > > Thanks Steve, > > For NN it all depends how fast you want a start-up. 1GB of NameNode memory > accommodates around 42T so if you are talking about 100GB of NN memory then > SSD may make sense to speed up the

Re: Spark configuration with 5 nodes

2016-03-19 Thread Steve Loughran
On 11 Mar 2016, at 16:25, Mich Talebzadeh > wrote: Hi Steve, My argument has always been that if one is going to use Solid State Disks (SSD), it makes sense to have it for NN disks start-up from fsimage etc. Obviously NN lives in

Re: Spark configuration with 5 nodes

2016-03-19 Thread Mich Talebzadeh
Thank you for info Steve. I always believed (IMO) that there is an optimal position where one can plot the projected NN memory (assuming 1GB--> 40TB of data) to the number of nodes. For example heuristically how many nodes would be sufficient for 1PB of storage with nodes each having 512GB of

Re: Spark configuration with 5 nodes

2016-03-11 Thread Mich Talebzadeh
Hi Steve, My argument has always been that if one is going to use Solid State Disks (SSD), it makes sense to have it for NN disks start-up from fsimage etc. Obviously NN lives in memory. Would you also rerommend RAID10 (mirroring & striping) for NN disks? Thanks Dr Mich Talebzadeh

Re: Spark configuration with 5 nodes

2016-03-11 Thread Steve Loughran
On 10 Mar 2016, at 22:15, Ashok Kumar > wrote: Hi, We intend to use 5 servers which will be utilized for building Bigdata Hadoop data warehouse system (not using any propriety distribution like Hortonworks or Cloudera or

Re: Spark configuration with 5 nodes

2016-03-10 Thread Mich Talebzadeh
Hi, Bear in mind that you typically need 1GB of NameNode memory for 1 million blocks. So if you have 128MB block size, you can store 128 * 1E6 / (3 *1024) = 41,666GB of data for every 1GB. Number 3 comes from the fact that the block is replicated three times. In other words just under 42TB of

Re: Spark configuration with 5 nodes

2016-03-10 Thread Prabhu Joseph
Ashok, Cluster nodes has enough memory but CPU cores are less. 512GB / 16 = 32 GB. For 1 core the cluster has 32GB memory. Either their should be more cores available to use efficiently the available memory or don't configure a higher executor memory which will cause lot of GC. Thanks,

Spark configuration with 5 nodes

2016-03-10 Thread Ashok Kumar
  Hi, We intend  to use 5servers which will be utilized for building Bigdata Hadoop data warehousesystem (not using any propriety distribution like Hortonworks or Cloudera orothers).All servers configurations are 512GB RAM, 30TB storageand 16 cores, Ubuntu Linux servers. Hadoop will be

Spark: configuration file 'metrics.properties'

2015-07-24 Thread allonsy
: http://apache-spark-user-list.1001560.n3.nabble.com/Spark-configuration-file-metrics-properties-tp23985.html Sent from the Apache Spark User List mailing list archive at Nabble.com. - To unsubscribe, e-mail: user-unsubscr

Retrieving Spark Configuration properties

2015-07-16 Thread RajG
(SQLConf.scala:283) at org.apache.spark.sql.SQLConf$$anonfun$getConf$1.apply(SQLConf.scala:283) *Am I retrieving the properties in the right way? -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/Retrieving-Spark-Configuration-properties-tp23881.html Sent from

Re: Retrieving Spark Configuration properties

2015-07-16 Thread Yanbo Liang
) at org.apache.spark.sql.SQLConf$$anonfun$getConf$1.apply(SQLConf.scala:283) *Am I retrieving the properties in the right way? -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/Retrieving-Spark-Configuration-properties-tp23881.html Sent from the Apache Spark

Spark Configuration of spark.worker.cleanup.appDataTtl

2015-06-16 Thread luohui20001
Hi guys: I added a parameter spark.worker.cleanup.appDataTtl 3 * 24 * 3600 in my conf/spark-default.conf, then I start my spark cluster. However I got an exception: 15/06/16 14:25:14 INFO util.Utils: Successfully started service 'sparkWorker' on port 43344. 15/06/16 14:25:14 ERROR

Re: Spark Configuration of spark.worker.cleanup.appDataTtl

2015-06-16 Thread Saisai Shao
I think you have to using 604800 instead of 7 * 24 * 3600, obviously SparkConf will not do multiplication for you.. The exception is quite obvious: Caused by: java.lang.NumberFormatException: For input string: 3 * 24 * 3600 2015-06-16 14:52 GMT+08:00 luohui20...@sina.com: Hi guys: I

回复:Re: Spark Configuration of spark.worker.cleanup.appDataTtl

2015-06-16 Thread luohui20001
:Re: Spark Configuration of spark.worker.cleanup.appDataTtl 日期:2015年06月16日 15点00分 I think you have to using 604800 instead of 7 * 24 * 3600, obviously SparkConf will not do multiplication for you.. The exception is quite obvious: Caused by: java.lang.NumberFormatException: For input string: 3

setting spark configuration properties problem

2015-05-05 Thread Hafiz Mujadid
related to cassandra so i declare sparkcontext earlier and then want to set this property at some later stage. Any suggestion? -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/setting-spark-configuration-properties-problem-tp22764.html Sent from the Apache

Spark configuration

2015-02-23 Thread King sami
Hi Experts, I am new in Spark, so I want manipulate it locally on my machine with Ubuntu as OS. I dowloaded the last version of Spark. I ran this command to start it : ./sbin/start-master.sh but an error is occured : *starting org.apache.spark.deploy.master.Master, logging to

Re: Spark configuration

2015-02-23 Thread Sean Owen
It sounds like you downloaded the source distribution perhaps, but have not built it. That's what the message is telling you. See http://spark.apache.org/docs/latest/building-spark.html Or maybe you intended to get a binary distribution. On Mon, Feb 23, 2015 at 10:40 PM, King sami

Re: Spark configuration

2015-02-23 Thread Shlomi Babluki
I guess you downloaded the source code. You can build it with the following command: mvn -DskipTests clean package Or just download a compiled version. Shlomi On 24 בפבר׳ 2015, at 00:40, King sami kgsam...@gmail.com wrote: Hi Experts, I am new in Spark, so I want manipulate it locally

Re: Passing Spark Configuration from Driver (Master) to all of the Slave nodes

2014-12-16 Thread Gerard Maas
. You can read about it in my Blog Post http://progexc.blogspot.co.il/2014/12/spark-configuration-mess-solved.html -- Enjoy, Demi Ben-Ari Senior Software Engineer Windward LTD.

Passing Spark Configuration from Driver (Master) to all of the Slave nodes

2014-12-12 Thread Demi Ben-Ari
/12/spark-configuration-mess-solved.html -- Enjoy, Demi Ben-Ari Senior Software Engineer Windward LTD.

Optimized spark configuration

2014-12-05 Thread vdiwakar.malladi
. Thanks in advance. -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/Optimized-spark-configuration-tp20495.html Sent from the Apache Spark User List mailing list archive at Nabble.com

R: Optimized spark configuration

2014-12-05 Thread Paolo Platter
: vdiwakar.malladimailto:vdiwakar.mall...@gmail.com Inviato: ‎05/‎12/‎2014 18:52 A: u...@spark.incubator.apache.orgmailto:u...@spark.incubator.apache.org Oggetto: Optimized spark configuration Hi Could any one help what would be better / optimized configuration for driver memory, worker memory

Move Spark configuration from SPARK_CLASSPATH to spark-default.conf , HiveContext went wrong with Class com.hadoop.compression.lzo.LzoCodec not found

2014-09-17 Thread Zhun Shen
Hi there, My product environment is AWS EMR with hadoop2.4.0 and spark1.0.2. I moved the spark configuration in SPARK_CLASSPATH to spark-default.conf,  then the hiveContext went wrong. I also found WARN info “WARN DataNucleus.General: Plugin (Bundle) org.datanucleus.store.rdbms is already

How to Reload Spark Configuration Files

2014-06-24 Thread Sirisha Devineni
Hi All, I am working with Spark to add new slaves automatically when there is more data to be processed by the cluster. During this process there is question arisen, after adding/removing new slave node to/from the spark cluster do we need to restart master and other existing slaves in the

Re: How to Reload Spark Configuration Files

2014-06-24 Thread Mayur Rustagi
Not really. You are better off using a cluster manager like Mesos or Yarn for this. Mayur Rustagi Ph: +1 (760) 203 3257 http://www.sigmoidanalytics.com @mayur_rustagi https://twitter.com/mayur_rustagi On Tue, Jun 24, 2014 at 11:35 AM, Sirisha Devineni sirisha_devin...@persistent.co.in wrote:

Re: How to Reload Spark Configuration Files

2014-06-24 Thread Peng Cheng
-Spark-Configuration-Files-tp8159p8219.html Sent from the Apache Spark User List mailing list archive at Nabble.com.