Re: Adding jars
This can help us to solve the immediate issue, however the ideally one should submit the jars in the beginning of the job. -- Sent from: http://apache-spark-user-list.1001560.n3.nabble.com/ - To unsubscribe e-mail: user-unsubscr...@spark.apache.org
Re: Adding jars
The purpose of broadcast variable is different. @Malveeka, could you please explain your usecase and issue. If the FAT/ Uber jar is not having required dependent jars then the spark job will fail at the start itself. What is your scenario in which you want to add new jars? Also, what you mean by adding spark.jars in middle? (Did you mean middle of processing) then you can not change spark jars in middle of execution. However, you can change them for next processing iteration. On Wed, May 23, 2018 at 5:39 PM, kedarsdixitwrote: > In case of already running jobs, you can make use of broadcasters which > will > broadcast the jars to workers, if you want to change it on the fly you can > rebroadcast it. > > You can explore broadcasters bit more to make use of. > > Regards, > Kedar Dixit > Data Science at Persistent Systems Ltd. > > > > -- > Sent from: http://apache-spark-user-list.1001560.n3.nabble.com/ > > - > To unsubscribe e-mail: user-unsubscr...@spark.apache.org > >
Re: Adding jars
In case of already running jobs, you can make use of broadcasters which will broadcast the jars to workers, if you want to change it on the fly you can rebroadcast it. You can explore broadcasters bit more to make use of. Regards, Kedar Dixit Data Science at Persistent Systems Ltd. -- Sent from: http://apache-spark-user-list.1001560.n3.nabble.com/ - To unsubscribe e-mail: user-unsubscr...@spark.apache.org
Re: Adding jars
Hi With spark-submit we can start a new spark job, but it will not add new jar files in already running job. ~Sushil On Wed, May 23, 2018, 17:28 kedarsdixitwrote: > Hi, > > You can add dependencies in spark-submit as below: > > ./bin/spark-submit \ > --class \ > --master \ > --deploy-mode \ > --conf = \ > *--jars * \ > ... # other options >\ > [application-arguments] > > Hope this helps. > > Regards, > > Kedar Dixit > Data Science at Persistent Systems Ltd > > > > -- > Sent from: http://apache-spark-user-list.1001560.n3.nabble.com/ > > - > To unsubscribe e-mail: user-unsubscr...@spark.apache.org > >
Re: Adding jars
Hi, You can add dependencies in spark-submit as below: ./bin/spark-submit \ --class \ --master \ --deploy-mode \ --conf = \ *--jars * \ ... # other options \ [application-arguments] Hope this helps. Regards, Kedar Dixit Data Science at Persistent Systems Ltd -- Sent from: http://apache-spark-user-list.1001560.n3.nabble.com/ - To unsubscribe e-mail: user-unsubscr...@spark.apache.org
Re: adding jars - hive on spark cdh 5.4.3
btw, this issue happens only with classes needed for the inputFormat. if the input format is org.apache.hadoop.mapred.TextInputFormat and the serde is from an additional jar it works just fine. I don't want to upgrade cdh for this. also, if it should work on cdh5.5 why is that. what patch fixes that? (cdh 5.5 is the same hive version as cdh5.4. is it spark related and not hive?) On Sun, Jan 10, 2016 at 9:26 AM, sandeep vurawrote: > Upgrade to CDH 5.5 for spark. It should work > > On Sat, Jan 9, 2016 at 12:17 AM, Ophir Etzion > wrote: > >> It didn't work. assuming I did the right thing. >> in the properties you could see >> >> {"key":"hive.aux.jars.path","value":"file:///data/loko/foursquare.web-hiverc/current/hadoop-hive-serde.jar,file:///data/loko/foursquare.web-hiverc/current/hadoop-hive-udf.jar","isFinal":false,"resource":"programatically"} >> which includes the jar that has the class I need but I still get >> >> org.apache.hive.com.esotericsoftware.kryo.KryoException: Unable to find >> class: com.foursquare.hadoop.hive.io.HiveThriftSequenceFileInputFormat >> >> >> >> On Fri, Jan 8, 2016 at 12:24 PM, Edward Capriolo >> wrote: >> >>> You can not 'add jar' input formats and serde's. They need to be part of >>> your auxlib. >>> >>> On Fri, Jan 8, 2016 at 12:19 PM, Ophir Etzion >>> wrote: >>> I tried now. still getting 16/01/08 16:37:34 ERROR exec.Utilities: Failed to load plan: hdfs://hadoop-alidoro-nn-vip/tmp/hive/hive/c2af9882-38a9-42b0-8d17-3f56708383e8/hive_2016-01-08_16-36-41_370_3307331506800215903-3/-mr-10004/3c90a796-47fc-4541-bbec-b196c40aefab/map.xml: org.apache.hive.com.esotericsoftware.kryo.KryoException: Unable to find class: com.foursquare.hadoop.hive.io.HiveThriftSequenceFileInputFormat Serialization trace: inputFileFormatClass (org.apache.hadoop.hive.ql.plan.PartitionDesc) aliasToPartnInfo (org.apache.hadoop.hive.ql.plan.MapWork) org.apache.hive.com.esotericsoftware.kryo.KryoException: Unable to find class: com.foursquare.hadoop.hive.io.HiveThriftSequenceFileInputFormat HiveThriftSequenceFileInputFormat is in one of the jars I'm trying to add. On Thu, Jan 7, 2016 at 9:58 PM, Prem Sure wrote: > did you try -- jars property in spark submit? if your jar is of huge > size, you can pre-load the jar on all executors in a common available > directory to avoid network IO. > > On Thu, Jan 7, 2016 at 4:03 PM, Ophir Etzion > wrote: > >> I' trying to add jars before running a query using hive on spark on >> cdh 5.4.3. >> I've tried applying the patch in >> https://issues.apache.org/jira/browse/HIVE-12045 (manually as the >> patch is done on a different hive version) but still hasn't succeeded. >> >> did anyone manage to do ADD JAR successfully with CDH? >> >> Thanks, >> Ophir >> > > >>> >> >
Re: adding jars - hive on spark cdh 5.4.3
Upgrade to CDH 5.5 for spark. It should work On Sat, Jan 9, 2016 at 12:17 AM, Ophir Etzionwrote: > It didn't work. assuming I did the right thing. > in the properties you could see > > {"key":"hive.aux.jars.path","value":"file:///data/loko/foursquare.web-hiverc/current/hadoop-hive-serde.jar,file:///data/loko/foursquare.web-hiverc/current/hadoop-hive-udf.jar","isFinal":false,"resource":"programatically"} > which includes the jar that has the class I need but I still get > > org.apache.hive.com.esotericsoftware.kryo.KryoException: Unable to find > class: com.foursquare.hadoop.hive.io.HiveThriftSequenceFileInputFormat > > > > On Fri, Jan 8, 2016 at 12:24 PM, Edward Capriolo > wrote: > >> You can not 'add jar' input formats and serde's. They need to be part of >> your auxlib. >> >> On Fri, Jan 8, 2016 at 12:19 PM, Ophir Etzion >> wrote: >> >>> I tried now. still getting >>> >>> 16/01/08 16:37:34 ERROR exec.Utilities: Failed to load plan: >>> hdfs://hadoop-alidoro-nn-vip/tmp/hive/hive/c2af9882-38a9-42b0-8d17-3f56708383e8/hive_2016-01-08_16-36-41_370_3307331506800215903-3/-mr-10004/3c90a796-47fc-4541-bbec-b196c40aefab/map.xml: >>> org.apache.hive.com.esotericsoftware.kryo.KryoException: Unable to find >>> class: com.foursquare.hadoop.hive.io.HiveThriftSequenceFileInputFormat >>> Serialization trace: >>> inputFileFormatClass (org.apache.hadoop.hive.ql.plan.PartitionDesc) >>> aliasToPartnInfo (org.apache.hadoop.hive.ql.plan.MapWork) >>> org.apache.hive.com.esotericsoftware.kryo.KryoException: Unable to find >>> class: com.foursquare.hadoop.hive.io.HiveThriftSequenceFileInputFormat >>> >>> >>> HiveThriftSequenceFileInputFormat is in one of the jars I'm trying to add. >>> >>> >>> On Thu, Jan 7, 2016 at 9:58 PM, Prem Sure wrote: >>> did you try -- jars property in spark submit? if your jar is of huge size, you can pre-load the jar on all executors in a common available directory to avoid network IO. On Thu, Jan 7, 2016 at 4:03 PM, Ophir Etzion wrote: > I' trying to add jars before running a query using hive on spark on > cdh 5.4.3. > I've tried applying the patch in > https://issues.apache.org/jira/browse/HIVE-12045 (manually as the > patch is done on a different hive version) but still hasn't succeeded. > > did anyone manage to do ADD JAR successfully with CDH? > > Thanks, > Ophir > >>> >> >
Re: adding jars - hive on spark cdh 5.4.3
It didn't work. assuming I did the right thing. in the properties you could see {"key":"hive.aux.jars.path","value":"file:///data/loko/foursquare.web-hiverc/current/hadoop-hive-serde.jar,file:///data/loko/foursquare.web-hiverc/current/hadoop-hive-udf.jar","isFinal":false,"resource":"programatically"} which includes the jar that has the class I need but I still get org.apache.hive.com.esotericsoftware.kryo.KryoException: Unable to find class: com.foursquare.hadoop.hive.io.HiveThriftSequenceFileInputFormat On Fri, Jan 8, 2016 at 12:24 PM, Edward Capriolowrote: > You can not 'add jar' input formats and serde's. They need to be part of > your auxlib. > > On Fri, Jan 8, 2016 at 12:19 PM, Ophir Etzion > wrote: > >> I tried now. still getting >> >> 16/01/08 16:37:34 ERROR exec.Utilities: Failed to load plan: >> hdfs://hadoop-alidoro-nn-vip/tmp/hive/hive/c2af9882-38a9-42b0-8d17-3f56708383e8/hive_2016-01-08_16-36-41_370_3307331506800215903-3/-mr-10004/3c90a796-47fc-4541-bbec-b196c40aefab/map.xml: >> org.apache.hive.com.esotericsoftware.kryo.KryoException: Unable to find >> class: com.foursquare.hadoop.hive.io.HiveThriftSequenceFileInputFormat >> Serialization trace: >> inputFileFormatClass (org.apache.hadoop.hive.ql.plan.PartitionDesc) >> aliasToPartnInfo (org.apache.hadoop.hive.ql.plan.MapWork) >> org.apache.hive.com.esotericsoftware.kryo.KryoException: Unable to find >> class: com.foursquare.hadoop.hive.io.HiveThriftSequenceFileInputFormat >> >> >> HiveThriftSequenceFileInputFormat is in one of the jars I'm trying to add. >> >> >> On Thu, Jan 7, 2016 at 9:58 PM, Prem Sure wrote: >> >>> did you try -- jars property in spark submit? if your jar is of huge >>> size, you can pre-load the jar on all executors in a common available >>> directory to avoid network IO. >>> >>> On Thu, Jan 7, 2016 at 4:03 PM, Ophir Etzion >>> wrote: >>> I' trying to add jars before running a query using hive on spark on cdh 5.4.3. I've tried applying the patch in https://issues.apache.org/jira/browse/HIVE-12045 (manually as the patch is done on a different hive version) but still hasn't succeeded. did anyone manage to do ADD JAR successfully with CDH? Thanks, Ophir >>> >>> >> >
Re: adding jars - hive on spark cdh 5.4.3
You can not 'add jar' input formats and serde's. They need to be part of your auxlib. On Fri, Jan 8, 2016 at 12:19 PM, Ophir Etzionwrote: > I tried now. still getting > > 16/01/08 16:37:34 ERROR exec.Utilities: Failed to load plan: > hdfs://hadoop-alidoro-nn-vip/tmp/hive/hive/c2af9882-38a9-42b0-8d17-3f56708383e8/hive_2016-01-08_16-36-41_370_3307331506800215903-3/-mr-10004/3c90a796-47fc-4541-bbec-b196c40aefab/map.xml: > org.apache.hive.com.esotericsoftware.kryo.KryoException: Unable to find > class: com.foursquare.hadoop.hive.io.HiveThriftSequenceFileInputFormat > Serialization trace: > inputFileFormatClass (org.apache.hadoop.hive.ql.plan.PartitionDesc) > aliasToPartnInfo (org.apache.hadoop.hive.ql.plan.MapWork) > org.apache.hive.com.esotericsoftware.kryo.KryoException: Unable to find > class: com.foursquare.hadoop.hive.io.HiveThriftSequenceFileInputFormat > > > HiveThriftSequenceFileInputFormat is in one of the jars I'm trying to add. > > > On Thu, Jan 7, 2016 at 9:58 PM, Prem Sure wrote: > >> did you try -- jars property in spark submit? if your jar is of huge >> size, you can pre-load the jar on all executors in a common available >> directory to avoid network IO. >> >> On Thu, Jan 7, 2016 at 4:03 PM, Ophir Etzion >> wrote: >> >>> I' trying to add jars before running a query using hive on spark on cdh >>> 5.4.3. >>> I've tried applying the patch in >>> https://issues.apache.org/jira/browse/HIVE-12045 (manually as the patch >>> is done on a different hive version) but still hasn't succeeded. >>> >>> did anyone manage to do ADD JAR successfully with CDH? >>> >>> Thanks, >>> Ophir >>> >> >> >
Re: adding jars - hive on spark cdh 5.4.3
I tried now. still getting 16/01/08 16:37:34 ERROR exec.Utilities: Failed to load plan: hdfs://hadoop-alidoro-nn-vip/tmp/hive/hive/c2af9882-38a9-42b0-8d17-3f56708383e8/hive_2016-01-08_16-36-41_370_3307331506800215903-3/-mr-10004/3c90a796-47fc-4541-bbec-b196c40aefab/map.xml: org.apache.hive.com.esotericsoftware.kryo.KryoException: Unable to find class: com.foursquare.hadoop.hive.io.HiveThriftSequenceFileInputFormat Serialization trace: inputFileFormatClass (org.apache.hadoop.hive.ql.plan.PartitionDesc) aliasToPartnInfo (org.apache.hadoop.hive.ql.plan.MapWork) org.apache.hive.com.esotericsoftware.kryo.KryoException: Unable to find class: com.foursquare.hadoop.hive.io.HiveThriftSequenceFileInputFormat HiveThriftSequenceFileInputFormat is in one of the jars I'm trying to add. On Thu, Jan 7, 2016 at 9:58 PM, Prem Surewrote: > did you try -- jars property in spark submit? if your jar is of huge size, > you can pre-load the jar on all executors in a common available directory > to avoid network IO. > > On Thu, Jan 7, 2016 at 4:03 PM, Ophir Etzion wrote: > >> I' trying to add jars before running a query using hive on spark on cdh >> 5.4.3. >> I've tried applying the patch in >> https://issues.apache.org/jira/browse/HIVE-12045 (manually as the patch >> is done on a different hive version) but still hasn't succeeded. >> >> did anyone manage to do ADD JAR successfully with CDH? >> >> Thanks, >> Ophir >> > >
Re: adding jars - hive on spark cdh 5.4.3
did you try -- jars property in spark submit? if your jar is of huge size, you can pre-load the jar on all executors in a common available directory to avoid network IO. On Thu, Jan 7, 2016 at 4:03 PM, Ophir Etzionwrote: > I' trying to add jars before running a query using hive on spark on cdh > 5.4.3. > I've tried applying the patch in > https://issues.apache.org/jira/browse/HIVE-12045 (manually as the patch > is done on a different hive version) but still hasn't succeeded. > > did anyone manage to do ADD JAR successfully with CDH? > > Thanks, > Ophir >