Re: PIO 0.12.1 with HDP Spark on YARN

2018-05-29 Thread Pat Ferrel
Yarn has to be started explicitly. Usually it is part of Hadoop and is started with Hadoop. Spark only contains the client for Yarn (afaik). From: Miller, Clifford Reply: user@predictionio.apache.org Date: May 29, 2018 at 6:45:43 PM To: user@predictionio.apache.org Subject: Re: PIO

Re: PIO 0.12.1 with HDP Spark on YARN

2018-05-29 Thread suyash kharade
I use 'pio train -- --master yarn' It works for me to train universal recommender On Tue, May 29, 2018 at 8:31 PM, Miller, Clifford < clifford.mil...@phoenix-opsgroup.com> wrote: > To add more details to this. When I attempt to execute my training job > using the command 'pio train -- --master

Re: Spark cluster error

2018-05-29 Thread Donald Szeto
I recall at one point Spark switched to use per-thread classpath so that each job would have its own isolated classpath. That was probably around Spark 1.5 though, so not likely the exact same case here. From what version of Spark to what version did you upgrade to? On Tue, May 29, 2018 at 2:39

Re: Spark cluster error

2018-05-29 Thread Pat Ferrel
Yes, the spark-submit --jars is where we started to find the missing class. The class isn’t found on the remote executor so we looked in the jars actually downloaded into the executor’s work dir. the PIO assembly jars are there are do have the classes. This would be in the classpath of the

Re: Viewed This Bought That Product Template - unresolved dependency

2018-05-29 Thread Donald Szeto
Hi Nasos, I believe that the template will need to be updated to use the "org.apache.predictionio" group name. "io.prediction" was before the project became an Apache one. Regards, Donald On Fri, May 25, 2018 at 5:04 AM, Nasos Papageorgiou < at.papageorg...@gmail.com> wrote: > Hi all, > > I am

Re: Spark cluster error

2018-05-29 Thread Donald Szeto
Sorry, what I meant was the actual spark-submit command that PIO was using. It should be in the log. What Spark version was that? I recall classpath issues with certain versions of Spark. On Thu, May 24, 2018 at 4:52 PM, Pat Ferrel wrote: > Thanks Donald, > > We have: > >- built pio with

Re: pio app new failed in hbase

2018-05-29 Thread Marco Goldin
Thank you very much for your explanation. It all makes sense of course, i guess as soon as i'll setup a cluster everything will be better (and more manageable/predictable). Messing with hadoop, hdfs and hbase all in the same machine seems not the way to go even in developement stage. I'm also

Re: pio app new failed in hbase

2018-05-29 Thread Pat Ferrel
No, this is as expected. When you run pseudo-distributed everything internally is configured as if the services were on separate machines. See clustered instructions here: http://actionml.com/docs/small_ha_cluster This is to setup 3 machines running different parts and is not really the best

Re: PIO 0.12.1 with HDP Spark on YARN

2018-05-29 Thread Miller, Clifford
To add more details to this. When I attempt to execute my training job using the command 'pio train -- --master yarn' I get the exception that I've included below. Can anyone tell me how to correctly submit the training job or what setting I need to change to make this work. I've made not

Re: pio app new failed in hbase

2018-05-29 Thread Marco Goldin
i was able to solve the issue deleting hbase folder in hdfs with "hdfs dfs -rm -r /hbase" and restarting hbase. now app creation in pio is working again. I still wonder why this problem happen though, i'm running hbase in pseudo-distributed mode (for testing purposes everything, from spark to