I just did ./make-distribution.sh from the pio source directory.

On Fri, Mar 29, 2019, 12:08 Pat Ferrel <[email protected]> wrote:

> Templates have their own build.sbt. This means that if you upgrade a
> version of PIO you need to upgrade the dependencies in ALL your templates.
> So what you are calling a regression may just be that the UR needs to have
> upgraded dependencies.
>
> I’d be interested in helping but let’s move back to PIO 0.14.0 first. When
> you build PIO what is the exact command line?
>
>
> From: Michael Zhou <[email protected]>
> <[email protected]>
> Reply: [email protected] <[email protected]>
> <[email protected]>
> Date: March 20, 2019 at 12:05:26 PM
> To: [email protected] <[email protected]>
> <[email protected]>
> Subject:  Re: Wrong FS: file:/home/aml/ur/engine.json expected:
> hdfs://localhost:9000
>
> Update: This seems like a regression introduced by pio 0.14.0. It worked
> after I downgraded to pio 0.13.0.
> In particular, I suspect this diff
> https://github.com/apache/predictionio/pull/494/files#diff-167f4e9c1445b1f87aad1dead8da208c
>  to
> have caused the issue.
> Would be better if a committer can confirm this.
>
> On Wed, Mar 20, 2019 at 10:57 AM Michael Zhou <[email protected]>
> wrote:
>
>> I'm trying to run the integration test for the Universal Recommender.
>> However, I've been getting this error when doing "pio deploy":
>>
>> 2019-03-20 17:44:32,856 ERROR akka.actor.OneForOneStrategy
>> [pio-server-akka.actor.default-dispatcher-2] - Wrong FS:
>> file:/home/aml/ur/engine.json, expected: hdfs://localhost:9000
>> java.lang.IllegalArgumentException: Wrong FS:
>> file:/home/aml/ur/engine.json, expected: hdfs://localhost:9000
>>         at org.apache.hadoop.fs.FileSystem.checkPath(FileSystem.java:649)
>>         at
>> org.apache.hadoop.hdfs.DistributedFileSystem.getPathName(DistributedFileSystem.java:194)
>>         at
>> org.apache.hadoop.hdfs.DistributedFileSystem.access$000(DistributedFileSystem.java:106)
>>         at
>> org.apache.hadoop.hdfs.DistributedFileSystem$22.doCall(DistributedFileSystem.java:1305)
>>         at
>> org.apache.hadoop.hdfs.DistributedFileSystem$22.doCall(DistributedFileSystem.java:1301)
>>         at
>> org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
>>         at
>> org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:1317)
>>         at org.apache.hadoop.fs.FileSystem.exists(FileSystem.java:1426)
>>         at
>> org.apache.predictionio.workflow.EngineServerPluginContext$.stringFromFile(EngineServerPluginContext.scala:85)
>>         at
>> org.apache.predictionio.workflow.EngineServerPluginContext$.apply(EngineServerPluginContext.scala:58)
>>         at
>> org.apache.predictionio.workflow.PredictionServer.<init>(CreateServer.scala:424)
>>         at
>> org.apache.predictionio.workflow.CreateServer$.createPredictionServerWithEngine(CreateServer.scala:237)
>>         at
>> org.apache.predictionio.workflow.MasterActor.createServer(CreateServer.scala:389)
>>         at
>> org.apache.predictionio.workflow.MasterActor$$anonfun$receive$1.applyOrElse(CreateServer.scala:317)
>>         at akka.actor.Actor$class.aroundReceive(Actor.scala:517)
>>         at
>> org.apache.predictionio.workflow.MasterActor.aroundReceive(CreateServer.scala:259)
>>         at akka.actor.ActorCell.receiveMessage(ActorCell.scala:588)
>>         at akka.actor.ActorCell.invoke(ActorCell.scala:557)
>>         at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:258)
>>         at akka.dispatch.Mailbox.run(Mailbox.scala:225)
>>         at akka.dispatch.Mailbox.exec(Mailbox.scala:235)
>>         at
>> akka.dispatch.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
>>         at
>> akka.dispatch.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
>>         at
>> akka.dispatch.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
>>         at
>> akka.dispatch.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
>>
>> My pio-env.sh is as follows:
>>
>> SPARK_HOME=/usr/local/spark
>> ES_CONF_DIR=/usr/local/elasticsearch
>> HADOOP_CONF_DIR=/usr/local/hadoop/etc/hadoop
>> HBASE_CONF_DIR=/usr/local/hbase/conf
>>
>> PIO_FS_BASEDIR=$HOME/.pio_store
>> PIO_FS_ENGINESDIR=$PIO_FS_BASEDIR/engines
>> PIO_FS_TMPDIR=$PIO_FS_BASEDIR/tmp
>>
>> PIO_STORAGE_REPOSITORIES_METADATA_NAME=pio_meta
>> PIO_STORAGE_REPOSITORIES_METADATA_SOURCE=ELASTICSEARCH
>>
>> PIO_STORAGE_REPOSITORIES_EVENTDATA_NAME=pio_event
>> PIO_STORAGE_REPOSITORIES_EVENTDATA_SOURCE=HBASE
>>
>> PIO_STORAGE_REPOSITORIES_MODELDATA_NAME=pio_model
>> PIO_STORAGE_REPOSITORIES_MODELDATA_SOURCE=HDFS
>>
>> PIO_STORAGE_SOURCES_ELASTICSEARCH_TYPE=elasticsearch
>> PIO_STORAGE_SOURCES_ELASTICSEARCH_CLUSTERNAME=my-cluster
>> PIO_STORAGE_SOURCES_ELASTICSEARCH_HOSTS=localhost
>> PIO_STORAGE_SOURCES_ELASTICSEARCH_PORTS=9200
>> PIO_STORAGE_SOURCES_ELASTICSEARCH_HOME=/usr/local/elasticsearch
>>
>> PIO_STORAGE_SOURCES_HDFS_TYPE=hdfs
>> PIO_STORAGE_SOURCES_HDFS_PATH=/models
>>
>> PIO_STORAGE_SOURCES_HBASE_TYPE=hbase
>> PIO_STORAGE_SOURCES_HBASE_HOME=/usr/local/hbase
>> PIO_STORAGE_SOURCES_HBASE_HOSTS=localhost
>>
>> Any help would be appreciated.
>>
>

Reply via email to