Update: This seems like a regression introduced by pio 0.14.0. It worked after I downgraded to pio 0.13.0. In particular, I suspect this diff https://github.com/apache/predictionio/pull/494/files#diff-167f4e9c1445b1f87aad1dead8da208c to have caused the issue. Would be better if a committer can confirm this.
On Wed, Mar 20, 2019 at 10:57 AM Michael Zhou <[email protected]> wrote: > I'm trying to run the integration test for the Universal Recommender. > However, I've been getting this error when doing "pio deploy": > > 2019-03-20 17:44:32,856 ERROR akka.actor.OneForOneStrategy > [pio-server-akka.actor.default-dispatcher-2] - Wrong FS: > file:/home/aml/ur/engine.json, expected: hdfs://localhost:9000 > java.lang.IllegalArgumentException: Wrong FS: > file:/home/aml/ur/engine.json, expected: hdfs://localhost:9000 > at org.apache.hadoop.fs.FileSystem.checkPath(FileSystem.java:649) > at > org.apache.hadoop.hdfs.DistributedFileSystem.getPathName(DistributedFileSystem.java:194) > at > org.apache.hadoop.hdfs.DistributedFileSystem.access$000(DistributedFileSystem.java:106) > at > org.apache.hadoop.hdfs.DistributedFileSystem$22.doCall(DistributedFileSystem.java:1305) > at > org.apache.hadoop.hdfs.DistributedFileSystem$22.doCall(DistributedFileSystem.java:1301) > at > org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) > at > org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:1317) > at org.apache.hadoop.fs.FileSystem.exists(FileSystem.java:1426) > at > org.apache.predictionio.workflow.EngineServerPluginContext$.stringFromFile(EngineServerPluginContext.scala:85) > at > org.apache.predictionio.workflow.EngineServerPluginContext$.apply(EngineServerPluginContext.scala:58) > at > org.apache.predictionio.workflow.PredictionServer.<init>(CreateServer.scala:424) > at > org.apache.predictionio.workflow.CreateServer$.createPredictionServerWithEngine(CreateServer.scala:237) > at > org.apache.predictionio.workflow.MasterActor.createServer(CreateServer.scala:389) > at > org.apache.predictionio.workflow.MasterActor$$anonfun$receive$1.applyOrElse(CreateServer.scala:317) > at akka.actor.Actor$class.aroundReceive(Actor.scala:517) > at > org.apache.predictionio.workflow.MasterActor.aroundReceive(CreateServer.scala:259) > at akka.actor.ActorCell.receiveMessage(ActorCell.scala:588) > at akka.actor.ActorCell.invoke(ActorCell.scala:557) > at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:258) > at akka.dispatch.Mailbox.run(Mailbox.scala:225) > at akka.dispatch.Mailbox.exec(Mailbox.scala:235) > at > akka.dispatch.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) > at > akka.dispatch.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339) > at > akka.dispatch.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) > at > akka.dispatch.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) > > My pio-env.sh is as follows: > > SPARK_HOME=/usr/local/spark > ES_CONF_DIR=/usr/local/elasticsearch > HADOOP_CONF_DIR=/usr/local/hadoop/etc/hadoop > HBASE_CONF_DIR=/usr/local/hbase/conf > > PIO_FS_BASEDIR=$HOME/.pio_store > PIO_FS_ENGINESDIR=$PIO_FS_BASEDIR/engines > PIO_FS_TMPDIR=$PIO_FS_BASEDIR/tmp > > PIO_STORAGE_REPOSITORIES_METADATA_NAME=pio_meta > PIO_STORAGE_REPOSITORIES_METADATA_SOURCE=ELASTICSEARCH > > PIO_STORAGE_REPOSITORIES_EVENTDATA_NAME=pio_event > PIO_STORAGE_REPOSITORIES_EVENTDATA_SOURCE=HBASE > > PIO_STORAGE_REPOSITORIES_MODELDATA_NAME=pio_model > PIO_STORAGE_REPOSITORIES_MODELDATA_SOURCE=HDFS > > PIO_STORAGE_SOURCES_ELASTICSEARCH_TYPE=elasticsearch > PIO_STORAGE_SOURCES_ELASTICSEARCH_CLUSTERNAME=my-cluster > PIO_STORAGE_SOURCES_ELASTICSEARCH_HOSTS=localhost > PIO_STORAGE_SOURCES_ELASTICSEARCH_PORTS=9200 > PIO_STORAGE_SOURCES_ELASTICSEARCH_HOME=/usr/local/elasticsearch > > PIO_STORAGE_SOURCES_HDFS_TYPE=hdfs > PIO_STORAGE_SOURCES_HDFS_PATH=/models > > PIO_STORAGE_SOURCES_HBASE_TYPE=hbase > PIO_STORAGE_SOURCES_HBASE_HOME=/usr/local/hbase > PIO_STORAGE_SOURCES_HBASE_HOSTS=localhost > > Any help would be appreciated. >
