Re: Universal Recommender and PredictionIO 0.12.0 incompatibility

2017-10-04 Thread Vaghawan Ojha
Thank you for letting us know.

Thanks

On Thu, Oct 5, 2017 at 1:49 AM, Pat Ferrel  wrote:

> It looks like PIO 0.12.0 will require a code change in the UR. PIO changed
> ES1 support drastically when ES5 support was added and broke the UR code.
>
> We will do a quick fix to the template to address this. In the meantime
> stay on PIO 0.11.0 if you need the UR.


Re: Deploying multiple Prediction IO engines simultaneously

2017-08-09 Thread Vaghawan Ojha
Hi,

Although, this page (https://predictionio.incubator.apache.org/deploy/) and
this page (https://predictionio.incubator.apache.org/deploy/enginevariants/)
both are silent in regarding how many different Prediction IO engines you
could deploy on a machine. I don't think there is a limitation on that, but
I'm not sure if a single machine can do all the jobs of multiple engines
with huge datasets streaming on.

However, I've had deployed 10 multiple engines in a production server, each
with different events and in some cases with different templates as well.


Thanks
Vaghawan

On Thu, Aug 10, 2017 at 5:28 AM, Anshu Rajendra 
wrote:

> Hi,
>
> Is there a limit to how many different Prediction-IO engines we can deploy
> on a machine simultaneously using different ports?
>
> Please let me know.
>
> Thanks,
> Anshu Rajendra
>


Re: Retrieve all events with a list of entityId(s) as input

2017-08-04 Thread Vaghawan Ojha
Hi,

I've done this by querying the EventServer, but you would have to loop with
the entityId(s) which means that it gonna be slow. If you have large list
to loop then there are more chances of timeout response.

But anyway this is possible. You would just have to query the EventServer

   #get the user data count.
> getuserdatacount =
> requests.get(settings.EVENTSERVER+'/events.json?accessKey=dyJsZXASyAwpLmMRKKLckPaApPNDYzTMKxGjYfCqizv9Y-C8843Xqnw8q37O1deQ='+checkuser+'=-1').text
> #print getuserdatacount
> try:
> loadedjson = json.loads(getuserdatacount)
> usereventcount = len(loadedjson)
> except ValueError, e:
> usereventcount=' Error Fetching'


This is a python code retrieving all the events of a particular user from
the Event Server. Hope this helps.


Thanks
Vaghawan

On Fri, Aug 4, 2017 at 1:34 PM, Dominik Bartkowski 
wrote:

> Hello,
>
> I would like to retrieve all events (items in my case) for the list of
> entityId(s). These items are returned by the Engine as itemScores. I
> thought that it is possible to query the EventClient with a list of item
> ids. But the Event Server does not support these queries.
>
> Is there a workaround?
>
> I know that we can modify the Data Source and the Engine to retrieve
> properties. Like in the example below:
> http://predictionio.incubator.apache.org/templates/similarpr
> oduct/return-item-properties/
>
> But on the other hand, the engine is not supposed to return properties -
> it knows nothing about the materials. It is not supposed to. My first
> thought would be to query the Event Server. Am I  getting it wrong?
>
> With best regards,
> Dominik Bartkowski
>


Re: how to Clear trained data in predictionIO 0.11.0

2017-07-31 Thread Vaghawan Ojha
Ok, which template were you using (Just curious). One other way would be
deleting the whole app, and creating a new one. But that too isn't an
efficient way.

When you delete the data and  retain the old model should be changed.
Something must be wrong.


Thanks

On Mon, Jul 31, 2017 at 2:36 PM, Mahesh Hegde <mahesh.he...@tracxn.com>
wrote:

> Does not seem to be permission issue as well, i verified it by, trying to
> run train just after deleting the event data, which gave error as expected.
> Yes i have verified app name in delete command.
>
> As a temporary fix i deleted all model data from elasticsearch, after
> which newly trained data was picked up just fine. But looks like this is
> not the proper way to do. Will post updates if I find any other findings.
>
> -Mahesh
>
> On Mon, Jul 31, 2017 at 2:11 PM, Vaghawan Ojha <vaghawan...@gmail.com>
> wrote:
>
>> Could you please make sure, you deleted the data of a right app, I meant
>> the app should be correct. Is there any permission issues?
>>
>> On Mon, Jul 31, 2017 at 2:19 PM, Mahesh Hegde <mahesh.he...@tracxn.com>
>> wrote:
>>
>>> Also after deleting the data, i built , trained and deployed, still i am
>>> getting old label in predicted result.
>>>
>>> -Mahesh
>>>
>>> On Mon, Jul 31, 2017 at 2:03 PM, Mahesh Hegde <mahesh.he...@tracxn.com>
>>> wrote:
>>>
>>>> I retrained with new data after deleting the old event data with pio
>>>> app data-delete my_app_name , even then my query is return old label "spam"
>>>> for predicted result, which is from email spam example.
>>>>
>>>> Also I tried to change the models directory used by localfs , but it
>>>> looks like its refrering to old model file and gave
>>>> error models/pio_modelAV2XQxw-3F2NATbCwlg3 file doesnt exist.
>>>>
>>>> -Mahesh
>>>>
>>>> On Mon, Jul 31, 2017 at 1:19 PM, Vaghawan Ojha <vaghawan...@gmail.com>
>>>> wrote:
>>>>
>>>>> I think, when you retrain the model, the trained model will
>>>>> automatically change.
>>>>>
>>>>> Thanks
>>>>>
>>>>>
>>>>> On Mon, Jul 31, 2017 at 1:33 PM, Mahesh Hegde <mahesh.he...@tracxn.com
>>>>> > wrote:
>>>>>
>>>>>> pio app delete-data deletes only event data,
>>>>>>
>>>>>> I wanted to know , how to clean the trained data.
>>>>>> I am using elasticsearch (metadata) , hbase (eventdata) and
>>>>>> localfs(model data) stack
>>>>>>
>>>>>> -Mahesh
>>>>>>
>>>>>> This email is subject to Tracxn's Email Policy
>>>>>> <https://tracxn.com/emailpolicy>
>>>>>
>>>>>
>>>>>
>>>>
>>>
>>> This email is subject to Tracxn's Email Policy
>>> <https://tracxn.com/emailpolicy>
>>>
>>
>>
>
> This email is subject to Tracxn's Email Policy
> <https://tracxn.com/emailpolicy>
>


Re: how to Clear trained data in predictionIO 0.11.0

2017-07-31 Thread Vaghawan Ojha
Could you please make sure, you deleted the data of a right app, I meant
the app should be correct. Is there any permission issues?

On Mon, Jul 31, 2017 at 2:19 PM, Mahesh Hegde <mahesh.he...@tracxn.com>
wrote:

> Also after deleting the data, i built , trained and deployed, still i am
> getting old label in predicted result.
>
> -Mahesh
>
> On Mon, Jul 31, 2017 at 2:03 PM, Mahesh Hegde <mahesh.he...@tracxn.com>
> wrote:
>
>> I retrained with new data after deleting the old event data with pio app
>> data-delete my_app_name , even then my query is return old label "spam" for
>> predicted result, which is from email spam example.
>>
>> Also I tried to change the models directory used by localfs , but it
>> looks like its refrering to old model file and gave
>> error models/pio_modelAV2XQxw-3F2NATbCwlg3 file doesnt exist.
>>
>> -Mahesh
>>
>> On Mon, Jul 31, 2017 at 1:19 PM, Vaghawan Ojha <vaghawan...@gmail.com>
>> wrote:
>>
>>> I think, when you retrain the model, the trained model will
>>> automatically change.
>>>
>>> Thanks
>>>
>>>
>>> On Mon, Jul 31, 2017 at 1:33 PM, Mahesh Hegde <mahesh.he...@tracxn.com>
>>> wrote:
>>>
>>>> pio app delete-data deletes only event data,
>>>>
>>>> I wanted to know , how to clean the trained data.
>>>> I am using elasticsearch (metadata) , hbase (eventdata) and
>>>> localfs(model data) stack
>>>>
>>>> -Mahesh
>>>>
>>>> This email is subject to Tracxn's Email Policy
>>>> <https://tracxn.com/emailpolicy>
>>>
>>>
>>>
>>
>
> This email is subject to Tracxn's Email Policy
> <https://tracxn.com/emailpolicy>
>


Re: how to Clear trained data in predictionIO 0.11.0

2017-07-31 Thread Vaghawan Ojha
I think, when you retrain the model, the trained model will automatically
change.

Thanks


On Mon, Jul 31, 2017 at 1:33 PM, Mahesh Hegde 
wrote:

> pio app delete-data deletes only event data,
>
> I wanted to know , how to clean the trained data.
> I am using elasticsearch (metadata) , hbase (eventdata) and localfs(model
> data) stack
>
> -Mahesh
>
> This email is subject to Tracxn's Email Policy
> 


Re: Programitically trigger pio train and deploy

2017-07-30 Thread Vaghawan Ojha
Hi,

It's possible by running the command with a shell script. For me, I trigger
shell script from the program if I would need the command to run. Also, I'm
using a cron job to regulate the pio train and deploy process using a shell
script which is being triggered by the application.

However, I've not done that in Java, but I would do similar with the java
application as well.

You could fine more about this in the following links:
https://predictionio.incubator.apache.org/deploy/
https://github.com/apache/incubator-predictionio/tree/develop/examples/redeploy-script

Hope this helps.

Thanks
Vaghawan

On Sun, Jul 30, 2017 at 12:19 AM, Anshu Rajendra 
wrote:

> Hi,
>
> Is it possible to programmatically trigger pio commands like 'pio train'
> and 'pio deploy' from within a java application or only from command line
> within the app? We can create events and predict using an sdk but can we
> also deploy any model programmatically?
>
> Any help would be very appreciated.
>
> Thanks,
> Anshu
>


Re: Pio build success with error, pio train is faililng.

2017-07-28 Thread Vaghawan Ojha
Hi Pat,

Actually UR has helped us a lot, and I've been making a guide along with
the troubles I encountered in a stand alone installation of UR. So there is
a use case like Restaurant/Food Recommendation system using UR. Once done,
I would be submitting a PR on the doc repo.

Thank you
Vaghawan

On Thu, Jul 27, 2017 at 9:45 PM, Pat Ferrel <p...@occamsmachete.com> wrote:

> Yes, a great article but it and the tapster demo do not use the UR.
>
>
> On Jul 27, 2017, at 4:52 AM, Vaghawan Ojha <vaghawan...@gmail.com> wrote:
>
> Cool,
>
> Now every working template has engine.json and appName field as far as I
> know.
>
> Great it worked.
>
> Thanks
> Vaghawan
>
> On Thu, Jul 27, 2017 at 5:13 PM, Darshan A.N. <darshanan...@gmail.com>
> wrote:
>
>> hi,
>> i think i kind of found where did i do mistake, its not my mistake
>> actually... its document makers mistake.
>> In DEMO-TAPSTER , they had mentioned about appID, which suits the old
>> template. right now i discontinued following that doc. I' m using
>> https://predictionio.incubator.apache.org/templates/
>> similarproduct/quickstart/ for further, until i get deserved output.
>>
>>
>> You may see appId in engine.json instead, which means you are using old
>> template. In this case, make sure the appId defined in the file match
>> your *App ID*. Alternatively, you can download the latest version of the
>> template or follow our upgrade instructions
>> <https://predictionio.incubator.apache.org/resources/upgrade/#upgrade-to-0.9.2>
>> to modify the template to use appName as parameter.
>>
>>
>>
>> thanks VO for sending me the QuickStart link.
>> -DAN
>>
>>
>>
>> On Thu, Jul 27, 2017 at 2:45 PM, Darshan A.N. <darshanan...@gmail.com>
>> wrote:
>>
>>> VO,
>>> thanks, i ll look into it. i just sent u chat request, could u please
>>> accept it. i ll not bother u much.
>>>
>>> -DAN
>>>
>>> On Thu, Jul 27, 2017 at 2:41 PM, Vaghawan Ojha <vaghawan...@gmail.com>
>>> wrote:
>>>
>>>> Hi Darshan,
>>>>
>>>> Thanks, yes I myself had undergone with so many difficulties during
>>>> installation so I had covered them in that article. Also I've been trying
>>>> to make one or two examples with ur in the pio docs.
>>>>
>>>> For your current error, you need to change the appName in engine.json
>>>> and rebuild your app again.
>>>>
>>>> On Thu, Jul 27, 2017 at 2:49 PM, Darshan A.N. <darshanan...@gmail.com>
>>>> wrote:
>>>>
>>>>> hi VO,
>>>>> thats really impressive article you have written. i need to install
>>>>> that DEMO-tapster. my repository is mysql, could you please help me 
>>>>> solving
>>>>> it?
>>>>> thanks and cheers for ur article, i suppose it must be helpful for lot
>>>>> of noobies like me.
>>>>> -DAN
>>>>>
>>>>> On Wed, Jul 26, 2017 at 3:37 PM, Vaghawan Ojha <vaghawan...@gmail.com>
>>>>> wrote:
>>>>>
>>>>>> Hi Darshan,
>>>>>>
>>>>>> I had written a brief article on medium to start from scratch, you
>>>>>> could follow that as well. https://medium.freecodec
>>>>>> amp.org/building-an-recommendation-engine-with-apache-predic
>>>>>> tion-io-ml-server-aed0319e0d8
>>>>>> here's the link. I myself had faced lots of problems while
>>>>>> installation, but pio and ur community is great for instant support. I've
>>>>>> included common errors and there troubleshooting as well.
>>>>>>
>>>>>> Hope this helps.
>>>>>>
>>>>>> Thanks
>>>>>> Vaghawan
>>>>>>
>>>>>> On Wed, Jul 26, 2017 at 3:37 PM, Tom Chan <yukhei.c...@gmail.com>
>>>>>> wrote:
>>>>>>
>>>>>>> darshan@darshu:~/PredictionIO/tapster-episode-similar/src/main/scala$
>>>>>>> pio train
>>>>>>>
>>>>>>> Can you try the command from the tapster-episode-similar directory?
>>>>>>>
>>>>>>> Tom
>>>>>>>
>>>>>>> On Jul 26, 2017 2:45 AM, "Darshan A.N." <darshanan...@gmail.com>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> hi team

Re: Deleting / Reversing events in the event server

2017-07-27 Thread Vaghawan Ojha
Yes Mattz, that's why I pointed that, it's more easy and effective I think.
You could just save the blocked items somewhere that may be even in your
relational database in which your original application resides, and update
them accordingly, and using them while querying. Or else you can save them
in eventserver as a user properties and retrive them while querying. Both
way it should be of help.

Thanks
Vaghawan

On Thu, Jul 27, 2017 at 11:44 AM, Mattz <eatingka...@gmail.com> wrote:

> Thanks Vaghawan. I think using the "blacklistItems" might be better for my
> use case rather than passing in the "block" events. I didn't think of this.
>
> On Thu, Jul 27, 2017 at 11:15 AM, Vaghawan Ojha <vaghawan...@gmail.com>
> wrote:
>
>> So you're not sending the blocklisted items in query like this right?
>>
>> "blacklistItems": ["itemId1", "itemId2", ...]
>>
>> So, you have those blocked events, then you must be able to update them
>> accordingly just the way you set them as blocked. Or else I'm still not
>> clear (I'm sorry in that case).
>>
>> Thanks
>> Vaghawan
>>
>> On Thu, Jul 27, 2017 at 11:26 AM, Mattz <eatingka...@gmail.com> wrote:
>>
>>> This is how I am using now. I send a blocked event if a user blocks an
>>> item. So, when recommending items for the user, this item does not show up.
>>> I am using the "universal recommender" template.
>>>
>>> On Thu, Jul 27, 2017 at 11:09 AM, Vaghawan Ojha <vaghawan...@gmail.com>
>>> wrote:
>>>
>>>> Oh Yes Mattz, I seems to have misunderstood your situation. I thought
>>>> that you were having that blocked prop as a properties or something. My
>>>> thought must have been totally wrong.
>>>>
>>>> I'm not clear if you're sending the blocked items in query or something
>>>> else? Also which template you have been using?
>>>>
>>>> Thanks
>>>>
>>>>
>>>>
>>>> On Thu, Jul 27, 2017 at 11:19 AM, Mattz <eatingka...@gmail.com> wrote:
>>>>
>>>>> Thanks Vaghawan. "$set" and "$unset" are used to set/unset item
>>>>> properties right? Since these are not like the other events, these items
>>>>> would still be returned unless I manually find all the things blocked and
>>>>> filter from list to get my result set. But, if I send "blocked" as a
>>>>> regular event, PIO automatically filters out items that were blocked by 
>>>>> the
>>>>> user.
>>>>>
>>>>> On Thu, Jul 27, 2017 at 10:59 AM, Vaghawan Ojha <vaghawan...@gmail.com
>>>>> > wrote:
>>>>>
>>>>>> Hi Mattz,
>>>>>>
>>>>>> You could just use $set, $unset to achieve this. When user wish to
>>>>>> unblock the item, just send in another request in the event server to
>>>>>> update the event. This would solve your problem from as much as I know 
>>>>>> from
>>>>>> using pio eventserver.
>>>>>>
>>>>>> you could find the examples here. https://predictionio.inc
>>>>>> ubator.apache.org/datacollection/eventmodel/
>>>>>>
>>>>>> Thanks
>>>>>> Vaghawan
>>>>>>
>>>>>> On Thu, Jul 27, 2017 at 11:09 AM, Mattz <eatingka...@gmail.com>
>>>>>> wrote:
>>>>>>
>>>>>>> Hello,
>>>>>>>
>>>>>>> I have a use case where I am sending a "blocked" event for a user
>>>>>>> for a particular item. When this happens, PIO eliminates this item from 
>>>>>>> the
>>>>>>> result set. But, the user can choose to "unblock" this item at a later
>>>>>>> time. I am wondering how to handle this. Should I somehow delete the 
>>>>>>> event
>>>>>>> from the even server? If yes, how? Is there any other way to solve this?
>>>>>>>
>>>>>>> Thanks!
>>>>>>>
>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>


Re: Deleting / Reversing events in the event server

2017-07-26 Thread Vaghawan Ojha
So you're not sending the blocklisted items in query like this right?

"blacklistItems": ["itemId1", "itemId2", ...]

So, you have those blocked events, then you must be able to update them
accordingly just the way you set them as blocked. Or else I'm still not
clear (I'm sorry in that case).

Thanks
Vaghawan

On Thu, Jul 27, 2017 at 11:26 AM, Mattz <eatingka...@gmail.com> wrote:

> This is how I am using now. I send a blocked event if a user blocks an
> item. So, when recommending items for the user, this item does not show up.
> I am using the "universal recommender" template.
>
> On Thu, Jul 27, 2017 at 11:09 AM, Vaghawan Ojha <vaghawan...@gmail.com>
> wrote:
>
>> Oh Yes Mattz, I seems to have misunderstood your situation. I thought
>> that you were having that blocked prop as a properties or something. My
>> thought must have been totally wrong.
>>
>> I'm not clear if you're sending the blocked items in query or something
>> else? Also which template you have been using?
>>
>> Thanks
>>
>>
>>
>> On Thu, Jul 27, 2017 at 11:19 AM, Mattz <eatingka...@gmail.com> wrote:
>>
>>> Thanks Vaghawan. "$set" and "$unset" are used to set/unset item
>>> properties right? Since these are not like the other events, these items
>>> would still be returned unless I manually find all the things blocked and
>>> filter from list to get my result set. But, if I send "blocked" as a
>>> regular event, PIO automatically filters out items that were blocked by the
>>> user.
>>>
>>> On Thu, Jul 27, 2017 at 10:59 AM, Vaghawan Ojha <vaghawan...@gmail.com>
>>> wrote:
>>>
>>>> Hi Mattz,
>>>>
>>>> You could just use $set, $unset to achieve this. When user wish to
>>>> unblock the item, just send in another request in the event server to
>>>> update the event. This would solve your problem from as much as I know from
>>>> using pio eventserver.
>>>>
>>>> you could find the examples here. https://predictionio.inc
>>>> ubator.apache.org/datacollection/eventmodel/
>>>>
>>>> Thanks
>>>> Vaghawan
>>>>
>>>> On Thu, Jul 27, 2017 at 11:09 AM, Mattz <eatingka...@gmail.com> wrote:
>>>>
>>>>> Hello,
>>>>>
>>>>> I have a use case where I am sending a "blocked" event for a user for
>>>>> a particular item. When this happens, PIO eliminates this item from the
>>>>> result set. But, the user can choose to "unblock" this item at a later
>>>>> time. I am wondering how to handle this. Should I somehow delete the event
>>>>> from the even server? If yes, how? Is there any other way to solve this?
>>>>>
>>>>> Thanks!
>>>>>
>>>>
>>>>
>>>
>>
>


Re: Deleting / Reversing events in the event server

2017-07-26 Thread Vaghawan Ojha
Oh Yes Mattz, I seems to have misunderstood your situation. I thought that
you were having that blocked prop as a properties or something. My thought
must have been totally wrong.

I'm not clear if you're sending the blocked items in query or something
else? Also which template you have been using?

Thanks



On Thu, Jul 27, 2017 at 11:19 AM, Mattz <eatingka...@gmail.com> wrote:

> Thanks Vaghawan. "$set" and "$unset" are used to set/unset item properties
> right? Since these are not like the other events, these items would still
> be returned unless I manually find all the things blocked and filter from
> list to get my result set. But, if I send "blocked" as a regular event, PIO
> automatically filters out items that were blocked by the user.
>
> On Thu, Jul 27, 2017 at 10:59 AM, Vaghawan Ojha <vaghawan...@gmail.com>
> wrote:
>
>> Hi Mattz,
>>
>> You could just use $set, $unset to achieve this. When user wish to
>> unblock the item, just send in another request in the event server to
>> update the event. This would solve your problem from as much as I know from
>> using pio eventserver.
>>
>> you could find the examples here. https://predictionio.inc
>> ubator.apache.org/datacollection/eventmodel/
>>
>> Thanks
>> Vaghawan
>>
>> On Thu, Jul 27, 2017 at 11:09 AM, Mattz <eatingka...@gmail.com> wrote:
>>
>>> Hello,
>>>
>>> I have a use case where I am sending a "blocked" event for a user for a
>>> particular item. When this happens, PIO eliminates this item from the
>>> result set. But, the user can choose to "unblock" this item at a later
>>> time. I am wondering how to handle this. Should I somehow delete the event
>>> from the even server? If yes, how? Is there any other way to solve this?
>>>
>>> Thanks!
>>>
>>
>>
>


Re: Deleting / Reversing events in the event server

2017-07-26 Thread Vaghawan Ojha
Hi Mattz,

You could just use $set, $unset to achieve this. When user wish to unblock
the item, just send in another request in the event server to update the
event. This would solve your problem from as much as I know from using pio
eventserver.

you could find the examples here.
https://predictionio.incubator.apache.org/datacollection/eventmodel/

Thanks
Vaghawan

On Thu, Jul 27, 2017 at 11:09 AM, Mattz  wrote:

> Hello,
>
> I have a use case where I am sending a "blocked" event for a user for a
> particular item. When this happens, PIO eliminates this item from the
> result set. But, the user can choose to "unblock" this item at a later
> time. I am wondering how to handle this. Should I somehow delete the event
> from the even server? If yes, how? Is there any other way to solve this?
>
> Thanks!
>


Re: Error while doing client.create in Prediction IO Java Client

2017-07-26 Thread Vaghawan Ojha
Hi,

Sorry, but I'm not comfortable with Java, I've only worked with python SDK,
and I can only guess from your error message, so I suppose it must lie here
"NoSuchMethodError" but only guessing.

So you would have to wait for someone who have worked through it.

Thanks

On Thu, Jul 27, 2017 at 9:47 AM, Anshu Rajendra <anshurajen...@gmail.com>
wrote:

> Hi,
>
> I am able to create events using curl. It seems like a dependency issue to
> me between io.netty and com.ning.httpckirnt but I don't know how to find
> the compatible versions for them to avoid this error. Any pointers?
>
> Thanks,
> Anshu
>
> On 26-Jul-2017 8:50 PM, "Vaghawan Ojha" <vaghawan...@gmail.com> wrote:
>
>> Hi,
>>
>> Are you able to import the events using curl?
>>
>> I think you could probably see this line "lang.NoSuchMethodError:
>> org.jboss.netty.handler.codec.http.HttpRequest.setHeader(Lja
>> va/lang/String;Ljava/lang/Object;)V "
>>
>> I've not worked with java app, but have worked with python sdk. Probably
>> something wrong with your request if you're able to import the events using
>> curl.
>>
>> Thanks
>>
>> On Thu, Jul 27, 2017 at 3:18 AM, Anshu Rajendra <anshurajen...@gmail.com>
>> wrote:
>>
>>> Hi,
>>>
>>> When I try to create a event in my java app using io.prediction client,
>>> I get this error :
>>>
>>> ```org.jboss.netty.handler.codec.http.HttpRequest.setHeader(
>>> Ljava/lang/String;Ljava/lang/Object;)V"java.lang.NoSuchMethodError:
>>> org.jboss.netty.handler.codec.http.HttpRequest.setHeader(Lja
>>> va/lang/String;Ljava/lang/Object;)V at com.ning.http.client.providers
>>> .netty.NettyAsyncHttpProvider.construct(NettyAsyncHttpProvider.java:704)
>>> at com.ning.http.client.providers.netty.NettyAsyncHttpProvider.
>>> buildRequest(NettyAsyncHttpProvider.java:655) at
>>> com.ning.http.client.providers.netty.NettyConnectListener$Bu
>>> ilder.build(NettyConnectListener.java:145) at
>>> com.ning.http.client.providers.netty.NettyAsyncHttpProvider.
>>> doConnect(NettyAsyncHttpProvider.java:1069) at
>>> com.ning.http.client.providers.netty.NettyAsyncHttpProvider.
>>> execute(NettyAsyncHttpProvider.java:938) at
>>> com.ning.http.client.AsyncHttpClient.executeRequest(AsyncHttpClient.java:499)
>>> at io.prediction.EventClient.createEventAsFuture(EventClient.java:116)
>>> at io.prediction.EventClient.createEvent(EventClient.java:131) ```
>>>
>>> It seems to be a library issue but I cannot figure out. Can someone help
>>> me debug this error?
>>>
>>> Thanks,
>>> Anshu
>>
>>
>>


Re: Exception: Job aborted due to stage failure: Task 0 in stage 3.0 failed 1 times,

2017-07-08 Thread Vaghawan Ojha
Hi,

As much as I can understand from the error message, the Null Exception
happens when you have wrong data or no data at all. Please make sure
everything is fine in your data.

Sometimes, the event store can store the data you provide, but the template
you might be using may need other kind of data, so please make sure you're
following the right doc and providing the right kind of data.

Thanks

On Sat, Jul 8, 2017 at 2:39 PM, Sebastian Fix 
wrote:

> Hey Keneth
>
> Thank for the answer.
>
> I’managed to import the following data:
>
>
> After running „pio build“ i get the following error during the „pio train“
> process:
>
> [ERROR] [Executor] Exception in task 0.0 in stage 3.0 (TID 9)
> [WARN] [TaskSetManager] Lost task 0.0 in stage 3.0 (TID 9, localhost):
> java.lang.NullPointerException
>
> [ERROR] [TaskSetManager] Task 0 in stage 3.0 failed 1 times; aborting job
> Exception in thread "main" org.apache.spark.SparkException: Job aborted
> due to stage failure: Task 0 in stage 3.0 failed 1 times, most recent
> failure: Lost task 0.0 in stage 3.0 (TID 9, localhost):
> java.lang.NullPointerException
>
> Driver stacktrace:
> at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$
> scheduler$DAGScheduler$$failJobAndIndependentStages(
> DAGScheduler.scala:1431)
> Caused by: java.lang.NullPointerException
>
>
> Best regards
> Sebastian
>
>
> Am 08.07.2017 um 09:52 schrieb Kenneth Chan :
>
> re: " i’m always getting an exception with the initial load from User &
> Product"
> what's the error you got?
>
> On Fri, Jul 7, 2017 at 12:46 PM, Sebastian Fix  > wrote:
>
> Hello together,
>
> I finally got the System running and would like to import some initial
> data now.
>
> I’m using the E-Commerce-Template: http://predictionio.incubator.apache.or
> g/templates/ecommercerecommendation/quickstart/
>
> I've the following data’s i would like to import:
>
>- User data
>
>
>- Products data
>
>
>- Transaction data
>
>
> When i try to use the *pio import --appid 1 --input import/data-user.json* 
> command
> i’m always getting an exception with the initial load from User & Product.
> My .json file is following the structure mentioned on the tutorial.
>
>
>- For User Data: {"event":"$set","entityType":"
>user","entityId":"user001","eventTime":““}
>- For Products  {"event":"$set","entityType“:
>"item","entityId“:"product001","eventTime":““}
>
>
>- For Transactions: {"entityId“:"user001","entityType":"user","eve
>nt":"buy","targetEntityType":"item","targetEntityId“:"product001“}
>
>
> Only the Import for Transactions is working I found this template here:
> https://gist.github.com/vaghawan/0a5fb8ddb85e03631dd50
> 0d7c8f0677d#file-data-sample-json
>
> So my quesiton, how can I import Data’s like
>
>
>- User Data
>- User Categories
>- Product Data
>- Product Categories
>
>
> with a .json file? What is the file structure?
>
> Thanks for any help,
> Sebastian
>
>
>
>
>


Re: Error while training : NegativeArraySizeException

2017-06-07 Thread Vaghawan Ojha
Also what version of UR you're into? Is it the latest one? I've only worked
with UR 0.50 .

On Wed, Jun 7, 2017 at 3:12 PM, Vaghawan Ojha <vaghawan...@gmail.com> wrote:

> Yes you need to build the app again when you change something in the
> engine.json. That is every time when you change something in engine.json.
>
> Make sure the data corresponds to the same app which you have provided in
> the engine.json.
>
> Yes you can test example instigation test in UR
> with ./examples/integration-test this command.
>
> You can find more in here http://actionml.com/docs/ur_quickstart .
>
> On Wed, Jun 7, 2017 at 3:07 PM, Bruno LEBON <b.le...@redfakir.fr> wrote:
>
>> Yes the three event types that I defined in the engine.json exist in my
>> dataset, facet is my primary, I checked that it exists.
>>
>> I think it is not needed to build again when changing something in the
>> engine.json, as the file is read in the process but I built it and tried
>> again and I still have the same error.
>>
>> What is this example-intrigration? I dont know about this. Where can I
>> find this script?
>>
>> 2017-06-07 11:11 GMT+02:00 Vaghawan Ojha <vaghawan...@gmail.com>:
>>
>>> Hi,
>>>
>>> For me this problem had happened when I had mistaken my primary events.
>>> The first eventName in the eventName array "eventNames":
>>> ["facet","view","search"] is primary. There is that event in your data.
>>>
>>> Did you make sure, you built the app again when you changed the
>>> eventName in engine.json?
>>>
>>> Also you could varify everything's fine with UR with
>>> ./example-intrigration.
>>>
>>> Thanks
>>>
>>> On Wed, Jun 7, 2017 at 2:49 PM, Bruno LEBON <b.le...@redfakir.fr> wrote:
>>>
>>>> Thanks for your answer.
>>>>
>>>> *You could explicitly do *
>>>>
>>>>
>>>> *pio train -- --master spark://localhost:7077 --driver-memory 16G
>>>> --executor-memory 24G *
>>>>
>>>> *and change the spark master url and the memories configuration. And
>>>> see if that works. *
>>>>
>>>> Yes that is the command I use to launch the train, except I am on a
>>>> cluster, so Spark is not local. Here is mine:
>>>>  pio train -- --master spark://master:7077 --driver-memory 4g
>>>> --executor-memory 10g
>>>>
>>>> The train works with different datasets, it also works with this
>>>> dataset when I skip the event type *view*. So my guess is that there
>>>> is something about this event type, either in the data but the data looks
>>>> fine to me, or maybe there is a problem when I use more than two types of
>>>> event (this is the first time I have more than two, however I can't believe
>>>> that the problem is related the a number of event types).
>>>>
>>>> The spelling is the same in the event sent to the eventserver ( *view *)
>>>> and in the engine.json ( *view *).
>>>>
>>>> I am reading the code to figure out where this error comes from.
>>>>
>>>>
>>>>
>>>> 2017-06-07 10:17 GMT+02:00 Vaghawan Ojha <vaghawan...@gmail.com>:
>>>>
>>>>> You could explicitly do
>>>>>
>>>>> pio train -- --master spark://localhost:7077 --driver-memory 16G
>>>>> --executor-memory 24G
>>>>>
>>>>> and change the spark master url and the memories configuration. And
>>>>> see if that works.
>>>>>
>>>>> Thanks
>>>>>
>>>>> On Wed, Jun 7, 2017 at 1:55 PM, Bruno LEBON <b.le...@redfakir.fr>
>>>>> wrote:
>>>>>
>>>>>> Hi,
>>>>>>
>>>>>> Using UR with PIO 0.10 I am trying to train my dataset. In return I
>>>>>> get the following error:
>>>>>>
>>>>>> *...*
>>>>>> *[INFO] [DataSource] Received events List(facet, view, search)*
>>>>>> *[INFO] [DataSource] Number of events List(5, 4, 6)*
>>>>>> *[INFO] [Engine$] org.template.TrainingData does not support data
>>>>>> sanity check. Skipping check.*
>>>>>> *[INFO] [Engine$] org.template.PreparedData does not support data
>>>>>> sanity check. Skipping check.*
>>>>>> *[INFO] [URAlgorit

Re: Error while training : NegativeArraySizeException

2017-06-07 Thread Vaghawan Ojha
Hi,

For me this problem had happened when I had mistaken my primary events. The
first eventName in the eventName array "eventNames":
["facet","view","search"] is primary. There is that event in your data.

Did you make sure, you built the app again when you changed the eventName
in engine.json?

Also you could varify everything's fine with UR with
./example-intrigration.

Thanks

On Wed, Jun 7, 2017 at 2:49 PM, Bruno LEBON <b.le...@redfakir.fr> wrote:

> Thanks for your answer.
>
> *You could explicitly do *
>
>
> *pio train -- --master spark://localhost:7077 --driver-memory 16G
> --executor-memory 24G *
>
> *and change the spark master url and the memories configuration. And see
> if that works. *
>
> Yes that is the command I use to launch the train, except I am on a
> cluster, so Spark is not local. Here is mine:
>  pio train -- --master spark://master:7077 --driver-memory 4g
> --executor-memory 10g
>
> The train works with different datasets, it also works with this dataset
> when I skip the event type *view*. So my guess is that there is something
> about this event type, either in the data but the data looks fine to me, or
> maybe there is a problem when I use more than two types of event (this is
> the first time I have more than two, however I can't believe that the
> problem is related the a number of event types).
>
> The spelling is the same in the event sent to the eventserver ( *view *)
> and in the engine.json ( *view *).
>
> I am reading the code to figure out where this error comes from.
>
>
>
> 2017-06-07 10:17 GMT+02:00 Vaghawan Ojha <vaghawan...@gmail.com>:
>
>> You could explicitly do
>>
>> pio train -- --master spark://localhost:7077 --driver-memory 16G
>> --executor-memory 24G
>>
>> and change the spark master url and the memories configuration. And see
>> if that works.
>>
>> Thanks
>>
>> On Wed, Jun 7, 2017 at 1:55 PM, Bruno LEBON <b.le...@redfakir.fr> wrote:
>>
>>> Hi,
>>>
>>> Using UR with PIO 0.10 I am trying to train my dataset. In return I get
>>> the following error:
>>>
>>> *...*
>>> *[INFO] [DataSource] Received events List(facet, view, search)*
>>> *[INFO] [DataSource] Number of events List(5, 4, 6)*
>>> *[INFO] [Engine$] org.template.TrainingData does not support data sanity
>>> check. Skipping check.*
>>> *[INFO] [Engine$] org.template.PreparedData does not support data sanity
>>> check. Skipping check.*
>>> *[INFO] [URAlgorithm] Actions read now creating correlators*
>>> *[WARN] [TaskSetManager] Lost task 0.0 in stage 56.0 (TID 50,
>>> ip-172-31-40-139.eu-west-1.com
>>> <http://ip-172-31-40-139.eu-west-1.com>pute.internal):
>>> java.lang.NegativeArraySizeException*
>>> *at
>>> org.apache.mahout.math.DenseVector.(DenseVector.java:57)*
>>> *at
>>> org.apache.mahout.sparkbindings.SparkEngine$$anonfun$5.apply(SparkEngine.scala:73)*
>>> *at
>>> org.apache.mahout.sparkbindings.SparkEngine$$anonfun$5.apply(SparkEngine.scala:72)*
>>> *at
>>> org.apache.spark.rdd.RDD$$anonfun$mapPartitions$1$$anonfun$apply$20.apply(RDD.scala:710)*
>>> *at
>>> org.apache.spark.rdd.RDD$$anonfun$mapPartitions$1$$anonfun$apply$20.apply(RDD.scala:710)*
>>> *at
>>> org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)*
>>> *at
>>> org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)*
>>> *at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)*
>>> *at
>>> org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)*
>>> *at org.apache.spark.scheduler.Task.run(Task.scala:89)*
>>> *at
>>> org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:227)*
>>> *at
>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)*
>>> *at
>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)*
>>> *at java.lang.Thread.run(Thread.java:748)*
>>>
>>> *[ERROR] [TaskSetManager] Task 0 in stage 56.0 failed 4 times; aborting
>>> job*
>>> *Exception in thread "main" org.apache.spark.SparkException: Job aborted
>>> due to stage failure: Task 0 in stage 56.0 failed 4 times, most recent
>>> failure: Lost task 0.3 in stage 56.0 (TID 56,
>>> ip-172-1-1-1.eu-west-1.compute.internal):
>>> java.lang.NegativeArraySizeE

Re: Error while training : NegativeArraySizeException

2017-06-07 Thread Vaghawan Ojha
You could explicitly do

pio train -- --master spark://localhost:7077 --driver-memory 16G
--executor-memory 24G

and change the spark master url and the memories configuration. And see if
that works.

Thanks

On Wed, Jun 7, 2017 at 1:55 PM, Bruno LEBON  wrote:

> Hi,
>
> Using UR with PIO 0.10 I am trying to train my dataset. In return I get
> the following error:
>
> *...*
> *[INFO] [DataSource] Received events List(facet, view, search)*
> *[INFO] [DataSource] Number of events List(5, 4, 6)*
> *[INFO] [Engine$] org.template.TrainingData does not support data sanity
> check. Skipping check.*
> *[INFO] [Engine$] org.template.PreparedData does not support data sanity
> check. Skipping check.*
> *[INFO] [URAlgorithm] Actions read now creating correlators*
> *[WARN] [TaskSetManager] Lost task 0.0 in stage 56.0 (TID 50,
> ip-172-31-40-139.eu-west-1.compute.internal):
> java.lang.NegativeArraySizeException*
> *at org.apache.mahout.math.DenseVector.(DenseVector.java:57)*
> *at
> org.apache.mahout.sparkbindings.SparkEngine$$anonfun$5.apply(SparkEngine.scala:73)*
> *at
> org.apache.mahout.sparkbindings.SparkEngine$$anonfun$5.apply(SparkEngine.scala:72)*
> *at
> org.apache.spark.rdd.RDD$$anonfun$mapPartitions$1$$anonfun$apply$20.apply(RDD.scala:710)*
> *at
> org.apache.spark.rdd.RDD$$anonfun$mapPartitions$1$$anonfun$apply$20.apply(RDD.scala:710)*
> *at
> org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)*
> *at
> org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)*
> *at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)*
> *at
> org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)*
> *at org.apache.spark.scheduler.Task.run(Task.scala:89)*
> *at
> org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:227)*
> *at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)*
> *at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)*
> *at java.lang.Thread.run(Thread.java:748)*
>
> *[ERROR] [TaskSetManager] Task 0 in stage 56.0 failed 4 times; aborting
> job*
> *Exception in thread "main" org.apache.spark.SparkException: Job aborted
> due to stage failure: Task 0 in stage 56.0 failed 4 times, most recent
> failure: Lost task 0.3 in stage 56.0 (TID 56,
> ip-172-1-1-1.eu-west-1.compute.internal):
> java.lang.NegativeArraySizeException*
> *at org.apache.mahout.math.DenseVector.(DenseVector.java:57)*
> *at
> org.apache.mahout.sparkbindings.SparkEngine$$anonfun$5.apply(SparkEngine.scala:73)*
> *at
> org.apache.mahout.sparkbindings.SparkEngine$$anonfun$5.apply(SparkEngine.scala:72)*
> *at
> org.apache.spark.rdd.RDD$$anonfun$mapPartitions$1$$anonfun$apply$20.apply(RDD.scala:710)*
> *at
> org.apache.spark.rdd.RDD$$anonfun$mapPartitions$1$$anonfun$apply$20.apply(RDD.scala:710)*
> *at
> org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)*
> *at
> org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)*
> *at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)*
> *at
> org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)*
> *at org.apache.spark.scheduler.Task.run(Task.scala:89)*
> *at
> org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:227)*
> *at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)*
> *at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)*
> *at java.lang.Thread.run(Thread.java:748)*
>
> *Driver stacktrace:*
> *at org.apache.spark.scheduler.DAGScheduler.org
> $apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1431)*
> *at
> org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1419)*
> *at
> org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1418)*
> *at
> scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)*
> *at
> scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)*
> *at
> org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1418)*
> *at
> org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799)*
> *at
> org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799)*
> *at scala.Option.foreach(Option.scala:236)*
> *at
> org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:799)*
> *at
> org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:1640)*
> *at
> 

Re: How $set events works together?

2017-05-04 Thread Vaghawan Ojha
Hi Pat,

Ok, thanks for the addition, I got the point.

Thanks

On Thu, May 4, 2017 at 9:41 PM, Pat Ferrel <p...@occamsmachete.com> wrote:

> Correct. The purchase is associated with the user so any properties would
> be attached to user objects and for the UR there are no user objects in the
> EventStore. Users are represented by the collection of their events, not
> properties.
>
>
> On May 3, 2017, at 10:47 PM, Vaghawan Ojha <vaghawan...@gmail.com> wrote:
>
> Hi Marius,
>
> Thank you very much for the quick reply. That helps, thank you.
>
>
>
> On Thu, May 4, 2017 at 11:23 AM, Marius Rabenarivo <mariusrabenarivo@
> gmail.com> wrote:
>
>> Hi Vaghawan,
>>
>> yes, you have to send two events : one for the purchase and one for the
>> $set property.
>>
>> 2017-05-04 9:25 GMT+04:00 Vaghawan Ojha <vaghawan...@gmail.com>:
>>
>>> Hi,
>>>
>>> It's quite a novice question, but I was curious how $set events are
>>> handled by the engine.
>>>
>>> For say I've a user1 buying item2 from the category fruits.
>>>
>>> so the event without $set would look like this:
>>> {"eventTime":"2016-04-01T00:03:16+05:41","entityType":"user"
>>> ,"targetEntityType":"item","event":"purchase","entityId":"us
>>> er5","targetEntityId":"apple","properties":{"category":"fruits"}}
>>>
>>> But if I use $set, that would be something like this:
>>> { "event" : "$set", "entityType" : "item", "entityId" : "apple",
>>> "properties" : { "category": ["fruits"] }, "eventTime" : "
>>> 2016-04-01T00:03:16+05:41" }
>>> So here I can't place the user, I think I'm wrong about how to implement
>>> $set events, in this case how the engine knows that the item was bought by
>>>  the user5 at that particular time?
>>>
>>> Am I supposed to send two events, one with user5 bought apple, and
>>> another $set event with category fruits?
>>>
>>> Someday who have already used this feature could let me know?
>>>
>>> Thanks
>>>
>>
>>
>
> --
> You received this message because you are subscribed to the Google Groups
> "actionml-user" group.
> To unsubscribe from this group and stop receiving emails from it, send an
> email to actionml-user+unsubscr...@googlegroups.com.
> To post to this group, send email to actionml-u...@googlegroups.com.
> To view this discussion on the web visit https://groups.google.
> com/d/msgid/actionml-user/CA%2B69RXaQyLtM2y18wWsKtds099YvAM
> VyjPB3OqjRnC5M_YahZg%40mail.gmail.com
> <https://groups.google.com/d/msgid/actionml-user/CA%2B69RXaQyLtM2y18wWsKtds099YvAMVyjPB3OqjRnC5M_YahZg%40mail.gmail.com?utm_medium=email_source=footer>
> .
> For more options, visit https://groups.google.com/d/optout.
>
>


Re: Missing lib folder and contents

2017-05-04 Thread Vaghawan Ojha
You need to 
PredictionIO-0.11.0-incubating/lib/pio-assembly-0.11.0-incubating.jar
untar this file, and proceed to the instillation.

On Thu, May 4, 2017 at 5:55 PM, Brian Sexton <briansex...@engager.ie> wrote:

> Ive ran it twice now and its still not creating the folder or contents. I
> can see it in the output though
>
> PredictionIO-0.11.0-incubating/
> PredictionIO-0.11.0-incubating/RELEASE
> PredictionIO-0.11.0-incubating/sbt/
> PredictionIO-0.11.0-incubating/sbt/sbt
> PredictionIO-0.11.0-incubating/bin/
> PredictionIO-0.11.0-incubating/bin/install.sh
> PredictionIO-0.11.0-incubating/bin/load-pio-env.sh
> PredictionIO-0.11.0-incubating/bin/compute-classpath.sh
> PredictionIO-0.11.0-incubating/bin/pio
> PredictionIO-0.11.0-incubating/bin/pio-stop-all
> PredictionIO-0.11.0-incubating/bin/pio-daemon
> PredictionIO-0.11.0-incubating/bin/cjson
> PredictionIO-0.11.0-incubating/bin/semver.sh
> PredictionIO-0.11.0-incubating/bin/pio-start-all
> PredictionIO-0.11.0-incubating/bin/pio-class
> PredictionIO-0.11.0-incubating/bin/pio-shell
> PredictionIO-0.11.0-incubating/lib/
> PredictionIO-0.11.0-incubating/lib/spark/
> PredictionIO-0.11.0-incubating/lib/spark/pio-data-hdfs-assembly-0.11.0-incubating.jar
> PredictionIO-0.11.0-incubating/lib/spark/pio-data-jdbc-assembly-0.11.0-incubating.jar
> PredictionIO-0.11.0-incubating/lib/spark/pio-data-elasticsearch1-assembly-0.11.0-incubating.jar
> PredictionIO-0.11.0-incubating/lib/spark/pio-data-hbase-assembly-0.11.0-incubating.jar
> PredictionIO-0.11.0-incubating/lib/spark/pio-data-localfs-assembly-0.11.0-incubating.jar
> PredictionIO-0.11.0-incubating/lib/pio-assembly-0.11.0-incubating.jar
>
>
> Brian Sexton
>
> CTO at Engager
>
> *M* +353 86 035 3453 <+353%2086%20035%203453>
>
> *E* briansex...@engager.ie <briansex...@engager.ie?subject=Hey%20there!>
>
> *W* www.engager.ie
> <https://www.facebook.com/EngagerIre/>   <https://twitter.com/engagerire>
><https://www.linkedin.com/company/7958977>
>
> On Thu, May 4, 2017 at 12:59 PM, Vaghawan Ojha <vaghawan...@gmail.com>
> wrote:
>
>> Seems like you haven't done ./make-distribution.sh . Follow the full
>> installation process
>>
>> https://predictionio.incubator.apache.org/install/install-sourcecode/
>>
>> On Thu, May 4, 2017 at 5:24 PM, Brian Sexton <briansex...@engager.ie>
>> wrote:
>>
>>>
>>>
>>> Hi
>>>
>>> Im getting the following error when i try to run pio status
>>>
>>> Error: Could not find or load main class 
>>> org.apache.predictionio.tools.console.Console
>>>
>>>
>>>
>>> when I downloaded apache-predictionio-0.11.0-incubating.tar.gz from apache 
>>> there is no lib folder inlcuded and i dont know where t
>>>
>>> to get a copy of the files that are meant to be in there
>>>
>>> Thanks
>>> Brian
>>>
>>
>>
>


Re: Missing lib folder and contents

2017-05-04 Thread Vaghawan Ojha
Seems like you haven't done ./make-distribution.sh . Follow the full
installation process

https://predictionio.incubator.apache.org/install/install-sourcecode/

On Thu, May 4, 2017 at 5:24 PM, Brian Sexton  wrote:

>
>
> Hi
>
> Im getting the following error when i try to run pio status
>
> Error: Could not find or load main class 
> org.apache.predictionio.tools.console.Console
>
>
>
> when I downloaded apache-predictionio-0.11.0-incubating.tar.gz from apache 
> there is no lib folder inlcuded and i dont know where t
>
> to get a copy of the files that are meant to be in there
>
> Thanks
> Brian
>


Re: Deploying Multiple App in Same Machine

2017-05-04 Thread Vaghawan Ojha
Sorry, it did work. the content was coming from the cache. Thank you very
much, I just changed the ES indexNames.

Thank you

On Thu, May 4, 2017 at 1:32 PM, Vaghawan Ojha <vaghawan...@gmail.com> wrote:

> Hi Pat,
>
> Thanks, I had the same Index names in engine.json. I changed that now and
> retrain them, appnames are different, so that shouldn't be the problem. But
> the problem still persists.
>
> It's quite weird.
>
> On Mon, May 1, 2017 at 11:33 PM, Pat Ferrel <p...@occamsmachete.com> wrote:
>
>> make sure the 2 apps have separate appNames in `pio app list` and have
>> different indexNames in their engine.jsons.
>>
>> 1) appNames are where to get input, if the same then only params will be
>> making different models, not different data. A PIO “app” is really a
>> dataset.
>> 2) indexNames tell the UR where to store the model in Elasticsearch, if 2
>> engin.jsons have the same indexName, the latest `pio train` will overwrite
>> the previous one.
>>
>> On Apr 30, 2017, at 10:13 PM, Vaghawan Ojha <vaghawan...@gmail.com>
>> wrote:
>>
>> Hi,
>>
>> I'm using two seperate Universal recommender app in same machine.
>> Everything is fine, but whenever I retrain one app, the score in the other
>> app always goes to zero. However recommendation seems fine as far as I
>> could see.
>>
>> Why this happens? I'm using different port to deploy the different apps.
>> Have anyone else gone through the same problem?
>>
>> --
>> You received this message because you are subscribed to the Google Groups
>> "actionml-user" group.
>> To unsubscribe from this group and stop receiving emails from it, send an
>> email to actionml-user+unsubscr...@googlegroups.com.
>> To post to this group, send email to actionml-u...@googlegroups.com.
>> To view this discussion on the web visit https://groups.google.com/d/ms
>> gid/actionml-user/CA%2B69RXY0ncgUngGai5ztBwsqSwt%3DZTyUcj%
>> 2B0ccqhzi_bqS%3D01A%40mail.gmail.com
>> <https://groups.google.com/d/msgid/actionml-user/CA%2B69RXY0ncgUngGai5ztBwsqSwt%3DZTyUcj%2B0ccqhzi_bqS%3D01A%40mail.gmail.com?utm_medium=email_source=footer>
>> .
>> For more options, visit https://groups.google.com/d/optout.
>>
>>
>


Re: How $set events works together?

2017-05-03 Thread Vaghawan Ojha
Hi Marius,

Thank you very much for the quick reply. That helps, thank you.



On Thu, May 4, 2017 at 11:23 AM, Marius Rabenarivo <
mariusrabenar...@gmail.com> wrote:

> Hi Vaghawan,
>
> yes, you have to send two events : one for the purchase and one for the
> $set property.
>
> 2017-05-04 9:25 GMT+04:00 Vaghawan Ojha <vaghawan...@gmail.com>:
>
>> Hi,
>>
>> It's quite a novice question, but I was curious how $set events are
>> handled by the engine.
>>
>> For say I've a user1 buying item2 from the category fruits.
>>
>> so the event without $set would look like this:
>> {"eventTime":"2016-04-01T00:03:16+05:41","entityType":"user"
>> ,"targetEntityType":"item","event":"purchase","entityId":"
>> user5","targetEntityId":"apple","properties":{"category":"fruits"}}
>>
>> But if I use $set, that would be something like this:
>> { "event" : "$set", "entityType" : "item", "entityId" : "apple",
>> "properties" : { "category": ["fruits"] }, "eventTime" : "
>> 2016-04-01T00:03:16+05:41" }
>> So here I can't place the user, I think I'm wrong about how to implement
>> $set events, in this case how the engine knows that the item was bought by
>>  the user5 at that particular time?
>>
>> Am I supposed to send two events, one with user5 bought apple, and
>> another $set event with category fruits?
>>
>> Someday who have already used this feature could let me know?
>>
>> Thanks
>>
>
>


Knowing the probability of purchase

2017-04-24 Thread Vaghawan Ojha
Hi,

I was following a research paper regarding the probability of a user buying
a particular item recommended by the recommendation system. It's here, if
you want to checkout as well
http://www.kecl.ntt.co.jp/as/members/iwata/doctor.pdf

I was wondering if there is a way or anybody has done with the current
templates of PIO, the calculation of the probability of a user buying an
item.

I think this should be possible with current templates as well, I am just
wondering, if anybody could provide me a brief way to do that, or any
documentation of the algorithms that could be used.


Thanks


Re: Cart/Complementary purchase recommendation with UR

2017-04-21 Thread Vaghawan Ojha
Ok, Thank you!

On Wed, Apr 19, 2017 at 8:23 PM, Pat Ferrel <p...@occamsmachete.com> wrote:

> The UR v0.6.0 will be out in days and make this much easier.
>
>
> On Apr 19, 2017, at 4:02 AM, Vaghawan Ojha <vaghawan...@gmail.com> wrote:
>
> Hi Pat,
>
> I was trying to accomplish the cart recommendation also with properties in
> the events, so my events looks something like this:
>
> {"eventTime":"2016-01-01T07:19:17+05:41","entityType":"
> user","targetEntityType":"item","event":"purchase","
> entityId":"firstcartid","targetEntityId":"product12","
> properties":{"store_id":"2"}}
> {"eventTime":"2016-01-01T07:06:51+05:41","entityType":"
> user","targetEntityType":"item","event":"purchase","
> entityId":"secondcartid","targetEntityId":"product23","
> properties":{"store_id":"3"}}
> {"eventTime":"2016-01-01T07:06:51+05:41","entityType":"
> user","targetEntityType":"item","event":"purchase","
> entityId":"firstcartid","targetEntityId":"product101","
> properties":{"store_id":"3"}}
>
> And I would query it like this:
>
> curl -H "Content-Type: application/json" -d '
> > {
> > "user": "CM/00074305/16",
> > "num":4,
> > "fields": [
> > {
> >   "name": "store_id",
> >   "values": ["2"]
> > }
> >   ]
> > }' http://10.10.10.2:8123/queries.json
>
> It works without bias. But I don't think it's a valid way. I meant, if I'm
> not wrong, I should have $set in the events, but if I use $set in the
> events, then I won't have option to put both cart id as user and product id
> as item.
>
> Can you please add your suggestion?
>
>
> On Sat, Apr 8, 2017 at 10:13 AM, Vaghawan Ojha <vaghawan...@gmail.com> w
> rote:
>
>> Hi Pat,
>>
>> Thanks I will try this in my existing app.
>>
>> Thanks
>> Vaghawan
>>
>> On Thu, Apr 6, 2017 at 9:49 PM, Pat Ferrel <p...@occamsmachete.com> wrote:
>>
>>> Yes it can be done but support is not ideal so UR v0.6.0 coming just
>>> after the release of PIO-0.11.0 (in RC now) has some enhancements that give
>>> you better options and easier queries.
>>>
>>> To get truly complimentary purchases you need to train on shopping
>>> carts, not users purchases. So the primary input would of the form:
>>>
>>> (shopping-cart-id, cart-purchase, product-id)
>>>
>>> Then in < 0.6.0 you query with the currently active shopping cart id as
>>> if it is the user-id
>>>
>>> {
>>> “user”: “some-shopping-cart-id”
>>> }
>>>
>>> In v0.6.0 you still need to train on shopping carts to get true
>>> “complimentary purchase” but the query may be easier:
>>>
>>> {
>>> “Item-set”: [“product1”, “product2”]
>>> }
>>>
>>> The query is the current contents of the shopping cart, no fake
>>> “user-id” required.
>>>
>>> UR 0.6.0 will also allow you to query against a model trained on user
>>> purchases with the same query but, though the results will be "things
>>> purchased by people with similar purchasing behavior as you query
>>> item-set", it won’t be strictly “complimentary purchases”.
>>>
>>> Since 2 methods are supported with UR 0.6.0 you could A/B test for which
>>> has the best results. If using user purchase (instead of shopping carts) is
>>> close to as good it would be much easier since you can have user-based,
>>> item-based and shopping carts recs from the same model. Otherwise one for
>>> user-based and item-based and another for “complimentary purchases” would
>>> make 2 trains and 2 models.
>>>
>>>
>>>
>>> On Apr 5, 2017, at 8:08 PM, Vaghawan Ojha <vaghawan...@gmail.com> wrote:
>>>
>>> Hi Pat,
>>>
>>> As far as the actionml docs tells me that I could be able to use the UR
>>> to recommend the cart/complementary purchase, but there seems to be no
>>> documentation regarding this.
>>>
>>> It would be helpful if there was a documentation for this feature.

Re: Cart/Complementary purchase recommendation with UR

2017-04-19 Thread Vaghawan Ojha
Hi Pat,

I was trying to accomplish the cart recommendation also with properties in
the events, so my events looks something like this:

{"eventTime":"2016-01-01T07:19:17+05:41","entityType":"user","targetEntityType":"item","event":"purchase","entityId":"firstcartid","targetEntityId":"product12","properties":{"store_id":"2"}}
{"eventTime":"2016-01-01T07:06:51+05:41","entityType":"user","targetEntityType":"item","event":"purchase","entityId":"secondcartid","targetEntityId":"product23","properties":{"store_id":"3"}}
{"eventTime":"2016-01-01T07:06:51+05:41","entityType":"user","targetEntityType":"item","event":"purchase","entityId":"firstcartid","targetEntityId":"product101","properties":{"store_id":"3"}}

And I would query it like this:

curl -H "Content-Type: application/json" -d '
> {
> "user": "CM/00074305/16",
> "num":4,
> "fields": [
> {
>   "name": "store_id",
>   "values": ["2"]
> }
>   ]
> }' http://10.10.10.2:8123/queries.json

It works without bias. But I don't think it's a valid way. I meant, if I'm
not wrong, I should have $set in the events, but if I use $set in the
events, then I won't have option to put both cart id as user and product id
as item.

Can you please add your suggestion?


On Sat, Apr 8, 2017 at 10:13 AM, Vaghawan Ojha <vaghawan...@gmail.com>
wrote:

> Hi Pat,
>
> Thanks I will try this in my existing app.
>
> Thanks
> Vaghawan
>
> On Thu, Apr 6, 2017 at 9:49 PM, Pat Ferrel <p...@occamsmachete.com> wrote:
>
>> Yes it can be done but support is not ideal so UR v0.6.0 coming just
>> after the release of PIO-0.11.0 (in RC now) has some enhancements that give
>> you better options and easier queries.
>>
>> To get truly complimentary purchases you need to train on shopping carts,
>> not users purchases. So the primary input would of the form:
>>
>> (shopping-cart-id, cart-purchase, product-id)
>>
>> Then in < 0.6.0 you query with the currently active shopping cart id as
>> if it is the user-id
>>
>> {
>> “user”: “some-shopping-cart-id”
>> }
>>
>> In v0.6.0 you still need to train on shopping carts to get true
>> “complimentary purchase” but the query may be easier:
>>
>> {
>> “Item-set”: [“product1”, “product2”]
>> }
>>
>> The query is the current contents of the shopping cart, no fake “user-id”
>> required.
>>
>> UR 0.6.0 will also allow you to query against a model trained on user
>> purchases with the same query but, though the results will be "things
>> purchased by people with similar purchasing behavior as you query
>> item-set", it won’t be strictly “complimentary purchases”.
>>
>> Since 2 methods are supported with UR 0.6.0 you could A/B test for which
>> has the best results. If using user purchase (instead of shopping carts) is
>> close to as good it would be much easier since you can have user-based,
>> item-based and shopping carts recs from the same model. Otherwise one for
>> user-based and item-based and another for “complimentary purchases” would
>> make 2 trains and 2 models.
>>
>>
>>
>> On Apr 5, 2017, at 8:08 PM, Vaghawan Ojha <vaghawan...@gmail.com> wrote:
>>
>> Hi Pat,
>>
>> As far as the actionml docs tells me that I could be able to use the UR
>> to recommend the cart/complementary purchase, but there seems to be no
>> documentation regarding this.
>>
>> It would be helpful if there was a documentation for this feature. I know
>> people have worked out and implemented the cart recommendation, but for
>> begineer like me, it would still be like "what to do now".
>>
>> So could you help me how can I do that? If documentation is not ready yet
>> for this, I could write one for this feature once I got what to do.
>>
>> Thanks
>> Vaghawan
>>
>> --
>> You received this message because you are subscribed to the Google Groups
>> "actionml-user" group.
>> To unsubscribe from this group and stop receiving emails from it, send an
>> email to actionml-user+unsubscr...@googlegroups.com.
>> To post to this group, send email to actionml-u...@googlegroups.com.
>> To view this discussion on the web visit https://groups.google.com/d/ms
>> gid/actionml-user/CA%2B69RXYDeMRXOG25v%3Dadr%3DSXLk6BO811DSC
>> YX0t6vp8EVPP1Xw%40mail.gmail.com
>> <https://groups.google.com/d/msgid/actionml-user/CA%2B69RXYDeMRXOG25v%3Dadr%3DSXLk6BO811DSCYX0t6vp8EVPP1Xw%40mail.gmail.com?utm_medium=email_source=footer>
>> .
>> For more options, visit https://groups.google.com/d/optout.
>>
>>
>


Re: Invalid app name INVALID_APP_NAME

2017-04-18 Thread Vaghawan Ojha
Ya I see! the pio path shouldn't be the problem. But you can easily export
the pio path.

Ya, you can stick with the same pio, but could create new app from scratch,
like downloading the template and creating a new pio app and changing
engine.json.

Hope  this should work.

On Wed, Apr 19, 2017 at 9:22 AM, Ravi Kiran <ravi.gun...@gmail.com> wrote:

>
> I am running "pio build" from template dir.  But pio path is not
> configured. So, everytime I have to give absolute path of pio. Could that
> be a problem (But most queries are working).
> For e.g.
>  ../bin/pio build
> ../bin/pio eval mlframework.AccuracyEvaluation mlframework.EngineParamsList
> ../bin app list
>
> I will made new app right from scratch, even then this issue persists :(
>
> On Wed, Apr 19, 2017 at 9:00 AM, Vaghawan Ojha <vaghawan...@gmail.com>
> wrote:
>
>> Hi,
>>
>> This seems fine, did you make sure to run pio build from that template
>> dir? In case still not working, I would suggest you to delete the
>> wineapp(if you haven't imported the events yet) and recreate a app, and
>> change the engine.json and build again.
>>
>> I had this problem with both MyRecommendation and Universal Recommender
>> as well, in case of MyRecommendation, I rebuilt the app and it worked.
>>
>> Please try that, it may work for you as well.
>>
>> Thanks
>>
>> On Wed, Apr 19, 2017 at 9:05 AM, Ravi Kiran <ravi.gun...@gmail.com>
>> wrote:
>>
>>> [INFO] [App$] Name |   ID |
>>>   Access Key | Allowed Event(s)
>>> [INFO] [App$]app10 |   10 |
>>>  mnbvcxz | (all)
>>> [INFO] [App$] sample_app10 |   11 |
>>> 1234 | (all)
>>> [INFO] [App$] sample_app21 |   13 |
>>> 4321 | (all)
>>> [INFO] [App$] sample_app22 |   14 |
>>> 4321abcd | (all)
>>> [INFO] [App$] sample_app23 |   15 |
>>>4321abcde | (all)
>>> [INFO] [App$] sample_app26 |   18 |
>>>  4321abcdefg | (all)
>>> [INFO] [App$]  wineapp |   19 |
>>>  wineappeine | (all)
>>> [INFO] [App$] Finished listing 19 app(s).
>>>
>>>
>>> This is output of pio app list
>>>
>>> On Wed, Apr 19, 2017 at 8:42 AM, Ravi Kiran <ravi.gun...@gmail.com>
>>> wrote:
>>>
>>>> *Contents of engine.json*
>>>>
>>>> {
>>>>   "id": "default",
>>>>   "description": "Default settings",
>>>>   "engineFactory": "mlframework.ClassificationEngine",
>>>>   "datasource": {
>>>> "params": {
>>>>   "appName": "wineapp"
>>>> }
>>>>   },
>>>>   "algorithms": [
>>>> {
>>>>   "name": "naive",
>>>>   "params": {
>>>> "lambda": 1.0
>>>>   }
>>>> }
>>>>   ]
>>>> }
>>>>
>>>>
>>>> *Location of engine.json*
>>>> /home/user_name/Documents/code/apache-predictionio-0.10.0-in
>>>> cubating/PredictionIO-0.10.0-incubating/sample_classifier/engine.json
>>>> I have installed template in sample_classifier folder
>>>>
>>>>
>>>>
>>>>
>>>> On Wed, Apr 19, 2017 at 8:32 AM, Vaghawan Ojha <vaghawan...@gmail.com>
>>>> wrote:
>>>>
>>>>> What does your pio app list produce?
>>>>>
>>>>> And can you also send a screenshot of your engine.json. This
>>>>> engine.json should be inside your template directory where you installed
>>>>> your template.
>>>>>
>>>>> I had faced  this issue, but resolved it too. Let me know with your
>>>>> app list screenshot and corresponding engine.json
>>>>>
>>>>> Thanks
>>>>>
>>>>> On Wed, Apr 19, 2017 at 8:45 AM, Ravi Kiran <ravi.gun...@gmail.com>
>>>>> wrote:
>>>>>
>>>>>> @Vaghawan, I have updated app_name in engine.json . And then 'pio
>>>>>> Build' and evaluation query. Still this issue exists
>>>>>>
>>>>>>
>>>>>> On 19-Apr-2017 8:14 AM, "Vaghawan Ojha" <vaghawan...@gmail.com>
>>>>>> wrote:
>>>>>>
>>>>>> Hi,
>>>>>>
>>>>>> It seems you haven't changed the app name in your engine.json. There
>>>>>> is an engine.json in your template directory, please make sure that you
>>>>>> have changed that name.
>>>>>>
>>>>>> And it will works. Make sure you input the right appname in your
>>>>>> engine.json.
>>>>>>
>>>>>> Thanks
>>>>>>
>>>>>>
>>>>>> On Wed, Apr 19, 2017 at 5:33 AM, Ravi Kiran <ravi.gun...@gmail.com>
>>>>>> wrote:
>>>>>>
>>>>>>> Hi,
>>>>>>> When I try to evaluate I am getting following error.
>>>>>>>
>>>>>>>
>>>>>>> [ERROR] [Common$] Invalid app name INVALID_APP_NAME
>>>>>>> Exception in thread "main" java.lang.IllegalArgumentException:
>>>>>>> Invalid app name INVALID_APP_NAME
>>>>>>>
>>>>>>>
>>>>>>> I have updated app name in engine.json
>>>>>>> This app is in 'pio app list'
>>>>>>> I have also done 'pio build', 'pio train' and 'pio deploy' before
>>>>>>> running evaluation query.
>>>>>>>
>>>>>>>
>>>>>>> Can you please help me resolve this query?
>>>>>>>
>>>>>>> Thanks
>>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>


Re: Invalid app name INVALID_APP_NAME

2017-04-18 Thread Vaghawan Ojha
Hi,

It seems you haven't changed the app name in your engine.json. There is an
engine.json in your template directory, please make sure that you have
changed that name.

And it will works. Make sure you input the right appname in your
engine.json.

Thanks


On Wed, Apr 19, 2017 at 5:33 AM, Ravi Kiran  wrote:

> Hi,
> When I try to evaluate I am getting following error.
>
>
> [ERROR] [Common$] Invalid app name INVALID_APP_NAME
> Exception in thread "main" java.lang.IllegalArgumentException: Invalid
> app name INVALID_APP_NAME
>
>
> I have updated app name in engine.json
> This app is in 'pio app list'
> I have also done 'pio build', 'pio train' and 'pio deploy' before running
> evaluation query.
>
>
> Can you please help me resolve this query?
>
> Thanks
>


Re: Similar product template

2017-04-14 Thread Vaghawan Ojha
Sorry the email sent accidentally without finishing, it would be really
helpful for me if you describe about in which case the multi model are
being used.

On Fri, Apr 14, 2017 at 7:01 PM, Vaghawan Ojha <vaghawan...@gmail.com>
wrote:

> Hi Pat,
>
> This is really a great explanation, I myself had tried ALS before CCO, but
> in my case CCO seems better. You had a nice presentation, but I was quite
> confused regarding multi-model recommendation.
>
> In what case does UR make use of multi model? For say, I've a location
> preference for every user event, and category preference as well. Let's say
> I trained the model and queried with the preference parameter, in that case
> is it using multi model for each preference?
>
> If you could describe a bit about this, it would be reall
>
> On Thu, Apr 13, 2017 at 9:15 PM, Pat Ferrel <p...@occamsmachete.com> wrote:
>
>> I’m surprised that ALS seemed clear because is is based on a complicated
>> matrix factorization algorithm that transforms the user vectors into a
>> smaller dimensional space that is composed of “important” features. These
>> are not interactions with items like “buys”, they can only be described as
>> defining a new feature space. The factorized matrices transform in and out
>> of that space. The factorized matrices are approximations of user x
>> features, and features x items.
>>
>> The user’s history is transformed into the feature space, which will be
>> dense, in other words indicating some preference for all features. Then
>> when this dense user vector is transformed back into item space the
>> approximation nature of ALS will give some preference value for all items.
>> At this point they can be ranked by score and the top few returned. This is
>> clearly wrong since user will never have a preference for all items and
>> would never purchase or convert on a large number of them no mater what the
>> circumstances. It does give good results for the top ranked though when you
>> have lots of “conversions” per user on average because ALS can only use
>> conversions as input. in other words it can use only one kind of behavior
>> data.
>>
>> The CCO (Correlated Cross-Occurrence) algorithm from Mahout that is
>> behind the Universal Recommender is multi-domain and multi-modal, in that
>> takes interactions of the user from many actions they perform and even
>> contextual data like profile info or location. It takes all this and finds
>> which “indicators”, a name for these interactions or other user info, and
>> compares them with the user’s conversions. It does this for all users and
>> so finds which of the indicators most often lead to conversion. These
>> highly correlated indicators are then associated with items as properties,
>> When a user recommendation is needed we see which items have the most
>> similar behavioral indicators as the user's history. This tells us that the
>> user probably has an affinity for the item—we can predict a preference for
>> these items.
>>
>> The differences:
>> 1) ALS can ingest only one type of behavior. This is not bad but also not
>> very flexible and requires a good number of these interactions per user.
>> 2) Cross-behavioral recommendations cannot be made with ALS since no
>> cross behavioral data is seen by it. This in turn means that users with few
>> or no conversions will not get recommendations. The Universal Recommender
>> can make recommendations to users with no conversions if they have other
>> behavior to draw from so it is generally said to handle cool-start for
>> user’s better. Another way to say this is that “cold-start” for ALS is only
>> “cool-start” for CCO (in the UR). The same goes for item-based
>> recommendations.
>> 3) CCO can also use content directly for similar item recommendations,
>> which helps solve the item “cold-start” problem. ALS cannot.
>> 4) CCO is more like a landscape of Predictive AI algorithms using all we
>> know about a user from multiple domains (conversions, page views, search
>> terms, category preferences, tag preferences, brand preferences, location,
>> device used, etc) to make predictions in some specific domain. It can also
>> work with conversions alone
>> 5) To do queries with ALS in the MLlib requires that the factorized
>> matrices be in-memory. They are much smaller than the input but this means
>> running Spark to make queries. This makes it rather heavy-weight for
>> queries and makes scaling a bit of a problem and fairly complicated (too
>> much to explain here). CCO on the other hand uses Spark only to create the
>> indicators model, which it puts in

Re: Similar product template

2017-04-14 Thread Vaghawan Ojha
Hi Pat,

This is really a great explanation, I myself had tried ALS before CCO, but
in my case CCO seems better. You had a nice presentation, but I was quite
confused regarding multi-model recommendation.

In what case does UR make use of multi model? For say, I've a location
preference for every user event, and category preference as well. Let's say
I trained the model and queried with the preference parameter, in that case
is it using multi model for each preference?

If you could describe a bit about this, it would be reall

On Thu, Apr 13, 2017 at 9:15 PM, Pat Ferrel  wrote:

> I’m surprised that ALS seemed clear because is is based on a complicated
> matrix factorization algorithm that transforms the user vectors into a
> smaller dimensional space that is composed of “important” features. These
> are not interactions with items like “buys”, they can only be described as
> defining a new feature space. The factorized matrices transform in and out
> of that space. The factorized matrices are approximations of user x
> features, and features x items.
>
> The user’s history is transformed into the feature space, which will be
> dense, in other words indicating some preference for all features. Then
> when this dense user vector is transformed back into item space the
> approximation nature of ALS will give some preference value for all items.
> At this point they can be ranked by score and the top few returned. This is
> clearly wrong since user will never have a preference for all items and
> would never purchase or convert on a large number of them no mater what the
> circumstances. It does give good results for the top ranked though when you
> have lots of “conversions” per user on average because ALS can only use
> conversions as input. in other words it can use only one kind of behavior
> data.
>
> The CCO (Correlated Cross-Occurrence) algorithm from Mahout that is behind
> the Universal Recommender is multi-domain and multi-modal, in that takes
> interactions of the user from many actions they perform and even contextual
> data like profile info or location. It takes all this and finds which
> “indicators”, a name for these interactions or other user info, and
> compares them with the user’s conversions. It does this for all users and
> so finds which of the indicators most often lead to conversion. These
> highly correlated indicators are then associated with items as properties,
> When a user recommendation is needed we see which items have the most
> similar behavioral indicators as the user's history. This tells us that the
> user probably has an affinity for the item—we can predict a preference for
> these items.
>
> The differences:
> 1) ALS can ingest only one type of behavior. This is not bad but also not
> very flexible and requires a good number of these interactions per user.
> 2) Cross-behavioral recommendations cannot be made with ALS since no cross
> behavioral data is seen by it. This in turn means that users with few or no
> conversions will not get recommendations. The Universal Recommender can
> make recommendations to users with no conversions if they have other
> behavior to draw from so it is generally said to handle cool-start for
> user’s better. Another way to say this is that “cold-start” for ALS is only
> “cool-start” for CCO (in the UR). The same goes for item-based
> recommendations.
> 3) CCO can also use content directly for similar item recommendations,
> which helps solve the item “cold-start” problem. ALS cannot.
> 4) CCO is more like a landscape of Predictive AI algorithms using all we
> know about a user from multiple domains (conversions, page views, search
> terms, category preferences, tag preferences, brand preferences, location,
> device used, etc) to make predictions in some specific domain. It can also
> work with conversions alone
> 5) To do queries with ALS in the MLlib requires that the factorized
> matrices be in-memory. They are much smaller than the input but this means
> running Spark to make queries. This makes it rather heavy-weight for
> queries and makes scaling a bit of a problem and fairly complicated (too
> much to explain here). CCO on the other hand uses Spark only to create the
> indicators model, which it puts in Elasticsearch. Elasticsearch finds the
> top ranked items compared to the user’s history at runtime in real-time.
> This makes scaling queries as easy as scaling Elasticsearch since it was
> meant to scale.
>
> I have done cross-validaton comparisons but they are a bit unfair and the
> winner depends on the dataset, In real-life CCO serves more users than ALS
> since it uses more behavior and so tends to win for this reason. It’s
> nearly impossible to compare this with cross-validation so A/B tests are
> our only metric.
>
> We have a slide deck showing some of these comparisons here:
> https://docs.google.com/presentation/d/1HpHZZiRmHpMKtu86rOKBJ70cd58Vy
> TOUM1a8OmKSMTo/edit?usp=sharing
>
>
> On Apr 13, 

Uses Of PIO in Image Processing

2017-04-12 Thread Vaghawan Ojha
Hi,

I was just wondering if anyone has used PIO for image processing, like
classification type. I can make use of tenserflow for image processing and
build a model, and work with that.

But I'm particular curious, if PIO is suitable for those  tasks  too? This
may be a dumb question to ask, but I'm curious.


Thanks


Re: Event server id

2017-04-11 Thread Vaghawan Ojha
Hi,

If you are talking about entityId, yes they are stored in string. Fore
more: http://predictionio.incubator.apache.org/datacollection/eventapi/

Thanks

On Tue, Apr 11, 2017 at 3:45 PM, Dennis Honders 
wrote:

> Hello,
>
> Are id's (like user id) always stored as a String?
> I added data (id as an integer) to the eventserver and when I did a GET
> request, I got a response with the id as a String.
>
> Greetings,
>
> Dennis
>


Re: How Can I count events In The Event Server?

2017-04-10 Thread Vaghawan Ojha
Hi Marius,

Thanks, I was wondering if there was already some method to just to count
the events, but that's Ok. I would be able to get the events of particular
entityId and count them inside my app on my own.

Thanks


On Tue, Apr 11, 2017 at 11:29 AM, Marius Rabenarivo <
mariusrabenar...@gmail.com> wrote:

> Hello,
>
> You can use the Event API : http://predictionio.incubator.
> apache.org/datacollection/eventapi/
>
> 2017-04-11 7:34 GMT+04:00 Vaghawan Ojha <vaghawan...@gmail.com>:
>
>> I suppose there should be a way to count the events in the event server,
>> which I might be unaware of.
>>
>> I'm trying to count all the events, and number of users in those events,
>> also how many events a particular user have. I suppose there should be
>> something that I can do with eventserver query.
>>
>> I found a way to count all the events in this doc
>> http://predictionio.incubator.apache.org/customize/troubleshooting/.
>> Have anyone tried to count the events of a particular user?
>>
>> If anyone has already done, could you please share it?
>>
>> Thanks
>>
>
>


Re: Existing javascript SDK for predicitonIO (0.10) ?

2017-04-10 Thread Vaghawan Ojha
Hi,

Ok, in  that case I am currently not aware of any js SDK like you wanted.
Someone else may help.

Thanks

On Mon, Apr 10, 2017 at 6:14 PM, Mohamed Zouga  wrote:

> @vaghawan : i don't want to go posting my accessKey allover my requests
> and writing verbose code when all i might use is a function with one
> parameter or so.
> @ Marius : Why have 3 dependencies (3 more files, supposing those files
> don't have additional dependencies...) when i can use just Ajax to send my
> events and requests
> so any JS SDK out there ?
>
> 2017-04-10 14:23 GMT+02:00 Marius Rabenarivo :
>
>> Hello,
>>
>> The NodeJS SDK has the following dependencies :
>>
>> lodash
>> request
>> bluebird
>>
>> You can download them if you want.
>>
>> Regards,
>>
>> Marius
>>
>> 2017-04-10 16:18 GMT+04:00 Mohamed Zouga :
>>
>>> I'am using PredictionIO in a web site, and i want to send the event
>>> using Javascript (instead of PHP) is there any know SDK JS for PredictionIO
>>> (lastest) ?
>>> i found some outdated ones in Github, i didn't want to rush into coding
>>> one from scratch until i'am sure there is no existing ones, thank you
>>> P.S : I know there is a NodeJS SDK, i don't have npm and i don't want
>>> any dependencies.
>>> --
>>>
>>> *Mohamed ZOUGA*Stagiaire Data Scientist, Wizaplace
>>> moha...@wizaplace.com  | http://www.wizaplace.com
>>> [image: https://www.linkedin.com/in/zouga-mohamed-44b02974/]
>>> 
>>>
>>
>>
>
>
> --
>
> *Mohamed ZOUGA*Stagiaire Data Scientist, Wizaplace
> moha...@wizaplace.com  | http://www.wizaplace.com
> [image: https://www.linkedin.com/in/zouga-mohamed-44b02974/]
> 
>


Re: Existing javascript SDK for predicitonIO (0.10) ?

2017-04-10 Thread Vaghawan Ojha
Hi,

You could just use the REST API in case you need the query to be responded
in plain java script. PIO provides REST api out of box already.

Thanks

On Mon, Apr 10, 2017 at 6:03 PM, Mohamed Zouga  wrote:

> I'am using PredictionIO in a web site, and i want to send the event using
> Javascript (instead of PHP) is there any know SDK JS for PredictionIO
> (lastest) ?
> i found some outdated ones in Github, i didn't want to rush into coding
> one from scratch until i'am sure there is no existing ones, thank you
> P.S : I know there is a NodeJS SDK, i don't have npm and i don't want any
> dependencies.
> --
>
> *Mohamed ZOUGA*Stagiaire Data Scientist, Wizaplace
> moha...@wizaplace.com  | http://www.wizaplace.com
> [image: https://www.linkedin.com/in/zouga-mohamed-44b02974/]
> 
>


Re: How Does The Score Works In UR

2017-04-09 Thread Vaghawan Ojha
Hi Pat,

I did some study and understood some of those complexities, and some are
yet far to be understood. You have put really nice presentation in the doc
site.

I really appreciate your work and help.

Thank you very much!

On Sat, Apr 8, 2017 at 10:12 AM, Vaghawan Ojha <vaghawan...@gmail.com>
wrote:

> Hi Pat,
>
> Thank you very much for the explanation, seems like I forgot to properly
> search this question in the group. By ranking you means the order of the
> product right?
>
> Also I wanted to replicate such dot products of all user history vector
> segments against the correlator fields in the model in a very very small
> data, so that I could have a test of what's really working behind. And also
> I wanted to learn about the implicating algorithms. I've seen the
> presentation in the doc site of UR. Should you have any resources which
> could lead me to better understand all these please point me.
>
> Thanks
> Vaghawan
>
> On Sat, Apr 8, 2017 at 12:16 AM, Pat Ferrel <p...@occamsmachete.com> wrote:
>
>> The limit depends on your data. I tell people to ignore the value as you
>> do when searching (no one uses the score when doing searches do they?) Only
>> the ranking/ordering makes any real difference.
>>
>> The score is the sum of dot products of all user history vector segments
>> against the correlator fields in the model. If you boost or add properties
>> to your queries they multiply times some of the dot products before
>> summation. So given any boosting the max is hard to state and we have found
>> no particular use for it. Other tuning parameters give you the best ways of
>> affecting the score so looking at the score afterwards is not very helpful
>> except to rank results.
>>
>>
>> On Apr 7, 2017, at 1:02 AM, Vaghawan Ojha <vaghawan...@gmail.com> wrote:
>>
>> Hi,
>>
>> I was just curious, how does the score works in the UR, I've seen the UR
>> presentation in the documentation site, but I couldn't be sure what it
>> really signifies.
>>
>> For example: what's the limit of the score? What's the possibly highest
>> score?
>>
>> Thanks
>>
>>
>> --
>> You received this message because you are subscribed to the Google Groups
>> "actionml-user" group.
>> To unsubscribe from this group and stop receiving emails from it, send an
>> email to actionml-user+unsubscr...@googlegroups.com.
>> To post to this group, send email to actionml-u...@googlegroups.com.
>> To view this discussion on the web visit https://groups.google.com/d/ms
>> gid/actionml-user/CA%2B69RXbuZU6ujHgQWpTizTigXpvahzMeR86c5MP
>> 6EL%2BCr32ZzA%40mail.gmail.com
>> <https://groups.google.com/d/msgid/actionml-user/CA%2B69RXbuZU6ujHgQWpTizTigXpvahzMeR86c5MP6EL%2BCr32ZzA%40mail.gmail.com?utm_medium=email_source=footer>
>> .
>> For more options, visit https://groups.google.com/d/optout.
>>
>>
>


Re: Cart/Complementary purchase recommendation with UR

2017-04-07 Thread Vaghawan Ojha
Hi Pat,

Thanks I will try this in my existing app.

Thanks
Vaghawan

On Thu, Apr 6, 2017 at 9:49 PM, Pat Ferrel <p...@occamsmachete.com> wrote:

> Yes it can be done but support is not ideal so UR v0.6.0 coming just after
> the release of PIO-0.11.0 (in RC now) has some enhancements that give you
> better options and easier queries.
>
> To get truly complimentary purchases you need to train on shopping carts,
> not users purchases. So the primary input would of the form:
>
> (shopping-cart-id, cart-purchase, product-id)
>
> Then in < 0.6.0 you query with the currently active shopping cart id as if
> it is the user-id
>
> {
> “user”: “some-shopping-cart-id”
> }
>
> In v0.6.0 you still need to train on shopping carts to get true
> “complimentary purchase” but the query may be easier:
>
> {
> “Item-set”: [“product1”, “product2”]
> }
>
> The query is the current contents of the shopping cart, no fake “user-id”
> required.
>
> UR 0.6.0 will also allow you to query against a model trained on user
> purchases with the same query but, though the results will be "things
> purchased by people with similar purchasing behavior as you query
> item-set", it won’t be strictly “complimentary purchases”.
>
> Since 2 methods are supported with UR 0.6.0 you could A/B test for which
> has the best results. If using user purchase (instead of shopping carts) is
> close to as good it would be much easier since you can have user-based,
> item-based and shopping carts recs from the same model. Otherwise one for
> user-based and item-based and another for “complimentary purchases” would
> make 2 trains and 2 models.
>
>
>
> On Apr 5, 2017, at 8:08 PM, Vaghawan Ojha <vaghawan...@gmail.com> wrote:
>
> Hi Pat,
>
> As far as the actionml docs tells me that I could be able to use the UR to
> recommend the cart/complementary purchase, but there seems to be no
> documentation regarding this.
>
> It would be helpful if there was a documentation for this feature. I know
> people have worked out and implemented the cart recommendation, but for
> begineer like me, it would still be like "what to do now".
>
> So could you help me how can I do that? If documentation is not ready yet
> for this, I could write one for this feature once I got what to do.
>
> Thanks
> Vaghawan
>
> --
> You received this message because you are subscribed to the Google Groups
> "actionml-user" group.
> To unsubscribe from this group and stop receiving emails from it, send an
> email to actionml-user+unsubscr...@googlegroups.com.
> To post to this group, send email to actionml-u...@googlegroups.com.
> To view this discussion on the web visit https://groups.google.com/d/
> msgid/actionml-user/CA%2B69RXYDeMRXOG25v%3Dadr%
> 3DSXLk6BO811DSCYX0t6vp8EVPP1Xw%40mail.gmail.com
> <https://groups.google.com/d/msgid/actionml-user/CA%2B69RXYDeMRXOG25v%3Dadr%3DSXLk6BO811DSCYX0t6vp8EVPP1Xw%40mail.gmail.com?utm_medium=email_source=footer>
> .
> For more options, visit https://groups.google.com/d/optout.
>
>


Re: How Does The Score Works In UR

2017-04-07 Thread Vaghawan Ojha
Hi Pat,

Thank you very much for the explanation, seems like I forgot to properly
search this question in the group. By ranking you means the order of the
product right?

Also I wanted to replicate such dot products of all user history vector
segments against the correlator fields in the model in a very very small
data, so that I could have a test of what's really working behind. And also
I wanted to learn about the implicating algorithms. I've seen the
presentation in the doc site of UR. Should you have any resources which
could lead me to better understand all these please point me.

Thanks
Vaghawan

On Sat, Apr 8, 2017 at 12:16 AM, Pat Ferrel <p...@occamsmachete.com> wrote:

> The limit depends on your data. I tell people to ignore the value as you
> do when searching (no one uses the score when doing searches do they?) Only
> the ranking/ordering makes any real difference.
>
> The score is the sum of dot products of all user history vector segments
> against the correlator fields in the model. If you boost or add properties
> to your queries they multiply times some of the dot products before
> summation. So given any boosting the max is hard to state and we have found
> no particular use for it. Other tuning parameters give you the best ways of
> affecting the score so looking at the score afterwards is not very helpful
> except to rank results.
>
>
> On Apr 7, 2017, at 1:02 AM, Vaghawan Ojha <vaghawan...@gmail.com> wrote:
>
> Hi,
>
> I was just curious, how does the score works in the UR, I've seen the UR
> presentation in the documentation site, but I couldn't be sure what it
> really signifies.
>
> For example: what's the limit of the score? What's the possibly highest
> score?
>
> Thanks
>
>
> --
> You received this message because you are subscribed to the Google Groups
> "actionml-user" group.
> To unsubscribe from this group and stop receiving emails from it, send an
> email to actionml-user+unsubscr...@googlegroups.com.
> To post to this group, send email to actionml-u...@googlegroups.com.
> To view this discussion on the web visit https://groups.google.com/d/
> msgid/actionml-user/CA%2B69RXbuZU6ujHgQWpTizTigXpvahz
> MeR86c5MP6EL%2BCr32ZzA%40mail.gmail.com
> <https://groups.google.com/d/msgid/actionml-user/CA%2B69RXbuZU6ujHgQWpTizTigXpvahzMeR86c5MP6EL%2BCr32ZzA%40mail.gmail.com?utm_medium=email_source=footer>
> .
> For more options, visit https://groups.google.com/d/optout.
>
>


How Does The Score Works In UR

2017-04-07 Thread Vaghawan Ojha
Hi,

I was just curious, how does the score works in the UR, I've seen the UR
presentation in the documentation site, but I couldn't be sure what it
really signifies.

For example: what's the limit of the score? What's the possibly highest
score?

Thanks


Cart/Complementary purchase recommendation with UR

2017-04-05 Thread Vaghawan Ojha
Hi Pat,

As far as the actionml docs tells me that I could be able to use the UR to
recommend the cart/complementary purchase, but there seems to be no
documentation regarding this.

It would be helpful if there was a documentation for this feature. I know
people have worked out and implemented the cart recommendation, but for
begineer like me, it would still be like "what to do now".

So could you help me how can I do that? If documentation is not ready yet
for this, I could write one for this feature once I got what to do.

Thanks
Vaghawan


Re: Empty result even though there are events

2017-04-04 Thread Vaghawan Ojha
Oh, Ok. That's my mistake then. I think creating a new app would solve the
problem.

Thanks


On Wed, Apr 5, 2017 at 9:04 AM, Donald Szeto <don...@apache.org> wrote:

> Are you using the appName in engine.json that corresponds to the appId you
> used when you did `pio import`? Setting a new appName in engine.json will
> not automatically create a new app. Apps are created when you do `pio app
> new`.
>
> On Tue, Apr 4, 2017 at 8:13 PM Vaghawan Ojha <vaghawan...@gmail.com>
> wrote:
>
> Hi,
>
> Yes, I couldn't use different appName. Even when I changed the appName in
> engine.json and did pio build, the app name in pio app list still reamins
> the same. Something to do with the pio itself, I guess.
>
> Thank you very much for your support.
>
> On Mon, Apr 3, 2017 at 6:58 PM, Pat Ferrel <p...@occamsmachete.com> wrote:
>
> I am unclear. What is the problem? You can’t use different datasets
> (different appNames)?
>
> What is the output of `pio app list`
>
>
> On Apr 2, 2017, at 9:06 PM, Vaghawan Ojha <vaghawan...@gmail.com> wrote:
>
> Hi,
>
> Thank you, I did the changes. But whenever I changed the appname and did
> pio build its Ok, but while training it says invalid appname for any other
> app name. I'm missing something, however when I left the appname as
> handmade, it works.
>
> Now the data and recommendation are all working. Thank you very much, I
> used engine.json.minimum. I will be testing other features, like cart
> recommendation. Here's my engine.json
>
> {
>   "comment":" This config file uses default settings for all but the
> required values see README.md for docs",
>   "id": "default",
>   "description": "Default settings",
>   "engineFactory": "org.template.RecommendationEngine",
>   "datasource": {
> "params" : {
>   "name": "sample-handmade-data.txt",
>   "appName": "handmade",
>   "eventNames": ["purchase", "view"]
> }
>   },
>   "sparkConf": {
> "spark.serializer": "org.apache.spark.serializer.KryoSerializer",
> "spark.kryo.registrator": "org.apache.mahout.sparkbindings.io
> .MahoutKryoRegistrator",
> "spark.kryo.referenceTracking": "false",
> "spark.kryoserializer.buffer.mb": "300",
> "spark.kryoserializer.buffer": "300m",
> "es.index.auto.create": "true"
>   },
>   "algorithms": [
> {
>   "comment": "simplest setup where all values are default, popularity
> based backfill, must add eventsNames",
>   "name": "ur",
>   "params": {
> "appName": "handmade",
> "indexName": "urindex",
> "typeName": "items",
> "comment": "must have data for the first event or the model will
> not build, other events are optional",
> "eventNames": ["purchase", "view"]
>   }
> }
>   ]
> }
>
> I was very curious, what does this means: "name":
> "sample-handmade-data.txt",
> Thanks
>
>
> On Sat, Apr 1, 2017 at 1:29 AM, Pat Ferrel <p...@occamsmachete.com> wrote:
>
> You have active date range filters and no dates in your data. This is
> because you are using the engine.json from the integration test, which uses
> that config.
>
> Start form engine.json.minimum and set the appName for your data and the
> eventNames you will use.
>
>
> On Mar 31, 2017, at 12:52 AM, Vaghawan Ojha <vaghawan...@gmail.com> wrote:
>
> More info: I don't have the secondary event like "View" in my data, I just
> don't collect the view event, since I'm using old purchased history of the
> user.
>
> With integration test data everything is Ok, but with my real data, it's
> not.  All the data looks like this with varying item_id and user_id. There
> is no view events.
>
> {"entityId":"00068116","entityType":"user","event":"
> purchase","targetEntityType":"item","targetEntityId":"BM8941"}
> {"entityId":"00068116","entityType":"user","event":"
> purchase","targetEntityType":"item","targetEntityId":"BM8920"}
> {"entityId":"00088894","entityType":"user","ev

Re: The recommendation engine shows only one result for all users

2017-04-03 Thread Vaghawan Ojha
Hi,

I've approx : 100 event data, all the data looks exactly like this:
{"entityId":"00021938","entityType":"user","event":"buy","targetEntityType":"item","targetEntityId":"BM55432"}
{"entityId":"00040844","entityType":"user","event":"buy","targetEntityType":"item","targetEntityId":"BM59126"}
{"entityId":"00040844","entityType":"user","event":"buy","targetEntityType":"item","targetEntityId":"BM50258"}
{"entityId":"00040844","entityType":"user","event":"buy","targetEntityType":"item","targetEntityId":"BM8920"}
{"entityId":"00040844","entityType":"user","event":"buy","targetEntityType":"item","targetEntityId":"BM11408"}
{"entityId":"00040844","entityType":"user","event":"buy","targetEntityType":"item","targetEntityId":"BM63982"}
{"entityId":"00041181","entityType":"user","event":"buy","targetEntityType":"item","targetEntityId":"BM41244"}
{"entityId":"00041181","entityType":"user","event":"buy","targetEntityType":"item","targetEntityId":"BM31483"}
{"entityId":"00041181","entityType":"user","event":"buy","targetEntityType":"item","targetEntityId":"BM66411"}
{"entityId":"00041181","entityType":"user","event":"buy","targetEntityType":"item","targetEntityId":"BM9068"}
{"entityId":"00040765","entityType":"user","event":"buy","targetEntityType":"item","targetEntityId":"BM8928"}
{"entityId":"00040765","entityType":"user","event":"buy","targetEntityType":"item","targetEntityId":"BM8925"}



It's fine at the size of approx 5 event, but not same with the large
data. Should you need anything else, let me know further.

Thanks

On Fri, Mar 31, 2017 at 7:17 PM, Donald Szeto <don...@apache.org> wrote:

> Hi,
>
> A statistics of your event data's distribution would be helpful.
>
> Regards,
> Donald
>
> On Tue, Mar 28, 2017 at 11:27 AM Vaghawan Ojha <vaghawan...@gmail.com>
> wrote:
>
>> Hi,
>>
>> I've added more than 30 events in the engine, and I'm using this
>> template (https://github.com/apache/incubator-predictionio-
>> template-recommender)
>>
>> When I had small data set, the engine was Ok and it was showing different
>> result for different users but when I've trained 30 of dataset it's
>> giving me the weird result.
>> For every users there is same product recommendation with just the
>> difference in the score.
>>
>> Why this happened, I deleted the data once and tested again but still the
>> same result. I don't know what's being so wrong here.
>>
>> Have anyone had experienced the same? Please let me know.
>>
>> Thank you
>>
>


Re: Empty result even though there are events

2017-03-31 Thread Vaghawan Ojha
More info: I don't have the secondary event like "View" in my data, I just
don't collect the view event, since I'm using old purchased history of the
user.

With integration test data everything is Ok, but with my real data, it's
not.  All the data looks like this with varying item_id and user_id. There
is no view events.

{"entityId":"00068116","entityType":"user","event":"purchase","targetEntityType":"item","targetEntityId":"BM8941"}
{"entityId":"00068116","entityType":"user","event":"purchase","targetEntityType":"item","targetEntityId":"BM8920"}
{"entityId":"00088894","entityType":"user","event":"purchase","targetEntityType":"item","targetEntityId":"BM66336"}
{"entityId":"00097710","entityType":"user","event":"purchase","targetEntityType":"item","targetEntityId":"BM42110"}
{"entityId":"00097710","entityType":"user","event":"purchase","targetEntityType":"item","targetEntityId":"BM61833"}
{"entityId":"00097710","entityType":"user","event":"purchase","targetEntityType":"item","targetEntityId":"BM65826"}
{"entityId":"00097710","entityType":"user","event":"purchase","targetEntityType":"item","targetEntityId":"BM40182"}
{"entityId":"00097710","entityType":"user","event":"purchase","targetEntityType":"item","targetEntityId":"BM5001"}

Is my data wrong? I am able to successfully  train and deploy the engine.

On Fri, Mar 31, 2017 at 11:40 AM, Vaghawan Ojha <vaghawan...@gmail.com>
wrote:

> I've 5 transection data for experiment. I don't  think it's the data
> because with the same data I could get recommendation using Recommendation
> template.
>
> On Fri, Mar 31, 2017 at 11:29 AM, Vaghawan Ojha <vaghawan...@gmail.com>
> wrote:
>
>> Hi,
>>
>> I imported the bulk data using pio import --appid (id) --input  data.json
>> . The import was successful, then I build and trained and deployed the UR,
>> but when I query the server, it always returns a empty
>> json {"itemScores":[]} .
>>
>>
>> I think,  I must be missing something in the config, here is my
>> engine.json
>>
>> {
>>   "comment":" This config file uses default settings for all but the
>> required values see README.md for docs",
>>   "id": "default",
>>   "description": "Default settings",
>>   "engineFactory": "org.template.RecommendationEngine",
>>   "datasource": {
>> "params" : {
>>   "name": "sample-handmade-data.txt",
>>   "appName": "handmade",
>>   "eventNames": ["purchase", "view"]
>> }
>>   },
>>   "sparkConf": {
>> "spark.serializer": "org.apache.spark.serializer.KryoSerializer",
>> "spark.kryo.registrator": "org.apache.mahout.sparkbindings.io
>> .MahoutKryoRegistrator",
>> "spark.kryo.referenceTracking": "false",
>> "spark.kryoserializer.buffer": "300m",
>> "es.index.auto.create": "true"
>>   },
>>   "algorithms": [
>> {
>>   "comment": "simplest setup where all values are default, popularity
>> based backfill, must add eventsNames",
>>   "name": "ur",
>>   "params": {
>> "appName": "handmade",
>> "indexName": "urindex",
>> "typeName": "items",
>> "comment": "must have data for the first event or the model will
>> not build, other events are optional",
>> "indicators": [
>>   {
>> "name": "purchase"
>>   },{
>> "name": "view",
>> "maxCorrelatorsPerItem": 50
>>   }
>> ],
>> "availableDateName": "available",
>> "expireDateName": "expires",
>> "dateName": "date",
>> "num": 4
>>   }
>> }
>>   ]
>> }
>>
>> is the data supposed to come from the sample-handmade-data.txt? I'm not
>> sure this is weird. My json data import and trains correctly without an
>> error. And example integration runs fine.
>>
>
>


Re: Empty result even though there are events

2017-03-30 Thread Vaghawan Ojha
I've 5 transection data for experiment. I don't  think it's the data
because with the same data I could get recommendation using Recommendation
template.

On Fri, Mar 31, 2017 at 11:29 AM, Vaghawan Ojha <vaghawan...@gmail.com>
wrote:

> Hi,
>
> I imported the bulk data using pio import --appid (id) --input  data.json
> . The import was successful, then I build and trained and deployed the UR,
> but when I query the server, it always returns a empty
> json {"itemScores":[]} .
>
>
> I think,  I must be missing something in the config, here is my
> engine.json
>
> {
>   "comment":" This config file uses default settings for all but the
> required values see README.md for docs",
>   "id": "default",
>   "description": "Default settings",
>   "engineFactory": "org.template.RecommendationEngine",
>   "datasource": {
> "params" : {
>   "name": "sample-handmade-data.txt",
>   "appName": "handmade",
>   "eventNames": ["purchase", "view"]
> }
>   },
>   "sparkConf": {
> "spark.serializer": "org.apache.spark.serializer.KryoSerializer",
> "spark.kryo.registrator": "org.apache.mahout.sparkbindings.io.
> MahoutKryoRegistrator",
> "spark.kryo.referenceTracking": "false",
> "spark.kryoserializer.buffer": "300m",
> "es.index.auto.create": "true"
>   },
>   "algorithms": [
> {
>   "comment": "simplest setup where all values are default, popularity
> based backfill, must add eventsNames",
>   "name": "ur",
>   "params": {
> "appName": "handmade",
> "indexName": "urindex",
> "typeName": "items",
> "comment": "must have data for the first event or the model will
> not build, other events are optional",
> "indicators": [
>   {
> "name": "purchase"
>   },{
> "name": "view",
> "maxCorrelatorsPerItem": 50
>   }
> ],
> "availableDateName": "available",
> "expireDateName": "expires",
> "dateName": "date",
> "num": 4
>   }
> }
>   ]
> }
>
> is the data supposed to come from the sample-handmade-data.txt? I'm not
> sure this is weird. My json data import and trains correctly without an
> error. And example integration runs fine.
>


Empty result even though there are events

2017-03-30 Thread Vaghawan Ojha
Hi,

I imported the bulk data using pio import --appid (id) --input  data.json .
The import was successful, then I build and trained and deployed the UR,
but when I query the server, it always returns a empty
json {"itemScores":[]} .


I think,  I must be missing something in the config, here is my engine.json

{
  "comment":" This config file uses default settings for all but the
required values see README.md for docs",
  "id": "default",
  "description": "Default settings",
  "engineFactory": "org.template.RecommendationEngine",
  "datasource": {
"params" : {
  "name": "sample-handmade-data.txt",
  "appName": "handmade",
  "eventNames": ["purchase", "view"]
}
  },
  "sparkConf": {
"spark.serializer": "org.apache.spark.serializer.KryoSerializer",
"spark.kryo.registrator":
"org.apache.mahout.sparkbindings.io.MahoutKryoRegistrator",
"spark.kryo.referenceTracking": "false",
"spark.kryoserializer.buffer": "300m",
"es.index.auto.create": "true"
  },
  "algorithms": [
{
  "comment": "simplest setup where all values are default, popularity
based backfill, must add eventsNames",
  "name": "ur",
  "params": {
"appName": "handmade",
"indexName": "urindex",
"typeName": "items",
"comment": "must have data for the first event or the model will
not build, other events are optional",
"indicators": [
  {
"name": "purchase"
  },{
"name": "view",
"maxCorrelatorsPerItem": 50
  }
],
"availableDateName": "available",
"expireDateName": "expires",
"dateName": "date",
"num": 4
  }
}
  ]
}

is the data supposed to come from the sample-handmade-data.txt? I'm not
sure this is weird. My json data import and trains correctly without an
error. And example integration runs fine.


Universal Recommender is recommending example data even after training the real data

2017-03-30 Thread Vaghawan Ojha
Hi,

I first tested the UR with example integration, and then I imported the my
real data and t rained the engine, Everything is Ok in deployment.

But when I make a real query, it stills shows me the example data, I'm
missing something, but what I'm missing I am not sure.

Can you please help me?

Thank you
Vaghawan


Re: Need Help setting up Universal Recommended with PIO

2017-03-30 Thread Vaghawan Ojha
Hi Thank yo very much for making me clear.


Thank you very much


On Thu, Mar 30, 2017 at 6:04 PM, Marius Rabenarivo <
mariusrabenar...@gmail.com> wrote:

> You have to run
>
> ./make-distribution.sh
>
> inside incubator-predictionio (the directory containing the repository where 
> you cloned PredictionIO)
>
> it will generate a
> PredictionIO-0.10.0-incubating.tar.gz
>
> that you will have to untar somewhere.
>
> For my case I make the command
>
> tar xvzf PredictionIO-0.10.0-incubating.tar.gz -C ~/
>
> to untar it to my home directory
>
> After that you have to add
> PredictionIO-0.10.0-incubating/bin to you PATH environment variable to be 
> able to run pio command from any directory.
>
> After that you have to cd to universal-recommender directory
>
> pio build
>
> (this generate the manifest.json)
>
> then
>
> pio train
>
> And
>
> pio deploy
>
>
> 2017-03-30 16:13 GMT+04:00 Vaghawan Ojha <vaghawan...@gmail.com>:
>
>> Ok, So universal recommender stays at a different dir independently, so I
>> can't see in the documentation where it tells me to build pio. Without
>> building how can I deploy.
>>
>> Also the quick start tells to pio deploy in the universal recommender
>> dir, in my case it doesn't work. Since there will be no manifest file.
>>
>> Thanks
>>
>> On Thu, Mar 30, 2017 at 5:53 PM, Marius Rabenarivo <
>> mariusrabenar...@gmail.com> wrote:
>>
>>> *universal-recommender
>>>
>>> 2017-03-30 16:08 GMT+04:00 Marius Rabenarivo <mariusrabenar...@gmail.com
>>> >:
>>>
>>>> You have just to clone it once and put it in a directory (I put it in
>>>> my home directory)
>>>>
>>>> You can name it like you want.
>>>> When cloning without providing a directory name you will have the repo
>>>> in univeral-recommender (It's what I do)
>>>>
>>>> 2017-03-30 15:58 GMT+04:00 Vaghawan Ojha <vaghawan...@gmail.com>:
>>>>
>>>>> Ok, I am gonna do that now.
>>>>>
>>>>> There are lots of confusions in the documentation for me, due to
>>>>> multiple options of installation. I'm trying a fresh install.
>>>>>
>>>>> "Clone Universal Recommender Template from its root repo into
>>>>> ~/universal or do similar for any other template."
>>>>>
>>>>> What does this means? am I supposed to clone the UR into pio dir or
>>>>> outside of it?
>>>>>
>>>>> And again in the quickstart_doc it says again to clone the UR in root.
>>>>>
>>>>> Am I not doing the same thing twice? as per apache encubator, the
>>>>> template should be inside the pio instllation.
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> On Thu, Mar 30, 2017 at 5:36 PM, Marius Rabenarivo <
>>>>> mariusrabenar...@gmail.com> wrote:
>>>>>
>>>>>> Hi,
>>>>>>
>>>>>> As you can see in : https://github.com/actionml/PredictionIO
>>>>>> it was forked from : https://github.com/apache/incubator-predictionio
>>>>>>
>>>>>> What I advise you is cloning the https://github.com/apache/incu
>>>>>> bator-predictionio repo and
>>>>>> git checkout release/0.10.0
>>>>>>
>>>>>> 2017-03-30 15:43 GMT+04:00 Vaghawan Ojha <vaghawan...@gmail.com>:
>>>>>>
>>>>>>> Hi,
>>>>>>>
>>>>>>> Are you telling me to use either one of them?
>>>>>>>
>>>>>>> So let's say I only cloned the first one, the merged branch from
>>>>>>> actionml. Am I supposed to skip the later one? like cloning 
>>>>>>> apacheecubator
>>>>>>> and so on?
>>>>>>>
>>>>>>> Why the documenation is like that? Since you've already installed
>>>>>>> it, can you please tell me which could be the best approach?
>>>>>>>
>>>>>>> Like installing encubator or action ml's? The documentation confused
>>>>>>> me badly. Please suggest me.
>>>>>>>
>>>>>>> On Thu, Mar 30, 2017 at 5:20 PM, Marius Rabenarivo <
>>>>>>> mariusrabenar...@gmail.com> wrote:
>>>>>>>
>>>>

Re: Need Help setting up Universal Recommended with PIO

2017-03-30 Thread Vaghawan Ojha
Ok, So universal recommender stays at a different dir independently, so I
can't see in the documentation where it tells me to build pio. Without
building how can I deploy.

Also the quick start tells to pio deploy in the universal recommender dir,
in my case it doesn't work. Since there will be no manifest file.

Thanks

On Thu, Mar 30, 2017 at 5:53 PM, Marius Rabenarivo <
mariusrabenar...@gmail.com> wrote:

> *universal-recommender
>
> 2017-03-30 16:08 GMT+04:00 Marius Rabenarivo <mariusrabenar...@gmail.com>:
>
>> You have just to clone it once and put it in a directory (I put it in my
>> home directory)
>>
>> You can name it like you want.
>> When cloning without providing a directory name you will have the repo in
>> univeral-recommender (It's what I do)
>>
>> 2017-03-30 15:58 GMT+04:00 Vaghawan Ojha <vaghawan...@gmail.com>:
>>
>>> Ok, I am gonna do that now.
>>>
>>> There are lots of confusions in the documentation for me, due to
>>> multiple options of installation. I'm trying a fresh install.
>>>
>>> "Clone Universal Recommender Template from its root repo into
>>> ~/universal or do similar for any other template."
>>>
>>> What does this means? am I supposed to clone the UR into pio dir or
>>> outside of it?
>>>
>>> And again in the quickstart_doc it says again to clone the UR in root.
>>>
>>> Am I not doing the same thing twice? as per apache encubator, the
>>> template should be inside the pio instllation.
>>>
>>>
>>>
>>>
>>>
>>> On Thu, Mar 30, 2017 at 5:36 PM, Marius Rabenarivo <
>>> mariusrabenar...@gmail.com> wrote:
>>>
>>>> Hi,
>>>>
>>>> As you can see in : https://github.com/actionml/PredictionIO
>>>> it was forked from : https://github.com/apache/incubator-predictionio
>>>>
>>>> What I advise you is cloning the https://github.com/apache/incu
>>>> bator-predictionio repo and
>>>> git checkout release/0.10.0
>>>>
>>>> 2017-03-30 15:43 GMT+04:00 Vaghawan Ojha <vaghawan...@gmail.com>:
>>>>
>>>>> Hi,
>>>>>
>>>>> Are you telling me to use either one of them?
>>>>>
>>>>> So let's say I only cloned the first one, the merged branch from
>>>>> actionml. Am I supposed to skip the later one? like cloning apacheecubator
>>>>> and so on?
>>>>>
>>>>> Why the documenation is like that? Since you've already installed it,
>>>>> can you please tell me which could be the best approach?
>>>>>
>>>>> Like installing encubator or action ml's? The documentation confused
>>>>> me badly. Please suggest me.
>>>>>
>>>>> On Thu, Mar 30, 2017 at 5:20 PM, Marius Rabenarivo <
>>>>> mariusrabenar...@gmail.com> wrote:
>>>>>
>>>>>> Hello,
>>>>>>
>>>>>> The first is the ActionML's fork but it was merged with the upstream
>>>>>> incubator predictionIO repo.
>>>>>> You just have to choose on of them.
>>>>>>
>>>>>> I think you should use de release branch instead of the master one.
>>>>>>
>>>>>> Regards,
>>>>>>
>>>>>> Marius
>>>>>>
>>>>>> 2017-03-30 15:19 GMT+04:00 Vaghawan Ojha <vaghawan...@gmail.com>:
>>>>>>
>>>>>>> I've ran into confusion as well as after wrong setup probably
>>>>>>> because I'm not getting this documentation correctly.
>>>>>>>
>>>>>>> Documentation URL (http://actionml.com/docs/single_machine)
>>>>>>>
>>>>>>> First the documentation tells me to do something like this:
>>>>>>>
>>>>>>> Clone PredictionIO-0.10.0 from its root repo into ~/pio
>>>>>>>
>>>>>>> git clone https://github.com/actionml/PredictionIO.git pio
>>>>>>> cd ~/pio
>>>>>>> git checkout master #get the latest stable version
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> I followed everything in the guide, and the guide again tells me
>>>>>>> something like this:
>>>>>>>
>>>>>>> Setup PredictionIO
>>>>>>>
>>>>>>> PredictionIO is a source only release so you will need to build it.
>>>>>>>
>>>>>>>-
>>>>>>>
>>>>>>>Build PredictionIO
>>>>>>>
>>>>>>>git clone https://github.com/apache/incubator-predictionio.git ~/pio
>>>>>>>cd ~/pio
>>>>>>>git checkout master # usually the version you want
>>>>>>>./make-distribution
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> So, I'm confused why the documentation is telling me something like
>>>>>>> this, since I already have the dir called pio. I also tried with the
>>>>>>> installation into a different dir, and completed the quickstart of UR, 
>>>>>>> but
>>>>>>> at the end I just wasted my time.
>>>>>>>
>>>>>>> So my question here is I can't make two dir with the same name pio.
>>>>>>> But why the documentation tells me this?
>>>>>>>
>>>>>>> If you have any other way of installation please let me know. I
>>>>>>> would be really grateful, if you could help me out.
>>>>>>>
>>>>>>> Thank you
>>>>>>> Vaghawan
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>


Re: Need Help setting up Universal Recommended with PIO

2017-03-30 Thread Vaghawan Ojha
Ok, I am gonna do that now.

There are lots of confusions in the documentation for me, due to multiple
options of installation. I'm trying a fresh install.

"Clone Universal Recommender Template from its root repo into ~/universal or
do similar for any other template."

What does this means? am I supposed to clone the UR into pio dir or outside
of it?

And again in the quickstart_doc it says again to clone the UR in root.

Am I not doing the same thing twice? as per apache encubator, the template
should be inside the pio instllation.





On Thu, Mar 30, 2017 at 5:36 PM, Marius Rabenarivo <
mariusrabenar...@gmail.com> wrote:

> Hi,
>
> As you can see in : https://github.com/actionml/PredictionIO
> it was forked from : https://github.com/apache/incubator-predictionio
>
> What I advise you is cloning the https://github.com/apache/
> incubator-predictionio repo and
> git checkout release/0.10.0
>
> 2017-03-30 15:43 GMT+04:00 Vaghawan Ojha <vaghawan...@gmail.com>:
>
>> Hi,
>>
>> Are you telling me to use either one of them?
>>
>> So let's say I only cloned the first one, the merged branch from
>> actionml. Am I supposed to skip the later one? like cloning apacheecubator
>> and so on?
>>
>> Why the documenation is like that? Since you've already installed it, can
>> you please tell me which could be the best approach?
>>
>> Like installing encubator or action ml's? The documentation confused me
>> badly. Please suggest me.
>>
>> On Thu, Mar 30, 2017 at 5:20 PM, Marius Rabenarivo <
>> mariusrabenar...@gmail.com> wrote:
>>
>>> Hello,
>>>
>>> The first is the ActionML's fork but it was merged with the upstream
>>> incubator predictionIO repo.
>>> You just have to choose on of them.
>>>
>>> I think you should use de release branch instead of the master one.
>>>
>>> Regards,
>>>
>>> Marius
>>>
>>> 2017-03-30 15:19 GMT+04:00 Vaghawan Ojha <vaghawan...@gmail.com>:
>>>
>>>> I've ran into confusion as well as after wrong setup probably because
>>>> I'm not getting this documentation correctly.
>>>>
>>>> Documentation URL (http://actionml.com/docs/single_machine)
>>>>
>>>> First the documentation tells me to do something like this:
>>>>
>>>> Clone PredictionIO-0.10.0 from its root repo into ~/pio
>>>>
>>>> git clone https://github.com/actionml/PredictionIO.git pio
>>>> cd ~/pio
>>>> git checkout master #get the latest stable version
>>>>
>>>>
>>>>
>>>> I followed everything in the guide, and the guide again tells me
>>>> something like this:
>>>>
>>>> Setup PredictionIO
>>>>
>>>> PredictionIO is a source only release so you will need to build it.
>>>>
>>>>-
>>>>
>>>>Build PredictionIO
>>>>
>>>>git clone https://github.com/apache/incubator-predictionio.git ~/pio
>>>>cd ~/pio
>>>>git checkout master # usually the version you want
>>>>./make-distribution
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> So, I'm confused why the documentation is telling me something like
>>>> this, since I already have the dir called pio. I also tried with the
>>>> installation into a different dir, and completed the quickstart of UR, but
>>>> at the end I just wasted my time.
>>>>
>>>> So my question here is I can't make two dir with the same name pio. But
>>>> why the documentation tells me this?
>>>>
>>>> If you have any other way of installation please let me know. I would
>>>> be really grateful, if you could help me out.
>>>>
>>>> Thank you
>>>> Vaghawan
>>>>
>>>>
>>>>
>>>>
>>>
>>
>


Re: ERROR: Remote Elastic search cluster

2017-03-28 Thread Vaghawan Ojha
Hi Malay,

Have you configured the configuration of the elasticsearch like hosts,
ports and clustername in elasticsearch.yaml?

(cluster.name: elasticsearch) for cluster name?

I had ran into similar kind of issues, but for me I forgot to put the
cluster name in configuration. That worked for me make sure you have your
configuration good to go.

Thanks
Vaghawan


On Wed, Mar 29, 2017 at 10:28 AM, Ambuj Sharma  wrote:

> make sure elasticSearch is running nd available from PIO Server. Try to
> send request to ES using Curl
>
>
> Thanks and Regards
> Ambuj Sharma
> Sunrise may late, But Morning is sure.
> Team ML
> Betaout
>
> On Tue, Mar 28, 2017 at 11:39 PM, Malay Tripathi  > wrote:
>
>> Hi Ambuj,
>>
>> I did try below configuration and ran "pio status", same error.
>>
>> PIO_STORAGE_SOURCES_ELASTICSEARCH_TYPE=elasticsearch
>>
>> PIO_STORAGE_SOURCES_ELASTICSEARCH_CLUSTERNAME=prod
>>
>> PIO_STORAGE_SOURCES_ELASTICSEARCH_HOSTS=aabbcc176
>>
>> PIO_STORAGE_SOURCES_ELASTICSEARCH_PORTS=9300
>>
>>
>> Thanks,
>>
>> Malay
>>
>> On Tue, Mar 28, 2017 at 2:34 AM, Ambuj Sharma 
>> wrote:
>>
>>> You need to configure only these properties
>>>
>>> PIO_STORAGE_SOURCES_ELASTICSEARCH_TYPE=elasticsearch
>>> PIO_STORAGE_SOURCES_ELASTICSEARCH_CLUSTERNAME=
>>> PIO_STORAGE_SOURCES_ELASTICSEARCH_HOSTS=
>>> PIO_STORAGE_SOURCES_ELASTICSEARCH_PORTS=9300
>>>
>>>
>>>
>>> Thanks and Regards
>>> Ambuj Sharma
>>> Sunrise may late, But Morning is sure.
>>> Team ML
>>> Betaout
>>>
>>> On Tue, Mar 28, 2017 at 12:47 PM, Malay Tripathi <
>>> malaytripat...@gmail.com> wrote:
>>>
 Hi,

 I have setup 3 node cluster for ES - aabbcc174, aabbcc175, aabbcc176.
 In my spark edge node, I have installed PredictionIO & given below
 properties in conf/pio-env.sh-

 PIO_STORAGE_SOURCES_ELASTICSEARCH_TYPE=elasticsearch

 PIO_STORAGE_SOURCES_ELASTICSEARCH_CLUSTERNAME=prod

 PIO_STORAGE_SOURCES_ELASTICSEARCH_HOSTS= aabbcc174, aabbcc175,
 aabbcc176

 PIO_STORAGE_SOURCES_ELASTICSEARCH_PORTS=9300,9300,9300


 But if i run *"pio status"* I keep getting error -

 [ERROR] [Console$] Unable to connect to all storage backends
 successfully. The following shows the error message from the storage
 backend.

 [ERROR] [Console$] None of the configured nodes are available: []
 (org.elasticsearch.client.transport.NoNodeAvailableException)

 [ERROR] [Console$] Dumping configuration of initialized storage backend
 sources. Please make sure they are correct.

 [ERROR] [Console$] Source Name: ELASTICSEARCH; Type: elasticsearch;
 Configuration: HOSTS -> aabbcc174, aabbcc175, aabbcc176, TYPE ->
 elasticsearch, CLUSTERNAME -> prod


 Although If i CURL my elastic search from this edge node where "pio
 status" is failing I get successful response-

 [xx@edgeNode conf]$ curl http://aabbcc174:9200/_nodes/p
 rocess?pretty

 {

   "cluster_name" : "prod",

   "nodes" : {

 "Jru48ImiSuSbSe95e0Qh8A" : {

   "name" : "prod-node1",

 .

  Thanks !!!

>>>
>>>
>>
>


The recommendation engine shows only one result for all users

2017-03-28 Thread Vaghawan Ojha
Hi,

I've added more than 30 events in the engine, and I'm using this
template (
https://github.com/apache/incubator-predictionio-template-recommender)

When I had small data set, the engine was Ok and it was showing different
result for different users but when I've trained 30 of dataset it's
giving me the weird result.
For every users there is same product recommendation with just the
difference in the score.

Why this happened, I deleted the data once and tested again but still the
same result. I don't know what's being so wrong here.

Have anyone had experienced the same? Please let me know.

Thank you


Re: Need Help Building

2017-03-26 Thread Vaghawan Ojha
Hi,

It too seems to be caused by the storage. When I'm
in www/cd/apache-predictionio-0.10.0-incubating/$

sudo PredictionIO-0.10.0-incubating/bin/pio status

 is working fine, but when I do just pio status it shows the error
something like this:
[INFO] [Storage$] Verifying Model Data Backend (Source: LOCALFS)...
[ERROR] [Storage$] Error initializing storage client for source LOCALFS

How can I solve the error I can't do sudo
PredictionIO-0.10.0-incubating/bin/pio train, because I would be in another
dir www/cd/apache-predictionio-0.10.0-incubating/MyRecommendation$

How could I solve this, this seems to be the problem of not having sodoers
for pio command.

On Sun, Mar 26, 2017 at 11:22 PM, Vaghawan Ojha <vaghawan...@gmail.com>
wrote:

> Hi,
> Thanks but the error was because I was not inside the template dir while
> running pio build. It builded now successfully, but it seems in every step
> there is some crazy errors awaiting for me. Now it actually fails at
> training. Can you suggest me anything from the train log?
> I'm sorry but they are really hard to grab unless I ask for help.
>
> Thank you very much
>
> On Sun, Mar 26, 2017 at 10:00 PM, Marius Rabenarivo <
> mariusrabenar...@gmail.com> wrote:
>
>> Hi,
>>
>> The error is :
>>
>> [ERROR] [Storage$] Error initializing storage client for source PGSQL
>>
>> I think you need to change it to HBASE if you want to use HBase
>>
>> PIO_STORAGE_REPOSITORIES_EVENTDATA_SOURCE=PGSQL
>> ->
>> PIO_STORAGE_REPOSITORIES_EVENTDATA_SOURCE=HBASE
>>
>> in your pio-env.sh
>>
>> And start HBase before if not using the pio-start-all script.
>>
>> If you want to use PostreSQL pio-start-all attempt to start it too.
>>
>>
>>
>>
>>
>>
>>
>> 2017-03-26 19:29 GMT+04:00 Vaghawan Ojha <vaghawan...@gmail.com>:
>>
>>> I followed the procedure of manual install, everything was fine until I
>>> stumbled into the pio build.
>>>
>>> I've a directory something like this /abc/pio0.0.10/pio and inside that
>>> another dir pio, in total it would be like :
>>> /abc/pio0.0.10/pio /
>>>
>>> where do I actually run build? inside /abc/pio0.0.10 or
>>> /abc/pio0.0.10/pio / ?
>>>
>>> I don't know but I get some weird errors which I can't properly
>>> diagnose. I"ve attached my log file here. I've followed to load the engine
>>> template. here http://predictionio.incubator.apache.org/templates/reco
>>> mmendation/quickstart/
>>>
>>>
>>>
>>
>


Re: Error while importing the event.json file.

2017-03-26 Thread Vaghawan Ojha
But when I restart the whole machine, it again works. Its quite weird.

On Sun, Mar 26, 2017 at 7:16 PM, Vaghawan Ojha <vaghawan...@gmail.com>
wrote:

> Sorry even pio status shows some errors now:
>
> [ERROR] [RecoverableZooKeeper] ZooKeeper exists failed after 1 attempts
> [ERROR] [ZooKeeperWatcher] hconnection-0x39ad12b6, quorum=localhost:2181,
> baseZNode=/hbase Received unexpected KeeperException, re-throwing exception
> [WARN] [ZooKeeperRegistry] Can't retrieve clusterId from Zookeeper
> [ERROR] [StorageClient] Cannot connect to ZooKeeper (ZooKeeper ensemble:
> localhost). Please make sure that the configuration is pointing at the
> correct ZooKeeper ensemble. By default, HBase manages its own ZooKeeper, so
> if you have not configured HBase to use an external ZooKeeper, that means
> your HBase is not started or configured properly.
> [ERROR] [Storage$] Error initializing storage client for source HBASE
> [ERROR] [Console$] Unable to connect to all storage backends successfully.
> The following shows the error message from the storage backend.
> [ERROR] [Console$] Data source HBASE was not properly initialized.
> (org.apache.predictionio.data.storage.StorageClientException)
> [ERROR] [Console$] Dumping configuration of initialized storage backend
> sources. Please make sure they are correct.
> [ERROR] [Console$] Source Name: ELASTICSEARCH; Type: elasticsearch;
> Configuration: HOME -> /var/www/abc/apache-predictionio-0.10.0-
> incubating/PredictionIO-0.10.0-incubating/vendors/elasticsearch-1.4.4,
> HOSTS -> localhost, PORTS -> 9300, CLUSTERNAME -> elasticsearch, TYPE ->
> elasticsearch
> [ERROR] [Console$] Source Name: LOCALFS; Type: localfs; Configuration:
> PATH -> /home/ekbana-php/.pio_store/models, TYPE -> localfs
> [ERROR] [Console$] Source Name: HBASE; Type: (error); Configuration:
> (error)
>
>
> How would I actually solve it?
>
> On Sun, Mar 26, 2017 at 7:13 PM, Vaghawan Ojha <vaghawan...@gmail.com>
> wrote:
>
>> While importing the event data, I get the following error, I am new and I
>> don't even understand  what does they means? While doing pio status it's
>> shows everything ok. What's wrong here.
>>
>> sudo PredictionIO-0.10.0-incubating/bin/pio import --appid 1 --input
>> my_events.json
>> [INFO] [Runner$] Submission command: /var/www/abc/apache-prediction
>> io-0.10.0-incubating/PredictionIO-0.10.0-incubating/vendors/spark-1.5.
>> 1-bin-hadoop2.6/bin/spark-submit --class 
>> org.apache.predictionio.tools.imprt.FileToEvents
>> --files file:/var/www/abc/apache-predictionio-0.10.0-incubating/
>> PredictionIO-0.10.0-incubating/conf/log4j.properti
>> es,file:/var/www/abc/apache-predictionio-0.10.0-incubating
>> /PredictionIO-0.10.0-incubating/vendors/hbase-1.0.0/conf/hbase-site.xml
>> --driver-class-path /var/www/abc/apache-predictionio-0.10.0-incubating/
>> PredictionIO-0.10.0-incubating/conf:/var/www/abc/
>> apache-predictionio-0.10.0-incubating/PredictionIO-0.10.
>> 0-incubating/vendors/elasticsearch-1.4.4/conf:/var/www/abc/
>> apache-predictionio-0.10.0-incubating/PredictionIO-0.10.
>> 0-incubating/lib/postgresql-9.4-1204.jdbc41.jar:/var/www/
>> abc/apache-predictionio-0.10.0-incubating/PredictionIO-0.
>> 10.0-incubating/lib/mysql-connector-java-5.1.37.jar:/var/
>> www/abc/apache-predictionio-0.10.0-incubating/PredictionIO-
>> 0.10.0-incubating/vendors/spark-1.5.1-bin-hadoop2.6/
>> conf:/var/www/abc/apache-predictionio-0.10.0-
>> incubating/PredictionIO-0.10.0-incubating/vendors/hbase-1.0.0/conf
>> file:/var/www/abc/apache-predictionio-0.10.0-incubating/
>> PredictionIO-0.10.0-incubating/lib/pio-assembly-0.10.0-incubating.jar
>> --appid 1 --input file:/var/www/abc/apache-predi
>> ctionio-0.10.0-incubating/my_events.json --env
>> PIO_STORAGE_SOURCES_HBASE_TYPE=hbase,PIO_ENV_LOADED=1,PIO_
>> STORAGE_REPOSITORIES_METADATA_NAME=pio_meta,PIO_FS_BASEDIR=/
>> home/ekbana-php/.pio_store,PIO_STORAGE_SOURCES_ELASTICSEA
>> RCH_HOSTS=localhost,PIO_STORAGE_SOURCES_HBASE_HOME=/
>> var/www/abc/apache-predictionio-0.10.0-incubating/PredictionIO-0.10.0-
>> incubating/vendors/hbase-1.0.0,PIO_HOME=/var/www/abc/apache
>> -predictionio-0.10.0-incubating/PredictionIO-0.10.0-
>> incubating,PIO_FS_ENGINESDIR=/home/ekbana-php/.pio_store/
>> engines,PIO_STORAGE_SOURCES_LOCALFS_PATH=/home/ekbana-php/
>> .pio_store/models,PIO_STORAGE_SOURCES_ELASTICSEARCH_TYPE=ela
>> sticsearch,PIO_STORAGE_REPOSITORIES_METADATA_SOURCE=ELASTICS
>> EARCH,PIO_STORAGE_REPOSITORIES_MODELDATA_SOURCE=LOCALFS,PIO_
>> STORAGE_REPOSITORIES_EVENTDATA_NAME=pio_event,PIO_
>> STORAGE_SOURCES_ELASTICSEARCH_CLUSTERNAME=elasticsearch,PIO_
>> STORAG

Error while importing the event.json file.

2017-03-26 Thread Vaghawan Ojha
While importing the event data, I get the following error, I am new and I
don't even understand  what does they means? While doing pio status it's
shows everything ok. What's wrong here.

sudo PredictionIO-0.10.0-incubating/bin/pio import --appid 1 --input
my_events.json
[INFO] [Runner$] Submission command:
/var/www/abc/apache-predictionio-0.10.0-incubating/PredictionIO-0.10.0-incubating/vendors/spark-1.5.1-bin-hadoop2.6/bin/spark-submit
--class org.apache.predictionio.tools.imprt.FileToEvents --files
file:/var/www/abc/apache-predictionio-0.10.0-incubating/PredictionIO-0.10.0-incubating/conf/log4j.properties,file:/var/www/abc/apache-predictionio-0.10.0-incubating/PredictionIO-0.10.0-incubating/vendors/hbase-1.0.0/conf/hbase-site.xml
--driver-class-path
/var/www/abc/apache-predictionio-0.10.0-incubating/PredictionIO-0.10.0-incubating/conf:/var/www/abc/apache-predictionio-0.10.0-incubating/PredictionIO-0.10.0-incubating/vendors/elasticsearch-1.4.4/conf:/var/www/abc/apache-predictionio-0.10.0-incubating/PredictionIO-0.10.0-incubating/lib/postgresql-9.4-1204.jdbc41.jar:/var/www/abc/apache-predictionio-0.10.0-incubating/PredictionIO-0.10.0-incubating/lib/mysql-connector-java-5.1.37.jar:/var/www/abc/apache-predictionio-0.10.0-incubating/PredictionIO-0.10.0-incubating/vendors/spark-1.5.1-bin-hadoop2.6/conf:/var/www/abc/apache-predictionio-0.10.0-incubating/PredictionIO-0.10.0-incubating/vendors/hbase-1.0.0/conf
file:/var/www/abc/apache-predictionio-0.10.0-incubating/PredictionIO-0.10.0-incubating/lib/pio-assembly-0.10.0-incubating.jar
--appid 1 --input
file:/var/www/abc/apache-predictionio-0.10.0-incubating/my_events.json
--env
PIO_STORAGE_SOURCES_HBASE_TYPE=hbase,PIO_ENV_LOADED=1,PIO_STORAGE_REPOSITORIES_METADATA_NAME=pio_meta,PIO_FS_BASEDIR=/home/ekbana-php/.pio_store,PIO_STORAGE_SOURCES_ELASTICSEARCH_HOSTS=localhost,PIO_STORAGE_SOURCES_HBASE_HOME=/var/www/abc/apache-predictionio-0.10.0-incubating/PredictionIO-0.10.0-incubating/vendors/hbase-1.0.0,PIO_HOME=/var/www/abc/apache-predictionio-0.10.0-incubating/PredictionIO-0.10.0-incubating,PIO_FS_ENGINESDIR=/home/ekbana-php/.pio_store/engines,PIO_STORAGE_SOURCES_LOCALFS_PATH=/home/ekbana-php/.pio_store/models,PIO_STORAGE_SOURCES_ELASTICSEARCH_TYPE=elasticsearch,PIO_STORAGE_REPOSITORIES_METADATA_SOURCE=ELASTICSEARCH,PIO_STORAGE_REPOSITORIES_MODELDATA_SOURCE=LOCALFS,PIO_STORAGE_REPOSITORIES_EVENTDATA_NAME=pio_event,PIO_STORAGE_SOURCES_ELASTICSEARCH_CLUSTERNAME=elasticsearch,PIO_STORAGE_SOURCES_ELASTICSEARCH_HOME=/var/www/abc/apache-predictionio-0.10.0-incubating/PredictionIO-0.10.0-incubating/vendors/elasticsearch-1.4.4,PIO_FS_TMPDIR=/home/ekbana-php/.pio_store/tmp,PIO_STORAGE_REPOSITORIES_MODELDATA_NAME=pio_model,PIO_STORAGE_REPOSITORIES_EVENTDATA_SOURCE=HBASE,PIO_CONF_DIR=/var/www/abc/apache-predictionio-0.10.0-incubating/PredictionIO-0.10.0-incubating/conf,PIO_STORAGE_SOURCES_ELASTICSEARCH_PORTS=9300,PIO_STORAGE_SOURCES_LOCALFS_TYPE=localfs
[WARN] [Utils] Your hostname, EK-LT-15 resolves to a loopback address:
127.0.1.1; using 192.168.10.8 instead (on interface wlp6s0)
[WARN] [Utils] Set SPARK_LOCAL_IP if you need to bind to another address
[INFO] [Remoting] Starting remoting
[INFO] [Remoting] Remoting started; listening on addresses :[akka.tcp://
sparkDriver@192.168.10.8:36713]
[WARN] [MetricsSystem] Using default name DAGScheduler for source because
spark.app.id is not set.
[ERROR] [RecoverableZooKeeper] ZooKeeper exists failed after 1 attempts
[ERROR] [ZooKeeperWatcher] hconnection-0x7bede4ea, quorum=localhost:2181,
baseZNode=/hbase Received unexpected KeeperException, re-throwing exception
[WARN] [ZooKeeperRegistry] Can't retrieve clusterId from Zookeeper
[ERROR] [StorageClient] Cannot connect to ZooKeeper (ZooKeeper ensemble:
localhost). Please make sure that the configuration is pointing at the
correct ZooKeeper ensemble. By default, HBase manages its own ZooKeeper, so
if you have not configured HBase to use an external ZooKeeper, that means
your HBase is not started or configured properly.
[ERROR] [Storage$] Error initializing storage client for source HBASE
Exception in thread "main"
org.apache.predictionio.data.storage.StorageClientException: Data source
HBASE was not properly initialized.
at
org.apache.predictionio.data.storage.Storage$$anonfun$10.apply(Storage.scala:282)
at
org.apache.predictionio.data.storage.Storage$$anonfun$10.apply(Storage.scala:282)
at scala.Option.getOrElse(Option.scala:120)
at
org.apache.predictionio.data.storage.Storage$.getDataObject(Storage.scala:281)
at
org.apache.predictionio.data.storage.Storage$.getPDataObject(Storage.scala:330)
at
org.apache.predictionio.data.storage.Storage$.getPDataObject(Storage.scala:273)
at
org.apache.predictionio.data.storage.Storage$.getPEvents(Storage.scala:394)
at
org.apache.predictionio.tools.imprt.FileToEvents$$anonfun$main$1.apply(FileToEvents.scala:98)
at
org.apache.predictionio.tools.imprt.FileToEvents$$anonfun$main$1.apply(FileToEvents.scala:68)
at scala.Option.map(Option.scala:145)
at

Re: Error while importing the event.json file.

2017-03-26 Thread Vaghawan Ojha
Sorry even pio status shows some errors now:

[ERROR] [RecoverableZooKeeper] ZooKeeper exists failed after 1 attempts
[ERROR] [ZooKeeperWatcher] hconnection-0x39ad12b6, quorum=localhost:2181,
baseZNode=/hbase Received unexpected KeeperException, re-throwing exception
[WARN] [ZooKeeperRegistry] Can't retrieve clusterId from Zookeeper
[ERROR] [StorageClient] Cannot connect to ZooKeeper (ZooKeeper ensemble:
localhost). Please make sure that the configuration is pointing at the
correct ZooKeeper ensemble. By default, HBase manages its own ZooKeeper, so
if you have not configured HBase to use an external ZooKeeper, that means
your HBase is not started or configured properly.
[ERROR] [Storage$] Error initializing storage client for source HBASE
[ERROR] [Console$] Unable to connect to all storage backends successfully.
The following shows the error message from the storage backend.
[ERROR] [Console$] Data source HBASE was not properly initialized.
(org.apache.predictionio.data.storage.StorageClientException)
[ERROR] [Console$] Dumping configuration of initialized storage backend
sources. Please make sure they are correct.
[ERROR] [Console$] Source Name: ELASTICSEARCH; Type: elasticsearch;
Configuration: HOME ->
/var/www/abc/apache-predictionio-0.10.0-incubating/PredictionIO-0.10.0-incubating/vendors/elasticsearch-1.4.4,
HOSTS -> localhost, PORTS -> 9300, CLUSTERNAME -> elasticsearch, TYPE ->
elasticsearch
[ERROR] [Console$] Source Name: LOCALFS; Type: localfs; Configuration: PATH
-> /home/ekbana-php/.pio_store/models, TYPE -> localfs
[ERROR] [Console$] Source Name: HBASE; Type: (error); Configuration: (error)


How would I actually solve it?

On Sun, Mar 26, 2017 at 7:13 PM, Vaghawan Ojha <vaghawan...@gmail.com>
wrote:

> While importing the event data, I get the following error, I am new and I
> don't even understand  what does they means? While doing pio status it's
> shows everything ok. What's wrong here.
>
> sudo PredictionIO-0.10.0-incubating/bin/pio import --appid 1 --input
> my_events.json
> [INFO] [Runner$] Submission command: /var/www/abc/apache-
> predictionio-0.10.0-incubating/PredictionIO-0.10.
> 0-incubating/vendors/spark-1.5.1-bin-hadoop2.6/bin/spark-submit --class
> org.apache.predictionio.tools.imprt.FileToEvents --files
> file:/var/www/abc/apache-predictionio-0.10.0-incubating/PredictionIO-0.10.
> 0-incubating/conf/log4j.properties,file:/var/www/abc/
> apache-predictionio-0.10.0-incubating/PredictionIO-0.10.
> 0-incubating/vendors/hbase-1.0.0/conf/hbase-site.xml --driver-class-path
> /var/www/abc/apache-predictionio-0.10.0-incubating/PredictionIO-0.10.
> 0-incubating/conf:/var/www/abc/apache-predictionio-0.10.
> 0-incubating/PredictionIO-0.10.0-incubating/vendors/
> elasticsearch-1.4.4/conf:/var/www/abc/apache-predictionio-0.
> 10.0-incubating/PredictionIO-0.10.0-incubating/lib/
> postgresql-9.4-1204.jdbc41.jar:/var/www/abc/apache-predictionio-0.10.0-
> incubating/PredictionIO-0.10.0-incubating/lib/mysql-
> connector-java-5.1.37.jar:/var/www/abc/apache-predictionio-0.10.0-
> incubating/PredictionIO-0.10.0-incubating/vendors/spark-1.
> 5.1-bin-hadoop2.6/conf:/var/www/abc/apache-predictionio-0.
> 10.0-incubating/PredictionIO-0.10.0-incubating/vendors/hbase-1.0.0/conf
> file:/var/www/abc/apache-predictionio-0.10.0-incubating/PredictionIO-0.10.
> 0-incubating/lib/pio-assembly-0.10.0-incubating.jar --appid 1 --input
> file:/var/www/abc/apache-predictionio-0.10.0-incubating/my_events.json
> --env PIO_STORAGE_SOURCES_HBASE_TYPE=hbase,PIO_ENV_LOADED=1,
> PIO_STORAGE_REPOSITORIES_METADATA_NAME=pio_meta,PIO_FS_
> BASEDIR=/home/ekbana-php/.pio_store,PIO_STORAGE_SOURCES_
> ELASTICSEARCH_HOSTS=localhost,PIO_STORAGE_SOURCES_HBASE_
> HOME=/var/www/abc/apache-predictionio-0.10.0-incubating/PredictionIO-0.10.
> 0-incubating/vendors/hbase-1.0.0,PIO_HOME=/var/www/abc/
> apache-predictionio-0.10.0-incubating/PredictionIO-0.10.
> 0-incubating,PIO_FS_ENGINESDIR=/home/ekbana-php/.
> pio_store/engines,PIO_STORAGE_SOURCES_LOCALFS_PATH=/home/
> ekbana-php/.pio_store/models,PIO_STORAGE_SOURCES_ELASTICSEARCH_TYPE=
> elasticsearch,PIO_STORAGE_REPOSITORIES_METADATA_SOURCE=
> ELASTICSEARCH,PIO_STORAGE_REPOSITORIES_MODELDATA_SOURCE=
> LOCALFS,PIO_STORAGE_REPOSITORIES_EVENTDATA_NAME=
> pio_event,PIO_STORAGE_SOURCES_ELASTICSEARCH_CLUSTERNAME=
> elasticsearch,PIO_STORAGE_SOURCES_ELASTICSEARCH_HOME=/var/www/abc/apache-
> predictionio-0.10.0-incubating/PredictionIO-0.10.0-incubating/vendors/
> elasticsearch-1.4.4,PIO_FS_TMPDIR=/home/ekbana-php/.pio_
> store/tmp,PIO_STORAGE_REPOSITORIES_MODELDATA_NAME=pio_model,PIO_STORAGE_
> REPOSITORIES_EVENTDATA_SOURCE=HBASE,PIO_CONF_DIR=/var/www/
> abc/apache-predictionio-0.10.0-incubating/PredictionIO-0.
> 10.0-incubating/conf,PIO_STORAGE_SOURCES_ELASTICSEARCH_
> PORTS=9300,PIO_STORAGE_SOURCES_LOCALFS_TYPE=localfs
> 

Using Recommendation Engine without product rating.

2017-03-26 Thread Vaghawan Ojha
Hi,

I'm seeing this template engine (
http://predictionio.incubator.apache.org/templates/recommendation/quickstart/
)

But it seems that it needs rating parameter by default, is it possible to
just to use user_id and item_id as the parameters.

Normally I would have to load the data in json, something like this:

[
{
bill_no: "CM/00038646/16",
bill_date: "2016-12-01 07:19:45",
icode: "BM55423",
customer_code: "00041077"
},
{
bill_no: "CM/00038646/16",
bill_date: "2016-12-01 07:19:45",
icode: "BM49665",
customer_code: "00041077"
},
{
bill_no: "CM/00038646/16",
bill_date: "2016-12-01 07:19:45",
icode: "BM60574",
customer_code: "00041077"
},
{
bill_no: "CM/00038647/16",
bill_date: "2016-12-01 07:21:26",
icode: "BM67206",
customer_code: "00041272"
},
]

all the events here are buy, so there is no other event  like rating or
view. Is it possible in this template engine. I had seen Universal
recommender to be my choice, but it seems it would need huge resources for
me. But this time I'm just working around various templates so that to
decide the very suitable one.

Thank you


Need Help

2017-03-26 Thread Vaghawan Ojha
HI,

I'm setting up prediction.io with elastic search and hbase, but there is a
option in pio-env.sh for postgresql, can I use hbase for this section?

# Default is to use PostgreSQL
PIO_STORAGE_REPOSITORIES_METADATA_NAME=pio_meta
PIO_STORAGE_REPOSITORIES_METADATA_SOURCE=PGSQL

PIO_STORAGE_REPOSITORIES_EVENTDATA_NAME=pio_event
PIO_STORAGE_REPOSITORIES_EVENTDATA_SOURCE=PGSQL

PIO_STORAGE_REPOSITORIES_MODELDATA_NAME=pio_model
PIO_STORAGE_REPOSITORIES_MODELDATA_SOURCE=PGSQL

Can I replace the PGSQL with HBASE, or I do have to set up the PGSQL?

Any help would be great.

Thank you


Re: Need a Suggessations

2017-03-23 Thread Vaghawan Ojha
Hi Pat,

Thank you so much for giving me such a clear idea, it really did help me
alot. This is the very first time I'm touching the big data, I hope It
wouldn't be that bad.

I would set it up as you recommended, and will come to ask you if something
I need to know, which will be very often.

Thank You
Vaghawan

On Fri, Mar 24, 2017 at 3:23 AM, Pat Ferrel <p...@occamsmachete.com> wrote:

> Think of the recommender as a single app. It is scalable to whatever your
> data size via the services it is built on. We often see that using a
> recommender is people’s first experience with really big data. Other tools
> and services you use outside of it are fine because they do not deal with
> such large data. Recommenders force you so process every interaction that
> all your users have made over perhaps a year and do it often. There are few
> other apps that require this. Welcome to Big-Data.
>
> MySQL is fine to run your app as you no doubt know. The “model” built in a
> recommender is generally not human readable but in the case of the UR you
> can understand it with some experience. It lives in Elasticsearch while the
> user interactions live in HBase. The user events can be looked at but not
> sure why you’d want too, they are condensed snippets of server logs.
>
> In any case it may help to think of the model in Elasticsearch as a
> product catalog. It will define what items can be recommended and have an
> entry for each item with Machine Learning calculated attributes attached
> that indicate the type of user that prefers each item. But the model also
> contains item properties/attributes that you may want to include for
> business rules.
>
> The Recommender is easily accessed from you app through the input and
> query API. You can change attributes of items by sending special input
> events. Queries are defined that match the type of things recommenders with
> business rules do and the model can be seen through Elasticsearch APIs but
> it is discouraged to do any direct manipulation of these since their
> meaning or format may change with any update.
>
> Plan to use the PIO query API, it will respond in real-time, with latency
> on the order of 25ms, and multiple simultaneous connections/queries. There
> would be no reason to pull out data from the UR and put it in a database or
> you would loose the ability to react to user’s real-time behavior, which is
> used to make recommendations. Stick to the input/query APIs and feed data
> into the UR in real-time and you’ll get the most benefit.
>
>
> On Mar 23, 2017, at 12:25 PM, Vaghawan Ojha <vaghawan...@gmail.com> wrote:
>
> Hi Pat,
>
> Thank you very much.Yes I will be following actionml instruction since I'm
> going to use UR. I think I should rather direct myself to HBASE rather than
> expensing time  in setting up Mysql. Part of my need is that once we train
> the dataset, the result should be easily available to the application which
> are running into Mysql.
>
> I'm fairly new to the concept itself. So basically I would always have a
> larage json file coming from the application which uses mysql(this
> shouldn't be the problem). Then I would use PIO and UR to do the hard work,
> and get back the result either like an API which I think already works in
> PIO or saved somewhere in database like mysql or something like that.
>
> Thanks
>
> On Fri, Mar 24, 2017 at 1:03 AM, Pat Ferrel <p...@occamsmachete.com> wrote:
>
>> The UR uses Elasticsearch for part of the Recommender algorithm, therefor
>> it must be configured as a storage backend. It is possible to use Postgres
>> or MySQL for the other stores but we have very little experience with this.
>> HBase is indefinitely scalable so we always use that. Single machine
>> deployments are rare with a reasonably sized data so Elasticsearch + Hbase
>> running separately or in clusters will always meet the data needs. The RDBs
>> will not and anyway, like I said you have to use Elasticsearch.
>>
>> Therefore for the UR follow instructions on the ActionML site since they
>> are specific to the UR. For other templates you may use other
>> configurations of PIO but if you use the UR config you can also use every
>> template too.
>>
>>
>>
>> On Mar 23, 2017, at 9:07 AM, Vaghawan Ojha <vaghawan...@gmail.com> wrote:
>>
>> Hi, Thank you!
>>
>> I came into further more confusion here, actually I installed prediction
>> IO version 0.10.0 from here http://predictionio.incub
>> ator.apache.org/install/install-sourcecode/  and have been fighting to
>> configure mysql as a storage in my local linux machine.
>>
>> But I see there is a different documentation of installing in actionml
&

Re: Need a Suggessations

2017-03-23 Thread Vaghawan Ojha
Hi Pat,

Thank you very much.Yes I will be following actionml instruction since I'm
going to use UR. I think I should rather direct myself to HBASE rather than
expensing time  in setting up Mysql. Part of my need is that once we train
the dataset, the result should be easily available to the application which
are running into Mysql.

I'm fairly new to the concept itself. So basically I would always have a
larage json file coming from the application which uses mysql(this
shouldn't be the problem). Then I would use PIO and UR to do the hard work,
and get back the result either like an API which I think already works in
PIO or saved somewhere in database like mysql or something like that.

Thanks

On Fri, Mar 24, 2017 at 1:03 AM, Pat Ferrel <p...@occamsmachete.com> wrote:

> The UR uses Elasticsearch for part of the Recommender algorithm, therefor
> it must be configured as a storage backend. It is possible to use Postgres
> or MySQL for the other stores but we have very little experience with this.
> HBase is indefinitely scalable so we always use that. Single machine
> deployments are rare with a reasonably sized data so Elasticsearch + Hbase
> running separately or in clusters will always meet the data needs. The RDBs
> will not and anyway, like I said you have to use Elasticsearch.
>
> Therefore for the UR follow instructions on the ActionML site since they
> are specific to the UR. For other templates you may use other
> configurations of PIO but if you use the UR config you can also use every
> template too.
>
>
>
> On Mar 23, 2017, at 9:07 AM, Vaghawan Ojha <vaghawan...@gmail.com> wrote:
>
> Hi, Thank you!
>
> I came into further more confusion here, actually I installed prediction
> IO version 0.10.0 from here http://predictionio.
> incubator.apache.org/install/install-sourcecode/  and have been fighting
> to configure mysql as a storage in my local linux machine.
>
> But I see there is a different documentation of installing in actionml
> website, I'm not sure for which I would have to go. Currently there is no "
> pio-env.sh".  file inside conf folder however there is
> pio-env.sh.template file. I commented the pgsql section and uncommented the
> mysql section with the username and password, but whenever I do . sudo
> PredictionIO-0.10.0-incubating/bin/pio eventserver there seems to be an
> error that says that authentication failed with pgsql, however I don't want
> to use pgsql.
>
> # Storage Repositories
>
> # Default is to use PostgreSQL
> PIO_STORAGE_REPOSITORIES_METADATA_NAME=pio_meta
> PIO_STORAGE_REPOSITORIES_METADATA_SOURCE=PGSQL
>
> PIO_STORAGE_REPOSITORIES_EVENTDATA_NAME=pio_event
> PIO_STORAGE_REPOSITORIES_EVENTDATA_SOURCE=PGSQL
>
> PIO_STORAGE_REPOSITORIES_MODELDATA_NAME=pio_model
> PIO_STORAGE_REPOSITORIES_MODELDATA_SOURCE=PGSQL
>
> # Storage Data Sources
>
> # PostgreSQL Default Settings
> # Please change "pio" to your database name in
> PIO_STORAGE_SOURCES_PGSQL_URL
> # Please change PIO_STORAGE_SOURCES_PGSQL_USERNAME and
> # PIO_STORAGE_SOURCES_PGSQL_PASSWORD accordingly
> #PIO_STORAGE_SOURCES_PGSQL_TYPE=jdbc
> #PIO_STORAGE_SOURCES_PGSQL_URL=jdbc:postgresql://localhost/pio
> #PIO_STORAGE_SOURCES_PGSQL_USERNAME=pio
> #PIO_STORAGE_SOURCES_PGSQL_PASSWORD=pio
>
> # MySQL Example
>  PIO_STORAGE_SOURCES_MYSQL_TYPE=jdbc
>  PIO_STORAGE_SOURCES_MYSQL_URL=jdbc:mysql://localhost/pio
>  PIO_STORAGE_SOURCES_MYSQL_USERNAME=root
>  PIO_STORAGE_SOURCES_MYSQL_PASSWORD=root
>
>
> This is how the pio-env.sh.template looks like. And again when I visited
> the actionml site, it suggests that I do have to have ELASTICSEARCH. but
> prediction.io site doesn't tells us the same. Which one should I follow
> and where would I find the current working version of installation guide. I
> actually wanaa use prediction.io in my production shortly after I
> implemented in local.
>
> Please help me, thank you very much for your help, I appreciate it so much.
> Vaghawan
>
>
> On Thu, Mar 23, 2017 at 9:27 PM, Pat Ferrel <p...@occamsmachete.com> wrote:
>
>> Since PIO has moved to Apache, the namespace of PIO code changed and so
>> all templates need to be updated. None of the ones in
>> https://github.com/PredictionIO/
>> <https://github.com/PredictionIO/template-scala-parallel-universal-recommendation>
>>  will
>> work with Apache PIO. For the upgraded UR see: https://github.com/action
>> ml/universal-recommender Docs for the UR are here:
>> http://actionml.com/docs/ur
>>
>> Also look on the Template gallery page here for a description of template
>> status. Some have not been moved to the new namespace and converted to run
>> with PIO but this is pretty easy to do yourse

Re: Need a Suggessations

2017-03-23 Thread Vaghawan Ojha
I get the same error even if I deleted the pio-env.sh . I am not sure from
where else it throws that error. But there is yet anotehr file called
pio-env.sh.travis, which seems a bit different then pio-env.sh.templat.

May be it's coming from there? I don't know but the code inside bin shows
it exactly reads the pio-env.sh.

On Fri, Mar 24, 2017 at 12:29 AM, Vaghawan Ojha <vaghawan...@gmail.com>
wrote:

> I think something else is wrong  that this, otherwise I should get a
> different error, I get the same error even if I deleted the whole
> pio-env.sh. This is weird.
>
> I'm supposed to get this error when there was no pio-env.sh file
>
>   else
> echo -e "\033[0;35mWarning: pio-env.sh was not found in
> ${use_conf_dir}. Using system environment variables instead.\033[0m\n"
>
>
> Thanks
>
> On Fri, Mar 24, 2017 at 12:18 AM, Vaghawan Ojha <vaghawan...@gmail.com>
> wrote:
>
>> Am I supposed to do something after I changed the pio-env.sh ? Because
>> it's not working at any rate. I don't know what's so wrong. The error was
>> there even when there was no pio-env.sh file. And it still is the same.
>>
>> ionIO-0.10.0-incubating/bin/pio status
>> [INFO] [Console$] Inspecting PredictionIO...
>> [INFO] [Console$] PredictionIO 0.10.0-incubating is installed at
>> /var/www/apache-predictionio-0.10.0-incubating/PredictionIO-
>> 0.10.0-incubating
>> [INFO] [Console$] Inspecting Apache Spark...
>> [INFO] [Console$] Apache Spark is installed at
>> /var/www/apache-predictionio-0.10.0-incubating/PredictionIO-
>> 0.10.0-incubating/vendors/spark-1.5.1-bin-hadoop2.6
>> [INFO] [Console$] Apache Spark 1.5.1 detected (meets minimum requirement
>> of 1.3.0)
>> [INFO] [Console$] Inspecting storage backend connections...
>> [INFO] [Storage$] Verifying Meta Data Backend (Source: PGSQL)...
>> [ERROR] [Console$] Unable to connect to all storage backends
>> successfully. The following shows the error message from the storage
>> backend.
>> [ERROR] [Console$] FATAL: password authentication failed for user "pio"
>> (org.postgresql.util.PSQLException)
>> [ERROR] [Console$] Dumping configuration of initialized storage backend
>> sources. Please make sure they are correct.
>> [ERROR] [Console$] Source Name: PGSQL; Type: jdbc; Configuration: URL ->
>> jdbc:postgresql://localhost/pio, PASSWORD -> pio, TYPE -> jdbc, USERNAME
>> -> pio
>>
>> Sorry!
>>
>> On Fri, Mar 24, 2017 at 12:10 AM, Marius Rabenarivo <
>> mariusrabenar...@gmail.com> wrote:
>>
>>> See below an example
>>>
>>> *.*
>>> It seems you copied the content twice in fact
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>> *#!/usr/bin/env bash## Copy this file as pio-env.sh and edit it for your
>>> site's configuration.## Licensed to the Apache Software Foundation (ASF)
>>> under one or more# contributor license agreements.  See the NOTICE file
>>> distributed with# this work for additional information regarding copyright
>>> ownership.# The ASF licenses this file to You under the Apache License,
>>> Version 2.0# (the "License"); you may not use this file except in
>>> compliance with# the License.  You may obtain a copy of the License at##
>>> http://www.apache.org/licenses/LICENSE-2.0
>>> <http://www.apache.org/licenses/LICENSE-2.0>## Unless required by
>>> applicable law or agreed to in writing, software# distributed und

Re: Need a Suggessations

2017-03-23 Thread Vaghawan Ojha
I got it, actually I should be into another pio dir rather than the parent
one. The sub dir inside the pio installation seems to be active, so there
was already a file named pio-env.sh and that was working. Now I made
changes and it is reflecting.

Thank you very much.

On Fri, Mar 24, 2017 at 12:42 AM, Vaghawan Ojha <vaghawan...@gmail.com>
wrote:

> The problem here isn't the pio-env.sh. Because I get the same thing even
> if I renamed the pio-env.sh to whatever else. It doesn't even matter even
> if I deleted the file, because I am getting the same message. There should
> have been some file else where this should have been defined.
>
> On Fri, Mar 24, 2017 at 12:36 AM, Vaghawan Ojha <vaghawan...@gmail.com>
> wrote:
>
>> I get the same error even if I deleted the pio-env.sh . I am not sure
>> from where else it throws that error. But there is yet anotehr file called
>> pio-env.sh.travis, which seems a bit different then pio-env.sh.templat.
>>
>> May be it's coming from there? I don't know but the code inside bin shows
>> it exactly reads the pio-env.sh.
>>
>> On Fri, Mar 24, 2017 at 12:29 AM, Vaghawan Ojha <vaghawan...@gmail.com>
>> wrote:
>>
>>> I think something else is wrong  that this, otherwise I should get a
>>> different error, I get the same error even if I deleted the whole
>>> pio-env.sh. This is weird.
>>>
>>> I'm supposed to get this error when there was no pio-env.sh file
>>>
>>>   else
>>> echo -e "\033[0;35mWarning: pio-env.sh was not found in
>>> ${use_conf_dir}. Using system environment variables instead.\033[0m\n"
>>>
>>>
>>> Thanks
>>>
>>> On Fri, Mar 24, 2017 at 12:18 AM, Vaghawan Ojha <vaghawan...@gmail.com>
>>> wrote:
>>>
>>>> Am I supposed to do something after I changed the pio-env.sh ? Because
>>>> it's not working at any rate. I don't know what's so wrong. The error was
>>>> there even when there was no pio-env.sh file. And it still is the same.
>>>>
>>>> ionIO-0.10.0-incubating/bin/pio status
>>>> [INFO] [Console$] Inspecting PredictionIO...
>>>> [INFO] [Console$] PredictionIO 0.10.0-incubating is installed at
>>>> /var/www/apache-predictionio-0.10.0-incubating/PredictionIO-
>>>> 0.10.0-incubating
>>>> [INFO] [Console$] Inspecting Apache Spark...
>>>> [INFO] [Console$] Apache Spark is installed at
>>>> /var/www/apache-predictionio-0.10.0-incubating/PredictionIO-
>>>> 0.10.0-incubating/vendors/spark-1.5.1-bin-hadoop2.6
>>>> [INFO] [Console$] Apache Spark 1.5.1 detected (meets minimum
>>>> requirement of 1.3.0)
>>>> [INFO] [Console$] Inspecting storage backend connections...
>>>> [INFO] [Storage$] Verifying Meta Data Backend (Source: PGSQL)...
>>>> [ERROR] [Console$] Unable to connect to all storage backends
>>>> successfully. The following shows the error message from the storage
>>>> backend.
>>>> [ERROR] [Console$] FATAL: password authentication failed for user "pio"
>>>> (org.postgresql.util.PSQLException)
>>>> [ERROR] [Console$] Dumping configuration of initialized storage backend
>>>> sources. Please make sure they are correct.
>>>> [ERROR] [Console$] Source Name: PGSQL; Type: jdbc; Configuration: URL
>>>> -> jdbc:postgresql://localhost/pio, PASSWORD -> pio, TYPE -> jdbc,
>>>> USERNAME -> pio
>>>>
>>>> Sorry!
>>>>
>>>> On Fri, Mar 24, 2017 at 12:10 AM, Marius Rabenarivo <
>>>> mariusrabenar...@gmail.com> wrote:
>>>>
>>>>> See below an example
>>>>>
>>>>> *.*
>>>>> It seems you copied the content twice in fact
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
&

Re: Need a Suggessations

2017-03-23 Thread Vaghawan Ojha
 Please change "pio" to
> your database name in PIO_STORAGE_SOURCES_PGSQL_URL# Please change
> PIO_STORAGE_SOURCES_PGSQL_USERNAME and# PIO_STORAGE_SOURCES_PGSQL_PASSWORD
> accordingly#PIO_STORAGE_SOURCES_PGSQL_TYPE=jdbc#PIO_STORAGE_SOURCES_PGSQL_URL=jdbc:postgresql://localhost/pio#PIO_STORAGE_SOURCES_PGSQL_USERNAME=pio#PIO_STORAGE_SOURCES_PGSQL_PASSWORD=pio#
> MySQL
> Example PIO_STORAGE_SOURCES_MYSQL_TYPE=jdbc 
> PIO_STORAGE_SOURCES_MYSQL_URL=jdbc:mysql://localhost/pio 
> PIO_STORAGE_SOURCES_MYSQL_USERNAME=pio PIO_STORAGE_SOURCES_MYSQL_PASSWORD=pio#
> Elasticsearch Example#
> PIO_STORAGE_SOURCES_ELASTICSEARCH_TYPE=elasticsearch#
> PIO_STORAGE_SOURCES_ELASTICSEARCH_CLUSTERNAME=#
> PIO_STORAGE_SOURCES_ELASTICSEARCH_HOSTS=localhost#
> PIO_STORAGE_SOURCES_ELASTICSEARCH_PORTS=9300#
> PIO_STORAGE_SOURCES_ELASTICSEARCH_HOME=$PIO_HOME/vendors/elasticsearch-1.4.4#
> Local File System
> ExamplePIO_STORAGE_SOURCES_LOCALFS_TYPE=localfsPIO_STORAGE_SOURCES_LOCALFS_PATH=$PIO_FS_BASEDIR/models#
> HBase Example# PIO_STORAGE_SOURCES_HBASE_TYPE=hbase#
> PIO_STORAGE_SOURCES_HBASE_HOME=$PIO_HOME/vendors/hbase-1.0.0*
>
> 2017-03-23 22:06 GMT+04:00 Ambuj Sharma <am...@getamplify.com>:
>
>> check your pio-env.sh. Every configuration is copied twice
>>
>>
>> Thanks and Regards
>> Ambuj Sharma
>> Sunrise may late, But Morning is sure.
>> Team ML
>> Betaout
>>
>> On Thu, Mar 23, 2017 at 11:26 PM, Vaghawan Ojha <vaghawan...@gmail.com>
>> wrote:
>>
>>> Still got the same unfortunately:
>>>
>>> this is pio status
>>>
>>> ionIO-0.10.0-incubating/bin/pio status
>>> [INFO] [Console$] Inspecting PredictionIO...
>>> [INFO] [Console$] PredictionIO 0.10.0-incubating is installed at
>>> /var/www/apache-predictionio-0.10.0-incubating/PredictionIO-
>>> 0.10.0-incubating
>>> [INFO] [Console$] Inspecting Apache Spark...
>>> [INFO] [Console$] Apache Spark is installed at
>>> /var/www/apache-predictionio-0.10.0-incubating/PredictionIO-
>>> 0.10.0-incubating/vendors/spark-1.5.1-bin-hadoop2.6
>>> [INFO] [Console$] Apache Spark 1.5.1 detected (meets minimum requirement
>>> of 1.3.0)
>>> [INFO] [Console$] Inspecting storage backend connections...
>>> [INFO] [Storage$] Verifying Meta Data Backend (Source: PGSQL)...
>>> [ERROR] [Console$] Unable to connect to all storage backends
>>> successfully. The following shows the error message from the storage
>>> backend.
>>> [ERROR] [Console$] FATAL: password authentication failed for user "pio"
>>> (org.postgresql.util.PSQLException)
>>> [ERROR] [Console$] Dumping configuration of initialized storage backend
>>> sources. Please make sure they are correct.
>>> [ERROR] [Console$] Source Name: PGSQL; Type: jdbc; Configuration: URL ->
>>> jdbc:postgresql://localhost/pio, PASSWORD -> pio, TYPE -> jdbc,
>>> USERNAME -> pio
>>>
>>> and I've attacted the pio-env.sh. Any help?
>>>
>>> On Thu, Mar 23, 2017 at 11:37 PM, Ambuj Sharma <am...@getamplify.com>
>>> wrote:
>>>
>>>> copy pio-env.sh.template to pio-env.sh.
>>>>
>>>>
>>>> Thanks and Regards
>>>> Ambuj Sharma
>>>> Sunrise may late, But Morning is sure.
>>>> Team ML
>>>> Betaout
>>>>
>>>> On Thu, Mar 23, 2017 at 11:13 PM, Vaghawan Ojha <vaghawan...@gmail.com>
>>>> wrote:
>>>>
>>>>> There is no file called pio-env.sh however there is
>>>>> pio-env.sh.template. Are they the same?
>>>>>
>>>>> On Thu, Mar 23, 2017 at 11:11 PM, Marius Rabenarivo <
>>>>> mariusrabenar...@gmail.com> wrote:
>>>>>
>>>>>> # Default is to use PostgreSQL
>>>>>> PIO_STORAGE_REPOSITORIES_METADATA_NAME=pio_meta
>>>>>> PIO_STORAGE_REPOSITORIES_METADATA_SOURCE=MYSQL
>>>>>>
>>>>>> PIO_STORAGE_REPOSITORIES_EVENTDATA_NAME=pio_event
>>>>>> PIO_STORAGE_REPOSITORIES_EVENTDATA_SOURCE=MYSQL
>>>>>>
>>>>>> PIO_STORAGE_REPOSITORIES_MODELDATA_NAME=pio_model
>>>>>> PIO_STORAGE_REPOSITORIES_MODELDATA_SOURCE=MYSQL
>>>>>>
>>>>>>
>>>>>> 2017-03-23 21:24 GMT+04:00 Marius Rabenarivo <
>>>>>> mariusrabenar...@gmail.com>:
>>>>>>
>>>>>>> You have to change the config in pio-env.sh to use MYSQL instead of
>>>>>>> PGSQL
>

Re: Need a Suggessations

2017-03-23 Thread Vaghawan Ojha
The problem here isn't the pio-env.sh. Because I get the same thing even if
I renamed the pio-env.sh to whatever else. It doesn't even matter even if I
deleted the file, because I am getting the same message. There should have
been some file else where this should have been defined.

On Fri, Mar 24, 2017 at 12:36 AM, Vaghawan Ojha <vaghawan...@gmail.com>
wrote:

> I get the same error even if I deleted the pio-env.sh . I am not sure from
> where else it throws that error. But there is yet anotehr file called
> pio-env.sh.travis, which seems a bit different then pio-env.sh.templat.
>
> May be it's coming from there? I don't know but the code inside bin shows
> it exactly reads the pio-env.sh.
>
> On Fri, Mar 24, 2017 at 12:29 AM, Vaghawan Ojha <vaghawan...@gmail.com>
> wrote:
>
>> I think something else is wrong  that this, otherwise I should get a
>> different error, I get the same error even if I deleted the whole
>> pio-env.sh. This is weird.
>>
>> I'm supposed to get this error when there was no pio-env.sh file
>>
>>   else
>> echo -e "\033[0;35mWarning: pio-env.sh was not found in
>> ${use_conf_dir}. Using system environment variables instead.\033[0m\n"
>>
>>
>> Thanks
>>
>> On Fri, Mar 24, 2017 at 12:18 AM, Vaghawan Ojha <vaghawan...@gmail.com>
>> wrote:
>>
>>> Am I supposed to do something after I changed the pio-env.sh ? Because
>>> it's not working at any rate. I don't know what's so wrong. The error was
>>> there even when there was no pio-env.sh file. And it still is the same.
>>>
>>> ionIO-0.10.0-incubating/bin/pio status
>>> [INFO] [Console$] Inspecting PredictionIO...
>>> [INFO] [Console$] PredictionIO 0.10.0-incubating is installed at
>>> /var/www/apache-predictionio-0.10.0-incubating/PredictionIO-
>>> 0.10.0-incubating
>>> [INFO] [Console$] Inspecting Apache Spark...
>>> [INFO] [Console$] Apache Spark is installed at
>>> /var/www/apache-predictionio-0.10.0-incubating/PredictionIO-
>>> 0.10.0-incubating/vendors/spark-1.5.1-bin-hadoop2.6
>>> [INFO] [Console$] Apache Spark 1.5.1 detected (meets minimum requirement
>>> of 1.3.0)
>>> [INFO] [Console$] Inspecting storage backend connections...
>>> [INFO] [Storage$] Verifying Meta Data Backend (Source: PGSQL)...
>>> [ERROR] [Console$] Unable to connect to all storage backends
>>> successfully. The following shows the error message from the storage
>>> backend.
>>> [ERROR] [Console$] FATAL: password authentication failed for user "pio"
>>> (org.postgresql.util.PSQLException)
>>> [ERROR] [Console$] Dumping configuration of initialized storage backend
>>> sources. Please make sure they are correct.
>>> [ERROR] [Console$] Source Name: PGSQL; Type: jdbc; Configuration: URL ->
>>> jdbc:postgresql://localhost/pio, PASSWORD -> pio, TYPE -> jdbc,
>>> USERNAME -> pio
>>>
>>> Sorry!
>>>
>>> On Fri, Mar 24, 2017 at 12:10 AM, Marius Rabenarivo <
>>> mariusrabenar...@gmail.com> wrote:
>>>
>>>> See below an example
>>>>
>>>> *.*
>>>> It seems you copied the content twice in fact
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>

Re: Need a Suggessations

2017-03-23 Thread Vaghawan Ojha
Oh ya, I fixed but with the same thing in my hand it still search for
PGSQL.

On Thu, Mar 23, 2017 at 11:51 PM, Ambuj Sharma <am...@getamplify.com> wrote:

> check your pio-env.sh. Every configuration is copied twice
>
>
> Thanks and Regards
> Ambuj Sharma
> Sunrise may late, But Morning is sure.
> Team ML
> Betaout
>
> On Thu, Mar 23, 2017 at 11:26 PM, Vaghawan Ojha <vaghawan...@gmail.com>
> wrote:
>
>> Still got the same unfortunately:
>>
>> this is pio status
>>
>> ionIO-0.10.0-incubating/bin/pio status
>> [INFO] [Console$] Inspecting PredictionIO...
>> [INFO] [Console$] PredictionIO 0.10.0-incubating is installed at
>> /var/www/apache-predictionio-0.10.0-incubating/PredictionIO-
>> 0.10.0-incubating
>> [INFO] [Console$] Inspecting Apache Spark...
>> [INFO] [Console$] Apache Spark is installed at
>> /var/www/apache-predictionio-0.10.0-incubating/PredictionIO-
>> 0.10.0-incubating/vendors/spark-1.5.1-bin-hadoop2.6
>> [INFO] [Console$] Apache Spark 1.5.1 detected (meets minimum requirement
>> of 1.3.0)
>> [INFO] [Console$] Inspecting storage backend connections...
>> [INFO] [Storage$] Verifying Meta Data Backend (Source: PGSQL)...
>> [ERROR] [Console$] Unable to connect to all storage backends
>> successfully. The following shows the error message from the storage
>> backend.
>> [ERROR] [Console$] FATAL: password authentication failed for user "pio"
>> (org.postgresql.util.PSQLException)
>> [ERROR] [Console$] Dumping configuration of initialized storage backend
>> sources. Please make sure they are correct.
>> [ERROR] [Console$] Source Name: PGSQL; Type: jdbc; Configuration: URL ->
>> jdbc:postgresql://localhost/pio, PASSWORD -> pio, TYPE -> jdbc, USERNAME
>> -> pio
>>
>> and I've attacted the pio-env.sh. Any help?
>>
>> On Thu, Mar 23, 2017 at 11:37 PM, Ambuj Sharma <am...@getamplify.com>
>> wrote:
>>
>>> copy pio-env.sh.template to pio-env.sh.
>>>
>>>
>>> Thanks and Regards
>>> Ambuj Sharma
>>> Sunrise may late, But Morning is sure.
>>> Team ML
>>> Betaout
>>>
>>> On Thu, Mar 23, 2017 at 11:13 PM, Vaghawan Ojha <vaghawan...@gmail.com>
>>> wrote:
>>>
>>>> There is no file called pio-env.sh however there is
>>>> pio-env.sh.template. Are they the same?
>>>>
>>>> On Thu, Mar 23, 2017 at 11:11 PM, Marius Rabenarivo <
>>>> mariusrabenar...@gmail.com> wrote:
>>>>
>>>>> # Default is to use PostgreSQL
>>>>> PIO_STORAGE_REPOSITORIES_METADATA_NAME=pio_meta
>>>>> PIO_STORAGE_REPOSITORIES_METADATA_SOURCE=MYSQL
>>>>>
>>>>> PIO_STORAGE_REPOSITORIES_EVENTDATA_NAME=pio_event
>>>>> PIO_STORAGE_REPOSITORIES_EVENTDATA_SOURCE=MYSQL
>>>>>
>>>>> PIO_STORAGE_REPOSITORIES_MODELDATA_NAME=pio_model
>>>>> PIO_STORAGE_REPOSITORIES_MODELDATA_SOURCE=MYSQL
>>>>>
>>>>>
>>>>> 2017-03-23 21:24 GMT+04:00 Marius Rabenarivo <
>>>>> mariusrabenar...@gmail.com>:
>>>>>
>>>>>> You have to change the config in pio-env.sh to use MYSQL instead of
>>>>>> PGSQL
>>>>>>
>>>>>> In the pasted response above I can read Verifying Meta Data Backend
>>>>>> (Source: PGSQL)...
>>>>>>
>>>>>> 2017-03-23 21:19 GMT+04:00 Vaghawan Ojha <vaghawan...@gmail.com>:
>>>>>>
>>>>>>> Yes following is the response of pio status
>>>>>>>
>>>>>>> sudo PredictionIO-0.10.0-incubating/bin/pio status
>>>>>>> [INFO] [Console$] Inspecting PredictionIO...
>>>>>>> [INFO] [Console$] PredictionIO 0.10.0-incubating is installed at
>>>>>>> /home/ekbana-php/Downloads/apache-predictionio-0.10.0-incuba
>>>>>>> ting/PredictionIO-0.10.0-incubating
>>>>>>> [INFO] [Console$] Inspecting Apache Spark...
>>>>>>> [INFO] [Console$] Apache Spark is installed at
>>>>>>> /home/ekbana-php/Downloads/apache-predictionio-0.10.0-incuba
>>>>>>> ting/PredictionIO-0.10.0-incubating/vendors/spark-1.5.1-bin-
>>>>>>> hadoop2.6
>>>>>>> [INFO] [Console$] Apache Spark 1.5.1 detected (meets minimum
>>>>>>> requirement of 1.3.0)
>>>>>>> [INFO] [Console$] In

Re: Need a Suggessations

2017-03-23 Thread Vaghawan Ojha
I think something else is wrong  that this, otherwise I should get a
different error, I get the same error even if I deleted the whole
pio-env.sh. This is weird.

I'm supposed to get this error when there was no pio-env.sh file

  else
echo -e "\033[0;35mWarning: pio-env.sh was not found in
${use_conf_dir}. Using system environment variables instead.\033[0m\n"


Thanks

On Fri, Mar 24, 2017 at 12:18 AM, Vaghawan Ojha <vaghawan...@gmail.com>
wrote:

> Am I supposed to do something after I changed the pio-env.sh ? Because
> it's not working at any rate. I don't know what's so wrong. The error was
> there even when there was no pio-env.sh file. And it still is the same.
>
> ionIO-0.10.0-incubating/bin/pio status
> [INFO] [Console$] Inspecting PredictionIO...
> [INFO] [Console$] PredictionIO 0.10.0-incubating is installed at
> /var/www/apache-predictionio-0.10.0-incubating/
> PredictionIO-0.10.0-incubating
> [INFO] [Console$] Inspecting Apache Spark...
> [INFO] [Console$] Apache Spark is installed at
> /var/www/apache-predictionio-0.10.0-incubating/PredictionIO-0.10.0-
> incubating/vendors/spark-1.5.1-bin-hadoop2.6
> [INFO] [Console$] Apache Spark 1.5.1 detected (meets minimum requirement
> of 1.3.0)
> [INFO] [Console$] Inspecting storage backend connections...
> [INFO] [Storage$] Verifying Meta Data Backend (Source: PGSQL)...
> [ERROR] [Console$] Unable to connect to all storage backends successfully.
> The following shows the error message from the storage backend.
> [ERROR] [Console$] FATAL: password authentication failed for user "pio"
> (org.postgresql.util.PSQLException)
> [ERROR] [Console$] Dumping configuration of initialized storage backend
> sources. Please make sure they are correct.
> [ERROR] [Console$] Source Name: PGSQL; Type: jdbc; Configuration: URL ->
> jdbc:postgresql://localhost/pio, PASSWORD -> pio, TYPE -> jdbc, USERNAME
> -> pio
>
> Sorry!
>
> On Fri, Mar 24, 2017 at 12:10 AM, Marius Rabenarivo <
> mariusrabenar...@gmail.com> wrote:
>
>> See below an example
>>
>> *.*
>> It seems you copied the content twice in fact
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> *#!/usr/bin/env bash## Copy this file as pio-env.sh and edit it for your
>> site's configuration.## Licensed to the Apache Software Foundation (ASF)
>> under one or more# contributor license agreements.  See the NOTICE file
>> distributed with# this work for additional information regarding copyright
>> ownership.# The ASF licenses this file to You under the Apache License,
>> Version 2.0# (the "License"); you may not use this file except in
>> compliance with# the License.  You may obtain a copy of the License at##
>> http://www.apache.org/licenses/LICENSE-2.0
>> <http://www.apache.org/licenses/LICENSE-2.0>## Unless required by
>> applicable law or agreed to in writing, software# distributed under the
>> License is distributed on an "AS IS" BASIS,# WITHOUT WARRANTIES OR
>> CONDITIONS OF ANY KIND, either express or implied.# See the License for the
>> specific language governing permissions and# limitations under the
>> License.## PredictionIO Main Configuration## This section controls core
>> behavior of PredictionIO. It is very likely that# you need to change these
>> to fit your site.# SPARK_HOME: Apache Spark is a hard dependency and must
>> be
>> configured.SPARK_HOME=$PIO_HOME/vendors/spark-1.5.1-bin-hadoop2.6POSTGRES_JDBC_DRIVER=$PIO_HOME/lib/postgresql-9.4-1204.jdbc41.jarMYSQL_JDBC_DRIVER=$PIO_HOME/lib/mysql-connector-java-5.1.37.jar#
>> ES_CONF_DIR: You must configure this if you have advanced configuration
>> for#  your Elasticsearch setup.#
>> ES_CONF_DIR=/opt/elasticsearch# HADOOP_CONF_DIR: You must configure this if
>> you intend to run PredictionIO#  with Hadoop 2.#
>> HADOOP_CONF_DIR=/opt/had

Re: Need a Suggessations

2017-03-23 Thread Vaghawan Ojha
Still got the same unfortunately:

this is pio status

ionIO-0.10.0-incubating/bin/pio status
[INFO] [Console$] Inspecting PredictionIO...
[INFO] [Console$] PredictionIO 0.10.0-incubating is installed at
/var/www/apache-predictionio-0.10.0-incubating/PredictionIO-0.10.0-incubating
[INFO] [Console$] Inspecting Apache Spark...
[INFO] [Console$] Apache Spark is installed at
/var/www/apache-predictionio-0.10.0-incubating/PredictionIO-0.10.0-incubating/vendors/spark-1.5.1-bin-hadoop2.6
[INFO] [Console$] Apache Spark 1.5.1 detected (meets minimum requirement of
1.3.0)
[INFO] [Console$] Inspecting storage backend connections...
[INFO] [Storage$] Verifying Meta Data Backend (Source: PGSQL)...
[ERROR] [Console$] Unable to connect to all storage backends successfully.
The following shows the error message from the storage backend.
[ERROR] [Console$] FATAL: password authentication failed for user "pio"
(org.postgresql.util.PSQLException)
[ERROR] [Console$] Dumping configuration of initialized storage backend
sources. Please make sure they are correct.
[ERROR] [Console$] Source Name: PGSQL; Type: jdbc; Configuration: URL ->
jdbc:postgresql://localhost/pio, PASSWORD -> pio, TYPE -> jdbc, USERNAME ->
pio

and I've attacted the pio-env.sh. Any help?

On Thu, Mar 23, 2017 at 11:37 PM, Ambuj Sharma <am...@getamplify.com> wrote:

> copy pio-env.sh.template to pio-env.sh.
>
>
> Thanks and Regards
> Ambuj Sharma
> Sunrise may late, But Morning is sure.
> Team ML
> Betaout
>
> On Thu, Mar 23, 2017 at 11:13 PM, Vaghawan Ojha <vaghawan...@gmail.com>
> wrote:
>
>> There is no file called pio-env.sh however there is pio-env.sh.template.
>> Are they the same?
>>
>> On Thu, Mar 23, 2017 at 11:11 PM, Marius Rabenarivo <
>> mariusrabenar...@gmail.com> wrote:
>>
>>> # Default is to use PostgreSQL
>>> PIO_STORAGE_REPOSITORIES_METADATA_NAME=pio_meta
>>> PIO_STORAGE_REPOSITORIES_METADATA_SOURCE=MYSQL
>>>
>>> PIO_STORAGE_REPOSITORIES_EVENTDATA_NAME=pio_event
>>> PIO_STORAGE_REPOSITORIES_EVENTDATA_SOURCE=MYSQL
>>>
>>> PIO_STORAGE_REPOSITORIES_MODELDATA_NAME=pio_model
>>> PIO_STORAGE_REPOSITORIES_MODELDATA_SOURCE=MYSQL
>>>
>>>
>>> 2017-03-23 21:24 GMT+04:00 Marius Rabenarivo <mariusrabenar...@gmail.com
>>> >:
>>>
>>>> You have to change the config in pio-env.sh to use MYSQL instead of
>>>> PGSQL
>>>>
>>>> In the pasted response above I can read Verifying Meta Data Backend
>>>> (Source: PGSQL)...
>>>>
>>>> 2017-03-23 21:19 GMT+04:00 Vaghawan Ojha <vaghawan...@gmail.com>:
>>>>
>>>>> Yes following is the response of pio status
>>>>>
>>>>> sudo PredictionIO-0.10.0-incubating/bin/pio status
>>>>> [INFO] [Console$] Inspecting PredictionIO...
>>>>> [INFO] [Console$] PredictionIO 0.10.0-incubating is installed at
>>>>> /home/ekbana-php/Downloads/apache-predictionio-0.10.0-incuba
>>>>> ting/PredictionIO-0.10.0-incubating
>>>>> [INFO] [Console$] Inspecting Apache Spark...
>>>>> [INFO] [Console$] Apache Spark is installed at
>>>>> /home/ekbana-php/Downloads/apache-predictionio-0.10.0-incuba
>>>>> ting/PredictionIO-0.10.0-incubating/vendors/spark-1.5.1-bin-hadoop2.6
>>>>> [INFO] [Console$] Apache Spark 1.5.1 detected (meets minimum
>>>>> requirement of 1.3.0)
>>>>> [INFO] [Console$] Inspecting storage backend connections...
>>>>> [INFO] [Storage$] Verifying Meta Data Backend (Source: PGSQL)...
>>>>> [ERROR] [Console$] Unable to connect to all storage backends
>>>>> successfully. The following shows the error message from the storage
>>>>> backend.
>>>>> [ERROR] [Console$] FATAL: password authentication failed for user
>>>>> "pio" (org.postgresql.util.PSQLException)
>>>>> [ERROR] [Console$] Dumping configuration of initialized storage
>>>>> backend sources. Please make sure they are correct.
>>>>> [ERROR] [Console$] Source Name: PGSQL; Type: jdbc; Configuration: URL
>>>>> -> jdbc:postgresql://localhost/pio, PASSWORD -> pio, TYPE -> jdbc,
>>>>> USERNAME -> pio
>>>>>
>>>>> Which says that postgresql authentication failed. Ok I'm trying to go
>>>>> for postgresql this time instead of mysql, I don't wanna lose much time
>>>>> fighting with this two sql. I can use any of them, it just need to work.
>>>>>
>>>>> Thanks
>>>&g

Re: Need a Suggessations

2017-03-23 Thread Vaghawan Ojha
Yes that's what I wanted to do, but couldn't come out to configure it. No
matter how much I change in the pio-env.sh.template, when I start the event
server it always pops up with the posgresql error of authentication.

I even tried with posgresql and it shows the same things. Party something
to do with my dumbness I guess.

Thank you very much Marius.

On Thu, Mar 23, 2017 at 10:39 PM, Marius Rabenarivo <
mariusrabenar...@gmail.com> wrote:

> You can use MySQL as event store but you need ElasticSearch to store the
> model metadata and index used by UR Engine.
>
> 2017-03-23 20:25 GMT+04:00 Vaghawan Ojha <vaghawan...@gmail.com>:
>
>> But I wanted to use mysql as default. Should I switch it? I can do that,
>> but I was just wondering since the documentation says that I could use
>> mysql.
>>
>> On Thu, Mar 23, 2017 at 10:03 PM, Marius Rabenarivo <
>> mariusrabenar...@gmail.com> wrote:
>>
>>> I think pio-start-all start PostgreSQL by default
>>>
>>> 2017-03-23 20:16 GMT+04:00 Vaghawan Ojha <vaghawan...@gmail.com>:
>>>
>>>> Hi,
>>>>
>>>> Yes I did that but still I get the same output, it's weird.
>>>>
>>>> Thanks
>>>>
>>>> On Thu, Mar 23, 2017 at 10:00 PM, Marius Rabenarivo <
>>>> mariusrabenar...@gmail.com> wrote:
>>>>
>>>>> You have to change this section
>>>>>
>>>>> # Default is to use PostgreSQL
>>>>> PIO_STORAGE_REPOSITORIES_METADATA_NAME=pio_meta
>>>>> PIO_STORAGE_REPOSITORIES_METADATA_SOURCE=PGSQL
>>>>>
>>>>> PIO_STORAGE_REPOSITORIES_EVENTDATA_NAME=pio_event
>>>>> PIO_STORAGE_REPOSITORIES_EVENTDATA_SOURCE=PGSQL
>>>>>
>>>>> PIO_STORAGE_REPOSITORIES_MODELDATA_NAME=pio_model
>>>>> PIO_STORAGE_REPOSITORIES_MODELDATA_SOURCE=PGSQL
>>>>>
>>>>> Put MYSQL in place of PGSQL
>>>>>
>>>>> 2017-03-23 20:07 GMT+04:00 Vaghawan Ojha <vaghawan...@gmail.com>:
>>>>>
>>>>>> Hi, Thank you!
>>>>>>
>>>>>> I came into further more confusion here, actually I installed
>>>>>> prediction IO version 0.10.0 from here http://predictionio.incub
>>>>>> ator.apache.org/install/install-sourcecode/  and have been fighting
>>>>>> to configure mysql as a storage in my local linux machine.
>>>>>>
>>>>>> But I see there is a different documentation of installing in
>>>>>> actionml website, I'm not sure for which I would have to go. Currently
>>>>>> there is no "pio-env.sh".  file inside conf folder however there is
>>>>>> pio-env.sh.template file. I commented the pgsql section and uncommented 
>>>>>> the
>>>>>> mysql section with the username and password, but whenever I do . sudo
>>>>>> PredictionIO-0.10.0-incubating/bin/pio eventserver there seems to be
>>>>>> an error that says that authentication failed with pgsql, however I don't
>>>>>> want to use pgsql.
>>>>>>
>>>>>> # Storage Repositories
>>>>>>
>>>>>> # Default is to use PostgreSQL
>>>>>> PIO_STORAGE_REPOSITORIES_METADATA_NAME=pio_meta
>>>>>> PIO_STORAGE_REPOSITORIES_METADATA_SOURCE=PGSQL
>>>>>>
>>>>>> PIO_STORAGE_REPOSITORIES_EVENTDATA_NAME=pio_event
>>>>>> PIO_STORAGE_REPOSITORIES_EVENTDATA_SOURCE=PGSQL
>>>>>>
>>>>>> PIO_STORAGE_REPOSITORIES_MODELDATA_NAME=pio_model
>>>>>> PIO_STORAGE_REPOSITORIES_MODELDATA_SOURCE=PGSQL
>>>>>>
>>>>>> # Storage Data Sources
>>>>>>
>>>>>> # PostgreSQL Default Settings
>>>>>> # Please change "pio" to your database name in
>>>>>> PIO_STORAGE_SOURCES_PGSQL_URL
>>>>>> # Please change PIO_STORAGE_SOURCES_PGSQL_USERNAME and
>>>>>> # PIO_STORAGE_SOURCES_PGSQL_PASSWORD accordingly
>>>>>> #PIO_STORAGE_SOURCES_PGSQL_TYPE=jdbc
>>>>>> #PIO_STORAGE_SOURCES_PGSQL_URL=jdbc:postgresql://localhost/pio
>>>>>> #PIO_STORAGE_SOURCES_PGSQL_USERNAME=pio
>>>>>> #PIO_STORAGE_SOURCES_PGSQL_PASSWORD=pio
>>>>>>
>>>>>> # MySQL Example
>>>>>>  PIO_STORAGE_SOURCES_MYSQL_TYPE=jdbc
>>>>>>  PIO_STORAGE_SO

Re: Need a Suggessations

2017-03-23 Thread Vaghawan Ojha
But I wanted to use mysql as default. Should I switch it? I can do that,
but I was just wondering since the documentation says that I could use
mysql.

On Thu, Mar 23, 2017 at 10:03 PM, Marius Rabenarivo <
mariusrabenar...@gmail.com> wrote:

> I think pio-start-all start PostgreSQL by default
>
> 2017-03-23 20:16 GMT+04:00 Vaghawan Ojha <vaghawan...@gmail.com>:
>
>> Hi,
>>
>> Yes I did that but still I get the same output, it's weird.
>>
>> Thanks
>>
>> On Thu, Mar 23, 2017 at 10:00 PM, Marius Rabenarivo <
>> mariusrabenar...@gmail.com> wrote:
>>
>>> You have to change this section
>>>
>>> # Default is to use PostgreSQL
>>> PIO_STORAGE_REPOSITORIES_METADATA_NAME=pio_meta
>>> PIO_STORAGE_REPOSITORIES_METADATA_SOURCE=PGSQL
>>>
>>> PIO_STORAGE_REPOSITORIES_EVENTDATA_NAME=pio_event
>>> PIO_STORAGE_REPOSITORIES_EVENTDATA_SOURCE=PGSQL
>>>
>>> PIO_STORAGE_REPOSITORIES_MODELDATA_NAME=pio_model
>>> PIO_STORAGE_REPOSITORIES_MODELDATA_SOURCE=PGSQL
>>>
>>> Put MYSQL in place of PGSQL
>>>
>>> 2017-03-23 20:07 GMT+04:00 Vaghawan Ojha <vaghawan...@gmail.com>:
>>>
>>>> Hi, Thank you!
>>>>
>>>> I came into further more confusion here, actually I installed
>>>> prediction IO version 0.10.0 from here http://predictionio.incub
>>>> ator.apache.org/install/install-sourcecode/  and have been fighting to
>>>> configure mysql as a storage in my local linux machine.
>>>>
>>>> But I see there is a different documentation of installing in actionml
>>>> website, I'm not sure for which I would have to go. Currently there is no "
>>>> pio-env.sh".  file inside conf folder however there is
>>>> pio-env.sh.template file. I commented the pgsql section and uncommented the
>>>> mysql section with the username and password, but whenever I do . sudo
>>>> PredictionIO-0.10.0-incubating/bin/pio eventserver there seems to be
>>>> an error that says that authentication failed with pgsql, however I don't
>>>> want to use pgsql.
>>>>
>>>> # Storage Repositories
>>>>
>>>> # Default is to use PostgreSQL
>>>> PIO_STORAGE_REPOSITORIES_METADATA_NAME=pio_meta
>>>> PIO_STORAGE_REPOSITORIES_METADATA_SOURCE=PGSQL
>>>>
>>>> PIO_STORAGE_REPOSITORIES_EVENTDATA_NAME=pio_event
>>>> PIO_STORAGE_REPOSITORIES_EVENTDATA_SOURCE=PGSQL
>>>>
>>>> PIO_STORAGE_REPOSITORIES_MODELDATA_NAME=pio_model
>>>> PIO_STORAGE_REPOSITORIES_MODELDATA_SOURCE=PGSQL
>>>>
>>>> # Storage Data Sources
>>>>
>>>> # PostgreSQL Default Settings
>>>> # Please change "pio" to your database name in
>>>> PIO_STORAGE_SOURCES_PGSQL_URL
>>>> # Please change PIO_STORAGE_SOURCES_PGSQL_USERNAME and
>>>> # PIO_STORAGE_SOURCES_PGSQL_PASSWORD accordingly
>>>> #PIO_STORAGE_SOURCES_PGSQL_TYPE=jdbc
>>>> #PIO_STORAGE_SOURCES_PGSQL_URL=jdbc:postgresql://localhost/pio
>>>> #PIO_STORAGE_SOURCES_PGSQL_USERNAME=pio
>>>> #PIO_STORAGE_SOURCES_PGSQL_PASSWORD=pio
>>>>
>>>> # MySQL Example
>>>>  PIO_STORAGE_SOURCES_MYSQL_TYPE=jdbc
>>>>  PIO_STORAGE_SOURCES_MYSQL_URL=jdbc:mysql://localhost/pio
>>>>  PIO_STORAGE_SOURCES_MYSQL_USERNAME=root
>>>>  PIO_STORAGE_SOURCES_MYSQL_PASSWORD=root
>>>>
>>>>
>>>> This is how the pio-env.sh.template looks like. And again when I
>>>> visited the actionml site, it suggests that I do have to have
>>>> ELASTICSEARCH. but prediction.io site doesn't tells us the same. Which
>>>> one should I follow and where would I find the current working version of
>>>> installation guide. I actually wanaa use prediction.io in my
>>>> production shortly after I implemented in local.
>>>>
>>>> Please help me, thank you very much for your help, I appreciate it so
>>>> much.
>>>> Vaghawan
>>>>
>>>>
>>>> On Thu, Mar 23, 2017 at 9:27 PM, Pat Ferrel <p...@occamsmachete.com>
>>>> wrote:
>>>>
>>>>> Since PIO has moved to Apache, the namespace of PIO code changed and
>>>>> so all templates need to be updated. None of the ones in
>>>>> https://github.com/PredictionIO/
>>>>> <https://github.com/PredictionIO/template-scala-parallel-universal-recommendation>
>>&

Re: Need a Suggessations

2017-03-23 Thread Vaghawan Ojha
Ok thanks Marius. I would configure that way then. Back to the problem, I
would need something like this:

First item based similarity from the consumption history and user based
similarity so that we could combine both to produce the best result
possible. Probably something like this after the product similarity:

   -

   Choose a product and check if the user already consumed that product.
   -

   Get the similarities of the product’s top (X) neighbours.
   -

   Get the consumption histories of the user of the top X neighbors.
   - Calculate the score, and sort them with the highest score first.


Thank You

On Thu, Mar 23, 2017 at 10:06 PM, Marius Rabenarivo <
mariusrabenar...@gmail.com> wrote:

> Pat, and what if I want items similar to user's taste in the
> recommendation without providing an item?
>
> Any time line for the release of UR 0.6?
>
> 2017-03-23 20:18 GMT+04:00 Marius Rabenarivo <mariusrabenar...@gmail.com>:
>
>> I think pio-start-all start PostgreSQL by default
>>
>> 2017-03-23 20:16 GMT+04:00 Vaghawan Ojha <vaghawan...@gmail.com>:
>>
>>> Hi,
>>>
>>> Yes I did that but still I get the same output, it's weird.
>>>
>>> Thanks
>>>
>>> On Thu, Mar 23, 2017 at 10:00 PM, Marius Rabenarivo <
>>> mariusrabenar...@gmail.com> wrote:
>>>
>>>> You have to change this section
>>>>
>>>> # Default is to use PostgreSQL
>>>> PIO_STORAGE_REPOSITORIES_METADATA_NAME=pio_meta
>>>> PIO_STORAGE_REPOSITORIES_METADATA_SOURCE=PGSQL
>>>>
>>>> PIO_STORAGE_REPOSITORIES_EVENTDATA_NAME=pio_event
>>>> PIO_STORAGE_REPOSITORIES_EVENTDATA_SOURCE=PGSQL
>>>>
>>>> PIO_STORAGE_REPOSITORIES_MODELDATA_NAME=pio_model
>>>> PIO_STORAGE_REPOSITORIES_MODELDATA_SOURCE=PGSQL
>>>>
>>>> Put MYSQL in place of PGSQL
>>>>
>>>> 2017-03-23 20:07 GMT+04:00 Vaghawan Ojha <vaghawan...@gmail.com>:
>>>>
>>>>> Hi, Thank you!
>>>>>
>>>>> I came into further more confusion here, actually I installed
>>>>> prediction IO version 0.10.0 from here http://predictionio.incub
>>>>> ator.apache.org/install/install-sourcecode/  and have been fighting
>>>>> to configure mysql as a storage in my local linux machine.
>>>>>
>>>>> But I see there is a different documentation of installing in actionml
>>>>> website, I'm not sure for which I would have to go. Currently there is no 
>>>>> "
>>>>> pio-env.sh".  file inside conf folder however there is
>>>>> pio-env.sh.template file. I commented the pgsql section and uncommented 
>>>>> the
>>>>> mysql section with the username and password, but whenever I do . sudo
>>>>> PredictionIO-0.10.0-incubating/bin/pio eventserver there seems to be
>>>>> an error that says that authentication failed with pgsql, however I don't
>>>>> want to use pgsql.
>>>>>
>>>>> # Storage Repositories
>>>>>
>>>>> # Default is to use PostgreSQL
>>>>> PIO_STORAGE_REPOSITORIES_METADATA_NAME=pio_meta
>>>>> PIO_STORAGE_REPOSITORIES_METADATA_SOURCE=PGSQL
>>>>>
>>>>> PIO_STORAGE_REPOSITORIES_EVENTDATA_NAME=pio_event
>>>>> PIO_STORAGE_REPOSITORIES_EVENTDATA_SOURCE=PGSQL
>>>>>
>>>>> PIO_STORAGE_REPOSITORIES_MODELDATA_NAME=pio_model
>>>>> PIO_STORAGE_REPOSITORIES_MODELDATA_SOURCE=PGSQL
>>>>>
>>>>> # Storage Data Sources
>>>>>
>>>>> # PostgreSQL Default Settings
>>>>> # Please change "pio" to your database name in
>>>>> PIO_STORAGE_SOURCES_PGSQL_URL
>>>>> # Please change PIO_STORAGE_SOURCES_PGSQL_USERNAME and
>>>>> # PIO_STORAGE_SOURCES_PGSQL_PASSWORD accordingly
>>>>> #PIO_STORAGE_SOURCES_PGSQL_TYPE=jdbc
>>>>> #PIO_STORAGE_SOURCES_PGSQL_URL=jdbc:postgresql://localhost/pio
>>>>> #PIO_STORAGE_SOURCES_PGSQL_USERNAME=pio
>>>>> #PIO_STORAGE_SOURCES_PGSQL_PASSWORD=pio
>>>>>
>>>>> # MySQL Example
>>>>>  PIO_STORAGE_SOURCES_MYSQL_TYPE=jdbc
>>>>>  PIO_STORAGE_SOURCES_MYSQL_URL=jdbc:mysql://localhost/pio
>>>>>  PIO_STORAGE_SOURCES_MYSQL_USERNAME=root
>>>>>  PIO_STORAGE_SOURCES_MYSQL_PASSWORD=root
>>>>>
>>>>>
>>>>> This is how the pio-env.sh.template loo