SSL is supported on the Event and PredictionServers but someone else will have 
to answer how. There is a Jira to add instructions to the site, not sure if 
that has been cleared but you might want to check and vote for the issue. 
https://issues.apache.org/jira/browse/PIO-7?jql=project%20%3D%20PIO 
<https://issues.apache.org/jira/browse/PIO-7?jql=project%20=%20PIO>

The key can be auto-generated or can be specified and is really only an ID for 
the dataset to send events into. It is not used for queries.


On Oct 21, 2016, at 9:37 AM, Georg Heiler <[email protected]> wrote:

Thanks a lot for this great answer. 

May I add an additional question regarding the api :
I know pio generates an api key. For which operations is this key required and 
is it possible to use encryption and a key with the api in oder to sort of 
force authentication in order to obtain a predicted result?


Cheers 
Georg 
Pat Ferrel <[email protected] <mailto:[email protected]>> schrieb am 
Fr. 21. Okt. 2016 um 18:17:
The command line for any pio command that is launched on Spark can specify the 
master so you can train on one cluster and deploy on another. This is typical 
when using the ALS recommenders, which use a big cluster to train but deploy 
with `pio deploy -- --master local[2]` which would use a local context to load 
and serve the model. Beware of memory use, wherever the pio command is run will 
also run the Spark driver, which can have large memory needs, as large as the 
executors, which run on the cluster. If you run 2 contexts on the same machine, 
one with a local master and one with a cluster master you will have 2 drivers 
and may have executors also.

Yarn allows you to run the driver on a cluster machine but is somewhat 
complicated to setup.



On Oct 21, 2016, at 4:53 AM, Georg Heiler <[email protected] 
<mailto:[email protected]>> wrote:

Hi,
I am curious if prediction.IO supports different environments e.g. is it 
possible to define a separate spark context for training and serving of the 
model in engine.json?

The idea is that a trained model e.g. xgboost could be evaluated very quickly 
outside of a cluster environment (no yarn, ... involved, only prediction.io 
<http://prediction.io/> in docker with a database + model in file system)

Cheers,
Georg


Reply via email to