Repository: incubator-unomi
Updated Branches:
  refs/heads/master bf3cd3cca -> ddbebd25f


Prepare site update and documentation for upcoming 1.2 release.

Signed-off-by: Serge Huber <[email protected]>


Project: http://git-wip-us.apache.org/repos/asf/incubator-unomi/repo
Commit: http://git-wip-us.apache.org/repos/asf/incubator-unomi/commit/ddbebd25
Tree: http://git-wip-us.apache.org/repos/asf/incubator-unomi/tree/ddbebd25
Diff: http://git-wip-us.apache.org/repos/asf/incubator-unomi/diff/ddbebd25

Branch: refs/heads/master
Commit: ddbebd25fb81b381f008fb3f8ccd1cad871aa4a4
Parents: bf3cd3c
Author: Serge Huber <[email protected]>
Authored: Wed Jan 11 15:45:22 2017 +0100
Committer: Serge Huber <[email protected]>
Committed: Wed Jan 11 15:45:22 2017 +0100

----------------------------------------------------------------------
 BUILDING                                        |   3 +-
 README.md                                       | 395 +------------------
 ...g.apache.unomi.persistence.elasticsearch.cfg |   4 +
 pom.xml                                         |   6 +-
 src/site/markdown/building-and-deploying.md     |  90 +++--
 src/site/markdown/clustering.md                 |  15 +-
 src/site/markdown/configuration.md              |  28 ++
 src/site/markdown/index.md                      |   4 +-
 8 files changed, 116 insertions(+), 429 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/incubator-unomi/blob/ddbebd25/BUILDING
----------------------------------------------------------------------
diff --git a/BUILDING b/BUILDING
index 6b95733..dddb2d8 100644
--- a/BUILDING
+++ b/BUILDING
@@ -21,9 +21,8 @@ Building Apache Unomi
 Initial Setup
 -------------
 
-1) Install J2SE 7.0 SDK (or later), which can be downloaded from
+1) Install J2SE 8.0 SDK (or later), which can be downloaded from
    http://www.oracle.com/technetwork/java/javase/downloads/index.html
-   Use version of "JDK 7.0 Update 55" (or later).
 
 2) Make sure that your JAVA_HOME environment variable is set to the newly 
installed
    JDK location, and that your PATH includes %JAVA_HOME%\bin (windows) or

http://git-wip-us.apache.org/repos/asf/incubator-unomi/blob/ddbebd25/README.md
----------------------------------------------------------------------
diff --git a/README.md b/README.md
index fd6e1d7..4965e0f 100644
--- a/README.md
+++ b/README.md
@@ -18,400 +18,15 @@
 Apache Unomi
 ============
 
-An Apache project that implements the Context Server specification
+Apache Unomi stores user profile information and is mostly used to provide a 
backend server for A/B testing and 
+personalization. To do so it implements the currently under development OASIS 
Context Server specification. 
 
 License
 -------
 The source code is available under the Apache License V2
 
-Building from source code
--------------------------
-
-See the instructions in the BUILDING file provided in the source code package
-
-Deploying the generated binary package
---------------------------------------
-
-The "package" sub-project generates a pre-configured Apache Karaf installation 
that is the simplest way to get started.
-Simply uncompress the package/target/unomi-VERSION.tar.gz (for Linux or Mac OS 
X) or
- package/target/unomi-VERSION.zip (for Windows) archive into the directory of 
your choice.
- 
-You can then start the server simply by using the command on UNIX/Linux/MacOS 
X : 
-
-    ./bin/karaf start    
-    
-or on Windows shell : 
-
-    bin\karaf.bat start
-    
-
-Deploying into an existing Karaf server
----------------------------------------
-
-This is only needed if you didn't use the generated package. Also, this is the 
preferred way to install a development
-environment if you intend to re-deploy the context server KAR iteratively.
-
-Additional requirements:
-* Apache Karaf 3.0.2+, http://karaf.apache.org
-* Local copy of the Elasticsearch ZIP package, available here : 
http://www.elasticsearch.org
-
-1. Before deploying, make sure that you have Apache Karaf properly installed. 
You will also have to increase the
-default maximum memory size and perm gen size by adjusting the following 
environment values in the bin/setenv(.bat)
-files (at the end of the file):
-
-    ```
-       MY_DIRNAME=`dirname $0`
-       MY_KARAF_HOME=`cd "$MY_DIRNAME/.."; pwd`
-       export KARAF_OPTS="-Djava.library.path=$MY_KARAF_HOME/lib/sigar"
-       export JAVA_MAX_MEM=3G
-       export JAVA_MAX_PERM_MEM=384M
-    ```
-    
-2. You will also need to have the Hyperic Sigar native libraries in your Karaf 
installation, so in order to this
-go to the Elasticsearch website (http://www.elasticsearch.org)  and download 
the ZIP package. Decompress it somewhere 
-on your disk and copy all the files from the lib/sigar directory into Karaf's 
lib/sigar directory 
-(must be created first) EXCEPT THE SIGAR.JAR file.
-
-3. Install the WAR support, CXF and CDI (OpenWebBeans) into Karaf by doing the 
following in the Karaf command line:
-
-    ```
-       feature:repo-add cxf 3.0.2
-       feature:repo-add cellar 3.0.3
-       feature:repo-add mvn:org.apache.unomi/unomi-kar/VERSION/xml/features
-       feature:install unomi-kar
-    ```
-
-4. Create a new $MY_KARAF_HOME/etc/org.apache.cxf.osgi.cfg file and put the 
following property inside :
-
-    ```
-       org.apache.cxf.servlet.context=/cxs
-    ```
-   
-5. If all went smoothly, you should be able to access the context script here 
: http://localhost:8181/cxs/cluster .
- You should be able to login with karaf / karaf and see basic server 
information. If not something went wrong during the install.
-
-Configuration
-----------------------------------
-
-Before starting your Unomi server, you need to properly configure it by 
editing configuration files.
-All unomi configuration files can be found in the $MY_KARAF_HOME/etc 
directory, named org.apache.unomi.*.cfg .
-
-The $MY_KARAF_HOME/etc/org.apache.unomi.web.cfg file defines the addresses and 
port where Unomi can be found :
-
-    contextserver.address=localhost
-    contextserver.port=8181
-    contextserver.secureAddress=localhost
-    contextserver.securePort=9443
-
-Addresses must be updated with reachable addresses, as they will be sent to 
the client by the cluster end point.
-
-If you need to specify an Elasticsearch cluster name that is different than 
the default, it is recommended to do this
-BEFORE you start the server for the first time, or you will loose all the data 
you have stored previously.
-
-To change the cluster name, first create a file called
-
-    $MY_KARAF_HOME/etc/org.apache.unomi.persistence.elasticsearch.cfg
-
-with the following contents:
-
-    cluster.name=contextElasticSearch
-    index.name=context
-
-And replace the cluster.name parameter here by your cluster name.
-    
-Secured events configuration
----------------------------
-
-If you need to secure some events, that will be sent only by a trusted third 
party server, you can update the file :
-
-    $MY_KARAF_HOME/etc/org.apache.unomi.thirdparty.cfg
-
-Ususally, login events, which operate on profiles and do merge on protected 
properties, must be secured. For each
-trusted third party server, you need to add these 3 lines :
-
-thirdparty.provider1.key=secret-key
-thirdparty.provider1.ipAddresses=127.0.0.1,::1
-thirdparty.provider1.allowedEvents=login,download
-
-The events set in allowedEvents will be secured and will only be accepted if 
the call comes from the specified IP
-address, and if the secret-key is passed in the X-Unomi-Peer header.
-
-Installing the MaxMind GeoIPLite2 IP lookup database
-----------------------------------------------------
-
-The Context Server requires an IP database in order to resolve IP addresses to 
user location.
-The GeoLite2 database can be downloaded from MaxMind here :
-http://dev.maxmind.com/geoip/geoip2/geolite2/
-
-Simply download the GeoLite2-City.mmdb file into the "etc" directory.
-
-Installing Geonames database
-----------------------------
-
-Context server includes a geocoding service based on the geonames database ( 
http://www.geonames.org/ ). It can be
-used to create conditions on countries or cities.
-
-In order to use it, you need to install the Geonames database into . Get the 
"allCountries.zip" database from here :
-http://download.geonames.org/export/dump/
-
-Download it and put it in the "etc" directory, without unzipping it.
-Edit $MY_KARAF_HOME/etc/org.apache.unomi.geonames.cfg and set 
request.geonamesDatabase.forceImport to true, import should start right away.
-Otherwise, import should start at the next startup. Import runs in background, 
but can take about 15 minutes.
-At the end, you should have about 4 million entries in the geonames index.
- 
-REST API Security
------------------
-
-The Context Server REST API is protected using JAAS authentication and using 
Basic or Digest HTTP auth.
-By default, the login/password for the REST API full administrative access is 
"karaf/karaf".
-
-The generated package is also configured with a default SSL certificate. You 
can change it by following these steps :
-
-1. Replace the existing keystore in $MY_KARAF_HOME/etc/keystore by your own 
certificate :
- 
-    http://wiki.eclipse.org/Jetty/Howto/Configure_SSL
-    
-2. Update the keystore and certificate password in 
$MY_KARAF_HOME/etc/custom.properties file :
- 
-```
-    org.osgi.service.http.secure.enabled = true
-    org.ops4j.pax.web.ssl.keystore=${karaf.etc}/keystore
-    org.ops4j.pax.web.ssl.password=changeme
-    org.ops4j.pax.web.ssl.keypassword=changeme
-    org.osgi.service.http.port.secure=9443
-```
-
-You should now have SSL setup on Karaf with your certificate, and you can test 
it by trying to access it on port 9443.
-
- 
-Running the integration tests
------------------------------
-
-The integration tests are not executed by default to make build time minimal, 
but it is recommended to run the 
-integration tests at least once before using the server to make sure that 
everything is ok in the build. Another way
-to use these tests is to run them from a continuous integration server such as 
Jenkins, Apache Gump, Atlassian Bamboo or
- others. 
- 
-Note : the integration tests require a JDK 7 or more recent !
-
-To run the tests simply activate the following profile : 
- 
-    mvn -P integration-tests clean install
-
-Running the performance tests
------------------------------
-
-Performance tests are based on Gatling. You need to have a running context 
server or cluster of servers before
-executing the tests.
-
-Test parameteres are editable in the 
performance-tests/src/test/scala/unomi/Parameters.scala file. baseUrls should
-contains the URLs of all your cluster nodes
-
-Run the test by using the gatling.conf file in 
performance-tests/src/test/resources :
-
-```
-    export GATLING_CONF=<path>/performance-tests/src/test/resources
-    gatling.sh
-```
-
-Reports are generated in performance-tests/target/results.
-
-
-Testing with an example page
-----------------------------
-
-A default test page is provided at the following URL:
-
-```
-   http://localhost:8181/index.html
-```
-
-This test page will trigger the loading of the /context.js script, which will 
try to retrieving the user context
-or create a new one if it doesn't exist yet. It also contains an experimental 
integration with Facebook Login, but it
-doesn't yet save the context back to the context server.
-
-Integrating onto a page
------------------------
-
- Simply reference the context script in your HTML as in the following example:
-
-```javascript
-<script type="text/javascript">
-    (function(){ var u=(("https:" == document.location.protocol) ? 
"https://localhost:8181/"; : "http://localhost:8181/";);
-    var d=document, g=d.createElement('script'), 
s=d.getElementsByTagName('script')[0]; g.type='text/javascript'; g.defer=true; 
g.async=true; g.src=u+'context.js';
-    s.parentNode.insertBefore(g,s); })();
-</script>
-```
-
-Cluster setup
+Documentation
 -------------
 
-Context server relies on Elasticsearch to discover and configure its cluster. 
You just need to install multiple context
-servers on the same network, and enable the discovery protocol in 
$MY_KARAF_HOME/etc/org.apache.unomi.persistence.elasticsearch.cfg file :
-
-    discovery.zen.ping.multicast.enabled=true
-
-All nodes on the same network, sharing the same cluster name will be part of 
the same cluster.
-
-#### 2 nodes  configuration
-One node dedicated to context server, 1 node for elasticsearch storage.
-
-Node A :
-
-    numberOfReplicas=0
-    monthlyIndex.numberOfReplicas=0
-
-Node B :
-
-    numberOfReplicas=0
-    monthlyIndex.numberOfReplicas=0
-
-#### 3 nodes configuration
-One node dedicated to context server, 2 nodes for elasticsearch storage with 
fault-tolerance
-
-Node A :
-
-    numberOfReplicas=1
-    monthlyIndex.numberOfReplicas=1
-
-Node B :
-
-    numberOfReplicas=1
-    monthlyIndex.numberOfReplicas=1
-
-Node C :
-
-    numberOfReplicas=1
-    monthlyIndex.numberOfReplicas=1
-
-
-JDK Selection on Mac OS X
--------------------------
-
-You might need to select the JDK to run the tests in the itests subproject. In 
order to do so you can list the 
-installed JDKs with the following command : 
-
-    /usr/libexec/java_home -V
-    
-which will output something like this : 
-
-    Matching Java Virtual Machines (7):
-        1.7.0_51, x86_64:      "Java SE 7"     
/Library/Java/JavaVirtualMachines/jdk1.7.0_51.jdk/Contents/Home
-        1.7.0_45, x86_64:      "Java SE 7"     
/Library/Java/JavaVirtualMachines/jdk1.7.0_45.jdk/Contents/Home
-        1.7.0_25, x86_64:      "Java SE 7"     
/Library/Java/JavaVirtualMachines/jdk1.7.0_25.jdk/Contents/Home
-        1.6.0_65-b14-462, x86_64:      "Java SE 6"     
/Library/Java/JavaVirtualMachines/1.6.0_65-b14-462.jdk/Contents/Home
-        1.6.0_65-b14-462, i386:        "Java SE 6"     
/Library/Java/JavaVirtualMachines/1.6.0_65-b14-462.jdk/Contents/Home
-        1.6.0_65-b14-462, x86_64:      "Java SE 6"     
/System/Library/Java/JavaVirtualMachines/1.6.0.jdk/Contents/Home
-        1.6.0_65-b14-462, i386:        "Java SE 6"     
/System/Library/Java/JavaVirtualMachines/1.6.0.jdk/Contents/Home
-
-
-You can then select the one you want using : 
-
-    export JAVA_HOME=`/usr/libexec/java_home -v 1.7.0_51`
-    
-and then check that it was correctly referenced using: 
-
-    java -version
-    
-which should give you a result such as this: 
-
-    java version "1.7.0_51"
-    Java(TM) SE Runtime Environment (build 1.7.0_51-b13)
-    Java HotSpot(TM) 64-Bit Server VM (build 24.51-b03, mixed mode)
-        
-Automatic profile merging
--------------------------
-
-The context server is capable of merging profiles based on a common property 
value. In order to use this, you must
-add the MergeProfileOnPropertyAction to a rule (such as a login rule for 
example), and configure it with the name
- of the property that will be used to identify the profiles to be merged. An 
example could be the "email" property,
- meaning that if two (or more) profiles are found to have the same value for 
the "email" property they will be merged
- by this action.
- 
-Upon merge, the old profiles are marked with a "mergedWith" property that will 
be used on next profile access to delete
-the original profile and replace it with the merged profile (aka "master" 
profile). Once this is done, all cookie tracking
-will use the merged profile.
-
-To test, simply configure the action in the "login" or "facebookLogin" rules 
and set it up on the "email" property. 
-Upon sending one of the events, all matching profiles will be merged.
-
-Securing a production environment
----------------------------------
-
-Before going live with a project, you should *absolutely* read the following 
section that will help you setup a proper 
-secure environment for running your context server.         
-
-Step 1: Install and configure a firewall 
-
-You should setup a firewall around your cluster of context servers and/or 
Elasticsearch nodes. If you have an 
-application-level firewall you should only allow the following connections 
open to the whole world : 
-
- - http://localhost:8181/context.js
- - http://localhost:8181/eventcollector
-
-All other ports should not be accessible to the world.
-
-For your Context Server client applications (such as the Jahia CMS), you will 
need to make the following ports 
-accessible : 
-
-    8181 (Context Server HTTP port) 
-    9443 (Context Server HTTPS port)
-    
-The context server actually requires HTTP Basic Auth for access to the Context 
Server administration REST API, so it is
-highly recommended that you design your client applications to use the HTTPS 
port for accessing the REST API.
-
-The user accounts to access the REST API are actually routed through Karaf's 
JAAS support, which you may find the
-documentation for here : 
-
- - http://karaf.apache.org/manual/latest/users-guide/security.html
-    
-The default username/password is 
-
-    karaf/karaf
-    
-You should really change this default username/password as soon as possible. 
To do so, simply modify the following
-file : 
-
-    $MY_KARAF_HOME/etc/users.properties
-
-For your context servers, and for any standalone Elasticsearch nodes you will 
need to open the following ports for proper
-node-to-node communication : 9200 (Elasticsearch REST API), 9300 
(Elasticsearch TCP transport)
-
-Of course any ports listed here are the default ports configured in each 
server, you may adjust them if needed.
-
-Step 2 : Adjust the Context Server IP filtering
-
-Step 3 : Follow industry recommended best practices for securing Elasticsearch
-
-You may find more valuable recommendations here : 
-
-- https://www.elastic.co/blog/found-elasticsearch-security
-- https://www.elastic.co/blog/scripting-security
-    
-Step 4 : Setup a proxy in front of the context server
-
-As an alternative to an application-level firewall, you could also route all 
traffic to the context server through
-a proxy, and use it to filter any communication.
-
-Checking the license status
----------------------------
-
-We have a new profile to check the license headers, you can perform a full 
project check with the following command:
- 
-    mvn clean install -P integration-tests,performance-tests,rat
-    
-Generating the Maven documentation site
----------------------------------------
-
-To generate the site use : 
-
-    mvn site site:stage
-    
-You will then find a full generate site in the target/staging directory
-
-To upload the site to the Apache website, simply run after the above command 
has completed:
-
-    mvn scm-publish:publish-scm
-    
-This operation takes a little bit of time, so don't interrupt it even if 
you're waiting for a while for it to complete
-(usually takes about 16 minutes !)
+You can find all the updated documentation, including building and deployment 
instructions, on the [Apache Unomi 
+web site](http://unomi.incubator.apache.org).

http://git-wip-us.apache.org/repos/asf/incubator-unomi/blob/ddbebd25/persistence-elasticsearch/core/src/main/resources/org.apache.unomi.persistence.elasticsearch.cfg
----------------------------------------------------------------------
diff --git 
a/persistence-elasticsearch/core/src/main/resources/org.apache.unomi.persistence.elasticsearch.cfg
 
b/persistence-elasticsearch/core/src/main/resources/org.apache.unomi.persistence.elasticsearch.cfg
index 55a24ea..e0c382c 100644
--- 
a/persistence-elasticsearch/core/src/main/resources/org.apache.unomi.persistence.elasticsearch.cfg
+++ 
b/persistence-elasticsearch/core/src/main/resources/org.apache.unomi.persistence.elasticsearch.cfg
@@ -33,5 +33,9 @@ bulkProcessor.bulkSize=5MB
 bulkProcessor.flushInterval=5s
 bulkProcessor.backoffPolicy=exponential
 
+# The following settings are used to perform version checks on the connected 
ElasticSearch cluster, to make sure that
+# appropriate versions are used. The check is performed like this :
+# for each node in the ElasticSearch cluster:
+#   minimalElasticSearchVersion <= ElasticSearch node version < 
maximalElasticSearchVersion
 minimalElasticSearchVersion=5.0.0
 maximalElasticSearchVersion=5.2.0
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/incubator-unomi/blob/ddbebd25/pom.xml
----------------------------------------------------------------------
diff --git a/pom.xml b/pom.xml
index 57da7e7..e47920a 100644
--- a/pom.xml
+++ b/pom.xml
@@ -29,9 +29,9 @@
     <groupId>org.apache.unomi</groupId>
     <artifactId>unomi-root</artifactId>
     <name>Apache Unomi :: Root Project</name>
-    <description>Unomi is a reference implementation of the OASIS Context 
Server specification currently being worked on
-        by the OASIS Context Server Technical Committee. It
-        provides a high-performance user profile and event tracking server.
+    <description>Apache Unomi is a reference implementation of the OASIS 
Context Server specification currently being
+        worked on by the OASIS Context Server Technical Committee. It provides 
a high-performance user profile and
+        event tracking server.
     </description>
     <url>http://unomi.incubator.apache.org</url>
     <version>1.2.0-incubating-SNAPSHOT</version>

http://git-wip-us.apache.org/repos/asf/incubator-unomi/blob/ddbebd25/src/site/markdown/building-and-deploying.md
----------------------------------------------------------------------
diff --git a/src/site/markdown/building-and-deploying.md 
b/src/site/markdown/building-and-deploying.md
index c7a18c2..68e7668 100644
--- a/src/site/markdown/building-and-deploying.md
+++ b/src/site/markdown/building-and-deploying.md
@@ -17,28 +17,68 @@
 Building
 ========
 
-Simply type at the root of the project:
+Initial Setup
+-------------
 
-    mvn clean install
+1) Install J2SE 8.0 SDK (or later), which can be downloaded from
+   http://www.oracle.com/technetwork/java/javase/downloads/index.html
 
-The Maven build process will generate both a standalone package you can use 
directly to start the context server
-(see "Deploying the generated package") or a KAR file that you can then deploy 
using a manual deployment process into
-an already installed Apache Karaf server (see "Deploying into an existing 
Karaf server")
+2) Make sure that your JAVA_HOME environment variable is set to the newly 
installed
+   JDK location, and that your PATH includes %JAVA_HOME%\bin (windows) or
+   $JAVA_HOME$/bin (unix).
 
-If you want to build and run the integration tests, you should instead use : 
+3) Install Maven 3.0.3 (or later), which can be downloaded from
+   http://maven.apache.org/download.html. Make sure that your PATH includes
+   the MVN_HOME/bin directory.
 
-    mvn -P integration-tests clean install
 
-Deploying the generated package
--------------------------------
+Building
+--------
+
+1) Change to the top level directory of Apache Unomi source distribution.
+2) Run
+
+         $> mvn clean install
+
+   This will compile Apache Unomi and run all of the tests in the
+   Apache Unomi source distribution. Alternatively, you can run
+
+         $> mvn -P \!integration-tests,\!performance-tests clean install
+
+   This will compile Apache Unomi without running the tests and takes less
+   time to build.
+
+3) The distributions will be available under "package/target" directory.
+
+Installing an ElasticSearch server
+----------------------------------
+
+Starting with version 1.2, Apache Unomi no longer embeds an ElasticSearch 
server as this is no longer supported by 
+the developers of ElasticSearch. Therefore you will need to install a 
standalone ElasticSearch using the following steps:
+
+1. Download an ElasticSearch 5.x version (5.1.1 or more recent, but not 6.x) 
from the following site:
+
+    https://www.elastic.co/downloads/elasticsearch
+    
+2. Uncompress the downloaded package into a directory and launch the server 
using
+
+    bin/elasticsearch (Mac, Linux)
+    bin\elasticsearch.bat (Windows)
+    
+3. Check that the ElasticSearch is up and running by accessing the following 
URL : 
+
+    http://localhost:9200    
+
+Deploying the generated binary package
+--------------------------------------
 
 The "package" sub-project generates a pre-configured Apache Karaf installation 
that is the simplest way to get started.
-Simply uncompress the `package/target/unomi-VERSION.tar.gz` (for Linux or Mac 
OS X) or
- `package/target/unomi-VERSION.zip` (for Windows) archive into the directory 
of your choice.
+Simply uncompress the package/target/unomi-VERSION.tar.gz (for Linux or Mac OS 
X) or
+ package/target/unomi-VERSION.zip (for Windows) archive into the directory of 
your choice.
  
 You can then start the server simply by using the command on UNIX/Linux/MacOS 
X : 
 
-    ./bin/karaf start
+    ./bin/karaf start    
     
 or on Windows shell : 
 
@@ -52,8 +92,7 @@ This is only needed if you didn't use the generated package. 
Also, this is the p
 environment if you intend to re-deploy the context server KAR iteratively.
 
 Additional requirements:
- - Apache Karaf 3.0.2+, http://karaf.apache.org
- - Local copy of the Elasticsearch ZIP package, available here : 
http://www.elasticsearch.org
+* Apache Karaf 3.x, http://karaf.apache.org
 
 1. Before deploying, make sure that you have Apache Karaf properly installed. 
You will also have to increase the
 default maximum memory size and perm gen size by adjusting the following 
environment values in the bin/setenv(.bat)
@@ -62,22 +101,17 @@ files (at the end of the file):
     ```
        MY_DIRNAME=`dirname $0`
        MY_KARAF_HOME=`cd "$MY_DIRNAME/.."; pwd`
-       export KARAF_OPTS="-Djava.library.path=$MY_KARAF_HOME/lib/sigar"
        export JAVA_MAX_MEM=3G
        export JAVA_MAX_PERM_MEM=384M
     ```
     
-2. You will also need to have the Hyperic Sigar native libraries in your Karaf 
installation, so in order to this
-go to the Elasticsearch website (http://www.elasticsearch.org)  and download 
the ZIP package. Decompress it somewhere 
-on your disk and copy all the files from the lib/sigar directory into Karaf's 
lib/sigar directory 
-(must be created first) EXCEPT THE SIGAR.JAR file.
-
-3. Install the WAR support, CXF into Karaf by doing the following in the Karaf 
command line:
+2. Install the WAR support, CXF and CDI (OpenWebBeans) into Karaf by doing the 
following in the Karaf command line:
 
     ```
-       feature:install -v war
-       feature:repo-add cxf 2.7.11
-       feature:install -v cxf/2.7.11
+       feature:repo-add cxf 3.0.2
+       feature:repo-add cellar 3.0.3
+       feature:repo-add mvn:org.apache.unomi/unomi-kar/VERSION/xml/features
+       feature:install unomi-kar
     ```
 
 4. Create a new $MY_KARAF_HOME/etc/org.apache.cxf.osgi.cfg file and put the 
following property inside :
@@ -85,14 +119,8 @@ on your disk and copy all the files from the lib/sigar 
directory into Karaf's li
     ```
        org.apache.cxf.servlet.context=/cxs
     ```
-
-5. Copy the following KAR to the Karaf deploy directory, as in this example 
line:
-
-    ```
-      cp kar/target/unomi-kar-1.0.0-SNAPSHOT.kar 
~/java/deployments/unomi/apache-karaf-3.0.1/deploy/
-    ```
    
-6. If all went smoothly, you should be able to access the context script here 
: http://localhost:8181/cxs/cluster .
+5. If all went smoothly, you should be able to access the context script here 
: http://localhost:8181/cxs/cluster .
  You should be able to login with karaf / karaf and see basic server 
information. If not something went wrong during the install.
  
 JDK Selection on Mac OS X

http://git-wip-us.apache.org/repos/asf/incubator-unomi/blob/ddbebd25/src/site/markdown/clustering.md
----------------------------------------------------------------------
diff --git a/src/site/markdown/clustering.md b/src/site/markdown/clustering.md
index 9e1a548..03e19fe 100644
--- a/src/site/markdown/clustering.md
+++ b/src/site/markdown/clustering.md
@@ -18,11 +18,21 @@
 Cluster setup
 =============
 
-Context server relies on Apache Karaf Cellar to discover and configure its 
cluster. You just need to install multiple context
-servers on the same network, and enable the discovery protocol in 
$MY_KARAF_HOME/etc/hazelcast.xml file.
+Apache Karaf relies on Apache Karaf Cellar, which in turn uses Hazelcast to 
discover and configure its cluster. 
+You just need to install multiple context servers on the same network, and 
then (optionally) change the Hazelcast 
+ configuration in the following file :
+
+    etc/hazelcast.xml
 
 All nodes on the same network, sharing the same cluster name will be part of 
the same cluster.
 
+For the actual ElasticSearch configuration however, this must be done using 
the following file:
+
+    etc/org.apache.unomi.persistence.elasticsearch.cfg
+    
+Depending on the cluster size, you will want to adjust the following 
parameters to make sure your setup is optimal in 
+terms of performance and safety.
+
 #### 2 nodes  configuration
 One node dedicated to context server, 1 node for elasticsearch storage.
 
@@ -53,3 +63,4 @@ Node C :
 
     numberOfReplicas=1
     monthlyIndex.numberOfReplicas=1
+

http://git-wip-us.apache.org/repos/asf/incubator-unomi/blob/ddbebd25/src/site/markdown/configuration.md
----------------------------------------------------------------------
diff --git a/src/site/markdown/configuration.md 
b/src/site/markdown/configuration.md
index 0bf595c..db3473d 100644
--- a/src/site/markdown/configuration.md
+++ b/src/site/markdown/configuration.md
@@ -43,6 +43,23 @@ with the following contents:
 
     cluster.name=contextElasticSearch
     index.name=context
+    
+Secured events configuration
+---------------------------
+
+If you need to secure some events, that will be sent only by a trusted third 
party server, you can update the file :
+
+    $MY_KARAF_HOME/etc/org.apache.unomi.thirdparty.cfg
+
+Ususally, login events, which operate on profiles and do merge on protected 
properties, must be secured. For each
+trusted third party server, you need to add these 3 lines :
+
+thirdparty.provider1.key=secret-key
+thirdparty.provider1.ipAddresses=127.0.0.1,::1
+thirdparty.provider1.allowedEvents=login,download
+
+The events set in allowedEvents will be secured and will only be accepted if 
the call comes from the specified IP
+address, and if the secret-key is passed in the X-Unomi-Peer header.    
 
 Installing the MaxMind GeoIPLite2 IP lookup database
 ----------------------------------------------------
@@ -91,6 +108,17 @@ The generated package is also configured with a default SSL 
certificate. You can
 
 You should now have SSL setup on Karaf with your certificate, and you can test 
it by trying to access it on port 9443.
 
+3. Changing the default Karaf password can be done by modifying the 
etc/users.properties file
+
+4. You will also need to change the user/password information in the 
org.apache.unomi.cluster.cfg file : 
+
+```
+    cluster.group=default
+    cluster.jmxUsername=karaf
+    cluster.jmxPassword=karaf
+    cluster.jmxPort=1099
+```
+
 Automatic profile merging
 -------------------------
 

http://git-wip-us.apache.org/repos/asf/incubator-unomi/blob/ddbebd25/src/site/markdown/index.md
----------------------------------------------------------------------
diff --git a/src/site/markdown/index.md b/src/site/markdown/index.md
index 548ca8d..ec98be1 100644
--- a/src/site/markdown/index.md
+++ b/src/site/markdown/index.md
@@ -17,7 +17,7 @@
   
 ![Unomi logo](images/apache-unomi-380x85.png)
 
-## A front end for personalization big data technologies (with built-in 
privacy management)
+## A personalization and A/B testing open source server (with built-in privacy 
management)
 
 ### In a few words
 
@@ -90,6 +90,8 @@ compliant server community.
 
 ### News
 
+- 2016-10-04 Released version 1.1.0-incubating
+- 2016-05-22 Released version 1.0.0-incubating
 - 2015-11-23 Initial code base import in [Git 
repository](source-repository.html) 
 - 2015-11-20 Added [Apache Maturity Model report 
page](maturity-model-report.html)
 - 2015-11-13 Initial web site created

Reply via email to