This is an automated email from the ASF dual-hosted git repository.

abhi pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/ranger-tools.git


The following commit(s) were added to refs/heads/main by this push:
     new a1d7880  RANGER-5506: Support Ranger Docker Image Releases from 
apache/ranger-tools (#8)
a1d7880 is described below

commit a1d788028cc7520f10f25889990d34a668996a5f
Author: Abhishek Kumar <[email protected]>
AuthorDate: Thu Mar 5 10:43:00 2026 -0800

    RANGER-5506: Support Ranger Docker Image Releases from apache/ranger-tools 
(#8)
---
 .github/workflows/build-and-tag-ranger-image.yaml  |  144 +++
 release/.dockerignore                              |    4 +
 release/.gitignore                                 |    7 +
 release/Dockerfile.ranger                          |   46 +
 release/Dockerfile.ranger-postgres                 |   27 +
 release/Dockerfile.ranger-solr                     |   26 +
 release/README.md                                  |   62 ++
 release/downloads/log4jdbc-1.2.jar                 |  Bin 0 -> 62913 bytes
 release/downloads/postgresql-42.2.16.jre7.jar      |  Bin 0 -> 980002 bytes
 release/scripts/create-ranger-services.py          |   75 ++
 release/scripts/download-ranger.sh                 |   53 +
 .../ranger-admin-install-postgres.properties       |   99 ++
 release/scripts/ranger.sh                          |   66 ++
 release/scripts/rdbms/init_postgres.sh             |   26 +
 .../scripts/solr/solr-ranger_audits/managed-schema |   95 ++
 .../scripts/solr/solr-ranger_audits/solrconfig.xml | 1148 ++++++++++++++++++++
 16 files changed, 1878 insertions(+)

diff --git a/.github/workflows/build-and-tag-ranger-image.yaml 
b/.github/workflows/build-and-tag-ranger-image.yaml
new file mode 100644
index 0000000..d657475
--- /dev/null
+++ b/.github/workflows/build-and-tag-ranger-image.yaml
@@ -0,0 +1,144 @@
+# Licensed to the Apache Software Foundation (ASF) under one or more
+# contributor license agreements.  See the NOTICE file distributed with
+# this work for additional information regarding copyright ownership.
+# The ASF licenses this file to You under the Apache License, Version 2.0
+# (the "License"); you may not use this file except in compliance with
+# the License.  You may obtain a copy of the License at
+#
+#     http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+name: build-and-tag-ranger-image
+
+# This workflow builds the following images: ranger, ranger-db, ranger-solr.
+# It also pushes the image to the GitHub Container Registry, tagging it based 
on the ranger version present in the release branch.
+# It pushes the images to DockerHub if an OAuth token is provided as input.
+
+# Use this command to generate a unique 11 character token:
+# code=$(uuidgen | tr A-Z a-z | cut -c 1-11)
+# then pass the code to state param here: 
https://oauth.apache.org/auth?redirect_uri=https://ranger.apache.org&state=code
+# On successful authentication, it generates an OAuth token on redirect_uri 
that can be used to trigger the workflow and push the images to DockerHub.
+
+# For more info, read ASF OAuth doc here: https://idm.apache.org/api.html
+on:
+  workflow_dispatch:
+    inputs:
+      token:
+        description: 'OAuth Access Token'
+        required: true
+        type: string
+      release-version:
+        description: 'Ranger Release Version'
+        required: true
+        type: string
+
+permissions:
+  contents: read
+  packages: write
+
+jobs:
+  build:
+    runs-on: ubuntu-latest
+    steps:
+      - name: Set up QEMU
+        uses: docker/setup-qemu-action@49b3bc8e6bdd4a60e6116a5414239cba5943d3cf
+
+      - name: Set up Docker Buildx
+        uses: 
docker/setup-buildx-action@c47758b77c9736f4b2ef4073d4d51994fabfe349
+
+      - name: Login to GitHub Container Registry
+        id: login
+        uses: docker/login-action@9780b0c442fbb1117ed29e0efdff1e18412f7567
+        with:
+          registry: ghcr.io
+          username: ${{ github.repository_owner }}
+          password: ${{ secrets.GITHUB_TOKEN }}
+
+      - name: Build and push images
+        id: build
+        working-directory: release
+        run: |
+          export RANGER_VERSION=${{ github.event.inputs.release-version }})
+          docker buildx build \
+            --build-arg RANGER_VERSION=${RANGER_VERSION} \
+            --file Dockerfile.ranger \
+            --platform linux/amd64,linux/arm64 \
+            --tag ghcr.io/${{ github.repository_owner 
}}/ranger:${RANGER_VERSION} \
+            --push .
+          docker buildx build \
+            --file Dockerfile.ranger-postgres \
+            --platform linux/amd64,linux/arm64 \
+            --tag ghcr.io/${{ github.repository_owner 
}}/ranger-db:${RANGER_VERSION} \
+            --push .
+          docker buildx build \
+            --file Dockerfile.ranger-solr \
+            --platform linux/amd64,linux/arm64 \
+            --tag ghcr.io/${{ github.repository_owner 
}}/ranger-solr:${RANGER_VERSION} \
+            --push .
+
+  tag:
+    needs: build
+    if: ${{ github.event.inputs.token != '' }}
+    runs-on: ubuntu-latest
+    env:
+      DOCKERHUB_USER: ${{ secrets.DOCKERHUB_USER }}
+      RANGER_VERSION: ${{ github.event.inputs.release-version }}
+      REGISTRIES: ghcr.io # docker.io is appended dynamically
+    steps:
+      - name: Verify OAuth Token
+        run: |
+          response=$(curl -LSs https://oauth.apache.org/token\?code\=${{ 
github.event.inputs.token }})
+          echo "$response" | jq -e . >/dev/null 2>&1
+          if [[ $? -eq 0 ]]; then
+            check=$(echo $response | jq -e --arg str "ranger" '.pmcs | 
index($str) != null')
+            if [[ $check == true ]]; then
+              echo "Authorized to push to Docker Hub"
+            else
+              echo "Not authorized to push to Docker Hub"
+              exit 1
+            fi
+          else
+            echo $response
+          fi
+
+      - name: Add Docker Hub to targets
+        if: ${{ env.DOCKERHUB_USER }}
+        run: |
+          echo "REGISTRIES=${{ env.REGISTRIES }} docker.io" >> $GITHUB_ENV
+
+      - name: Pull image
+        run: |
+          docker pull ghcr.io/${{ github.repository_owner 
}}/ranger:${RANGER_VERSION}
+          docker pull ghcr.io/${{ github.repository_owner 
}}/ranger-db:${RANGER_VERSION}
+          docker pull ghcr.io/${{ github.repository_owner 
}}/ranger-solr:${RANGER_VERSION}
+
+      - name: Login to GitHub Container Registry
+        uses: docker/login-action@9780b0c442fbb1117ed29e0efdff1e18412f7567
+        with:
+          registry: ghcr.io
+          username: ${{ github.repository_owner }}
+          password: ${{ secrets.GITHUB_TOKEN }}
+
+      - name: Login to Docker Hub
+        if: ${{ env.DOCKERHUB_USER }}
+        uses: docker/login-action@9780b0c442fbb1117ed29e0efdff1e18412f7567
+        with:
+          username: ${{ env.DOCKERHUB_USER }}
+          password: ${{ secrets.DOCKERHUB_TOKEN }}
+
+      - name: Apply tags to existing image
+        run: |
+          set -x
+          for registry in $REGISTRIES; do
+            for service in ranger ranger-db ranger-solr; do
+              opts="$(echo "$RANGER_VERSION" | sed "s@^@--tag $registry/${{ 
github.repository_owner }}/$service:@g" | xargs echo)"
+              if [[ -n "$opts" ]]; then
+                docker buildx imagetools create $opts ghcr.io/${{ 
github.repository_owner }}/$service:${RANGER_VERSION}
+              fi
+            done
+          done
\ No newline at end of file
diff --git a/release/.dockerignore b/release/.dockerignore
new file mode 100644
index 0000000..b423db3
--- /dev/null
+++ b/release/.dockerignore
@@ -0,0 +1,4 @@
+*
+!config
+!downloads/*
+!scripts/*
diff --git a/release/.gitignore b/release/.gitignore
new file mode 100644
index 0000000..ed44abb
--- /dev/null
+++ b/release/.gitignore
@@ -0,0 +1,7 @@
+*.class
+*.iml
+.settings/
+.metadata
+.classpath
+.project
+/target/
diff --git a/release/Dockerfile.ranger b/release/Dockerfile.ranger
new file mode 100644
index 0000000..4cf55a5
--- /dev/null
+++ b/release/Dockerfile.ranger
@@ -0,0 +1,46 @@
+# Licensed to the Apache Software Foundation (ASF) under one
+# or more contributor license agreements.  See the NOTICE file
+# distributed with this work for additional information
+# regarding copyright ownership.  The ASF licenses this file
+# to you under the Apache License, Version 2.0 (the
+# "License"); you may not use this file except in compliance
+# with the License.  You may obtain a copy of the License at
+#
+#     http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+ARG RANGER_BASE_IMAGE=apache/ranger-base
+ARG RANGER_BASE_VERSION=20260123-2-8
+
+FROM ${RANGER_BASE_IMAGE}:${RANGER_BASE_VERSION} AS ranger
+
+ARG RANGER_VERSION
+ARG TARGETARCH
+
+COPY ./scripts/download-ranger.sh                                
${RANGER_SCRIPTS}/
+COPY ./scripts/ranger.sh                                         
${RANGER_SCRIPTS}/
+COPY ./scripts/ranger-admin-install-postgres.properties          
${RANGER_SCRIPTS}/ranger-admin-install.properties
+COPY ./scripts/create-ranger-services.py                         
${RANGER_SCRIPTS}/
+
+# Download released ranger-admin.tar.gz
+WORKDIR /home/ranger
+RUN chmod +x ${RANGER_SCRIPTS}/download-ranger.sh && 
${RANGER_SCRIPTS}/download-ranger.sh
+WORKDIR /
+
+RUN    tar xvfz ${RANGER_DIST}/ranger-${RANGER_VERSION}-admin.tar.gz 
--directory=${RANGER_HOME} \
+    && ln -s ${RANGER_HOME}/ranger-${RANGER_VERSION}-admin 
${RANGER_HOME}/admin \
+    && rm -rf ${RANGER_DIST}/* \
+    && mkdir -p /var/run/ranger /var/log/ranger \
+    && chown -R ranger:ranger ${RANGER_HOME}/admin/ ${RANGER_SCRIPTS}/ 
/var/run/ranger/ /var/log/ranger/ \
+    && chmod 755 ${RANGER_SCRIPTS}/ranger.sh 
${RANGER_SCRIPTS}/create-ranger-services.py \
+    && mkdir -p /usr/share/java/
+
+COPY ./downloads/postgresql-42.2.16.jre7.jar /usr/share/java/postgresql.jar
+
+USER ranger
+
+ENTRYPOINT [ "/home/ranger/scripts/ranger.sh" ]
\ No newline at end of file
diff --git a/release/Dockerfile.ranger-postgres 
b/release/Dockerfile.ranger-postgres
new file mode 100644
index 0000000..ab36ef9
--- /dev/null
+++ b/release/Dockerfile.ranger-postgres
@@ -0,0 +1,27 @@
+# Licensed to the Apache Software Foundation (ASF) under one
+# or more contributor license agreements.  See the NOTICE file
+# distributed with this work for additional information
+# regarding copyright ownership.  The ASF licenses this file
+# to you under the Apache License, Version 2.0 (the
+# "License"); you may not use this file except in compliance
+# with the License.  You may obtain a copy of the License at
+#
+#     http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+ARG POSTGRES_VERSION=13.16
+FROM postgres:${POSTGRES_VERSION}
+
+# Copy DB init script
+USER 0
+
+RUN  mkdir -p /docker-entrypoint-initdb.d
+
+COPY ./scripts/rdbms/init_postgres.sh /docker-entrypoint-initdb.d/
+
+RUN chown -R postgres:postgres /docker-entrypoint-initdb.d/
diff --git a/release/Dockerfile.ranger-solr b/release/Dockerfile.ranger-solr
new file mode 100644
index 0000000..4256c89
--- /dev/null
+++ b/release/Dockerfile.ranger-solr
@@ -0,0 +1,26 @@
+# Licensed to the Apache Software Foundation (ASF) under one
+# or more contributor license agreements.  See the NOTICE file
+# distributed with this work for additional information
+# regarding copyright ownership.  The ASF licenses this file
+# to you under the Apache License, Version 2.0 (the
+# "License"); you may not use this file except in compliance
+# with the License.  You may obtain a copy of the License at
+#
+#     http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+ARG SOLR_VERSION=8.11.3
+FROM solr:${SOLR_VERSION}
+
+# Copy audit config set
+USER 0
+RUN  mkdir -p /opt/solr/server/solr/configsets/ranger_audits/conf
+COPY scripts/solr/solr-ranger_audits/* 
/opt/solr/server/solr/configsets/ranger_audits/conf/
+RUN chown -R solr:solr /opt/solr/server/solr/configsets/ranger_audits/
+
+USER solr
\ No newline at end of file
diff --git a/release/README.md b/release/README.md
new file mode 100644
index 0000000..4cad44e
--- /dev/null
+++ b/release/README.md
@@ -0,0 +1,62 @@
+<!--
+  Licensed to the Apache Software Foundation (ASF) under one or more
+  contributor license agreements.  See the NOTICE file distributed with
+  this work for additional information regarding copyright ownership.
+  The ASF licenses this file to You under the Apache License, Version 2.0
+  (the "License"); you may not use this file except in compliance with
+  the License.  You may obtain a copy of the License at
+
+      http://www.apache.org/licenses/LICENSE-2.0
+
+  Unless required by applicable law or agreed to in writing, software
+  distributed under the License is distributed on an "AS IS" BASIS,
+  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+  See the License for the specific language governing permissions and
+  limitations under the License.
+-->
+
+# Apache Ranger Images
+
+[ranger-tools/release](https://github.com/apache/ranger-tools/release) 
contains utilities to publish docker images for released Ranger versions. 
+
+## Usage
+
+### Build Images
+
+To build the images, run the following commands from the `release` directory:
+
+```bash
+export RANGER_VERSION=2.8.0
+docker build -f Dockerfile.ranger-postgres -t ranger-db:latest .
+docker build -f Dockerfile.ranger-solr -t ranger-solr:latest .
+docker build --build-arg RANGER_VERSION=${RANGER_VERSION} -f Dockerfile.ranger 
-t ranger:latest .
+```
+
+### Run Containers
+
+To run the containers, use the following commands:
+
+```bash
+docker network create rangernw
+
+export ZK_VERSION=3.9.2
+docker run -d --name ranger-zk --hostname ranger-zk.rangernw --network 
rangernw -p 2181:2181 zookeeper:${ZK_VERSION}
+
+docker run -d --name ranger-solr --hostname ranger-solr.rangernw --network 
rangernw -p 8983:8983 ranger-solr:latest \
+  solr-precreate ranger_audits /opt/solr/server/solr/configsets/ranger_audits/
+
+docker run -d \
+  -e POSTGRES_PASSWORD=rangerR0cks! \
+  -e RANGER_DB_USER=rangeradmin \
+  -e RANGER_DB_PASSWORD=rangerR0cks! \
+  --name ranger-db --hostname ranger-db.rangernw --network rangernw 
--health-cmd='su -c "pg_isready -q" postgres' --health-interval=10s 
--health-timeout=2s --health-retries=30 ranger-db:latest
+
+docker run -d \
+  -e POSTGRES_PASSWORD=rangerR0cks! \
+  -e RANGER_DB_USER=rangeradmin \
+  -e RANGER_DB_PASSWORD=rangerR0cks! \
+  --name ranger-admin --hostname ranger-admin.rangernw --network rangernw -p 
6080:6080 ranger:latest
+```
+### Access Ranger Admin UI
+
+Once the containers are running, you can access the Ranger Admin UI by 
navigating to `http://localhost:6080` in your web browser. The default 
credentials are: `admin/rangerR0cks!`
\ No newline at end of file
diff --git a/release/downloads/log4jdbc-1.2.jar 
b/release/downloads/log4jdbc-1.2.jar
new file mode 100644
index 0000000..8208a44
Binary files /dev/null and b/release/downloads/log4jdbc-1.2.jar differ
diff --git a/release/downloads/postgresql-42.2.16.jre7.jar 
b/release/downloads/postgresql-42.2.16.jre7.jar
new file mode 100644
index 0000000..02243ef
Binary files /dev/null and b/release/downloads/postgresql-42.2.16.jre7.jar 
differ
diff --git a/release/scripts/create-ranger-services.py 
b/release/scripts/create-ranger-services.py
new file mode 100644
index 0000000..6909bf1
--- /dev/null
+++ b/release/scripts/create-ranger-services.py
@@ -0,0 +1,75 @@
+from apache_ranger.model.ranger_service import RangerService
+from apache_ranger.client.ranger_client import RangerClient
+from json import JSONDecodeError
+
+ranger_client = RangerClient('http://ranger-admin.rangernw:6080', ('admin', 
'rangerR0cks!'))
+
+
+def service_not_exists(service):
+    try:
+        svc = ranger_client.get_service(service.name)
+    except JSONDecodeError:
+        return 1
+    return 0 if svc is not None else 1
+
+
+hdfs = RangerService({'name': 'dev_hdfs', 'type': 'hdfs',
+                      'configs': {'username': 'hdfs', 'password': 'hdfs',
+                                  'fs.default.name': 
'hdfs://ranger-hadoop:9000',
+                                  'hadoop.security.authentication': 'simple',
+                                  'hadoop.security.authorization': 'true'}})
+
+hive = RangerService({'name': 'dev_hive', 'type': 'hive',
+                      'configs': {'username': 'hive', 'password': 'hive',
+                                  'jdbc.driverClassName': 
'org.apache.hive.jdbc.HiveDriver',
+                                  'jdbc.url': 'jdbc:hive2://ranger-hive:10000',
+                                  'hadoop.security.authorization': 'true'}})
+
+kafka = RangerService({'name': 'dev_kafka', 'type': 'kafka',
+                       'configs': {'username': 'kafka', 'password': 'kafka',
+                                   'zookeeper.connect': 
'ranger-zk.example.com:2181'}})
+
+knox = RangerService({'name': 'dev_knox', 'type': 'knox',
+                      'configs': {'username': 'knox', 'password': 'knox', 
'knox.url': 'https://ranger-knox:8443'}})
+
+yarn = RangerService({'name': 'dev_yarn', 'type': 'yarn',
+                      'configs': {'username': 'yarn', 'password': 'yarn',
+                                  'yarn.url': 'http://ranger-hadoop:8088'}})
+
+hbase = RangerService({'name': 'dev_hbase', 'type': 'hbase',
+                       'configs': {'username': 'hbase', 'password': 'hbase',
+                                   'hadoop.security.authentication': 'simple',
+                                   'hbase.security.authentication': 'simple',
+                                   'hadoop.security.authorization': 'true',
+                                   'hbase.zookeeper.property.clientPort': 
'2181',
+                                   'hbase.zookeeper.quorum': 'ranger-zk',
+                                   'zookeeper.znode.parent': '/hbase'}})
+
+kms = RangerService({'name': 'dev_kms', 'type': 'kms',
+                     'configs': {'username': 'keyadmin', 'password': 
'rangerR0cks!',
+                                 'provider': 'http://ranger-kms:9292'}})
+
+trino = RangerService({'name': 'dev_trino',
+                       'type': 'trino',
+                       'configs': {
+                           'username': 'trino',
+                           'password': 'trino',
+                           'jdbc.driverClassName': 'io.trino.jdbc.TrinoDriver',
+                           'jdbc.url': 'jdbc:trino://ranger-trino:8080',
+                       }})
+
+ozone = RangerService({'name': 'dev_ozone',
+                       'type': 'ozone',
+                       'displayName': 'dev_ozone',
+                       'configs': {'username': 'hdfs', 'password': 'hdfs',
+                                   'ozone.om.http-address': 'http://om:9874',
+                                   'hadoop.security.authentication': 
'simple'}})
+
+services = [hdfs, yarn, hive, hbase, kafka, knox, kms, trino, ozone]
+for service in services:
+    try:
+        if service_not_exists(service):
+            ranger_client.create_service(service)
+            print(f" {service.name} service created!")
+    except Exception as e:
+        print(f"An exception occured: {e}")
diff --git a/release/scripts/download-ranger.sh 
b/release/scripts/download-ranger.sh
new file mode 100644
index 0000000..0fa20ad
--- /dev/null
+++ b/release/scripts/download-ranger.sh
@@ -0,0 +1,53 @@
+#!/usr/bin/env bash
+
+# Licensed to the Apache Software Foundation (ASF) under one
+# or more contributor license agreements.  See the NOTICE file
+# distributed with this work for additional information
+# regarding copyright ownership.  The ASF licenses this file
+# to you under the Apache License, Version 2.0 (the
+# "License"); you may not use this file except in compliance
+# with the License.  You may obtain a copy of the License at
+#
+#     http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+set -u -o pipefail
+
+: 
${BASE_URL:="https://www.apache.org/dyn/closer.lua?action=download&filename="}
+: ${CHECKSUM_BASE_URL:="https://downloads.apache.org/"}
+: ${RANGER_VERSION:=2.8.0}
+
+download_if_not_exists() {
+  local url="$1"
+  local file="$2"
+
+  if [[ -e "${file}" ]]; then
+    echo "${file} already downloaded"
+  else
+    echo "Downloading ${file} from ${url}"
+    curl --fail --location --output "${file}" --show-error --silent "${url}" 
|| rm -fv "${file}"
+  fi
+}
+
+download_and_verify() {
+  local remote_path="$1"
+  local file="$(basename "${remote_path}")"
+
+  pushd dist
+  download_if_not_exists "${BASE_URL}${remote_path}" "${file}"
+  download_if_not_exists "${CHECKSUM_BASE_URL}${remote_path}.asc" "${file}.asc"
+  curl -LO https://downloads.apache.org/ranger/KEYS
+  gpg --import KEYS
+  gpg --verify "${file}.asc" "${file}" || exit 1
+  popd
+}
+
+mkdir -p dist
+# source
+download_and_verify 
"ranger/${RANGER_VERSION}/apache-ranger-${RANGER_VERSION}.tar.gz"
+# binary
+download_and_verify 
"ranger/${RANGER_VERSION}/services/admin/ranger-${RANGER_VERSION}-admin.tar.gz"
diff --git a/release/scripts/ranger-admin-install-postgres.properties 
b/release/scripts/ranger-admin-install-postgres.properties
new file mode 100644
index 0000000..aae8002
--- /dev/null
+++ b/release/scripts/ranger-admin-install-postgres.properties
@@ -0,0 +1,99 @@
+# Licensed to the Apache Software Foundation (ASF) under one or more
+# contributor license agreements.  See the NOTICE file distributed with
+# this work for additional information regarding copyright ownership.
+# The ASF licenses this file to You under the Apache License, Version 2.0
+# (the "License"); you may not use this file except in compliance with
+# the License.  You may obtain a copy of the License at
+#
+#     http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+#
+# This file provides a list of the deployment variables for the Policy Manager 
Web Application
+#
+
+PYTHON_COMMAND_INVOKER=python3
+RANGER_ADMIN_LOG_DIR=/var/log/ranger
+RANGER_PID_DIR_PATH=/var/run/ranger
+DB_FLAVOR=POSTGRES
+SQL_CONNECTOR_JAR=/usr/share/java/postgresql.jar
+RANGER_ADMIN_LOGBACK_CONF_FILE=/opt/ranger/admin/ews/webapp/WEB-INF/classes/conf/logback.xml
+
+db_root_user=postgres
+db_host=ranger-db
+db_name=ranger
+
+postgres_core_file=db/postgres/optimized/current/ranger_core_db_postgres.sql
+postgres_audit_file=db/postgres/xa_audit_db_postgres.sql
+mysql_core_file=db/mysql/optimized/current/ranger_core_db_mysql.sql
+mysql_audit_file=db/mysql/xa_audit_db.sql
+
+audit_store=solr
+audit_solr_urls=http://ranger-solr:8983/solr/ranger_audits
+audit_solr_collection_name=ranger_audits
+
+# audit_store=elasticsearch
+audit_elasticsearch_urls=
+audit_elasticsearch_port=9200
+audit_elasticsearch_protocol=http
+audit_elasticsearch_user=elastic
+audit_elasticsearch_password=elasticsearch
+audit_elasticsearch_index=ranger_audits
+audit_elasticsearch_bootstrap_enabled=true
+
+policymgr_external_url=http://ranger-admin:6080
+policymgr_http_enabled=true
+
+unix_user=ranger
+unix_user_pwd=ranger
+unix_group=ranger
+
+# Following variables are referenced in db_setup.py. Do not remove these
+oracle_core_file=
+sqlserver_core_file=
+sqlanywhere_core_file=
+cred_keystore_filename=
+
+# #################  DO NOT MODIFY ANY VARIABLES BELOW 
#########################
+#
+# --- These deployment variables are not to be modified unless you understand 
the full impact of the changes
+#
+################################################################################
+XAPOLICYMGR_DIR=$PWD
+app_home=$PWD/ews/webapp
+TMPFILE=$PWD/.fi_tmp
+LOGFILE=$PWD/logfile
+LOGFILES="$LOGFILE"
+
+JAVA_BIN='java'
+JAVA_VERSION_REQUIRED='1.8'
+
+ranger_admin_max_heap_size=1g
+#retry DB and Java patches after the given time in seconds.
+PATCH_RETRY_INTERVAL=120
+STALE_PATCH_ENTRY_HOLD_TIME=10
+
+hadoop_conf=
+authentication_method=UNIX
+
+#------------ Kerberos Config -----------------
+spnego_principal=
+spnego_keytab=
+token_valid=
+admin_principal=
+admin_keytab=
+lookup_principal=
+lookup_keytab=
+audit_jaas_client_loginModuleName=
+audit_jaas_client_loginModuleControlFlag=
+audit_jaas_client_option_useKeyTab=
+audit_jaas_client_option_storeKey=
+audit_jaas_client_option_useTicketCache=
+audit_jaas_client_option_serviceName=
+audit_jaas_client_option_keyTab=
+audit_jaas_client_option_principal=
diff --git a/release/scripts/ranger.sh b/release/scripts/ranger.sh
new file mode 100644
index 0000000..d566dd2
--- /dev/null
+++ b/release/scripts/ranger.sh
@@ -0,0 +1,66 @@
+#!/bin/bash
+
+# Licensed to the Apache Software Foundation (ASF) under one
+# or more contributor license agreements.  See the NOTICE file
+# distributed with this work for additional information
+# regarding copyright ownership.  The ASF licenses this file
+# to you under the Apache License, Version 2.0 (the
+# "License"); you may not use this file except in compliance
+# with the License.  You may obtain a copy of the License at
+#
+#     http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+
+if [ ! -e ${RANGER_HOME}/.setupDone ]
+then
+  SETUP_RANGER=true
+else
+  SETUP_RANGER=false
+fi
+
+if [ "${SETUP_RANGER}" == "true" ]
+then
+  cp ${RANGER_SCRIPTS}/ranger-admin-install.properties 
${RANGER_HOME}/admin/install.properties
+  {
+         echo "db_root_password=${POSTGRES_PASSWORD}"
+         echo "db_user=${RANGER_DB_USER}"
+         echo "db_password=${RANGER_DB_PASSWORD}"
+         echo "rangerAdmin_password=${RANGER_DB_PASSWORD}"
+         echo "rangerTagsync_password=${RANGER_DB_PASSWORD}"
+         echo "rangerUsersync_password=${RANGER_DB_PASSWORD}"
+         echo "keyadmin_password=${RANGER_DB_PASSWORD}"
+  } >> ${RANGER_HOME}/admin/install.properties
+  cd ${RANGER_HOME}/admin || exit
+  if ./setup.sh;
+  then
+       rm -f ${RANGER_HOME}/admin/install.properties
+    touch "${RANGER_HOME}"/.setupDone
+  else
+    echo "Ranger Admin Setup Script didn't complete proper execution."
+  fi
+fi
+
+cd ${RANGER_HOME}/admin && ./ews/ranger-admin-services.sh start
+
+if [ "${SETUP_RANGER}" == "true" ]
+then
+  # Wait for Ranger Admin to become ready
+  sleep 30
+  python3 ${RANGER_SCRIPTS}/create-ranger-services.py
+fi
+
+RANGER_ADMIN_PID=`ps -ef  | grep -v grep | grep -i 
"org.apache.ranger.server.tomcat.EmbeddedServer" | awk '{print $2}'`
+
+# prevent the container from exiting
+if [ -z "$RANGER_ADMIN_PID" ]
+then
+  echo "Ranger Admin process probably exited, no process id found!"
+else
+  tail --pid=$RANGER_ADMIN_PID -f /dev/null
+fi
diff --git a/release/scripts/rdbms/init_postgres.sh 
b/release/scripts/rdbms/init_postgres.sh
new file mode 100644
index 0000000..f1aa705
--- /dev/null
+++ b/release/scripts/rdbms/init_postgres.sh
@@ -0,0 +1,26 @@
+#!/bin/bash
+
+# Licensed to the Apache Software Foundation (ASF) under one
+# or more contributor license agreements.  See the NOTICE file
+# distributed with this work for additional information
+# regarding copyright ownership.  The ASF licenses this file
+# to you under the Apache License, Version 2.0 (the
+# "License"); you may not use this file except in compliance
+# with the License.  You may obtain a copy of the License at
+#
+#     http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+
+set -e
+
+psql -v ON_ERROR_STOP=1 --username "$POSTGRES_USER" --dbname "$POSTGRES_DB" 
<<-EOSQL
+    CREATE USER ${RANGER_DB_USER} WITH PASSWORD '${RANGER_DB_PASSWORD}';
+    CREATE DATABASE ranger;
+    GRANT ALL PRIVILEGES ON DATABASE ranger TO ${RANGER_DB_USER};
+EOSQL
diff --git a/release/scripts/solr/solr-ranger_audits/managed-schema 
b/release/scripts/solr/solr-ranger_audits/managed-schema
new file mode 100644
index 0000000..e470725
--- /dev/null
+++ b/release/scripts/solr/solr-ranger_audits/managed-schema
@@ -0,0 +1,95 @@
+<?xml version="1.0" encoding="UTF-8"?>
+<!--
+  Licensed to the Apache Software Foundation (ASF) under one or more
+  contributor license agreements.  See the NOTICE file distributed with
+  this work for additional information regarding copyright ownership.
+  The ASF licenses this file to You under the Apache License, Version 2.0
+  (the "License"); you may not use this file except in compliance with
+  the License.  You may obtain a copy of the License at
+
+      http://www.apache.org/licenses/LICENSE-2.0
+
+  Unless required by applicable law or agreed to in writing, software
+  distributed under the License is distributed on an "AS IS" BASIS,
+  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+  See the License for the specific language governing permissions and
+  limitations under the License.
+-->
+<schema name="ranger-audit-schema" version="1.6">
+  <uniqueKey>id</uniqueKey>
+  <fieldType name="binary" class="solr.BinaryField"/>
+  <fieldType name="boolean" class="solr.BoolField" sortMissingLast="true"/>
+  <fieldType name="booleans" class="solr.BoolField" multiValued="true" 
sortMissingLast="true"/>
+  <fieldType name="date" class="solr.TrieDateField" docValues="true" 
precisionStep="0" positionIncrementGap="0"/>
+  <fieldType name="double" class="solr.TrieDoubleField" docValues="true" 
precisionStep="0" positionIncrementGap="0"/>
+  <fieldType name="float" class="solr.TrieFloatField" docValues="true" 
precisionStep="0" positionIncrementGap="0"/>
+  <fieldType name="ignored" class="solr.StrField" multiValued="true" 
indexed="false" stored="false"/>
+  <fieldType name="int" class="solr.TrieIntField" docValues="true" 
precisionStep="0" positionIncrementGap="0"/>
+  <fieldType name="key_lower_case" class="solr.TextField" 
sortMissingLast="true" omitNorms="true">
+    <analyzer>
+      <tokenizer class="solr.KeywordTokenizerFactory"/>
+      <filter class="solr.LowerCaseFilterFactory"/>
+      <filter class="solr.LengthFilterFactory" min="0" max="2500"/>
+    </analyzer>
+  </fieldType>
+  <fieldType name="long" class="solr.TrieLongField" docValues="true" 
precisionStep="0" positionIncrementGap="0"/>
+  <fieldType name="random" class="solr.RandomSortField" indexed="true"/>
+  <fieldType name="string" class="solr.StrField" sortMissingLast="true"/>
+  <fieldType name="tdate" class="solr.TrieDateField" docValues="true" 
precisionStep="6" positionIncrementGap="0"/>
+  <fieldType name="tdates" class="solr.TrieDateField" docValues="true" 
precisionStep="6" multiValued="true" positionIncrementGap="0"/>
+  <fieldType name="tdouble" class="solr.TrieDoubleField" docValues="true" 
precisionStep="8" positionIncrementGap="0"/>
+  <fieldType name="tdoubles" class="solr.TrieDoubleField" docValues="true" 
precisionStep="8" multiValued="true" positionIncrementGap="0"/>
+  <fieldType name="text_std_token_lower_case" class="solr.TextField" 
multiValued="true" positionIncrementGap="100">
+    <analyzer>
+      <tokenizer class="solr.StandardTokenizerFactory"/>
+      <filter class="solr.LowerCaseFilterFactory"/>
+    </analyzer>
+  </fieldType>
+  <fieldType name="text_ws" class="solr.TextField" positionIncrementGap="100">
+    <analyzer>
+      <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+    </analyzer>
+  </fieldType>
+  <fieldType name="tfloat" class="solr.TrieFloatField" docValues="true" 
precisionStep="8" positionIncrementGap="0"/>
+  <fieldType name="tfloats" class="solr.TrieFloatField" docValues="true" 
precisionStep="8" multiValued="true" positionIncrementGap="0"/>
+  <fieldType name="tint" class="solr.TrieIntField" docValues="true" 
precisionStep="8" positionIncrementGap="0"/>
+  <fieldType name="tints" class="solr.TrieIntField" docValues="true" 
precisionStep="8" multiValued="true" positionIncrementGap="0"/>
+  <fieldType name="tlong" class="solr.TrieLongField" docValues="true" 
precisionStep="8" positionIncrementGap="0"/>
+  <fieldType name="tlongs" class="solr.TrieLongField" docValues="true" 
precisionStep="8" multiValued="true" positionIncrementGap="0"/>
+  <field name="_expire_at_" type="tdate" multiValued="false" stored="true" 
docValues="true"/>
+  <field name="_ttl_" type="string" multiValued="false" indexed="true" 
stored="true"/>
+  <field name="_version_" type="long" indexed="false" stored="true"/>
+  <field name="access" type="key_lower_case" multiValued="false"/>
+  <field name="action" type="key_lower_case" multiValued="false"/>
+  <field name="agent" type="key_lower_case" multiValued="false"/>
+  <field name="agentHost" type="key_lower_case" multiValued="false"/>
+  <field name="cliIP" type="key_lower_case" multiValued="false"/>
+  <field name="cliType" type="key_lower_case" multiValued="false"/>
+  <field name="cluster" type="key_lower_case" multiValued="false"/>
+  <field name="reqContext" type="key_lower_case" multiValued="true"/>
+  <field name="enforcer" type="key_lower_case" multiValued="false"/>
+  <field name="event_count" type="tlong" multiValued="false" docValues="true" 
default="1"/>
+  <field name="event_dur_ms" type="tlong" multiValued="false" 
docValues="true"/>
+  <field name="evtTime" type="tdate" docValues="true"/>
+  <field name="id" type="string" multiValued="false" indexed="true" 
required="true" stored="true"/>
+  <field name="logType" type="key_lower_case" multiValued="false"/>
+  <field name="policy" type="tlong" docValues="true"/>
+  <field name="proxyUsers" type="key_lower_case" multiValued="true"/>
+  <field name="reason" type="text_std_token_lower_case" multiValued="false" 
omitNorms="false"/>
+  <field name="repo" type="key_lower_case" multiValued="false"/>
+  <field name="repoType" type="tint" multiValued="false" docValues="true"/>
+  <field name="req_caller_id" type="key_lower_case" multiValued="false"/>
+  <field name="req_self_id" type="key_lower_case" multiValued="false"/>
+  <field name="reqData" type="text_std_token_lower_case" multiValued="false"/>
+  <field name="reqUser" type="key_lower_case" multiValued="false"/>
+  <field name="resType" type="key_lower_case" multiValued="false"/>
+  <field name="resource" type="key_lower_case" multiValued="false"/>
+  <field name="result" type="tint" multiValued="false"/>
+  <field name="seq_num" type="tlong" multiValued="false" docValues="true" 
default="0"/>
+  <field name="sess" type="key_lower_case" multiValued="false"/>
+  <field name="tags" type="key_lower_case" multiValued="true"/>
+  <field name="tags_str" type="text_std_token_lower_case" multiValued="false"/>
+  <field name="text" type="text_std_token_lower_case" multiValued="true" 
indexed="true" stored="false"/>
+  <field name="zoneName" type="key_lower_case" multiValued="false"/>
+  <field name="policyVersion" type="tlong" multiValued="false"/>
+</schema>
diff --git a/release/scripts/solr/solr-ranger_audits/solrconfig.xml 
b/release/scripts/solr/solr-ranger_audits/solrconfig.xml
new file mode 100644
index 0000000..8876406
--- /dev/null
+++ b/release/scripts/solr/solr-ranger_audits/solrconfig.xml
@@ -0,0 +1,1148 @@
+<?xml version="1.0" encoding="UTF-8" ?>
+<!--
+ Licensed to the Apache Software Foundation (ASF) under one or more
+ contributor license agreements.  See the NOTICE file distributed with
+ this work for additional information regarding copyright ownership.
+ The ASF licenses this file to You under the Apache License, Version 2.0
+ (the "License"); you may not use this file except in compliance with
+ the License.  You may obtain a copy of the License at
+     http://www.apache.org/licenses/LICENSE-2.0
+ Unless required by applicable law or agreed to in writing, software
+ distributed under the License is distributed on an "AS IS" BASIS,
+ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ See the License for the specific language governing permissions and
+ limitations under the License.
+-->
+
+<!--
+     For more details about configurations options that may appear in
+     this file, see http://wiki.apache.org/solr/SolrConfigXml.
+-->
+<config>
+    <!-- In all configuration below, a prefix of "solr." for class names
+         is an alias that causes solr to search appropriate packages,
+         including org.apache.solr.(search|update|request|core|analysis)
+         You may also specify a fully qualified Java classname if you
+         have your own custom plugins.
+      -->
+
+    <!-- Controls what version of Lucene various components of Solr
+         adhere to.  Generally, you want to use the latest version to
+         get all bug fixes and improvements. It is highly recommended
+         that you fully re-index after changing this setting as it can
+         affect both how text is indexed and queried.
+    -->
+    <luceneMatchVersion>8.4.1</luceneMatchVersion>
+
+    <!-- <lib/> directives can be used to instruct Solr to load any Jars
+         identified and use them to resolve any "plugins" specified in
+         your solrconfig.xml or schema.xml (ie: Analyzers, Request
+         Handlers, etc...).
+         All directories and paths are resolved relative to the
+         instanceDir.
+         Please note that <lib/> directives are processed in the order
+         that they appear in your solrconfig.xml file, and are "stacked"
+         on top of each other when building a ClassLoader - so if you have
+         plugin jars with dependencies on other jars, the "lower level"
+         dependency jars should be loaded first.
+         If a "./lib" directory exists in your instanceDir, all files
+         found in it are included as if you had used the following
+         syntax...
+                <lib dir="./lib" />
+      -->
+
+    <!-- A 'dir' option by itself adds any files found in the directory
+         to the classpath, this is useful for including all jars in a
+         directory.
+         When a 'regex' is specified in addition to a 'dir', only the
+         files in that directory which completely match the regex
+         (anchored on both ends) will be included.
+         If a 'dir' option (with or without a regex) is used and nothing
+         is found that matches, a warning will be logged.
+         The example below can be used to load a solr-contrib along
+         with their external dependencies.
+      -->
+    <!-- <lib dir="${solr.install.dir:../../../..}/dist/" 
regex="solr-ltr-\d.*\.jar" /> -->
+
+    <!-- an exact 'path' can be used instead of a 'dir' to specify a
+         specific jar file.  This will cause a serious error to be logged
+         if it can't be loaded.
+      -->
+    <!--
+       <lib path="../a-jar-that-does-not-exist.jar" />
+    -->
+
+    <lib dir="${solr.install.dir:../../../..}/dist/" 
regex="solr-dataimporthandler-.*\.jar" />
+
+    <!-- Data Directory
+         Used to specify an alternate directory to hold all index data
+         other than the default ./data under the Solr home.  If
+         replication is in use, this should match the replication
+         configuration.
+      -->
+    <dataDir>${solr.data.dir:}</dataDir>
+
+
+    <!-- The DirectoryFactory to use for indexes.
+         solr.StandardDirectoryFactory is filesystem
+         based and tries to pick the best implementation for the current
+         JVM and platform.  solr.NRTCachingDirectoryFactory, the default,
+         wraps solr.StandardDirectoryFactory and caches small files in memory
+         for better NRT performance.
+         One can force a particular implementation via 
solr.MMapDirectoryFactory,
+         solr.NIOFSDirectoryFactory, or solr.SimpleFSDirectoryFactory.
+         solr.RAMDirectoryFactory is memory based and not persistent.
+      -->
+
+    <directoryFactory name="DirectoryFactory"
+                      
class="${solr.directoryFactory:solr.NRTCachingDirectoryFactory}">
+
+
+        <!-- These will be used if you are using the solr.HdfsDirectoryFactory,
+             otherwise they will be ignored. If you don't plan on using hdfs,
+             you can safely remove this section. -->
+        <!-- The root directory that collection data should be written to. -->
+        <str name="solr.hdfs.home">${solr.hdfs.home:}</str>
+        <!-- The hadoop configuration files to use for the hdfs client. -->
+        <str name="solr.hdfs.confdir">${solr.hdfs.confdir:}</str>
+        <!-- Enable/Disable the hdfs cache. -->
+        <str 
name="solr.hdfs.blockcache.enabled">${solr.hdfs.blockcache.enabled:true}</str>
+        <!-- Enable/Disable using one global cache for all SolrCores.
+             The settings used will be from the first HdfsDirectoryFactory 
created. -->
+        <str 
name="solr.hdfs.blockcache.global">${solr.hdfs.blockcache.global:true}</str>
+
+    </directoryFactory>
+
+    <!-- The CodecFactory for defining the format of the inverted index.
+         The default implementation is SchemaCodecFactory, which is the 
official Lucene
+         index format, but hooks into the schema to provide per-field 
customization of
+         the postings lists and per-document values in the fieldType element
+         (postingsFormat/docValuesFormat). Note that most of the alternative 
implementations
+         are experimental, so if you choose to customize the index format, 
it's a good
+         idea to convert back to the official format e.g. via 
IndexWriter.addIndexes(IndexReader)
+         before upgrading to a newer version to avoid unnecessary reindexing.
+         A "compressionMode" string element can be added to <codecFactory> to 
choose
+         between the existing compression modes in the default codec: 
"BEST_SPEED" (default)
+         or "BEST_COMPRESSION".
+    -->
+    <codecFactory class="solr.SchemaCodecFactory"/>
+
+    <schemaFactory class="ManagedIndexSchemaFactory">
+        <bool name="mutable">true</bool>
+        <str name="managedSchemaResourceName">managed-schema</str>
+    </schemaFactory>
+
+    <!-- ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
+         Index Config - These settings control low-level behavior of indexing
+         Most example settings here show the default value, but are commented
+         out, to more easily see where customizations have been made.
+         Note: This replaces <indexDefaults> and <mainIndex> from older 
versions
+         ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 
-->
+    <indexConfig>
+        <!-- maxFieldLength was removed in 4.0. To get similar behavior, 
include a
+             LimitTokenCountFilterFactory in your fieldType definition. E.g.
+         <filter class="solr.LimitTokenCountFilterFactory" 
maxTokenCount="10000"/>
+        -->
+        <!-- Maximum time to wait for a write lock (ms) for an IndexWriter. 
Default: 1000 -->
+        <!-- <writeLockTimeout>1000</writeLockTimeout>  -->
+
+        <!-- Expert: Enabling compound file will use less files for the index,
+             using fewer file descriptors on the expense of performance 
decrease.
+             Default in Lucene is "true". Default in Solr is "false" (since 
3.6) -->
+        <!-- <useCompoundFile>false</useCompoundFile> -->
+
+        <!-- ramBufferSizeMB sets the amount of RAM that may be used by Lucene
+             indexing for buffering added documents and deletions before they 
are
+             flushed to the Directory.
+             maxBufferedDocs sets a limit on the number of documents buffered
+             before flushing.
+             If both ramBufferSizeMB and maxBufferedDocs is set, then
+             Lucene will flush based on whichever limit is hit first.  -->
+        <ramBufferSizeMB>128</ramBufferSizeMB>
+        <!-- <maxBufferedDocs>1000</maxBufferedDocs> -->
+
+        <!-- Expert: ramPerThreadHardLimitMB sets the maximum amount of RAM 
that can be consumed
+             per thread before they are flushed. When limit is exceeded, this 
triggers a forced
+             flush even if ramBufferSizeMB has not been exceeded.
+             This is a safety limit to prevent Lucene's 
DocumentsWriterPerThread from address space
+             exhaustion due to its internal 32 bit signed integer based memory 
addressing.
+             The specified value should be greater than 0 and less than 
2048MB. When not specified,
+             Solr uses Lucene's default value 1945. -->
+        <!-- <ramPerThreadHardLimitMB>1945</ramPerThreadHardLimitMB> -->
+
+        <!-- Expert: Merge Policy
+             The Merge Policy in Lucene controls how merging of segments is 
done.
+             The default since Solr/Lucene 3.3 is TieredMergePolicy.
+             The default since Lucene 2.3 was the LogByteSizeMergePolicy,
+             Even older versions of Lucene used LogDocMergePolicy.
+          -->
+        <!--
+            <mergePolicyFactory 
class="org.apache.solr.index.TieredMergePolicyFactory">
+              <int name="maxMergeAtOnce">10</int>
+              <int name="segmentsPerTier">10</int>
+              <double name="noCFSRatio">0.1</double>
+            </mergePolicyFactory>
+          -->
+
+        <!-- Expert: Merge Scheduler
+             The Merge Scheduler in Lucene controls how merges are
+             performed.  The ConcurrentMergeScheduler (Lucene 2.3 default)
+             can perform merges in the background using separate threads.
+             The SerialMergeScheduler (Lucene 2.2 default) does not.
+         -->
+        <!--
+           <mergeScheduler 
class="org.apache.lucene.index.ConcurrentMergeScheduler"/>
+           -->
+
+        <!-- LockFactory
+             This option specifies which Lucene LockFactory implementation
+             to use.
+             single = SingleInstanceLockFactory - suggested for a
+                      read-only index or when there is no possibility of
+                      another process trying to modify the index.
+             native = NativeFSLockFactory - uses OS native file locking.
+                      Do not use when multiple solr webapps in the same
+                      JVM are attempting to share a single index.
+             simple = SimpleFSLockFactory  - uses a plain file for locking
+             Defaults: 'native' is default for Solr3.6 and later, otherwise
+                       'simple' is the default
+             More details on the nuances of each LockFactory...
+             http://wiki.apache.org/lucene-java/AvailableLockFactories
+        -->
+        <lockType>${solr.lock.type:native}</lockType>
+
+        <!-- Commit Deletion Policy
+             Custom deletion policies can be specified here. The class must
+             implement org.apache.lucene.index.IndexDeletionPolicy.
+             The default Solr IndexDeletionPolicy implementation supports
+             deleting index commit points on number of commits, age of
+             commit point and optimized status.
+             The latest commit point should always be preserved regardless
+             of the criteria.
+        -->
+        <!--
+        <deletionPolicy class="solr.SolrDeletionPolicy">
+        -->
+        <!-- The number of commit points to be kept -->
+        <!-- <str name="maxCommitsToKeep">1</str> -->
+        <!-- The number of optimized commit points to be kept -->
+        <!-- <str name="maxOptimizedCommitsToKeep">0</str> -->
+        <!--
+            Delete all commit points once they have reached the given age.
+            Supports DateMathParser syntax e.g.
+          -->
+        <!--
+           <str name="maxCommitAge">30MINUTES</str>
+           <str name="maxCommitAge">1DAY</str>
+        -->
+        <!--
+        </deletionPolicy>
+        -->
+
+        <!-- Lucene Infostream
+             To aid in advanced debugging, Lucene provides an "InfoStream"
+             of detailed information when indexing.
+             Setting The value to true will instruct the underlying Lucene
+             IndexWriter to write its debugging info the specified file
+          -->
+        <!-- <infoStream file="INFOSTREAM.txt">false</infoStream> -->
+    </indexConfig>
+
+
+    <!-- JMX
+         This example enables JMX if and only if an existing MBeanServer
+         is found, use this if you want to configure JMX through JVM
+         parameters. Remove this to disable exposing Solr configuration
+         and statistics to JMX.
+         For more details see http://wiki.apache.org/solr/SolrJmx
+      -->
+    <jmx />
+    <!-- If you want to connect to a particular server, specify the
+         agentId
+      -->
+    <!-- <jmx agentId="myAgent" /> -->
+    <!-- If you want to start a new MBeanServer, specify the serviceUrl -->
+    <!-- <jmx serviceUrl="service:jmx:rmi:///jndi/rmi://localhost:9999/solr"/>
+      -->
+
+    <!-- The default high-performance update handler -->
+    <updateHandler class="solr.DirectUpdateHandler2">
+
+        <!-- Enables a transaction log, used for real-time get, durability, and
+             and solr cloud replica recovery.  The log can grow as big as
+             uncommitted changes to the index, so use of a hard autoCommit
+             is recommended (see below).
+             "dir" - the target directory for transaction logs, defaults to the
+                    solr data directory.
+             "numVersionBuckets" - sets the number of buckets used to keep
+                    track of max version values when checking for re-ordered
+                    updates; increase this value to reduce the cost of
+                    synchronizing access to version buckets during high-volume
+                    indexing, this requires 8 bytes (long) * numVersionBuckets
+                    of heap space per Solr core.
+        -->
+        <updateLog>
+            <str name="dir">${solr.ulog.dir:}</str>
+            <int 
name="tlogDfsReplication">${solr.ulog.tlogDfsReplication:3}</int>
+            <int 
name="numVersionBuckets">${solr.ulog.numVersionBuckets:65536}</int>
+        </updateLog>
+
+        <!-- AutoCommit
+             Perform a hard commit automatically under certain conditions.
+             Instead of enabling autoCommit, consider using "commitWithin"
+             when adding documents.
+             http://wiki.apache.org/solr/UpdateXmlMessages
+             maxDocs - Maximum number of documents to add since the last
+                       commit before automatically triggering a new commit.
+             maxTime - Maximum amount of time in ms that is allowed to pass
+                       since a document was added before automatically
+                       triggering a new commit.
+             openSearcher - if false, the commit causes recent index changes
+               to be flushed to stable storage, but does not cause a new
+               searcher to be opened to make those changes visible.
+             If the updateLog is enabled, then it's highly recommended to
+             have some sort of hard autoCommit to limit the log size.
+          -->
+        <autoCommit>
+            <maxTime>${solr.autoCommit.maxTime:60000}</maxTime>
+            <openSearcher>false</openSearcher>
+        </autoCommit>
+
+        <!-- softAutoCommit is like autoCommit except it causes a
+             'soft' commit which only ensures that changes are visible
+             but does not ensure that data is synced to disk.  This is
+             faster and more near-realtime friendly than a hard commit.
+          -->
+
+        <autoSoftCommit>
+            <maxTime>${solr.autoSoftCommit.maxTime:15000}</maxTime>
+        </autoSoftCommit>
+
+        <!-- Update Related Event Listeners
+             Various IndexWriter related events can trigger Listeners to
+             take actions.
+             postCommit - fired after every commit or optimize command
+             postOptimize - fired after every optimize command
+          -->
+        <!-- The RunExecutableListener executes an external command from a
+             hook such as postCommit or postOptimize.
+             exe - the name of the executable to run
+             dir - dir to use as the current working directory. (default=".")
+             wait - the calling thread waits until the executable returns.
+                    (default="true")
+             args - the arguments to pass to the program.  (default is none)
+             env - environment variables to set.  (default is none)
+          -->
+        <!-- This example shows how RunExecutableListener could be used
+             with the script based replication...
+             http://wiki.apache.org/solr/CollectionDistribution
+          -->
+        <!--
+           <listener event="postCommit" class="solr.RunExecutableListener">
+             <str name="exe">solr/bin/snapshooter</str>
+             <str name="dir">.</str>
+             <bool name="wait">true</bool>
+             <arr name="args"> <str>arg1</str> <str>arg2</str> </arr>
+             <arr name="env"> <str>MYVAR=val1</str> </arr>
+           </listener>
+          -->
+
+    </updateHandler>
+
+    <!-- IndexReaderFactory
+         Use the following format to specify a custom IndexReaderFactory,
+         which allows for alternate IndexReader implementations.
+         ** Experimental Feature **
+         Please note - Using a custom IndexReaderFactory may prevent
+         certain other features from working. The API to
+         IndexReaderFactory may change without warning or may even be
+         removed from future releases if the problems cannot be
+         resolved.
+         ** Features that may not work with custom IndexReaderFactory **
+         The ReplicationHandler assumes a disk-resident index. Using a
+         custom IndexReader implementation may cause incompatibility
+         with ReplicationHandler and may cause replication to not work
+         correctly. See SOLR-1366 for details.
+      -->
+    <!--
+    <indexReaderFactory name="IndexReaderFactory" class="package.class">
+      <str name="someArg">Some Value</str>
+    </indexReaderFactory >
+    -->
+
+    <!-- ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
+         Query section - these settings control query time things like caches
+         ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 
-->
+    <query>
+
+        <!-- Maximum number of clauses allowed when parsing a boolean query 
string.
+
+             This limit only impacts boolean queries specified by a user as 
part of a query string,
+             and provides per-collection controls on how complex user 
specified boolean queries can
+             be.  Query strings that specify more clauses then this will 
result in an error.
+
+             If this per-collection limit is greater then the global 
`maxBooleanClauses` limit
+             specified in `solr.xml`, it will have no effect, as that setting 
also limits the size
+             of user specified boolean queries.
+          -->
+        <maxBooleanClauses>${solr.max.booleanClauses:1024}</maxBooleanClauses>
+
+        <!-- Solr Internal Query Caches
+             There are two implementations of cache available for Solr,
+             LRUCache, based on a synchronized LinkedHashMap, and
+             FastLRUCache, based on a ConcurrentHashMap.
+             FastLRUCache has faster gets and slower puts in single
+             threaded operation and thus is generally faster than LRUCache
+             when the hit ratio of the cache is high (> 75%), and may be
+             faster under other scenarios on multi-cpu systems.
+        -->
+
+        <!-- Filter Cache
+             Cache used by SolrIndexSearcher for filters (DocSets),
+             unordered sets of *all* documents that match a query.  When a
+             new searcher is opened, its caches may be prepopulated or
+             "autowarmed" using data from caches in the old searcher.
+             autowarmCount is the number of items to prepopulate.  For
+             LRUCache, the autowarmed items will be the most recently
+             accessed items.
+             Parameters:
+               class - the SolrCache implementation LRUCache or
+                   (LRUCache or FastLRUCache)
+               size - the maximum number of entries in the cache
+               initialSize - the initial capacity (number of entries) of
+                   the cache.  (see java.util.HashMap)
+               autowarmCount - the number of entries to prepopulate from
+                   and old cache.
+               maxRamMB - the maximum amount of RAM (in MB) that this cache is 
allowed
+                          to occupy. Note that when this option is specified, 
the size
+                          and initialSize parameters are ignored.
+          -->
+        <filterCache class="solr.FastLRUCache"
+                     size="512"
+                     initialSize="512"
+                     autowarmCount="0"/>
+
+        <!-- Query Result Cache
+             Caches results of searches - ordered lists of document ids
+             (DocList) based on a query, a sort, and the range of documents 
requested.
+             Additional supported parameter by LRUCache:
+                maxRamMB - the maximum amount of RAM (in MB) that this cache 
is allowed
+                           to occupy
+          -->
+        <queryResultCache class="solr.LRUCache"
+                          size="512"
+                          initialSize="512"
+                          autowarmCount="0"/>
+
+        <!-- Document Cache
+             Caches Lucene Document objects (the stored fields for each
+             document).  Since Lucene internal document ids are transient,
+             this cache will not be autowarmed.
+          -->
+        <documentCache class="solr.LRUCache"
+                       size="512"
+                       initialSize="512"
+                       autowarmCount="0"/>
+
+        <!-- custom cache currently used by block join -->
+        <cache name="perSegFilter"
+               class="solr.search.LRUCache"
+               size="10"
+               initialSize="0"
+               autowarmCount="10"
+               regenerator="solr.NoOpRegenerator" />
+
+        <!-- Field Value Cache
+             Cache used to hold field values that are quickly accessible
+             by document id.  The fieldValueCache is created by default
+             even if not configured here.
+          -->
+        <!--
+           <fieldValueCache class="solr.FastLRUCache"
+                            size="512"
+                            autowarmCount="128"
+                            showItems="32" />
+          -->
+
+        <!-- Custom Cache
+             Example of a generic cache.  These caches may be accessed by
+             name through SolrIndexSearcher.getCache(),cacheLookup(), and
+             cacheInsert().  The purpose is to enable easy caching of
+             user/application level data.  The regenerator argument should
+             be specified as an implementation of solr.CacheRegenerator
+             if autowarming is desired.
+          -->
+        <!--
+           <cache name="myUserCache"
+                  class="solr.LRUCache"
+                  size="4096"
+                  initialSize="1024"
+                  autowarmCount="1024"
+                  regenerator="com.mycompany.MyRegenerator"
+                  />
+          -->
+
+
+        <!-- Lazy Field Loading
+             If true, stored fields that are not requested will be loaded
+             lazily.  This can result in a significant speed improvement
+             if the usual case is to not load all stored fields,
+             especially if the skipped fields are large compressed text
+             fields.
+        -->
+        <enableLazyFieldLoading>true</enableLazyFieldLoading>
+
+        <!-- Use Filter For Sorted Query
+             A possible optimization that attempts to use a filter to
+             satisfy a search.  If the requested sort does not include
+             score, then the filterCache will be checked for a filter
+             matching the query. If found, the filter will be used as the
+             source of document ids, and then the sort will be applied to
+             that.
+             For most situations, this will not be useful unless you
+             frequently get the same search repeatedly with different sort
+             options, and none of them ever use "score"
+          -->
+        <!--
+           <useFilterForSortedQuery>true</useFilterForSortedQuery>
+          -->
+
+        <!-- Result Window Size
+             An optimization for use with the queryResultCache.  When a search
+             is requested, a superset of the requested number of document ids
+             are collected.  For example, if a search for a particular query
+             requests matching documents 10 through 19, and queryWindowSize is 
50,
+             then documents 0 through 49 will be collected and cached.  Any 
further
+             requests in that range can be satisfied via the cache.
+          -->
+        <queryResultWindowSize>20</queryResultWindowSize>
+
+        <!-- Maximum number of documents to cache for any entry in the
+             queryResultCache.
+          -->
+        <queryResultMaxDocsCached>200</queryResultMaxDocsCached>
+
+        <!-- Query Related Event Listeners
+             Various IndexSearcher related events can trigger Listeners to
+             take actions.
+             newSearcher - fired whenever a new searcher is being prepared
+             and there is a current searcher handling requests (aka
+             registered).  It can be used to prime certain caches to
+             prevent long request times for certain requests.
+             firstSearcher - fired whenever a new searcher is being
+             prepared but there is no current registered searcher to handle
+             requests or to gain autowarming data from.
+          -->
+        <!-- QuerySenderListener takes an array of NamedList and executes a
+             local query request for each NamedList in sequence.
+          -->
+        <listener event="newSearcher" class="solr.QuerySenderListener">
+            <arr name="queries">
+                <!--
+                   <lst><str name="q">solr</str><str name="sort">price 
asc</str></lst>
+                   <lst><str name="q">rocks</str><str name="sort">weight 
asc</str></lst>
+                  -->
+            </arr>
+        </listener>
+        <listener event="firstSearcher" class="solr.QuerySenderListener">
+            <arr name="queries">
+                <!--
+                <lst>
+                  <str name="q">static firstSearcher warming in 
solrconfig.xml</str>
+                </lst>
+                -->
+            </arr>
+        </listener>
+
+        <!-- Use Cold Searcher
+             If a search request comes in and there is no current
+             registered searcher, then immediately register the still
+             warming searcher and use it.  If "false" then all requests
+             will block until the first searcher is done warming.
+          -->
+        <useColdSearcher>false</useColdSearcher>
+
+        <!-- Slow Query Request Logging
+             Any queries that take longer than the specified threshold
+             will be logged as "slow" queries.
+             To disable slow request logging for this Solr config,
+             set the value to -1
+          -->
+        <slowQueryThresholdMillis>5000</slowQueryThresholdMillis>
+
+    </query>
+
+
+    <!-- Request Dispatcher
+         This section contains instructions for how the SolrDispatchFilter
+         should behave when processing requests for this SolrCore.
+      -->
+    <requestDispatcher>
+        <!-- Request Parsing
+             These settings indicate how Solr Requests may be parsed, and
+             what restrictions may be placed on the ContentStreams from
+             those requests
+             enableRemoteStreaming - enables use of the stream.file
+             and stream.url parameters for specifying remote streams.
+             multipartUploadLimitInKB - specifies the max size (in KiB) of
+             Multipart File Uploads that Solr will allow in a Request.
+             formdataUploadLimitInKB - specifies the max size (in KiB) of
+             form data (application/x-www-form-urlencoded) sent via
+             POST. You can use POST to pass request parameters not
+             fitting into the URL.
+             addHttpRequestToContext - if set to true, it will instruct
+             the requestParsers to include the original HttpServletRequest
+             object in the context map of the SolrQueryRequest under the
+             key "httpRequest". It will not be used by any of the existing
+             Solr components, but may be useful when developing custom
+             plugins.
+             *** WARNING ***
+             Before enabling remote streaming, you should make sure your
+             system has authentication enabled.
+          -->
+        <requestParsers enableRemoteStreaming="true"
+                        multipartUploadLimitInKB="2048000"
+                        formdataUploadLimitInKB="2048"
+                        addHttpRequestToContext="true"/>
+
+        <!-- HTTP Caching
+             Set HTTP caching related parameters (for proxy caches and 
clients).
+             The options below instruct Solr not to output any HTTP Caching
+             related headers
+          -->
+        <httpCaching never304="true" />
+        <!-- If you include a <cacheControl> directive, it will be used to
+             generate a Cache-Control header (as well as an Expires header
+             if the value contains "max-age=")
+             By default, no Cache-Control header is generated.
+             You can use the <cacheControl> option even if you have set
+             never304="true"
+          -->
+        <!--
+           <httpCaching never304="true" >
+             <cacheControl>max-age=30, public</cacheControl>
+           </httpCaching>
+          -->
+        <!-- To enable Solr to respond with automatically generated HTTP
+             Caching headers, and to response to Cache Validation requests
+             correctly, set the value of never304="false"
+             This will cause Solr to generate Last-Modified and ETag
+             headers based on the properties of the Index.
+             The following options can also be specified to affect the
+             values of these headers...
+             lastModFrom - the default value is "openTime" which means the
+             Last-Modified value (and validation against If-Modified-Since
+             requests) will all be relative to when the current Searcher
+             was opened.  You can change it to lastModFrom="dirLastMod" if
+             you want the value to exactly correspond to when the physical
+             index was last modified.
+             etagSeed="..." is an option you can change to force the ETag
+             header (and validation against If-None-Match requests) to be
+             different even if the index has not changed (ie: when making
+             significant changes to your config file)
+             (lastModifiedFrom and etagSeed are both ignored if you use
+             the never304="true" option)
+          -->
+        <!--
+           <httpCaching lastModifiedFrom="openTime"
+                        etagSeed="Solr">
+             <cacheControl>max-age=30, public</cacheControl>
+           </httpCaching>
+          -->
+    </requestDispatcher>
+
+    <!-- Request Handlers
+         http://wiki.apache.org/solr/SolrRequestHandler
+         Incoming queries will be dispatched to a specific handler by name
+         based on the path specified in the request.
+         If a Request Handler is declared with startup="lazy", then it will
+         not be initialized until the first request that uses it.
+      -->
+    <!-- SearchHandler
+         http://wiki.apache.org/solr/SearchHandler
+         For processing Search Queries, the primary Request Handler
+         provided with Solr is "SearchHandler" It delegates to a sequent
+         of SearchComponents (see below) and supports distributed
+         queries across multiple shards
+      -->
+    <requestHandler name="/select" class="solr.SearchHandler">
+        <!-- default values for query parameters can be specified, these
+             will be overridden by parameters in the request
+          -->
+        <lst name="defaults">
+            <str name="echoParams">explicit</str>
+            <int name="rows">10</int>
+            <!-- Default search field
+               <str name="df">text</str>
+              -->
+            <!-- Change from JSON to XML format (the default prior to Solr 7.0)
+               <str name="wt">xml</str>
+              -->
+        </lst>
+        <!-- In addition to defaults, "appends" params can be specified
+             to identify values which should be appended to the list of
+             multi-val params from the query (or the existing "defaults").
+          -->
+        <!-- In this example, the param "fq=instock:true" would be appended to
+             any query time fq params the user may specify, as a mechanism for
+             partitioning the index, independent of any user selected filtering
+             that may also be desired (perhaps as a result of faceted 
searching).
+             NOTE: there is *absolutely* nothing a client can do to prevent 
these
+             "appends" values from being used, so don't use this mechanism
+             unless you are sure you always want it.
+          -->
+        <!--
+           <lst name="appends">
+             <str name="fq">inStock:true</str>
+           </lst>
+          -->
+        <!-- "invariants" are a way of letting the Solr maintainer lock down
+             the options available to Solr clients.  Any params values
+             specified here are used regardless of what values may be specified
+             in either the query, the "defaults", or the "appends" params.
+             In this example, the facet.field and facet.query params would
+             be fixed, limiting the facets clients can use.  Faceting is
+             not turned on by default - but if the client does specify
+             facet=true in the request, these are the only facets they
+             will be able to see counts for; regardless of what other
+             facet.field or facet.query params they may specify.
+             NOTE: there is *absolutely* nothing a client can do to prevent 
these
+             "invariants" values from being used, so don't use this mechanism
+             unless you are sure you always want it.
+          -->
+        <!--
+           <lst name="invariants">
+             <str name="facet.field">cat</str>
+             <str name="facet.field">manu_exact</str>
+             <str name="facet.query">price:[* TO 500]</str>
+             <str name="facet.query">price:[500 TO *]</str>
+           </lst>
+          -->
+        <!-- If the default list of SearchComponents is not desired, that
+             list can either be overridden completely, or components can be
+             prepended or appended to the default list.  (see below)
+          -->
+        <!--
+           <arr name="components">
+             <str>nameOfCustomComponent1</str>
+             <str>nameOfCustomComponent2</str>
+           </arr>
+          -->
+    </requestHandler>
+
+    <!-- A request handler that returns indented JSON by default -->
+    <requestHandler name="/query" class="solr.SearchHandler">
+        <lst name="defaults">
+            <str name="echoParams">explicit</str>
+            <str name="wt">json</str>
+            <str name="indent">true</str>
+        </lst>
+    </requestHandler>
+
+    <initParams path="/update/**,/query,/select,/spell">
+        <lst name="defaults">
+            <str name="df">_text_</str>
+        </lst>
+    </initParams>
+
+    <!-- Search Components
+         Search components are registered to SolrCore and used by
+         instances of SearchHandler (which can access them by name)
+         By default, the following components are available:
+         <searchComponent name="query"     class="solr.QueryComponent" />
+         <searchComponent name="facet"     class="solr.FacetComponent" />
+         <searchComponent name="mlt"       class="solr.MoreLikeThisComponent" 
/>
+         <searchComponent name="highlight" class="solr.HighlightComponent" />
+         <searchComponent name="stats"     class="solr.StatsComponent" />
+         <searchComponent name="debug"     class="solr.DebugComponent" />
+         Default configuration in a requestHandler would look like:
+         <arr name="components">
+           <str>query</str>
+           <str>facet</str>
+           <str>mlt</str>
+           <str>highlight</str>
+           <str>stats</str>
+           <str>debug</str>
+         </arr>
+         If you register a searchComponent to one of the standard names,
+         that will be used instead of the default.
+         To insert components before or after the 'standard' components, use:
+         <arr name="first-components">
+           <str>myFirstComponentName</str>
+         </arr>
+         <arr name="last-components">
+           <str>myLastComponentName</str>
+         </arr>
+         NOTE: The component registered with the name "debug" will
+         always be executed after the "last-components"
+       -->
+
+    <!-- Spell Check
+         The spell check component can return a list of alternative spelling
+         suggestions.
+         http://wiki.apache.org/solr/SpellCheckComponent
+      -->
+    <searchComponent name="spellcheck" class="solr.SpellCheckComponent">
+
+        <str name="queryAnalyzerFieldType">text_general</str>
+
+        <!-- Multiple "Spell Checkers" can be declared and used by this
+             component
+          -->
+
+        <!-- a spellchecker built from a field of the main index -->
+        <lst name="spellchecker">
+            <str name="name">default</str>
+            <str name="field">_text_</str>
+            <str name="classname">solr.DirectSolrSpellChecker</str>
+            <!-- the spellcheck distance measure used, the default is the 
internal levenshtein -->
+            <str name="distanceMeasure">internal</str>
+            <!-- minimum accuracy needed to be considered a valid spellcheck 
suggestion -->
+            <float name="accuracy">0.5</float>
+            <!-- the maximum #edits we consider when enumerating terms: can be 
1 or 2 -->
+            <int name="maxEdits">2</int>
+            <!-- the minimum shared prefix when enumerating terms -->
+            <int name="minPrefix">1</int>
+            <!-- maximum number of inspections per result. -->
+            <int name="maxInspections">5</int>
+            <!-- minimum length of a query term to be considered for 
correction -->
+            <int name="minQueryLength">4</int>
+            <!-- maximum threshold of documents a query term can appear to be 
considered for correction -->
+            <float name="maxQueryFrequency">0.01</float>
+            <!-- uncomment this to require suggestions to occur in 1% of the 
documents
+              <float name="thresholdTokenFrequency">.01</float>
+            -->
+        </lst>
+
+        <!-- a spellchecker that can break or combine words.  See "/spell" 
handler below for usage -->
+        <!--
+        <lst name="spellchecker">
+          <str name="name">wordbreak</str>
+          <str name="classname">solr.WordBreakSolrSpellChecker</str>
+          <str name="field">name</str>
+          <str name="combineWords">true</str>
+          <str name="breakWords">true</str>
+          <int name="maxChanges">10</int>
+        </lst>
+        -->
+    </searchComponent>
+
+    <!-- A request handler for demonstrating the spellcheck component.
+         NOTE: This is purely as an example.  The whole purpose of the
+         SpellCheckComponent is to hook it into the request handler that
+         handles your normal user queries so that a separate request is
+         not needed to get suggestions.
+         IN OTHER WORDS, THERE IS REALLY GOOD CHANCE THE SETUP BELOW IS
+         NOT WHAT YOU WANT FOR YOUR PRODUCTION SYSTEM!
+         See http://wiki.apache.org/solr/SpellCheckComponent for details
+         on the request parameters.
+      -->
+    <requestHandler name="/spell" class="solr.SearchHandler" startup="lazy">
+        <lst name="defaults">
+            <!-- Solr will use suggestions from both the 'default' spellchecker
+                 and from the 'wordbreak' spellchecker and combine them.
+                 collations (re-written queries) can include a combination of
+                 corrections from both spellcheckers -->
+            <str name="spellcheck.dictionary">default</str>
+            <str name="spellcheck">on</str>
+            <str name="spellcheck.extendedResults">true</str>
+            <str name="spellcheck.count">10</str>
+            <str name="spellcheck.alternativeTermCount">5</str>
+            <str name="spellcheck.maxResultsForSuggest">5</str>
+            <str name="spellcheck.collate">true</str>
+            <str name="spellcheck.collateExtendedResults">true</str>
+            <str name="spellcheck.maxCollationTries">10</str>
+            <str name="spellcheck.maxCollations">5</str>
+        </lst>
+        <arr name="last-components">
+            <str>spellcheck</str>
+        </arr>
+    </requestHandler>
+
+    <!-- Terms Component
+         http://wiki.apache.org/solr/TermsComponent
+         A component to return terms and document frequency of those
+         terms
+      -->
+    <searchComponent name="terms" class="solr.TermsComponent"/>
+
+    <!-- A request handler for demonstrating the terms component -->
+    <requestHandler name="/terms" class="solr.SearchHandler" startup="lazy">
+        <lst name="defaults">
+            <bool name="terms">true</bool>
+            <bool name="distrib">false</bool>
+        </lst>
+        <arr name="components">
+            <str>terms</str>
+        </arr>
+    </requestHandler>
+
+    <!-- Highlighting Component
+         http://wiki.apache.org/solr/HighlightingParameters
+      -->
+    <searchComponent class="solr.HighlightComponent" name="highlight">
+        <highlighting>
+            <!-- Configure the standard fragmenter -->
+            <!-- This could most likely be commented out in the "default" case 
-->
+            <fragmenter name="gap"
+                        default="true"
+                        class="solr.highlight.GapFragmenter">
+                <lst name="defaults">
+                    <int name="hl.fragsize">100</int>
+                </lst>
+            </fragmenter>
+
+            <!-- A regular-expression-based fragmenter
+                 (for sentence extraction)
+              -->
+            <fragmenter name="regex"
+                        class="solr.highlight.RegexFragmenter">
+                <lst name="defaults">
+                    <!-- slightly smaller fragsizes work better because of 
slop -->
+                    <int name="hl.fragsize">70</int>
+                    <!-- allow 50% slop on fragment sizes -->
+                    <float name="hl.regex.slop">0.5</float>
+                    <!-- a basic sentence pattern -->
+                    <str name="hl.regex.pattern">[-\w 
,/\n\&quot;&apos;]{20,200}</str>
+                </lst>
+            </fragmenter>
+
+            <!-- Configure the standard formatter -->
+            <formatter name="html"
+                       default="true"
+                       class="solr.highlight.HtmlFormatter">
+                <lst name="defaults">
+                    <str name="hl.simple.pre"><![CDATA[<em>]]></str>
+                    <str name="hl.simple.post"><![CDATA[</em>]]></str>
+                </lst>
+            </formatter>
+
+            <!-- Configure the standard encoder -->
+            <encoder name="html"
+                     class="solr.highlight.HtmlEncoder" />
+
+            <!-- Configure the standard fragListBuilder -->
+            <fragListBuilder name="simple"
+                             class="solr.highlight.SimpleFragListBuilder"/>
+
+            <!-- Configure the single fragListBuilder -->
+            <fragListBuilder name="single"
+                             class="solr.highlight.SingleFragListBuilder"/>
+
+            <!-- Configure the weighted fragListBuilder -->
+            <fragListBuilder name="weighted"
+                             default="true"
+                             class="solr.highlight.WeightedFragListBuilder"/>
+
+            <!-- default tag FragmentsBuilder -->
+            <fragmentsBuilder name="default"
+                              default="true"
+                              
class="solr.highlight.ScoreOrderFragmentsBuilder">
+                <!--
+                <lst name="defaults">
+                  <str name="hl.multiValuedSeparatorChar">/</str>
+                </lst>
+                -->
+            </fragmentsBuilder>
+
+            <!-- multi-colored tag FragmentsBuilder -->
+            <fragmentsBuilder name="colored"
+                              
class="solr.highlight.ScoreOrderFragmentsBuilder">
+                <lst name="defaults">
+                    <str name="hl.tag.pre"><![CDATA[
+               <b style="background:yellow">,<b style="background:lawgreen">,
+               <b style="background:aquamarine">,<b 
style="background:magenta">,
+               <b style="background:palegreen">,<b style="background:coral">,
+               <b style="background:wheat">,<b style="background:khaki">,
+               <b style="background:lime">,<b 
style="background:deepskyblue">]]></str>
+                    <str name="hl.tag.post"><![CDATA[</b>]]></str>
+                </lst>
+            </fragmentsBuilder>
+
+            <boundaryScanner name="default"
+                             default="true"
+                             class="solr.highlight.SimpleBoundaryScanner">
+                <lst name="defaults">
+                    <str name="hl.bs.maxScan">10</str>
+                    <str name="hl.bs.chars">.,!? &#9;&#10;&#13;</str>
+                </lst>
+            </boundaryScanner>
+
+            <boundaryScanner name="breakIterator"
+                             
class="solr.highlight.BreakIteratorBoundaryScanner">
+                <lst name="defaults">
+                    <!-- type should be one of CHARACTER, WORD(default), LINE 
and SENTENCE -->
+                    <str name="hl.bs.type">WORD</str>
+                    <!-- language and country are used when constructing 
Locale object.  -->
+                    <!-- And the Locale object will be used when getting 
instance of BreakIterator -->
+                    <str name="hl.bs.language">en</str>
+                    <str name="hl.bs.country">US</str>
+                </lst>
+            </boundaryScanner>
+        </highlighting>
+    </searchComponent>
+
+    <!-- Update Processors
+         Chains of Update Processor Factories for dealing with Update
+         Requests can be declared, and then used by name in Update
+         Request Processors
+         http://wiki.apache.org/solr/UpdateRequestProcessor
+      -->
+
+    <!-- Add unknown fields to the schema
+         Field type guessing update processors that will
+         attempt to parse string-typed field values as Booleans, Longs,
+         Doubles, or Dates, and then add schema fields with the guessed
+         field types. Text content will be indexed as "text_general" as
+         well as a copy to a plain string version in *_str.
+         These require that the schema is both managed and mutable, by
+         declaring schemaFactory as ManagedIndexSchemaFactory, with
+         mutable specified as true.
+         See http://wiki.apache.org/solr/GuessingFieldTypes
+      -->
+    <updateProcessor class="solr.UUIDUpdateProcessorFactory" name="uuid"/>
+    <updateProcessor class="solr.RemoveBlankFieldUpdateProcessorFactory" 
name="remove-blank"/>
+    <updateProcessor class="solr.FieldNameMutatingUpdateProcessorFactory" 
name="field-name-mutating">
+        <str name="pattern">[^\w-\.]</str>
+        <str name="replacement">_</str>
+    </updateProcessor>
+    <updateProcessor class="solr.ParseBooleanFieldUpdateProcessorFactory" 
name="parse-boolean"/>
+    <updateProcessor class="solr.ParseLongFieldUpdateProcessorFactory" 
name="parse-long"/>
+    <updateProcessor class="solr.ParseDoubleFieldUpdateProcessorFactory" 
name="parse-double"/>
+    <updateProcessor class="solr.ParseDateFieldUpdateProcessorFactory" 
name="parse-date">
+        <arr name="format">
+            <str>yyyy-MM-dd['T'[HH:mm[:ss[.SSS]][z</str>
+            <str>yyyy-MM-dd['T'[HH:mm[:ss[,SSS]][z</str>
+            <str>yyyy-MM-dd HH:mm[:ss[.SSS]][z</str>
+            <str>yyyy-MM-dd HH:mm[:ss[,SSS]][z</str>
+            <str>[EEE, ]dd MMM yyyy HH:mm[:ss] z</str>
+            <str>EEEE, dd-MMM-yy HH:mm:ss z</str>
+            <str>EEE MMM ppd HH:mm:ss [z ]yyyy</str>
+        </arr>
+    </updateProcessor>
+    <updateProcessor class="solr.AddSchemaFieldsUpdateProcessorFactory" 
name="add-schema-fields">
+        <str name="defaultFieldType">key_lower_case</str>
+        <lst name="typeMapping">
+            <str name="valueClass">java.lang.Boolean</str>
+            <str name="fieldType">boolean</str>
+        </lst>
+        <lst name="typeMapping">
+            <str name="valueClass">java.util.Date</str>
+            <str name="fieldType">tdate</str>
+        </lst>
+        <lst name="typeMapping">
+            <str name="valueClass">java.lang.Long</str>
+            <str name="valueClass">java.lang.Integer</str>
+            <str name="fieldType">tlong</str>
+        </lst>
+        <lst name="typeMapping">
+            <str name="valueClass">java.lang.Number</str>
+            <str name="fieldType">tdouble</str>
+        </lst>
+    </updateProcessor>
+
+    <!-- The update.autoCreateFields property can be turned to false to 
disable schemaless mode -->
+    <updateRequestProcessorChain name="add-unknown-fields-to-the-schema" 
default="${update.autoCreateFields:true}"
+                                 
processor="uuid,remove-blank,field-name-mutating,parse-boolean,parse-long,parse-double,parse-date,add-schema-fields">
+        <processor class="solr.LogUpdateProcessorFactory"/>
+        <processor class="solr.DistributedUpdateProcessorFactory"/>
+        <processor class="solr.RunUpdateProcessorFactory"/>
+    </updateRequestProcessorChain>
+
+    <!-- Deduplication
+         An example dedup update processor that creates the "id" field
+         on the fly based on the hash code of some other fields.  This
+         example has overwriteDupes set to false since we are using the
+         id field as the signatureField and Solr will maintain
+         uniqueness based on that anyway.
+      -->
+    <!--
+       <updateRequestProcessorChain name="dedupe">
+         <processor class="solr.processor.SignatureUpdateProcessorFactory">
+           <bool name="enabled">true</bool>
+           <str name="signatureField">id</str>
+           <bool name="overwriteDupes">false</bool>
+           <str name="fields">name,features,cat</str>
+           <str name="signatureClass">solr.processor.Lookup3Signature</str>
+         </processor>
+         <processor class="solr.LogUpdateProcessorFactory" />
+         <processor class="solr.RunUpdateProcessorFactory" />
+       </updateRequestProcessorChain>
+      -->
+
+    <!-- Response Writers
+         http://wiki.apache.org/solr/QueryResponseWriter
+         Request responses will be written using the writer specified by
+         the 'wt' request parameter matching the name of a registered
+         writer.
+         The "default" writer is the default and will be used if 'wt' is
+         not specified in the request.
+      -->
+    <!-- The following response writers are implicitly configured unless
+         overridden...
+      -->
+    <!--
+       <queryResponseWriter name="xml"
+                            default="true"
+                            class="solr.XMLResponseWriter" />
+       <queryResponseWriter name="json" class="solr.JSONResponseWriter"/>
+       <queryResponseWriter name="python" class="solr.PythonResponseWriter"/>
+       <queryResponseWriter name="ruby" class="solr.RubyResponseWriter"/>
+       <queryResponseWriter name="php" class="solr.PHPResponseWriter"/>
+       <queryResponseWriter name="phps" 
class="solr.PHPSerializedResponseWriter"/>
+       <queryResponseWriter name="csv" class="solr.CSVResponseWriter"/>
+       <queryResponseWriter name="schema.xml" 
class="solr.SchemaXmlResponseWriter"/>
+      -->
+
+    <queryResponseWriter name="json" class="solr.JSONResponseWriter">
+        <!-- For the purposes of the tutorial, JSON responses are written as
+         plain text so that they are easy to read in *any* browser.
+         If you expect a MIME type of "application/json" just remove this 
override.
+        -->
+        <str name="content-type">text/plain; charset=UTF-8</str>
+    </queryResponseWriter>
+
+    <!-- Query Parsers
+         https://lucene.apache.org/solr/guide/query-syntax-and-parsing.html
+         Multiple QParserPlugins can be registered by name, and then
+         used in either the "defType" param for the QueryComponent (used
+         by SearchHandler) or in LocalParams
+      -->
+    <!-- example of registering a query parser -->
+    <!--
+       <queryParser name="myparser" class="com.mycompany.MyQParserPlugin"/>
+      -->
+
+    <!-- Function Parsers
+         http://wiki.apache.org/solr/FunctionQuery
+         Multiple ValueSourceParsers can be registered by name, and then
+         used as function names when using the "func" QParser.
+      -->
+    <!-- example of registering a custom function parser  -->
+    <!--
+       <valueSourceParser name="myfunc"
+                          class="com.mycompany.MyValueSourceParser" />
+      -->
+
+
+    <!-- Document Transformers
+         http://wiki.apache.org/solr/DocTransformers
+      -->
+    <!--
+       Could be something like:
+       <transformer name="db" 
class="com.mycompany.LoadFromDatabaseTransformer" >
+         <int name="connection">jdbc://....</int>
+       </transformer>
+       To add a constant value to all docs, use:
+       <transformer name="mytrans2" 
class="org.apache.solr.response.transform.ValueAugmenterFactory" >
+         <int name="value">5</int>
+       </transformer>
+       If you want the user to still be able to change it with 
_value:something_ use this:
+       <transformer name="mytrans3" 
class="org.apache.solr.response.transform.ValueAugmenterFactory" >
+         <double name="defaultValue">5</double>
+       </transformer>
+        If you are using the QueryElevationComponent, you may wish to mark 
documents that get boosted.  The
+        EditorialMarkerFactory will do exactly that:
+       <transformer name="qecBooster" 
class="org.apache.solr.response.transform.EditorialMarkerFactory" />
+      -->
+</config>


Reply via email to