Repository: incubator-pirk
Updated Branches:
  refs/heads/asf-site 492e11d99 -> 6250eea1e


Jekyll build from gh-pages:828f346

closes apache/incubator-pirk#32


Project: http://git-wip-us.apache.org/repos/asf/incubator-pirk/repo
Commit: http://git-wip-us.apache.org/repos/asf/incubator-pirk/commit/a672b316
Tree: http://git-wip-us.apache.org/repos/asf/incubator-pirk/tree/a672b316
Diff: http://git-wip-us.apache.org/repos/asf/incubator-pirk/diff/a672b316

Branch: refs/heads/asf-site
Commit: a672b31644f639b7030cb6ac8958b2e97269e767
Parents: 492e11d
Author: eawilliams <[email protected]>
Authored: Mon Aug 1 09:45:49 2016 -0400
Committer: eawilliams <[email protected]>
Committed: Mon Aug 1 09:45:51 2016 -0400

----------------------------------------------------------------------
 feed.xml                            |   4 +-
 target/classes/data-schema.xsd      |  84 -----------
 target/classes/log4j2.xml           |  49 -------
 target/classes/pirk.properties      | 239 -------------------------------
 target/classes/querier.properties   | 123 ----------------
 target/classes/query-schema.xsd     | 119 ---------------
 target/classes/responder.properties | 137 ------------------
 7 files changed, 2 insertions(+), 753 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/incubator-pirk/blob/a672b316/feed.xml
----------------------------------------------------------------------
diff --git a/feed.xml b/feed.xml
index 18d304b..33278cc 100644
--- a/feed.xml
+++ b/feed.xml
@@ -6,8 +6,8 @@
 </description>
     <link>http://pirk.incubator.apache.org/</link>
     <atom:link href="http://pirk.incubator.apache.org/feed.xml"; rel="self" 
type="application/rss+xml"/>
-    <pubDate>Mon, 01 Aug 2016 09:44:07 -0400</pubDate>
-    <lastBuildDate>Mon, 01 Aug 2016 09:44:07 -0400</lastBuildDate>
+    <pubDate>Mon, 01 Aug 2016 09:45:50 -0400</pubDate>
+    <lastBuildDate>Mon, 01 Aug 2016 09:45:50 -0400</lastBuildDate>
     <generator>Jekyll v3.0.5</generator>
     
   </channel>

http://git-wip-us.apache.org/repos/asf/incubator-pirk/blob/a672b316/target/classes/data-schema.xsd
----------------------------------------------------------------------
diff --git a/target/classes/data-schema.xsd b/target/classes/data-schema.xsd
deleted file mode 100644
index 7384409..0000000
--- a/target/classes/data-schema.xsd
+++ /dev/null
@@ -1,84 +0,0 @@
-<?xml version="1.0" encoding="UTF-8" ?>
-<!--
- ~ Licensed to the Apache Software Foundation (ASF) under one or more
- ~ contributor license agreements.  See the NOTICE file distributed with
- ~ this work for additional information regarding copyright ownership.
- ~ The ASF licenses this file to You under the Apache License, Version 2.0
- ~ (the "License"); you may not use this file except in compliance with
- ~ the License.  You may obtain a copy of the License at
- ~
- ~    http://www.apache.org/licenses/LICENSE-2.0
- ~
- ~ Unless required by applicable law or agreed to in writing, software
- ~ distributed under the License is distributed on an "AS IS" BASIS,
- ~ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- ~ See the License for the specific language governing permissions and
- ~ limitations under the License.
- -->
- <xs:schema xmlns:xs="http://www.w3.org/2001/XMLSchema";
-    targetNamespace="http://pirk.apache.org"; xmlns="http://pirk.apache.org";
-    elementFormDefault="qualified">
-
-    <xs:element name="schema">
-        <xs:complexType>
-            <xs:sequence>
-
-                <xs:element name="schemaName" type="xs:string">
-                    <xs:annotation>
-                        <xs:documentation>The name of the data schema.
-                            The name omits leading and trailing
-                            whitespace, and is case sensitive.
-                        </xs:documentation>
-                    </xs:annotation>
-                </xs:element>
-
-                <xs:element name="element" maxOccurs="unbounded">
-                    <xs:complexType>
-                        <xs:sequence>
-
-                            <xs:element name="name" type="xs:string">
-                                <xs:annotation>
-                                    <xs:documentation>The name of the
-                                        data element.
-                                        The name omits
-                                        leading and trailing whitespace,
-                                        and is case sensitive.
-                                    </xs:documentation>
-                                </xs:annotation>
-                            </xs:element>
-
-                            <xs:element name="type" type="xs:string">
-                                <xs:annotation>
-                                    <xs:documentation>The type of the data 
element.
-                                        The type name is the fully qualified 
class name,
-                                        or the primitive Java type of the 
element.
-                                    </xs:documentation>
-                                </xs:annotation>
-                            </xs:element>
-
-                            <xs:element name="isArray" type="xs:boolean"
-                                default="true" minOccurs="0">
-                                <xs:annotation>
-                                    <xs:documentation>Whether or not the 
schema element is
-                                        an array within the data.  False if 
unspecified.
-                                    </xs:documentation>
-                                </xs:annotation>
-                            </xs:element>
-
-                            <xs:element name="partitioner"
-                                type="xs:string" minOccurs="0">
-                                <xs:annotation>
-                                    <xs:documentation>Partitioner type for the 
element.
-                                        The type name is the fully qualified 
class name.
-                                        Uses the primitive Java type 
partitioner if unspecified.
-                                    </xs:documentation>
-                                </xs:annotation>
-                            </xs:element>
-                        </xs:sequence>
-                    </xs:complexType>
-                </xs:element>
-            </xs:sequence>
-        </xs:complexType>
-    </xs:element>
-
-</xs:schema>
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/incubator-pirk/blob/a672b316/target/classes/log4j2.xml
----------------------------------------------------------------------
diff --git a/target/classes/log4j2.xml b/target/classes/log4j2.xml
deleted file mode 100644
index 81549db..0000000
--- a/target/classes/log4j2.xml
+++ /dev/null
@@ -1,49 +0,0 @@
-<?xml version="1.0" encoding="UTF-8"?>
-<!--
-  ~ Licensed to the Apache Software Foundation (ASF) under one or more
-  ~ contributor license agreements.  See the NOTICE file distributed with
-  ~ this work for additional information regarding copyright ownership.
-  ~ The ASF licenses this file to You under the Apache License, Version 2.0
-  ~ (the "License"); you may not use this file except in compliance with
-  ~ the License.  You may obtain a copy of the License at
-  ~
-  ~    http://www.apache.org/licenses/LICENSE-2.0
-  ~
-  ~ Unless required by applicable law or agreed to in writing, software
-  ~ distributed under the License is distributed on an "AS IS" BASIS,
-  ~ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-  ~ See the License for the specific language governing permissions and
-  ~ limitations under the License.
-  -->
-<Configuration>
-       <Properties>
-               <Property name="projectPrefix">pirk</Property>
-               <Property name="Directory">./logs</Property>
-               <Property name="Filename">${projectPrefix}.log</Property>
-       </Properties>
-
-       <Appenders>
-               <Console name="STDOUT" target="SYSTEM_OUT">
-                       <PatternLayout pattern="%5p [%t] %d (%F:%L) - %m%n"/>
-               </Console>
-
-               <RollingFile name="RollingFile" 
fileName="${Directory}/${Filename}" filePattern="${projectPrefix}-%i.log">
-            <PatternLayout pattern="%d %-5p %-17c{2} %3x - %m%n"/>
-                       <Policies>
-                               <SizeBasedTriggeringPolicy size="1 MB"/>
-                       </Policies>
-                       <DefaultRolloverStrategy max="2"/>
-               </RollingFile>
-       </Appenders>
-
-       <Loggers>
-               <Logger name="org.apache.hadoop.hbase" level="info" 
additivity="false"/>
-               <Logger name="org.apache.hadoop.hbase.zookeeper" level="warn" 
additivity="false"/>
-               <Logger name="org.apache.zookeeper" level="error" 
additivity="false"/>
-               <Root level="info">
-                       <AppenderRef ref="STDOUT"/>
-                       <AppenderRef ref="RollingFile"/>
-               </Root>
-       </Loggers>
-
-</Configuration>
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/incubator-pirk/blob/a672b316/target/classes/pirk.properties
----------------------------------------------------------------------
diff --git a/target/classes/pirk.properties b/target/classes/pirk.properties
deleted file mode 100644
index aee87ad..0000000
--- a/target/classes/pirk.properties
+++ /dev/null
@@ -1,239 +0,0 @@
-###############################################################################
-# Licensed to the Apache Software Foundation (ASF) under one
-# or more contributor license agreements.  See the NOTICE file
-# distributed with this work for additional information
-# regarding copyright ownership.  The ASF licenses this file
-# to you under the Apache License, Version 2.0 (the
-# "License"); you may not use this file except in compliance
-# with the License.  You may obtain a copy of the License at
-#
-#   http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing,
-# software distributed under the License is distributed on an
-# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
-# KIND, either express or implied.  See the License for the
-# specific language governing permissions and limitations
-# under the License.
-###############################################################################
-
-##
-## Default System Configuration Properties
-##
-
-#Name of the local properties file - used when running with the
-#hadoop jar command
-local.pirk.properties=/root/
-
-##
-##Spark path for SparkLauncher
-##
-spark.home = /usr
-
-##
-## Data schema properties
-##
-## Each data schema should be specified in an xml file of the form:
-##
-##<schema>
-##  <schemaName> name of the schema </schemaName>
-##  <element>
-##      <name> element name </name>
-##      <type> class name or type name (if Java primitive type) of the element 
</type>
-##      <isArray> true or false -- whether or not the schema element is an 
array within the data </isArray>
-##      <partitioner> optional - Partitioner class for the element; defaults 
to primitive java type partitioner </partitioner> 
-##  </element>
-## </schema>
-##
-## Primitive Java types must be one of the following: "byte", "short", "int", 
"long", "float", 
-##                  "double", "char", "string", "boolean"
-##
-
-#Comma separated list of local data schema files to load, fully qualified file 
names
-data.schemas = none
-
-##
-## Query schema properties
-##
-## Each query schema should be specified in an xml file of the form; 
-##             all items are treated in a case insensitive manner:
-##
-##<schema>
-##  <schemaName> name of the schema </schemaName>
-##     <dataSchemaName> name of the data schema over which this query is run 
</dataSchemaName>
-##     <selectorName> name of the element in the data schema that will be the 
selector </selectorName>
-##  <elements>
-##             <name> element name of element in the data schema to include in 
the query response </name>
-##     </elements>
-##  <filter> (optional) name of the filter class to use to filter the data 
</filter>
-##  <filterNames>
-##       <name> (optional) element name of element in the data schema to apply 
pre-processing filters </name>
-##  </filterNames>
-## </schema>
-##
-##
-
-#Comma separated list of local query schema files to load, fully qualified 
file names
-query.schemas = none
-
-##
-##Properties for ElasticSearch compatibility
-##
-
-#ES host address - One Elasticsearch node in the cluster - may include port 
specification
-es.nodes= none
-          
-#Default HTTP/REST port used for connecting to Elasticsearch 
-es.port=9200
-
-#Number of results/items returned by each individual scroll
-es.scroll.size = 1000
-
-#Whether elasticsearch-hadoop will treat empty fields as null
-es.field.read.empty.as.null=yes
-
-
-##
-##Properties for functional testing
-##
-
-#Test index to create in ES (created and deleted within the tests)
-#Need final form to be: <host>:<port>/<test.esTestIndex>
-test.es.index = testindex
-test.pir.es.index = testindexpir
-
-#Type of elements to insert in ES
-test.es.type = pkt
-
-#Elasticsearch resource - Elasticsearch resource location where data is read 
and written to. 
-#Requires the format <index>/<type> 
-test.es.resource= none
-test.pir.es.resource = none
-
-#Pathname in hdfs to place input JSON file testing
-test.inputJSONFile = none
-test.pir.inputJSONFile = /tmp/testJSONInput
-
-#Pathname in hdfs to place output file for testing
-test.outputHDFSFile = /tmp/testOutput
-
-#PIR query input dir in hdfs for testing
-test.pir.queryInputDir = none
-
-#PIR stoplist file
-test.pir.stopListFile = /tmp/testStopListFile
-
-#Whether or not we are running PIR testing (used as a flag to dump 
intermediate RDDs for checking)
-#This should default to false; it is changed to true in the test suite, as 
applicable
-pir.test = false
-
-#HDFS output dir for PIR intermediate testing
-#Should default to none; changed to a real hdfs path in the test suite, as 
applicable
-pir.testOut = none
-
-
-##
-## Properties to enable/disable JNA-GMP modPow acceleration for Paillier 
-##
-
-paillier.useGMPForModPow = true
-
-# The JNA-GMP library we use to invoke the much faster, natively compiled GMP
-# can be called in a way that tries to make all modPow calls take a constant 
amount 
-# of time to execute. This will slow down the modPow performance (still better 
than
-# Java's BigInteger.modPow() ). 
-# If you are using this package in a multi-tenant computing environment and 
have 
-# concerns about other tenants being able to closely inspect the runtime of 
your
-# software, you may want to enable this property. 
-paillier.GMPConstantTimeMode = false
-
-# This property controls the more rigorous prime generation checks in 
PrimeMaker.java
-# which are based on FIPS SP 800-56B and FIPS 186-4 (extra Miller-Rabin 
rounds, limits
-# on how close the primes p and q can be, and bounds on the values of the 
primes)
-# These checks slow down prime generation considerably
-pallier.FIPSPrimeGenerationChecks = true
-
-## These properties control the secure random number generator algorithm and 
provider.
-## You can specify just the algorithm, or both algorithm and provider.  The 
system's
-## default secure random is used when the algorithm is left unspecified. 
-pallier.secureRandom.algorithm=NativePRNG
-#pallier.secureRandom.provider=SUN
-
-##
-## Properties for PIR query and response
-##
-
-#Number of bits to return when encoding/returning string values as part of 
return data elements
-pir.stringBits = 64
-
-#Number of array elements to return when encoding/returning array valued 
elements
-pir.numReturnArrayElements = 2
-
-#Default prime certainty
-pir.primeCertainty = 128
-
-#Fully qualified dir in hdfs of Query files
-pir.queryInput = none
-
-#Data input format type -- 'base' or 'elasticsearch' (given in 
InputFormatsConst class)
-pir.dataInputFormat = base
-
-#Fully qualified name of input file/directory in hdfs; used if 
pir.dataInputFormat = base
-pir.inputData = none
-
-#Inputformat for 'base' data input format type -- must extend BaseInputFormat
-pir.baseInputFormat = none
-
-#ElasticSearch-like query if using 'base' input format
-pir.baseQuery = none
-
-#ES resource for input data
-pir.esResource = null
-
-#ES query for input data
-pir.esQuery = none
-
-#Fully qualified name of output file in hdfs
-pir.outputFile = none
-
-#Fully qualified dir in hdfs of file containing stoplist terms
-pir.stopListFile = stopListFile
-
-#Number of reduce tasks
-pir.numReduceTasks = 100
-
-#Whether or not to use the local cache during PIR computations
-pir.useLocalCache = true
-
-#Whether or not to limit the hits for each query term
-pir.limitHitsPerSelector = true
-
-#Number of hits to limit for each query term, if pir.limitHitsPerSelector = 
true
-pir.maxHitsPerSelector = 100
-
-#Whether or not to embed the selector in the results for false positive 
reduction
-pir.embedSelector = true
-
-#Whether or not to generate and use the HDFS modular exponentiation lookup 
table
-pir.useHDFSLookupTable = false
-
-#Number of partitions to coalesce the input data into in Spark
-pir.numDataPartitions = 1500
-
-#Mapreduce memory options
-mapreduce.map.memory.mb = 3000
-mapreduce.reduce.memory.mb = 3000
-mapreduce.map.java.opts = -Xmx2800m
-mapreduce.reduce.java.opts = -Xmx2800m
-
-#HDFS directory for the expLookupTable 
-pir.expDir = none
-               
-#Parallelism for expLookupTable creation in hdfs 
-pir.expCreationSplits = 600
-
-
-
-
-
-

http://git-wip-us.apache.org/repos/asf/incubator-pirk/blob/a672b316/target/classes/querier.properties
----------------------------------------------------------------------
diff --git a/target/classes/querier.properties 
b/target/classes/querier.properties
deleted file mode 100644
index 2051d46..0000000
--- a/target/classes/querier.properties
+++ /dev/null
@@ -1,123 +0,0 @@
-###############################################################################
-# Licensed to the Apache Software Foundation (ASF) under one
-# or more contributor license agreements.  See the NOTICE file
-# distributed with this work for additional information
-# regarding copyright ownership.  The ASF licenses this file
-# to you under the Apache License, Version 2.0 (the
-# "License"); you may not use this file except in compliance
-# with the License.  You may obtain a copy of the License at
-#
-#   http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing,
-# software distributed under the License is distributed on an
-# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
-# KIND, either express or implied.  See the License for the
-# specific language governing permissions and limitations
-# under the License.
-###############################################################################
-
-##
-## Required Properties
-##
-
-#action -- required - 'encrypt' or 'decrypt' -- The action performed by the 
QuerierDriver
-querier.action=
-
-#inputFile - required - Fully qualified file containing input
-#The input is either:
-#(1) For Encryption: A query file - Contains the query selectors, one per line;
-#the first line must be the query number
-#OR
-#(2) For Decryption: A response file - Contains the serialized Response object
-querier.inputFile=
-
-#outputFile -- required - Fully qualified file for the result output.
-#The output file specifies either:
-#(1) For encryption:
-#(a) A file to contain the serialized Querier object named: 
<outputFile>-querier
-#AND
-#(b) A file to contain the serialized Query object named: <outputFile>-query
-#OR
-#(2) A file to contain the decryption results where each line is where each 
line
-#corresponds to one hit and is a JSON object with the schema QuerySchema
-querier.outputFile=
-
-#numThreads -- required -- Number of threads to use for encryption/decryption
-querier.numThreads=
-
-##
-## Optional Args
-##
-
-#dataSchemas -- optional -- Comma separated list of data schema file names to 
load
-#querier.dataSchemas=
-
-#querySchemas -- optional -- Comma separated list of query schema file names 
to load
-#querier.dataSchemas=
-
-
-## Optional, but required for Encryption (ignored if not encrypting)
-
-#bitset -- required for encryption -- Ensure that this bit position is set in 
the Paillier
-#modulus (will generate Paillier moduli until finding one in which this bit is 
set)
-#querier.bitSet=
-
-#certainty -- required for encryption -- Certainty of prime generation for 
Paillier
-#must be greater than or equal to 128
-#querier.certainty=
-
-#dataPartitionBitSize -- required for encryption -- Partition bit size in data 
partitioning
-#querier.dataPartitionBitSize=
-
-#embedSelector - required for encryption -- 'true' or 'false'
-#Whether or not to embed the selector in the results to reduce false positives
-#Defaults to 'true'
-#querier.embedSelector=
-
-#embedQuerySchema - true or false
-#Whether or not to embed the QuerySchema in the Query (via QueryInfo)
-#Defaults to 'false'
-#querier.embedQuerySchema=
- 
-#hashBitSize - required for encryption-- Bit size of keyed hash
-#querier.hashBitSize=
-
-#hashKey -- required for encryption -- String key for the keyed hash 
functionality
-#querier.hashKey=
-
-#useHDFSLookupTable -- required for encryption -- 'true' or 'false'
-#Whether or not to generate and use a hdfs modular exponentation lookup table
-#Defaults to 'false'
-#querier.useHDFSLookupTable=
-
-#memLookupTable -- required for encryption -- 'true' or 'false'
-#Whether or not to generate and use an in memory modular exponentation lookup 
table - only for
-#standalone/testing right now...
-#Defaults to 'false'
-#querier.memLookupTable=
-
-#paillierBitSize -- required for encryption -- Paillier modulus size N
-#querier.paillierBitSize=
- 
-#queryName -- required for encryption -- Name of the query
-#querier.queryName=
-
-#queryType -- required for encryption
-#Type of the query as defined in the 'schemaName' tag of the corresponding 
query schema file
-#querier.queryType=
-
-#secureRandomAlg -- specify the SecureRandom algorithm
-#Defaults to NativePRNG
-#querier.secureRandomAlg=
-
-#secureRandomProvider -- specify the SecureRandom provider
-#Defaults to SUN
-#querier.secureRandomProvider=
-  
-## Optional, but required for Decryption (ignored if not decrypting)
-
-#querierFile -- required for decryption
-#Fully qualified file containing the serialized Querier object
-#querier.querierFile=
-  
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/incubator-pirk/blob/a672b316/target/classes/query-schema.xsd
----------------------------------------------------------------------
diff --git a/target/classes/query-schema.xsd b/target/classes/query-schema.xsd
deleted file mode 100644
index 65a36ce..0000000
--- a/target/classes/query-schema.xsd
+++ /dev/null
@@ -1,119 +0,0 @@
-<?xml version="1.0" encoding="UTF-8" ?>
-<!--
- ~ Licensed to the Apache Software Foundation (ASF) under one or more
- ~ contributor license agreements.  See the NOTICE file distributed with
- ~ this work for additional information regarding copyright ownership.
- ~ The ASF licenses this file to You under the Apache License, Version 2.0
- ~ (the "License"); you may not use this file except in compliance with
- ~ the License.  You may obtain a copy of the License at
- ~
- ~    http://www.apache.org/licenses/LICENSE-2.0
- ~
- ~ Unless required by applicable law or agreed to in writing, software
- ~ distributed under the License is distributed on an "AS IS" BASIS,
- ~ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- ~ See the License for the specific language governing permissions and
- ~ limitations under the License.
- -->
-<xs:schema xmlns:xs="http://www.w3.org/2001/XMLSchema";
-    targetNamespace="http://pirk.apache.org"; xmlns="http://pirk.apache.org";
-    elementFormDefault="qualified">
-
-    <xs:element name="schema">
-        <xs:complexType>
-            <xs:sequence>
-                <xs:element name="schemaName" type="xs:string">
-                    <xs:annotation>
-                        <xs:documentation>
-                            The name of the query schema.
-                            The name omits leading and trailing
-                            whitespace, and is case sensitive.
-                        </xs:documentation>
-                    </xs:annotation>
-                </xs:element>
-
-                <xs:element name="dataSchemaName" type="xs:string">
-                    <xs:annotation>
-                        <xs:documentation>
-                            The name of the data schema
-                            over which this query is run. The name omits
-                            leading and trailing whitespace, and is case
-                            sensitive.
-                        </xs:documentation>
-                    </xs:annotation>
-                </xs:element>
-
-                <xs:element name="selectorName" type="xs:string">
-                    <xs:annotation>
-                        <xs:documentation>The name of the name of the
-                            element in the data schema that will be the
-                            selector for this query.
-                        </xs:documentation>
-                    </xs:annotation>
-                </xs:element>
-
-                <xs:element name="elements">
-                    <xs:annotation>
-                        <xs:documentation>
-                            The set of element names to
-                            include in the query response.
-                        </xs:documentation>
-                    </xs:annotation>
-                    <xs:complexType>
-                        <xs:sequence>
-
-                            <xs:element name="name" type="xs:string"
-                                maxOccurs="unbounded">
-                                <xs:annotation>
-                                    <xs:documentation>
-                                        The name of an
-                                        element in the data schema to
-                                        include in the query response.
-                                    </xs:documentation>
-                                </xs:annotation>
-                            </xs:element>
-
-                        </xs:sequence>
-                    </xs:complexType>
-                </xs:element>
-
-                <xs:element name="filter" type="xs:string"
-                    minOccurs="0">
-                    <xs:annotation>
-                        <xs:documentation>
-                            The name of a class used to
-                            filter the query response data.
-                        </xs:documentation>
-                    </xs:annotation>
-                </xs:element>
-
-                <xs:element name="filterNames" minOccurs="0"
-                    maxOccurs="unbounded">
-                    <xs:annotation>
-                        <xs:documentation>
-                            The set of data element names
-                            over which the
-                            response filter is applied.
-                        </xs:documentation>
-                    </xs:annotation>
-                    <xs:complexType>
-                        <xs:sequence>
-
-                            <xs:element name="name" type="xs:string">
-                                <xs:annotation>
-                                    <xs:documentation>
-                                        The name of an
-                                        element in the data schema over
-                                        which to apply the filter.
-                                    </xs:documentation>
-                                </xs:annotation>
-                            </xs:element>
-
-                        </xs:sequence>
-                    </xs:complexType>
-                </xs:element>
-            </xs:sequence>
-        </xs:complexType>
-    </xs:element>
-
-</xs:schema>

http://git-wip-us.apache.org/repos/asf/incubator-pirk/blob/a672b316/target/classes/responder.properties
----------------------------------------------------------------------
diff --git a/target/classes/responder.properties 
b/target/classes/responder.properties
deleted file mode 100644
index 11ad7f6..0000000
--- a/target/classes/responder.properties
+++ /dev/null
@@ -1,137 +0,0 @@
-###############################################################################
-# Licensed to the Apache Software Foundation (ASF) under one
-# or more contributor license agreements.  See the NOTICE file
-# distributed with this work for additional information
-# regarding copyright ownership.  The ASF licenses this file
-# to you under the Apache License, Version 2.0 (the
-# "License"); you may not use this file except in compliance
-# with the License.  You may obtain a copy of the License at
-#
-#   http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing,
-# software distributed under the License is distributed on an
-# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
-# KIND, either express or implied.  See the License for the
-# specific language governing permissions and limitations
-# under the License.
-###############################################################################
-
-##
-## Required Properties
-##
-
-#dataInputFormat -- required -- 'base', 'elasticsearch', or 'standalone' -- 
Specify the input format
-pir.dataInputFormat=
-
-#outputFile -- required -- Fully qualified name of output file in hdfs
-pir.outputFile=
-
-#platform -- required -- 'mapreduce', 'spark', or 'standalone'
-#Processing platform technology for the responder                
-platform= 
-
-#queryInput -- required -- Fully qualified dir in hdfs of Query files
-pir.queryInput=
-
-
-##
-## Optional Args - Leave empty if not using/not changing default values
-##
-
-#inputData -- required if baseInputFormat = 'base'
-#Fully qualified name of input file/directory in hdfs; used if inputFormat = 
'base'
-#pir.inputData=
- 
-#dataSchemas -- optional -- Comma separated list of data schema file names to 
load
-#responder.dataSchemas=
-
-#querySchemas -- optional -- Comma separated list of query schema file names 
to load
-#responder.querySchemas=                                    
-
-#allowAdHocQuerySchemas -- 'true' or 'false'
-#If true, allows embedded QuerySchemas for a query.
-#Defaults to 'false'
-#pir.allowEmbeddedQuerySchemas=
-
-#colMultReduceByKey -- 'true' or 'false' -- Spark only
-#If true, uses reduceByKey in performing column multiplication; if false, uses 
groupByKey -> reduce
-#Defaults to 'false' 
-#pir.colMultReduceByKey=
-
-#baseInputFormat -- required if baseInputFormat = 'base'
-#Full class name of the InputFormat to use when reading in the data - must 
extend BaseInputFormat
-#pir.baseInputFormat=
-
-#esQuery -- required if baseInputFormat = 'elasticsearch' -- ElasticSearch 
query
-#if using 'elasticsearch' input format
-#pir.esQuery=
-
-#esResource --  required if baseInputFormat = 'elasticsearch'
-#Requires the format <index>/<type> : Elasticsearch resource where data is 
read and written to
-#pir.esResource=
-
-#useHDFSLookupTable -- 'true' or 'false' - Whether or not to generate and use 
the
-#hdfs lookup table for modular exponentiation
-#Defaults to 'false'
-#pir.useHDFSLookupTable=
-
-#baseQuery -- ElasticSearch-like query if using 'base' input format -
-#used to filter records in the RecordReader
-#Defaults to ?q=*
-#pir.baseQuery=
-
-#limitHitsPerSelector -- 'true' or 'false'
-#Whether or not to limit the number of hits per selector
-#Defaults to 'true'
-#pir.limitHitsPerSelector=
-
-#mapreduceMapJavaOpts -- Amount of heap (in MB) to allocate per map task
-#Defaults to -Xmx2800m
-#mapreduce.map.java.opts=
-
-#mapreduceMapMemoryMb -- Amount of memory (in MB) to allocate per map task
-#Defaults to 3000
-#mapreduce.map.memory.mb=
-     
-#mapreduceReduceJavaOpts
-#Amount of heap (in MB) to allocate per reduce task
-#Defaults to -Xmx2800m
-#mapreduce.reduce.java.opts=
-
-#mapreduceReduceMemoryMb
-#Amount of memory (in MB) to allocate per reduce task
-#Defaults to 3000
-#mapreduce.reduce.memory.mb=
-
-#stopListFile -- optional (unless using StopListFilter) -- Fully qualified 
file in hdfs
-#containing stoplist terms; used by the StopListFilter
-#pir.stopListFile=
-
-#useLocalCache -- 'true' or 'false'
-#Whether or not to use the local cache for modular exponentiation
-#Defaults to 'true'
-#pir.useLocalCache=
-
-#useModExpJoin -- 'true' or 'false' -- Spark only
-#Whether or not to pre-compute the modular exponentiation table and join it to 
the data
-#partitions when performing the encrypted row calculations
-#Defaults to 'false'
-#pir.useModExpJoin=
-
-#numReduceTasks -- optional -- Number of reduce tasks
-#pir.numReduceTasks=
-
-#numColMultPartitions -- optional, Spark only
-#Number of partitions to use when performing  column multiplication
-#pir.numColMultPartitions=        
-
-#maxHitsPerSelector --  optional -- Max number of hits encrypted per selector
-#pir.maxHitsPerSelector=
-
-#dataParts -- optional -- Number of partitions for the input data
-#pir.numDataPartitions=
-
-#numExpLookupPartitions -- optional -- Number of partitions for the exp lookup 
table
-#pir.numExpLookupPartitions=
- 
\ No newline at end of file

Reply via email to