Repository: incubator-pirk Updated Branches: refs/heads/asf-site 492e11d99 -> 6250eea1e
Jekyll build from gh-pages:828f346 closes apache/incubator-pirk#32 Project: http://git-wip-us.apache.org/repos/asf/incubator-pirk/repo Commit: http://git-wip-us.apache.org/repos/asf/incubator-pirk/commit/a672b316 Tree: http://git-wip-us.apache.org/repos/asf/incubator-pirk/tree/a672b316 Diff: http://git-wip-us.apache.org/repos/asf/incubator-pirk/diff/a672b316 Branch: refs/heads/asf-site Commit: a672b31644f639b7030cb6ac8958b2e97269e767 Parents: 492e11d Author: eawilliams <[email protected]> Authored: Mon Aug 1 09:45:49 2016 -0400 Committer: eawilliams <[email protected]> Committed: Mon Aug 1 09:45:51 2016 -0400 ---------------------------------------------------------------------- feed.xml | 4 +- target/classes/data-schema.xsd | 84 ----------- target/classes/log4j2.xml | 49 ------- target/classes/pirk.properties | 239 ------------------------------- target/classes/querier.properties | 123 ---------------- target/classes/query-schema.xsd | 119 --------------- target/classes/responder.properties | 137 ------------------ 7 files changed, 2 insertions(+), 753 deletions(-) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/incubator-pirk/blob/a672b316/feed.xml ---------------------------------------------------------------------- diff --git a/feed.xml b/feed.xml index 18d304b..33278cc 100644 --- a/feed.xml +++ b/feed.xml @@ -6,8 +6,8 @@ </description> <link>http://pirk.incubator.apache.org/</link> <atom:link href="http://pirk.incubator.apache.org/feed.xml" rel="self" type="application/rss+xml"/> - <pubDate>Mon, 01 Aug 2016 09:44:07 -0400</pubDate> - <lastBuildDate>Mon, 01 Aug 2016 09:44:07 -0400</lastBuildDate> + <pubDate>Mon, 01 Aug 2016 09:45:50 -0400</pubDate> + <lastBuildDate>Mon, 01 Aug 2016 09:45:50 -0400</lastBuildDate> <generator>Jekyll v3.0.5</generator> </channel> http://git-wip-us.apache.org/repos/asf/incubator-pirk/blob/a672b316/target/classes/data-schema.xsd ---------------------------------------------------------------------- diff --git a/target/classes/data-schema.xsd b/target/classes/data-schema.xsd deleted file mode 100644 index 7384409..0000000 --- a/target/classes/data-schema.xsd +++ /dev/null @@ -1,84 +0,0 @@ -<?xml version="1.0" encoding="UTF-8" ?> -<!-- - ~ Licensed to the Apache Software Foundation (ASF) under one or more - ~ contributor license agreements. See the NOTICE file distributed with - ~ this work for additional information regarding copyright ownership. - ~ The ASF licenses this file to You under the Apache License, Version 2.0 - ~ (the "License"); you may not use this file except in compliance with - ~ the License. You may obtain a copy of the License at - ~ - ~ http://www.apache.org/licenses/LICENSE-2.0 - ~ - ~ Unless required by applicable law or agreed to in writing, software - ~ distributed under the License is distributed on an "AS IS" BASIS, - ~ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - ~ See the License for the specific language governing permissions and - ~ limitations under the License. - --> - <xs:schema xmlns:xs="http://www.w3.org/2001/XMLSchema" - targetNamespace="http://pirk.apache.org" xmlns="http://pirk.apache.org" - elementFormDefault="qualified"> - - <xs:element name="schema"> - <xs:complexType> - <xs:sequence> - - <xs:element name="schemaName" type="xs:string"> - <xs:annotation> - <xs:documentation>The name of the data schema. - The name omits leading and trailing - whitespace, and is case sensitive. - </xs:documentation> - </xs:annotation> - </xs:element> - - <xs:element name="element" maxOccurs="unbounded"> - <xs:complexType> - <xs:sequence> - - <xs:element name="name" type="xs:string"> - <xs:annotation> - <xs:documentation>The name of the - data element. - The name omits - leading and trailing whitespace, - and is case sensitive. - </xs:documentation> - </xs:annotation> - </xs:element> - - <xs:element name="type" type="xs:string"> - <xs:annotation> - <xs:documentation>The type of the data element. - The type name is the fully qualified class name, - or the primitive Java type of the element. - </xs:documentation> - </xs:annotation> - </xs:element> - - <xs:element name="isArray" type="xs:boolean" - default="true" minOccurs="0"> - <xs:annotation> - <xs:documentation>Whether or not the schema element is - an array within the data. False if unspecified. - </xs:documentation> - </xs:annotation> - </xs:element> - - <xs:element name="partitioner" - type="xs:string" minOccurs="0"> - <xs:annotation> - <xs:documentation>Partitioner type for the element. - The type name is the fully qualified class name. - Uses the primitive Java type partitioner if unspecified. - </xs:documentation> - </xs:annotation> - </xs:element> - </xs:sequence> - </xs:complexType> - </xs:element> - </xs:sequence> - </xs:complexType> - </xs:element> - -</xs:schema> \ No newline at end of file http://git-wip-us.apache.org/repos/asf/incubator-pirk/blob/a672b316/target/classes/log4j2.xml ---------------------------------------------------------------------- diff --git a/target/classes/log4j2.xml b/target/classes/log4j2.xml deleted file mode 100644 index 81549db..0000000 --- a/target/classes/log4j2.xml +++ /dev/null @@ -1,49 +0,0 @@ -<?xml version="1.0" encoding="UTF-8"?> -<!-- - ~ Licensed to the Apache Software Foundation (ASF) under one or more - ~ contributor license agreements. See the NOTICE file distributed with - ~ this work for additional information regarding copyright ownership. - ~ The ASF licenses this file to You under the Apache License, Version 2.0 - ~ (the "License"); you may not use this file except in compliance with - ~ the License. You may obtain a copy of the License at - ~ - ~ http://www.apache.org/licenses/LICENSE-2.0 - ~ - ~ Unless required by applicable law or agreed to in writing, software - ~ distributed under the License is distributed on an "AS IS" BASIS, - ~ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - ~ See the License for the specific language governing permissions and - ~ limitations under the License. - --> -<Configuration> - <Properties> - <Property name="projectPrefix">pirk</Property> - <Property name="Directory">./logs</Property> - <Property name="Filename">${projectPrefix}.log</Property> - </Properties> - - <Appenders> - <Console name="STDOUT" target="SYSTEM_OUT"> - <PatternLayout pattern="%5p [%t] %d (%F:%L) - %m%n"/> - </Console> - - <RollingFile name="RollingFile" fileName="${Directory}/${Filename}" filePattern="${projectPrefix}-%i.log"> - <PatternLayout pattern="%d %-5p %-17c{2} %3x - %m%n"/> - <Policies> - <SizeBasedTriggeringPolicy size="1 MB"/> - </Policies> - <DefaultRolloverStrategy max="2"/> - </RollingFile> - </Appenders> - - <Loggers> - <Logger name="org.apache.hadoop.hbase" level="info" additivity="false"/> - <Logger name="org.apache.hadoop.hbase.zookeeper" level="warn" additivity="false"/> - <Logger name="org.apache.zookeeper" level="error" additivity="false"/> - <Root level="info"> - <AppenderRef ref="STDOUT"/> - <AppenderRef ref="RollingFile"/> - </Root> - </Loggers> - -</Configuration> \ No newline at end of file http://git-wip-us.apache.org/repos/asf/incubator-pirk/blob/a672b316/target/classes/pirk.properties ---------------------------------------------------------------------- diff --git a/target/classes/pirk.properties b/target/classes/pirk.properties deleted file mode 100644 index aee87ad..0000000 --- a/target/classes/pirk.properties +++ /dev/null @@ -1,239 +0,0 @@ -############################################################################### -# Licensed to the Apache Software Foundation (ASF) under one -# or more contributor license agreements. See the NOTICE file -# distributed with this work for additional information -# regarding copyright ownership. The ASF licenses this file -# to you under the Apache License, Version 2.0 (the -# "License"); you may not use this file except in compliance -# with the License. You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, -# software distributed under the License is distributed on an -# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY -# KIND, either express or implied. See the License for the -# specific language governing permissions and limitations -# under the License. -############################################################################### - -## -## Default System Configuration Properties -## - -#Name of the local properties file - used when running with the -#hadoop jar command -local.pirk.properties=/root/ - -## -##Spark path for SparkLauncher -## -spark.home = /usr - -## -## Data schema properties -## -## Each data schema should be specified in an xml file of the form: -## -##<schema> -## <schemaName> name of the schema </schemaName> -## <element> -## <name> element name </name> -## <type> class name or type name (if Java primitive type) of the element </type> -## <isArray> true or false -- whether or not the schema element is an array within the data </isArray> -## <partitioner> optional - Partitioner class for the element; defaults to primitive java type partitioner </partitioner> -## </element> -## </schema> -## -## Primitive Java types must be one of the following: "byte", "short", "int", "long", "float", -## "double", "char", "string", "boolean" -## - -#Comma separated list of local data schema files to load, fully qualified file names -data.schemas = none - -## -## Query schema properties -## -## Each query schema should be specified in an xml file of the form; -## all items are treated in a case insensitive manner: -## -##<schema> -## <schemaName> name of the schema </schemaName> -## <dataSchemaName> name of the data schema over which this query is run </dataSchemaName> -## <selectorName> name of the element in the data schema that will be the selector </selectorName> -## <elements> -## <name> element name of element in the data schema to include in the query response </name> -## </elements> -## <filter> (optional) name of the filter class to use to filter the data </filter> -## <filterNames> -## <name> (optional) element name of element in the data schema to apply pre-processing filters </name> -## </filterNames> -## </schema> -## -## - -#Comma separated list of local query schema files to load, fully qualified file names -query.schemas = none - -## -##Properties for ElasticSearch compatibility -## - -#ES host address - One Elasticsearch node in the cluster - may include port specification -es.nodes= none - -#Default HTTP/REST port used for connecting to Elasticsearch -es.port=9200 - -#Number of results/items returned by each individual scroll -es.scroll.size = 1000 - -#Whether elasticsearch-hadoop will treat empty fields as null -es.field.read.empty.as.null=yes - - -## -##Properties for functional testing -## - -#Test index to create in ES (created and deleted within the tests) -#Need final form to be: <host>:<port>/<test.esTestIndex> -test.es.index = testindex -test.pir.es.index = testindexpir - -#Type of elements to insert in ES -test.es.type = pkt - -#Elasticsearch resource - Elasticsearch resource location where data is read and written to. -#Requires the format <index>/<type> -test.es.resource= none -test.pir.es.resource = none - -#Pathname in hdfs to place input JSON file testing -test.inputJSONFile = none -test.pir.inputJSONFile = /tmp/testJSONInput - -#Pathname in hdfs to place output file for testing -test.outputHDFSFile = /tmp/testOutput - -#PIR query input dir in hdfs for testing -test.pir.queryInputDir = none - -#PIR stoplist file -test.pir.stopListFile = /tmp/testStopListFile - -#Whether or not we are running PIR testing (used as a flag to dump intermediate RDDs for checking) -#This should default to false; it is changed to true in the test suite, as applicable -pir.test = false - -#HDFS output dir for PIR intermediate testing -#Should default to none; changed to a real hdfs path in the test suite, as applicable -pir.testOut = none - - -## -## Properties to enable/disable JNA-GMP modPow acceleration for Paillier -## - -paillier.useGMPForModPow = true - -# The JNA-GMP library we use to invoke the much faster, natively compiled GMP -# can be called in a way that tries to make all modPow calls take a constant amount -# of time to execute. This will slow down the modPow performance (still better than -# Java's BigInteger.modPow() ). -# If you are using this package in a multi-tenant computing environment and have -# concerns about other tenants being able to closely inspect the runtime of your -# software, you may want to enable this property. -paillier.GMPConstantTimeMode = false - -# This property controls the more rigorous prime generation checks in PrimeMaker.java -# which are based on FIPS SP 800-56B and FIPS 186-4 (extra Miller-Rabin rounds, limits -# on how close the primes p and q can be, and bounds on the values of the primes) -# These checks slow down prime generation considerably -pallier.FIPSPrimeGenerationChecks = true - -## These properties control the secure random number generator algorithm and provider. -## You can specify just the algorithm, or both algorithm and provider. The system's -## default secure random is used when the algorithm is left unspecified. -pallier.secureRandom.algorithm=NativePRNG -#pallier.secureRandom.provider=SUN - -## -## Properties for PIR query and response -## - -#Number of bits to return when encoding/returning string values as part of return data elements -pir.stringBits = 64 - -#Number of array elements to return when encoding/returning array valued elements -pir.numReturnArrayElements = 2 - -#Default prime certainty -pir.primeCertainty = 128 - -#Fully qualified dir in hdfs of Query files -pir.queryInput = none - -#Data input format type -- 'base' or 'elasticsearch' (given in InputFormatsConst class) -pir.dataInputFormat = base - -#Fully qualified name of input file/directory in hdfs; used if pir.dataInputFormat = base -pir.inputData = none - -#Inputformat for 'base' data input format type -- must extend BaseInputFormat -pir.baseInputFormat = none - -#ElasticSearch-like query if using 'base' input format -pir.baseQuery = none - -#ES resource for input data -pir.esResource = null - -#ES query for input data -pir.esQuery = none - -#Fully qualified name of output file in hdfs -pir.outputFile = none - -#Fully qualified dir in hdfs of file containing stoplist terms -pir.stopListFile = stopListFile - -#Number of reduce tasks -pir.numReduceTasks = 100 - -#Whether or not to use the local cache during PIR computations -pir.useLocalCache = true - -#Whether or not to limit the hits for each query term -pir.limitHitsPerSelector = true - -#Number of hits to limit for each query term, if pir.limitHitsPerSelector = true -pir.maxHitsPerSelector = 100 - -#Whether or not to embed the selector in the results for false positive reduction -pir.embedSelector = true - -#Whether or not to generate and use the HDFS modular exponentiation lookup table -pir.useHDFSLookupTable = false - -#Number of partitions to coalesce the input data into in Spark -pir.numDataPartitions = 1500 - -#Mapreduce memory options -mapreduce.map.memory.mb = 3000 -mapreduce.reduce.memory.mb = 3000 -mapreduce.map.java.opts = -Xmx2800m -mapreduce.reduce.java.opts = -Xmx2800m - -#HDFS directory for the expLookupTable -pir.expDir = none - -#Parallelism for expLookupTable creation in hdfs -pir.expCreationSplits = 600 - - - - - - http://git-wip-us.apache.org/repos/asf/incubator-pirk/blob/a672b316/target/classes/querier.properties ---------------------------------------------------------------------- diff --git a/target/classes/querier.properties b/target/classes/querier.properties deleted file mode 100644 index 2051d46..0000000 --- a/target/classes/querier.properties +++ /dev/null @@ -1,123 +0,0 @@ -############################################################################### -# Licensed to the Apache Software Foundation (ASF) under one -# or more contributor license agreements. See the NOTICE file -# distributed with this work for additional information -# regarding copyright ownership. The ASF licenses this file -# to you under the Apache License, Version 2.0 (the -# "License"); you may not use this file except in compliance -# with the License. You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, -# software distributed under the License is distributed on an -# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY -# KIND, either express or implied. See the License for the -# specific language governing permissions and limitations -# under the License. -############################################################################### - -## -## Required Properties -## - -#action -- required - 'encrypt' or 'decrypt' -- The action performed by the QuerierDriver -querier.action= - -#inputFile - required - Fully qualified file containing input -#The input is either: -#(1) For Encryption: A query file - Contains the query selectors, one per line; -#the first line must be the query number -#OR -#(2) For Decryption: A response file - Contains the serialized Response object -querier.inputFile= - -#outputFile -- required - Fully qualified file for the result output. -#The output file specifies either: -#(1) For encryption: -#(a) A file to contain the serialized Querier object named: <outputFile>-querier -#AND -#(b) A file to contain the serialized Query object named: <outputFile>-query -#OR -#(2) A file to contain the decryption results where each line is where each line -#corresponds to one hit and is a JSON object with the schema QuerySchema -querier.outputFile= - -#numThreads -- required -- Number of threads to use for encryption/decryption -querier.numThreads= - -## -## Optional Args -## - -#dataSchemas -- optional -- Comma separated list of data schema file names to load -#querier.dataSchemas= - -#querySchemas -- optional -- Comma separated list of query schema file names to load -#querier.dataSchemas= - - -## Optional, but required for Encryption (ignored if not encrypting) - -#bitset -- required for encryption -- Ensure that this bit position is set in the Paillier -#modulus (will generate Paillier moduli until finding one in which this bit is set) -#querier.bitSet= - -#certainty -- required for encryption -- Certainty of prime generation for Paillier -#must be greater than or equal to 128 -#querier.certainty= - -#dataPartitionBitSize -- required for encryption -- Partition bit size in data partitioning -#querier.dataPartitionBitSize= - -#embedSelector - required for encryption -- 'true' or 'false' -#Whether or not to embed the selector in the results to reduce false positives -#Defaults to 'true' -#querier.embedSelector= - -#embedQuerySchema - true or false -#Whether or not to embed the QuerySchema in the Query (via QueryInfo) -#Defaults to 'false' -#querier.embedQuerySchema= - -#hashBitSize - required for encryption-- Bit size of keyed hash -#querier.hashBitSize= - -#hashKey -- required for encryption -- String key for the keyed hash functionality -#querier.hashKey= - -#useHDFSLookupTable -- required for encryption -- 'true' or 'false' -#Whether or not to generate and use a hdfs modular exponentation lookup table -#Defaults to 'false' -#querier.useHDFSLookupTable= - -#memLookupTable -- required for encryption -- 'true' or 'false' -#Whether or not to generate and use an in memory modular exponentation lookup table - only for -#standalone/testing right now... -#Defaults to 'false' -#querier.memLookupTable= - -#paillierBitSize -- required for encryption -- Paillier modulus size N -#querier.paillierBitSize= - -#queryName -- required for encryption -- Name of the query -#querier.queryName= - -#queryType -- required for encryption -#Type of the query as defined in the 'schemaName' tag of the corresponding query schema file -#querier.queryType= - -#secureRandomAlg -- specify the SecureRandom algorithm -#Defaults to NativePRNG -#querier.secureRandomAlg= - -#secureRandomProvider -- specify the SecureRandom provider -#Defaults to SUN -#querier.secureRandomProvider= - -## Optional, but required for Decryption (ignored if not decrypting) - -#querierFile -- required for decryption -#Fully qualified file containing the serialized Querier object -#querier.querierFile= - \ No newline at end of file http://git-wip-us.apache.org/repos/asf/incubator-pirk/blob/a672b316/target/classes/query-schema.xsd ---------------------------------------------------------------------- diff --git a/target/classes/query-schema.xsd b/target/classes/query-schema.xsd deleted file mode 100644 index 65a36ce..0000000 --- a/target/classes/query-schema.xsd +++ /dev/null @@ -1,119 +0,0 @@ -<?xml version="1.0" encoding="UTF-8" ?> -<!-- - ~ Licensed to the Apache Software Foundation (ASF) under one or more - ~ contributor license agreements. See the NOTICE file distributed with - ~ this work for additional information regarding copyright ownership. - ~ The ASF licenses this file to You under the Apache License, Version 2.0 - ~ (the "License"); you may not use this file except in compliance with - ~ the License. You may obtain a copy of the License at - ~ - ~ http://www.apache.org/licenses/LICENSE-2.0 - ~ - ~ Unless required by applicable law or agreed to in writing, software - ~ distributed under the License is distributed on an "AS IS" BASIS, - ~ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - ~ See the License for the specific language governing permissions and - ~ limitations under the License. - --> -<xs:schema xmlns:xs="http://www.w3.org/2001/XMLSchema" - targetNamespace="http://pirk.apache.org" xmlns="http://pirk.apache.org" - elementFormDefault="qualified"> - - <xs:element name="schema"> - <xs:complexType> - <xs:sequence> - <xs:element name="schemaName" type="xs:string"> - <xs:annotation> - <xs:documentation> - The name of the query schema. - The name omits leading and trailing - whitespace, and is case sensitive. - </xs:documentation> - </xs:annotation> - </xs:element> - - <xs:element name="dataSchemaName" type="xs:string"> - <xs:annotation> - <xs:documentation> - The name of the data schema - over which this query is run. The name omits - leading and trailing whitespace, and is case - sensitive. - </xs:documentation> - </xs:annotation> - </xs:element> - - <xs:element name="selectorName" type="xs:string"> - <xs:annotation> - <xs:documentation>The name of the name of the - element in the data schema that will be the - selector for this query. - </xs:documentation> - </xs:annotation> - </xs:element> - - <xs:element name="elements"> - <xs:annotation> - <xs:documentation> - The set of element names to - include in the query response. - </xs:documentation> - </xs:annotation> - <xs:complexType> - <xs:sequence> - - <xs:element name="name" type="xs:string" - maxOccurs="unbounded"> - <xs:annotation> - <xs:documentation> - The name of an - element in the data schema to - include in the query response. - </xs:documentation> - </xs:annotation> - </xs:element> - - </xs:sequence> - </xs:complexType> - </xs:element> - - <xs:element name="filter" type="xs:string" - minOccurs="0"> - <xs:annotation> - <xs:documentation> - The name of a class used to - filter the query response data. - </xs:documentation> - </xs:annotation> - </xs:element> - - <xs:element name="filterNames" minOccurs="0" - maxOccurs="unbounded"> - <xs:annotation> - <xs:documentation> - The set of data element names - over which the - response filter is applied. - </xs:documentation> - </xs:annotation> - <xs:complexType> - <xs:sequence> - - <xs:element name="name" type="xs:string"> - <xs:annotation> - <xs:documentation> - The name of an - element in the data schema over - which to apply the filter. - </xs:documentation> - </xs:annotation> - </xs:element> - - </xs:sequence> - </xs:complexType> - </xs:element> - </xs:sequence> - </xs:complexType> - </xs:element> - -</xs:schema> http://git-wip-us.apache.org/repos/asf/incubator-pirk/blob/a672b316/target/classes/responder.properties ---------------------------------------------------------------------- diff --git a/target/classes/responder.properties b/target/classes/responder.properties deleted file mode 100644 index 11ad7f6..0000000 --- a/target/classes/responder.properties +++ /dev/null @@ -1,137 +0,0 @@ -############################################################################### -# Licensed to the Apache Software Foundation (ASF) under one -# or more contributor license agreements. See the NOTICE file -# distributed with this work for additional information -# regarding copyright ownership. The ASF licenses this file -# to you under the Apache License, Version 2.0 (the -# "License"); you may not use this file except in compliance -# with the License. You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, -# software distributed under the License is distributed on an -# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY -# KIND, either express or implied. See the License for the -# specific language governing permissions and limitations -# under the License. -############################################################################### - -## -## Required Properties -## - -#dataInputFormat -- required -- 'base', 'elasticsearch', or 'standalone' -- Specify the input format -pir.dataInputFormat= - -#outputFile -- required -- Fully qualified name of output file in hdfs -pir.outputFile= - -#platform -- required -- 'mapreduce', 'spark', or 'standalone' -#Processing platform technology for the responder -platform= - -#queryInput -- required -- Fully qualified dir in hdfs of Query files -pir.queryInput= - - -## -## Optional Args - Leave empty if not using/not changing default values -## - -#inputData -- required if baseInputFormat = 'base' -#Fully qualified name of input file/directory in hdfs; used if inputFormat = 'base' -#pir.inputData= - -#dataSchemas -- optional -- Comma separated list of data schema file names to load -#responder.dataSchemas= - -#querySchemas -- optional -- Comma separated list of query schema file names to load -#responder.querySchemas= - -#allowAdHocQuerySchemas -- 'true' or 'false' -#If true, allows embedded QuerySchemas for a query. -#Defaults to 'false' -#pir.allowEmbeddedQuerySchemas= - -#colMultReduceByKey -- 'true' or 'false' -- Spark only -#If true, uses reduceByKey in performing column multiplication; if false, uses groupByKey -> reduce -#Defaults to 'false' -#pir.colMultReduceByKey= - -#baseInputFormat -- required if baseInputFormat = 'base' -#Full class name of the InputFormat to use when reading in the data - must extend BaseInputFormat -#pir.baseInputFormat= - -#esQuery -- required if baseInputFormat = 'elasticsearch' -- ElasticSearch query -#if using 'elasticsearch' input format -#pir.esQuery= - -#esResource -- required if baseInputFormat = 'elasticsearch' -#Requires the format <index>/<type> : Elasticsearch resource where data is read and written to -#pir.esResource= - -#useHDFSLookupTable -- 'true' or 'false' - Whether or not to generate and use the -#hdfs lookup table for modular exponentiation -#Defaults to 'false' -#pir.useHDFSLookupTable= - -#baseQuery -- ElasticSearch-like query if using 'base' input format - -#used to filter records in the RecordReader -#Defaults to ?q=* -#pir.baseQuery= - -#limitHitsPerSelector -- 'true' or 'false' -#Whether or not to limit the number of hits per selector -#Defaults to 'true' -#pir.limitHitsPerSelector= - -#mapreduceMapJavaOpts -- Amount of heap (in MB) to allocate per map task -#Defaults to -Xmx2800m -#mapreduce.map.java.opts= - -#mapreduceMapMemoryMb -- Amount of memory (in MB) to allocate per map task -#Defaults to 3000 -#mapreduce.map.memory.mb= - -#mapreduceReduceJavaOpts -#Amount of heap (in MB) to allocate per reduce task -#Defaults to -Xmx2800m -#mapreduce.reduce.java.opts= - -#mapreduceReduceMemoryMb -#Amount of memory (in MB) to allocate per reduce task -#Defaults to 3000 -#mapreduce.reduce.memory.mb= - -#stopListFile -- optional (unless using StopListFilter) -- Fully qualified file in hdfs -#containing stoplist terms; used by the StopListFilter -#pir.stopListFile= - -#useLocalCache -- 'true' or 'false' -#Whether or not to use the local cache for modular exponentiation -#Defaults to 'true' -#pir.useLocalCache= - -#useModExpJoin -- 'true' or 'false' -- Spark only -#Whether or not to pre-compute the modular exponentiation table and join it to the data -#partitions when performing the encrypted row calculations -#Defaults to 'false' -#pir.useModExpJoin= - -#numReduceTasks -- optional -- Number of reduce tasks -#pir.numReduceTasks= - -#numColMultPartitions -- optional, Spark only -#Number of partitions to use when performing column multiplication -#pir.numColMultPartitions= - -#maxHitsPerSelector -- optional -- Max number of hits encrypted per selector -#pir.maxHitsPerSelector= - -#dataParts -- optional -- Number of partitions for the input data -#pir.numDataPartitions= - -#numExpLookupPartitions -- optional -- Number of partitions for the exp lookup table -#pir.numExpLookupPartitions= - \ No newline at end of file
