Author: aching
Date: Tue Apr 10 06:06:55 2012
New Revision: 1311583
URL: http://svn.apache.org/viewvc?rev=1311583&view=rev
Log:
GIRAPH-168: Simplify munge directive usage with new munge flag
HADOOP_SECURE (rather than HADOOP_FACEBOOK) and remove usage of
HADOOP (ekoontz via aching).
Modified:
incubator/giraph/trunk/CHANGELOG
incubator/giraph/trunk/README
incubator/giraph/trunk/pom.xml
incubator/giraph/trunk/src/main/java/org/apache/giraph/bsp/ImmutableOutputCommitter.java
incubator/giraph/trunk/src/main/java/org/apache/giraph/comm/BasicRPCCommunications.java
incubator/giraph/trunk/src/main/java/org/apache/giraph/comm/CommunicationsInterface.java
incubator/giraph/trunk/src/main/java/org/apache/giraph/comm/RPCCommunications.java
incubator/giraph/trunk/src/test/java/org/apache/giraph/TestBspBasic.java
Modified: incubator/giraph/trunk/CHANGELOG
URL:
http://svn.apache.org/viewvc/incubator/giraph/trunk/CHANGELOG?rev=1311583&r1=1311582&r2=1311583&view=diff
==============================================================================
--- incubator/giraph/trunk/CHANGELOG (original)
+++ incubator/giraph/trunk/CHANGELOG Tue Apr 10 06:06:55 2012
@@ -2,6 +2,10 @@ Giraph Change Log
Release 0.2.0 - unreleased
+ GIRAPH-168: Simplify munge directive usage with new munge flag
+ HADOOP_SECURE (rather than HADOOP_FACEBOOK) and remove usage of
+ HADOOP (ekoontz via aching).
+
GIRAPH-85: Simplify return expression in
RPCCommunications::getRPCProxy (Eli Reisman via jghoman)
Modified: incubator/giraph/trunk/README
URL:
http://svn.apache.org/viewvc/incubator/giraph/trunk/README?rev=1311583&r1=1311582&r2=1311583&view=diff
==============================================================================
--- incubator/giraph/trunk/README (original)
+++ incubator/giraph/trunk/README Tue Apr 10 06:06:55 2012
@@ -35,17 +35,37 @@ automatically take over if the current a
Hadoop versions for use with Giraph:
Secure Hadoop versions:
-- Apache Hadoop 0.20.203, 0.20.204, other secure versions may work as well
--- Other versions reported working include:
---- Cloudera CDH3u0, CDH3u1
+
+- Apache Hadoop 0.20.203.0
+
+ This is the default version used by Giraph: if you do not specify a
+profile with the -P flag, maven will use this version. You may also
+explicitly specify it with "mvn -Phadoop_0.20.203 <goals>".
+
+-Apache Hadoop 0.23.1
+
+ You may tell maven to use this version with "mvn -Phadoop_0.23 <goals>".
+
+-Apache Hadoop 3.0.0-SNAPSHOT
+
+ You may tell maven to use this version with "mvn -Phadoop_trunk <goals>".
Unsecure Hadoop versions:
+
- Apache Hadoop 0.20.1, 0.20.2, 0.20.3
-Facebook Hadoop release (https://github.com/facebook/hadoop-20-warehouse):
-- GitHub master
+ You may tell maven to use 0.20.2 with "mvn -Phadoop_non_secure <goals>".
+
+- Facebook Hadoop releases: https://github.com/facebook/hadoop-20, Master
branch
+
+ You may tell maven to use this version with:
+
+mvn -Phadoop_facebook
-Dhadoop.jar.path=/path/to/repo/build/hadoop-0.20.1-dev-core.jar <goals>
+
+-- Other versions reported working include:
+--- Cloudera CDH3u0, CDH3u1
-While we provide support for the unsecure and Facebook versions of Hadoop
+While we provide support for unsecure and Facebook versions of Hadoop
with the maven profiles 'hadoop_non_secure' and 'hadoop_facebook',
respectively, we have been primarily focusing on secure Hadoop releases
at this time.
@@ -67,14 +87,15 @@ Use the maven commands with secure Hadoo
- test (i.e. mvn test)
For the non-secure versions of Hadoop, run the maven commands with the
-additional argument '-Dhadoop=non_secure' to enable the maven profile
-'hadoop_non_secure'. An example compilation command is
-'mvn -Dhadoop=non_secure compile'.
+additional argument '-Dhadoop=non_secure' or '-Phadoop_non_secure' to enable
+ the maven profile 'hadoop_non_secure'. Example compilation commands are
+'mvn -Dhadoop=non_secure compile' and 'mvn -Phadoop_non_secure compile'.
For the Facebook Hadoop release, run the maven commands with the
-additional arguments '-Dhadoop=facebook' to enable the maven profile
-'hadoop_facebook' as well as a location for the hadoop core jar file. An
-example compilation command is 'mvn -Dhadoop=facebook
+additional arguments '-Dhadoop=facebook' or '-Phadoop_facebook' to enable
+the maven profile 'hadoop_facebook' as well as a location for the hadoop
+core jar file. Example compilation commands are 'mvn -Dhadoop=facebook
+-Dhadoop.jar.path=/tmp/hadoop-0.20.1-core.jar compile' or 'mvn
-Phadoop_facebook
-Dhadoop.jar.path=/tmp/hadoop-0.20.1-core.jar compile'.
Modified: incubator/giraph/trunk/pom.xml
URL:
http://svn.apache.org/viewvc/incubator/giraph/trunk/pom.xml?rev=1311583&r1=1311582&r2=1311583&view=diff
==============================================================================
--- incubator/giraph/trunk/pom.xml (original)
+++ incubator/giraph/trunk/pom.xml Tue Apr 10 06:06:55 2012
@@ -455,6 +455,42 @@ under the License.
<profiles>
<profile>
+ <id>hadoop_0.20.203</id>
+ <activation>
+ <activeByDefault>true</activeByDefault>
+ </activation>
+ <dependencies>
+ <dependency>
+ <groupId>org.apache.hadoop</groupId>
+ <artifactId>hadoop-core</artifactId>
+ <version>${hadoop.version}</version>
+ <scope>provided</scope>
+ </dependency>
+ </dependencies>
+
+ <build>
+ <plugins>
+ <plugin>
+ <groupId>org.sonatype.plugins</groupId>
+ <artifactId>munge-maven-plugin</artifactId>
+ <version>${munge-maven-plugin.version}</version>
+ <executions>
+ <execution>
+ <id>munge</id>
+ <phase>generate-sources</phase>
+ <goals>
+ <goal>munge</goal>
+ </goals>
+ </execution>
+ </executions>
+ <configuration>
+
<symbols>HADOOP_NON_SASL_RPC,HADOOP_NON_INTERVERSIONED_RPC</symbols>
+ </configuration>
+ </plugin>
+ </plugins>
+ </build>
+ </profile>
+ <profile>
<id>hadoop_non_secure</id>
<activation>
<property>
@@ -465,15 +501,15 @@ under the License.
<properties>
<hadoop.version>0.20.2</hadoop.version>
</properties>
+ <dependencies>
+ <dependency>
+ <groupId>org.apache.hadoop</groupId>
+ <artifactId>hadoop-core</artifactId>
+ <version>${hadoop.version}</version>
+ <scope>provided</scope>
+ </dependency>
+ </dependencies>
<build>
- <resources>
- <resource>
- <directory>src/main/java/org/apache/giraph/hadoop</directory>
- <excludes>
- <exclude>BspTokenSelector.java</exclude>
- </excludes>
- </resource>
- </resources>
<plugins>
<plugin>
<groupId>org.sonatype.plugins</groupId>
@@ -489,7 +525,7 @@ under the License.
</execution>
</executions>
<configuration>
- <symbols>HADOOP_NON_SECURE</symbols>
+
<symbols>HADOOP_NON_SECURE,HADOOP_NON_SASL_RPC,HADOOP_NON_INTERVERSIONED_RPC</symbols>
</configuration>
</plugin>
<plugin>
@@ -502,22 +538,8 @@ under the License.
</excludes>
<source>${compileSource}</source>
<target>${compileSource}</target>
- <showWarnings>true</showWarnings>
</configuration>
</plugin>
- <plugin>
- <groupId>org.apache.maven.plugins</groupId>
- <artifactId>maven-surefire-plugin</artifactId>
- <version>2.6</version>
- <configuration>
- <systemProperties>
- <property>
- <name>prop.jarLocation</name>
-
<value>target/munged/giraph-${project.version}-jar-with-dependencies.jar</value>
- </property>
- </systemProperties>
- </configuration>
- </plugin>
</plugins>
</build>
</profile>
@@ -538,6 +560,18 @@ under the License.
<scope>system</scope>
<systemPath>${hadoop.jar.path}</systemPath>
</dependency>
+ <dependency>
+ <groupId>commons-logging</groupId>
+ <artifactId>commons-logging-api</artifactId>
+ <version>1.0.4</version>
+ <scope>runtime</scope>
+ </dependency>
+ <dependency>
+ <groupId>commons-httpclient</groupId>
+ <artifactId>commons-httpclient</artifactId>
+ <version>3.0.1</version>
+ <scope>runtime</scope>
+ </dependency>
</dependencies>
<build>
<resources>
@@ -550,6 +584,18 @@ under the License.
</resources>
<plugins>
<plugin>
+ <groupId>org.apache.maven.plugins</groupId>
+ <artifactId>maven-compiler-plugin</artifactId>
+ <version>${maven-compiler-plugin.version}</version>
+ <configuration>
+ <excludes>
+ <exclude>**/BspTokenSelector.java</exclude>
+ </excludes>
+ <source>${compileSource}</source>
+ <target>${compileSource}</target>
+ </configuration>
+ </plugin>
+ <plugin>
<groupId>org.sonatype.plugins</groupId>
<artifactId>munge-maven-plugin</artifactId>
<version>${munge-maven-plugin.version}</version>
@@ -563,24 +609,107 @@ under the License.
</execution>
</executions>
<configuration>
- <symbols>HADOOP_FACEBOOK</symbols>
- </configuration>
- </plugin>
- <plugin>
- <groupId>org.apache.maven.plugins</groupId>
- <artifactId>maven-compiler-plugin</artifactId>
- <version>${maven-compiler-plugin.version}</version>
- <configuration>
- <excludes>
- <exclude>**/BspTokenSelector.java</exclude>
- </excludes>
- <source>${compileSource}</source>
- <target>${compileSource}</target>
+ <symbols>HADOOP_NON_SECURE,HADOOP_NON_SASL_RPC</symbols>
</configuration>
</plugin>
+ <plugin>
+ <groupId>org.apache.maven.plugins</groupId>
+ <artifactId>maven-surefire-plugin</artifactId>
+ <version>2.6</version>
+ <configuration>
+ <systemProperties>
+ <property>
+ <name>prop.jarLocation</name>
+
<value>../target/giraph-${project.version}-jar-with-dependencies.jar</value>
+ </property>
+ </systemProperties>
+ </configuration>
+ </plugin>
</plugins>
</build>
</profile>
+
+ <profile>
+ <id>hadoop_0.23</id>
+ <activation>
+ <property>
+ <name>hadoop</name>
+ <value>0.23</value>
+ </property>
+ </activation>
+ <properties>
+ <hadoop.version>0.23.1</hadoop.version>
+ </properties>
+ <dependencies>
+ <dependency>
+ <groupId>org.apache.hadoop</groupId>
+ <artifactId>hadoop-common</artifactId>
+ <version>${hadoop.version}</version>
+ <scope>provided</scope>
+ </dependency>
+ <dependency>
+ <groupId>org.apache.hadoop</groupId>
+ <artifactId>hadoop-mapreduce-client-core</artifactId>
+ <version>${hadoop.version}</version>
+ <scope>provided</scope>
+ </dependency>
+ <dependency>
+ <groupId>org.apache.hadoop</groupId>
+ <artifactId>hadoop-mapreduce-client-common</artifactId>
+ <version>${hadoop.version}</version>
+ </dependency>
+ </dependencies>
+ </profile>
+ <profile>
+ <id>hadoop_trunk</id>
+ <activation>
+ <property>
+ <name>hadoop</name>
+ <value>trunk</value>
+ </property>
+ </activation>
+ <properties>
+ <hadoop.version>3.0.0-SNAPSHOT</hadoop.version>
+ </properties>
+ <dependencies>
+ <dependency>
+ <groupId>org.apache.hadoop</groupId>
+ <artifactId>hadoop-common</artifactId>
+ <version>${hadoop.version}</version>
+ <scope>provided</scope>
+ </dependency>
+ <dependency>
+ <groupId>org.apache.hadoop</groupId>
+ <artifactId>hadoop-mapreduce-client-core</artifactId>
+ <version>${hadoop.version}</version>
+ <scope>provided</scope>
+ </dependency>
+ <dependency>
+ <groupId>org.apache.hadoop</groupId>
+ <artifactId>hadoop-mapreduce-client-common</artifactId>
+ <version>${hadoop.version}</version>
+ <scope>provided</scope>
+ </dependency>
+ <dependency>
+ <groupId>org.apache.hadoop</groupId>
+ <artifactId>hadoop-auth</artifactId>
+ <version>${hadoop.version}</version>
+ <scope>provided</scope>
+ </dependency>
+ <dependency>
+ <groupId>commons-configuration</groupId>
+ <artifactId>commons-configuration</artifactId>
+ <version>1.6</version>
+ <scope>runtime</scope>
+ </dependency>
+ <dependency>
+ <groupId>commons-httpclient</groupId>
+ <artifactId>commons-httpclient</artifactId>
+ <version>3.0.1</version>
+ <scope>runtime</scope>
+ </dependency>
+ </dependencies>
+ </profile>
</profiles>
<dependencies>
@@ -591,12 +720,6 @@ under the License.
<scope>test</scope>
</dependency>
<dependency>
- <groupId>org.apache.hadoop</groupId>
- <artifactId>hadoop-core</artifactId>
- <version>${hadoop.version}</version>
- <scope>provided</scope>
- </dependency>
- <dependency>
<groupId>org.codehaus.jackson</groupId>
<artifactId>jackson-core-asl</artifactId>
<version>${jackson.version}</version>
@@ -661,5 +784,5 @@ under the License.
<version>1.8.5</version>
<scope>test</scope>
</dependency>
- </dependencies>
+ </dependencies>
</project>
Modified:
incubator/giraph/trunk/src/main/java/org/apache/giraph/bsp/ImmutableOutputCommitter.java
URL:
http://svn.apache.org/viewvc/incubator/giraph/trunk/src/main/java/org/apache/giraph/bsp/ImmutableOutputCommitter.java?rev=1311583&r1=1311582&r2=1311583&view=diff
==============================================================================
---
incubator/giraph/trunk/src/main/java/org/apache/giraph/bsp/ImmutableOutputCommitter.java
(original)
+++
incubator/giraph/trunk/src/main/java/org/apache/giraph/bsp/ImmutableOutputCommitter.java
Tue Apr 10 06:06:55 2012
@@ -52,12 +52,11 @@ public class ImmutableOutputCommitter ex
public void setupTask(TaskAttemptContext context) throws IOException {
}
- /*if[HADOOP_NON_SECURE]
- @Override
- public void cleanupJob(JobContext jobContext) throws IOException {
- }
- else[HADOOP_NON_SECURE]*/
@Override
+ /*if[HADOOP_NON_SECURE]
+ public void cleanupJob(JobContext jobContext) throws IOException {
+ }
+ else[HADOOP_NON_SECURE]*/
/*end[HADOOP_NON_SECURE]*/
public void commitJob(JobContext jobContext) throws IOException {
}
Modified:
incubator/giraph/trunk/src/main/java/org/apache/giraph/comm/BasicRPCCommunications.java
URL:
http://svn.apache.org/viewvc/incubator/giraph/trunk/src/main/java/org/apache/giraph/comm/BasicRPCCommunications.java?rev=1311583&r1=1311582&r2=1311583&view=diff
==============================================================================
---
incubator/giraph/trunk/src/main/java/org/apache/giraph/comm/BasicRPCCommunications.java
(original)
+++
incubator/giraph/trunk/src/main/java/org/apache/giraph/comm/BasicRPCCommunications.java
Tue Apr 10 06:06:55 2012
@@ -61,9 +61,10 @@ import org.apache.giraph.utils.MemoryUti
import com.google.common.collect.Iterables;
-/*if[HADOOP_FACEBOOK]
+/*if[HADOOP_NON_INTERVERSIONED_RPC]
+else[HADOOP_NON_INTERVERSIONED_RPC]*/
import org.apache.hadoop.ipc.ProtocolSignature;
-end[HADOOP_FACEBOOK]*/
+/*end[HADOOP_NON_INTERVERSIONED_RPC]*/
/**
* Basic RPC communications object that implements the lower level operations
@@ -693,14 +694,24 @@ public abstract class BasicRPCCommunicat
return VERSION_ID;
}
- /*if[HADOOP_FACEBOOK]
- public ProtocolSignature getProtocolSignature(
- String protocol,
- long clientVersion,
- int clientMethodsHash) throws IOException {
- return new ProtocolSignature(versionID, null);
- }
-end[HADOOP_FACEBOOK]*/
+ /*if[HADOOP_NON_INTERVERSIONED_RPC]
+ else[HADOOP_NON_INTERVERSIONED_RPC]*/
+ /**
+ * Get the Protocol Signature for the given protocol,
+ * client version and method.
+ *
+ * @param protocol Protocol.
+ * @param clientVersion Version of Client.
+ * @param clientMethodsHash Hash of Client methods.
+ * @return ProtocolSignature for input parameters.
+ */
+ public ProtocolSignature getProtocolSignature(
+ String protocol,
+ long clientVersion,
+ int clientMethodsHash) throws IOException {
+ return new ProtocolSignature(VERSION_ID, null);
+ }
+ /*end[HADOOP_NON_INTERVERSIONED_RPC]*/
@Override
public void closeConnections() throws IOException {
Modified:
incubator/giraph/trunk/src/main/java/org/apache/giraph/comm/CommunicationsInterface.java
URL:
http://svn.apache.org/viewvc/incubator/giraph/trunk/src/main/java/org/apache/giraph/comm/CommunicationsInterface.java?rev=1311583&r1=1311582&r2=1311583&view=diff
==============================================================================
---
incubator/giraph/trunk/src/main/java/org/apache/giraph/comm/CommunicationsInterface.java
(original)
+++
incubator/giraph/trunk/src/main/java/org/apache/giraph/comm/CommunicationsInterface.java
Tue Apr 10 06:06:55 2012
@@ -22,11 +22,11 @@ import java.io.IOException;
import org.apache.giraph.graph.Edge;
import org.apache.giraph.graph.BasicVertex;
-/*if_not[HADOOP]
- else[HADOOP]*/
+/*if[HADOOP_NON_SECURE]
+ else[HADOOP_NON_SECURE]*/
import org.apache.giraph.hadoop.BspTokenSelector;
import org.apache.hadoop.security.token.TokenInfo;
-/*end[HADOOP]*/
+/*end[HADOOP_NON_SECURE]*/
import org.apache.hadoop.io.Writable;
import org.apache.hadoop.io.WritableComparable;
import org.apache.hadoop.ipc.VersionedProtocol;
@@ -40,10 +40,10 @@ import org.apache.hadoop.ipc.VersionedPr
* @param <M> Message data
*/
@SuppressWarnings("rawtypes")
-/*if_not[HADOOP]
- else[HADOOP]*/
+/*if[HADOOP_NON_SECURE]
+ else[HADOOP_NON_SECURE]*/
@TokenInfo(BspTokenSelector.class)
-/*end[HADOOP]*/
+/*end[HADOOP_NON_SECURE]*/
public interface CommunicationsInterface<I extends WritableComparable,
V extends Writable, E extends Writable, M extends Writable>
extends VersionedProtocol {
Modified:
incubator/giraph/trunk/src/main/java/org/apache/giraph/comm/RPCCommunications.java
URL:
http://svn.apache.org/viewvc/incubator/giraph/trunk/src/main/java/org/apache/giraph/comm/RPCCommunications.java?rev=1311583&r1=1311582&r2=1311583&view=diff
==============================================================================
---
incubator/giraph/trunk/src/main/java/org/apache/giraph/comm/RPCCommunications.java
(original)
+++
incubator/giraph/trunk/src/main/java/org/apache/giraph/comm/RPCCommunications.java
Tue Apr 10 06:06:55 2012
@@ -22,9 +22,10 @@ import java.io.IOException;
import java.net.InetSocketAddress;
-/*if_not[HADOOP]
-else[HADOOP]*/
+/*if[HADOOP_NON_SECURE]
+else[HADOOP_NON_SECURE]*/
import java.security.PrivilegedExceptionAction;
+import org.apache.hadoop.mapred.JobConf;
import org.apache.hadoop.mapreduce.security.TokenCache;
import org.apache.hadoop.mapreduce.security.token.JobTokenIdentifier;
import org.apache.hadoop.mapreduce.security.token.JobTokenSecretManager;
@@ -32,15 +33,21 @@ import org.apache.hadoop.security.Creden
import org.apache.hadoop.security.UserGroupInformation;
import org.apache.hadoop.security.authorize.ServiceAuthorizationManager;
import org.apache.hadoop.security.token.Token;
-/*end[HADOOP]*/
+/*end[HADOOP_NON_SECURE]*/
import org.apache.log4j.Logger;
import org.apache.giraph.bsp.CentralizedServiceWorker;
import org.apache.giraph.graph.GraphState;
+/*if[HADOOP_NON_SECURE]
+else[HADOOP_NON_SECURE]*/
import org.apache.giraph.hadoop.BspPolicyProvider;
+/*end[HADOOP_NON_SECURE]*/
import org.apache.hadoop.conf.Configuration;
+/*if[HADOOP_NON_SECURE]
+else[HADOOP_NON_SECURE]*/
import org.apache.hadoop.io.Text;
+/*end[HADOOP_NON_SECURE]*/
import org.apache.hadoop.io.Writable;
import org.apache.hadoop.io.WritableComparable;
import org.apache.hadoop.ipc.RPC;
@@ -58,11 +65,11 @@ import org.apache.hadoop.mapreduce.Mappe
@SuppressWarnings("rawtypes")
public class RPCCommunications<I extends WritableComparable,
V extends Writable, E extends Writable, M extends Writable>
- /*if_not[HADOOP]
+ /*if[HADOOP_NON_SASL_RPC]
extends BasicRPCCommunications<I, V, E, M, Object> {
- else[HADOOP]*/
+ else[HADOOP_NON_SASL_RPC]*/
extends BasicRPCCommunications<I, V, E, M, Token<JobTokenIdentifier>> {
- /*end[HADOOP]*/
+ /*end[HADOOP_NON_SASL_RPC]*/
/** Class logger */
public static final Logger LOG = Logger.getLogger(RPCCommunications.class);
@@ -84,27 +91,30 @@ public class RPCCommunications<I extends
super(context, service);
}
- /*if_not[HADOOP]
- protected Object createJobToken() throws IOException {
- return null;
- }
- else[HADOOP]*/
/**
- * Create the job token.
- *
- * @return Job token.
- */
- protected Token<JobTokenIdentifier> createJobToken() throws IOException {
+ * Create the job token.
+ *
+ * @return Job token.
+ */
+ protected
+ /*if[HADOOP_NON_SECURE]
+ Object createJobToken() throws IOException {
+ else[HADOOP_NON_SECURE]*/
+ Token<JobTokenIdentifier> createJobToken() throws IOException {
+ /*end[HADOOP_NON_SECURE]*/
+ /*if[HADOOP_NON_SECURE]
+ else[HADOOP_NON_SECURE]*/
String localJobTokenFile = System.getenv().get(
UserGroupInformation.HADOOP_TOKEN_FILE_LOCATION);
if (localJobTokenFile != null) {
+ JobConf jobConf = new JobConf(conf);
Credentials credentials =
- TokenCache.loadTokens(localJobTokenFile, conf);
+ TokenCache.loadTokens(localJobTokenFile, jobConf);
return TokenCache.getJobToken(credentials);
}
+ /*end[HADOOP_NON_SECURE]*/
return null;
}
- /*end[HADOOP]*/
/**
* Get the RPC server.
@@ -115,23 +125,16 @@ public class RPCCommunications<I extends
* @param jt Jobtoken indentifier.
* @return RPC server.
*/
+ @Override
protected Server getRPCServer(
InetSocketAddress myAddress, int numHandlers, String jobId,
- /*if_not[HADOOP]
- Object jt) throws IOException {
- return RPC.getServer(this, myAddress.getHostName(),
myAddress.getPort(),
- numHandlers, false, conf);
- }
- else[HADOOP]*/
+ /*if[HADOOP_NON_SASL_RPC]
+ Object jt) throws IOException {
+ return RPC.getServer(this, myAddress.getHostName(), myAddress.getPort(),
+ numHandlers, false, conf);
+ else[HADOOP_NON_SASL_RPC]*/
Token<JobTokenIdentifier> jt) throws IOException {
@SuppressWarnings("deprecation")
- String hadoopSecurityAuthorization =
- ServiceAuthorizationManager.SERVICE_AUTHORIZATION_CONFIG;
- if (conf.getBoolean(
- hadoopSecurityAuthorization,
- false)) {
- ServiceAuthorizationManager.refresh(conf, new BspPolicyProvider());
- }
JobTokenSecretManager jobTokenSecretManager =
new JobTokenSecretManager();
if (jt != null) { //could be null in the case of some unit tests
@@ -140,10 +143,18 @@ public class RPCCommunications<I extends
LOG.info("getRPCServer: Added jobToken " + jt);
}
}
- return RPC.getServer(this, myAddress.getHostName(), myAddress.getPort(),
- numHandlers, false, conf, jobTokenSecretManager);
+ Server server = RPC.getServer(RPCCommunications.class, this,
+ myAddress.getHostName(), myAddress.getPort(),
+ numHandlers, false, conf, jobTokenSecretManager);
+ String hadoopSecurityAuthorization =
+ ServiceAuthorizationManager.SERVICE_AUTHORIZATION_CONFIG;
+ if (conf.getBoolean(hadoopSecurityAuthorization, false)) {
+ server.refreshServiceAcl(conf, new BspPolicyProvider());
+ }
+ return server;
+ /*end[HADOOP_NON_SASL_RPC]*/
}
- /*end[HADOOP]*/
+
/**
* Get the RPC proxy.
@@ -157,17 +168,17 @@ public class RPCCommunications<I extends
protected CommunicationsInterface<I, V, E, M> getRPCProxy(
final InetSocketAddress addr,
String jobId,
- /*if_not[HADOOP]
+ /*if[HADOOP_NON_SASL_RPC]
Object jt)
- else[HADOOP]*/
+ else[HADOOP_NON_SASL_RPC]*/
Token<JobTokenIdentifier> jt)
- /*end[HADOOP]*/
+ /*end[HADOOP_NON_SASL_RPC]*/
throws IOException, InterruptedException {
final Configuration config = new Configuration(conf);
- /*if_not[HADOOP]
+ /*if[HADOOP_NON_SASL_RPC]
return (CommunicationsInterface<I, V, E, M>)RPC.getProxy(
CommunicationsInterface.class, VERSION_ID, addr, config);
- else[HADOOP]*/
+ else[HADOOP_NON_SASL_RPC]*/
if (jt == null) {
return (CommunicationsInterface<I, V, E, M>) RPC.getProxy(
CommunicationsInterface.class, VERSION_ID, addr, config);
@@ -190,6 +201,6 @@ public class RPCCommunications<I extends
CommunicationsInterface.class, VERSION_ID, addr, config);
}
});
- /*end[HADOOP]*/
+ /*end[HADOOP_NON_SASL_RPC]*/
}
}
Modified:
incubator/giraph/trunk/src/test/java/org/apache/giraph/TestBspBasic.java
URL:
http://svn.apache.org/viewvc/incubator/giraph/trunk/src/test/java/org/apache/giraph/TestBspBasic.java?rev=1311583&r1=1311582&r2=1311583&view=diff
==============================================================================
--- incubator/giraph/trunk/src/test/java/org/apache/giraph/TestBspBasic.java
(original)
+++ incubator/giraph/trunk/src/test/java/org/apache/giraph/TestBspBasic.java
Tue Apr 10 06:06:55 2012
@@ -52,7 +52,10 @@ import org.apache.hadoop.io.Writable;
import org.apache.hadoop.mapreduce.InputSplit;
import org.apache.hadoop.mapreduce.JobContext;
import org.apache.hadoop.mapreduce.JobID;
-
+/*if[HADOOP_NON_SASL_RPC]
+else[HADOOP_NON_SASL_RPC]*/
+import org.apache.hadoop.mapreduce.task.JobContextImpl;
+/*end[HADOOP_NON_SASL_RPC]*/
import java.io.ByteArrayOutputStream;
import java.io.DataOutputStream;
import java.io.IOException;
@@ -112,9 +115,15 @@ public class TestBspBasic extends BspCas
", graphState" + gs);
VertexInputFormat<LongWritable, IntWritable, FloatWritable, IntWritable>
inputFormat = BspUtils.createVertexInputFormat(job.getConfiguration());
- List<InputSplit> splitArray =
- inputFormat.getSplits(
- new JobContext(new Configuration(), new JobID()), 1);
+ /*if[HADOOP_NON_SASL_RPC]
+ List<InputSplit> splitArray =
+ inputFormat.getSplits(
+ new JobContext(new Configuration(), new JobID()), 1);
+ else[HADOOP_NON_SASL_RPC]*/
+ List<InputSplit> splitArray =
+ inputFormat.getSplits(
+ new JobContextImpl(new Configuration(), new JobID()), 1);
+ /*end[HADOOP_NON_SASL_RPC]*/
ByteArrayOutputStream byteArrayOutputStream =
new ByteArrayOutputStream();
DataOutputStream outputStream =