Running now, will let you know. 2012/1/9 Daniel Dai <[email protected]>
> I just rolled back 205 changes. Can you try if tests pass? > > Thanks, > Daniel > > On Mon, Jan 9, 2012 at 6:57 PM, Jonathan Coveney <[email protected]> > wrote: > > > In my /etc/hosts, oauth.localhost.twitter.com is redirected to > > 127.0.0.1...but maybe the presence of this mapping makes it think that > the > > cluster is remote, and not local? I should note that I've always had this > > set in my /etc/hosts, but only Friday started having issues. > > > > I removed just that line, and it chose a random other one. I removed all > of > > the mappings to 127.0.0.1 except localhost, and here is the error I got: > > > > org.apache.pig.backend.executionengine.ExecException: ERROR 2118: Wrong > FS: > > hdfs://localhost:60145/user/jcoveney, expected: file:/// > > > > Just to see, I got rid of all of my mappings in /etc/hosts (though > mapping > > localhost to 127.0.0.1 is totally valid), and got this error: > > > > org.apache.pig.backend.executionengine.ExecException: ERROR 2118: Wrong > FS: > > hdfs://127.0.0.1:60420/user/jcoveney, expected: file:/// > > > > So It's not related to that. For whatever reason, it's treating it like > an > > hdfs file system instead of just going for file:/// > > > > 2012/1/9 Daniel Dai <[email protected]> > > > > > Thanks Jonathan, > > > I saw this in the log: > > > Wrong FS: hdfs://oauth.localhost.twitter.com:55237/user/jcoveney, > > > expected: > > > file:/// > > > > > > Where does this host come from? Is there a hadoop config file in your > > > CLASSPATH? > > > > > > Daniel > > > > > > On Mon, Jan 9, 2012 at 5:46 PM, Jonathan Coveney <[email protected]> > > > wrote: > > > > > > > Attached is the log for ant -Dtestcase=TestOrderBy test on a clean > > clone > > > > of trunk. > > > > > > > > 2012/1/9 Daniel Dai <[email protected]> > > > > > > > >> Hi, Jonathan, can you paste your error message? > > > >> > > > >> Daniel > > > >> > > > >> On Mon, Jan 9, 2012 at 4:47 PM, Jonathan Coveney < > [email protected]> > > > >> wrote: > > > >> > > > >> > Aha! I've been having a bunch of unit tests fail mysteriously, and > > it > > > >> > started Friday..and they've been giving local filesystem > permissions > > > >> > errors. Seems like it is related to that. I was going crazy. > > > >> > > > > >> > 2012/1/9 Daniel Dai <[email protected]> > > > >> > > > > >> > > Is that something new? Are you testing trunk? What's the > revision > > > >> number? > > > >> > > Could that relate to 205 upgrade? > > > >> > > (PIG-2431<https://issues.apache.org/jira/browse/PIG-2431>). > > > >> > > > > > >> > > > > > >> > > Daniel > > > >> > > > > > >> > > On Mon, Jan 9, 2012 at 3:54 PM, Bill Graham < > [email protected] > > > > > > >> > wrote: > > > >> > > > > > >> > > > Is anyone else seeing a bunch of errors when trying to run the > > pig > > > >> or > > > >> > > > piggybank unit tests? I had to change ivy to use xercesImpl > > 2.9.1 > > > >> > instead > > > >> > > > of xerces 2.4.4 to make this exception go away: > > > >> > > > > > > >> > > > 12/01/09 15:47:52 ERROR conf.Configuration: Failed to set > > > >> > > > setXIncludeAware(true) for parser > > > >> > > > org.apache.xerces.jaxp.DocumentBuilderFactoryImpl@358b3364 > > > >> > > > :java.lang.UnsupportedOperationException: > > > >> > > > This parser does not support specification "null" version > "null" > > > >> > > > java.lang.UnsupportedOperationException: This parser does not > > > >> support > > > >> > > > specification "null" version "null" > > > >> > > > at > > > >> > > > > > > >> > > > > > > >> > > > > > >> > > > > >> > > > > > > javax.xml.parsers.DocumentBuilderFactory.setXIncludeAware(DocumentBuilderFactory.java:590) > > > >> > > > at > > > >> > > > > > > >> > > > > > >> > > > > >> > > > > > > org.apache.hadoop.conf.Configuration.loadResource(Configuration.java:1117) > > > >> > > > at > > > >> > > > > > > >> > > > > > >> > > > > >> > > > > > > org.apache.hadoop.conf.Configuration.loadResources(Configuration.java:1103) > > > >> > > > at > > > >> > > > > > > >> > org.apache.hadoop.conf.Configuration.getProps(Configuration.java:1037) > > > >> > > > at > > > >> > > > > > > >> > org.apache.hadoop.conf.Configuration.iterator(Configuration.java:1079) > > > >> > > > at > > > >> > > > > > > >> > > > > > > >> > > > > > >> > > > > >> > > > > > > org.apache.pig.backend.hadoop.executionengine.HExecutionEngine.recomputeProperties(HExecutionEngine.java:366) > > > >> > > > at > > > >> > > > > > > >> > > > > > > >> > > > > > >> > > > > >> > > > > > > org.apache.pig.backend.hadoop.executionengine.HExecutionEngine.init(HExecutionEngine.java:177) > > > >> > > > at > > > >> > > > > > > >> > > > > > > >> > > > > > >> > > > > >> > > > > > > org.apache.pig.backend.hadoop.executionengine.HExecutionEngine.init(HExecutionEngine.java:119) > > > >> > > > at > > > >> org.apache.pig.impl.PigContext.connect(PigContext.java:206) > > > >> > > > at org.apache.pig.PigServer.<init>(PigServer.java:246) > > > >> > > > at org.apache.pig.PigServer.<init>(PigServer.java:231) > > > >> > > > at org.apache.pig.PigServer.<init>(PigServer.java:227) > > > >> > > > at org.apache.pig.PigServer.<init>(PigServer.java:223) > > > >> > > > ... > > > >> > > > > > > >> > > > thanks, > > > >> > > > Bill > > > >> > > > > > > >> > > > > > >> > > > > >> > > > > > > > > > > > > > >
