Ack, I had wondered if it was something like that :) Is there a place where that is documented? (I suppose it's just not included in the ant test or ant test-commit suite?)
Thanks for the update on that. You just unwound a lot of consternation! 2012/1/11 Daniel Dai <[email protected]> > Hi, Jonathan, > TestOrderBy is an excluded test. It does not seems to be written in a right > way. It runs against MiniCluster, but it uses local file as input. This is > not supported even in current code. All regular tests runs fine for me. > > Daniel > > On Mon, Jan 9, 2012 at 9:54 PM, Jon Coveney <[email protected]> wrote: > > > Err, I mistyped. I did see the file permission error mentioned in the > Jura > > ticket, but setting umask 0022 fixed that and led to the wrongfs > exception > > > > Sent from my iPhone > > > > On Jan 9, 2012, at 9:42 PM, Daniel Dai <[email protected]> wrote: > > > > > Hi, Jonathan, > > > What is the file permissions error you saw? Your previous log only > shows > > > WrongFS exception. > > > > > > Daniel > > > > > > On Mon, Jan 9, 2012 at 8:01 PM, Jonathan Coveney <[email protected]> > > wrote: > > > > > >> I'm not getting the file permissions error anymore. I tried > TestOrderBy > > >> (which still has some failing tests, but not related to this) and > > TestUDF. > > >> > > >> 2012/1/9 Jonathan Coveney <[email protected]> > > >> > > >>> Running now, will let you know. > > >>> > > >>> > > >>> 2012/1/9 Daniel Dai <[email protected]> > > >>> > > >>>> I just rolled back 205 changes. Can you try if tests pass? > > >>>> > > >>>> Thanks, > > >>>> Daniel > > >>>> > > >>>> On Mon, Jan 9, 2012 at 6:57 PM, Jonathan Coveney < > [email protected]> > > >>>> wrote: > > >>>> > > >>>>> In my /etc/hosts, oauth.localhost.twitter.com is redirected to > > >>>>> 127.0.0.1...but maybe the presence of this mapping makes it think > > that > > >>>> the > > >>>>> cluster is remote, and not local? I should note that I've always > had > > >>>> this > > >>>>> set in my /etc/hosts, but only Friday started having issues. > > >>>>> > > >>>>> I removed just that line, and it chose a random other one. I > removed > > >>>> all of > > >>>>> the mappings to 127.0.0.1 except localhost, and here is the error I > > >> got: > > >>>>> > > >>>>> org.apache.pig.backend.executionengine.ExecException: ERROR 2118: > > >> Wrong > > >>>> FS: > > >>>>> hdfs://localhost:60145/user/jcoveney, expected: file:/// > > >>>>> > > >>>>> Just to see, I got rid of all of my mappings in /etc/hosts (though > > >>>> mapping > > >>>>> localhost to 127.0.0.1 is totally valid), and got this error: > > >>>>> > > >>>>> org.apache.pig.backend.executionengine.ExecException: ERROR 2118: > > >> Wrong > > >>>> FS: > > >>>>> hdfs://127.0.0.1:60420/user/jcoveney, expected: file:/// > > >>>>> > > >>>>> So It's not related to that. For whatever reason, it's treating it > > >> like > > >>>> an > > >>>>> hdfs file system instead of just going for file:/// > > >>>>> > > >>>>> 2012/1/9 Daniel Dai <[email protected]> > > >>>>> > > >>>>>> Thanks Jonathan, > > >>>>>> I saw this in the log: > > >>>>>> Wrong FS: hdfs://oauth.localhost.twitter.com:55237/user/jcoveney, > > >>>>>> expected: > > >>>>>> file:/// > > >>>>>> > > >>>>>> Where does this host come from? Is there a hadoop config file in > > >> your > > >>>>>> CLASSPATH? > > >>>>>> > > >>>>>> Daniel > > >>>>>> > > >>>>>> On Mon, Jan 9, 2012 at 5:46 PM, Jonathan Coveney < > > >> [email protected]> > > >>>>>> wrote: > > >>>>>> > > >>>>>>> Attached is the log for ant -Dtestcase=TestOrderBy test on a > clean > > >>>>> clone > > >>>>>>> of trunk. > > >>>>>>> > > >>>>>>> 2012/1/9 Daniel Dai <[email protected]> > > >>>>>>> > > >>>>>>>> Hi, Jonathan, can you paste your error message? > > >>>>>>>> > > >>>>>>>> Daniel > > >>>>>>>> > > >>>>>>>> On Mon, Jan 9, 2012 at 4:47 PM, Jonathan Coveney < > > >>>> [email protected]> > > >>>>>>>> wrote: > > >>>>>>>> > > >>>>>>>>> Aha! I've been having a bunch of unit tests fail mysteriously, > > >>>> and > > >>>>> it > > >>>>>>>>> started Friday..and they've been giving local filesystem > > >>>> permissions > > >>>>>>>>> errors. Seems like it is related to that. I was going crazy. > > >>>>>>>>> > > >>>>>>>>> 2012/1/9 Daniel Dai <[email protected]> > > >>>>>>>>> > > >>>>>>>>>> Is that something new? Are you testing trunk? What's the > > >>>> revision > > >>>>>>>> number? > > >>>>>>>>>> Could that relate to 205 upgrade? > > >>>>>>>>>> (PIG-2431<https://issues.apache.org/jira/browse/PIG-2431>). > > >>>>>>>>>> > > >>>>>>>>>> > > >>>>>>>>>> Daniel > > >>>>>>>>>> > > >>>>>>>>>> On Mon, Jan 9, 2012 at 3:54 PM, Bill Graham < > > >>>> [email protected] > > >>>>>> > > >>>>>>>>> wrote: > > >>>>>>>>>> > > >>>>>>>>>>> Is anyone else seeing a bunch of errors when trying to run > > >>>> the > > >>>>> pig > > >>>>>>>> or > > >>>>>>>>>>> piggybank unit tests? I had to change ivy to use xercesImpl > > >>>>> 2.9.1 > > >>>>>>>>> instead > > >>>>>>>>>>> of xerces 2.4.4 to make this exception go away: > > >>>>>>>>>>> > > >>>>>>>>>>> 12/01/09 15:47:52 ERROR conf.Configuration: Failed to set > > >>>>>>>>>>> setXIncludeAware(true) for parser > > >>>>>>>>>>> org.apache.xerces.jaxp.DocumentBuilderFactoryImpl@358b3364 > > >>>>>>>>>>> :java.lang.UnsupportedOperationException: > > >>>>>>>>>>> This parser does not support specification "null" version > > >>>> "null" > > >>>>>>>>>>> java.lang.UnsupportedOperationException: This parser does > > >> not > > >>>>>>>> support > > >>>>>>>>>>> specification "null" version "null" > > >>>>>>>>>>> at > > >>>>>>>>>>> > > >>>>>>>>>>> > > >>>>>>>>>> > > >>>>>>>>> > > >>>>>>>> > > >>>>>> > > >>>>> > > >>>> > > >> > > > javax.xml.parsers.DocumentBuilderFactory.setXIncludeAware(DocumentBuilderFactory.java:590) > > >>>>>>>>>>> at > > >>>>>>>>>>> > > >>>>>>>>>> > > >>>>>>>>> > > >>>>>>>> > > >>>>>> > > >>>>> > > >>>> > > >> > > > org.apache.hadoop.conf.Configuration.loadResource(Configuration.java:1117) > > >>>>>>>>>>> at > > >>>>>>>>>>> > > >>>>>>>>>> > > >>>>>>>>> > > >>>>>>>> > > >>>>>> > > >>>>> > > >>>> > > >> > > > org.apache.hadoop.conf.Configuration.loadResources(Configuration.java:1103) > > >>>>>>>>>>> at > > >>>>>>>>>>> > > >>>>>>>> > > >>>> > org.apache.hadoop.conf.Configuration.getProps(Configuration.java:1037) > > >>>>>>>>>>> at > > >>>>>>>>>>> > > >>>>>>>> > > >>>> > org.apache.hadoop.conf.Configuration.iterator(Configuration.java:1079) > > >>>>>>>>>>> at > > >>>>>>>>>>> > > >>>>>>>>>>> > > >>>>>>>>>> > > >>>>>>>>> > > >>>>>>>> > > >>>>>> > > >>>>> > > >>>> > > >> > > > org.apache.pig.backend.hadoop.executionengine.HExecutionEngine.recomputeProperties(HExecutionEngine.java:366) > > >>>>>>>>>>> at > > >>>>>>>>>>> > > >>>>>>>>>>> > > >>>>>>>>>> > > >>>>>>>>> > > >>>>>>>> > > >>>>>> > > >>>>> > > >>>> > > >> > > > org.apache.pig.backend.hadoop.executionengine.HExecutionEngine.init(HExecutionEngine.java:177) > > >>>>>>>>>>> at > > >>>>>>>>>>> > > >>>>>>>>>>> > > >>>>>>>>>> > > >>>>>>>>> > > >>>>>>>> > > >>>>>> > > >>>>> > > >>>> > > >> > > > org.apache.pig.backend.hadoop.executionengine.HExecutionEngine.init(HExecutionEngine.java:119) > > >>>>>>>>>>> at > > >>>>>>>> org.apache.pig.impl.PigContext.connect(PigContext.java:206) > > >>>>>>>>>>> at > > >> org.apache.pig.PigServer.<init>(PigServer.java:246) > > >>>>>>>>>>> at > > >> org.apache.pig.PigServer.<init>(PigServer.java:231) > > >>>>>>>>>>> at > > >> org.apache.pig.PigServer.<init>(PigServer.java:227) > > >>>>>>>>>>> at > > >> org.apache.pig.PigServer.<init>(PigServer.java:223) > > >>>>>>>>>>> ... > > >>>>>>>>>>> > > >>>>>>>>>>> thanks, > > >>>>>>>>>>> Bill > > >>>>>>>>>>> > > >>>>>>>>>> > > >>>>>>>>> > > >>>>>>>> > > >>>>>>> > > >>>>>>> > > >>>>>> > > >>>>> > > >>>> > > >>> > > >>> > > >> > > >
