[jira] [Commented] (NUTCH-1129) Any23 Nutch plugin
[ https://issues.apache.org/jira/browse/NUTCH-1129?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13208309#comment-13208309 ] Markus Jelsma commented on NUTCH-1129: -- This is a parser plugin right? How will this work if we for example would like to parse microdata with any23 and use Tika's BoilerpipeContentHandler to extraction? In the current BP patch we use multiple content handlers to parse all in one go so i wonder if this could be implemented as such. Please correct me when wrong :) Any23 Nutch plugin -- Key: NUTCH-1129 URL: https://issues.apache.org/jira/browse/NUTCH-1129 Project: Nutch Issue Type: New Feature Components: parser Reporter: Lewis John McGibbney Assignee: Lewis John McGibbney Priority: Minor Fix For: 1.5 Attachments: NUTCH-1129.patch This plugin should build on the Any23 library to provide us with a plugin which extracts RDF data from HTTP and file resources. Although as of writing Any23 not part of the ASF, the project is working towards integration into the Apache Incubator. Once the project proves its value, this would be an excellent addition to the Nutch 1.X codebase. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (NUTCH-1279) Check if limit has been reached in GeneraterReducer must be the first check performance-wise.
[ https://issues.apache.org/jira/browse/NUTCH-1279?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Ferdy Galema updated NUTCH-1279: Attachment: NUTCH-1279.txt Attached patch and committed. Check if limit has been reached in GeneraterReducer must be the first check performance-wise. - Key: NUTCH-1279 URL: https://issues.apache.org/jira/browse/NUTCH-1279 Project: Nutch Issue Type: Improvement Components: generator Reporter: Ferdy Galema Priority: Minor Fix For: nutchgora Attachments: NUTCH-1279.txt The (count = limit) should be put up front in the reduce method of the generator, because that way when the limit is reached the reduce method will return faster. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (NUTCH-1279) Check if limit has been reached in GeneraterReducer must be the first check performance-wise.
[ https://issues.apache.org/jira/browse/NUTCH-1279?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13208341#comment-13208341 ] Lewis John McGibbney commented on NUTCH-1279: - Hi Ferdy, have you checked whether this is the case in trunk as well? I know the fetcher architecture is slightly different. Check if limit has been reached in GeneraterReducer must be the first check performance-wise. - Key: NUTCH-1279 URL: https://issues.apache.org/jira/browse/NUTCH-1279 Project: Nutch Issue Type: Improvement Components: generator Reporter: Ferdy Galema Priority: Minor Fix For: nutchgora Attachments: NUTCH-1279.txt The (count = limit) should be put up front in the reduce method of the generator, because that way when the limit is reached the reduce method will return faster. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
Re: Detecting Encoding with plugins
Hi Markus, I've been vaguely keeping up with yourself and Julien's work on this. I would really like to get a test case for this though! I'll try working towards this as a sub-target of another issue. For reference, there is a Tika mimeType test case here [1] and Tika document encoding test here [2]. Which we may or may not be interested in porting over to o.a.n? wdyt? Thanks Lewis [1] https://svn.apache.org/viewvc/incubator/any23/trunk/core/src/test/java/org/apache/any23/mime/TikaMIMETypeDetectorTest.java?view=markup [2] https://svn.apache.org/viewvc/incubator/any23/trunk/core/src/test/java/org/apache/any23/encoding/TikaEncodingDetectorTest.java?view=markup On Tue, Feb 14, 2012 at 11:51 PM, Markus Jelsma mar...@apache.org wrote: Hi, This was indeed an issue until today. The detected type is in the crawl datum metadata. https://issues.apache.org/jira/browse/NUTCH-1259 Hi, I can't see anywhere within our parser plugins where we detect encoding of documents. I've also begun looking through the o.a.n.p package but again I can't see anything. Can anyone provide some detail on this please? Thank you Lewis -- *Lewis*
[jira] [Commented] (NUTCH-1279) Check if limit has been reached in GeneraterReducer must be the first check performance-wise.
[ https://issues.apache.org/jira/browse/NUTCH-1279?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13208354#comment-13208354 ] Ferdy Galema commented on NUTCH-1279: - Hi Lewis, I can confirm that trunk already uses the same optimization. Check if limit has been reached in GeneraterReducer must be the first check performance-wise. - Key: NUTCH-1279 URL: https://issues.apache.org/jira/browse/NUTCH-1279 Project: Nutch Issue Type: Improvement Components: generator Reporter: Ferdy Galema Priority: Minor Fix For: nutchgora Attachments: NUTCH-1279.txt The (count = limit) should be put up front in the reduce method of the generator, because that way when the limit is reached the reduce method will return faster. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
Re: Detecting Encoding with plugins
The mimetype is not the same thing as the encoding. As Ken pointed out this is done at the individual parser level On 14 February 2012 23:51, Markus Jelsma mar...@apache.org wrote: Hi, This was indeed an issue until today. The detected type is in the crawl datum metadata. https://issues.apache.org/jira/browse/NUTCH-1259 Hi, I can't see anywhere within our parser plugins where we detect encoding of documents. I've also begun looking through the o.a.n.p package but again I can't see anything. Can anyone provide some detail on this please? Thank you Lewis -- * *Open Source Solutions for Text Engineering http://digitalpebble.blogspot.com/ http://www.digitalpebble.com http://twitter.com/digitalpebble
Re: Detecting Encoding with plugins
Yes this is correct, but we still don't test for either of the two. On Wed, Feb 15, 2012 at 10:59 AM, Julien Nioche lists.digitalpeb...@gmail.com wrote: The mimetype is not the same thing as the encoding. As Ken pointed out this is done at the individual parser level On 14 February 2012 23:51, Markus Jelsma mar...@apache.org wrote: Hi, This was indeed an issue until today. The detected type is in the crawl datum metadata. https://issues.apache.org/jira/browse/NUTCH-1259 Hi, I can't see anywhere within our parser plugins where we detect encoding of documents. I've also begun looking through the o.a.n.p package but again I can't see anything. Can anyone provide some detail on this please? Thank you Lewis -- * *Open Source Solutions for Text Engineering http://digitalpebble.blogspot.com/ http://www.digitalpebble.com http://twitter.com/digitalpebble -- *Lewis*
[jira] [Commented] (NUTCH-1278) Fetch Improvement in threads per host
[ https://issues.apache.org/jira/browse/NUTCH-1278?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13208413#comment-13208413 ] Lewis John McGibbney commented on NUTCH-1278: - Hi Behnam. Do you have a patch for trunk? Thank you Fetch Improvement in threads per host - Key: NUTCH-1278 URL: https://issues.apache.org/jira/browse/NUTCH-1278 Project: Nutch Issue Type: New Feature Components: fetcher Affects Versions: 1.4 Reporter: behnam nikbakht the value of maxThreads is equal to fetcher.threads.per.host and is constant for every host there is a possibility with using of dynamic values for every host that influeced with number of blocked requests. this means that if number of blocked requests for one host increased, then we most decrease this value and increase http.timeout -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
Re: Detecting Encoding with plugins
I assume Tika does already - why should we duplicate the tests in Nutch? we delegate the functionality to Tika, IMHO this means delegating the testing as well. What we could do to contribute tests to Tika instead if it does not have any. Re-any23 : why not handling it as a Tika parser instead of a Nutch one? This could be useful to other Tika users who do not necessarily use Nutch Julien On 15 February 2012 12:17, Lewis John Mcgibbney lewis.mcgibb...@gmail.comwrote: Yes this is correct, but we still don't test for either of the two. On Wed, Feb 15, 2012 at 10:59 AM, Julien Nioche lists.digitalpeb...@gmail.com wrote: The mimetype is not the same thing as the encoding. As Ken pointed out this is done at the individual parser level On 14 February 2012 23:51, Markus Jelsma mar...@apache.org wrote: Hi, This was indeed an issue until today. The detected type is in the crawl datum metadata. https://issues.apache.org/jira/browse/NUTCH-1259 Hi, I can't see anywhere within our parser plugins where we detect encoding of documents. I've also begun looking through the o.a.n.p package but again I can't see anything. Can anyone provide some detail on this please? Thank you Lewis -- * *Open Source Solutions for Text Engineering http://digitalpebble.blogspot.com/ http://www.digitalpebble.com http://twitter.com/digitalpebble -- *Lewis* -- * *Open Source Solutions for Text Engineering http://digitalpebble.blogspot.com/ http://www.digitalpebble.com http://twitter.com/digitalpebble
Re: Detecting Encoding with plugins
Hi Julien, On Wed, Feb 15, 2012 at 12:27 PM, Julien Nioche lists.digitalpeb...@gmail.com wrote: I assume Tika does already - why should we duplicate the tests in Nutch? We don't want to I suppose. However the point I was trying to make was that as NUTCH-1259 detects the encoding type, however we don't have an automated test to cover this, I assume the case is somewhat important or else the ticket for NUTCH-1259 wouldn't have been opened originally? I agree with you that general cases should be dealt with further upstream within Tika development itself, however as the encoding detection is done in Nutch within the cd metadata we may wish to get some test case to check... it's not a huge thing I suppose. we delegate the functionality to Tika, IMHO this means delegating the testing as well. What we could do to contribute tests to Tika instead if it does not have any. Yeah this is correct. I'm expecting you guys will know better than me but I would assume that Tika is mimetype and encoding detection compliant ;0) Re-any23 : why not handling it as a Tika parser instead of a Nutch one? This could be useful to other Tika users who do not necessarily use Nutch OK so I suppose this is completely open for discussion and I really welcome it as well. On one hand I see working with Any23 as a parse-any23 plugin within Nutch as the first step in the road to answering this question. Regardless of whether Any23 graduates and is integrated into Tika itself or as a TLP you are completely right that it should be made as openly available to as many people. Personally I agree with you Julien. One last thing, I know this if off topic... but with regards to our microformats-reltag plugin... I think the RelTagParser could and should be move over to Any23. Any23 already supports extraction of an number of microformats. wdyt? Thanks
Re: Detecting Encoding with plugins
Hi Lewis I assume Tika does already - why should we duplicate the tests in Nutch? We don't want to I suppose. However the point I was trying to make was that as NUTCH-1259 detects the encoding type, however we don't have an automated test to cover this, I assume the case is somewhat important or else the ticket for NUTCH-1259 wouldn't have been opened originally? nope. NUTCH-1259 is about storing the mime-type value detected by Tika. It is not the same as the encoding. This specific JIRA is not whether or not we get the correct value but a purely functional one about where we store it. There is not much to test wrt it I agree with you that general cases should be dealt with further upstream within Tika development itself, however as the encoding detection is done in Nutch within the cd metadata we may wish to get some test case to check... it's not a huge thing I suppose. we do have tests for the EncodingDetector (TestEncodingDetector), which is used by parse-html already. It is Ok to have that as it is our own parser. As explained earlier, for the Tika parser the detection is delegated to the Tika parser implementations and as such should be tested there. we delegate the functionality to Tika, IMHO this means delegating the testing as well. What we could do to contribute tests to Tika instead if it does not have any. Yeah this is correct. I'm expecting you guys will know better than me but I would assume that Tika is mimetype and encoding detection compliant ;0) I definitely do not pretend to know more than anyone else BTW :-) I don't understand what you mean by 'compliant'. Perfect? Probably not. There was an interesting experiment made by Ken on measuring the accuracy of the charset detection in the Tika book - which anyone remotely interested in Nutch should get BTW. There has been an interesting blog entry recently on comparing the language detection in Tika and other libraries (cant find ref and am in a hurry - sorry) Re-any23 : why not handling it as a Tika parser instead of a Nutch one? This could be useful to other Tika users who do not necessarily use Nutch OK so I suppose this is completely open for discussion and I really welcome it as well. On one hand I see working with Any23 as a parse-any23 plugin within Nutch as the first step in the road to answering this question. Regardless of whether Any23 graduates and is integrated into Tika itself or as a TLP you are completely right that it should be made as openly available to as many people. Personally I agree with you Julien. One last thing, I know this if off topic... but with regards to our microformats-reltag plugin... I think the RelTagParser could and should be move over to Any23. Any23 already supports extraction of an number of microformats. wdyt? it would probably make sense as an initial step if you don't want to venture in trying to wrap it as a Tika parser :-) Julien -- * *Open Source Solutions for Text Engineering http://digitalpebble.blogspot.com/ http://www.digitalpebble.com http://twitter.com/digitalpebble
[jira] [Updated] (NUTCH-1215) UpdateDB should not require segment as input
[ https://issues.apache.org/jira/browse/NUTCH-1215?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Markus Jelsma updated NUTCH-1215: - Attachment: NUTCH-1215-1.5-1.patch Patch for 1.5. Couldn't be simpler. UpdateDB should not require segment as input Key: NUTCH-1215 URL: https://issues.apache.org/jira/browse/NUTCH-1215 Project: Nutch Issue Type: Bug Components: linkdb Affects Versions: 1.4 Reporter: Markus Jelsma Fix For: 1.5 Attachments: NUTCH-1215-1.5-1.patch UpdateDB requires an input segment. This causes the metrics for the records of the segment to change, e.g. from fetched to not_modified and changes an adaptive fetch schedule accordingly. This should not happen when one needs to update for filtering of normalizing or other maintenance. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (NUTCH-1210) DomainBlacklistFilter
[ https://issues.apache.org/jira/browse/NUTCH-1210?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13208459#comment-13208459 ] Lewis John McGibbney commented on NUTCH-1210: - This looks really nice Markus. I like the documentation and test as well. I would like to try it out with another couple of test scenarios before passing my full opinion, which I will be able to do this afternoon. DomainBlacklistFilter - Key: NUTCH-1210 URL: https://issues.apache.org/jira/browse/NUTCH-1210 Project: Nutch Issue Type: New Feature Reporter: Markus Jelsma Assignee: Markus Jelsma Fix For: 1.5 Attachments: NUTCH-1210-1.5-1.patch The current DomainFilter acts as a white list. We also need a filter that acts as a black list so we can allow tld's and/or domains with DomainFilter but blacklist specific subdomains. If we would patch the current DomainFilter for this behaviour it would break current semantics such as it's precedence. Therefore i would propose a new filter instead. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (NUTCH-1262) Map `duplicating` content-types to a single type
[ https://issues.apache.org/jira/browse/NUTCH-1262?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13208461#comment-13208461 ] Markus Jelsma commented on NUTCH-1262: -- Is this issue still subject to debate? Opinions? Map `duplicating` content-types to a single type Key: NUTCH-1262 URL: https://issues.apache.org/jira/browse/NUTCH-1262 Project: Nutch Issue Type: Improvement Affects Versions: 1.4 Reporter: Markus Jelsma Assignee: Markus Jelsma Priority: Minor Fix For: 1.5 Attachments: NUTCH-1262-1.5-1.patch Similar or duplicating content-types can end-up differently in an index. With, for example, both application/xhtml+xml and text/html it is impossible to use a single filter to select `web pages`. See also: http://lucene.472066.n3.nabble.com/application-xhtml-xml-gt-text-html-td3699942.html -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
Re: XSD for Solr Schema
On Tuesday 14 February 2012 19:23:25 Lewis John Mcgibbney wrote: Hi, Whilst we were chatting about XSD, XLST etc the other night, I started thinking about using xalan or saxon to validate Solr schema's against some XSD. This would mean that a specific error is thrown which highlights where in the schema the problem(s) lie, I think this would provide a better level of user functionality for situations where you wish to have custom schema implementations (as many of us do). Do we currently have measures in place to catch an invalid schema case? Actually, yes. Solr usually throws a useful exception when something goes wrong. Thanks -- Markus Jelsma - CTO - Openindex
[jira] [Commented] (NUTCH-1210) DomainBlacklistFilter
[ https://issues.apache.org/jira/browse/NUTCH-1210?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13208881#comment-13208881 ] Lewis John McGibbney commented on NUTCH-1210: - Hi Markus. 1) I would ask one tiny change in ivy.xml from {code} configurations include file=${nutch.root}/ivy/ivy-configurations.xml/ /configurations {code} to {code} configurations include file=../../..//ivy/ivy-configurations.xml/ /configurations {code} this is purely for consistency as I think it's easier to configure in Eclipse as the ${nutch.root} variable hasn't been specified. 2) Also domainblacklist-urlfilter.txt is not included in the patch under /conf. Would it be possible to have a file there with some commented out documentation so users at least have something to go on? 3) Your documentation in the main class also mentions that the property can be overridden in nutch-*.xml, however no property exists in nutch-default for people to go on meaning that it is likely people will become confused when trying to set the property from nutch-site.xml. My tests seemt obe failing with trunk therefore there is something up with my trunk co, so I'll go get that sorted then test a bit more. Thanks DomainBlacklistFilter - Key: NUTCH-1210 URL: https://issues.apache.org/jira/browse/NUTCH-1210 Project: Nutch Issue Type: New Feature Reporter: Markus Jelsma Assignee: Markus Jelsma Fix For: 1.5 Attachments: NUTCH-1210-1.5-1.patch The current DomainFilter acts as a white list. We also need a filter that acts as a black list so we can allow tld's and/or domains with DomainFilter but blacklist specific subdomains. If we would patch the current DomainFilter for this behaviour it would break current semantics such as it's precedence. Therefore i would propose a new filter instead. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
Re: how linkdb impact on scores
hi Please submit your question to the user mailing list. There's more public there and this list is for Nutch development only. thanks there is a question about impact of linkdb on solrindex with test of solrindex, found that boost calculated from indexerScore, and pass inlinks to scoring plugins, but these plugins dont use of inlinks so, the results of solrindex with linkdb are equal to results without it. and with large linkdb this cause to slow solrindex. my question is where we can influence scoring with linkdb? -- View this message in context: http://lucene.472066.n3.nabble.com/how-linkdb-impact-on-scores-tp3746051p3 746051.html Sent from the Nutch - Dev mailing list archive at Nabble.com.
Build failed in Jenkins: Nutch-nutchgora #162
See https://builds.apache.org/job/Nutch-nutchgora/162/ -- Started by timer Building remotely on solaris1 in workspace https://builds.apache.org/job/Nutch-nutchgora/ws/ hudson.util.IOException2: remote file operation failed: https://builds.apache.org/job/Nutch-nutchgora/ws/ at hudson.remoting.Channel@50e71521:solaris1 at hudson.FilePath.act(FilePath.java:784) at hudson.FilePath.act(FilePath.java:770) at hudson.scm.SubversionSCM.checkout(SubversionSCM.java:742) at hudson.scm.SubversionSCM.checkout(SubversionSCM.java:684) at hudson.model.AbstractProject.checkout(AbstractProject.java:1195) at hudson.model.AbstractBuild$AbstractRunner.checkout(AbstractBuild.java:576) at hudson.model.AbstractBuild$AbstractRunner.run(AbstractBuild.java:465) at hudson.model.Run.run(Run.java:1409) at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:46) at hudson.model.ResourceController.execute(ResourceController.java:88) at hudson.model.Executor.run(Executor.java:238) Caused by: java.io.IOException: Remote call on solaris1 failed at hudson.remoting.Channel.call(Channel.java:690) at hudson.FilePath.act(FilePath.java:777) ... 10 more Caused by: java.lang.ClassFormatError: Failed to load javax.servlet.ServletException at hudson.remoting.RemoteClassLoader.loadClassFile(RemoteClassLoader.java:154) at hudson.remoting.RemoteClassLoader.findClass(RemoteClassLoader.java:131) at java.lang.ClassLoader.loadClass(ClassLoader.java:307) at java.lang.ClassLoader.loadClass(ClassLoader.java:252) at java.lang.ClassLoader.loadClassInternal(ClassLoader.java:320) at hudson.scm.SubversionWorkspaceSelector.syncWorkspaceFormatFromMaster(SubversionWorkspaceSelector.java:85) at hudson.scm.SubversionSCM.createSvnClientManager(SubversionSCM.java:822) at hudson.scm.SubversionSCM$CheckOutTask.invoke(SubversionSCM.java:765) at hudson.scm.SubversionSCM$CheckOutTask.invoke(SubversionSCM.java:752) at hudson.FilePath$FileCallableWrapper.call(FilePath.java:2099) at hudson.remoting.UserRequest.perform(UserRequest.java:118) at hudson.remoting.UserRequest.perform(UserRequest.java:48) at hudson.remoting.Request$2.run(Request.java:287) at hudson.remoting.InterceptingExecutorService$1.call(InterceptingExecutorService.java:72) at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:269) at java.util.concurrent.FutureTask.run(FutureTask.java:123) at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:651) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:676) at java.lang.Thread.run(Thread.java:595) Caused by: java.lang.UnsupportedClassVersionError: Bad version number in .class file at java.lang.ClassLoader.defineClass1(Native Method) at java.lang.ClassLoader.defineClass(ClassLoader.java:621) at java.lang.ClassLoader.defineClass(ClassLoader.java:466) at hudson.remoting.RemoteClassLoader.loadClassFile(RemoteClassLoader.java:152) ... 18 more Publishing Javadoc FATAL: Unable to copy Javadoc from https://builds.apache.org/job/Nutch-nutchgora/ws/nutchgora/build/docs/api to /home/hudson/hudson/jobs/Nutch-nutchgora/builds/2012-02-16_04-01-08/javadoc hudson.util.IOException2: java.lang.ClassFormatError: Failed to load javax.servlet.ServletRequest at hudson.FilePath.copyRecursiveTo(FilePath.java:1701) at hudson.FilePath.copyRecursiveTo(FilePath.java:1593) at hudson.tasks.JavadocArchiver.perform(JavadocArchiver.java:101) at hudson.tasks.BuildStepMonitor$1.perform(BuildStepMonitor.java:19) at hudson.model.AbstractBuild$AbstractRunner.perform(AbstractBuild.java:700) at hudson.model.AbstractBuild$AbstractRunner.performAllBuildSteps(AbstractBuild.java:675) at hudson.model.AbstractBuild$AbstractRunner.performAllBuildSteps(AbstractBuild.java:653) at hudson.model.Build$RunnerImpl.post2(Build.java:162) at hudson.model.AbstractBuild$AbstractRunner.post(AbstractBuild.java:622) at hudson.model.Run.run(Run.java:1434) at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:46) at hudson.model.ResourceController.execute(ResourceController.java:88) at hudson.model.Executor.run(Executor.java:238) Caused by: java.util.concurrent.ExecutionException: java.lang.ClassFormatError: Failed to load javax.servlet.ServletRequest at hudson.remoting.Channel$2.adapt(Channel.java:714) at hudson.remoting.Channel$2.adapt(Channel.java:709) at hudson.remoting.FutureAdapter.get(FutureAdapter.java:55) at hudson.FilePath.copyRecursiveTo(FilePath.java:1699) ... 12 more Caused by: java.lang.ClassFormatError: Failed to load javax.servlet.ServletRequest
Build failed in Jenkins: Nutch-trunk #1758
See https://builds.apache.org/job/Nutch-trunk/1758/ -- Started by timer Building remotely on solaris1 in workspace https://builds.apache.org/job/Nutch-trunk/ws/ hudson.util.IOException2: remote file operation failed: https://builds.apache.org/job/Nutch-trunk/ws/ at hudson.remoting.Channel@50e71521:solaris1 at hudson.FilePath.act(FilePath.java:784) at hudson.FilePath.act(FilePath.java:770) at hudson.scm.SubversionSCM.checkout(SubversionSCM.java:742) at hudson.scm.SubversionSCM.checkout(SubversionSCM.java:684) at hudson.model.AbstractProject.checkout(AbstractProject.java:1195) at hudson.model.AbstractBuild$AbstractRunner.checkout(AbstractBuild.java:576) at hudson.model.AbstractBuild$AbstractRunner.run(AbstractBuild.java:465) at hudson.model.Run.run(Run.java:1409) at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:46) at hudson.model.ResourceController.execute(ResourceController.java:88) at hudson.model.Executor.run(Executor.java:238) Caused by: java.io.IOException: Remote call on solaris1 failed at hudson.remoting.Channel.call(Channel.java:690) at hudson.FilePath.act(FilePath.java:777) ... 10 more Caused by: java.lang.ClassFormatError: Failed to load javax.servlet.ServletException at hudson.remoting.RemoteClassLoader.loadClassFile(RemoteClassLoader.java:154) at hudson.remoting.RemoteClassLoader.findClass(RemoteClassLoader.java:131) at java.lang.ClassLoader.loadClass(ClassLoader.java:307) at java.lang.ClassLoader.loadClass(ClassLoader.java:252) at java.lang.ClassLoader.loadClassInternal(ClassLoader.java:320) at hudson.scm.SubversionWorkspaceSelector.syncWorkspaceFormatFromMaster(SubversionWorkspaceSelector.java:85) at hudson.scm.SubversionSCM.createSvnClientManager(SubversionSCM.java:822) at hudson.scm.SubversionSCM$CheckOutTask.invoke(SubversionSCM.java:765) at hudson.scm.SubversionSCM$CheckOutTask.invoke(SubversionSCM.java:752) at hudson.FilePath$FileCallableWrapper.call(FilePath.java:2099) at hudson.remoting.UserRequest.perform(UserRequest.java:118) at hudson.remoting.UserRequest.perform(UserRequest.java:48) at hudson.remoting.Request$2.run(Request.java:287) at hudson.remoting.InterceptingExecutorService$1.call(InterceptingExecutorService.java:72) at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:269) at java.util.concurrent.FutureTask.run(FutureTask.java:123) at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:651) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:676) at java.lang.Thread.run(Thread.java:595) Caused by: java.lang.UnsupportedClassVersionError: Bad version number in .class file at java.lang.ClassLoader.defineClass1(Native Method) at java.lang.ClassLoader.defineClass(ClassLoader.java:621) at java.lang.ClassLoader.defineClass(ClassLoader.java:466) at hudson.remoting.RemoteClassLoader.loadClassFile(RemoteClassLoader.java:152) ... 18 more Recording test results ERROR: Failed to archive test reports hudson.util.IOException2: remote file operation failed: https://builds.apache.org/job/Nutch-trunk/ws/ at hudson.remoting.Channel@50e71521:solaris1 at hudson.FilePath.act(FilePath.java:784) at hudson.FilePath.act(FilePath.java:770) at hudson.tasks.junit.JUnitParser.parse(JUnitParser.java:83) at hudson.tasks.junit.JUnitResultArchiver.parse(JUnitResultArchiver.java:122) at hudson.tasks.junit.JUnitResultArchiver.perform(JUnitResultArchiver.java:134) at hudson.tasks.BuildStepMonitor$1.perform(BuildStepMonitor.java:19) at hudson.model.AbstractBuild$AbstractRunner.perform(AbstractBuild.java:700) at hudson.model.AbstractBuild$AbstractRunner.performAllBuildSteps(AbstractBuild.java:675) at hudson.model.AbstractBuild$AbstractRunner.performAllBuildSteps(AbstractBuild.java:653) at hudson.model.Build$RunnerImpl.post2(Build.java:162) at hudson.model.AbstractBuild$AbstractRunner.post(AbstractBuild.java:622) at hudson.model.Run.run(Run.java:1434) at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:46) at hudson.model.ResourceController.execute(ResourceController.java:88) at hudson.model.Executor.run(Executor.java:238) Caused by: java.io.IOException: Remote call on solaris1 failed at hudson.remoting.Channel.call(Channel.java:690) at hudson.FilePath.act(FilePath.java:777) ... 14 more Caused by: java.lang.ClassFormatError: Failed to load javax.servlet.ServletException at hudson.remoting.RemoteClassLoader.loadClassFile(RemoteClassLoader.java:154) at hudson.remoting.RemoteClassLoader.findClass(RemoteClassLoader.java:131)