Thanks Jeffrey,

since I had running the pre-compiled version of capture-HPC server
(capture-server-2.5.1-389-withLinuxRevert.zip)
I needed download from svn the capture directory with all java classes and the
compile_revert_linux.sh script and after solved some minor issues all the 
compilation was succesfull.

I have configured the plugin and executed the first time and I could see the 
crawler output for a url working
right but the capture-HPC server  looks like stay in a constant reverting state 
and not working with the
crawled urls:

--cut--
..
Depth=2  Crawling http://www.domain.com/resource.asp
Finished Crawling http://www.domain.com
Waiting for input URLs...

[sep 4, 2009 11:27:18 AM-172.21.1.44:902-11546362] Finished processing VM item: 
revert
[sep 4, 2009 11:27:39 AM-172.21.1.44:902-11546362] Client inactivity, reverting 
VM
[sep 4, 2009 11:27:39 AM-172.21.1.44:902-11546362] VMSetState: 
WAITING_TO_BE_REVERTED
[sep 4, 2009 11:27:40 AM-172.21.1.44:902-11546362] VMSetState: REVERTING
Waiting for input URLs...
[sep 4, 2009 11:27:59 AM-172.21.1.44:902-11546362] VMSetState: RUNNING
Reverting same VM...just waiting a bit
[sep 4, 2009 11:28:05 AM-172.21.1.44:902-11546362] Finished processing VM item: 
revert
[sep 4, 2009 11:28:45 AM-172.21.1.44:902-11546362] Client inactivity, reverting 
VM
[sep 4, 2009 11:28:45 AM-172.21.1.44:902-11546362] VMSetState: 
WAITING_TO_BE_REVERTED
[sep 4, 2009 11:28:45 AM-172.21.1.44:902-11546362] VMSetState: REVERTING
Waiting for input URLs...
[sep 4, 2009 11:29:07 AM-172.21.1.44:902-11546362] VMSetState: RUNNING
Reverting same VM...just waiting a bit
[sep 4, 2009 11:29:13 AM-172.21.1.44:902-11546362] Finished processing VM item: 
revert
[sep 4, 2009 11:29:53 AM-172.21.1.44:902-11546362] Client inactivity, reverting 
VM
[sep 4, 2009 11:29:53 AM-172.21.1.44:902-11546362] VMSetState: 
WAITING_TO_BE_REVERTED
[sep 4, 2009 11:29:53 AM-172.21.1.44:902-11546362] VMSetState: REVERTING
Waiting for input URLs...
--end--


Thanks
Emilio






>
>De: JEFFREY S STEWART <jss1...@esu.edu>
>Para: Emilio Casbas <ecasb...@yahoo.es>; capture-hpc@public.honeynet.org
>Enviado: jueves, 3 de septiembre, 2009 17:33:18
>Asunto: RE: [Capture-HPC] Capture-HPC Crawler Preprocessor
>
>RE: [Capture-HPC] Capture-HPC Crawler Preprocessor >
>
>
>Emilio,
>
>>Please reply via the mailing list so that others can find the solution if 
>>they have the same problem.
>
>>The errors lead me to believe that the Preprocessor.java file is not being 
>>found when you build it.  Please check to make sure that the 
>>Preprocessor.java is in the capture directory along with the source files 
>>from the Crawler.tar I sent.
>
>>Thanks,
>>Jeff
>> 
>
>
>>-----Original Message-----
>>From: Emilio Casbas [mailto:ecasb...@yahoo.es]
>>Sent: Thu 9/3/2009 4:11 AM
>>To: JEFFREY S STEWART
>>Subject: Re: [Capture-HPC] Capture-HPC Crawler Preprocessor
>
>>Hi Jeffrey,
>
>>congratulations for your support and excellent job with the capture-hpc 
>>project.
>
>>I am interested in testing this feature but since I'm not a developer I'm 
>>having
>>some problems installing it.
>
>>Following the instructions, in the step 4, I run the "ant" command and after 
>>solved some
>>issues I get this:
>
>>compile:
>>    [javac] Compiling 3 source files to 
>> /home/machine/capture-HPC/capture-with-crawl/build
>>    [javac] 
>> /home/machine/capture-HPC/capture-with-crawl/source/Crawler.java:14: cannot 
>> find symbol
>>    [javac] symbol  : class Preprocessor
>>    [javac] location: package capture
>>    [javac] public class Crawler extends capture.Preprocessor
>>    [javac]                                     ^
>>    [javac] 
>> /home/machine/capture-HPC/capture-with-crawl/source/Crawler.java:472: cannot 
>> find symbol
>>    [javac] symbol  : method addUrlToCaptureQueue(java.lang.String)
>>    [javac] location: class capture.Crawler
>>    [javac]             addUrlToCaptureQueue(url + "::" + program + "::" + 
>> delay + priority);
>>    [javac]             ^
>>    [javac] 2 errors
>
>>BUILD FAILED
>>/home/machine/capture-HPC/capture-with-crawl/build.xml:34: Compile failed; 
>>see the compiler error output for details.
>
>>Total time: 5 seconds
>>machine@pam-inv-03:~/capture-HPC/capture-with-crawl$
>
>>Previously I had the capture-HPC program running successfully but I
>>didn't compile the software I had installed a pre-configured version.
>>Could you point me to some solution?
>
>>I could help you in testing and troubleshooting the plugin.
>
>>TIA
>>Emilio
>
>
>
>>>
>>>De: JEFFREY S STEWART <jss1...@esu.edu>
>>>Para: General discussion list for Capture-HPC users 
>>><capture-hpc@public.honeynet.org>
>>>Enviado: lunes, 17 de agosto, 2009 15:11:41
>>>Asunto: [Capture-HPC] Capture-HPC Crawler Preprocessor
>>>
>>>Capture-HPC Crawler Preprocessor >
>>>
>>>
>>>All,
>>>
>>>>Attached is a preprocessor that I've made to add web crawler support to 
>>>>capture-HPC.  It only does http right now.  It works by finding links in 
>>>>the href field of the input page's HTML that you specify.  It has a bunch 
>>>>of options to configure that let you determine where it crawls, view the 
>>>>Crawler.README for a list of them.
>>>
>>>>One of the features, not really a crawler function, but decided it fit in 
>>>>nicely with scrapping pages.  (It queries google for 
>>>>site:safebrowsing.clients.google.com "the last time suspicious content was 
>>>>found on this site was on ) plus yesterday's date.  This has the result of 
>>>>getting all the malicious urls that google identified and crawled from 
>>>>yesterday.  (Good for when you don't have any malicious urls to crawl).  
>>>>Note, this feature doesn't fall google's TOS.
>>>
>>>>There are some more specific build instructions because the classes that I 
>>>>used have to be built with the project.  Take a look at the enclosed 
>>>>build.README.
>>>
>>>>If there are any questions or feedback, let me know.
>>>
>>>>Thanks
>>>>Jeff
>>>
>>>
>
>
>>     
>
> 


      
_______________________________________________
Capture-HPC mailing list
Capture-HPC@public.honeynet.org
https://public.honeynet.org/mailman/listinfo/capture-hpc

Reply via email to