Because we have a separate non Java legacy process that would take care of the parsing, and it requires raw html. It's more of a process reasoning than anything else. On Jun 30, 2013 8:06 AM, "Tejas Patil" <[email protected]> wrote:
> I am curious to know why do needed the raw html content instead of parsed > stuff. Search engines are meant to index parsed text. The data to be stored > and indexed reduces after parsing. > > > On Sat, Jun 29, 2013 at 9:20 PM, h b <[email protected]> wrote: > > > Thanks Tejas, > > I have just 2 urls in my seed file, and the second run of fetch ran for a > > few hours. I will verify if I got what I wanted. > > > > Regarding the raw html, its a ugly hack, so I did not really create a > > patch. But this is what I did > > > > > > In > > > src/plugin/parse-html/src/java/org/apache/nutch/parse/html/HtmlParser.java > > getParse method, > > > > //text = sb.toString(); > > text = new String(page.getContent().array()); > > > > Would be nice to make this as a configuration in the plugin xml. > > > > Other thing I will try soon is to extract the content only for a specific > > depth. > > > > > > > > On Sat, Jun 29, 2013 at 12:49 AM, Tejas Patil <[email protected] > > >wrote: > > > > > Yes. Nutch would parse the HTML and extract the content out of it. > > Tweaking > > > around the code surrounding the parser would have made that happen. If > > you > > > did something else, would you mind sharing it ? > > > > > > The "depth" is used by the Crawl class in 1.x which is deprecated in > 2.x. > > > Use bin/crawl instead. > > > While running the "bin/crawl" script, the "<numberOfRounds>" option is > > > nothing but the depth till which you want the crawling to be performed. > > > > > > If you want to use the individual commands instead, run generate -> > fetch > > > -> parse -> update multiple times. The crawl script internally does the > > > same thing. > > > eg. If you want to fetch till depth 3, this is how you could do: > > > inject -> (generate -> fetch -> parse -> update) > > > -> (generate -> fetch -> parse -> update) > > > -> (generate -> fetch -> parse -> update) > > > -> solrindex > > > > > > On Fri, Jun 28, 2013 at 7:24 PM, h b <[email protected]> wrote: > > > > > > > Ok, I tweaked the code a bit to extract the html as is from the > parser, > > > to > > > > realize that it is too much of a text and too much depth of crawling. > > So > > > I > > > > am looking to see if I can somehow limit the depth. Nutch 1.x docs > > > mention > > > > about the -depth parameter. However, I do not see this in the > > > > nutch-default.xml under Nutch 2.x. The -topN is used for number of > > links > > > > per depth. So for Nutch 2.x where/how do I set the depth? > > > > > > > > > > > > On Fri, Jun 28, 2013 at 11:32 AM, h b <[email protected]> wrote: > > > > > > > > > Ok, SO i also got this work with Solr 4 no errors, I think the key > > was > > > > not > > > > > using a crawl id. > > > > > I had to comment the updatelog in solrconfig.xml because I got some > > > > > "_version_" related error. > > > > > > > > > > My next questions is, my solr document, or for that matter even the > > > hbase > > > > > value of the html content is 'not html'. It appears that nutch is > > > > > extracting out text only. How do I retain the html content "as is". > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > On Fri, Jun 28, 2013 at 10:54 AM, Tejas Patil < > > > [email protected] > > > > >wrote: > > > > > > > > > >> Kewl !! > > > > >> > > > > >> I wonder why "org.apache.solr.common.SolrException: undefined > field > > > > text" > > > > >> happens.. Anybody who can throw light on this ? > > > > >> > > > > >> > > > > >> On Fri, Jun 28, 2013 at 10:45 AM, h b <[email protected]> wrote: > > > > >> > > > > >> > Thanks Tejas > > > > >> > I tried these steps, One step I added, was updatedb > > > > >> > > > > > >> > *bin/nutch updatedb* > > > > >> > > > > > >> > Just to be consistent with the doc, and your suggestion on some > > > other > > > > >> > thread, I used solr 3.6 instead of 4.x > > > > >> > I copied the schema.xml from nutch/conf (rootlevel) and started > > > solr. > > > > It > > > > >> > failed with > > > > >> > > > > > >> > SEVERE: org.apache.solr.common.SolrException: undefined field > text > > > > >> > > > > > >> > > > > > >> > One of the google thread, suggested I ignore this error, so I > > > ignored > > > > >> and > > > > >> > indexed anyway > > > > >> > > > > > >> > So now I got it to work. Playing some more with the queries > > > > >> > > > > > >> > > > > > >> > > > > > >> > > > > > >> > On Fri, Jun 28, 2013 at 9:52 AM, Tejas Patil < > > > > [email protected] > > > > >> > >wrote: > > > > >> > > > > > >> > > The "storage.schema.webpage" seems messed up but I don't have > > > ample > > > > >> time > > > > >> > > now to look into it. Here is what I would suggest to get > things > > > > >> working: > > > > >> > > * > > > > >> > > * > > > > >> > > *[1] Remove all the old data from HBase* > > > > >> > > > > > > >> > > (I assume that HBase is running while you do this) > > > > >> > > *cd $HBASE_HOME* > > > > >> > > *./bin/hbase shell > > > > >> > > * > > > > >> > > In the HBase shell, use "list" to see all the tables, delete > all > > > of > > > > >> those > > > > >> > > related to Nutch (ones named as *webpage). > > > > >> > > Remove them using "disable" and "drop" commands. > > > > >> > > > > > > >> > > eg. if one of the tables is "webpage", you would run this: > > > > >> > > *disable 'webpage' > > > > >> > > * > > > > >> > > *drop 'webpage'* > > > > >> > > * * > > > > >> > > > > > > >> > > *[2] Run crawl* > > > > >> > > I assume that you have not changed "storage.schema.webpage" is > > > > >> > > nutch-site.xml and nutch-default.xml. If yes, revert it to: > > > > >> > > > > > > >> > > *<property>* > > > > >> > > * <name>storage.schema.webpage</**name>* > > > > >> > > * <value>webpage</value>* > > > > >> > > * <description>This value holds the schema name used for > Nutch > > > web > > > > >> db.* > > > > >> > > * Note that Nutch ignores the value in the gora mapping > files, > > > and > > > > >> uses* > > > > >> > > * this as the webpage schema name.* > > > > >> > > * </description>* > > > > >> > > *</property>* > > > > >> > > > > > > >> > > Run crawl commands: > > > > >> > > *bin/nutch inject urls/* > > > > >> > > *bin/nutch generate -topN 50000 -noFilter -adddays 0* > > > > >> > > *bin/nutch fetch -all -threads 5 * > > > > >> > > *bin/nutch parse -all * > > > > >> > > > > > > >> > > *[3] Perform indexing* > > > > >> > > I assume that you have Solr setup and > NUTCH_HOME/conf/schema.xml > > > > >> copied > > > > >> > in > > > > >> > > ${SOLR_HOME}/example/solr/conf/. See bullets 4-6 in [0] for > > > details. > > > > >> > > Start solr and run the indexing command: > > > > >> > > *bin/nutch solrindex $SOLR_URL -all * > > > > >> > > > > > > >> > > [0] : http://wiki.apache.org/nutch/NutchTutorial > > > > >> > > > > > > >> > > Thanks, > > > > >> > > Tejas > > > > >> > > > > > > >> > > On Thu, Jun 27, 2013 at 1:47 PM, h b <[email protected]> > wrote: > > > > >> > > > > > > >> > > > Ok, so avro did not work quite well for me, I got a test > grid > > > with > > > > >> > hbase, > > > > >> > > > and I started using that for now. All steps ran without > errors > > > > and I > > > > >> > see > > > > >> > > my > > > > >> > > > crawled doc in hbase. > > > > >> > > > However, after running the solr integration, and querying > > solr, > > > I > > > > >> get > > > > >> > > back > > > > >> > > > nothing. Index files look very tiny. The one thing I noted > is > > a > > > > >> message > > > > >> > > > during almost every step > > > > >> > > > > > > > >> > > > 13/06/27 20:37:53 INFO store.HBaseStore: Keyclass and > > nameclass > > > > >> match > > > > >> > but > > > > >> > > > mismatching table names mappingfile schema is 'webpage' vs > > > actual > > > > >> > schema > > > > >> > > > 'crawl2_webpage' , assuming they are the same. > > > > >> > > > > > > > >> > > > This looks suspicious and I think this is the one causing > the > > > solr > > > > >> > index > > > > >> > > to > > > > >> > > > be empty. Googling suggested I should edit the > > > nutch-default,xml, > > > > I > > > > >> > tried > > > > >> > > > and rebuilt the job but no luck with this message. > > > > >> > > > > > > > >> > > > > > > > >> > > > > > > > >> > > > On Thu, Jun 27, 2013 at 10:30 AM, h b <[email protected]> > > wrote: > > > > >> > > > > > > > >> > > > > Ok, I ran a ant, ant jar and ant job and that seems to > have > > > > >> picked up > > > > >> > > the > > > > >> > > > > config changes. > > > > >> > > > > Now, the inject output shows that it is using AvroStore as > > > Gora > > > > >> > > storage. > > > > >> > > > > > > > > >> > > > > Now I am getting Nullpointer on > > > > >> > > > > > > > > >> > > > > java.lang.NullPointerException > > > > >> > > > > at > > > > >> > > > > > > > > >> > > > > > > > >> > > > > > > >> > > > > > >> > > > > > > > > > > org.apache.gora.mapreduce.GoraOutputFormat.setOutputPath(GoraOutputFormat.java:70) > > > > >> > > > > at > > > > >> > > > > > > > > >> > > > > > > > >> > > > > > > >> > > > > > >> > > > > > > > > > > org.apache.gora.mapreduce.GoraOutputFormat.getRecordWriter(GoraOutputFormat.java:91) > > > > >> > > > > at > > > > >> > > > > > > > > >> > > > > > > > >> > > > > > > >> > > > > > >> > > > > > > > > > > org.apache.hadoop.mapred.MapTask$NewDirectOutputCollector.<init>(MapTask.java:521) > > > > >> > > > > at > > > > >> > > > > > org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:636) > > > > >> > > > > at > > > > org.apache.hadoop.mapred.MapTask.run(MapTask.java:323) > > > > >> > > > > at > > > org.apache.hadoop.mapred.Child$4.run(Child.java:270) > > > > >> > > > > at > > java.security.AccessController.doPrivileged(Native > > > > >> Method) > > > > >> > > > > at > > javax.security.auth.Subject.doAs(Subject.java:396) > > > > >> > > > > at > > > > >> > > > > > > > > >> > > > > > > > >> > > > > > > >> > > > > > >> > > > > > > > > > > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1177) > > > > >> > > > > at > > org.apache.hadoop.mapred.Child.main(Child.java:264) > > > > >> > > > > > > > > >> > > > > which does not look like nutch related. I will work on > this > > > and > > > > >> write > > > > >> > > > back > > > > >> > > > > if I get stuck on something else, or will write back if I > > > > succeed. > > > > >> > > > > > > > > >> > > > > > > > > >> > > > > On Thu, Jun 27, 2013 at 10:18 AM, h b <[email protected]> > > > wrote: > > > > >> > > > > > > > > >> > > > >> Hi Lewis, > > > > >> > > > >> > > > > >> > > > >> Sorry for missing that one. So I update the top level > conf > > > and > > > > >> > rebuild > > > > >> > > > >> the job. > > > > >> > > > >> > > > > >> > > > >> cat ~/nutch/apache-nutch-2.2/conf/nutch-site.xml > > > > >> > > > >> > > > > >> > > > >> ...... > > > > >> > > > >> <property> > > > > >> > > > >> <name>storage.data.store.class</name> > > > > >> > > > >> <value>org.apache.gora.avro.store.AvroStore</value> > > > > >> > > > >> </property> > > > > >> > > > >> ...... > > > > >> > > > >> > > > > >> > > > >> cd ~/nutch/apache-nutch-2.2/ > > > > >> > > > >> ant job > > > > >> > > > >> cd ~/nutch/apache-nutch-2.2/runtime/deploy/ > > > > >> > > > >> > > > > >> > > > >> > > > > >> > > > >> bin/nutch inject urls -crawlId crawl1 > > > > >> > > > >> 13/06/27 17:12:01 INFO crawl.InjectorJob: InjectorJob: > > > starting > > > > >> at > > > > >> > > > >> 2013-06-27 17:12:01 > > > > >> > > > >> 13/06/27 17:12:01 INFO crawl.InjectorJob: InjectorJob: > > > > Injecting > > > > >> > > urlDir: > > > > >> > > > >> urls > > > > >> > > > >> 13/06/27 17:12:01 INFO crawl.InjectorJob: InjectorJob: > > Using > > > > >> class > > > > >> > > > >> org.apache.gora.memory.store.MemStore as the Gora storage > > > > class. > > > > >> > > > >> > > > > >> > > > >> It still shows me MemStore. > > > > >> > > > >> > > > > >> > > > >> In the jobtracker I see a [crawl1]inject urls job does > not > > > have > > > > >> > > > >> urls_injected property > > > > >> > > > >> I have a *db.score.injected* 1.0, but dont think that is > > > > >> anything to > > > > >> > > say > > > > >> > > > >> about urls injected. > > > > >> > > > >> > > > > >> > > > >> > > > > >> > > > >> > > > > >> > > > >> On Thu, Jun 27, 2013 at 7:09 AM, Lewis John Mcgibbney < > > > > >> > > > >> [email protected]> wrote: > > > > >> > > > >> > > > > >> > > > >>> Hi, > > > > >> > > > >>> Please re-read my mail. > > > > >> > > > >>> If you are using the deploy directory e.g. running on a > > > hadoop > > > > >> > > cluster, > > > > >> > > > >>> then make sure to edit nutch-site.xml from within the > top > > > > level > > > > >> > conf > > > > >> > > > >>> directory _not_ the conf directory in runtime/local. > > > > >> > > > >>> If you look at the ant runtime target in the build > script > > > you > > > > >> will > > > > >> > > see > > > > >> > > > >>> the > > > > >> > > > >>> code which generates the runtime directory structure. > > > > >> > > > >>> Make changes to conf/nutch-site.xml, build the job jar, > > > > >> navigate to > > > > >> > > > >>> runtime/deploy, run the code. > > > > >> > > > >>> It's easier to make the job jar and scripts in deploy > > > > available > > > > >> to > > > > >> > > the > > > > >> > > > >>> job > > > > >> > > > >>> tracker. > > > > >> > > > >>> You also didn't comment on the counters for the inject > > job. > > > Do > > > > >> you > > > > >> > > see > > > > >> > > > >>> any? > > > > >> > > > >>> Best > > > > >> > > > >>> Lewis > > > > >> > > > >>> > > > > >> > > > >>> On Wednesday, June 26, 2013, h b <[email protected]> > > wrote: > > > > >> > > > >>> > Here is an example of what I am saying about the > config > > > > >> changes > > > > >> > not > > > > >> > > > >>> taking > > > > >> > > > >>> > effect. > > > > >> > > > >>> > > > > > >> > > > >>> > cd runtime/deploy > > > > >> > > > >>> > cat ../local/conf/nutch-site.xml > > > > >> > > > >>> > ...... > > > > >> > > > >>> > > > > > >> > > > >>> > <property> > > > > >> > > > >>> > <name>storage.data.store.class</name> > > > > >> > > > >>> > > <value>org.apache.gora.avro.store.AvroStore</value> > > > > >> > > > >>> > </property> > > > > >> > > > >>> > ..... > > > > >> > > > >>> > > > > > >> > > > >>> > cd ../.. > > > > >> > > > >>> > > > > > >> > > > >>> > ant job > > > > >> > > > >>> > > > > > >> > > > >>> > cd runtime/deploy > > > > >> > > > >>> > bin/nutch inject urls -crawlId crawl1 > > > > >> > > > >>> > ..... > > > > >> > > > >>> > 13/06/27 06:34:29 INFO crawl.InjectorJob: InjectorJob: > > > Using > > > > >> > class > > > > >> > > > >>> > org.apache.gora.memory.store.MemStore as the Gora > > storage > > > > >> class. > > > > >> > > > >>> > ..... > > > > >> > > > >>> > > > > > >> > > > >>> > So the nutch-site.xml was changed to use AvroStore as > > > > storage > > > > >> > class > > > > >> > > > and > > > > >> > > > >>> job > > > > >> > > > >>> > was rebuilt, and I reran inject, the output of which > > still > > > > >> shows > > > > >> > > that > > > > >> > > > >>> it > > > > >> > > > >>> is > > > > >> > > > >>> > trying to use Memstore. > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > On Wed, Jun 26, 2013 at 11:05 PM, Lewis John > Mcgibbney < > > > > >> > > > >>> > [email protected]> wrote: > > > > >> > > > >>> > > > > > >> > > > >>> >> The Gora MemStore was introduced to deal > predominantly > > > with > > > > >> test > > > > >> > > > >>> scenarios. > > > > >> > > > >>> >> This is justified as the 2.x code is pulled nightly > and > > > > after > > > > >> > > every > > > > >> > > > >>> commit > > > > >> > > > >>> >> and tested. > > > > >> > > > >>> >> It is nnot thread safe and should not be used (until > we > > > fix > > > > >> some > > > > >> > > > >>> issues) > > > > >> > > > >>> >> for any kind of serious deployment. > > > > >> > > > >>> >> From your inject task on the job tracker, you will be > > > able > > > > to > > > > >> > see > > > > >> > > > >>> >> 'urls_injected' counters which represent the number > of > > > urls > > > > >> > > actually > > > > >> > > > >>> >> persisted through Gora into the datastore. > > > > >> > > > >>> >> I understand that HBase is not an option. Gora should > > > also > > > > >> > support > > > > >> > > > >>> writing > > > > >> > > > >>> >> the output into Avro sequence files... which can be > > > pumped > > > > >> into > > > > >> > > > hdfs. > > > > >> > > > >>> We > > > > >> > > > >>> >> have done some work on this so I suppose that right > now > > > is > > > > as > > > > >> > > good a > > > > >> > > > >>> time > > > > >> > > > >>> >> as any for you to try it out. > > > > >> > > > >>> >> use the default datastore as > > > > >> > org.apache.gora.avro.store.AvroStore > > > > >> > > I > > > > >> > > > >>> think. > > > > >> > > > >>> >> You can double check by looking into gora.properties > > > > >> > > > >>> >> As a note, youu should use nutch-site.xml within the > > top > > > > >> level > > > > >> > > conf > > > > >> > > > >>> >> directory for all your Nutch configuration. You > should > > > then > > > > >> > > create a > > > > >> > > > >>> new > > > > >> > > > >>> >> job jar for use in hadoop by calling 'ant job' after > > the > > > > >> changes > > > > >> > > are > > > > >> > > > >>> made. > > > > >> > > > >>> >> hth > > > > >> > > > >>> >> Lewis > > > > >> > > > >>> >> > > > > >> > > > >>> >> On Wednesday, June 26, 2013, h b <[email protected]> > > > wrote: > > > > >> > > > >>> >> > The quick responses flowing are very encouraging. > > > Thanks > > > > >> > Tejas. > > > > >> > > > >>> >> > Tejas, as I mentioned earlier, in fact I actually > ran > > > it > > > > >> step > > > > >> > by > > > > >> > > > >>> step. > > > > >> > > > >>> >> > > > > > >> > > > >>> >> > So first I ran the inject command and then the > readdb > > > > with > > > > >> > dump > > > > >> > > > >>> option > > > > >> > > > >>> >> and > > > > >> > > > >>> >> > did not see anything in the dump files, that leads > me > > > to > > > > >> say > > > > >> > > that > > > > >> > > > >>> the > > > > >> > > > >>> >> > inject did not work.I verified the regex-urlfilter > > and > > > > made > > > > >> > sure > > > > >> > > > >>> that > > > > >> > > > >>> my > > > > >> > > > >>> >> > url is not getting filtered. > > > > >> > > > >>> >> > > > > > >> > > > >>> >> > I agree that the second link is about configuring > > HBase > > > > as > > > > >> a > > > > >> > > > >>> storageDB. > > > > >> > > > >>> >> > However, I do not have Hbase installed and dont > > foresee > > > > >> > getting > > > > >> > > it > > > > >> > > > >>> >> > installed any sooner, hence using HBase for storage > > is > > > > not > > > > >> a > > > > >> > > > option, > > > > >> > > > >>> so I > > > > >> > > > >>> >> > am going to have to stick to Gora with memory > store. > > > > >> > > > >>> >> > > > > > >> > > > >>> >> > > > > > >> > > > >>> >> > > > > > >> > > > >>> >> > > > > > >> > > > >>> >> > On Wed, Jun 26, 2013 at 10:02 PM, Tejas Patil < > > > > >> > > > >>> [email protected] > > > > >> > > > >>> >> >wrote: > > > > >> > > > >>> >> > > > > > >> > > > >>> >> >> On Wed, Jun 26, 2013 at 9:53 PM, h b < > > > [email protected]> > > > > >> > wrote: > > > > >> > > > >>> >> >> > > > > >> > > > >>> >> >> > Thanks for the response Lewis. > > > > >> > > > >>> >> >> > I did read these links, I mostly followed the > > first > > > > link > > > > >> > and > > > > >> > > > >>> tried > > > > >> > > > >>> >> both > > > > >> > > > >>> >> >> the > > > > >> > > > >>> >> >> > 3.2 and 3.3 sections. Using the bin/crawl gave > me > > > null > > > > >> > > pointer > > > > >> > > > >>> >> exception > > > > >> > > > >>> >> >> on > > > > >> > > > >>> >> >> > solr, so I figured that I should first deal with > > > > getting > > > > >> > the > > > > >> > > > >>> crawl > > > > >> > > > >>> >> part > > > > >> > > > >>> >> >> to > > > > >> > > > >>> >> >> > work and then deal with solr indexing. Hence I > > went > > > > >> back to > > > > >> > > > >>> trying > > > > >> > > > >>> it > > > > >> > > > >>> >> >> > stepwise. > > > > >> > > > >>> >> >> > > > > > >> > > > >>> >> >> > > > > >> > > > >>> >> >> You should try running the crawl using individual > > > > commands > > > > >> > and > > > > >> > > > see > > > > >> > > > >>> where > > > > >> > > > >>> >> >> the problem is. The nutch tutorial which Lewis > > pointed > > > > >> you to > > > > >> > > had > > > > >> > > > >>> those > > > > >> > > > >>> >> >> commands. Even peeking into the bin/crawl script > > would > > > > >> also > > > > >> > > help > > > > >> > > > >>> as it > > > > >> > > > >>> >> >> calls the nutch commands. > > > > >> > > > >>> >> >> > > > > >> > > > >>> >> >> > > > > > >> > > > >>> >> >> > As for the second link, it is more about using > > HBase > > > > as > > > > >> > store > > > > >> > > > >>> instead > > > > >> > > > >>> >> of > > > > >> > > > >>> >> >> > gora. This is not really a option for me yet, > > cause > > > my > > > > >> grid > > > > >> > > > does > > > > >> > > > >>> not > > > > >> > > > >>> >> have > > > > >> > > > >>> >> >> > hbase installed yet. Getting it done is not much > > > under > > > > >> my > > > > >> > > > control > > > > >> > > > >>> >> >> > > > > > >> > > > >>> >> >> > > > > >> > > > >>> >> >> HBase is one of the datastores supported by Apache > > > Gora. > > > > >> That > > > > >> > > > >>> tutorial > > > > >> > > > >>> >> >> speaks about how to configure Nutch (actually > Gora) > > to > > > > use > > > > >> > > HBase > > > > >> > > > >>> as a > > > > >> > > > >>> >> >> backend. So, its wrong to say that the tutorial > was > > > > about > > > > >> > HBase > > > > >> > > > and > > > > >> > > > >>> not > > > > >> > > > >>> >> >> Gora. > > > > >> > > > >>> >> >> > > > > >> > > > >>> >> >> > > > > > >> > > > >>> >> >> > the FAQ link is the one I had not gone through > > > until I > > > > >> > > checked > > > > >> > > > >>> your > > > > >> > > > >>> >> >> > response, but I do not find answers to any of my > > > > >> questions > > > > >> > > > >>> >> >> > (directly/indirectly) in it. > > > > >> > > > >>> >> >> > > > > > >> > > > >>> >> >> > > > > >> > > > >>> >> >> Ok > > > > >> > > > >>> >> >> > > > > >> > > > >>> >> >> > > > > > >> > > > >>> >> >> > > > > > >> > > > >>> >> >> > > > > > >> > > > >>> >> >> > > > > > >> > > > >>> >> >> > On Wed, Jun 26, 2013 at 7:44 PM, Lewis John > > > Mcgibbney > > > > < > > > > >> > > > >>> >> >> > > *Lewis* > > > > >> > > > >>> >> > > > > >> > > > >>> > > > > > >> > > > >>> > > > > >> > > > >>> -- > > > > >> > > > >>> *Lewis* > > > > >> > > > >>> > > > > >> > > > >> > > > > >> > > > >> > > > > >> > > > > > > > > >> > > > > > > > >> > > > > > > >> > > > > > >> > > > > > > > > > > > > > > > > > > > >

