Hi Jim
        The job I left running before the weekend had some (other) problems,
mainly about hadoop API change.
        Anyway I'm running it again right now and at first glance its working
(I'll know for sure in about 1 hour), on a different note there was a
problem with RegExpRowFilter where if it received more that one
conditional in the constructor map it would filter out records it
shouldn't, and that problem is solved.
        As Friday before I got your response I had already upgraded the cluster
to the hadoop and hbase trunk versions I'm currently testing with these
versions instead of 0.1, hope there is no problem there.
        I'll send another email soon.

Regards
David Alves

On Mon, 2008-04-07 at 08:31 -0700, Jim Kellerman wrote:
> David,
> 
> Any luck running this patch either against head or against the 0.1 branch?
> 
> Thanks.
> 
> ---
> Jim Kellerman, Senior Engineer; Powerset
> 
> 
> > -----Original Message-----
> > From: David Alves [mailto:[EMAIL PROTECTED]
> > Sent: Friday, April 04, 2008 10:05 AM
> > To: [email protected]
> > Subject: RE: StackOverFlow Error in HBase
> >
> > Hi Jim
> >
> >         Of course, my questions was regarding whether I
> > should use HEAD or some branch or tag.
> >         Anyways I currently running Hbase HEAD patched
> > against Hadoop HEAD, I'll know if its ok soon.
> >
> > Regards
> > David Alves
> > On Fri, 2008-04-04 at 09:18 -0700, Jim Kellerman wrote:
> > > After applying the patch, you have to rebuild and deploy on your
> > > cluster, run your test that was failing and verify that it
> > now works.
> > > See
> > >
> > http://hadoop.apache.org/hbase/docs/current/api/overview-summary.html#
> > > overview_description
> > >
> > >
> > >
> > > ---
> > > Jim Kellerman, Senior Engineer; Powerset
> > >
> > >
> > > > -----Original Message-----
> > > > From: David Alves [mailto:[EMAIL PROTECTED]
> > > > Sent: Friday, April 04, 2008 6:29 AM
> > > > To: [email protected]
> > > > Subject: RE: StackOverFlow Error in HBase
> > > >
> > > > Hi all again
> > > >
> > > >         I've never used the patch system you guys use, so
> > in order
> > > > to test the patch submitted by Clint what do I have to do? I mean
> > > > I've updated HEAD and applied the patch, is this it?
> > > >
> > > > Regards
> > > > David Alves
> > > >
> > > >
> > > >
> > > > On Thu, 2008-04-03 at 10:18 -0700, Jim Kellerman wrote:
> > > > > Thanks David. I'll add 554 as a blocker for 0.1.1
> > > > >
> > > > > ---
> > > > > Jim Kellerman, Senior Engineer; Powerset
> > > > >
> > > > >
> > > > > > -----Original Message-----
> > > > > > From: David Alves [mailto:[EMAIL PROTECTED]
> > > > > > Sent: Thursday, April 03, 2008 9:21 AM
> > > > > > To: [EMAIL PROTECTED]
> > > > > > Subject: RE: StackOverFlow Error in HBase
> > > > > >
> > > > > > Hi Jim and all
> > > > > >
> > > > > >         I'll commit to test the patch under the same
> > > > conditions as
> > > > > > it failed before, (with around 36000 records) but in this
> > > > > > precise moment I preparing my next development
> > iteration, which
> > > > means a lot
> > > > > > of meetings.
> > > > > >         By the end of the day tomorrow (friday) I
> > should have a
> > > > > > confirmation whether the patch worked (or not).
> > > > > >
> > > > > > Regards
> > > > > > David Alves
> > > > > >
> > > > > > On Thu, 2008-04-03 at 09:12 -0700, Jim Kellerman wrote:
> > > > > > > David,
> > > > > > >
> > > > > > > Have you had a chance to try this patch? We are about to
> > > > > > release hbase-0.1.1 and until we receive a confirmation in
> > > > > > HBASE-554 from another person who has tried it and
> > > > verifies that it
> > > > > > works, we cannot include it in this release. If it is not in
> > > > > > this release, there will be a significant wait for it
> > to appear
> > > > > > in an hbase release. hbase-0.1.2 will not happen anytime soon
> > > > unless there
> > > > > > are critical issues that arise that have not been
> > fixed in 0.1.1.
> > > > > > hbase-0.2.0 is also some time in the future. There are a
> > > > significant
> > > > > > number of issues to address before that release is ready.
> > > > > > >
> > > > > > > Frankly, I'd like to see this patch in 0.1.1, because it is
> > > > > > an issue for people that use filters.
> > > > > > >
> > > > > > > The alternative would be for Clint to supply a test case
> > > > > > that fails without the patch but passes with the patch.
> > > > > > >
> > > > > > > We will hold up the release, but need a commitment either
> > > > > > from David to test the patch or for Clint to supply a test.
> > > > > > We need that commitment by the end of the day today
> > > > > > 2008/04/03 along with an eta as to when it will be completed.
> > > > > > >
> > > > > > > ---
> > > > > > > Jim Kellerman, Senior Engineer; Powerset
> > > > > > >
> > > > > > >
> > > > > > > > -----Original Message-----
> > > > > > > > From: David Alves [mailto:[EMAIL PROTECTED]
> > > > > > > > Sent: Tuesday, April 01, 2008 2:36 PM
> > > > > > > > To: [EMAIL PROTECTED]
> > > > > > > > Subject: RE: StackOverFlow Error in HBase
> > > > > > > >
> > > > > > > > Hi
> > > > > > > >
> > > > > > > >         I just deployed the unpatched version.
> > > > > > > >         Tomorrow I'll rebuild the system with the patch
> > > > > > and try it
> > > > > > > > out.
> > > > > > > >         Thanks again.
> > > > > > > >
> > > > > > > > Regards
> > > > > > > > David Alves
> > > > > > > >
> > > > > > > > > -----Original Message-----
> > > > > > > > > From: Jim Kellerman [mailto:[EMAIL PROTECTED]
> > > > > > > > > Sent: Tuesday, April 01, 2008 10:04 PM
> > > > > > > > > To: [EMAIL PROTECTED]
> > > > > > > > > Subject: RE: StackOverFlow Error in HBase
> > > > > > > > >
> > > > > > > > > David,
> > > > > > > > >
> > > > > > > > > Have you tried this patch and does it work for
> > you? If so
> > > > > > > > > we'll include it
> > > > > > > > > hbase-0.1.1
> > > > > > > > >
> > > > > > > > > ---
> > > > > > > > > Jim Kellerman, Senior Engineer; Powerset
> > > > > > > > >
> > > > > > > > >
> > > > > > > > > > -----Original Message-----
> > > > > > > > > > From: David Alves
> > [mailto:[EMAIL PROTECTED]
> > > > > > > > > > Sent: Tuesday, April 01, 2008 10:44 AM
> > > > > > > > > > To: [EMAIL PROTECTED]
> > > > > > > > > > Subject: RE: StackOverFlow Error in HBase
> > > > > > > > > >
> > > > > > > > > > Hi
> > > > > > > > > >         Thanks for the prompt path Clint, St.Ack and
> > > > > > all you guys.
> > > > > > > > > >
> > > > > > > > > > Regards
> > > > > > > > > > David Alves
> > > > > > > > > >
> > > > > > > > > > > -----Original Message-----
> > > > > > > > > > > From: [EMAIL PROTECTED]
> > [mailto:[EMAIL PROTECTED]
> > > > > > > > On Behalf
> > > > > > > > > > > Of Clint Morgan
> > > > > > > > > > > Sent: Tuesday, April 01, 2008 2:04 AM
> > > > > > > > > > > To: [EMAIL PROTECTED]
> > > > > > > > > > > Subject: Re: StackOverFlow Error in HBase
> > > > > > > > > > >
> > > > > > > > > > > Try the patch at
> > > > > > > > https://issues.apache.org/jira/browse/HBASE-554.
> > > > > > > > > > >
> > > > > > > > > > > cheers,
> > > > > > > > > > > -clint
> > > > > > > > > > >
> > > > > > > > > > > On Mon, Mar 31, 2008 at 5:39 AM, David Alves
> > > > > > > > > > > <[EMAIL PROTECTED]> wrote:
> > > > > > > > > > > > Hi ... again
> > > > > > > > > > > >
> > > > > > > > > > > >         In my previous mail I stated that
> > > > increasing the
> > > > > > > > > > stack size
> > > > > > > > > > > solved the
> > > > > > > > > > > >  problem, well I jumped a little bit to the
> > > > conclusion,
> > > > > > > > > > in fact it
> > > > > > > > > > > > didn't, the StackOverFlowError always occurs
> > > > at the end
> > > > > > > > > > of the cycle
> > > > > > > > > > > > when no more records match the filter. Anyway I've
> > > > > > > > rewritten my
> > > > > > > > > > > > application to use a normal scanner and and do the
> > > > > > > > > > "filtering" after
> > > > > > > > > > > > which is not optimal but it works.
> > > > > > > > > > > >         I'm just saying this because it might
> > > > be a clue,
> > > > > > > > > > in previous
> > > > > > > > > > > versions
> > > > > > > > > > > >  (!= 0.1.0) even though a more serious
> > > > problem happened
> > > > > > > > > > > > (regionservers  became irresponsive after so many
> > > > > > > > records) this
> > > > > > > > > > > > didn't happen. Btw in  current version I
> > notice no,
> > > > > > > > > > > > or
> > > > > > > > > > very small,
> > > > > > > > > > > > decrease of thoughput with  time, great work!
> > > > > > > > > > > >
> > > > > > > > > > > >  Regards
> > > > > > > > > > > >  David Alves
> > > > > > > > > > > >
> > > > > > > > > > > >
> > > > > > > > > > > >
> > > > > > > > > > > >
> > > > > > > > > > > >
> > > > > > > > > > > >
> > > > > > > > > > > >
> > > > > > > > > > > >  On Mon, 2008-03-31 at 05:18 +0100, David
> > Alves wrote:
> > > > > > > > > > > >  > Hi again
> > > > > > > > > > > >  >
> > > > > > > > > > > >  >       As I was almost at the end (80%)
> > of indexable
> > > > > > > > > > docs, for the
> > > > > > > > > > > time
> > > > > > > > > > > >  > being I simply increased the stack size, which
> > > > > > > > seemed to work.
> > > > > > > > > > > >  >       Thanks for your input St.Ack
> > really helped me
> > > > > > > > > > solve the problem
> > > > > > > > > > > at
> > > > > > > > > > > >  > least for the moment.
> > > > > > > > > > > >  >       On another note in the same
> > method I changed
> > > > > > > > the way the
> > > > > > > > > > > scanner was
> > > > > > > > > > > >  > obtained when htable.getStartKeys()
> > would be more
> > > > > > > > > > > > than
> > > > > > > > > > 1, so that
> > > > > > > > > > > > I
> > > > > > > > > > > could
> > > > > > > > > > > >  > limit the records read each time to a single
> > > > > > > > region, and the
> > > > > > > > > > > > scanning
> > > > > > > > > > > would
> > > > > > > > > > > >  > start at the last region, strangely
> > the number of
> > > > > > > > > > > > keys
> > > > > > > > > > obtained
> > > > > > > > > > > > by  > htable.getStartKeys() was always 1
> > even though
> > > > > > > > by the end
> > > > > > > > > > > > there are
> > > > > > > > > > > already
> > > > > > > > > > > >  > 21 regions.
> > > > > > > > > > > >  >       Any thoughts?
> > > > > > > > > > > >  >
> > > > > > > > > > > >  > Regards
> > > > > > > > > > > >  > David Alves
> > > > > > > > > > > >  >
> > > > > > > > > > > >  > > -----Original Message-----  > > From: stack
> > > > > > > > > > > > [mailto:[EMAIL PROTECTED]  > > Sent:
> > > > > > > > > > Sunday, March
> > > > > > > > > > > > 30, 2008 9:36 PM  > > To:
> > > > [EMAIL PROTECTED]
> > > > > > > > > > > > >
> > > > > > > > > > > Subject:
> > > > > > > > > > > > Re: StackOverFlow Error in HBase  > >  > > You're
> > > > > > > > doing nothing
> > > > > > > > > > > > wrong.
> > > > > > > > > > > >  > >
> > > > > > > > > > > >  > > The filters as written recurse until
> > they find
> > > > > > > > > > > > a
> > > > > > > > match.  If
> > > > > > > > > > > > long  > > stretches between matching rows,
> > > > then you will
> > > > > > > > > > get a  > >
> > > > > > > > > > > > StackOverflowError.  Filters need to be changed.
> > > > > > Thanks for
> > > > > > > > > > > pointing
> > > > > > > > > > > >  > > this out.  Can you do without them for the
> > > > > > > > > > > > moment
> > > > > > > > > > until we get
> > > > > > > > > > > > a
> > > > > > > > > > > chance
> > > > > > > > > > > >  > > to fix it?  (HBASE-554)  > >  > > Thanks,  >
> > > > > > > St.Ack
> > > > > > > > > > > > > >  > >  > >  > > David Alves wrote:
> > > > > > > > > > > >  > > > Hi St.Ack and all  > > >
> > > > > > > > > > > >  > > >   The error always occurs when
> > trying to see if
> > > > > > > > > > there are more
> > > > > > > > > > > rows to
> > > > > > > > > > > >  > > > process.
> > > > > > > > > > > >  > > >   Yes I'm using a
> > filter(RegExpRowFilter) to
> > > > > > > > > > select only the rows
> > > > > > > > > > > (any
> > > > > > > > > > > >  > > > row key) that match a specific
> > value in one
> > > > > > > > > > > > of
> > > > > > > > the columns.
> > > > > > > > > > > >  > > >   Then I obtain the scanner just test
> > > > the hasNext
> > > > > > > > > > method, close
> > > > > > > > > > > the
> > > > > > > > > > > >  > > > scanner and return.
> > > > > > > > > > > >  > > >   Am I doing something wrong?
> > > > > > > > > > > >  > > >   Still StackOverflowError is not
> > supposed to
> > > > > > > > happen right?
> > > > > > > > > > > >  > > >
> > > > > > > > > > > >  > > > Regards
> > > > > > > > > > > >  > > > David Alves
> > > > > > > > > > > >  > > > On Thu, 2008-03-27 at 12:36 -0700,
> > stack wrote:
> > > > > > > > > > > >  > > >
> > > > > > > > > > > >  > > >> You are using a filter?  If so, tell us
> > > > > > more about it.
> > > > > > > > > > > >  > > >> St.Ack
> > > > > > > > > > > >  > > >>
> > > > > > > > > > > >  > > >> David Alves wrote:
> > > > > > > > > > > >  > > >>
> > > > > > > > > > > >  > > >>> Hi guys
> > > > > > > > > > > >  > > >>>
> > > > > > > > > > > >  > > >>>         I 'm using HBase to keep
> > data that is
> > > > > > > > > > later indexed.
> > > > > > > > > > > >  > > >>>         The data is indexed in
> > chunks so the
> > > > > > > > > > cycle is get XXXX
> > > > > > > > > > > records index
> > > > > > > > > > > >  > > >>> them check for more records etc...
> > > > > > > > > > > >  > > >>>         When I tryed the candidate-2
> > > > instead of
> > > > > > > > > > the old 0.16.0
> > > > > > > > > > > (which I
> > > > > > > > > > > >  > > >>> switched to do to the regionservers
> > > > > > > > > > > > becoming
> > > > > > > > > > unresponsive)
> > > > > > > > > > > > I
> > > > > > > > > > > got the
> > > > > > > > > > > >  > > >>> error in the end of this email
> > well into an
> > > > > > > > indexing job.
> > > > > > > > > > > >  > > >>>         So you have any idea
> > why? Am I doing
> > > > > > > > > > something wrong?
> > > > > > > > > > > >  > > >>>
> > > > > > > > > > > >  > > >>> David Alves  > > >>>  > > >>>
> > > > > > > > > > > > java.lang.RuntimeException:
> > > > > > > > > > > org.apache.hadoop.ipc.RemoteException:
> > > > > > > > > > > >  > > >>> java.io.IOException:
> > > > java.lang.StackOverflowError
> > > > > > > > > > > >  > > >>>         at
> > > > > > > > > > >
> > > > java.io.DataInputStream.readFully(DataInputStream.java:178
> > > > > > > > > > > )
> > > > > > > > > > > >  > > >>>         at
> > > > > > > > > > >
> > > > java.io.DataInputStream.readLong(DataInputStream.java:399)
> > > > > > > > > > > >  > > >>>         at
> > org.apache.hadoop.dfs.DFSClient
> > > > > > > > > > > >  > > >>>
> > $BlockReader.readChunk(DFSClient.java:735)
> > > > > > > > > > > >  > > >>>         at
> > > > > > > > > > > >  > > >>>
> > > > > > > > > > > >  > >
> > > > > > > > > > >
> > > > > > > > > >
> > > > org.apache.hadoop.fs.FSInputChecker.readChecksumChunk(FSInpu
> > > > > > > > > > tC
> > > > > > > > > > hecker.java:
> > > > > > > > > > > >  > > 234)
> > > > > > > > > > > >  > > >>>         at
> > > > > > > > > > > >  > > >>>
> > > > > > > > > > >
> > > > > > > >
> > > > org.apache.hadoop.fs.FSInputChecker.fill(FSInputChecker.java:176
> > > > > > > > )
> > > > > > > > > > > >  > > >>>         at
> > > > > > > > > > > >  > > >>>
> > > > > > > > > > >
> > > > > > > >
> > > > org.apache.hadoop.fs.FSInputChecker.read1(FSInputChecker.java:19
> > > > > > > > 3)
> > > > > > > > > > > >  > > >>>         at
> > > > > > > > > > > >  > > >>>
> > > > > > > > > > >
> > > > > > > >
> > > > org.apache.hadoop.fs.FSInputChecker.read(FSInputChecker.java:157
> > > > > > > > )
> > > > > > > > > > > >  > > >>>         at
> > org.apache.hadoop.dfs.DFSClient
> > > > > > > > > > > >  > > >>> $BlockReader.read(DFSClient.java:658)
> > > > > > > > > > > >  > > >>>         at
> > org.apache.hadoop.dfs.DFSClient
> > > > > > > > > > > >  > > >>>
> > > > $DFSInputStream.readBuffer(DFSClient.java:1130)
> > > > > > > > > > > >  > > >>>         at
> > org.apache.hadoop.dfs.DFSClient
> > > > > > > > > > > >  > > >>> $DFSInputStream.read(DFSClient.java:1166)
> > > > > > > > > > > >  > > >>>         at
> > > > > > > > > > >
> > > > java.io.DataInputStream.readFully(DataInputStream.java:178
> > > > > > > > > > > )
> > > > > > > > > > > >  > > >>>         at
> > > > org.apache.hadoop.io.DataOutputBuffer
> > > > > > > > > > > >  > > >>> $Buffer.write(DataOutputBuffer.java:56)
> > > > > > > > > > > >  > > >>>         at
> > > > > > > > > > > >  > > >>>
> > > > > > > > > > >
> > > > > > > > > >
> > > > > > > >
> > > > > >
> > > >
> > org.apache.hadoop.io.DataOutputBuffer.write(DataOutputBuffer.java:90
> > > > > > > > > > )
> > > > > > > > > > > >  > > >>>         at
> > org.apache.hadoop.io.SequenceFile
> > > > > > > > > > > >  > > >>> $Reader.next(SequenceFile.java:1829)
> > > > > > > > > > > >  > > >>>         at
> > org.apache.hadoop.io.SequenceFile
> > > > > > > > > > > >  > > >>> $Reader.next(SequenceFile.java:1729)
> > > > > > > > > > > >  > > >>>         at
> > org.apache.hadoop.io.SequenceFile
> > > > > > > > > > > >  > > >>> $Reader.next(SequenceFile.java:1775)
> > > > > > > > > > > >  > > >>>         at
> > > > > > > > > > >
> > > > org.apache.hadoop.io.MapFile$Reader.next(MapFile.java:461)
> > > > > > > > > > > >  > > >>>         at org.apache.hadoop.hbase.HStore
> > > > > > > > > > > >  > > >>>
> > $StoreFileScanner.getNext(HStore.java:2350)
> > > > > > > > > > > >  > > >>>         at
> > > > > > > > > > > >  > > >>>
> > > > > > > > > > > >  > >
> > > > > > > > > > >
> > > > > > > > > >
> > > > > > > >
> > > > > >
> > > >
> > org.apache.hadoop.hbase.HAbstractScanner.next(HAbstractScanner.java:
> > > > > > > > > > 25
> > > > > > > > > > > 6)
> > > > > > > > > > > >  > > >>>         at org.apache.hadoop.hbase.HStore
> > > > > > > > > > > >  > > >>> $HStoreScanner.next(HStore.java:2561)
> > > > > > > > > > > >  > > >>>         at
> > org.apache.hadoop.hbase.HRegion
> > > > > > > > > > > >  > > >>> $HScanner.next(HRegion.java:1807)
> > > > > > > > > > > >  > > >>>         at
> > org.apache.hadoop.hbase.HRegion
> > > > > > > > > > > >  > > >>> $HScanner.next(HRegion.java:1843)
> > > > > > > > > > > >  > > >>>         at
> > org.apache.hadoop.hbase.HRegion
> > > > > > > > > > > >  > > >>> $HScanner.next(HRegion.java:1843)
> > > > > > > > > > > >  > > >>>         at
> > org.apache.hadoop.hbase.HRegion
> > > > > > > > > > > >  > > >>> $HScanner.next(HRegion.java:1843)
> > > > > > > > > > > >  > > >>>         at
> > org.apache.hadoop.hbase.HRegion
> > > > > > > > > > > >  > > >>> $HScanner.next(HRegion.java:1843)
> > > > > > > > > > > >  > > >>>         at
> > org.apache.hadoop.hbase.HRegion
> > > > > > > > > > > >  > > >>> $HScanner.next(HRegion.java:1843)
> > > > > > > > > > > >  > > >>>         at
> > org.apache.hadoop.hbase.HRegion
> > > > > > > > > > > >  > > >>> $HScanner.next(HRegion.java:1843)
> > > > > > > > > > > >  > > >>>         at
> > org.apache.hadoop.hbase.HRegion
> > > > > > > > > > > >  > > >>> $HScanner.next(HRegion.java:1843)
> > > > > > > > > > > >  > > >>>         at
> > org.apache.hadoop.hbase.HRegion
> > > > > > > > > > > >  > > >>> $HScanner.next(HRegion.java:1843)
> > > > > > > > > > > >  > > >>>         at
> > org.apache.hadoop.hbase.HRegion
> > > > > > > > > > > >  > > >>> $HScanner.next(HRegion.java:1843)
> > > > > > > > > > > >  > > >>>         at
> > org.apache.hadoop.hbase.HRegion
> > > > > > > > > > > >  > > >>> $HScanner.next(HRegion.java:1843)
> > > > > > > > > > > >  > > >>>         at
> > org.apache.hadoop.hbase.HRegion
> > > > > > > > > > > >  > > >>> $HScanner.next(HRegion.java:1843)
> > > > > > > > > > > >  > > >>>         at
> > org.apache.hadoop.hbase.HRegion
> > > > > > > > > > > >  > > >>> $HScanner.next(HRegion.java:1843)
> > > > > > > > > > > >  > > >>>         at
> > org.apache.hadoop.hbase.HRegion
> > > > > > > > > > > >  > > >>> $HScanner.next(HRegion.java:1843)
> > > > > > > > > > > >  > > >>>         at
> > org.apache.hadoop.hbase.HRegion
> > > > > > > > > > > >  > > >>> $HScanner.next(HRegion.java:1843)
> > > > > > > > > > > >  > > >>>         at
> > org.apache.hadoop.hbase.HRegion
> > > > > > > > > > > >  > > >>> $HScanner.next(HRegion.java:1843)
> > > > > > > > > > > >  > > >>>         at
> > org.apache.hadoop.hbase.HRegion
> > > > > > > > > > > >  > > >>> $HScanner.next(HRegion.java:1843)
> > > > > > > > > > > >  > > >>>         at
> > org.apache.hadoop.hbase.HRegion
> > > > > > > > > > > >  > > >>> $HScanner.next(HRegion.java:1843)  >
> > > > > >>> ...
> > > > > > > > > > > >  > > >>>
> > > > > > > > > > > >  > > >>>
> > > > > > > > > > > >  > > >>>
> > > > > > > > > > > >  > > >>>
> > > > > > > > > > > >  > > >
> > > > > > > > > > > >  > > >
> > > > > > > > > > > >  >
> > > > > > > > > > > >
> > > > > > > > > > > >
> > > > > > > > > >
> > > > > > > > > >
> > > > > > > > > > No virus found in this incoming message.
> > > > > > > > > > Checked by AVG.
> > > > > > > > > > Version: 7.5.519 / Virus Database: 269.22.3/1354 -
> > > > > > > > > > Release
> > > > > > > > > > Date: 4/1/2008 5:38 AM
> > > > > > > > > >
> > > > > > > > > >
> > > > > > > > >
> > > > > > > > > No virus found in this outgoing message.
> > > > > > > > > Checked by AVG.
> > > > > > > > > Version: 7.5.519 / Virus Database: 269.22.3/1354 -
> > > > Release Date:
> > > > > > > > > 4/1/2008
> > > > > > > > > 5:38 AM
> > > > > > > >
> > > > > > > >
> > > > > > > > No virus found in this incoming message.
> > > > > > > > Checked by AVG.
> > > > > > > > Version: 7.5.519 / Virus Database: 269.22.3/1354 - Release
> > > > > > > > Date: 4/1/2008 5:38 AM
> > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > > > No virus found in this outgoing message.
> > > > > > > Checked by AVG.
> > > > > > > Version: 7.5.519 / Virus Database: 269.22.5/1357 -
> > Release Date:
> > > > > > > 4/3/2008 10:48 AM
> > > > > > >
> > > > > >
> > > > > >
> > > > > > No virus found in this incoming message.
> > > > > > Checked by AVG.
> > > > > > Version: 7.5.519 / Virus Database: 269.22.5/1357 - Release
> > > > > > Date: 4/3/2008 10:48 AM
> > > > > >
> > > > > >
> > > > >
> > > > > No virus found in this outgoing message.
> > > > > Checked by AVG.
> > > > > Version: 7.5.519 / Virus Database: 269.22.5/1357 - Release Date:
> > > > > 4/3/2008 10:48 AM
> > > > >
> > > >
> > > >
> > > > No virus found in this incoming message.
> > > > Checked by AVG.
> > > > Version: 7.5.519 / Virus Database: 269.22.5/1359 - Release
> > > > Date: 4/4/2008 8:23 AM
> > > >
> > > >
> > >
> > > No virus found in this outgoing message.
> > > Checked by AVG.
> > > Version: 7.5.519 / Virus Database: 269.22.5/1359 - Release Date:
> > > 4/4/2008 8:23 AM
> > >
> >
> >
> > No virus found in this incoming message.
> > Checked by AVG.
> > Version: 7.5.519 / Virus Database: 269.22.5/1359 - Release
> > Date: 4/4/2008 8:23 AM
> >
> >
> 
> No virus found in this outgoing message.
> Checked by AVG.
> Version: 7.5.519 / Virus Database: 269.22.8/1362 - Release Date: 4/6/2008 
> 11:12 AM
> 

Reply via email to