Hmm,

It'd be good to find one other person having problems with it on Cloudera's
distro as I could just be making a mistake somewhere and it has nothing to
do with the fact that I'm using cloudera's distro.

~Ed

On Thu, Sep 30, 2010 at 8:47 AM, Rohan Rai <rohan....@inmobi.com> wrote:

> The validation done by me was on
> Apache 0.20.2 and Apache Pig 0.7..
>
> I haven't tried it with Cloudera's version.
>
> Can we verify that it doesnt work  with them too
>
>
> Regards
> Rohan
>
> ed wrote:
>
>> Hello,
>>
>> I'm using Cloudera's Hadoop CDH3B2--Hadoop-0.20.2+320 (based on Apache
>> Hadoop 20.2) with Pig 0.7 (from Cloudera's distro).
>>
>> Thank you!
>>
>> ~Ed
>>
>>
>> On Wed, Sep 29, 2010 at 11:56 PM, Rohan Rai <rohan....@inmobi.com> wrote:
>>
>>
>>  Hi
>>>
>>> Which Hadoop/ PIg version are you using ??
>>>
>>> Regards
>>> Rohan
>>>
>>>
>>> ed wrote:
>>>
>>>
>>>  Hello,
>>>>
>>>> I tested the newest push to the hirohanin elephant-bird branch (for pig
>>>> 0.7)
>>>> and had an error when trying to use LzoTokenizedLoader with the
>>>> following
>>>> pig script:
>>>>
>>>>    REGISTER elephant-bird-1.0.jar
>>>>    REGISTER /usr/lib/elephant-bird/lib/google-collect-1.0.jar
>>>>    A = load '/usr/foo/input/test_input_chars.txt.lzo' USING
>>>> com.twitter.elephantbird.pig.load.LzoTokenizedLoader('\t');
>>>>    DUMP A;
>>>>
>>>> The error I get is in the mapper logs and is as follows:
>>>>
>>>> INFO com.hadoop.compression.lzo.GPLNativeCodeLoader: Loaded native gpl
>>>> library
>>>> INFO com.hadoop.compression.lzo.LzoCodec: Succesfully loaded &
>>>> initialized
>>>> native-lzo library
>>>> INFO com.twitter.elephantbird.pig.load.LzoTokenizedLoader:
>>>> LzoTokenizedLoader with given delimiter [     ]
>>>> INFO com.twitter.elephantbird.mapreduce.input.LzoRecordReader: Seeking
>>>> to
>>>> split start at pos 0
>>>> FATAL org.apache.hadoop.mapred.TaskTracker: Error running child :
>>>> java.lang.NoSuchMethodError:
>>>>
>>>>
>>>> org.apache.pig.backend.executionengine.mapReduceLayer.PigHadoopLogger.getTaskIOCContext()Lorg/apache/hadoop/mapreduce/TaskInputOutputContext;
>>>>    at com.twitter.elephantbird.pig.util.PigCounterHelper.getTIOC
>>>> (Unknown
>>>> Source)
>>>>    at com.twitter.elephantbird.pig.util.PigCounterHelper.incrCounter
>>>> (Unknown Source)
>>>>    at com.twitter.elephantbird.pig.load.LzoBaseLoadFunc.incrCounter
>>>> (Unknown Source)
>>>>    at com.twitter.elephantbird.pig.load.LzoTokenizedLoader.getNext
>>>> (Unknown Source)
>>>>    at
>>>>
>>>>
>>>> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigRecordReader.nextKeyValue
>>>> (PigRecordReader.java:142)
>>>>    at
>>>>
>>>>
>>>> org.apache.hadoop.mapred.MapTask$NewTrackignRecordReader.nextKeyValue(MapTask.java:423)
>>>>    at org.apache.hadoop.mapreduce.MapContext.nextKeyValue
>>>> (MapContent.java:67)
>>>>    at org.apache.hadoop.mapreduce.Mapper.run (Mapper.java:143)
>>>>    at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:621)
>>>>    at org.apache.hadoop.mapred.MapTask.run(MapTask.java:305)
>>>>    at org.apache.hadoop.mapred.Child.main(Child.java:170)
>>>>
>>>> Do you think I'm forgetting some required library?
>>>>
>>>> Thank you!
>>>>
>>>> ~Ed
>>>>
>>>> On Tue, Sep 28, 2010 at 2:10 PM, ed <hadoopn...@gmail.com> wrote:
>>>>
>>>>
>>>>  Thank you Rohan,  I really appreciate your help!  I'll give it shot and
>>>>
>>>>  post back if it works.
>>>>>
>>>>> ~Ed
>>>>>
>>>>>
>>>>> On Mon, Sep 27, 2010 at 11:51 PM, Rohan Rai <rohan....@inmobi.com>
>>>>> wrote:
>>>>>
>>>>>
>>>>>  Just corrected/tested and pushed LzoTokenizedLoader to the personal
>>>>> fork
>>>>>
>>>>>  Hopefully it works now
>>>>>>
>>>>>>
>>>>>> Regards
>>>>>> Rohan
>>>>>>
>>>>>> Dmitriy Ryaboy wrote:
>>>>>>
>>>>>>
>>>>>>  lzop should work.
>>>>>>
>>>>>>  On Mon, Sep 27, 2010 at 10:59 AM, Rohan Rai <rohan....@inmobi.com>
>>>>>>> wrote:
>>>>>>>
>>>>>>>
>>>>>>>  Well
>>>>>>>
>>>>>>>  I haven't tried (rather I don't remember) compressing via lzop and
>>>>>>>
>>>>>>>  then
>>>>>>>> putting on cluster...
>>>>>>>> So cant tell you about that...Here is what works for me.
>>>>>>>>
>>>>>>>> I do it by first putting the file on cluster and then doing Stream
>>>>>>>> Compression.
>>>>>>>>
>>>>>>>> And yes it need not be indexed (I guess it doesn't matter for  small
>>>>>>>> test file, otherwise it is unwise
>>>>>>>> for one loses the benefit of parallelism)
>>>>>>>>
>>>>>>>> Regards
>>>>>>>> Rohan
>>>>>>>>
>>>>>>>>
>>>>>>>> pig wrote:
>>>>>>>>
>>>>>>>>
>>>>>>>>  Hi Rohan,
>>>>>>>>
>>>>>>>>  The test file (test_input_chars.txt.lzo) is not indexed.  I created
>>>>>>>>
>>>>>>>>  it
>>>>>>>>> using
>>>>>>>>> the command
>>>>>>>>>
>>>>>>>>> 'lzop test_input_chars.txt'
>>>>>>>>>
>>>>>>>>> It's a really small file (only 6 lines) so I didn't think it needed
>>>>>>>>> to
>>>>>>>>> be
>>>>>>>>> index.  Do all files regardless of size need to be indexed for the
>>>>>>>>> LzoTokenizedLoader to work?
>>>>>>>>>
>>>>>>>>> Thank you!
>>>>>>>>>
>>>>>>>>> ~Ed
>>>>>>>>>
>>>>>>>>> On Mon, Sep 27, 2010 at 1:25 AM, Rohan Rai <rohan....@inmobi.com>
>>>>>>>>> wrote:
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>  Oh  Sorry I am completely out of  sync...
>>>>>>>>>
>>>>>>>>>  Can you tell how did you lzo'ed and indexed  the file
>>>>>>>>>
>>>>>>>>>  Regards
>>>>>>>>>
>>>>>>>>>  Rohan
>>>>>>>>>>
>>>>>>>>>> Rohan Rai wrote:
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>  Oh Sorry I did not see this mail ...
>>>>>>>>>>
>>>>>>>>>>  Its not an official patch/release
>>>>>>>>>>
>>>>>>>>>>  But here is a fork on elephant-bird which works with pig 0.7
>>>>>>>>>>
>>>>>>>>>>  for  normal LZOText Loading etc
>>>>>>>>>>>
>>>>>>>>>>> (NOt HbaseLoader)
>>>>>>>>>>>
>>>>>>>>>>> Regards
>>>>>>>>>>> Rohan
>>>>>>>>>>>
>>>>>>>>>>> Dmitriy Ryaboy wrote:
>>>>>>>>>>>
>>>>>>>>>>>  The 0.7 branch is not tested.. it's quite likely it doesn't
>>>>>>>>>>> actually
>>>>>>>>>>> work
>>>>>>>>>>>
>>>>>>>>>>>  :).
>>>>>>>>>>>
>>>>>>>>>>>  Rohan Rai was working on it.. Rohan, think you can take a look
>>>>>>>>>>> and
>>>>>>>>>>>
>>>>>>>>>>>  help
>>>>>>>>>>>
>>>>>>>>>>>  Ed
>>>>>>>>>>>> out?
>>>>>>>>>>>>
>>>>>>>>>>>> Ed, you may want to check if the same input works when you use
>>>>>>>>>>>> Pig
>>>>>>>>>>>> 0.6
>>>>>>>>>>>> (and
>>>>>>>>>>>> the official elephant-bird, on Kevin Weil's github).
>>>>>>>>>>>>
>>>>>>>>>>>> -D
>>>>>>>>>>>>
>>>>>>>>>>>> On Thu, Sep 23, 2010 at 6:49 AM, pig <hadoopn...@gmail.com>
>>>>>>>>>>>> wrote:
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>  Hello,
>>>>>>>>>>>>
>>>>>>>>>>>>  After getting all the errors to go away with LZO libraries not
>>>>>>>>>>>> being
>>>>>>>>>>>>
>>>>>>>>>>>>  found
>>>>>>>>>>>>
>>>>>>>>>>>>  and missing jar files for elephant-bird I've run into a new
>>>>>>>>>>>>
>>>>>>>>>>>>  problem
>>>>>>>>>>>>> when
>>>>>>>>>>>>> using the elephant-bird branch for pig 0.7
>>>>>>>>>>>>>
>>>>>>>>>>>>> The following simple pig script works as expected
>>>>>>>>>>>>>
>>>>>>>>>>>>>  REGISTER elephant-bird-1.0.jar
>>>>>>>>>>>>>  REGISTER /usr/lib/elephant-bird/lib/google-collect-1.0.jar
>>>>>>>>>>>>>  A = load '/usr/foo/input/test_input_chars.txt';
>>>>>>>>>>>>>  DUMP A;
>>>>>>>>>>>>>
>>>>>>>>>>>>> This just dumps out the contents of the test_input_chars.txt
>>>>>>>>>>>>> file
>>>>>>>>>>>>> which
>>>>>>>>>>>>> is
>>>>>>>>>>>>> tab delimited. The output looks like:
>>>>>>>>>>>>>
>>>>>>>>>>>>>  (1,a,a,a,a,a,a)
>>>>>>>>>>>>>  (2,b,b,b,b,b,b)
>>>>>>>>>>>>>  (3,c,c,c,c,c,c)
>>>>>>>>>>>>>  (4,d,d,d,d,d,d)
>>>>>>>>>>>>>  (5,e,e,e,e,e,e)
>>>>>>>>>>>>>
>>>>>>>>>>>>> I then lzop the test file to get test_input_chars.txt.lzo (I
>>>>>>>>>>>>> decompressed
>>>>>>>>>>>>> this with lzop -d to make sure the compression worked fine and
>>>>>>>>>>>>> everything
>>>>>>>>>>>>> looks good).
>>>>>>>>>>>>> If I run the exact same script provided above on the lzo file
>>>>>>>>>>>>> it
>>>>>>>>>>>>> works
>>>>>>>>>>>>> fine.  However, this file is really small and doesn't need to
>>>>>>>>>>>>> use
>>>>>>>>>>>>> indexes.
>>>>>>>>>>>>> As a result, I wanted to
>>>>>>>>>>>>> have LZO support that worked with indexes.  Based on this I
>>>>>>>>>>>>> decided
>>>>>>>>>>>>> to
>>>>>>>>>>>>> try
>>>>>>>>>>>>> out the elephant-bird branch for pig 0.7 located here (
>>>>>>>>>>>>> http://github.com/hirohanin/elephant-bird/) as
>>>>>>>>>>>>> recommended by Dimitriy.
>>>>>>>>>>>>>
>>>>>>>>>>>>> I created the following pig script that mirrors the above
>>>>>>>>>>>>> script
>>>>>>>>>>>>> but
>>>>>>>>>>>>> should
>>>>>>>>>>>>> hopefully work on LZO files (including indexed ones)
>>>>>>>>>>>>>
>>>>>>>>>>>>>  REGISTER elephant-bird-1.0.jar
>>>>>>>>>>>>>  REGISTER /usr/lib/elephant-bird/lib/google-collect-1.0.jar
>>>>>>>>>>>>>  A = load '/usr/foo/input/test_input_chars.txt.lzo' USING
>>>>>>>>>>>>> com.twitter.elephantbird.pig.load.LzoTokenizedLoader('\t');
>>>>>>>>>>>>>  DUMP A;
>>>>>>>>>>>>>
>>>>>>>>>>>>> When I run this script which uses the LzoTokenizedLoader there
>>>>>>>>>>>>> is
>>>>>>>>>>>>> no
>>>>>>>>>>>>> output.  The script appears to run without errors but there are
>>>>>>>>>>>>> zero
>>>>>>>>>>>>> Records
>>>>>>>>>>>>> Written and 0 Bytes Written.
>>>>>>>>>>>>>
>>>>>>>>>>>>> Here is the exact output:
>>>>>>>>>>>>>
>>>>>>>>>>>>> grunt > DUMP A;
>>>>>>>>>>>>> [main] INFO
>>>>>>>>>>>>> com.twitter.elephantbird.pig.load.LzoTokenizedLoader
>>>>>>>>>>>>> -
>>>>>>>>>>>>> LzoTokenizedLoader with given delimited [     ]
>>>>>>>>>>>>> [main] INFO
>>>>>>>>>>>>> com.twitter.elephantbird.pig.load.LzoTokenizedLoader
>>>>>>>>>>>>> -
>>>>>>>>>>>>> LzoTokenizedLoader with given delimited [     ]
>>>>>>>>>>>>> [main] INFO
>>>>>>>>>>>>> com.twitter.elephantbird.pig.load.LzoTokenizedLoader
>>>>>>>>>>>>> -
>>>>>>>>>>>>> LzoTokenizedLoader with given delimited [     ]
>>>>>>>>>>>>> [main] INFO
>>>>>>>>>>>>> org.apache.pig.backend.hadoop.executionengine.HExecutionEngine
>>>>>>>>>>>>> -
>>>>>>>>>>>>> (Name:
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> Store(hdfs://master:9000/tmp/temp-2052828736/tmp-1533645117:org.apache.pig.builtin.BinStorage)
>>>>>>>>>>>>> - 1-4 Operator Key: 1-4
>>>>>>>>>>>>> [main] INFO
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MultiQueryOptimizer
>>>>>>>>>>>>> - MR plan size before optimization: 1
>>>>>>>>>>>>> [main] INFO
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MultiQueryOptimizer
>>>>>>>>>>>>> - MR plan size after optimization: 1
>>>>>>>>>>>>> [main] INFO
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.JobControlCompiler
>>>>>>>>>>>>> - mapred.job.reduce.markreset.buffer.percent is not set, set to
>>>>>>>>>>>>> default
>>>>>>>>>>>>> 0.3
>>>>>>>>>>>>> [main] INFO
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.JobControlCompiler
>>>>>>>>>>>>> - Setting up single store job
>>>>>>>>>>>>> [main] INFO
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher
>>>>>>>>>>>>> - 1 map-reduce job(s) waiting for submission.
>>>>>>>>>>>>> [Thread-12] WARN org.apache.hadoop.mapred.JobClient - Use
>>>>>>>>>>>>> GenericOptionsParser for parsing the arguments.  Applications
>>>>>>>>>>>>> should
>>>>>>>>>>>>> implement Tool for the same.
>>>>>>>>>>>>> [Thread-12] INFO
>>>>>>>>>>>>> com.twitter.elephantbird.pig.load.LzoTokenizedLoader
>>>>>>>>>>>>> -
>>>>>>>>>>>>> LzoTokenizedLoader with given delimiter [     ]
>>>>>>>>>>>>> [Thread-12] INFO
>>>>>>>>>>>>> org.apache.hadoop.mapreduce.lib.input.FileInputFormat
>>>>>>>>>>>>> -
>>>>>>>>>>>>> Total input paths to process : 1
>>>>>>>>>>>>> [main] INFO
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher
>>>>>>>>>>>>> - 0% complete
>>>>>>>>>>>>> [main] INFO
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher
>>>>>>>>>>>>> - HadoopJobId: job_201009101108_0151
>>>>>>>>>>>>> [main] INFO
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher
>>>>>>>>>>>>> - More information at
>>>>>>>>>>>>> http://master:50030/jobdetails.jsp?jobid=job_201009101108_0151
>>>>>>>>>>>>> [main] INFO
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher
>>>>>>>>>>>>> - 50% complete
>>>>>>>>>>>>> [main] INFO
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher
>>>>>>>>>>>>> - 100% complete
>>>>>>>>>>>>> [main] INFO
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher
>>>>>>>>>>>>> - Succesfully stored result in
>>>>>>>>>>>>> "hdfs://amb-hadoop-01:9000/tmp/temp-2052828736/tmp-1533645117
>>>>>>>>>>>>> [main] INFO
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher
>>>>>>>>>>>>> - Records written: 0
>>>>>>>>>>>>> [main] INFO
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher
>>>>>>>>>>>>> - Bytes written: 0
>>>>>>>>>>>>> [main] INFO
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher
>>>>>>>>>>>>> - Spillable Memory Manager spill count : 0
>>>>>>>>>>>>> [main] INFO
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher
>>>>>>>>>>>>> - Proactive spill count : 0
>>>>>>>>>>>>> [main] INFO
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher
>>>>>>>>>>>>> - Success!
>>>>>>>>>>>>> [main] INFO
>>>>>>>>>>>>> org.apache.hadoop.mapreduce.lib.input.FileInputFormat
>>>>>>>>>>>>> -
>>>>>>>>>>>>> Total
>>>>>>>>>>>>> input paths to process: 1
>>>>>>>>>>>>> [main] INFO
>>>>>>>>>>>>> org.apache.pig.backend.hadoop.executionengine.util.MapRedUtil -
>>>>>>>>>>>>> Total input paths to process: 1
>>>>>>>>>>>>> grunt >
>>>>>>>>>>>>>
>>>>>>>>>>>>> I'm not sure if I'm doing something wrong in my use of
>>>>>>>>>>>>> LzoTokenizedLoader
>>>>>>>>>>>>> or
>>>>>>>>>>>>> if there is a problem with the class itself (most likely the
>>>>>>>>>>>>> problem
>>>>>>>>>>>>> is
>>>>>>>>>>>>> with
>>>>>>>>>>>>> my code heh)  Thank you for any help!
>>>>>>>>>>>>>
>>>>>>>>>>>>> ~Ed
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>  .
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>  The information contained in this communication is intended
>>>>>>>>>>>>>
>>>>>>>>>>>>>  solely for
>>>>>>>>>>>>>
>>>>>>>>>>>>>   the
>>>>>>>>>>>>
>>>>>>>>>>>>  use of the individual or entity to whom it is addressed and
>>>>>>>>>>>>
>>>>>>>>>>>>  others
>>>>>>>>>>> authorized to receive it. It may contain confidential or legally
>>>>>>>>>>> privileged
>>>>>>>>>>> information. If you are not the intended recipient you are hereby
>>>>>>>>>>> notified
>>>>>>>>>>> that any disclosure, copying, distribution or taking any action
>>>>>>>>>>> in
>>>>>>>>>>> reliance
>>>>>>>>>>> on the contents of this information is strictly prohibited and
>>>>>>>>>>> may
>>>>>>>>>>> be
>>>>>>>>>>> unlawful. If you have received this communication in error,
>>>>>>>>>>> please
>>>>>>>>>>> notify us
>>>>>>>>>>> immediately by responding to this email and then delete it from
>>>>>>>>>>> your
>>>>>>>>>>> system.
>>>>>>>>>>> The firm is neither liable for the proper and complete
>>>>>>>>>>> transmission
>>>>>>>>>>> of
>>>>>>>>>>> the
>>>>>>>>>>> information contained in this communication nor for any delay in
>>>>>>>>>>> its
>>>>>>>>>>> receipt.
>>>>>>>>>>> .
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>  The information contained in this communication is intended
>>>>>>>>>>> solely
>>>>>>>>>>> for
>>>>>>>>>>>
>>>>>>>>>>>  the
>>>>>>>>>>>
>>>>>>>>>>>  use of the individual or entity to whom it is addressed and
>>>>>>>>>>> others
>>>>>>>>>>>
>>>>>>>>>>>  authorized to receive it. It may contain confidential or legally
>>>>>>>>>> privileged
>>>>>>>>>> information. If you are not the intended recipient you are hereby
>>>>>>>>>> notified
>>>>>>>>>> that any disclosure, copying, distribution or taking any action in
>>>>>>>>>> reliance
>>>>>>>>>> on the contents of this information is strictly prohibited and may
>>>>>>>>>> be
>>>>>>>>>> unlawful. If you have received this communication in error, please
>>>>>>>>>> notify
>>>>>>>>>> us
>>>>>>>>>> immediately by responding to this email and then delete it from
>>>>>>>>>> your
>>>>>>>>>> system.
>>>>>>>>>> The firm is neither liable for the proper and complete
>>>>>>>>>> transmission
>>>>>>>>>> of
>>>>>>>>>> the
>>>>>>>>>> information contained in this communication nor for any delay in
>>>>>>>>>> its
>>>>>>>>>> receipt.
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>  .
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>  The information contained in this communication is intended
>>>>>>>>>> solely
>>>>>>>>>>
>>>>>>>>>>  for
>>>>>>>>>
>>>>>>>>>  the
>>>>>>>>>
>>>>>>>>>  use of the individual or entity to whom it is addressed and others
>>>>>>>> authorized to receive it. It may contain confidential or legally
>>>>>>>> privileged
>>>>>>>> information. If you are not the intended recipient you are hereby
>>>>>>>> notified
>>>>>>>> that any disclosure, copying, distribution or taking any action in
>>>>>>>> reliance
>>>>>>>> on the contents of this information is strictly prohibited and may
>>>>>>>> be
>>>>>>>> unlawful. If you have received this communication in error, please
>>>>>>>> notify us
>>>>>>>> immediately by responding to this email and then delete it from your
>>>>>>>> system.
>>>>>>>> The firm is neither liable for the proper and complete transmission
>>>>>>>> of
>>>>>>>> the
>>>>>>>> information contained in this communication nor for any delay in its
>>>>>>>> receipt.
>>>>>>>>
>>>>>>>>
>>>>>>>>  .
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>   The information contained in this communication is intended solely
>>>>>>> for
>>>>>>>
>>>>>>>  the
>>>>>> use of the individual or entity to whom it is addressed and others
>>>>>> authorized to receive it. It may contain confidential or legally
>>>>>> privileged
>>>>>> information. If you are not the intended recipient you are hereby
>>>>>> notified
>>>>>> that any disclosure, copying, distribution or taking any action in
>>>>>> reliance
>>>>>> on the contents of this information is strictly prohibited and may be
>>>>>> unlawful. If you have received this communication in error, please
>>>>>> notify us
>>>>>> immediately by responding to this email and then delete it from your
>>>>>> system.
>>>>>> The firm is neither liable for the proper and complete transmission of
>>>>>> the
>>>>>> information contained in this communication nor for any delay in its
>>>>>> receipt.
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>   .
>>>>>
>>>>>
>>>>
>>>>  The information contained in this communication is intended solely for
>>> the
>>> use of the individual or entity to whom it is addressed and others
>>> authorized to receive it. It may contain confidential or legally
>>> privileged
>>> information. If you are not the intended recipient you are hereby
>>> notified
>>> that any disclosure, copying, distribution or taking any action in
>>> reliance
>>> on the contents of this information is strictly prohibited and may be
>>> unlawful. If you have received this communication in error, please notify
>>> us
>>> immediately by responding to this email and then delete it from your
>>> system.
>>> The firm is neither liable for the proper and complete transmission of
>>> the
>>> information contained in this communication nor for any delay in its
>>> receipt.
>>>
>>>
>>>  .
>>
>>
>>
>
> The information contained in this communication is intended solely for the
> use of the individual or entity to whom it is addressed and others
> authorized to receive it. It may contain confidential or legally privileged
> information. If you are not the intended recipient you are hereby notified
> that any disclosure, copying, distribution or taking any action in reliance
> on the contents of this information is strictly prohibited and may be
> unlawful. If you have received this communication in error, please notify us
> immediately by responding to this email and then delete it from your system.
> The firm is neither liable for the proper and complete transmission of the
> information contained in this communication nor for any delay in its
> receipt.
>

Reply via email to