Yes! Reproduced on single-node cluster:

10/04/28 16:30:24 INFO mapred.JobClient:     ROWS=274884
10/04/28 16:30:24 INFO mapred.JobClient:     TOMBSTONES=951083

10/04/28 16:42:49 INFO mapred.JobClient:     ROWS=166580
10/04/28 16:42:49 INFO mapred.JobClient:     TOMBSTONES=1059387

On Wed, Apr 28, 2010 at 10:43 AM, Jonathan Ellis <jbel...@gmail.com> wrote:
> It sounds like either there is a fairly obvious bug, or you're doing
> something wrong. :)
>
> Can you reproduce against a single node?
>
> On Tue, Apr 27, 2010 at 5:14 PM, Joost Ouwerkerk <jo...@openplaces.org> wrote:
>> Update: I ran a test whereby I deleted ALL the rows in a column
>> family, using a consistency level of ALL.  To do this, I mapped the
>> ColumnFamily and called remove on each row id.  There were 1.5 million
>> rows, so 1.5 million rows were deleted.
>>
>> I ran a counter job immediately after.  This job maps the same column
>> family and tests if any data is returned.  If not, it considers the
>> row a "tombstone".  If yes, it considers the row not deleted.  Below
>> are the hadoop counters for those jobs.  Note the fluctuation in the
>> number of rows with data over time, and the increase in time to map
>> the column family after the destroy job.  No other clients were
>> accessing cassandra during this time.
>>
>> I'm thoroughly confused.
>>
>> Count: started 13:02:30 EDT, finished 13:11:33 EDT (9 minutes 2 seconds):
>>   ROWS:        1,542,479
>>   TOMBSTONES:  69
>>
>> Destroy: started 16:48:45 EDT, finished 17:07:36 EDT (18 minutes 50 seconds)
>>   DESTROYED:  1,542,548
>>
>> Count: started 17:15:42 EDT, finished 17:31:03 EDT (15 minutes 21 seconds)
>>   ROWS 876,464
>>   TOMBSTONES   666,084
>>
>> Count: started 17:31:32, finished 17:47:16 (15mins, 44 seconds)
>>   ROWS 1,451,665
>>   TOMBSTONES   90,883
>>
>> Count: started 17:52:34, finished 18:10:28 (17mins, 53 seconds)
>>   ROWS 1,425,644
>>   TOMBSTONES   116,904
>>
>> On Tue, Apr 27, 2010 at 5:37 PM, Joost Ouwerkerk <jo...@openplaces.org> 
>> wrote:
>>> Clocks are in sync:
>>>
>>> cluster04:~/cassandra$ dsh -g development "date"
>>> Tue Apr 27 17:36:33 EDT 2010
>>> Tue Apr 27 17:36:33 EDT 2010
>>> Tue Apr 27 17:36:33 EDT 2010
>>> Tue Apr 27 17:36:33 EDT 2010
>>> Tue Apr 27 17:36:34 EDT 2010
>>> Tue Apr 27 17:36:34 EDT 2010
>>> Tue Apr 27 17:36:34 EDT 2010
>>> Tue Apr 27 17:36:34 EDT 2010
>>> Tue Apr 27 17:36:34 EDT 2010
>>> Tue Apr 27 17:36:35 EDT 2010
>>> Tue Apr 27 17:36:35 EDT 2010
>>> Tue Apr 27 17:36:35 EDT 2010
>>>
>>> On Tue, Apr 27, 2010 at 5:35 PM, Nathan McCall <n...@vervewireless.com> 
>>> wrote:
>>>> Have you confirmed that your clocks are all synced in the cluster?
>>>> This may be the result of an unintentional read-repair occurring if
>>>> that were the case.
>>>>
>>>> -Nate
>>>>
>>>> On Tue, Apr 27, 2010 at 2:20 PM, Joost Ouwerkerk <jo...@openplaces.org> 
>>>> wrote:
>>>>> Hmm... Even after deleting with cl.ALL, I'm getting data back for some
>>>>> rows after having deleted them.  Which rows return data is
>>>>> inconsistent from one run of the job to the next.
>>>>>
>>>>> On Tue, Apr 27, 2010 at 1:44 PM, Joost Ouwerkerk <jo...@openplaces.org> 
>>>>> wrote:
>>>>>> To check that rows are gone, I check that KeySlice.columns is empty.  
>>>>>> And as
>>>>>> I mentioned, immediately after the delete job, this returns the expected
>>>>>> number.
>>>>>> Unfortunately I reproduced with QUORUM this morning.  No node outages.  
>>>>>> I am
>>>>>> going to try ALL to see if that changes anything, but I am starting to
>>>>>> wonder if I'm doing something else wrong.
>>>>>> On Mon, Apr 26, 2010 at 9:45 PM, Jonathan Ellis <jbel...@gmail.com> 
>>>>>> wrote:
>>>>>>>
>>>>>>> How are you checking that the rows are gone?
>>>>>>>
>>>>>>> Are you experiencing node outages during this?
>>>>>>>
>>>>>>> DC_QUORUM is unfinished code right now, you should avoid using it.
>>>>>>> Can you reproduce with normal QUORUM?
>>>>>>>
>>>>>>> On Sat, Apr 24, 2010 at 12:23 PM, Joost Ouwerkerk <jo...@openplaces.org>
>>>>>>> wrote:
>>>>>>> > I'm having trouble deleting rows in Cassandra.  After running a job 
>>>>>>> > that
>>>>>>> > deletes hundreds of rows, I run another job that verifies that the 
>>>>>>> > rows
>>>>>>> > are
>>>>>>> > gone.  Both jobs run correctly.  However, when I run the verification
>>>>>>> > job an
>>>>>>> > hour later, the rows have re-appeared.  This is not a case of 
>>>>>>> > "ghosting"
>>>>>>> > because the verification job actually checks that there is data in the
>>>>>>> > columns.
>>>>>>> >
>>>>>>> > I am running a cluster with 12 nodes and a replication factor of 3.  I
>>>>>>> > am
>>>>>>> > using DC_QUORUM consistency when deleting.
>>>>>>> >
>>>>>>> > Any ideas?
>>>>>>> > Joost.
>>>>>>> >
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> --
>>>>>>> Jonathan Ellis
>>>>>>> Project Chair, Apache Cassandra
>>>>>>> co-founder of Riptano, the source for professional Cassandra support
>>>>>>> http://riptano.com
>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>
>
>
> --
> Jonathan Ellis
> Project Chair, Apache Cassandra
> co-founder of Riptano, the source for professional Cassandra support
> http://riptano.com
>

Reply via email to