Can you include your read code? On Aug 18, 2015 5:50 AM, "Hervé Rivière" <herve.rivi...@zenika.com> wrote:
> Hello, > > > > > > I have an issue with a <ErrorMessage code=0000 [Server error] > message="java.lang.NullPointerException"> when I query a table with static > fields (without where clause) with Cassandra 2.1.8 / 2 nodes clusters. > > > > No more indication in the log : > > ERROR [SharedPool-Worker-1] 2015-08-18 10:39:02,549 QueryMessage.java:132 > - Unexpected error during query > > java.lang.NullPointerException: null > > ERROR [SharedPool-Worker-1] 2015-08-18 10:39:02,550 ErrorMessage.java:251 > - Unexpected exception during request > > java.lang.NullPointerException: null > > > > > > The scenario was : > > 1) loading data inside the table with spark (~12 million rows) > > 2) Make some deletes with the primary keys and use the static fields to > keep a certain state for each partition. > > > > The null pointer exception occurs when I query all the table after I made > some deletions. > > > > I observed that : > > - Before delete statement the table is perfectly readable > > - It's repeatable (I achieved to isolate ~20 delete statements that > create a null pointer exception when they are executed by cqlsh) > > - it occurs only with some rows (nothing special in these rows compared > to others) > > - Didn't succeed to repeat the problem with the problematic rows inside a > toy table > > - repair/compact and scrub on each node before and after the deletes > statements didn't change anything (always the null pointer exception after > the delete) > > - Maybe related with static columns ? > > > > The table structure is : > > CREATE TABLE my_table ( > > pk1 text, > > pk2 text, > > ck1 timestamp, > > ck2 text, > > ck3 text, > > valuefield text, > > staticField1 text static, > > staticField2 text static, > > PRIMARY KEY ((pk1, pk2), ck1, ck2, ck3) > > ) WITH CLUSTERING ORDER BY (pk1 DESC, pk2 ASC, ck1 ASC) > > AND bloom_filter_fp_chance = 0.01 > > AND caching = '{"keys":"ALL", "rows_per_partition":"NONE"}' > > AND compaction = {'class': > 'org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy'} > > AND compression = {'sstable_compression': > 'org.apache.cassandra.io.compress.LZ4Compressor'} > > AND dclocal_read_repair_chance = 0.1 > > AND default_time_to_live = 0 > > AND gc_grace_seconds = 0 > > AND max_index_interval = 2048 > > AND memtable_flush_period_in_ms = 0 > > AND min_index_interval = 128 > > AND read_repair_chance = 0.0 > > AND speculative_retry = '99.0PERCENTILE'; > > > > > > > > > > Is someone already met this issue or has an idea to solve/investigate this > exceptions ? > > > > > > Thank you > > > > > > Regards > > > > > > -- > > Hervé >