this is what I get from the trace log (repeated twice for each insert):
INFO  [SharedPool-Worker-1] 2017-05-26 23:10:12,587  SLF4JAuditWriter.java:96 - 
host:/*.*.*.*|source:/*.*.*.*|user:*****|timestamp:1495840212587|category:DML|type:CQL_UPDATE|ks:myks|cf:mytab|operation:INSERT
 INTO myks.mytab (id, day, month, email, e_at, fn, ln, last_upd_dt, pc, pn, 
ptype, rs, sm) 
VALUES
('TSTUSERID', 4, 6, 'tstemail_lt...@crmqa.com', 1495840211904, 
'tstFirst_LVMF4', 'tstLast_PVAKZ', 1495840212584, null, null, [('5208096456', 
'5208096456', 'Registered'), ('11230103039', '11230103039', 'Registered'), 
('4649861872074', 'Garbage Phone number', 'Unknown')], 'Web', 'US') USING 
TIMESTAMP 1495840211904000;|consistency level:LOCAL_QUORUM 

On Saturday, May 27, 2017, 5:58:21 PM PDT, Jeff Jirsa <jji...@gmail.com> 
wrote:Can you upload the trace somewhere ?



-- 
Jeff Jirsa


> On May 27, 2017, at 5:34 PM, Subroto Barua <sbarua...@yahoo.com.INVALID> 
> wrote:
> 
> 
> i have a table with one of the column as Tuple datatype.
> 
> when I insert rows with "using timestamp", I see multiple inserts in the 
> trace log and I see multiple data sets per partition key.
> 
> 
> 
> 
> for eg: for each partition key, I get extra tuple for each insert as opposed 
> to 1 row for the key, which keeps growing with new insert for the same key.
> 
> this issue happens when I insert via api using Phantom driver; works fine via 
> cqlsh.
> not sure whether there is any limitation with "using timestamp" clause or a 
> bug with Phantom driver.
> appreciate any advice/comments on Tuple or "using timestamp"...
> 

---------------------------------------------------------------------
To unsubscribe, e-mail: dev-unsubscr...@cassandra.apache.org
For additional commands, e-mail: dev-h...@cassandra.apache.org

Reply via email to