Hello, I have a table as below
CREATE TABLE analytics_db.ml_forecast_tbl ( "MetricID" int, "TimeStamp" timestamp, "ResourceID" timeuuid "Value" double, PRIMARY KEY ("MetricID", "TimeStamp", "ResourceID") ) select * from ml_forecast_tbl where "MetricID" = 1 and "TimeStamp" > '2016-01-22 00:00:00+0000' and "TimeStamp" <= '2016-01-22 00:30:00+0000' ; MetricID | TimeStamp | ResourceID | Value| ----------+---------------------------------+--------------------------------------+----------+ | 1 | 2016-01-22 00:30:00.000000+0000 | 4a925190-3b13-11e7-83c6-a32261219d03 | 32.16177 | | 23.74124 | 15.2371 1 | 2016-01-22 00:30:00.000000+0000 | 4a92c6c0-3b13-11e7-83c6-a32261219d03 | 32.16177 | | 23.74124 | 15.2371 1 | 2016-01-22 00:30:00.000000+0000 | 4a936300-3b13-11e7-83c6-a32261219d03 | 32.16177 | | 23.74124 | 15.2371 1 | 2016-01-22 00:30:00.000000+0000 | 4a93d830-3b13-11e7-83c6-a32261219d03 | 32.16177 | | 23.74124 | 15.2371 This query runs perfectly fine from cqlsh, but not with Spark SQL, it just emits empty results, Is there a catch to think about on querying timestamp ranges with cassandra spark connector Any inputs on this ?.. Thanks, Sujeet