Hey everyone, I'm trying to use the Python phoenixdb adapter work with JSON serialization on PQS.
I'm using Phoenix 4.14 and the adapter works fine with protobuf, but when I try making it work with an older version of phoenixdb (before the JSON to protobuf switch was introduced), it just returns 0 rows. I don't see anything in particular wrong with the HTTP requests itself, and they seem to conform to the Avatica JSON spec ( http://calcite.apache.org/avatica/docs/json_reference.html). Here's the result (with some debug statements) that returns 0 rows. Notice the *"firstFrame":{"offset":0,"done":true,"rows":[]* below: request body = {"maxRowCount": -2, "connectionId": "68c05d12-5770-47d6-b3e4-dba556db4790", "request": "prepareAndExecute", "statementId": 3, "sql": "SELECT col1, col2 from table limit 20"} request headers = {'content-type': 'application/json'} _post_request: got response {'fp': <socket._fileobject object at 0x7f858330b9d0>, 'status': 200, 'will_close': False, 'chunk_left': 'UNKNOWN', 'length': 1395, 'strict': 0, 'reason': 'OK', 'version': 11, 'debuglevel': 0, 'msg': <httplib.HTTPMessage instance at 0x7f84fb50be18>, 'chunked': 0, '_method': 'POST'} response.read(): body = {"response":"executeResults","missingStatement":false,"rpcMetadata":{"response":"rpcMetadata","serverAddress":"ip-10-55-6-247:8765"},"results":[{"response":"resultSet","connectionId":"68c05d12-5770-47d6-b3e4-dba556db4790","statementId":3,"ownStatement":true,"signature":{"columns":[{"ordinal":0,"autoIncrement":false,"caseSensitive":false,"searchable":true,"currency":false,"nullable ":0,"signed":true,"displaySize":40,"label":"COL1","columnName":"COL1","schemaName":"","precision":0,"scale":0,"tableName":"TABLE","catalogName":"","type":{"type":"scalar","id":4,"name":"INTEGER","rep":"PRIMITIVE_INT"},"readOnly":true,"writable":false,"definitelyWritable":false,"columnClassName":"java.lang.Integer"},{"ordinal":1,"autoIncrement":false,"caseSensitive":false,"searchable":true,"currency":false,"nullable":0,"signed":true,"displaySize":40,"label":"COL2","columnName":"COL2","schemaName":"","precision":0,"scale":0,"tableName":"TABLE","catalogName":"","type":{"type":"scalar","id":4,"name":"INTEGER","rep":"PRIMITIVE_INT"},"readOnly":true,"writable":false,"definitelyWritable":false,"columnClassName":"java.lang.Integer"}],"sql":null,"parameters":[],"cursorFactory":{"style":"LIST","clazz":null,"fieldNames":null},"statementType":null}, *"firstFrame":{"offset":0,"done":true,"rows":[]* },"updateCount":-1,"rpcMetadata":{"response":"rpcMetadata","serverAddress":"ip-10-55-6-247:8765"}}]} The same query issued against a PQS started with PROTOBUF serialization and using a newer phoenixdb adapter returns the correct number of rows. Has anyone had luck making this work? Thanks, Manoj