Hey, I am having a parquet dir and a table mounted on it. the table is showing sample view , via hue fine but a simple query like select * from tablename gives this error :
- Bad status for request TFetchResultsReq(fetchType=0, operationHandle=TOperationHandle(hasResultSet=True, modifiedRowCount=None, operationType=0, operationId=THandleIdentifier(secret='\xf7\xe7\x90\x0e\x85\x91E{\x99\xd1\xdf>v\xf7\x8c`', guid='\xcc\xd6$^\xac{M\xaf\x9c{\xc2\xcf\xf3\xc6\xe7/')), orientation=4, maxRows=100): TFetchResultsResp(status=TStatus(errorCode=0, errorMessage='java.io.IOException: parquet.io.ParquetDecodingException: Can not read value at 0 in block -1 in file hdfs://askmehadoop/parquet1_mpdm_mpdm_store/partitioned_on_seller_mailer_flag=1/part-r-00000-a77c308f-c088-4f41-ab07-0c8e0557dbe1.gz.parquet', sqlState=None, infoMessages=['*org.apache.hive.service.cli.HiveSQLException:java.io.IOException: parquet.io.ParquetDecodingException: Can not read value at 0 in block -1 in file hdfs://askmehadoop/parquet1_mpdm_mpdm_store/partitioned_on_seller_mailer_flag=1/part-r-00000-a77c308f-c088-4f41-ab07-0c8e0557dbe1.gz.parquet:25:24', 'org.apache.hive.service.cli.operation.SQLOperation:getNextRowSet:SQLOperation.java:352', 'org.apache.hive.service.cli.operation.OperationManager:getOperationNextRowSet:OperationManager.java:220', 'org.apache.hive.service.cli.session.HiveSessionImpl:fetchResults:HiveSessionImpl.java:685', 'sun.reflect.GeneratedMethodAccessor63:invoke::-1', 'sun.reflect.DelegatingMethodAccessorImpl:invoke:DelegatingMethodAccessorImpl.java:43', 'java.lang.reflect.Method:invoke:Method.java:498', 'org.apache.hive.service.cli.session.HiveSessionProxy:invoke:HiveSessionProxy.java:78', 'org.apache.hive.service.cli.session.HiveSessionProxy:access$000:HiveSessionProxy.java:36', 'org.apache.hive.service.cli.session.HiveSessionProxy$1:run:HiveSessionProxy.java:63', 'java.security.AccessController:doPrivileged:AccessController.java:-2', 'javax.security.auth.Subject:doAs:Subject.java:422', 'org.apache.hadoop.security.UserGroupInformation:doAs:UserGroupInformation.java:1657', 'org.apache.hive.service.cli.session.HiveSessionProxy:invoke:HiveSessionProxy.java:59', 'com.sun.proxy.$Proxy22:fetchResults::-1', 'org.apache.hive.service.cli.CLIService:fetchResults:CLIService.java:454', 'org.apache.hive.service.cli.thrift.ThriftCLIService:FetchResults:ThriftCLIService.java:672', 'org.apache.hive.service.cli.thrift.TCLIService$Processor$FetchResults:getResult:TCLIService.java:1553', 'org.apache.hive.service.cli.thrift.TCLIService$Processor$FetchResults:getResult:TCLIService.java:1538', 'org.apache.thrift.ProcessFunction:process:ProcessFunction.java:39', 'org.apache.thrift.TBaseProcessor:process:TBaseProcessor.java:39', 'org.apache.hive.service.auth.TSetIpAddressProcessor:process:TSetIpAddressProcessor.java:56', 'org.apache.thrift.server.TThreadPoolServer$WorkerProcess:run:TThreadPoolServer.java:285', 'java.util.concurrent.ThreadPoolExecutor:runWorker:ThreadPoolExecutor.java:1142', 'java.util.concurrent.ThreadPoolExecutor$Worker:run:ThreadPoolExecutor.java:617', 'java.lang.Thread:run:Thread.java:745', '*java.io.IOException:parquet.io.ParquetDecodingException: Can not read value at 0 in block -1 in file hdfs://askmehadoop/parquet1_mpdm_mpdm_store/partitioned_on_seller_mailer_flag=1/part-r-00000-a77c308f-c088-4f41-ab07-0c8e0557dbe1.gz.parquet:29:4', 'org.apache.hadoop.hive.ql.exec.FetchOperator:getNextRow:FetchOperator.java:507', 'org.apache.hadoop.hive.ql.exec.FetchOperator:pushRow:FetchOperator.java:414', 'org.apache.hadoop.hive.ql.exec.FetchTask:fetch:FetchTask.java:140', 'org.apache.hadoop.hive.ql.Driver:getResults:Driver.java:1670', 'org.apache.hive.service.cli.operation.SQLOperation:getNextRowSet:SQLOperation.java:347', '*parquet.io.ParquetDecodingException:Can not read value at 0 in block -1 in file hdfs://askmehadoop/parquet1_mpdm_mpdm_store/partitioned_on_seller_mailer_flag=1/part-r-00000-a77c308f-c088-4f41-ab07-0c8e0557dbe1.gz.parquet:36:7', 'parquet.hadoop.InternalParquetRecordReader:nextKeyValue:InternalParquetRecordReader.java:228', 'parquet.hadoop.ParquetRecordReader:nextKeyValue:ParquetRecordReader.java:201', 'org.apache.hadoop.hive.ql.io.parquet.read.ParquetRecordReaderWrapper:<init>:ParquetRecordReaderWrapper.java:122', 'org.apache.hadoop.hive.ql.io.parquet.read.ParquetRecordReaderWrapper:<init>:ParquetRecordReaderWrapper.java:85', 'org.apache.hadoop.hive.ql.io.parquet.MapredParquetInputFormat:getRecordReader:MapredParquetInputFormat.java:72', 'org.apache.hadoop.hive.ql.exec.FetchOperator$FetchInputFormatSplit:getRecordReader:FetchOperator.java:673', 'org.apache.hadoop.hive.ql.exec.FetchOperator:getRecordReader:FetchOperator.java:323', 'org.apache.hadoop.hive.ql.exec.FetchOperator:getNextRow:FetchOperator.java:445', '*java.lang.UnsupportedOperationException:parquet.column.values.dictionary.PlainValuesDictionary$PlainLongDictionary:47:11', 'parquet.column.Dictionary:decodeToBinary:Dictionary.java:44', 'org.apache.hadoop.hive.ql.io.parquet.convert.ETypeConverter$BinaryConverter:setDictionary:ETypeConverter.java:227', 'parquet.column.impl.ColumnReaderImpl:<init>:ColumnReaderImpl.java:339', 'parquet.column.impl.ColumnReadStoreImpl:newMemColumnReader:ColumnReadStoreImpl.java:66', 'parquet.column.impl.ColumnReadStoreImpl:getColumnReader:ColumnReadStoreImpl.java:61', 'parquet.io.RecordReaderImplementation:<init>:RecordReaderImplementation.java:270', 'parquet.io.MessageColumnIO$1:visit:MessageColumnIO.java:134', 'parquet.io.MessageColumnIO$1:visit:MessageColumnIO.java:99', 'parquet.filter2.compat.FilterCompat$NoOpFilter:accept:FilterCompat.java:154', 'parquet.io.MessageColumnIO:getRecordReader:MessageColumnIO.java:99', 'parquet.hadoop.InternalParquetRecordReader:checkRead:InternalParquetRecordReader.java:137', 'parquet.hadoop.InternalParquetRecordReader:nextKeyValue:InternalParquetRecordReader.java:208'], statusCode=3), results=None, hasMoreRows=None) ls that something to do with hive? or a parquet error as such ? I have posted it in both these groups, but I am afraid it mustnt be parquet because data is showing fine (viewed it in hue). Has anyone experienced similar errors before? Kindly let me know.. Awaiting Your Reply, Thanks Sumit