[ https://issues.apache.org/jira/browse/STORM-2841?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Jungtaek Lim resolved STORM-2841. --------------------------------- Resolution: Fixed Fix Version/s: 2.0.0 1.0.7 1.1.3 1.2.1 Merged into master, 1.x, 1.1.x, 1.0.x branches. > testNoAcksIfFlushFails UT fails with NullPointerException > --------------------------------------------------------- > > Key: STORM-2841 > URL: https://issues.apache.org/jira/browse/STORM-2841 > Project: Apache Storm > Issue Type: Test > Reporter: Yesha Vora > Assignee: Jungtaek Lim > Priority: Major > Labels: pull-request-available > Fix For: 1.2.1, 1.1.3, 1.0.7, 2.0.0 > > Time Spent: 0.5h > Remaining Estimate: 0h > > testNoAcksIfFlushFails UT fails with NPE > {code} > java.lang.NullPointerException: null > at > org.apache.storm.hive.bolt.HiveBolt.flushAllWriters(HiveBolt.java:207) > at > org.apache.storm.hive.bolt.TestHiveBolt.testNoAcksIfFlushFails(TestHiveBolt.java:322){code} > {code} > Standard Output > 60488 [main] INFO h.metastore - Mestastore configuration > hive.metastore.filter.hook changed from > org.apache.hadoop.hive.ql.security.authorization.plugin.AuthorizationMetaStoreFilterHook > to org.apache.hadoop.hive.metastore.DefaultMetaStoreFilterHookImpl > 60489 [main] INFO o.a.h.h.m.HiveMetaStore - 0: Shutting down the object > store... > 60489 [main] INFO o.a.h.h.m.H.audit - ugi=nobody ip=unknown-ip-addr > cmd=Shutting down the object store... > 60489 [main] INFO o.a.h.h.m.HiveMetaStore - 0: Metastore shutdown complete. > 60489 [main] INFO o.a.h.h.m.H.audit - ugi=nobody ip=unknown-ip-addr > cmd=Metastore shutdown complete. > 60497 [main] INFO o.a.h.h.q.s.SessionState - Created local directory: > /tmp/75acc3af-f90e-4ab2-888c-a8ad29383128_resources > 60499 [main] INFO o.a.h.h.q.s.SessionState - Created HDFS directory: > /tmp/hive/nobody/75acc3af-f90e-4ab2-888c-a8ad29383128 > 60501 [main] INFO o.a.h.h.q.s.SessionState - Created local directory: > /tmp/nobody/75acc3af-f90e-4ab2-888c-a8ad29383128 > 60503 [main] INFO o.a.h.h.q.s.SessionState - Created HDFS directory: > /tmp/hive/nobody/75acc3af-f90e-4ab2-888c-a8ad29383128/_tmp_space.db > 60793 [main] INFO o.a.h.h.m.HiveMetaStore - 0: get_table_names_by_filter: db > = testdb, filter = > 60793 [main] INFO o.a.h.h.m.H.audit - ugi=nobody ip=unknown-ip-addr > cmd=get_table_names_by_filter: db = testdb, filter = > 60847 [main] INFO o.a.h.h.m.HiveMetaStore - 0: Opening raw store with > implementation class:org.apache.hadoop.hive.metastore.ObjectStore > 60848 [main] INFO o.a.h.h.m.ObjectStore - ObjectStore, initialize called > 60864 [main] INFO D.Persistence - Property > hive.metastore.integral.jdo.pushdown unknown - will be ignored > 60864 [main] INFO D.Persistence - Property datanucleus.cache.level2 unknown > - will be ignored > 61370 [main] INFO o.a.h.h.m.ObjectStore - Setting MetaStore object pin > classes with > hive.metastore.cache.pinobjtypes="Table,StorageDescriptor,SerDeInfo,Partition,Database,Type,FieldSchema,Order" > 61711 [main] INFO o.a.h.h.m.MetaStoreDirectSql - Using direct SQL, > underlying DB is DERBY > 61711 [main] INFO o.a.h.h.m.ObjectStore - Initialized ObjectStore > 61716 [main] INFO o.a.h.h.m.HiveMetaStore - 0: get_table : db=testdb > tbl=test_table > 61716 [main] INFO o.a.h.h.m.H.audit - ugi=nobody ip=unknown-ip-addr > cmd=get_table : db=testdb tbl=test_table > 61738 [main] INFO o.a.h.h.m.HiveMetaStore - 0: drop_table : db=testdb > tbl=test_table > 61738 [main] INFO o.a.h.h.m.H.audit - ugi=nobody ip=unknown-ip-addr > cmd=drop_table : db=testdb tbl=test_table > 62053 [main] INFO h.m.hivemetastoressimpl - deleting > raw:/tmp/junit2860976527054408251/testdb.db/test_table > 62062 [main] INFO o.a.h.h.m.HiveMetaStore - 0: get_database: testdb > 62063 [main] INFO o.a.h.h.m.H.audit - ugi=nobody ip=unknown-ip-addr > cmd=get_database: testdb > 62068 [main] INFO o.a.h.h.m.HiveMetaStore - 0: drop_database: testdb > 62068 [main] INFO o.a.h.h.m.H.audit - ugi=nobody ip=unknown-ip-addr > cmd=drop_database: testdb > 62069 [main] INFO o.a.h.h.m.HiveMetaStore - 0: get_all_tables: db=testdb > 62069 [main] INFO o.a.h.h.m.H.audit - ugi=nobody ip=unknown-ip-addr > cmd=get_all_tables: db=testdb > 62073 [main] INFO o.a.h.h.m.HiveMetaStore - 0: get_functions: db=testdb pat=* > 62073 [main] INFO o.a.h.h.m.H.audit - ugi=nobody ip=unknown-ip-addr > cmd=get_functions: db=testdb pat=* > 62103 [main] INFO o.a.h.h.m.ObjectStore - Dropping database testdb along > with all tables > 62141 [main] INFO h.m.hivemetastoressimpl - deleting > raw:/tmp/junit2860976527054408251/testdb.db > 62148 [main] INFO o.a.h.h.q.l.PerfLogger - <PERFLOG method=Driver.run > from=org.apache.hadoop.hive.ql.Driver> > 62148 [main] INFO o.a.h.h.q.l.PerfLogger - <PERFLOG method=compile > from=org.apache.hadoop.hive.ql.Driver> > 62149 [main] INFO o.a.h.h.q.Driver - We are setting the hadoop caller > context from to nobody_20171128150454_9a730ff0-d279-4037-a1c9-4b3c28ff0a90 > 62149 [main] INFO o.a.h.h.q.l.PerfLogger - <PERFLOG method=parse > from=org.apache.hadoop.hive.ql.Driver> > 62149 [main] INFO h.q.p.ParseDriver - Parsing command: create database IF > NOT EXISTS testdb location 'raw:///tmp/junit8935893841909704896/testdb.db' > 62149 [main] INFO h.q.p.ParseDriver - Parse Completed > 62149 [main] INFO o.a.h.h.q.l.PerfLogger - </PERFLOG method=parse > start=1511881494579 end=1511881494579 duration=0 > from=org.apache.hadoop.hive.ql.Driver> > 62149 [main] INFO o.a.h.h.q.l.PerfLogger - <PERFLOG method=semanticAnalyze > from=org.apache.hadoop.hive.ql.Driver> > 62151 [main] INFO o.a.h.h.q.Driver - Semantic Analysis Completed > 62151 [main] INFO o.a.h.h.q.l.PerfLogger - </PERFLOG method=semanticAnalyze > start=1511881494579 end=1511881494581 duration=2 > from=org.apache.hadoop.hive.ql.Driver> > 62151 [main] INFO o.a.h.h.q.Driver - Returning Hive schema: > Schema(fieldSchemas:null, properties:null) > 62151 [main] INFO o.a.h.h.q.l.PerfLogger - </PERFLOG method=compile > start=1511881494578 end=1511881494581 duration=3 > from=org.apache.hadoop.hive.ql.Driver> > 62151 [main] INFO o.a.h.h.q.Driver - We are resetting the hadoop caller > context to > 62151 [main] INFO o.a.h.h.q.l.PerfLogger - <PERFLOG > method=acquireReadWriteLocks from=org.apache.hadoop.hive.ql.Driver> > 62151 [main] INFO o.a.h.h.q.l.DbTxnManager - Setting lock request > transaction to txnid:0 for > queryId=nobody_20171128150454_9a730ff0-d279-4037-a1c9-4b3c28ff0a90 > 62151 [main] INFO o.a.h.h.q.l.PerfLogger - </PERFLOG > method=acquireReadWriteLocks start=1511881494581 end=1511881494581 duration=0 > from=org.apache.hadoop.hive.ql.Driver> > 62151 [main] INFO o.a.h.h.q.l.PerfLogger - <PERFLOG method=Driver.execute > from=org.apache.hadoop.hive.ql.Driver> > 62152 [main] INFO o.a.h.h.q.Driver - Setting caller context to query id > nobody_20171128150454_9a730ff0-d279-4037-a1c9-4b3c28ff0a90 > 62152 [main] INFO o.a.h.h.q.Driver - Starting > command(queryId=nobody_20171128150454_9a730ff0-d279-4037-a1c9-4b3c28ff0a90): > create database IF NOT EXISTS testdb location > 'raw:///tmp/junit8935893841909704896/testdb.db' > 62152 [main] INFO o.a.h.h.q.s.a.p.s.SQLStdHiveAccessController - Created > SQLStdHiveAccessController for session context : HiveAuthzSessionContext > [sessionString=75acc3af-f90e-4ab2-888c-a8ad29383128, clientType=HIVECLI] > 62152 [main] INFO h.metastore - Mestastore configuration > hive.metastore.filter.hook changed from > org.apache.hadoop.hive.metastore.DefaultMetaStoreFilterHookImpl to > org.apache.hadoop.hive.ql.security.authorization.plugin.AuthorizationMetaStoreFilterHook > 62152 [main] INFO o.a.h.h.m.HiveMetaStore - 0: Shutting down the object > store... > 62152 [main] INFO o.a.h.h.m.H.audit - ugi=nobody ip=unknown-ip-addr > cmd=Shutting down the object store... > 62153 [main] INFO o.a.h.h.m.HiveMetaStore - 0: Metastore shutdown complete. > 62153 [main] INFO o.a.h.h.m.H.audit - ugi=nobody ip=unknown-ip-addr > cmd=Metastore shutdown complete. > 62153 [main] INFO o.a.h.h.q.l.PerfLogger - <PERFLOG method=runTasks > from=org.apache.hadoop.hive.ql.Driver> > 62153 [main] INFO o.a.h.h.q.Driver - Starting task [Stage-0:DDL] in serial > mode > 62153 [main] INFO o.a.h.h.m.HiveMetaStore - 0: create_database: > Database(name:testdb, description:null, > locationUri:raw:/tmp/junit8935893841909704896/testdb.db, parameters:null, > ownerName:nobody, ownerType:USER) > 62154 [main] INFO o.a.h.h.m.H.audit - ugi=nobody ip=unknown-ip-addr > cmd=create_database: Database(name:testdb, description:null, > locationUri:raw:/tmp/junit8935893841909704896/testdb.db, parameters:null, > ownerName:nobody, ownerType:USER) > 62199 [main] WARN o.a.h.h.c.HiveConf - HiveConf of name > hive.internal.ss.authz.settings.applied.marker does not exist > 62200 [main] INFO o.a.h.h.m.HiveMetaStore - 0: Opening raw store with > implementation class:org.apache.hadoop.hive.metastore.ObjectStore > 62201 [main] INFO o.a.h.h.m.ObjectStore - ObjectStore, initialize called > 62244 [main] WARN o.a.h.h.c.HiveConf - HiveConf of name > hive.internal.ss.authz.settings.applied.marker does not exist > 62246 [main] INFO o.a.h.h.m.MetaStoreDirectSql - Using direct SQL, > underlying DB is DERBY > 62246 [main] INFO o.a.h.h.m.ObjectStore - Initialized ObjectStore > 62247 [main] WARN o.a.h.h.m.ObjectStore - Failed to get database testdb, > returning NoSuchObjectException > 62247 [main] INFO o.a.h.h.m.HiveMetaStore - create_database_core testdb > 62247 [main] INFO o.a.h.h.m.HiveMetaStore - create_database_core preEvent > testdb > 62247 [main] INFO o.a.h.h.m.HiveMetaStore - create_database_core filesystem > testdb > 62247 [main] INFO o.a.h.h.m.HiveMetaStore - create_database_core rdbms testdb > 62256 [main] INFO o.a.h.h.m.HiveMetaStore - create_database_core rdbms > committed testdb > 62257 [main] INFO o.a.h.h.m.HiveMetaStore - create_database_core listeners > testdb > 62257 [main] INFO o.a.h.h.m.HiveMetaStore - create_database_core listeners > done testdb > 62257 [main] INFO o.a.h.h.q.l.PerfLogger - </PERFLOG method=runTasks > start=1511881494583 end=1511881494687 duration=104 > from=org.apache.hadoop.hive.ql.Driver> > 62257 [main] INFO o.a.h.h.q.Driver - Resetting the caller context to > 62257 [main] INFO o.a.h.h.q.l.PerfLogger - </PERFLOG method=Driver.execute > start=1511881494581 end=1511881494687 duration=106 > from=org.apache.hadoop.hive.ql.Driver> > 62257 [main] INFO o.a.h.h.q.Driver - OK > 62257 [main] INFO o.a.h.h.q.l.PerfLogger - <PERFLOG method=releaseLocks > from=org.apache.hadoop.hive.ql.Driver> > 62257 [main] INFO o.a.h.h.q.l.PerfLogger - </PERFLOG method=releaseLocks > start=1511881494687 end=1511881494687 duration=0 > from=org.apache.hadoop.hive.ql.Driver> > 62257 [main] INFO o.a.h.h.q.l.PerfLogger - </PERFLOG method=Driver.run > start=1511881494578 end=1511881494687 duration=109 > from=org.apache.hadoop.hive.ql.Driver> > 62257 [main] INFO o.a.h.h.q.l.PerfLogger - <PERFLOG method=Driver.run > from=org.apache.hadoop.hive.ql.Driver> > 62257 [main] INFO o.a.h.h.q.l.PerfLogger - <PERFLOG method=compile > from=org.apache.hadoop.hive.ql.Driver> > 62257 [main] INFO o.a.h.h.q.Driver - We are setting the hadoop caller > context from to nobody_20171128150454_8b6fa569-5a71-41c5-b207-df78c7b738c2 > 62257 [main] INFO o.a.h.h.q.l.PerfLogger - <PERFLOG method=parse > from=org.apache.hadoop.hive.ql.Driver> > 62258 [main] INFO h.q.p.ParseDriver - Parsing command: use testdb > 62258 [main] INFO h.q.p.ParseDriver - Parse Completed > 62258 [main] INFO o.a.h.h.q.l.PerfLogger - </PERFLOG method=parse > start=1511881494687 end=1511881494688 duration=1 > from=org.apache.hadoop.hive.ql.Driver> > 62258 [main] INFO o.a.h.h.q.l.PerfLogger - <PERFLOG method=semanticAnalyze > from=org.apache.hadoop.hive.ql.Driver> > 62258 [main] INFO o.a.h.h.m.HiveMetaStore - 0: get_database: testdb > 62258 [main] INFO o.a.h.h.m.H.audit - ugi=nobody ip=unknown-ip-addr > cmd=get_database: testdb > 62259 [main] INFO o.a.h.h.q.Driver - Semantic Analysis Completed > 62259 [main] INFO o.a.h.h.q.l.PerfLogger - </PERFLOG method=semanticAnalyze > start=1511881494688 end=1511881494689 duration=1 > from=org.apache.hadoop.hive.ql.Driver> > 62259 [main] INFO o.a.h.h.q.Driver - Returning Hive schema: > Schema(fieldSchemas:null, properties:null) > 62259 [main] INFO o.a.h.h.q.l.PerfLogger - </PERFLOG method=compile > start=1511881494687 end=1511881494689 duration=2 > from=org.apache.hadoop.hive.ql.Driver> > 62259 [main] INFO o.a.h.h.q.Driver - We are resetting the hadoop caller > context to > 62259 [main] INFO o.a.h.h.q.l.PerfLogger - <PERFLOG > method=acquireReadWriteLocks from=org.apache.hadoop.hive.ql.Driver> > 62259 [main] INFO o.a.h.h.q.l.DbTxnManager - Setting lock request > transaction to txnid:0 for > queryId=nobody_20171128150454_8b6fa569-5a71-41c5-b207-df78c7b738c2 > 62259 [main] INFO o.a.h.h.q.l.PerfLogger - </PERFLOG > method=acquireReadWriteLocks start=1511881494689 end=1511881494689 duration=0 > from=org.apache.hadoop.hive.ql.Driver> > 62259 [main] INFO o.a.h.h.q.l.PerfLogger - <PERFLOG method=Driver.execute > from=org.apache.hadoop.hive.ql.Driver> > 62259 [main] INFO o.a.h.h.q.Driver - Setting caller context to query id > nobody_20171128150454_8b6fa569-5a71-41c5-b207-df78c7b738c2 > 62259 [main] INFO o.a.h.h.q.Driver - Starting > command(queryId=nobody_20171128150454_8b6fa569-5a71-41c5-b207-df78c7b738c2): > create database IF NOT EXISTS testdb location > 'raw:///tmp/junit8935893841909704896/testdb.db' > 62259 [main] INFO o.a.h.h.q.l.PerfLogger - <PERFLOG method=runTasks > from=org.apache.hadoop.hive.ql.Driver> > 62259 [main] INFO o.a.h.h.q.Driver - Starting task [Stage-0:DDL] in serial > mode > 62260 [main] INFO o.a.h.h.m.HiveMetaStore - 0: get_database: testdb > 62260 [main] INFO o.a.h.h.m.H.audit - ugi=nobody ip=unknown-ip-addr > cmd=get_database: testdb > 62260 [main] INFO o.a.h.h.m.HiveMetaStore - 0: get_database: testdb > 62260 [main] INFO o.a.h.h.m.H.audit - ugi=nobody ip=unknown-ip-addr > cmd=get_database: testdb > 62261 [main] INFO o.a.h.h.q.l.PerfLogger - </PERFLOG method=runTasks > start=1511881494689 end=1511881494691 duration=2 > from=org.apache.hadoop.hive.ql.Driver> > 62261 [main] INFO o.a.h.h.q.Driver - Resetting the caller context to > 62261 [main] INFO o.a.h.h.q.l.PerfLogger - </PERFLOG method=Driver.execute > start=1511881494689 end=1511881494691 duration=2 > from=org.apache.hadoop.hive.ql.Driver> > 62261 [main] INFO o.a.h.h.q.Driver - OK > 62261 [main] INFO o.a.h.h.q.l.PerfLogger - <PERFLOG method=releaseLocks > from=org.apache.hadoop.hive.ql.Driver> > 62261 [main] INFO o.a.h.h.q.l.PerfLogger - </PERFLOG method=releaseLocks > start=1511881494691 end=1511881494691 duration=0 > from=org.apache.hadoop.hive.ql.Driver> > 62261 [main] INFO o.a.h.h.q.l.PerfLogger - </PERFLOG method=Driver.run > start=1511881494687 end=1511881494691 duration=4 > from=org.apache.hadoop.hive.ql.Driver> > 62261 [main] INFO o.a.h.h.q.l.PerfLogger - <PERFLOG method=Driver.run > from=org.apache.hadoop.hive.ql.Driver> > 62261 [main] INFO o.a.h.h.q.l.PerfLogger - <PERFLOG method=compile > from=org.apache.hadoop.hive.ql.Driver> > 62261 [main] INFO o.a.h.h.q.Driver - We are setting the hadoop caller > context from to nobody_20171128150454_f8e917c9-4cbc-4512-b93a-65db6c890fa3 > 62261 [main] INFO o.a.h.h.q.l.PerfLogger - <PERFLOG method=parse > from=org.apache.hadoop.hive.ql.Driver> > 62261 [main] INFO h.q.p.ParseDriver - Parsing command: create table > test_table ( id int,msg string ) partitioned by (city string,state string ) > clustered by ( id ) into 10 buckets stored as orc location > 'raw:///tmp/junit8935893841909704896/testdb.db/test_table' TBLPROPERTIES > ('transactional'='true') > 62262 [main] INFO h.q.p.ParseDriver - Parse Completed > 62262 [main] INFO o.a.h.h.q.l.PerfLogger - </PERFLOG method=parse > start=1511881494691 end=1511881494692 duration=1 > from=org.apache.hadoop.hive.ql.Driver> > 62262 [main] INFO o.a.h.h.q.l.PerfLogger - <PERFLOG method=semanticAnalyze > from=org.apache.hadoop.hive.ql.Driver> > 62262 [main] INFO o.a.h.h.q.p.CalcitePlanner - Starting Semantic Analysis > 62262 [main] INFO o.a.h.h.q.p.CalcitePlanner - Creating table > testdb.test_table position=13 > 62263 [main] INFO o.a.h.h.m.HiveMetaStore - 0: get_database: testdb > 62263 [main] INFO o.a.h.h.m.H.audit - ugi=nobody ip=unknown-ip-addr > cmd=get_database: testdb > 62263 [main] INFO o.a.h.h.q.Driver - Semantic Analysis Completed > 62263 [main] INFO o.a.h.h.q.l.PerfLogger - </PERFLOG method=semanticAnalyze > start=1511881494692 end=1511881494693 duration=1 > from=org.apache.hadoop.hive.ql.Driver> > 62263 [main] INFO o.a.h.h.q.Driver - Returning Hive schema: > Schema(fieldSchemas:null, properties:null) > 62263 [main] INFO o.a.h.h.q.l.PerfLogger - </PERFLOG method=compile > start=1511881494691 end=1511881494693 duration=2 > from=org.apache.hadoop.hive.ql.Driver> > 62264 [main] INFO o.a.h.h.q.Driver - We are resetting the hadoop caller > context to > 62264 [main] INFO o.a.h.h.q.l.PerfLogger - <PERFLOG > method=acquireReadWriteLocks from=org.apache.hadoop.hive.ql.Driver> > 62264 [main] INFO o.a.h.h.q.l.DbTxnManager - Setting lock request > transaction to txnid:0 for > queryId=nobody_20171128150454_f8e917c9-4cbc-4512-b93a-65db6c890fa3 > 62264 [main] INFO o.a.h.h.q.l.DbLockManager - Requesting: > queryId=nobody_20171128150454_f8e917c9-4cbc-4512-b93a-65db6c890fa3 > LockRequest(component:[LockComponent(type:SHARED_READ, level:DB, > dbname:testdb, operationType:NO_TXN, isDynamicPartitionWrite:false)], > txnid:0, user:nobody, > hostname:ctr-e134-1499953498516-345859-01-000003.hwx.site, > agentInfo:nobody_20171128150454_f8e917c9-4cbc-4512-b93a-65db6c890fa3) > 62283 [main] INFO o.a.h.h.q.l.DbLockManager - Response to > queryId=nobody_20171128150454_f8e917c9-4cbc-4512-b93a-65db6c890fa3 > LockResponse(lockid:1, state:ACQUIRED) > 62283 [main] INFO o.a.h.h.q.l.DbTxnManager - Started heartbeat with > delay/interval = 150000/150000 MILLISECONDS for query: > nobody_20171128150454_f8e917c9-4cbc-4512-b93a-65db6c890fa3 > 62283 [main] INFO o.a.h.h.q.l.PerfLogger - </PERFLOG > method=acquireReadWriteLocks start=1511881494694 end=1511881494713 > duration=19 from=org.apache.hadoop.hive.ql.Driver> > 62283 [main] INFO o.a.h.h.q.l.PerfLogger - <PERFLOG method=Driver.execute > from=org.apache.hadoop.hive.ql.Driver> > 62283 [main] INFO o.a.h.h.q.Driver - Setting caller context to query id > nobody_20171128150454_f8e917c9-4cbc-4512-b93a-65db6c890fa3 > 62283 [main] INFO o.a.h.h.q.Driver - Starting > command(queryId=nobody_20171128150454_f8e917c9-4cbc-4512-b93a-65db6c890fa3): > create database IF NOT EXISTS testdb location > 'raw:///tmp/junit8935893841909704896/testdb.db' > 62283 [main] INFO o.a.h.h.q.l.PerfLogger - <PERFLOG method=runTasks > from=org.apache.hadoop.hive.ql.Driver> > 62283 [main] INFO o.a.h.h.q.Driver - Starting task [Stage-0:DDL] in serial > mode > 62285 [main] INFO o.a.h.h.m.HiveMetaStore - 0: create_table: > Table(tableName:test_table, dbName:testdb, owner:nobody, > createTime:1511881494, lastAccessTime:0, retention:0, > sd:StorageDescriptor(cols:[FieldSchema(name:id, type:int, comment:null), > FieldSchema(name:msg, type:string, comment:null)], > location:raw:/tmp/junit8935893841909704896/testdb.db/test_table, > inputFormat:org.apache.hadoop.hive.ql.io.orc.OrcInputFormat, > outputFormat:org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat, > compressed:false, numBuckets:10, serdeInfo:SerDeInfo(name:null, > serializationLib:org.apache.hadoop.hive.ql.io.orc.OrcSerde, > parameters:{serialization.format=1}), bucketCols:[id], sortCols:[], > parameters:{}, skewedInfo:SkewedInfo(skewedColNames:[], skewedColValues:[], > skewedColValueLocationMaps:{}), storedAsSubDirectories:false), > partitionKeys:[FieldSchema(name:city, type:string, comment:null), > FieldSchema(name:state, type:string, comment:null)], > parameters:{transactional=true}, viewOriginalText:null, > viewExpandedText:null, tableType:MANAGED_TABLE, > privileges:PrincipalPrivilegeSet(userPrivileges:{nobody=[PrivilegeGrantInfo(privilege:INSERT, > createTime:-1, grantor:nobody, grantorType:USER, grantOption:true), > PrivilegeGrantInfo(privilege:SELECT, createTime:-1, grantor:nobody, > grantorType:USER, grantOption:true), PrivilegeGrantInfo(privilege:UPDATE, > createTime:-1, grantor:nobody, grantorType:USER, grantOption:true), > PrivilegeGrantInfo(privilege:DELETE, createTime:-1, grantor:nobody, > grantorType:USER, grantOption:true)]}, groupPrivileges:null, > rolePrivileges:null), temporary:false) > 62285 [main] INFO o.a.h.h.m.H.audit - ugi=nobody ip=unknown-ip-addr > cmd=create_table: Table(tableName:test_table, dbName:testdb, owner:nobody, > createTime:1511881494, lastAccessTime:0, retention:0, > sd:StorageDescriptor(cols:[FieldSchema(name:id, type:int, comment:null), > FieldSchema(name:msg, type:string, comment:null)], > location:raw:/tmp/junit8935893841909704896/testdb.db/test_table, > inputFormat:org.apache.hadoop.hive.ql.io.orc.OrcInputFormat, > outputFormat:org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat, > compressed:false, numBuckets:10, serdeInfo:SerDeInfo(name:null, > serializationLib:org.apache.hadoop.hive.ql.io.orc.OrcSerde, > parameters:{serialization.format=1}), bucketCols:[id], sortCols:[], > parameters:{}, skewedInfo:SkewedInfo(skewedColNames:[], skewedColValues:[], > skewedColValueLocationMaps:{}), storedAsSubDirectories:false), > partitionKeys:[FieldSchema(name:city, type:string, comment:null), > FieldSchema(name:state, type:string, comment:null)], > parameters:{transactional=true}, viewOriginalText:null, > viewExpandedText:null, tableType:MANAGED_TABLE, > privileges:PrincipalPrivilegeSet(userPrivileges:{nobody=[PrivilegeGrantInfo(privilege:INSERT, > createTime:-1, grantor:nobody, grantorType:USER, grantOption:true), > PrivilegeGrantInfo(privilege:SELECT, createTime:-1, grantor:nobody, > grantorType:USER, grantOption:true), PrivilegeGrantInfo(privilege:UPDATE, > createTime:-1, grantor:nobody, grantorType:USER, grantOption:true), > PrivilegeGrantInfo(privilege:DELETE, createTime:-1, grantor:nobody, > grantorType:USER, grantOption:true)]}, groupPrivileges:null, > rolePrivileges:null), temporary:false) > 62285 [main] INFO o.a.h.h.m.HiveMetaStore - create_table_core > testdb.test_table > 62285 [main] INFO o.a.h.h.m.HiveMetaStore - create_table_core preEvent > testdb.test_table > 62285 [main] INFO o.a.h.h.m.HiveMetaStore - create_table_core rdbms checks > testdb.test_table > 62286 [main] INFO o.a.h.h.m.HiveMetaStore - create_table_core filesystem > testdb.test_table > 62286 [main] WARN o.a.h.h.m.HiveMetaStore - Location: > raw:/tmp/junit8935893841909704896/testdb.db/test_table specified for > non-external table:test_table > 62286 [main] INFO o.a.h.h.c.FileUtils - Creating directory if it doesn't > exist: raw:/tmp/junit8935893841909704896/testdb.db/test_table > 62288 [main] INFO o.a.h.h.m.HiveMetaStore - create_table_core stats > testdb.test_table > 62288 [main] INFO o.a.h.h.m.HiveMetaStore - create_table_core rdbms create > testdb.test_table > 62324 [main] INFO o.a.h.h.m.HiveMetaStore - create_table_core rdbms > committed testdb.test_table > 62324 [main] INFO o.a.h.h.m.HiveMetaStore - create_table_core rdbms > listeners testdb.test_table > 62324 [main] INFO o.a.h.h.m.HiveMetaStore - create_table_core rdbms > listeners done testdb.test_table > 62324 [main] INFO o.a.h.h.q.l.PerfLogger - </PERFLOG method=runTasks > start=1511881494713 end=1511881494754 duration=41 > from=org.apache.hadoop.hive.ql.Driver> > 62324 [main] INFO o.a.h.h.q.Driver - Resetting the caller context to > 62324 [main] INFO o.a.h.h.q.l.PerfLogger - </PERFLOG method=Driver.execute > start=1511881494713 end=1511881494754 duration=41 > from=org.apache.hadoop.hive.ql.Driver> > 62324 [main] INFO o.a.h.h.q.Driver - OK > 62324 [main] INFO o.a.h.h.q.l.PerfLogger - <PERFLOG method=releaseLocks > from=org.apache.hadoop.hive.ql.Driver> > 62324 [main] INFO o.a.h.h.q.l.DbTxnManager - Stopped heartbeat for query: > nobody_20171128150454_f8e917c9-4cbc-4512-b93a-65db6c890fa3 > 62324 [main] INFO o.a.h.h.q.l.DbLockManager - releaseLocks: [lockid:1 > queryId=nobody_20171128150454_f8e917c9-4cbc-4512-b93a-65db6c890fa3 txnid:0] > 62328 [main] INFO o.a.h.h.q.l.PerfLogger - </PERFLOG method=releaseLocks > start=1511881494754 end=1511881494758 duration=4 > from=org.apache.hadoop.hive.ql.Driver> > 62328 [main] INFO o.a.h.h.q.l.PerfLogger - </PERFLOG method=Driver.run > start=1511881494691 end=1511881494758 duration=67 > from=org.apache.hadoop.hive.ql.Driver> > 62328 [main] INFO o.a.h.h.q.l.PerfLogger - <PERFLOG method=Driver.run > from=org.apache.hadoop.hive.ql.Driver> > 62329 [main] INFO o.a.h.h.q.l.PerfLogger - <PERFLOG method=compile > from=org.apache.hadoop.hive.ql.Driver> > 62329 [main] INFO o.a.h.h.q.Driver - We are setting the hadoop caller > context from to nobody_20171128150454_ae26eba9-11ee-41a6-9632-4b64134ecb73 > 62329 [main] INFO o.a.h.h.q.l.PerfLogger - <PERFLOG method=parse > from=org.apache.hadoop.hive.ql.Driver> > 62329 [main] INFO h.q.p.ParseDriver - Parsing command: alter table > test_table add partition ( city = 'sunnyvale',state = 'ca' ) > 62329 [main] INFO h.q.p.ParseDriver - Parse Completed > 62329 [main] INFO o.a.h.h.q.l.PerfLogger - </PERFLOG method=parse > start=1511881494759 end=1511881494759 duration=0 > from=org.apache.hadoop.hive.ql.Driver> > 62329 [main] INFO o.a.h.h.q.l.PerfLogger - <PERFLOG method=semanticAnalyze > from=org.apache.hadoop.hive.ql.Driver> > 62330 [main] INFO o.a.h.h.m.HiveMetaStore - 0: get_table : db=testdb > tbl=test_table > 62330 [main] INFO o.a.h.h.m.H.audit - ugi=nobody ip=unknown-ip-addr > cmd=get_table : db=testdb tbl=test_table > 62374 [main] INFO o.a.h.h.q.Driver - Semantic Analysis Completed > 62374 [main] INFO o.a.h.h.q.l.PerfLogger - </PERFLOG method=semanticAnalyze > start=1511881494759 end=1511881494804 duration=45 > from=org.apache.hadoop.hive.ql.Driver> > 62374 [main] INFO o.a.h.h.q.Driver - Returning Hive schema: > Schema(fieldSchemas:null, properties:null) > 62374 [main] INFO o.a.h.h.q.l.PerfLogger - </PERFLOG method=compile > start=1511881494759 end=1511881494804 duration=45 > from=org.apache.hadoop.hive.ql.Driver> > 62374 [main] INFO o.a.h.h.q.Driver - We are resetting the hadoop caller > context to > 62374 [main] INFO o.a.h.h.q.l.PerfLogger - <PERFLOG > method=acquireReadWriteLocks from=org.apache.hadoop.hive.ql.Driver> > 62374 [main] INFO o.a.h.h.q.l.DbTxnManager - Setting lock request > transaction to txnid:0 for > queryId=nobody_20171128150454_ae26eba9-11ee-41a6-9632-4b64134ecb73 > 62374 [main] INFO o.a.h.h.q.l.DbLockManager - Requesting: > queryId=nobody_20171128150454_ae26eba9-11ee-41a6-9632-4b64134ecb73 > LockRequest(component:[LockComponent(type:SHARED_READ, level:TABLE, > dbname:testdb, tablename:test_table, operationType:NO_TXN, isAcid:true, > isDynamicPartitionWrite:false)], txnid:0, user:nobody, > hostname:ctr-e134-1499953498516-345859-01-000003.hwx.site, > agentInfo:nobody_20171128150454_ae26eba9-11ee-41a6-9632-4b64134ecb73) > 62386 [main] INFO o.a.h.h.q.l.DbLockManager - Response to > queryId=nobody_20171128150454_ae26eba9-11ee-41a6-9632-4b64134ecb73 > LockResponse(lockid:2, state:ACQUIRED) > 62386 [main] INFO o.a.h.h.q.l.DbTxnManager - Started heartbeat with > delay/interval = 150000/150000 MILLISECONDS for query: > nobody_20171128150454_ae26eba9-11ee-41a6-9632-4b64134ecb73 > 62386 [main] INFO o.a.h.h.q.l.PerfLogger - </PERFLOG > method=acquireReadWriteLocks start=1511881494804 end=1511881494816 > duration=12 from=org.apache.hadoop.hive.ql.Driver> > 62386 [main] INFO o.a.h.h.q.l.PerfLogger - <PERFLOG method=Driver.execute > from=org.apache.hadoop.hive.ql.Driver> > 62386 [main] INFO o.a.h.h.q.Driver - Setting caller context to query id > nobody_20171128150454_ae26eba9-11ee-41a6-9632-4b64134ecb73 > 62386 [main] INFO o.a.h.h.q.Driver - Starting > command(queryId=nobody_20171128150454_ae26eba9-11ee-41a6-9632-4b64134ecb73): > create database IF NOT EXISTS testdb location > 'raw:///tmp/junit8935893841909704896/testdb.db' > 62386 [main] INFO o.a.h.h.q.l.PerfLogger - <PERFLOG method=runTasks > from=org.apache.hadoop.hive.ql.Driver> > 62387 [main] INFO o.a.h.h.q.Driver - Starting task [Stage-0:DDL] in serial > mode > 62387 [main] INFO o.a.h.h.m.HiveMetaStore - 0: get_table : db=testdb > tbl=test_table > 62387 [main] INFO o.a.h.h.m.H.audit - ugi=nobody ip=unknown-ip-addr > cmd=get_table : db=testdb tbl=test_table > 62393 [main] INFO o.a.h.h.m.HiveMetaStore - 0: add_partitions > 62393 [main] INFO o.a.h.h.m.H.audit - ugi=nobody ip=unknown-ip-addr > cmd=add_partitions > 62404 [HMSHandler #3] INFO o.a.h.h.c.FileUtils - Creating directory if it > doesn't exist: > raw:/tmp/junit8935893841909704896/testdb.db/test_table/city=sunnyvale/state=ca > 62422 [main] INFO o.a.h.h.q.l.PerfLogger - </PERFLOG method=runTasks > start=1511881494816 end=1511881494852 duration=36 > from=org.apache.hadoop.hive.ql.Driver> > 62422 [main] INFO o.a.h.h.q.Driver - Resetting the caller context to > 62422 [main] INFO o.a.h.h.q.l.PerfLogger - </PERFLOG method=Driver.execute > start=1511881494816 end=1511881494852 duration=36 > from=org.apache.hadoop.hive.ql.Driver> > 62422 [main] INFO o.a.h.h.q.Driver - OK > 62422 [main] INFO o.a.h.h.q.l.PerfLogger - <PERFLOG method=releaseLocks > from=org.apache.hadoop.hive.ql.Driver> > 62422 [main] INFO o.a.h.h.q.l.DbTxnManager - Stopped heartbeat for query: > nobody_20171128150454_ae26eba9-11ee-41a6-9632-4b64134ecb73 > 62422 [main] INFO o.a.h.h.q.l.DbLockManager - releaseLocks: [lockid:2 > queryId=nobody_20171128150454_ae26eba9-11ee-41a6-9632-4b64134ecb73 txnid:0] > 62426 [main] INFO o.a.h.h.q.l.PerfLogger - </PERFLOG method=releaseLocks > start=1511881494852 end=1511881494856 duration=4 > from=org.apache.hadoop.hive.ql.Driver> > 62426 [main] INFO o.a.h.h.q.l.PerfLogger - </PERFLOG method=Driver.run > start=1511881494758 end=1511881494856 duration=98 > from=org.apache.hadoop.hive.ql.Driver> > Standard Error > OK > OK > OK > OK{code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)