[ https://issues.apache.org/jira/browse/ASTERIXDB-1724?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Wenhai updated ASTERIXDB-1724: ------------------------------ Description: Hi, all Yestoday, after I applying the latest master onto the branch. I got a series of error (13 testcases) that should not be posed by my branch, like the following: {noformat} java.lang.Exception: Test "src/test/resources/runtimets/queries/api/cluster_state_1/cluster_state_1.1.cstate.aql" FAILED! at org.apache.asterix.test.aql.TestExecutor.executeTest(TestExecutor.java:1028) at org.apache.asterix.test.runtime.ExecutionTest.test(ExecutionTest.java:127) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:497) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57) at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288) at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268) at org.junit.runners.ParentRunner.run(ParentRunner.java:363) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288) at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) at org.junit.runners.ParentRunner.run(ParentRunner.java:363) at org.eclipse.jdt.internal.junit4.runner.JUnit4TestReference.run(JUnit4TestReference.java:86) at org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:38) at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:459) at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:675) at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:382) at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:192) Caused by: org.apache.asterix.test.base.ComparisonException: Result for src/test/resources/runtimets/queries/api/cluster_state_1/cluster_state_1.1.cstate.aql changed at line 3: < "configUri": "http://127.0.0.1:19002/admin/cluster/cc/config", > "statsUri": "http://127.0.0.1:19002/admin/cluster/cc/stats", at org.apache.asterix.test.aql.TestExecutor.throwLineChanged(TestExecutor.java:203) at org.apache.asterix.test.aql.TestExecutor.runScriptAndCompareWithResult(TestExecutor.java:163) at org.apache.asterix.test.aql.TestExecutor.executeTest(TestExecutor.java:870) at org.apache.asterix.test.aql.TestExecutor.executeTest(TestExecutor.java:1000) ... 34 more {noformat} The excepted result of runtimets/results/api/cluster_state_1 is {noformat} { "cc": { "configUri": "http://127.0.0.1:19002/admin/cluster/cc/config", "statsUri": "http://127.0.0.1:19002/admin/cluster/cc/stats", "threadDumpUri": "http://127.0.0.1:19002/admin/cluster/cc/threaddump" }, "config": { "api.port": 19002, "cluster.partitions": { "0": "ID:0, Original Node: asterix_nc1, IODevice: 0, Active Node: asterix_nc1", "1": "ID:1, Original Node: asterix_nc1, IODevice: 1, Active Node: asterix_nc1", "2": "ID:2, Original Node: asterix_nc2, IODevice: 0, Active Node: asterix_nc2", "3": "ID:3, Original Node: asterix_nc2, IODevice: 1, Active Node: asterix_nc2" }, "compiler.framesize": 32768, "compiler.groupmemory": 163840, "compiler.joinmemory": 163840, "compiler.pregelix.home": "~/pregelix", "compiler.sortmemory": 327680, "core.dump.paths": {}, "feed.central.manager.port": 4500, "feed.max.threshold.period": 5, "feed.memory.available.wait.timeout": 10, "feed.memory.global.budget": 67108864, "feed.pending.work.threshold": 50, "feed.port": 19003, "instance.name": null, "log.level": "INFO", "max.wait.active.cluster": 60, "metadata.callback.port": 0, "metadata.node": "asterix_nc1", "metadata.partition": "ID:0, Original Node: asterix_nc1, IODevice: 0, Active Node: asterix_nc1", "metadata.port": 0, "metadata.registration.timeout.secs": 60, "node.partitions": { "asterix_nc1": [ "ID:0, Original Node: asterix_nc1, IODevice: 0, Active Node: asterix_nc1", "ID:1, Original Node: asterix_nc1, IODevice: 1, Active Node: asterix_nc1" ], "asterix_nc2": [ "ID:2, Original Node: asterix_nc2, IODevice: 0, Active Node: asterix_nc2", "ID:3, Original Node: asterix_nc2, IODevice: 1, Active Node: asterix_nc2" ] }, "node.stores": { "asterix_nc1": [ "iodevice0", "iodevice1" ], "asterix_nc2": [ "iodevice0", "iodevice1" ] }, "plot.activate": false, "storage.buffercache.maxopenfiles": 2147483647, "storage.buffercache.pagesize": 32768, "storage.buffercache.size": 33554432, "storage.lsm.bloomfilter.falsepositiverate": 0.01, "storage.memorycomponent.globalbudget": 536870912, "storage.memorycomponent.numcomponents": 2, "storage.memorycomponent.numpages": 8, "storage.memorycomponent.pagesize": 131072, "storage.metadata.memorycomponent.numpages": 256, "transaction.log.dirs": { "asterix_nc1": "target/txnLogDir/asterix_nc1", "asterix_nc2": "target/txnLogDir/asterix_nc2" }, "txn.commitprofiler.reportinterval": 5, "txn.job.recovery.memorysize": 67108864, "txn.lock.escalationthreshold": 1000, "txn.lock.shrinktimer": 5000, "txn.lock.timeout.sweepthreshold": 10000, "txn.lock.timeout.waitthreshold": 60000, "txn.log.buffer.numpages": 8, "txn.log.buffer.pagesize": 131072, "txn.log.checkpoint.history": 0, "txn.log.checkpoint.lsnthreshold": 67108864, "txn.log.checkpoint.pollfrequency": 120, "txn.log.partitionsize": 268435456, "web.port": 19001, "web.queryinterface.port": 19006, "web.secondary.port": 19005 }, "diagnosticsUri": "http://127.0.0.1:19002/admin/diagnostics", "fullShutdownUri": "http://127.0.0.1:19002/admin/shutdown?all=true", "metadata_node": "asterix_nc1", "ncs": [ { "configUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc1/config", "node_id": "asterix_nc1", "partitions": [ { "active": true, "partition_id": "partition_0" }, { "active": true, "partition_id": "partition_1" } ], "state": "ACTIVE", "statsUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc1/stats", "threadDumpUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc1/threaddump" }, { "configUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc2/config", "node_id": "asterix_nc2", "partitions": [ { "active": true, "partition_id": "partition_2" }, { "active": true, "partition_id": "partition_3" } ], "state": "ACTIVE", "statsUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc2/stats", "threadDumpUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc2/threaddump" } ], "replicationUri": "http://127.0.0.1:19002/admin/cluster/replication", "shutdownUri": "http://127.0.0.1:19002/admin/shutdown", "state": "ACTIVE", "versionUri": "http://127.0.0.1:19002/admin/version" {noformat} But I got: {noformat} { "cc": { "statsUri": "http://127.0.0.1:19002/admin/cluster/cc/stats", "configUri": "http://127.0.0.1:19002/admin/cluster/cc/config", "threadDumpUri": "http://127.0.0.1:19002/admin/cluster/cc/threaddump" }, "replicationUri": "http://127.0.0.1:19002/admin/cluster/replication", "fullShutdownUri": "http://127.0.0.1:19002/admin/shutdown?all=true", "state": "ACTIVE", "metadata_node": "asterix_nc1", "shutdownUri": "http://127.0.0.1:19002/admin/shutdown", "versionUri": "http://127.0.0.1:19002/admin/version", "config": { "feed.port": 19003, "txn.log.checkpoint.history": 0, "compiler.joinmemory": 163840, "feed.central.manager.port": 4500, "compiler.sortmemory": 327680, "node.partitions": { "asterix_nc1": [ { "partitionId": 0, "IODeviceNum": 0, "active": true, "activeNodeId": "asterix_nc1", "nodeId": "asterix_nc1" }, { "partitionId": 1, "IODeviceNum": 1, "active": true, "activeNodeId": "asterix_nc1", "nodeId": "asterix_nc1" } ], "asterix_nc2": [ { "partitionId": 2, "IODeviceNum": 0, "active": true, "activeNodeId": "asterix_nc2", "nodeId": "asterix_nc2" }, { "partitionId": 3, "IODeviceNum": 1, "active": true, "activeNodeId": "asterix_nc2", "nodeId": "asterix_nc2" } ] }, "txn.lock.shrinktimer": 5000, "metadata.node": "asterix_nc1", "metadata.port": 0, "log.level": "INFO", "api.port": 19002, "cluster.partitions": { "0": { "partitionId": 0, "IODeviceNum": 0, "active": true, "activeNodeId": "asterix_nc1", "nodeId": "asterix_nc1" }, "1": { "partitionId": 1, "IODeviceNum": 1, "active": true, "activeNodeId": "asterix_nc1", "nodeId": "asterix_nc1" }, "2": { "partitionId": 2, "IODeviceNum": 0, "active": true, "activeNodeId": "asterix_nc2", "nodeId": "asterix_nc2" }, "3": { "partitionId": 3, "IODeviceNum": 1, "active": true, "activeNodeId": "asterix_nc2", "nodeId": "asterix_nc2" } }, "storage.buffercache.size": 33554432, "web.port": 19001, "txn.log.checkpoint.lsnthreshold": 67108864, "txn.log.partitionsize": 268435456, "compiler.groupmemory": 163840, "max.wait.active.cluster": 60, "metadata.registration.timeout.secs": 60, "storage.lsm.bloomfilter.falsepositiverate": 0.01, "plot.activate": false, "storage.memorycomponent.globalbudget": 536870912, "web.queryinterface.port": 19006, "txn.log.buffer.numpages": 8, "web.secondary.port": 19005, "feed.memory.global.budget": 67108864, "compiler.pregelix.home": "~/pregelix", "compiler.framesize": 32768, "feed.memory.available.wait.timeout": 10, "core.dump.paths": {}, "txn.log.buffer.pagesize": 131072, "txn.job.recovery.memorysize": 67108864, "storage.metadata.memorycomponent.numpages": 256, "metadata.partition": { "partitionId": 0, "IODeviceNum": 0, "active": true, "activeNodeId": "asterix_nc1", "nodeId": "asterix_nc1" }, "txn.log.checkpoint.pollfrequency": 120, "feed.pending.work.threshold": 50, "feed.max.threshold.period": 5, "storage.buffercache.maxopenfiles": 2147483647, "txn.lock.escalationthreshold": 1000, "txn.lock.timeout.waitthreshold": 60000, "storage.memorycomponent.numcomponents": 2, "storage.buffercache.pagesize": 32768, "storage.memorycomponent.numpages": 8, "transaction.log.dirs": { "asterix_nc1": "target/txnLogDir/asterix_nc1", "asterix_nc2": "target/txnLogDir/asterix_nc2" }, "txn.commitprofiler.reportinterval": 5, "metadata.callback.port": 0, "txn.lock.timeout.sweepthreshold": 10000, "storage.memorycomponent.pagesize": 131072, "node.stores": { "asterix_nc1": [ "iodevice0", "iodevice1" ], "asterix_nc2": [ "iodevice0", "iodevice1" ] } }, "diagnosticsUri": "http://127.0.0.1:19002/admin/diagnostics", "ncs": [ { "partitions": [ { "partition_id": "partition_0", "active": true }, { "partition_id": "partition_1", "active": true } ], "statsUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc1/stats", "configUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc1/config", "state": "ACTIVE", "node_id": "asterix_nc1", "threadDumpUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc1/threaddump" }, { "partitions": [ { "partition_id": "partition_2", "active": true }, { "partition_id": "partition_3", "active": true } ], "statsUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc2/stats", "configUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc2/config", "state": "ACTIVE", "node_id": "asterix_nc2", "threadDumpUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc2/threaddump" } ] } {noformat} was: Hi, all Yestoday, after I applying the latest master onto the branch. I got a series of error (13 testcases) that should not be posed by my branch, like the following: {noformat} java.lang.Exception: Test "src/test/resources/runtimets/queries/api/cluster_state_1/cluster_state_1.1.cstate.aql" FAILED! at org.apache.asterix.test.aql.TestExecutor.executeTest(TestExecutor.java:1028) at org.apache.asterix.test.runtime.ExecutionTest.test(ExecutionTest.java:127) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:497) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57) at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288) at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268) at org.junit.runners.ParentRunner.run(ParentRunner.java:363) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288) at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) at org.junit.runners.ParentRunner.run(ParentRunner.java:363) at org.eclipse.jdt.internal.junit4.runner.JUnit4TestReference.run(JUnit4TestReference.java:86) at org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:38) at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:459) at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:675) at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:382) at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:192) Caused by: org.apache.asterix.test.base.ComparisonException: Result for src/test/resources/runtimets/queries/api/cluster_state_1/cluster_state_1.1.cstate.aql changed at line 3: < "configUri": "http://127.0.0.1:19002/admin/cluster/cc/config", > "statsUri": "http://127.0.0.1:19002/admin/cluster/cc/stats", at org.apache.asterix.test.aql.TestExecutor.throwLineChanged(TestExecutor.java:203) at org.apache.asterix.test.aql.TestExecutor.runScriptAndCompareWithResult(TestExecutor.java:163) at org.apache.asterix.test.aql.TestExecutor.executeTest(TestExecutor.java:870) at org.apache.asterix.test.aql.TestExecutor.executeTest(TestExecutor.java:1000) ... 34 more {noformat} > Unexpected exception in ExecutionTest > ------------------------------------- > > Key: ASTERIXDB-1724 > URL: https://issues.apache.org/jira/browse/ASTERIXDB-1724 > Project: Apache AsterixDB > Issue Type: Test > Environment: General system. > Reporter: Wenhai > > Hi, all > Yestoday, after I applying the latest master onto the branch. I got a series > of error (13 testcases) that should not be posed by my branch, like the > following: > {noformat} > java.lang.Exception: Test > "src/test/resources/runtimets/queries/api/cluster_state_1/cluster_state_1.1.cstate.aql" > FAILED! > at > org.apache.asterix.test.aql.TestExecutor.executeTest(TestExecutor.java:1028) > at > org.apache.asterix.test.runtime.ExecutionTest.test(ExecutionTest.java:127) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:497) > at > org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50) > at > org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) > at > org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47) > at > org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) > at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325) > at > org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78) > at > org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57) > at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290) > at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71) > at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288) > at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58) > at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268) > at org.junit.runners.ParentRunner.run(ParentRunner.java:363) > at org.junit.runners.Suite.runChild(Suite.java:128) > at org.junit.runners.Suite.runChild(Suite.java:27) > at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290) > at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71) > at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288) > at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58) > at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268) > at > org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) > at > org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) > at org.junit.runners.ParentRunner.run(ParentRunner.java:363) > at > org.eclipse.jdt.internal.junit4.runner.JUnit4TestReference.run(JUnit4TestReference.java:86) > at > org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:38) > at > org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:459) > at > org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:675) > at > org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:382) > at > org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:192) > Caused by: org.apache.asterix.test.base.ComparisonException: Result for > src/test/resources/runtimets/queries/api/cluster_state_1/cluster_state_1.1.cstate.aql > changed at line 3: > < "configUri": "http://127.0.0.1:19002/admin/cluster/cc/config", > > "statsUri": "http://127.0.0.1:19002/admin/cluster/cc/stats", > at > org.apache.asterix.test.aql.TestExecutor.throwLineChanged(TestExecutor.java:203) > at > org.apache.asterix.test.aql.TestExecutor.runScriptAndCompareWithResult(TestExecutor.java:163) > at > org.apache.asterix.test.aql.TestExecutor.executeTest(TestExecutor.java:870) > at > org.apache.asterix.test.aql.TestExecutor.executeTest(TestExecutor.java:1000) > ... 34 more > {noformat} > The excepted result of runtimets/results/api/cluster_state_1 is > {noformat} > { > "cc": { > "configUri": "http://127.0.0.1:19002/admin/cluster/cc/config", > "statsUri": "http://127.0.0.1:19002/admin/cluster/cc/stats", > "threadDumpUri": "http://127.0.0.1:19002/admin/cluster/cc/threaddump" > }, > "config": { > "api.port": 19002, > "cluster.partitions": { > "0": "ID:0, Original Node: asterix_nc1, IODevice: 0, Active Node: > asterix_nc1", > "1": "ID:1, Original Node: asterix_nc1, IODevice: 1, Active Node: > asterix_nc1", > "2": "ID:2, Original Node: asterix_nc2, IODevice: 0, Active Node: > asterix_nc2", > "3": "ID:3, Original Node: asterix_nc2, IODevice: 1, Active Node: > asterix_nc2" > }, > "compiler.framesize": 32768, > "compiler.groupmemory": 163840, > "compiler.joinmemory": 163840, > "compiler.pregelix.home": "~/pregelix", > "compiler.sortmemory": 327680, > "core.dump.paths": {}, > "feed.central.manager.port": 4500, > "feed.max.threshold.period": 5, > "feed.memory.available.wait.timeout": 10, > "feed.memory.global.budget": 67108864, > "feed.pending.work.threshold": 50, > "feed.port": 19003, > "instance.name": null, > "log.level": "INFO", > "max.wait.active.cluster": 60, > "metadata.callback.port": 0, > "metadata.node": "asterix_nc1", > "metadata.partition": "ID:0, Original Node: asterix_nc1, IODevice: 0, > Active Node: asterix_nc1", > "metadata.port": 0, > "metadata.registration.timeout.secs": 60, > "node.partitions": { > "asterix_nc1": [ > "ID:0, Original Node: asterix_nc1, IODevice: 0, Active Node: > asterix_nc1", > "ID:1, Original Node: asterix_nc1, IODevice: 1, Active Node: > asterix_nc1" > ], > "asterix_nc2": [ > "ID:2, Original Node: asterix_nc2, IODevice: 0, Active Node: > asterix_nc2", > "ID:3, Original Node: asterix_nc2, IODevice: 1, Active Node: > asterix_nc2" > ] > }, > "node.stores": { > "asterix_nc1": [ > "iodevice0", > "iodevice1" > ], > "asterix_nc2": [ > "iodevice0", > "iodevice1" > ] > }, > "plot.activate": false, > "storage.buffercache.maxopenfiles": 2147483647, > "storage.buffercache.pagesize": 32768, > "storage.buffercache.size": 33554432, > "storage.lsm.bloomfilter.falsepositiverate": 0.01, > "storage.memorycomponent.globalbudget": 536870912, > "storage.memorycomponent.numcomponents": 2, > "storage.memorycomponent.numpages": 8, > "storage.memorycomponent.pagesize": 131072, > "storage.metadata.memorycomponent.numpages": 256, > "transaction.log.dirs": { > "asterix_nc1": "target/txnLogDir/asterix_nc1", > "asterix_nc2": "target/txnLogDir/asterix_nc2" > }, > "txn.commitprofiler.reportinterval": 5, > "txn.job.recovery.memorysize": 67108864, > "txn.lock.escalationthreshold": 1000, > "txn.lock.shrinktimer": 5000, > "txn.lock.timeout.sweepthreshold": 10000, > "txn.lock.timeout.waitthreshold": 60000, > "txn.log.buffer.numpages": 8, > "txn.log.buffer.pagesize": 131072, > "txn.log.checkpoint.history": 0, > "txn.log.checkpoint.lsnthreshold": 67108864, > "txn.log.checkpoint.pollfrequency": 120, > "txn.log.partitionsize": 268435456, > "web.port": 19001, > "web.queryinterface.port": 19006, > "web.secondary.port": 19005 > }, > "diagnosticsUri": "http://127.0.0.1:19002/admin/diagnostics", > "fullShutdownUri": "http://127.0.0.1:19002/admin/shutdown?all=true", > "metadata_node": "asterix_nc1", > "ncs": [ > { > "configUri": > "http://127.0.0.1:19002/admin/cluster/node/asterix_nc1/config", > "node_id": "asterix_nc1", > "partitions": [ > { > "active": true, > "partition_id": "partition_0" > }, > { > "active": true, > "partition_id": "partition_1" > } > ], > "state": "ACTIVE", > "statsUri": > "http://127.0.0.1:19002/admin/cluster/node/asterix_nc1/stats", > "threadDumpUri": > "http://127.0.0.1:19002/admin/cluster/node/asterix_nc1/threaddump" > }, > { > "configUri": > "http://127.0.0.1:19002/admin/cluster/node/asterix_nc2/config", > "node_id": "asterix_nc2", > "partitions": [ > { > "active": true, > "partition_id": "partition_2" > }, > { > "active": true, > "partition_id": "partition_3" > } > ], > "state": "ACTIVE", > "statsUri": > "http://127.0.0.1:19002/admin/cluster/node/asterix_nc2/stats", > "threadDumpUri": > "http://127.0.0.1:19002/admin/cluster/node/asterix_nc2/threaddump" > } > ], > "replicationUri": "http://127.0.0.1:19002/admin/cluster/replication", > "shutdownUri": "http://127.0.0.1:19002/admin/shutdown", > "state": "ACTIVE", > "versionUri": "http://127.0.0.1:19002/admin/version" > {noformat} > But I got: > {noformat} > { > "cc": { > "statsUri": "http://127.0.0.1:19002/admin/cluster/cc/stats", > "configUri": "http://127.0.0.1:19002/admin/cluster/cc/config", > "threadDumpUri": "http://127.0.0.1:19002/admin/cluster/cc/threaddump" > }, > "replicationUri": "http://127.0.0.1:19002/admin/cluster/replication", > "fullShutdownUri": "http://127.0.0.1:19002/admin/shutdown?all=true", > "state": "ACTIVE", > "metadata_node": "asterix_nc1", > "shutdownUri": "http://127.0.0.1:19002/admin/shutdown", > "versionUri": "http://127.0.0.1:19002/admin/version", > "config": { > "feed.port": 19003, > "txn.log.checkpoint.history": 0, > "compiler.joinmemory": 163840, > "feed.central.manager.port": 4500, > "compiler.sortmemory": 327680, > "node.partitions": { > "asterix_nc1": [ > { > "partitionId": 0, > "IODeviceNum": 0, > "active": true, > "activeNodeId": "asterix_nc1", > "nodeId": "asterix_nc1" > }, > { > "partitionId": 1, > "IODeviceNum": 1, > "active": true, > "activeNodeId": "asterix_nc1", > "nodeId": "asterix_nc1" > } > ], > "asterix_nc2": [ > { > "partitionId": 2, > "IODeviceNum": 0, > "active": true, > "activeNodeId": "asterix_nc2", > "nodeId": "asterix_nc2" > }, > { > "partitionId": 3, > "IODeviceNum": 1, > "active": true, > "activeNodeId": "asterix_nc2", > "nodeId": "asterix_nc2" > } > ] > }, > "txn.lock.shrinktimer": 5000, > "metadata.node": "asterix_nc1", > "metadata.port": 0, > "log.level": "INFO", > "api.port": 19002, > "cluster.partitions": { > "0": { > "partitionId": 0, > "IODeviceNum": 0, > "active": true, > "activeNodeId": "asterix_nc1", > "nodeId": "asterix_nc1" > }, > "1": { > "partitionId": 1, > "IODeviceNum": 1, > "active": true, > "activeNodeId": "asterix_nc1", > "nodeId": "asterix_nc1" > }, > "2": { > "partitionId": 2, > "IODeviceNum": 0, > "active": true, > "activeNodeId": "asterix_nc2", > "nodeId": "asterix_nc2" > }, > "3": { > "partitionId": 3, > "IODeviceNum": 1, > "active": true, > "activeNodeId": "asterix_nc2", > "nodeId": "asterix_nc2" > } > }, > "storage.buffercache.size": 33554432, > "web.port": 19001, > "txn.log.checkpoint.lsnthreshold": 67108864, > "txn.log.partitionsize": 268435456, > "compiler.groupmemory": 163840, > "max.wait.active.cluster": 60, > "metadata.registration.timeout.secs": 60, > "storage.lsm.bloomfilter.falsepositiverate": 0.01, > "plot.activate": false, > "storage.memorycomponent.globalbudget": 536870912, > "web.queryinterface.port": 19006, > "txn.log.buffer.numpages": 8, > "web.secondary.port": 19005, > "feed.memory.global.budget": 67108864, > "compiler.pregelix.home": "~/pregelix", > "compiler.framesize": 32768, > "feed.memory.available.wait.timeout": 10, > "core.dump.paths": {}, > "txn.log.buffer.pagesize": 131072, > "txn.job.recovery.memorysize": 67108864, > "storage.metadata.memorycomponent.numpages": 256, > "metadata.partition": { > "partitionId": 0, > "IODeviceNum": 0, > "active": true, > "activeNodeId": "asterix_nc1", > "nodeId": "asterix_nc1" > }, > "txn.log.checkpoint.pollfrequency": 120, > "feed.pending.work.threshold": 50, > "feed.max.threshold.period": 5, > "storage.buffercache.maxopenfiles": 2147483647, > "txn.lock.escalationthreshold": 1000, > "txn.lock.timeout.waitthreshold": 60000, > "storage.memorycomponent.numcomponents": 2, > "storage.buffercache.pagesize": 32768, > "storage.memorycomponent.numpages": 8, > "transaction.log.dirs": { > "asterix_nc1": "target/txnLogDir/asterix_nc1", > "asterix_nc2": "target/txnLogDir/asterix_nc2" > }, > "txn.commitprofiler.reportinterval": 5, > "metadata.callback.port": 0, > "txn.lock.timeout.sweepthreshold": 10000, > "storage.memorycomponent.pagesize": 131072, > "node.stores": { > "asterix_nc1": [ > "iodevice0", > "iodevice1" > ], > "asterix_nc2": [ > "iodevice0", > "iodevice1" > ] > } > }, > "diagnosticsUri": "http://127.0.0.1:19002/admin/diagnostics", > "ncs": [ > { > "partitions": [ > { > "partition_id": "partition_0", > "active": true > }, > { > "partition_id": "partition_1", > "active": true > } > ], > "statsUri": > "http://127.0.0.1:19002/admin/cluster/node/asterix_nc1/stats", > "configUri": > "http://127.0.0.1:19002/admin/cluster/node/asterix_nc1/config", > "state": "ACTIVE", > "node_id": "asterix_nc1", > "threadDumpUri": > "http://127.0.0.1:19002/admin/cluster/node/asterix_nc1/threaddump" > }, > { > "partitions": [ > { > "partition_id": "partition_2", > "active": true > }, > { > "partition_id": "partition_3", > "active": true > } > ], > "statsUri": > "http://127.0.0.1:19002/admin/cluster/node/asterix_nc2/stats", > "configUri": > "http://127.0.0.1:19002/admin/cluster/node/asterix_nc2/config", > "state": "ACTIVE", > "node_id": "asterix_nc2", > "threadDumpUri": > "http://127.0.0.1:19002/admin/cluster/node/asterix_nc2/threaddump" > } > ] > } > {noformat} -- This message was sent by Atlassian JIRA (v6.3.4#6332)