BTW, the expected result:
{
"cc": {
"configUri": "http://127.0.0.1:19002/admin/cluster/cc/config",
"statsUri": "http://127.0.0.1:19002/admin/cluster/cc/stats",
"threadDumpUri": "http://127.0.0.1:19002/admin/cluster/cc/threaddump"
},
"config": {
"api.port": 19002,
"cluster.partitions": {
"0": "ID:0, Original Node: asterix_nc1, IODevice: 0, Active Node:
asterix_nc1",
"1": "ID:1, Original Node: asterix_nc1, IODevice: 1, Active Node:
asterix_nc1",
"2": "ID:2, Original Node: asterix_nc2, IODevice: 0, Active Node:
asterix_nc2",
"3": "ID:3, Original Node: asterix_nc2, IODevice: 1, Active Node:
asterix_nc2"
},
"compiler.framesize": 32768,
"compiler.groupmemory": 163840,
"compiler.joinmemory": 163840,
"compiler.pregelix.home": "~/pregelix",
"compiler.sortmemory": 327680,
"core.dump.paths": {},
"feed.central.manager.port": 4500,
"feed.max.threshold.period": 5,
"feed.memory.available.wait.timeout": 10,
"feed.memory.global.budget": 67108864,
"feed.pending.work.threshold": 50,
"feed.port": 19003,
"instance.name": null,
"log.level": "INFO",
"max.wait.active.cluster": 60,
"metadata.callback.port": 0,
"metadata.node": "asterix_nc1",
"metadata.partition": "ID:0, Original Node: asterix_nc1, IODevice: 0,
Active Node: asterix_nc1",
"metadata.port": 0,
"metadata.registration.timeout.secs": 60,
"node.partitions": {
"asterix_nc1": [
"ID:0, Original Node: asterix_nc1, IODevice: 0, Active Node:
asterix_nc1",
"ID:1, Original Node: asterix_nc1, IODevice: 1, Active Node:
asterix_nc1"
],
"asterix_nc2": [
"ID:2, Original Node: asterix_nc2, IODevice: 0, Active Node:
asterix_nc2",
"ID:3, Original Node: asterix_nc2, IODevice: 1, Active Node:
asterix_nc2"
]
},
"node.stores": {
"asterix_nc1": [
"iodevice0",
"iodevice1"
],
"asterix_nc2": [
"iodevice0",
"iodevice1"
]
},
"plot.activate": false,
"storage.buffercache.maxopenfiles": 2147483647,
"storage.buffercache.pagesize": 32768,
"storage.buffercache.size": 33554432,
"storage.lsm.bloomfilter.falsepositiverate": 0.01,
"storage.memorycomponent.globalbudget": 536870912,
"storage.memorycomponent.numcomponents": 2,
"storage.memorycomponent.numpages": 8,
"storage.memorycomponent.pagesize": 131072,
"storage.metadata.memorycomponent.numpages": 256,
"transaction.log.dirs": {
"asterix_nc1": "target/txnLogDir/asterix_nc1",
"asterix_nc2": "target/txnLogDir/asterix_nc2"
},
"txn.commitprofiler.reportinterval": 5,
"txn.job.recovery.memorysize": 67108864,
"txn.lock.escalationthreshold": 1000,
"txn.lock.shrinktimer": 5000,
"txn.lock.timeout.sweepthreshold": 10000,
"txn.lock.timeout.waitthreshold": 60000,
"txn.log.buffer.numpages": 8,
"txn.log.buffer.pagesize": 131072,
"txn.log.checkpoint.history": 0,
"txn.log.checkpoint.lsnthreshold": 67108864,
"txn.log.checkpoint.pollfrequency": 120,
"txn.log.partitionsize": 268435456,
"web.port": 19001,
"web.queryinterface.port": 19006,
"web.secondary.port": 19005
},
"diagnosticsUri": "http://127.0.0.1:19002/admin/diagnostics",
"fullShutdownUri": "http://127.0.0.1:19002/admin/shutdown?all=true",
"metadata_node": "asterix_nc1",
"ncs": [
{
"configUri":
"http://127.0.0.1:19002/admin/cluster/node/asterix_nc1/config",
"node_id": "asterix_nc1",
"partitions": [
{
"active": true,
"partition_id": "partition_0"
},
{
"active": true,
"partition_id": "partition_1"
}
],
"state": "ACTIVE",
"statsUri":
"http://127.0.0.1:19002/admin/cluster/node/asterix_nc1/stats",
"threadDumpUri":
"http://127.0.0.1:19002/admin/cluster/node/asterix_nc1/threaddump"
},
{
"configUri":
"http://127.0.0.1:19002/admin/cluster/node/asterix_nc2/config",
"node_id": "asterix_nc2",
"partitions": [
{
"active": true,
"partition_id": "partition_2"
},
{
"active": true,
"partition_id": "partition_3"
}
],
"state": "ACTIVE",
"statsUri":
"http://127.0.0.1:19002/admin/cluster/node/asterix_nc2/stats",
"threadDumpUri":
"http://127.0.0.1:19002/admin/cluster/node/asterix_nc2/threaddump"
}
],
"replicationUri": "http://127.0.0.1:19002/admin/cluster/replication",
"shutdownUri": "http://127.0.0.1:19002/admin/shutdown",
"state": "ACTIVE",
"versionUri": "http://127.0.0.1:19002/admin/version"
}
But I got:
{
"cc": {
"statsUri": "http://127.0.0.1:19002/admin/cluster/cc/stats",
"configUri": "http://127.0.0.1:19002/admin/cluster/cc/config",
"threadDumpUri": "http://127.0.0.1:19002/admin/cluster/cc/threaddump"
},
"replicationUri": "http://127.0.0.1:19002/admin/cluster/replication",
"fullShutdownUri": "http://127.0.0.1:19002/admin/shutdown?all=true",
"state": "ACTIVE",
"metadata_node": "asterix_nc1",
"shutdownUri": "http://127.0.0.1:19002/admin/shutdown",
"versionUri": "http://127.0.0.1:19002/admin/version",
"config": {
"feed.port": 19003,
"txn.log.checkpoint.history": 0,
"compiler.joinmemory": 163840,
"feed.central.manager.port": 4500,
"compiler.sortmemory": 327680,
"node.partitions": {
"asterix_nc1": [
{
"partitionId": 0,
"IODeviceNum": 0,
"active": true,
"activeNodeId": "asterix_nc1",
"nodeId": "asterix_nc1"
},
{
"partitionId": 1,
"IODeviceNum": 1,
"active": true,
"activeNodeId": "asterix_nc1",
"nodeId": "asterix_nc1"
}
],
"asterix_nc2": [
{
"partitionId": 2,
"IODeviceNum": 0,
"active": true,
"activeNodeId": "asterix_nc2",
"nodeId": "asterix_nc2"
},
{
"partitionId": 3,
"IODeviceNum": 1,
"active": true,
"activeNodeId": "asterix_nc2",
"nodeId": "asterix_nc2"
}
]
},
"txn.lock.shrinktimer": 5000,
"metadata.node": "asterix_nc1",
"metadata.port": 0,
"log.level": "INFO",
"api.port": 19002,
"cluster.partitions": {
"0": {
"partitionId": 0,
"IODeviceNum": 0,
"active": true,
"activeNodeId": "asterix_nc1",
"nodeId": "asterix_nc1"
},
"1": {
"partitionId": 1,
"IODeviceNum": 1,
"active": true,
"activeNodeId": "asterix_nc1",
"nodeId": "asterix_nc1"
},
"2": {
"partitionId": 2,
"IODeviceNum": 0,
"active": true,
"activeNodeId": "asterix_nc2",
"nodeId": "asterix_nc2"
},
"3": {
"partitionId": 3,
"IODeviceNum": 1,
"active": true,
"activeNodeId": "asterix_nc2",
"nodeId": "asterix_nc2"
}
},
"storage.buffercache.size": 33554432,
"web.port": 19001,
"txn.log.checkpoint.lsnthreshold": 67108864,
"txn.log.partitionsize": 268435456,
"compiler.groupmemory": 163840,
"max.wait.active.cluster": 60,
"metadata.registration.timeout.secs": 60,
"storage.lsm.bloomfilter.falsepositiverate": 0.01,
"plot.activate": false,
"storage.memorycomponent.globalbudget": 536870912,
"web.queryinterface.port": 19006,
"txn.log.buffer.numpages": 8,
"web.secondary.port": 19005,
"feed.memory.global.budget": 67108864,
"compiler.pregelix.home": "~/pregelix",
"compiler.framesize": 32768,
"feed.memory.available.wait.timeout": 10,
"core.dump.paths": {},
"txn.log.buffer.pagesize": 131072,
"txn.job.recovery.memorysize": 67108864,
"storage.metadata.memorycomponent.numpages": 256,
"metadata.partition": {
"partitionId": 0,
"IODeviceNum": 0,
"active": true,
"activeNodeId": "asterix_nc1",
"nodeId": "asterix_nc1"
},
"txn.log.checkpoint.pollfrequency": 120,
"feed.pending.work.threshold": 50,
"feed.max.threshold.period": 5,
"storage.buffercache.maxopenfiles": 2147483647,
"txn.lock.escalationthreshold": 1000,
"txn.lock.timeout.waitthreshold": 60000,
"storage.memorycomponent.numcomponents": 2,
"storage.buffercache.pagesize": 32768,
"storage.memorycomponent.numpages": 8,
"transaction.log.dirs": {
"asterix_nc1": "target/txnLogDir/asterix_nc1",
"asterix_nc2": "target/txnLogDir/asterix_nc2"
},
"txn.commitprofiler.reportinterval": 5,
"metadata.callback.port": 0,
"txn.lock.timeout.sweepthreshold": 10000,
"storage.memorycomponent.pagesize": 131072,
"node.stores": {
"asterix_nc1": [
"iodevice0",
"iodevice1"
],
"asterix_nc2": [
"iodevice0",
"iodevice1"
]
}
},
"diagnosticsUri": "http://127.0.0.1:19002/admin/diagnostics",
"ncs": [
{
"partitions": [
{
"partition_id": "partition_0",
"active": true
},
{
"partition_id": "partition_1",
"active": true
}
],
"statsUri":
"http://127.0.0.1:19002/admin/cluster/node/asterix_nc1/stats",
"configUri":
"http://127.0.0.1:19002/admin/cluster/node/asterix_nc1/config",
"state": "ACTIVE",
"node_id": "asterix_nc1",
"threadDumpUri":
"http://127.0.0.1:19002/admin/cluster/node/asterix_nc1/threaddump"
},
{
"partitions": [
{
"partition_id": "partition_2",
"active": true
},
{
"partition_id": "partition_3",
"active": true
}
],
"statsUri":
"http://127.0.0.1:19002/admin/cluster/node/asterix_nc2/stats",
"configUri":
"http://127.0.0.1:19002/admin/cluster/node/asterix_nc2/config",
"state": "ACTIVE",
"node_id": "asterix_nc2",
"threadDumpUri":
"http://127.0.0.1:19002/admin/cluster/node/asterix_nc2/threaddump"
}
]
}
-----Original Messages-----
From: "李文海" <[email protected]>
Sent Time: 2016-11-02 11:13:58 (Wednesday)
To: [email protected]
Cc:
Subject: Re: Re: Question about the execution test issue.
Hi, Till.
This is a local error. Since it didn't pass the execution test, I have not
push it onto gerrit.
I remembered that before it got error, I have added org.reflection into
asterix-app/pom.xml. But after that,
I removed this dependency, but the error always happens. I addition, in another
local branch, this error
does not exist. I am very confused what on earth happened on my local branch.
Best,
Wenhai
> -----Original Messages-----
> From: "Till Westmann" <[email protected]>
> Sent Time: Wednesday, November 2, 2016
> To: [email protected]
> Cc:
> Subject: Re: Question about the execution test issue.
>
> Hi Wenhai,
>
> these are relatively simple tests to validate the result of some HTTP
> APIs that provide information
> on the status of the cluster. AFAIK They seem to be working reliably in
> the regression tests.
> Not sure why that would change in your branch ..
> Did you get those errors locally or on Jenkins?
>
> Cheers,
> Till
>
> On 1 Nov 2016, at 8:51, 李文海 wrote:
>
> > Hi, all
> > In the recent merging from merge to one local branch, I got the
> > following error that should not to be. Who knows why this happened?
> >
> > Best,
> > Wenhai
> >
> > Hi, all
> > Yestoday, after I applying the latest master onto the branch. I got a
> > series of error (13 testcases) that should not be posed by my branch,
> > like the following:
> > {noformat}
> > java.lang.Exception: Test
> > "src/test/resources/runtimets/queries/api/cluster_state_1/cluster_state_1.1.cstate.aql"
> >
> > FAILED!
> > at
> > org.apache.asterix.test.aql.TestExecutor.executeTest(TestExecutor.java:1028)
> > at
> > org.apache.asterix.test.runtime.ExecutionTest.test(ExecutionTest.java:127)
> > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> > at
> > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> > at
> > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> > at java.lang.reflect.Method.invoke(Method.java:497)
> > at
> > org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
> > at
> > org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
> > at
> > org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
> > at
> > org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
> > at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325)
> > at
> > org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78)
> > at
> > org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57)
> > at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
> > at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
> > at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
> > at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
> > at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
> > at org.junit.runners.ParentRunner.run(ParentRunner.java:363)
> > at org.junit.runners.Suite.runChild(Suite.java:128)
> > at org.junit.runners.Suite.runChild(Suite.java:27)
> > at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
> > at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
> > at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
> > at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
> > at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
> > at
> > org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
> > at
> > org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
> > at org.junit.runners.ParentRunner.run(ParentRunner.java:363)
> > at
> > org.eclipse.jdt.internal.junit4.runner.JUnit4TestReference.run(JUnit4TestReference.java:86)
> > at
> > org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:38)
> > at
> > org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:459)
> > at
> > org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:675)
> > at
> > org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:382)
> > at
> > org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:192)
> > Caused by: org.apache.asterix.test.base.ComparisonException: Result
> > for
> > src/test/resources/runtimets/queries/api/cluster_state_1/cluster_state_1.1.cstate.aql
> >
> > changed at line 3:
> > < "configUri":
> > "http://127.0.0.1:19002/admin/cluster/cc/config",
> >> "statsUri": "http://127.0.0.1:19002/admin/cluster/cc/stats",
> > at
> > org.apache.asterix.test.aql.TestExecutor.throwLineChanged(TestExecutor.java:203)
> > at
> > org.apache.asterix.test.aql.TestExecutor.runScriptAndCompareWithResult(TestExecutor.java:163)
> > at
> > org.apache.asterix.test.aql.TestExecutor.executeTest(TestExecutor.java:870)
> > at
> > org.apache.asterix.test.aql.TestExecutor.executeTest(TestExecutor.java:1000)
> > ... 34 more
> >
> > {noformat}