asterixdb-dev mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From 李文海 <...@whu.edu.cn>
Subject Fw: Re: Re: Question about the execution test issue.
Date Wed, 02 Nov 2016 03:16:01 GMT
BTW, the expected result:

{
    "cc": {
        "configUri": "http://127.0.0.1:19002/admin/cluster/cc/config",
        "statsUri": "http://127.0.0.1:19002/admin/cluster/cc/stats",
        "threadDumpUri": "http://127.0.0.1:19002/admin/cluster/cc/threaddump"
    },
    "config": {
        "api.port": 19002,
        "cluster.partitions": {
            "0": "ID:0, Original Node: asterix_nc1, IODevice: 0, Active Node: asterix_nc1",
            "1": "ID:1, Original Node: asterix_nc1, IODevice: 1, Active Node: asterix_nc1",
            "2": "ID:2, Original Node: asterix_nc2, IODevice: 0, Active Node: asterix_nc2",
            "3": "ID:3, Original Node: asterix_nc2, IODevice: 1, Active Node: asterix_nc2"
        },
        "compiler.framesize": 32768,
        "compiler.groupmemory": 163840,
        "compiler.joinmemory": 163840,
        "compiler.pregelix.home": "~/pregelix",
        "compiler.sortmemory": 327680,
        "core.dump.paths": {},
        "feed.central.manager.port": 4500,
        "feed.max.threshold.period": 5,
        "feed.memory.available.wait.timeout": 10,
        "feed.memory.global.budget": 67108864,
        "feed.pending.work.threshold": 50,
        "feed.port": 19003,
        "instance.name": null,
        "log.level": "INFO",
        "max.wait.active.cluster": 60,
        "metadata.callback.port": 0,
        "metadata.node": "asterix_nc1",
        "metadata.partition": "ID:0, Original Node: asterix_nc1, IODevice: 0, Active Node:
asterix_nc1",
        "metadata.port": 0,
        "metadata.registration.timeout.secs": 60,
        "node.partitions": {
            "asterix_nc1": [
                "ID:0, Original Node: asterix_nc1, IODevice: 0, Active Node: asterix_nc1",
                "ID:1, Original Node: asterix_nc1, IODevice: 1, Active Node: asterix_nc1"
            ],
            "asterix_nc2": [
                "ID:2, Original Node: asterix_nc2, IODevice: 0, Active Node: asterix_nc2",
                "ID:3, Original Node: asterix_nc2, IODevice: 1, Active Node: asterix_nc2"
            ]
        },
        "node.stores": {
            "asterix_nc1": [
                "iodevice0",
                "iodevice1"
            ],
            "asterix_nc2": [
                "iodevice0",
                "iodevice1"
            ]
        },
        "plot.activate": false,
        "storage.buffercache.maxopenfiles": 2147483647,
        "storage.buffercache.pagesize": 32768,
        "storage.buffercache.size": 33554432,
        "storage.lsm.bloomfilter.falsepositiverate": 0.01,
        "storage.memorycomponent.globalbudget": 536870912,
        "storage.memorycomponent.numcomponents": 2,
        "storage.memorycomponent.numpages": 8,
        "storage.memorycomponent.pagesize": 131072,
        "storage.metadata.memorycomponent.numpages": 256,
        "transaction.log.dirs": {
            "asterix_nc1": "target/txnLogDir/asterix_nc1",
            "asterix_nc2": "target/txnLogDir/asterix_nc2"
        },
        "txn.commitprofiler.reportinterval": 5,
        "txn.job.recovery.memorysize": 67108864,
        "txn.lock.escalationthreshold": 1000,
        "txn.lock.shrinktimer": 5000,
        "txn.lock.timeout.sweepthreshold": 10000,
        "txn.lock.timeout.waitthreshold": 60000,
        "txn.log.buffer.numpages": 8,
        "txn.log.buffer.pagesize": 131072,
        "txn.log.checkpoint.history": 0,
        "txn.log.checkpoint.lsnthreshold": 67108864,
        "txn.log.checkpoint.pollfrequency": 120,
        "txn.log.partitionsize": 268435456,
        "web.port": 19001,
        "web.queryinterface.port": 19006,
        "web.secondary.port": 19005
    },
    "diagnosticsUri": "http://127.0.0.1:19002/admin/diagnostics",
    "fullShutdownUri": "http://127.0.0.1:19002/admin/shutdown?all=true",
    "metadata_node": "asterix_nc1",
    "ncs": [
        {
            "configUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc1/config",
            "node_id": "asterix_nc1",
            "partitions": [
                {
                    "active": true,
                    "partition_id": "partition_0"
                },
                {
                    "active": true,
                    "partition_id": "partition_1"
                }
            ],
            "state": "ACTIVE",
            "statsUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc1/stats",
            "threadDumpUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc1/threaddump"
        },
        {
            "configUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc2/config",
            "node_id": "asterix_nc2",
            "partitions": [
                {
                    "active": true,
                    "partition_id": "partition_2"
                },
                {
                    "active": true,
                    "partition_id": "partition_3"
                }
            ],
            "state": "ACTIVE",
            "statsUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc2/stats",
            "threadDumpUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc2/threaddump"
        }
    ],
    "replicationUri": "http://127.0.0.1:19002/admin/cluster/replication",
    "shutdownUri": "http://127.0.0.1:19002/admin/shutdown",
    "state": "ACTIVE",
    "versionUri": "http://127.0.0.1:19002/admin/version"
}

But I got:

{
    "cc": {
        "statsUri": "http://127.0.0.1:19002/admin/cluster/cc/stats",
        "configUri": "http://127.0.0.1:19002/admin/cluster/cc/config",
        "threadDumpUri": "http://127.0.0.1:19002/admin/cluster/cc/threaddump"
    },
    "replicationUri": "http://127.0.0.1:19002/admin/cluster/replication",
    "fullShutdownUri": "http://127.0.0.1:19002/admin/shutdown?all=true",
    "state": "ACTIVE",
    "metadata_node": "asterix_nc1",
    "shutdownUri": "http://127.0.0.1:19002/admin/shutdown",
    "versionUri": "http://127.0.0.1:19002/admin/version",
    "config": {
        "feed.port": 19003,
        "txn.log.checkpoint.history": 0,
        "compiler.joinmemory": 163840,
        "feed.central.manager.port": 4500,
        "compiler.sortmemory": 327680,
        "node.partitions": {
            "asterix_nc1": [
                {
                    "partitionId": 0,
                    "IODeviceNum": 0,
                    "active": true,
                    "activeNodeId": "asterix_nc1",
                    "nodeId": "asterix_nc1"
                },
                {
                    "partitionId": 1,
                    "IODeviceNum": 1,
                    "active": true,
                    "activeNodeId": "asterix_nc1",
                    "nodeId": "asterix_nc1"
                }
            ],
            "asterix_nc2": [
                {
                    "partitionId": 2,
                    "IODeviceNum": 0,
                    "active": true,
                    "activeNodeId": "asterix_nc2",
                    "nodeId": "asterix_nc2"
                },
                {
                    "partitionId": 3,
                    "IODeviceNum": 1,
                    "active": true,
                    "activeNodeId": "asterix_nc2",
                    "nodeId": "asterix_nc2"
                }
            ]
        },
        "txn.lock.shrinktimer": 5000,
        "metadata.node": "asterix_nc1",
        "metadata.port": 0,
        "log.level": "INFO",
        "api.port": 19002,
        "cluster.partitions": {
            "0": {
                "partitionId": 0,
                "IODeviceNum": 0,
                "active": true,
                "activeNodeId": "asterix_nc1",
                "nodeId": "asterix_nc1"
            },
            "1": {
                "partitionId": 1,
                "IODeviceNum": 1,
                "active": true,
                "activeNodeId": "asterix_nc1",
                "nodeId": "asterix_nc1"
            },
            "2": {
                "partitionId": 2,
                "IODeviceNum": 0,
                "active": true,
                "activeNodeId": "asterix_nc2",
                "nodeId": "asterix_nc2"
            },
            "3": {
                "partitionId": 3,
                "IODeviceNum": 1,
                "active": true,
                "activeNodeId": "asterix_nc2",
                "nodeId": "asterix_nc2"
            }
        },
        "storage.buffercache.size": 33554432,
        "web.port": 19001,
        "txn.log.checkpoint.lsnthreshold": 67108864,
        "txn.log.partitionsize": 268435456,
        "compiler.groupmemory": 163840,
        "max.wait.active.cluster": 60,
        "metadata.registration.timeout.secs": 60,
        "storage.lsm.bloomfilter.falsepositiverate": 0.01,
        "plot.activate": false,
        "storage.memorycomponent.globalbudget": 536870912,
        "web.queryinterface.port": 19006,
        "txn.log.buffer.numpages": 8,
        "web.secondary.port": 19005,
        "feed.memory.global.budget": 67108864,
        "compiler.pregelix.home": "~/pregelix",
        "compiler.framesize": 32768,
        "feed.memory.available.wait.timeout": 10,
        "core.dump.paths": {},
        "txn.log.buffer.pagesize": 131072,
        "txn.job.recovery.memorysize": 67108864,
        "storage.metadata.memorycomponent.numpages": 256,
        "metadata.partition": {
            "partitionId": 0,
            "IODeviceNum": 0,
            "active": true,
            "activeNodeId": "asterix_nc1",
            "nodeId": "asterix_nc1"
        },
        "txn.log.checkpoint.pollfrequency": 120,
        "feed.pending.work.threshold": 50,
        "feed.max.threshold.period": 5,
        "storage.buffercache.maxopenfiles": 2147483647,
        "txn.lock.escalationthreshold": 1000,
        "txn.lock.timeout.waitthreshold": 60000,
        "storage.memorycomponent.numcomponents": 2,
        "storage.buffercache.pagesize": 32768,
        "storage.memorycomponent.numpages": 8,
        "transaction.log.dirs": {
            "asterix_nc1": "target/txnLogDir/asterix_nc1",
            "asterix_nc2": "target/txnLogDir/asterix_nc2"
        },
        "txn.commitprofiler.reportinterval": 5,
        "metadata.callback.port": 0,
        "txn.lock.timeout.sweepthreshold": 10000,
        "storage.memorycomponent.pagesize": 131072,
        "node.stores": {
            "asterix_nc1": [
                "iodevice0",
                "iodevice1"
            ],
            "asterix_nc2": [
                "iodevice0",
                "iodevice1"
            ]
        }
    },
    "diagnosticsUri": "http://127.0.0.1:19002/admin/diagnostics",
    "ncs": [
        {
            "partitions": [
                {
                    "partition_id": "partition_0",
                    "active": true
                },
                {
                    "partition_id": "partition_1",
                    "active": true
                }
            ],
            "statsUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc1/stats",
            "configUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc1/config",
            "state": "ACTIVE",
            "node_id": "asterix_nc1",
            "threadDumpUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc1/threaddump"
        },
        {
            "partitions": [
                {
                    "partition_id": "partition_2",
                    "active": true
                },
                {
                    "partition_id": "partition_3",
                    "active": true
                }
            ],
            "statsUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc2/stats",
            "configUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc2/config",
            "state": "ACTIVE",
            "node_id": "asterix_nc2",
            "threadDumpUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc2/threaddump"
        }
    ]
}


-----Original Messages-----
From: "李文海" <lwh@whu.edu.cn>
Sent Time: 2016-11-02 11:13:58 (Wednesday)
To: dev@asterixdb.apache.org
Cc: 
Subject: Re: Re: Question about the execution test issue.

Hi, Till.
    This is a local error. Since it didn't pass the execution test, I have not push it onto
gerrit.
I remembered that before it got error, I have added org.reflection into asterix-app/pom.xml.
But after that,
I removed this dependency, but the error always happens. I addition, in another local branch,
this error
does not exist. I am very confused what on earth happened on my local branch.
Best,
Wenhai


> -----Original Messages-----
> From: "Till Westmann" <tillw@apache.org>
> Sent Time: Wednesday, November 2, 2016
> To: dev@asterixdb.apache.org
> Cc: 
> Subject: Re: Question about the execution test issue.
> 
> Hi Wenhai,
> 
> these are relatively simple tests to validate the result of some HTTP 
> APIs that provide information
> on the status of the cluster. AFAIK They seem to be working reliably in 
> the regression tests.
> Not sure why that would change in your branch ..
> Did you get those errors locally or on Jenkins?
> 
> Cheers,
> Till
> 
> On 1 Nov 2016, at 8:51, 李文海 wrote:
> 
> > Hi, all
> >     In the recent merging from merge to one local branch, I got the 
> > following error that should not to be. Who knows why this happened?
> >
> > Best,
> > Wenhai
> >
> > Hi, all
> > Yestoday, after I applying the latest master onto the branch. I got a 
> > series of error (13 testcases) that should not be posed by my branch, 
> > like the following:
> > {noformat}
> > java.lang.Exception: Test 
> > "src/test/resources/runtimets/queries/api/cluster_state_1/cluster_state_1.1.cstate.aql"

> > FAILED!
> > 	at 
> > org.apache.asterix.test.aql.TestExecutor.executeTest(TestExecutor.java:1028)
> > 	at 
> > org.apache.asterix.test.runtime.ExecutionTest.test(ExecutionTest.java:127)
> > 	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> > 	at 
> > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> > 	at 
> > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> > 	at java.lang.reflect.Method.invoke(Method.java:497)
> > 	at 
> > org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
> > 	at 
> > org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
> > 	at 
> > org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
> > 	at 
> > org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
> > 	at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325)
> > 	at 
> > org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78)
> > 	at 
> > org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57)
> > 	at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
> > 	at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
> > 	at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
> > 	at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
> > 	at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
> > 	at org.junit.runners.ParentRunner.run(ParentRunner.java:363)
> > 	at org.junit.runners.Suite.runChild(Suite.java:128)
> > 	at org.junit.runners.Suite.runChild(Suite.java:27)
> > 	at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
> > 	at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
> > 	at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
> > 	at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
> > 	at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
> > 	at 
> > org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
> > 	at 
> > org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
> > 	at org.junit.runners.ParentRunner.run(ParentRunner.java:363)
> > 	at 
> > org.eclipse.jdt.internal.junit4.runner.JUnit4TestReference.run(JUnit4TestReference.java:86)
> > 	at 
> > org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:38)
> > 	at 
> > org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:459)
> > 	at 
> > org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:675)
> > 	at 
> > org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:382)
> > 	at 
> > org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:192)
> > Caused by: org.apache.asterix.test.base.ComparisonException: Result 
> > for 
> > src/test/resources/runtimets/queries/api/cluster_state_1/cluster_state_1.1.cstate.aql

> > changed at line 3:
> > <         "configUri": 
> > "http://127.0.0.1:19002/admin/cluster/cc/config",
> >>         "statsUri": "http://127.0.0.1:19002/admin/cluster/cc/stats",
> > 	at 
> > org.apache.asterix.test.aql.TestExecutor.throwLineChanged(TestExecutor.java:203)
> > 	at 
> > org.apache.asterix.test.aql.TestExecutor.runScriptAndCompareWithResult(TestExecutor.java:163)
> > 	at 
> > org.apache.asterix.test.aql.TestExecutor.executeTest(TestExecutor.java:870)
> > 	at 
> > org.apache.asterix.test.aql.TestExecutor.executeTest(TestExecutor.java:1000)
> > 	... 34 more
> >
> > {noformat}


Mime
View raw message