I started thinking whether passing the test of Mr. Jenkins is a mindless 
dreaming….:)  What does the “TOX” say? :)

Shixiong



Begin forwarded message:

> From: Shixiong Shang <sparkofwisdom.cl...@gmail.com>
> Subject: Re: [openstack-dev] [Neutron] tox run forever
> Date: February 27, 2014 at 9:41:34 PM EST
> To: "OpenStack Development Mailing List (not for usage questions)" 
> <openstack-dev@lists.openstack.org>
> 
> Hi, Clark:
> 
> Thanks a lot for the prompt response! I added the OS_TEST_TIMEOUT value (300 
> sec) and was tailing the tmp file. It turned out that the TOX run stopped at 
> the following point. My machine was tossed so badly that it became 
> unresponsive and I had to hard reboot it….I am pulling my teeth off now…..Is 
> it normal to see Traceback?
> 
> 2014-02-27 21:33:51,212     INFO [neutron.api.extensions] Extension 'agent' 
> provides no backward compatibility map for extended attributes
> 2014-02-27 21:33:51,212     INFO [neutron.api.extensions] Extension 'Allowed 
> Address Pairs' provides no backward compatibility map for extended attributes
> 2014-02-27 21:33:51,212     INFO [neutron.api.extensions] Extension 'Neutron 
> Extra Route' provides no backward compatibility map for extended attributes
> 2014-02-27 21:33:51,522    ERROR 
> [neutron.api.rpc.agentnotifiers.dhcp_rpc_agent_api] No DHCP agents are 
> associated with network '397fab50-26aa-4cb7-8aa4-c4d43909a00b'. Unable to 
> send notification for 'network_create_end' with payload: {'network': 
> {'status': 'ACTIVE', 'subnets': [], 'name': 'net1', 
> 'provider:physical_network': u'physnet1', 'admin_state_up': True, 
> 'tenant_id': 'test-tenant', 'provider:network_type': 'vlan', 'shared': False, 
> 'id': '397fab50-26aa-4cb7-8aa4-c4d43909a00b', 'provider:segmentation_id': 
> 1000}}
> 2014-02-27 21:33:51,567    ERROR [neutron.api.v2.resource] create failed
> Traceback (most recent call last):
>   File "neutron/api/v2/resource.py", line 84, in resource
>     result = method(request=request, **args)
>   File "neutron/api/v2/base.py", line 347, in create
>     allow_bulk=self._allow_bulk)
>   File "neutron/api/v2/base.py", line 600, in prepare_request_body
>     raise webob.exc.HTTPBadRequest(msg)
> HTTPBadRequest: Invalid input for cidr. Reason: '10.0.2.0' isn't a recognized 
> IP subnet cidr, '10.0.2.0/32' is recommended.
> 
> 
> Thanks again!
> 
> Shixiong
> 
> 
> 
> 
> 
> Shixiong Shang
> 
> !--- Stay Hungry, Stay Foolish ---!
> 
> On Feb 27, 2014, at 8:28 PM, Clark Boylan <clark.boy...@gmail.com> wrote:
> 
>> On Thu, Feb 27, 2014 at 4:43 PM, Shixiong Shang
>> <sparkofwisdom.cl...@gmail.com> wrote:
>>> Hi, guys:
>>> 
>>> I created a fresh local repository and pulled the most recent Neutron code. 
>>> Before I put in my own code, I did TOX run. However, seems like it is stuck 
>>> to the following condition for over a hour and didn't go any further. 
>>> Yesterday, the TOX had been running with a fresh copy of Neutron, but 
>>> didn't return SUCCESS after the entire night.
>>> 
>>> I assume the copy from MASTER BRANCH should already be 
>>> sanitized.....However, what I saw in the past 48 hours told me different 
>>> story. Did I do anything wrong?
>>> 
>>> 
>>> shshang@net-ubuntu2:~/github/neutron$ tox -e py27
>>> py27 create: /home/shshang/github/neutron/.tox/py27
>>> py27 installdeps: -r/home/shshang/github/neutron/requirements.txt, 
>>> -r/home/shshang/github/neutron/test-requirements.txt, setuptools_git>=0.4
>>> py27 develop-inst: /home/shshang/github/neutron
>>> py27 runtests: commands[0] | python -m neutron.openstack.common.lockutils 
>>> python setup.py testr --slowest --testr-args=
>>> [pbr] Excluding argparse: Python 2.6 only dependency
>>> running testr
>>> running=OS_STDOUT_CAPTURE=1 OS_STDERR_CAPTURE=1 OS_LOG_CAPTURE=1 
>>> ${PYTHON:-python} -m subunit.run discover -t ./ 
>>> ${OS_TEST_PATH:-./neutron/tests/unit} --list
>>> running=OS_STDOUT_CAPTURE=1 OS_STDERR_CAPTURE=1 OS_LOG_CAPTURE=1 
>>> ${PYTHON:-python} -m subunit.run discover -t ./ 
>>> ${OS_TEST_PATH:-./neutron/tests/unit}  --load-list /tmp/tmpbZwLwg
>>> running=OS_STDOUT_CAPTURE=1 OS_STDERR_CAPTURE=1 OS_LOG_CAPTURE=1 
>>> ${PYTHON:-python} -m subunit.run discover -t ./ 
>>> ${OS_TEST_PATH:-./neutron/tests/unit}  --load-list /tmp/tmp39qJYM
>>> running=OS_STDOUT_CAPTURE=1 OS_STDERR_CAPTURE=1 OS_LOG_CAPTURE=1 
>>> ${PYTHON:-python} -m subunit.run discover -t ./ 
>>> ${OS_TEST_PATH:-./neutron/tests/unit}  --load-list /tmp/tmpppXiTc
>>> running=OS_STDOUT_CAPTURE=1 OS_STDERR_CAPTURE=1 OS_LOG_CAPTURE=1 
>>> ${PYTHON:-python} -m subunit.run discover -t ./ 
>>> ${OS_TEST_PATH:-./neutron/tests/unit}  --load-list /tmp/tmpPhJZDc
>>> 
>>> Thanks!
>>> 
>>> Shixiong
>>> _______________________________________________
>>> OpenStack-dev mailing list
>>> OpenStack-dev@lists.openstack.org
>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
>> 
>> I think there are two potential problems here. Either a test is
>> deadlocking due to something it has done or
>> neutron.openstack.common.lockutils is deadlocking. In either case
>> OS_TEST_TIMEOUT is not set in .testr.conf so the test suite will not
>> timeout individual tests if necessary. I would start by setting that
>> in the .testr.conf next to OS_STDOUT_CAPTURE and you probably want a
>> value of like 300 (that is seconds).
>> 
>> The other thing you can do to debug this is grab the subunit log file
>> out of .testrepository. While tests are running it will have a tmp
>> random generated name. After tests have run it will be moved to a file
>> named after the most recent test run eg 1 for the first run. The
>> subunit log should offer clues to what was running at the time of
>> deadlocking.
>> 
>> Clark
>> 
>> _______________________________________________
>> OpenStack-dev mailing list
>> OpenStack-dev@lists.openstack.org
>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
> 

_______________________________________________
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev

Reply via email to