Further testing shows that this only seems to affect an upgraded
agent. I can reproduce if I install agent 4.1.1, then upgrade to 4.2,
but not if I start fresh with agent 4.2. If I had to guess, the
upgrade maybe makes the local storage pool re-register(maybe it
creates a new host in the table or something like that, haven't
looked), where it looks like a duplicate.

On Mon, Sep 23, 2013 at 12:36 AM, Marcus Sorensen <shadow...@gmail.com> wrote:
> I think this does have to do with the new default storage plugin. I
> think it only affects local storage, as the agent attempts to register
> it's local storage pool uuid as a pool with the mgmt server, who says
> 'sorry, this pool already exists in the database'.
>
> On Mon, Sep 23, 2013 at 12:24 AM, Marcus Sorensen <shadow...@gmail.com> wrote:
>> +0 (binding), created 4.1.1 zone with vms and a vpc, upgraded to 4.2
>> via RPMs. Restarted vpc. Then wiped database and redeployed zone, vms
>> to test 4.2 basic functionality. All on CentOS 6.4
>>
>> The reason for the +0 is that I found a regression when going beyond
>> the basic testing and trying things like maintenance mode, I'm not
>> sure if it's related to the new storage framework or not, but it
>> definitely didn't exist in the previous release. I'm including the bug
>> below.  Basically the agent fails to connect to the management server
>> on restart because we are trying to register the same pool multiple
>> times. We should just check for the pool, match the uuid/properties,
>> and say 'success' if it already exists. There is a hokey workaround,
>> which is why I'm not going to block, if you stop libvirtd after taking
>> the host out of maintenance and wait for the agent to reconnect, then
>> you can start libvirtd again and everything will jumpstart into
>> action. That can be done in order to complete a successful upgrade,
>> and upon deploying a fresh zone I don't think anyone will notice until
>> they go to use maintenance mode.
>>
>> https://issues.apache.org/jira/browse/CLOUDSTACK-4725
>>
>> Should be easy to fix, but I'm not going to look at it right this
>> second. Maybe Edison can take a peek in the meantime.
>>
>> Looks like nobody has voted yet, I'm not sure if that means nobody has
>> tested over the weekend or if they're being more thorough. If we do
>> find that nobody has tested, and decide to fix this, I'd request we
>> pull in 39f7ddbb8f7eedb050da2991cdc1fb72a9e97f5f from 4.2-forward as
>> well.
>>
>> On Fri, Sep 20, 2013 at 9:36 PM, Animesh Chaturvedi
>> <animesh.chaturv...@citrix.com> wrote:
>>>
>>>
>>>
>>>
>>> I've created a 4.2.0 release, with the following artifacts up for a vote:
>>>
>>> Git Branch and Commit SH:
>>> https://git-wip-us.apache.org/repos/asf?p=cloudstack.git;a=shortlog;h=refs/heads/4.2
>>> Commit: 69c459342c568e2400d57ee88572b301603d8686
>>>
>>> List of changes:
>>> https://git-wip-us.apache.org/repos/asf?p=cloudstack.git;a=blob_plain;f=CHANGES;hb=4.2
>>>
>>> Source release (checksums and signatures are available at the same
>>> location):
>>> https://dist.apache.org/repos/dist/dev/cloudstack/4.2.0/
>>>
>>> PGP release keys (signed using 94BE0D7C):
>>> https://dist.apache.org/repos/dist/release/cloudstack/KEYS
>>>
>>> Testing instructions are here:
>>> https://cwiki.apache.org/confluence/display/CLOUDSTACK/Release+test+procedure
>>>
>>> Vote will be open for 72 hours (Monday 9/23 PST EOD).
>>>
>>> For sanity in tallying the vote, can PMC members please be sure to indicate 
>>> "(binding)" with their vote?
>>>
>>> [ ] +1  approve
>>> [ ] +0  no opinion
>>> [ ] -1  disapprove (and reason why)
>>>
>>>

Reply via email to