[vdsm] Cancelled: vdsm call

2014-11-25 Thread Dan Kenigsberg
BEGIN:VCALENDAR
PRODID:Zimbra-Calendar-Provider
VERSION:2.0
METHOD:CANCEL
BEGIN:VTIMEZONE
TZID:Europe/London
BEGIN:STANDARD
DTSTART:16010101T02
TZOFFSETTO:-
TZOFFSETFROM:+0100
RRULE:FREQ=YEARLY;WKST=MO;INTERVAL=1;BYMONTH=10;BYDAY=-1SU
TZNAME:GMT/BST
END:STANDARD
BEGIN:DAYLIGHT
DTSTART:16010101T01
TZOFFSETTO:+0100
TZOFFSETFROM:-
RRULE:FREQ=YEARLY;WKST=MO;INTERVAL=1;BYMONTH=3;BYDAY=-1SU
TZNAME:GMT/BST
END:DAYLIGHT
END:VTIMEZONE
BEGIN:VEVENT
UID:95fbe6ce-3950-4413-b63a-d2c5701bc3a7
SUMMARY:Cancelled: vdsm call
COMMENT:A single instance of a recurring meeting has been cancelled.
ATTENDEE;CN=vdsm-devel;ROLE=REQ-PARTICIPANT;PARTSTAT=NEEDS-ACTION;RSVP=TRUE:
 mailto:vdsm-devel@lists.fedorahosted.org
ATTENDEE;CN=Otavio Luiz Ferranti;PARTSTAT=DECLINED:mailto:otavio.ferranti@el
 dorado.org.br
ATTENDEE;CN=Saggi Mizrahi;PARTSTAT=ACCEPTED:mailto:smizr...@redhat.com
ATTENDEE;CN=Antoni Segura Puimedon;PARTSTAT=ACCEPTED:mailto:asegurap@redhat.
 com
ATTENDEE;CN=Dima Kuznetsov;PARTSTAT=ACCEPTED:mailto:dkuzn...@redhat.com
ATTENDEE;CN=Yaniv Bronheim;PARTSTAT=ACCEPTED:mailto:ybron...@redhat.com
ATTENDEE;CN=Barak Azulay;PARTSTAT=TENTATIVE:mailto:bazu...@redhat.com
ATTENDEE;CN=Francesco Romani;PARTSTAT=ACCEPTED:mailto:from...@redhat.com
ATTENDEE;CN=Federico Simoncelli;PARTSTAT=ACCEPTED:mailto:fsimo...@redhat.com
ATTENDEE;CN=Sean Cohen;PARTSTAT=TENTATIVE:mailto:sco...@redhat.com
ATTENDEE;CN=Oved Ourfalli;PARTSTAT=TENTATIVE:mailto:ov...@redhat.com
ATTENDEE;CN=Piotr Kliczewski;PARTSTAT=ACCEPTED:mailto:pklic...@redhat.com
ATTENDEE;CN=Ido Barkan;PARTSTAT=ACCEPTED:mailto:ibar...@redhat.com
ATTENDEE;CN=Adam Litke;PARTSTAT=ACCEPTED:mailto:ali...@redhat.com
ORGANIZER;CN=Dan Kenigsberg:mailto:dkeni...@redhat.com
DTSTART;TZID=Europe/London:20141125T143000
DTEND;TZID=Europe/London:20141125T153000
STATUS:CANCELLED
CLASS:PUBLIC
X-MICROSOFT-CDO-INTENDEDSTATUS:BUSY
TRANSP:OPAQUE
RECURRENCE-ID;TZID=Europe/London:20141125T143000
LAST-MODIFIED:20141125T134313Z
DTSTAMP:20141125T134313Z
SEQUENCE:3
DESCRIPTION:A single instance of the following meeting has been cancelled:\n
 \nSubject: vdsm call \nOrganizer: Dan Kenigsberg dkeni...@redhat.com \n\
 nTime: Tuesday\, November 25\, 2014\, 2:30:00 PM - 3:30:00 PM GMT +00:00 Bri
 tain\, Ireland\, Portugal\n \nInvitees: vdsm-devel@lists.fedorahosted.org\; 
 otavio.ferra...@eldorado.org.br\; smizr...@redhat.com\; asegu...@redhat.com\
 ; dkuzn...@redhat.com\; ybron...@redhat.com\; bazu...@redhat.com\; fromani@r
 edhat.com\; fsimo...@redhat.com\; sco...@redhat.com\; ov...@redhat.com ... \
 n\n\n*~*~*~*~*~*~*~*~*~*\n\nSorry for the late notice\, but I just noticed t
 hat most of us have a conflicting meeting today.
END:VEVENT
END:VCALENDAR___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


[vdsm] Cancelled: vdsm call

2014-10-14 Thread Dan Kenigsberg
BEGIN:VCALENDAR
PRODID:Zimbra-Calendar-Provider
VERSION:2.0
METHOD:CANCEL
BEGIN:VTIMEZONE
TZID:Europe/London
BEGIN:STANDARD
DTSTART:16010101T02
TZOFFSETTO:-
TZOFFSETFROM:+0100
RRULE:FREQ=YEARLY;WKST=MO;INTERVAL=1;BYMONTH=10;BYDAY=-1SU
TZNAME:GMT/BST
END:STANDARD
BEGIN:DAYLIGHT
DTSTART:16010101T01
TZOFFSETTO:+0100
TZOFFSETFROM:-
RRULE:FREQ=YEARLY;WKST=MO;INTERVAL=1;BYMONTH=3;BYDAY=-1SU
TZNAME:GMT/BST
END:DAYLIGHT
END:VTIMEZONE
BEGIN:VEVENT
UID:95fbe6ce-3950-4413-b63a-d2c5701bc3a7
SUMMARY:Cancelled: vdsm call
COMMENT:A single instance of a recurring meeting has been cancelled.
ATTENDEE;CN=vdsm-devel;ROLE=REQ-PARTICIPANT;PARTSTAT=NEEDS-ACTION;RSVP=TRUE:
 mailto:vdsm-devel@lists.fedorahosted.org
ATTENDEE;CN=Otavio Luiz Ferranti;PARTSTAT=DECLINED:mailto:otavio.ferranti@el
 dorado.org.br
ATTENDEE;CN=Saggi Mizrahi;PARTSTAT=ACCEPTED:mailto:smizr...@redhat.com
ATTENDEE;CN=Antoni Segura Puimedon;PARTSTAT=ACCEPTED:mailto:asegurap@redhat.
 com
ATTENDEE;CN=Dima Kuznetsov;PARTSTAT=ACCEPTED:mailto:dkuzn...@redhat.com
ATTENDEE;CN=Yaniv Bronheim;PARTSTAT=ACCEPTED:mailto:ybron...@redhat.com
ATTENDEE;CN=Barak Azulay;PARTSTAT=TENTATIVE:mailto:bazu...@redhat.com
ATTENDEE;CN=Francesco Romani;PARTSTAT=ACCEPTED:mailto:from...@redhat.com
ATTENDEE;CN=Federico Simoncelli;PARTSTAT=ACCEPTED:mailto:fsimo...@redhat.com
ATTENDEE;CN=Sean Cohen;PARTSTAT=TENTATIVE:mailto:sco...@redhat.com
ATTENDEE;CN=Oved Ourfalli;PARTSTAT=TENTATIVE:mailto:ov...@redhat.com
ATTENDEE;CN=Piotr Kliczewski;PARTSTAT=ACCEPTED:mailto:pklic...@redhat.com
ATTENDEE;CN=Ido Barkan;PARTSTAT=ACCEPTED:mailto:ibar...@redhat.com
ORGANIZER;CN=Dan Kenigsberg:mailto:dkeni...@redhat.com
DTSTART;TZID=Europe/London:20141014T143000
DTEND;TZID=Europe/London:20141014T153000
STATUS:CANCELLED
CLASS:PUBLIC
X-MICROSOFT-CDO-INTENDEDSTATUS:BUSY
TRANSP:OPAQUE
RECURRENCE-ID;TZID=Europe/London:20141014T143000
LAST-MODIFIED:20141014T101625Z
DTSTAMP:20141014T101625Z
SEQUENCE:3
DESCRIPTION:A single instance of the following meeting has been cancelled:\n
 \nSubject: vdsm call \nOrganizer: Dan Kenigsberg dkeni...@redhat.com \n\
 nTime: Tuesday\, October 14\, 2014\, 2:30:00 PM - 3:30:00 PM GMT +00:00 Brit
 ain\, Ireland\, Portugal\n \nInvitees: vdsm-devel@lists.fedorahosted.org\; o
 tavio.ferra...@eldorado.org.br\; smizr...@redhat.com\; asegu...@redhat.com\;
  dkuzn...@redhat.com\; ybron...@redhat.com\; bazu...@redhat.com\; fromani@re
 dhat.com\; fsimo...@redhat.com\; sco...@redhat.com\; ov...@redhat.com ... \n
 \n\n*~*~*~*~*~*~*~*~*~*\n\nIf you receive this e-mail\, it means that you ar
 e invited to the \nbi-weekly phone meeting of Vdsm developers. This is an op
 portunity \nto discuss design problems\, pending patches\, and persistent bu
 gs. \n\nEven if you do not join\, you may solicit subjects for discussion on
  \nthe mailing list or over #v...@irc.freenode.net. \n\nNote the new meeting
  time and conference code. \n\nConference code: 353-86-075-901. \n\nDial-in 
 information: \nhttps://www.intercallonline.com/listNumbersByCode.action?conf
 Code=35386075901 
X-ALT-DESC;FMTTYPE=text/html:htmlbody id='htmlmode'h3A single instance
  of the following meeting has been cancelled:/h3\n\np\ntable border='0'
 \ntrth align=leftSubject:/thtdvdsm call /td/tr\ntrth align=l
 eftOrganizer:/thtdDan Kenigsberg lt\;dkeni...@redhat.comgt\; /td
 /tr\n/table\np\ntable border='0'\ntrth align=leftTime:/thtdTu
 esday\, October 14\, 2014\, 2:30:00 PM - 3:30:00 PM GMT +00:00 Britain\, Ire
 land\, Portugal\n /td/tr/table\np\ntable border='0'\ntrth align
 =leftInvitees:/thtdvdsm-devel@lists.fedorahosted.org\; otavio.ferranti@
 eldorado.org.br\; smizr...@redhat.com\; asegu...@redhat.com\; dkuznets@redha
 t.com\; ybron...@redhat.com\; bazu...@redhat.com\; from...@redhat.com\; fsim
 o...@redhat.com\; sco...@redhat.com\; ov...@redhat.com ... /td/tr\n/tab
 le\ndiv*~*~*~*~*~*~*~*~*~*/divbrIf you receive this e-mail\, it means
  that you are invited to the brbi-weekly phone meeting of Vdsm developers.
  This is an opportunity brto discuss design problems\, pending patches\, a
 nd persistent bugs. brbrEven if you do not join\, you may solicit subjec
 ts for discussion on brthe mailing list or over #v...@irc.freenode.net. b
 rbrNote the new meeting time and conference code. brbrConference code
 : 353-86-075-901. brbrDial-in information: brhttps://www.intercallonli
 ne.com/listNumbersByCode.action?confCode=35386075901/body/html
END:VEVENT
END:VCALENDAR___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] [VDSM] cleaning statistics retrieval

2014-04-10 Thread Dan Kenigsberg
On Wed, Apr 09, 2014 at 08:40:24AM -0400, Francesco Romani wrote:
 - Original Message -
  From: Dan Kenigsberg dan...@redhat.com
  To: Francesco Romani from...@redhat.com
  Cc: vdsm-devel vdsm-devel@lists.fedorahosted.org, de...@ovirt.org
  Sent: Tuesday, April 8, 2014 6:57:15 PM
  Subject: Re: cleaning statistics retrieval
  
  On Tue, Apr 08, 2014 at 06:52:50AM -0400, Francesco Romani wrote:
   Hello VDSM developers,
  
  Please use devel@ovirt, and mention vdsm at the subject. This thread
  in particular involves Engine as well.
 
 Right, I forgot. Sorry about that.
 
   I'd like to discuss and explain the plans for cleaning up Vm.getStats()
   in vdsm/virt/vm.py, and how it affects a bug we have:
   https://bugzilla.redhat.com/show_bug.cgi?id=1073478
   
   Let's start from the bug.
   
   To make a long story short, there is a (small) race in VDSM, probably
   introduced by commit
   8fedf8bde3c28edb07add23c3e9b72681cb48e49
   The race can actually be triggered only if the VM is shutting down, so it
   is not that bad.
   
   Fixing the reported issue in the specific case can be done with a trivial
   if, and that it what I did
   in my initial http://gerrit.ovirt.org/#/c/25803/1/vdsm/vm.py,cm
  
  Could you explain how an AttributeError there moved the VM to Down?
 
 This should actually be this bug of engine 
 https://bugzilla.redhat.com/show_bug.cgi?id=1072282
 if GetVmStats fails for whatever reason, engine thinks the VM is down.

Could you rename the Vdsm bug to exprese the in-vdsm problem, and make
clear that it confuses older Engines to think the Vm is Down?

 
   And this is the core of the issue.
   My initial idea is to either return success and a complete, well formed
   statistics set, or return an error.
   However current engine seems to not cope properly with this, and we cannot
   break backward compatibility.
  Would you be more precise? If getAllVmStats returns an errCode for one
  of the Vms, what happens?
 
 Of course.
 
 For GetAllVmStats, AFAIK, but please correct me if I am wrong, because I'm 
 not really expert on engine side,
 engine simply does not expects anything different from a list
 of either a RunningVmStats or an ExitedVmStats.
 
 So not sure (will verify just after this mail) if engine can cope with mixed 
 content,
 meaning [ stats, errCode, stats, stats ... ]
 
 For GetVmStats, like Michal said, the engine expects the call to succeed 
 otherwise
 it puts the VM into the Unknown state.
 
  
   
   Looks like the only way to go is to always return success and to add a
   field to describe the content of the
   statistics (partial, complete...). THis is admittedly a far cry from the
   ideal solution, but it is dictated
   by the need to preserve the compatibility with current/old engines.
  
  I don't think that I understand your suggestion, but it does not sound
  right to send a partial dictionary and to appologize about its
  paritiality elsewhere. The dictionary may be partial for engine-4.0
  yet complete for engine-3.5. It's not for Vdsm to grade its own
  output.
 
 I see your point (that's one of the reasons I'm not happy about this 
 solution).
 Please see below for the detauls.
 
   please note that I'm not really happy about this solution, but, given the
   constraint, I don't see better alternatives.
  
  Please explain the benefits of describing the partial content, as I do
  not see them.
 
 The root issue here is the getStats must always succeed, because the engine 
 doesn't
 expect otherwise and thus we guarantee this to cope with old engines;
 but inside VDSM, getStats can actually fail in a lot of places
 (like in this case getBalloonInfo)
 
 So, in VDSM we can end up with a partial response, and now the question
 is: what to do with this partial response? And if there are optional fields
 in the response, how can the engine distinguish between
 
 * full response, but with an optional field missing
 
 and
 
 * partial response (because of an exception inside VDSM),
   without some field, incidentally including an optional one


 
 ?
 
 The VDSM 'grading' was an hint from VDSM to help engine to distinguish
 between those cases.
 
 Even if we agree this grading idea is bad, the core issue remains open:
 what to do if we end up with a partial response?
 For example, let's say we handle the getBalloonInfo exception 
 (http://gerrit.ovirt.org/#/c/26539/),
 the stats object to be returned will lack
 
 * the (mandatory, expected) balloon stats
 * the (optional) migration progress - ok, bad example because this makes 
 sense only during migrations,
   but other optional fields may be added later and the issue remains
 
 Again, anyone feel free to correct me if I misunderstood something about 
 engine
 (or VDSM = engine communication) and to suggest better alternatives :\

Currently, we have way too many try-except-Exception clauses in our
code. They swallow everything: from expected libvirt errors to
unexpected syntax errors. We should eliminate

Re: [vdsm] cleaning statistics retrieval

2014-04-09 Thread Dan Kenigsberg
On Wed, Apr 09, 2014 at 01:11:22PM +0200, Michal Skrivanek wrote:
 
 On Apr 8, 2014, at 18:57 , Dan Kenigsberg dan...@redhat.com wrote:
 
  On Tue, Apr 08, 2014 at 06:52:50AM -0400, Francesco Romani wrote:
  Hello VDSM developers,
  
  Please use devel@ovirt, and mention vdsm at the subject. This thread
  in particular involves Engine as well.
  
  
  I'd like to discuss and explain the plans for cleaning up Vm.getStats()
  in vdsm/virt/vm.py, and how it affects a bug we have: 
  https://bugzilla.redhat.com/show_bug.cgi?id=1073478
  
  Let's start from the bug.
  
  To make a long story short, there is a (small) race in VDSM, probably 
  introduced by commit
  8fedf8bde3c28edb07add23c3e9b72681cb48e49
  The race can actually be triggered only if the VM is shutting down, so it 
  is not that bad.
  
  Fixing the reported issue in the specific case can be done with a trivial 
  if, and that it what I did
  in my initial http://gerrit.ovirt.org/#/c/25803/1/vdsm/vm.py,cm
  
  Could you explain how an AttributeError there moved the VM to Down?
  
  
  However, this is just a bandaid and a proper solution is needed. This is 
  the reason why
  the following versions of http://gerrit.ovirt.org/#/c/25803 changed 
  direction toward a more comprehensive
  approach.
  
  And this is the core of the issue.
  My initial idea is to either return success and a complete, well formed 
  statistics set, or return an error.
  However current engine seems to not cope properly with this, and we cannot 
  break backward compatibility.
  
  Would you be more precise? If getAllVmStats returns an errCode for one
  of the Vms, what happens?
 
 VM moves to Unknown after some time
 
  
  
  Looks like the only way to go is to always return success and to add a 
  field to describe the content of the
  statistics (partial, complete...). THis is admittedly a far cry from the 
  ideal solution, but it is dictated
  by the need to preserve the compatibility with current/old engines.
  
  I don't think that I understand your suggestion, but it does not sound
  right to send a partial dictionary and to appologize about its
  paritiality elsewhere. The dictionary may be partial for engine-4.0
  yet complete for engine-3.5. It's not for Vdsm to grade its own
  output.
  
  
  Moreover, I would like to take the chance and cleanup/refactor the 
  statistics collection. I already started
  adding the test infrastructure: http://gerrit.ovirt.org/#/c/26536/
  
  To summarize, what I suggest to do is:
  
  * fix https://bugzilla.redhat.com/show_bug.cgi?id=1073478 using a simple 
  ugly fix like the original 
   http://gerrit.ovirt.org/#/c/25803/1/vdsm/vm.py,cm (which I'll resubmit as 
  new change)
  * refactor and clean getStats() and friends
  * on the cleaner base, properly fix the statistics collection by let 
  getStats() always succeed and return
   possibly partial stats, with a new field describing the content
  
  please note that I'm not really happy about this solution, but, given the 
  constraint, I don't see better
  alternatives.
  
  Please explain the benefits of describing the partial content, as I do
  not see them.
 
 the main reason for me is that currently if we fail to fetch some parts of 
 the stats we still return what we could get. E.g.
 try:
 stats.update(self.guestAgent.getGuestInfo())
 except Exception:
 return stats
 
 currently in engine we can only tell it's partial by not seeing the missing 
 bits…
 it makes sense since for lifecycle decisions we care for the VM status, but 
 we don't necessarily care for other stuff 

Engine knows that the data is partial, since it sees only parts of it.
What is the benefit of Vdsm declaring it as partial?
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] vdsm sync meeting minutes April 1st, 2014

2014-04-09 Thread Dan Kenigsberg
On Wed, Apr 02, 2014 at 09:29:09AM +0100, dan...@redhat.com wrote:
 
 - We had a very (too) heated debate about ignoring failures of
   setDomainRegularRole() in http://gerrit.ovirt.org/24495/ and
   http://gerrit.ovirt.org/25424.
 
   The pain point is that relying on domain role (master/regular) is
   faulty by design. We cannot avoid the cases where a pool has more than
   one domain with a master role written in its metadata.
 
   One side argued that oVirt should be fixed to handle this unescapable
   truth, or at least enumerate the places where Vdsm and Engine, both
   current and old, depend on master role uniqueness.
 
   The other side argued that this is not a priority task, and that we
   should try to garbage-collect known-bad master roles as a courtesy
   to people digging into domain metadata, and as a means to lessen the
   problem until we kill the pool concept in the upcoming version.
 
   I hope that I present the debate fairly enough.

In order to move these two patches forward, how about:

- Limit the usage of the catching-all except Exception and replace
  it with swalling only the expected IO error

- Add a comment about setDomainRegularRole() being called only as a
  courtesy garbage-collection attempt.

- Conduct a survey on whether migrateMaster is used by anyone. No
  supported Engine has it, but I there was a suggestion that it was
  still expected via the command line.

What do you think?
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] cleaning statistics retrieval

2014-04-09 Thread Dan Kenigsberg
On Wed, Apr 09, 2014 at 02:32:57PM +0200, Michal Skrivanek wrote:
 
 On Apr 9, 2014, at 14:21 , Dan Kenigsberg dan...@redhat.com wrote:
 
  On Wed, Apr 09, 2014 at 01:11:22PM +0200, Michal Skrivanek wrote:
  
  On Apr 8, 2014, at 18:57 , Dan Kenigsberg dan...@redhat.com wrote:
  
  On Tue, Apr 08, 2014 at 06:52:50AM -0400, Francesco Romani wrote:
  Hello VDSM developers,
  
  Please use devel@ovirt, and mention vdsm at the subject. This thread
  in particular involves Engine as well.
  
  
  I'd like to discuss and explain the plans for cleaning up Vm.getStats()
  in vdsm/virt/vm.py, and how it affects a bug we have: 
  https://bugzilla.redhat.com/show_bug.cgi?id=1073478
  
  Let's start from the bug.
  
  To make a long story short, there is a (small) race in VDSM, probably 
  introduced by commit
  8fedf8bde3c28edb07add23c3e9b72681cb48e49
  The race can actually be triggered only if the VM is shutting down, so 
  it is not that bad.
  
  Fixing the reported issue in the specific case can be done with a 
  trivial if, and that it what I did
  in my initial http://gerrit.ovirt.org/#/c/25803/1/vdsm/vm.py,cm
  
  Could you explain how an AttributeError there moved the VM to Down?
  
  
  However, this is just a bandaid and a proper solution is needed. This is 
  the reason why
  the following versions of http://gerrit.ovirt.org/#/c/25803 changed 
  direction toward a more comprehensive
  approach.
  
  And this is the core of the issue.
  My initial idea is to either return success and a complete, well formed 
  statistics set, or return an error.
  However current engine seems to not cope properly with this, and we 
  cannot break backward compatibility.
  
  Would you be more precise? If getAllVmStats returns an errCode for one
  of the Vms, what happens?
  
  VM moves to Unknown after some time
  
  
  
  Looks like the only way to go is to always return success and to add a 
  field to describe the content of the
  statistics (partial, complete...). THis is admittedly a far cry from the 
  ideal solution, but it is dictated
  by the need to preserve the compatibility with current/old engines.
  
  I don't think that I understand your suggestion, but it does not sound
  right to send a partial dictionary and to appologize about its
  paritiality elsewhere. The dictionary may be partial for engine-4.0
  yet complete for engine-3.5. It's not for Vdsm to grade its own
  output.
  
  
  Moreover, I would like to take the chance and cleanup/refactor the 
  statistics collection. I already started
  adding the test infrastructure: http://gerrit.ovirt.org/#/c/26536/
  
  To summarize, what I suggest to do is:
  
  * fix https://bugzilla.redhat.com/show_bug.cgi?id=1073478 using a simple 
  ugly fix like the original 
  http://gerrit.ovirt.org/#/c/25803/1/vdsm/vm.py,cm (which I'll resubmit 
  as new change)
  * refactor and clean getStats() and friends
  * on the cleaner base, properly fix the statistics collection by let 
  getStats() always succeed and return
  possibly partial stats, with a new field describing the content
  
  please note that I'm not really happy about this solution, but, given 
  the constraint, I don't see better
  alternatives.
  
  Please explain the benefits of describing the partial content, as I do
  not see them.
  
  the main reason for me is that currently if we fail to fetch some parts of 
  the stats we still return what we could get. E.g.
 try:
 stats.update(self.guestAgent.getGuestInfo())
 except Exception:
 return stats
  
  currently in engine we can only tell it's partial by not seeing the 
  missing bits…
  it makes sense since for lifecycle decisions we care for the VM status, 
  but we don't necessarily care for other stuff 
  
  Engine knows that the data is partial, since it sees only parts of it.
  What is the benefit of Vdsm declaring it as partial?
 
 there are a lot of variable fields so it's not that easy to say if it's 
 partial or not. It does make sense to let someone know that there is an 
 issue and you're getting partial data, yet you don't want to make any 
 automatic actions...

But Vdsm cannot make this decision. Soon, Vdsm is to report the host's
boot time. Now assume that Vdsm fails to do so. Is the stats partial?
It's partial for engine-3.5, but it's complete for engine-3.4.

Vdsm should tell as much of the truth that it knows.

We could extend the alerts mechanism to report non-lethal errors in
getVmStats (we currently have it only in for storage domain status),
where Engine is told what's missing and why. I'm not sure if this is
really needed, though.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] Help Needed

2014-04-08 Thread Dan Kenigsberg
On Tue, Apr 08, 2014 at 02:02:37AM -0400, Darshan Narayana Murthy wrote:
 
 
 - Original Message -
 From: Darshan Narayana Murthy dnara...@redhat.com
 To: vdsm-de...@fedorahosted.org
 Sent: Tuesday, April 8, 2014 11:21:27 AM
 Subject: Re: [vdsm] Help Needed
 
 
 
 - Original Message -
 From: Dan Kenigsberg dan...@redhat.com
 To: Darshan Narayana Murthy dnara...@redhat.com
 Sent: Monday, April 7, 2014 4:53:30 PM
 Subject: Re: Help Needed
 
 On Mon, Apr 07, 2014 at 06:52:36AM -0400, Darshan Narayana Murthy wrote:
  Hi Dan,
  
 I sent a patch for vdsm to get the gluster volume  capacity
  statistics using libgf api ( patch : http://gerrit.ovirt.org/#/c/26343 ),
  This patch requires glusterfs-devel package for build.
  
 It looks like jenkins does not have this package and so the build
  for this patch is failing. Does jenkins automatically pull the required
  package or is there anything to be done to get this package in jenkins ?.
  
  Can you please help me to resolve this.
 
 Generally speaking, if you want a new package installed, you should ask
 that on in...@ovirt.org.
 
 However, I am not at all happy with adding C code into Vdsm. What is it?
 Python binding for glfs_statvfs ? Could this be implemented elsewhere
 (such as an independent python-glfs package)?
 
 Dan.
 
 
 Hi,
 
We are making use of libgf-api for getting the statistics related
 to a glusterfs volume, as it is more efficient than we mounting a volume
 and getting the statistics.
 
libgfapi is a c api. Initially we tried using ctypes to wrap the
 required functions in libgfapi. But because of a limitation in glusterfs
 when these functions were invoked through supervdsm it would break.

I hope Toni can help out here, he found outr that when using threads,
you must declare the function prototype explicitly, but I'm not at all
sure this is your issue.

Issue with the above approach was:
 https://lists.fedorahosted.org/pipermail/vdsm-devel/2013-August/002537.html
 
So we thought of having an extension module that makes use of libgfapi
 and provides the statistics, which can be used in vdsm.
 
what would be the better approach to resolve this ? Please provide us
 your suggestions.

In my opionion, the python module that you suggest makes sense - but it
should be part of libgf - it is their's python binding (or part
thereof).
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] Modeling graphics framebuffer device in VDSM

2014-04-08 Thread Dan Kenigsberg
On Fri, Mar 28, 2014 at 08:06:17AM -0400, Frantisek Kobzik wrote:
 Dear VDSM devels,
 
 I've been working on refactoring graphics devices in engine and VDSM for some
 time now and I'd like know your opinion of that.
 
 The aim of this refactoring is to model graphics framebuffer (SPICE, VNC) as
 device in the engine and VDSM. This which is quite natural since libvirt 
 treats
 graphics as a device and we have some kind of devices infrastructure in both
 projects. Another advantage (and actually the main reason for refactoring)  is
 simplified support for multiple graphics framebuffers on a single vm.
 
 Currently, passing information about graphics from engine to VDSM is done via
 'display' param in conf. In the other direction VDSM informs the engine about
 graphics parameters ('displayPort', 'displaySecurePort', 'displayIp' and
 'displayNetwork') in conf as well.
 
 What I'd like to achieve is to encapsulate all this information in specParams
 of the new graphics device and use specParams as a place for transfering data
 about graphics device between engine and vdsm. What do you think?
 
 the draft patch is here:
 http://gerrit.ovirt.org/#/c/23555/ (it's currently marked with '-1' but it 
 puts
 some light on what the solution looks like so feel free to take a look).

Thanks for the heads up. I'd appreciate if you could split this patch to
smaller ones, and define well-named functions as much as possible. It's
not your fault, but vm.py is in such a shape that adding more complexity
into it should be done with great care.

So even if the patches are interdependent, the review process would be
simpler.

Another high-level note that I have is that you should care for
reporting displayPort on the vm level only for those Engines that care
for that, so if you have multiple graphics devices - you should not.

Regards,
Dan.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] cleaning statistics retrieval

2014-04-08 Thread Dan Kenigsberg
On Tue, Apr 08, 2014 at 06:52:50AM -0400, Francesco Romani wrote:
 Hello VDSM developers,

Please use devel@ovirt, and mention vdsm at the subject. This thread
in particular involves Engine as well.

 
 I'd like to discuss and explain the plans for cleaning up Vm.getStats()
 in vdsm/virt/vm.py, and how it affects a bug we have: 
 https://bugzilla.redhat.com/show_bug.cgi?id=1073478
 
 Let's start from the bug.
 
 To make a long story short, there is a (small) race in VDSM, probably 
 introduced by commit
 8fedf8bde3c28edb07add23c3e9b72681cb48e49
 The race can actually be triggered only if the VM is shutting down, so it is 
 not that bad.
 
 Fixing the reported issue in the specific case can be done with a trivial if, 
 and that it what I did
 in my initial http://gerrit.ovirt.org/#/c/25803/1/vdsm/vm.py,cm

Could you explain how an AttributeError there moved the VM to Down?

 
 However, this is just a bandaid and a proper solution is needed. This is the 
 reason why
 the following versions of http://gerrit.ovirt.org/#/c/25803 changed direction 
 toward a more comprehensive
 approach.
 
 And this is the core of the issue.
 My initial idea is to either return success and a complete, well formed 
 statistics set, or return an error.
 However current engine seems to not cope properly with this, and we cannot 
 break backward compatibility.

Would you be more precise? If getAllVmStats returns an errCode for one
of the Vms, what happens?

 
 Looks like the only way to go is to always return success and to add a field 
 to describe the content of the
 statistics (partial, complete...). THis is admittedly a far cry from the 
 ideal solution, but it is dictated
 by the need to preserve the compatibility with current/old engines.

I don't think that I understand your suggestion, but it does not sound
right to send a partial dictionary and to appologize about its
paritiality elsewhere. The dictionary may be partial for engine-4.0
yet complete for engine-3.5. It's not for Vdsm to grade its own
output.

 
 Moreover, I would like to take the chance and cleanup/refactor the statistics 
 collection. I already started
 adding the test infrastructure: http://gerrit.ovirt.org/#/c/26536/
 
 To summarize, what I suggest to do is:
 
 * fix https://bugzilla.redhat.com/show_bug.cgi?id=1073478 using a simple ugly 
 fix like the original 
   http://gerrit.ovirt.org/#/c/25803/1/vdsm/vm.py,cm (which I'll resubmit as 
 new change)
 * refactor and clean getStats() and friends
 * on the cleaner base, properly fix the statistics collection by let 
 getStats() always succeed and return
   possibly partial stats, with a new field describing the content
 
 please note that I'm not really happy about this solution, but, given the 
 constraint, I don't see better
 alternatives.

Please explain the benefits of describing the partial content, as I do
not see them.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] Addming whole application profile

2014-03-26 Thread Dan Kenigsberg
On Wed, Mar 26, 2014 at 09:13:16AM -0400, Nir Soffer wrote:
 Hi all,
 
 Please review this patch:
 http://gerrit.ovirt.org/26113
 
 The short term goal of this patch is to allow debugging of this nasty bug:
 https://bugzilla.redhat.com/1074097
 
 The long term goal is having an easy way to profile vdsm in the field and
 during development, initiative started by Francesco.
 
 Here are some docs to help you get started with the profiler.
 
 Installing yappi
 
 
 To try this patch, you must install yappi:
 
 1. wget http://yappi.googlecode.com/files/yappi-0.82.tar.gz
 2. tar xzf yappi-0.82.tar.gz
 3. cd yappi-0.82
 4. sudo python setup.py install
 
 If you want to build an rpm, the quickest way is:
 
 1. Fix MANIFEST.in so it looks like this:
 
 include *.h
 include ez_setup.py   
   
 
 
 2. Build rpm
 
 $ python setup.py bdist_rpm
 
 You rpm is located in dist.
 
 I built rpm packages for fedora and rhel (looking for a place to share them)

How about making it a Fedora package?
https://fedoraproject.org/wiki/Join_the_package_collection_maintainers#Create_Your_Review_Request

(caveat - it can be a long long process)
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] if ipv6 ready and hot to use it?

2014-03-24 Thread Dan Kenigsberg
On Mon, Mar 24, 2014 at 12:32:50PM +, Sven Kieske wrote:
 Well would you maybe consider priorising this feature set a bit?
 
 Because the world really runs out of IPv4 addresses..fast.

The planning phase of ovirt-3.5 is almost done...
But it would be nice if you join that ovirt-3.5 planning thread on
ovirt-users and explain why it's important to have Engine send ipv6
addresses. And better yet, get someone to send the patches upstream...

Dan.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] VDSM profiling results, round 1

2014-03-19 Thread Dan Kenigsberg
On Wed, Mar 19, 2014 at 08:04:18AM -0400, Francesco Romani wrote:
 another cpickle patch round. http://paste.fedoraproject.org/86653/
 The only change is:
 - DO NOT restart vdsmd between runs (just because it was faster to test)
 
 the timings/profile is now much more closer to what I was expecting.
 The benefits of switching to cPickle are indeed small, but the change is
 quick easy and safe, so I'm still all for it.

I'm confused. What is the numerical benefit?
http://paste.fedoraproject.org/86611/ suggests that it's 36/1230=3% of
GIL-holding time, but the former pastebin saw a total *worsening*.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] creating ovirtmgmt bridge using tagged vlans

2014-03-19 Thread Dan Kenigsberg
On Wed, Mar 19, 2014 at 08:00:48AM -0400, Antoni Segura Puimedon wrote:
 
 
 - Original Message -
  From: Sandro Bonazzola sbona...@redhat.com
  To: VDSM Project Development vdsm-devel@lists.fedorahosted.org
  Cc: Antoni Segura Puimedon asegu...@redhat.com
  Sent: Wednesday, March 19, 2014 12:54:45 PM
  Subject: creating ovirtmgmt bridge using tagged vlans
  
  Hi,
  I'm working on: Bug 1072027 - hosted-engine setup fails when using VLAN
  tagged interfaces
  
  Provided the following caps:
  
  # vdsClient -s localhost getVdsHardwareInfo
  systemFamily = 'Not Specified'
  systemManufacturer = 'Dell Inc.'
  systemProductName = 'OptiPlex 7010'
  systemSerialNumber = 'DYC5YX1'
  systemUUID = '4C4C4544-0059-4310-8035-C4C04F595831'
  systemVersion = '01'
  [root@dellserver network-scripts]# vdsClient -s localhost getVdsCap
  getVdsCapabilities  getVdsCaps
  [root@dellserver network-scripts]# vdsClient -s localhost getVdsCaps
  HBAInventory = {'FC': [], 'iSCSI': [{'InitiatorName':
  'iqn.1994-05.com.redhat:babababa'}]}
  ISCSIInitiatorName = 'iqn.1994-05.com.redhat:babababa'
  bondings = {'bond0': {'addr': '',
'cfg': {},
'hwaddr': '5a:46:b1:db:ec:9b',
'ipv6addrs': [],
'mtu': '1500',
'netmask': '',
'slaves': []}}
  bridges = {';vdsmdummy;': {'addr': '',
 'cfg': {},
 'gateway': '',
 'ipv6addrs': 
  ['fe80::8c8a:edff:fe92:b332/64'],
 'ipv6gateway': '::',
 'mtu': '1500',
 'netmask': '',
 'ports': [],
 'stp': 'off'}}
  clusterLevels = ['3.0', '3.1', '3.2', '3.3', '3.4']
  cpuCores = '4'
  cpuFlags =
  'fpu,vme,de,pse,tsc,msr,pae,mce,cx8,apic,sep,mtrr,pge,mca,cmov,pat,pse36,clflush,dts,acpi,mmx,fxsr,sse,sse2,ss,ht,tm,pbe,syscall,nx,rdtscp,lm,constant_tsc,arch_perfmon,pebs,bts,rep_good,nopl,xtopology,nonstop_tsc,aperfmperf,eagerfpu,pni,pclmulqdq,dtes64,monitor,ds_cpl,vmx,smx,est,tm2,ssse3,cx16,xtpr,pdcm,pcid,sse4_1,sse4_2,x2apic,popcnt,tsc_deadline_timer,aes,xsave,avx,f16c,rdrand,lahf_lm,ida,arat,epb,xsaveopt,pln,pts,dtherm,tpr_shadow,vnmi,flexpriority,ept,vpid,fsgsbase,smep,erms,model_Nehalem,model_Conroe,model_coreduo,model_core2duo,model_Penryn,model_Westmere,model_n270,model_SandyBridge'
  cpuModel = 'Intel(R) Core(TM) i7-3770 CPU @ 3.40GHz'
  cpuSockets = '1'
  cpuSpeed = '1864.156'
  cpuThreads = '8'
  emulatedMachines = ['pc',
  'pc-q35-1.4',
  'pc-q35-1.5',
  'q35',
  'isapc',
  'pc-0.10',
  'pc-0.11',
  'pc-0.12',
  'pc-0.13',
  'pc-0.14',
  'pc-0.15',
  'pc-1.0',
  'pc-1.1',
  'pc-1.2',
  'pc-1.3',
  'pc-i440fx-1.4',
  'pc-i440fx-1.5',
  'none']
  guestOverhead = '65'
  hooks = {}
  kvmEnabled = 'true'
  lastClient = '127.0.0.1'
  lastClientIface = 'lo'
  management_ip = '0.0.0.0'
  memSize = '15936'
  netConfigDirty = 'False'
  networks = {}
  nics = {'em1': {'addr': '192.168.1.105',
  'cfg': {'DEVICE': 'em1',
  'HWADDR': 'b8:ca:3a:76:9a:43',
  'MTU': '1500',
  'ONBOOT': 'yes'},
  'hwaddr': 'b8:ca:3a:76:9a:43',
  'ipv6addrs': ['fe80::baca:3aff:fe76:9a43/64'],
  'mtu': '1500',
  'netmask': '255.255.255.0',
  'speed': 1000}}
  operatingSystem = {'name': 'Fedora', 'release': '8', 'version': '19'}
  packages2 = {'kernel': {'buildtime': 1394207804.0,
  'release': '100.fc19.x86_64',
  'version': '3.13.6'},
   'libvirt': {'buildtime': 1387094943,
   'release': '1.fc19',
   'version': '1.1.3.2'},
   'mom': {'buildtime': 1391183561, 'release': '1.fc19',
   'version': '0.4.0'},
   'qemu-img': {'buildtime': 1384762225,
'release': '2.fc19',
'version': '1.6.1'},
   'qemu-kvm': {'buildtime': 1384762225,
'release': '2.fc19',
'version': '1.6.1'},
 

[vdsm] Vdsm Coding Guidelines

2014-03-18 Thread Dan Kenigsberg
Hi folks,

I've patched up http://www.ovirt.org/Vdsm_Coding_Guidelines. Please keep
it in mind when coding and reviewing.

If you'd like to amend that suggestion, let's discuss it here.

Regards,
Dan.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


[vdsm] vdsm call

2014-03-18 Thread Dan Kenigsberg
BEGIN:VCALENDAR
PRODID:Zimbra-Calendar-Provider
VERSION:2.0
METHOD:REQUEST
BEGIN:VTIMEZONE
TZID:Europe/London
BEGIN:STANDARD
DTSTART:16010101T02
TZOFFSETTO:-
TZOFFSETFROM:+0100
RRULE:FREQ=YEARLY;WKST=MO;INTERVAL=1;BYMONTH=10;BYDAY=-1SU
TZNAME:GMT/BST
END:STANDARD
BEGIN:DAYLIGHT
DTSTART:16010101T01
TZOFFSETTO:+0100
TZOFFSETFROM:-
RRULE:FREQ=YEARLY;WKST=MO;INTERVAL=1;BYMONTH=3;BYDAY=-1SU
TZNAME:GMT/BST
END:DAYLIGHT
END:VTIMEZONE
BEGIN:VEVENT
UID:95fbe6ce-3950-4413-b63a-d2c5701bc3a7
SUMMARY:vdsm call
ATTENDEE;CN=vdsm-devel;ROLE=REQ-PARTICIPANT;PARTSTAT=NEEDS-ACTION;RSVP=TRUE:
 mailto:vdsm-devel@lists.fedorahosted.org
ORGANIZER;CN=Dan Kenigsberg:mailto:dkeni...@redhat.com
DTSTART;TZID=Europe/London:20140318T143000
DTEND;TZID=Europe/London:20140318T153000
STATUS:CONFIRMED
CLASS:PUBLIC
X-MICROSOFT-CDO-INTENDEDSTATUS:BUSY
TRANSP:OPAQUE
LAST-MODIFIED:20140318T160917Z
DTSTAMP:20140318T160917Z
SEQUENCE:1
DESCRIPTION:The following is a new meeting request:\n\nSubject: vdsm call \n
 Organizer: Dan Kenigsberg dkeni...@redhat.com \n\nTime: Tuesday\, March 
 18\, 2014\, 2:30:00 PM - 3:30:00 PM GMT +00:00 Britain\, Ireland\, Portugal\
 n \nInvitees: vdsm-devel@lists.fedorahosted.org \n\n\n*~*~*~*~*~*~*~*~*~*\n\
 nIf you receive this e-mail\, it means that you are invited to the\nbi-weekl
 y phone meeting of Vdsm developers. This is an opportunity\nto discuss desig
 n problems\, pending patches\, and persistent bugs.\n\nEven if you do not jo
 in\, you may solicit subjects for discussion on\nthe mailing list or over #v
 d...@irc.freenode.net.\n\nNote the new meeting time and conference code.\n\nC
 onference code: 353-86-075-901.\n\nDial-in information:\nhttps://www.interca
 llonline.com/listNumbersByCode.action?confCode=35386075901
BEGIN:VALARM
ACTION:DISPLAY
TRIGGER;RELATED=START:-PT5M
DESCRIPTION:Reminder
END:VALARM
END:VEVENT
END:VCALENDAR___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


[vdsm] vdsm call

2014-03-18 Thread Dan Kenigsberg
BEGIN:VCALENDAR
PRODID:Zimbra-Calendar-Provider
VERSION:2.0
METHOD:REQUEST
BEGIN:VTIMEZONE
TZID:Europe/London
BEGIN:STANDARD
DTSTART:16010101T02
TZOFFSETTO:-
TZOFFSETFROM:+0100
RRULE:FREQ=YEARLY;WKST=MO;INTERVAL=1;BYMONTH=10;BYDAY=-1SU
TZNAME:GMT/BST
END:STANDARD
BEGIN:DAYLIGHT
DTSTART:16010101T01
TZOFFSETTO:+0100
TZOFFSETFROM:-
RRULE:FREQ=YEARLY;WKST=MO;INTERVAL=1;BYMONTH=3;BYDAY=-1SU
TZNAME:GMT/BST
END:DAYLIGHT
END:VTIMEZONE
BEGIN:VEVENT
UID:95fbe6ce-3950-4413-b63a-d2c5701bc3a7
RRULE:FREQ=WEEKLY;INTERVAL=2;BYDAY=TU
SUMMARY:vdsm call
ATTENDEE;CN=vdsm-devel;ROLE=REQ-PARTICIPANT;PARTSTAT=NEEDS-ACTION;RSVP=TRUE:
 mailto:vdsm-devel@lists.fedorahosted.org
ORGANIZER;CN=Dan Kenigsberg:mailto:dkeni...@redhat.com
DTSTART;TZID=Europe/London:20140318T143000
DTEND;TZID=Europe/London:20140318T153000
STATUS:CONFIRMED
CLASS:PUBLIC
X-MICROSOFT-CDO-INTENDEDSTATUS:BUSY
TRANSP:OPAQUE
LAST-MODIFIED:20140318T161008Z
DTSTAMP:20140318T161008Z
SEQUENCE:2
DESCRIPTION:The following meeting has been modified:\n\nSubject: vdsm call \
 nOrganizer: Dan Kenigsberg dkeni...@redhat.com \n\nTime: 2:30:00 PM - 3:
 30:00 PM GMT +00:00 Britain\, Ireland\, Portugal\n Recurrence : Every 2 week
 (s) on Tuesday No end date Effective Mar 18\, 2014\n\nInvitees: vdsm-devel@l
 ists.fedorahosted.org \n\n\n*~*~*~*~*~*~*~*~*~*\n\nIf you receive this e-mai
 l\, it means that you are invited to the \nbi-weekly phone meeting of Vdsm d
 evelopers. This is an opportunity \nto discuss design problems\, pending pat
 ches\, and persistent bugs. \n\nEven if you do not join\, you may solicit su
 bjects for discussion on \nthe mailing list or over #v...@irc.freenode.net. 
 \n\nNote the new meeting time and conference code. \n\nConference code: 353-
 86-075-901. \n\nDial-in information: \nhttps://www.intercallonline.com/listN
 umbersByCode.action?confCode=35386075901 
BEGIN:VALARM
ACTION:DISPLAY
TRIGGER;RELATED=START:-PT5M
DESCRIPTION:Reminder
END:VALARM
END:VEVENT
END:VCALENDAR___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] deepcopy error when saveState

2014-03-16 Thread Dan Kenigsberg
On Thu, Mar 13, 2014 at 03:27:29PM +0800, bigclouds wrote:
 what is the reason of this error?
  
 Thread-20::DEBUG::2014-03-12 19:04:23,191::BindingXMLRPC::161::vds::(wrapper) 
 client [192.168.0.113]
 Thread-20::DEBUG::2014-03-12 
 19:04:23,191::task::563::TaskManager.Task::(_updateState) 
 Task=`30301b19-55ba-4152-98fd-0301d1c3975a`::moving from state init - state 
 preparing
   File /usr/lib64/python2.7/copy.py, line 257, in _deepcopy_dict
   File /usr/lib64/python2.7/copy.py, line 190, in deepcopy
   File /usr/lib64/python2.7/copy.py, line 334, in _reconstruct
   File /usr/lib64/python2.7/copy.py, line 163, in deepcopy
   File /usr/lib64/python2.7/copy.py, line 257, in _deepcopy_dict
   File /usr/lib64/python2.7/copy.py, line 190, in deepcopy
   File /usr/lib64/python2.7/copy.py, line 334, in _reconstruct
   File /usr/lib64/python2.7/copy.py, line 163, in deepcopy
   File /usr/lib64/python2.7/copy.py, line 257, in _deepcopy_dict
   File /usr/lib64/python2.7/copy.py, line 190, in deepcopy
   File /usr/lib64/python2.7/copy.py, line 329, in _reconstruct
   File /usr/lib64/python2.7/copy_reg.py, line 93, in __newobj__
 TypeError: object.__new__(PyCapsule) is not safe, use PyCapsule.__new__()
 Thread-13::DEBUG::2014-03-12 
 19:04:21,526::vm::2226::vm.Vm::(_startUnderlyingVm) 
 vmId=`6d32200d-43ee-48a5-8ee9-fe6beaf2ca9d`::_ongoingCreations released
 Thread-13::INFO::2014-03-12 
 19:04:21,550::vm::2250::vm.Vm::(_startUnderlyingVm) 
 vmId=`6d32200d-43ee-48a5-8ee9-fe6beaf2ca9d`::Skipping errors on recovery

Could you attach 6d32200d-43ee-48a5-8ee9-fe6beaf2ca9d.recovery file? It
might have been corrupted (e.g. due to a filesystem error) which most
probably means the untimely death of your VM.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] [PATCH] Monkeypatch python's thread.allocate_lock

2014-03-04 Thread Dan Kenigsberg
On Sun, Mar 02, 2014 at 12:01:21PM +0200, Yaniv Bronhaim wrote:
 In python's thread package implementation thread.allocate_lock is being used
 in each creation of new thread (python.__bootstrap_inner).
 Vdsm faced few scenarios where the thread run was stuck in this function.
 We assume that this lock implementation contains a deadlock bug. Therefore

I did not see enough evidence for this assumption, however, using
pthreading.Lock everywhere is just cleaner. Could you say we suspect
instead?

 this patch replaces thread.allocate_lock with python-pthreading Lock object
 (which uses pthread mutex).

python-pthreading Lock - pthreading.Lock


Other than this - ACK. Please take it into git.

___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] question about faqemu hook

2014-02-25 Thread Dan Kenigsberg
On Tue, Feb 25, 2014 at 02:51:54PM +0200, ybronhei wrote:
 Hey,
 
 does someone know why do we set memory size value to 20480 (20mb) in
 before_vm_start hook when using fake qemu?

That's the whole point that makes it a *fake* qemu. The hook is intended
to be used in cases when you do not care about the guest - you just want
it out there, bothering the management stack with its existence. It is
not expected to run anything, and it should consume as little resources
as possible, so that a quality engineer can condense as many fake VMs on
a single host (most porbably a lightweight virtualized host).
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] fedora19 needs the new glusterfs repo for certain CI jobs

2014-02-17 Thread Dan Kenigsberg
On Thu, Feb 13, 2014 at 09:30:21AM -0500, Eyal Edri wrote:
 
 
 - Original Message -
  From: Ohad Basan oba...@redhat.com
  To: Vered Volansky ve...@redhat.com
  Cc: infra in...@ovirt.org
  Sent: Thursday, February 13, 2014 3:04:13 PM
  Subject: Re: fedora19 needs the new glusterfs repo for certain CI jobs
  
  I think that what has to be done here is that we need to add to puppet the
  gluster upstream repository.
  opinions anyone?
 
 you're talking about the glusterfs epel repos? 
 fine by me, but adding vdsm devel to confirm.
 
 e.
  
  
  - Original Message -
   From: Vered Volansky ve...@redhat.com
   To: infra in...@ovirt.org
   Sent: Thursday, February 13, 2014 3:03:24 PM
   Subject: fedora19 needs the new glusterfs repo for certain CI jobs
   
   Hi,
   
   When http://jenkins.ovirt.org/job/vdsm_storage_functional_tests_localfs
   runs
   on a fedora19 machine, it seems as though the glusterfs repository is
   stale.
   Can it please be updated?
   
   Please note the following error in
   http://jenkins.ovirt.org/job/vdsm_storage_functional_tests_localfs/170/consoleFull
   .
   
   Error: Package: glusterfs-cli-3.4.0-8.fc19.x86_64 (@updates)
  Requires: glusterfs-libs = 3.4.0-8.fc19
   
   and so on...

Ohad, any reason why
http://download.fedoraproject.org/pub/fedora/linux/updates/19/x86_64/glusterfs-3.4.2-1.fc19.x86_64.rpm
is not installed on the slave?
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] pylint in make check-local

2014-02-17 Thread Dan Kenigsberg
On Mon, Feb 17, 2014 at 01:21:11PM +0100, Ewoud Kohl van Wijngaarden wrote:
 On Mon, Feb 17, 2014 at 11:44:37AM +, Dan Kenigsberg wrote:
  While I'm waiting for more acks/nacks, could infra add pylint to our
  Jenkins slaves?
 
 Is the version in EL6 recent enough or do we have to ship our own
 version? (EL6 is 0.21.1)
 
 Please review http://gerrit.ovirt.org/24556

Let's take the standard one, and move to something more modern (such as
Fedora's 1.0.0) when the need arrises.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] fedora19 needs the new glusterfs repo for certain CI jobs

2014-02-16 Thread Dan Kenigsberg
On Sun, Feb 16, 2014 at 09:59:52AM -0500, Ohad Basan wrote:
 I see that the only repository that exists there is aarch64 (ARM) 
 which is irrelevant for us.

Sorry for misleading you into this.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] VDSM Storage tests @ Jenkins

2014-02-10 Thread Dan Kenigsberg
On Mon, Feb 10, 2014 at 03:10:56PM +0100, Vinzenz Feenstra wrote:
 Hi,
 
 Could we please disable the storage tests until they are really
 working? These random failures are really creating more noise than
 being helpful.

As much as I think that they are valuable to have, I think Vinzenz is
correct - they are not yet dependable.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] suggested patch for python-pthreading

2014-02-04 Thread Dan Kenigsberg
On Tue, Feb 04, 2014 at 04:04:37AM -0500, Yaniv Bronheim wrote:
 according to coredumps we found in the scope of the bug [1] we opened [2] 
 that suggested to override python's implementation of thread.allocate_lock
 in each coredump we saw few threads stuck with the bt:
 
 #16 0x7fcb69288c93 in PyEval_CallObjectWithKeywords (func=0x2527820, 
 arg=0x7fcb6972f050, kw=value optimized out) at Python/ceval.c:3663
 #17 0x7fcb692ba7ba in t_bootstrap (boot_raw=0x250a820) at 
 Modules/threadmodule.c:428
 #18 0x7fcb68fa3851 in start_thread (arg=0x7fcb1bfff700) at 
 pthread_create.c:301
 #19 0x7fcb6866694d in clone () at 
 ../sysdeps/unix/sysv/linux/x86_64/clone.S:115
 
 in pystack the threads were stuck in  /usr/lib64/python2.6/threading.py 
 (513): __bootstrap_inner
 
 in bootstrap_inner we use thread.allocate_lock which python-pthreading does 
 not override.
 
 we suggest the following commit:
 
 From 9d89e9be1a379b3d93b23dd54a381b9ca0973ebc Mon Sep 17 00:00:00 2001
 From: Yaniv Bronhaim ybron...@redhat.com
 Date: Mon, 3 Feb 2014 19:24:30 +0200
 Subject: [PATCH] Mocking thread.allocate_lock with Lock imp
 
 Signed-off-by: Yaniv Bronhaim ybron...@redhat.com
 ---
  pthreading.py | 4 
  1 file changed, 4 insertions(+)
 
 diff --git a/pthreading.py b/pthreading.py
 index 916ca7f..96df42c 100644
 --- a/pthreading.py
 +++ b/pthreading.py
 @@ -132,6 +132,10 @@ def monkey_patch():
  Thus, Queue and SocketServer can easily enjoy them.
  
 
 +import thread
 +
 +thread.allocate_lock = Lock
 +
  import threading
 
  threading.Condition = Condition
 -- 
 1.8.3.1
 
 [1] https://bugzilla.redhat.com/show_bug.cgi?id=1022036
 [2] https://bugzilla.redhat.com/show_bug.cgi?id=1060749

It makes sense to use pthreading.Lock for thread.allocate_lock instead of the 
standard
threading.Lock CPU hog. However, I do not understand its
relevance to the deadlock sited above: pthreading.Lock fixes performance
issues, but not correctness issues, of threading.Lock.

Would you explain, in the commit message of the pthreading patch, why
you believe that the implementation of thread.allocate_lock() is buggy?
Do you know if the bug is fixed in Python 3?

Regards,
Dan.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] mom RPMs for 3.4

2014-02-01 Thread Dan Kenigsberg
On Fri, Jan 31, 2014 at 04:56:12PM -0500, Adam Litke wrote:
 On 31/01/14 08:36 +0100, Sandro Bonazzola wrote:
 Il 30/01/2014 19:30, Adam Litke ha scritto:
 On 30/01/14 18:13 +, Dan Kenigsberg wrote:
 On Thu, Jan 30, 2014 at 11:49:42AM -0500, Adam Litke wrote:
 Hi Sandro,
 
 After updating the MOM project's build system, I have used jenkins to
 produce a set of RPMs that I would like to tag into the oVirt 3.4
 release.  Please see the jenkins job [1] for the relevant artifacts
 for EL6[2], F19[3], and F20[4].
 
 Dan, should I submit a patch to vdsm to make it require mom = 0.4.0?
 I want to be careful to not break people's environments this late in
 the 3.4 release cycle.  What is the best way to minimize that damage?
 
 Hey, we're during beta. I prefer making this requirement explicit now
 over having users with supervdsmd.log retate due to log spam.
 
 In that case, Sandro, can you let me know when those RPMs hit the
 ovirt repos (for master and 3.4) and then I will submit a patch to
 vdsm to require the new version.
 
 
 mom 0.4.0 has been built in last night nightly job [1] and published to 
 nightly by publisher job [2]
 so it's already available on nightly [3]
 
 For 3.4.0, it has been planned [4] a beta 2 release on 2014-02-06 so we'll 
 include your builds in that release.
 
 I presume the scripting for 3.4 release rpms will produce a version
 without the git-rev based suffix: ie. mom-0.4.0-1.rpm?
 
 I need to figure out how to handle a problem that might be a bit
 unique to mom.  MOM is used by non-oVirt users who install it from the
 main Fedora repository.  I think it's fine that we are producing our
 own rpms in oVirt (that may have additional patches applied and may
 resync to upstream mom code more frequently than would be desired for
 the main Fedora repository).  Given this, I think it makes sense to
 tag the oVirt RPMs with a special version suffix to indicate that
 these are oVirt produced and not upstream Fedora.
 
 For example:
 The next Fedora update will be mom-0.4.0-1.f20.rpm.
 The next oVirt update will be mom-0.4.0-1ovirt.f20.rpm.
 
 Is this the best practice for accomplishing my goals?  One other thing
 I'd like to have the option of doing is to make vdsm depend on an
 ovirt distribution of mom so that the upstream Fedora version will not
 satisfy the dependency for vdsm.

What is the motivation for this? You would not like to bother Fedora
users with updates that are required only for oVirt?

Vdsm itself is built, signed, and distributed via Fedora.
It is also copied into the ovirt repo, for completeness sake. Could MoM
do the same?

Dan.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] [Users] ovirt-3.3.3 release postponed due to blockers

2014-02-01 Thread Dan Kenigsberg
On Fri, Jan 31, 2014 at 04:16:54PM -0500, Douglas Schilling Landgraf wrote:
 On 01/31/2014 05:17 AM, Dan Kenigsberg wrote:
 On Fri, Jan 31, 2014 at 09:36:48AM +0100, Sandro Bonazzola wrote:
 Il 30/01/2014 22:38, Robert Story ha scritto:
 Can we revert these packages to previous versions in the 3.3.2 stable repo
 so those of us who want/need to install new hosts in our clusters aren't
 dead in the water waiting for 3.3.3?
 
 Hi Robert, I think you can still install 3.3.2 on your clusters with the 
 requirement of adding
 manually oython-cpopen before trying to install vdsm.
 
 About 3.3.3, I think vdsm should really drop dependency on 
 vdsm-python-cpopen:
 it's a package obsoleted by python-cpopen so there's no point in still 
 requiring it especially if keeping that requirement still break dependency
 resolution.
 
 I really wanted to avoid eliminating a subpackage during a micro
 release. That's impolite and surprising.
 But given the awkward yum bug, persistent dependency problems, and
 the current release delay, I give up.
 
 Let's eliminate vdsm-python-cpopen from ovirt-3.3 branch, and require
 python-cpopen. Yaniv, Douglas: could you handle it?
 
 
 Sure. Done: http://gerrit.ovirt.org/#/c/23942/

Acked. Could you cherry-pick it into dist-git and rebuild the
ovirt-3.3.3 candidate (and without other changes that can wait for
ovirt-3.3.4).
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] [Users] ovirt-3.3.3 release postponed due to blockers

2014-01-31 Thread Dan Kenigsberg
On Fri, Jan 31, 2014 at 09:36:48AM +0100, Sandro Bonazzola wrote:
 Il 30/01/2014 22:38, Robert Story ha scritto:
  Can we revert these packages to previous versions in the 3.3.2 stable repo
  so those of us who want/need to install new hosts in our clusters aren't
  dead in the water waiting for 3.3.3?
 
 Hi Robert, I think you can still install 3.3.2 on your clusters with the 
 requirement of adding
 manually oython-cpopen before trying to install vdsm.
 
 About 3.3.3, I think vdsm should really drop dependency on vdsm-python-cpopen:
 it's a package obsoleted by python-cpopen so there's no point in still 
 requiring it especially if keeping that requirement still break dependency
 resolution.

I really wanted to avoid eliminating a subpackage during a micro
release. That's impolite and surprising.
But given the awkward yum bug, persistent dependency problems, and
the current release delay, I give up.

Let's eliminate vdsm-python-cpopen from ovirt-3.3 branch, and require
python-cpopen. Yaniv, Douglas: could you handle it?
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] [Users] ovirt-3.3.3 release postponed due to blockers

2014-01-30 Thread Dan Kenigsberg
On Thu, Jan 30, 2014 at 08:52:36AM +, Sven Kieske wrote:
 Hi,
 
 any news regarding my questions?
 
 Am 29.01.2014 09:23, schrieb Sandro Bonazzola:
  Il 29/01/2014 09:21, Sven Kieske ha scritto:
  Hi,
 
  I wanted to try it the other way around, installing vdsm-python-cpopen
  and check if it runs without python-cpopen .
 
  But that leads me to a question:
 
  Is there any difference between these packages beside their different
  name?
 
  If yes, what is the difference and which package should be installed?
  I no, why is there a packet vdsm-python-cpopen ?
  
  CCing VDSM

python-cpopen includes some improvements and bug fixes, that are going
to be needed in ovirt-3.4. vdsm-python-cpopen is shipped by ovirt-3.3.

python-cpopen wast intended to deprecate and replace vdsm-python-cpopen,
but we have had way too many issues trying to do that properly in
rpm/yum. Most of the bugs are ours, at least one is yum's.

At the moment, the existence of python-cpopen in Fedora confuses `yum
install vdsm`.

To avoid this unfortunate delay, we can either include python-cpopen in
ovirt-stable or exclude vdsm-python-cpopen from there. Both options are
unfavorable, but so is continuing to wait.

Regards,
Dan.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] [Users] ovirt-3.3.3 release postponed due to blockers

2014-01-30 Thread Dan Kenigsberg
On Thu, Jan 30, 2014 at 10:27:34AM +0100, Sandro Bonazzola wrote:
 Il 30/01/2014 10:20, Dan Kenigsberg ha scritto:
  On Thu, Jan 30, 2014 at 08:52:36AM +, Sven Kieske wrote:
  Hi,
 
  any news regarding my questions?
 
  Am 29.01.2014 09:23, schrieb Sandro Bonazzola:
  Il 29/01/2014 09:21, Sven Kieske ha scritto:
  Hi,
 
  I wanted to try it the other way around, installing vdsm-python-cpopen
  and check if it runs without python-cpopen .
 
  But that leads me to a question:
 
  Is there any difference between these packages beside their different
  name?
 
  If yes, what is the difference and which package should be installed?
  I no, why is there a packet vdsm-python-cpopen ?
 
  CCing VDSM
  
  python-cpopen includes some improvements and bug fixes, that are going
  to be needed in ovirt-3.4. vdsm-python-cpopen is shipped by ovirt-3.3.
  
  python-cpopen wast intended to deprecate and replace vdsm-python-cpopen,
  but we have had way too many issues trying to do that properly in
  rpm/yum. Most of the bugs are ours, at least one is yum's.
  
  At the moment, the existence of python-cpopen in Fedora confuses `yum
  install vdsm`.
  
  To avoid this unfortunate delay, we can either include python-cpopen in
  ovirt-stable or exclude vdsm-python-cpopen from there. Both options are
  unfavorable, but so is continuing to wait.
 
 If it's enough just to add python-cpopen to ovirt-stable, I'm fine with that.
 I just need the link to the build to be included there.

Yaniv, have you tried if shipping python-cpopen hides this issue?

http://kojipkgs.fedoraproject.org//packages/python-cpopen/1.3/1.fc20/x86_64/python-cpopen-1.3-1.fc20.x86_64.rpm
http://kojipkgs.fedoraproject.org//packages/python-cpopen/1.3/1.fc19/x86_64/python-cpopen-1.3-1.fc19.x86_64.rpm
http://kojipkgs.fedoraproject.org//packages/python-cpopen/1.3/1.el6/ppc64/python-cpopen-1.3-1.el6.ppc64.rpm

Dropping vdsm-python-cpopen from a stable version seems impolite, but
should work, too.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] mom RPMs for 3.4

2014-01-30 Thread Dan Kenigsberg
On Thu, Jan 30, 2014 at 11:49:42AM -0500, Adam Litke wrote:
 Hi Sandro,
 
 After updating the MOM project's build system, I have used jenkins to
 produce a set of RPMs that I would like to tag into the oVirt 3.4
 release.  Please see the jenkins job [1] for the relevant artifacts
 for EL6[2], F19[3], and F20[4].
 
 Dan, should I submit a patch to vdsm to make it require mom = 0.4.0?
 I want to be careful to not break people's environments this late in
 the 3.4 release cycle.  What is the best way to minimize that damage?

Hey, we're during beta. I prefer making this requirement explicit now
over having users with supervdsmd.log retate due to log spam.

 
 [1] http://jenkins.ovirt.org/view/All/job/manual-build-tarball/179/
 [2] 
 http://jenkins.ovirt.org/view/All/job/manual-build-tarball/179/label=centos6-host/artifact/exported-artifacts/mom-0.4.0-1.el6.noarch.rpm
 [3] 
 http://jenkins.ovirt.org/view/All/job/manual-build-tarball/179/label=fedora19-host/artifact/exported-artifacts/mom-0.4.0-1.fc19.noarch.rpm
 [4] 
 http://jenkins.ovirt.org/view/All/job/manual-build-tarball/179/label=fedora20-host/artifact/exported-artifacts/mom-0.4.0-1.fc20.noarch.rpm
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] [Users] ovirtmgmt vanishes after reboot

2014-01-29 Thread Dan Kenigsberg
On Wed, Jan 29, 2014 at 11:07:08AM +0100, Fabian Deutsch wrote:
 Am Mittwoch, den 29.01.2014, 05:02 -0500 schrieb Antoni Segura Puimedon:
 Right.
 I believe the bug you are seeing is:
 http://gerrit.ovirt.org/#/c/20068/
 
 This has been merged into out stable branch, but the release of a new
 oVirt Node ISO for 3.3 is pending because of the vdsm-python-cpopen
 dependency problem.

Would you agree to patch `yum` on your build machine? I do not fully
understand the issue, but with the following patch, the dependency
problem is resolved for me:

diff --git a/yum/depsolve.py b/yum/depsolve.py
index 95c21bc..57cf379 100644
--- a/yum/depsolve.py
+++ b/yum/depsolve.py
@@ -720,7 +720,7 @@ class Depsolve(object):
 else:
 self.verbose_logger.debug(_('TSINFO: Marking %s as install for 
%s'), best,
 requiringPo)
-reqtuple = misc.string_to_prco_tuple(needname + str(needflags) + 
needversion)
+reqtuple = misc.string_to_prco_tuple(requirement)
 txmbrs = self.install(best, provides_for=reqtuple)
 for txmbr in txmbrs:
 txmbr.setAsDep(po=requiringPo)

___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] Introduction of VDSM - Guest Agent API Versioning

2014-01-27 Thread Dan Kenigsberg
On Wed, Jan 22, 2014 at 12:12:53PM +0100, Vinzenz Feenstra wrote:
 Hi,
 
 With the increasing complexity on different version of the guest
 agent and vdsm
 I have proposed a patchset for each the guest agent and vdsm to
 implement it on
 top of the current protocol, in a way that it will not affect the
 existing protocol.
 This patchset has created some controversy about its implementation
 and design
 and I was asked to kick off a discussion to design a better
 implementation or
 validate the suggested approach.
 
 
 My goals for the API versioning:
 ===
 - Make VDSM and the Guest Agent both aware of what the other side
 understands
   (which messages) to avoid logging errors
 
 - The used version of the API should be automatically and
 immediately be agreed
   upon in case any change happened. And in case one end doesn't
 support it but
   the other does, the supporting end must revert back to a
 compatible state.
 
   Possible scenarios:
- a VM gets migrated from a VDSM version with API version support
 to a VDSM
  without API version support (or lower version)
 
- VDSM gets downgraded or upgraded and the API version changed or is no
  longer supported
 
- The ovirt-guest-agent gets up or downgraded

and Vdsm upgrade, too.

 
 - Make use of existing messages which allow extending without
 generating errors
   or would result in VDSM API changes due to direct export of the
 message, to be
   backwards compatible on both ends
 
 From my side not considered as goals for the API Versioning:
 ==

(sorry, I do not understand this section and its title. are these things
you need? nice-to-have? would like to avoid?)

 - Deprecation of messages, because you can send alternatives if the
 other side
   supports it
 - Raising the lowest supported API version, as this would actually
 break backwards
   compatibility
 
 
 The proposed solution:
 ==
 
 VDSM:
   - Add a field to the 'refresh' message reporting the highest API version
 supported by VDSM
 
   - Upon receiving the 'heartbeat' message check for the `apiVersion` field
 to know if the guest agent supports api versioning.
 - If the fields is not present:
   The guest agent won't support api versioning. And it needs to be
   disabled on the VDSM side and the version 0 has to be assumed. That
   simply means only messages can be sent which were supported
 before the
   API versioning was introduced.
 - If the field is present:
   The value of the field is supposed to represent the maximum version
   supported by the guest agent.
 
   VDSM then makes a `min(vdsmMaxApiVersion, guestAgentMaxApiVersion)`
   to determine the highest common value and uses this value as
 supported
   API version.
   When the value has changed since the last time a heartbeat was
 received.
   VDSM will send a message 'api-version' to update the guest
 agent about
   the determined value. And sets the internally stored apiVersion value
   to this new value.
 
 Guest Agent:
 - Adds the field `apiVersion` to the heartbeat containing the
 highest API
 version supported by the guest agent
 
   - Upon receiving the `refresh` message without a `apiVersion` field, the
 guest agent immediately will revert immediately fall back to version 0
 To avoid sending any messages which are not yet supported.
 
   - Upon receiving the `api-version` message, the guest agent will
 verify the
 value received and sets the value to min(received, maxAgentApiVersion)
 which should, of course, result in setting the received value (this is
 just an additional step to ensure it does not send more than
 supported by
 VDSM.
 
 
 Once having the API version, we can use some lookup table for the messages
 before sending and check which version is required to send it. And we would
 even be able to give some feedback if we support a feature or not on
 the guest
 side.
 
 Please let me know your opinion and I would be welcoming suggestions or any
 other comment related to this.

I find this reasonable solution - despite several comments that I had
regarding the implementation's clarity.

___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] Introduction of VDSM - Guest Agent API Versioning

2014-01-27 Thread Dan Kenigsberg
On Mon, Jan 27, 2014 at 03:00:55PM +0100, Vinzenz Feenstra wrote:
 On 01/27/2014 02:57 PM, Dan Kenigsberg wrote:
 On Wed, Jan 22, 2014 at 12:12:53PM +0100, Vinzenz Feenstra wrote:
 
Possible scenarios:
 - a VM gets migrated from a VDSM version with API version support
 to a VDSM
   without API version support (or lower version)
 
 - VDSM gets downgraded or upgraded and the API version changed or is no
   longer supported
 
 - The ovirt-guest-agent gets up or downgraded
 and Vdsm upgrade, too.
 Second point VDSM gets downgraded or upgraded...

oops, dunno how I've missed it.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] Find reviewers for codes of Numa feature

2014-01-27 Thread Dan Kenigsberg
On Sun, Jan 26, 2014 at 06:51:50AM +, Shi, Xiao-Lei (Bruce, HP 
Servers-PSC-CQ) wrote:
 Hi all,
 
 I have submitted a draft patch about Features/NUMA and Virtual 
 NUMAhttp://www.ovirt.org/Features/NUMA_and_Virtual_NUMA.
 Here is the gerrit request: http://gerrit.ovirt.org/#/c/23703/
 Please help me to review the codes.

Thanks for your contrinution!

I see that Francesco gave it a preliminary review. Since this patch
borders virt and sla, I think that Martin could be of help, too.

When you introduce new API functionality, you most usually need to
document it in vdsm_api/vdsmapi-schema.json and explain in the commit
message, how would the new response to this verb would look like.
Unit tests serve as a good example of how to use the new code.

Dan.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] Vdsm network functional tests taking slave offline

2014-01-23 Thread Dan Kenigsberg
On Thu, Jan 23, 2014 at 01:17:21PM +0100, David Caro wrote:
 Hi!
 
 
 We have been seeing the slave that is used for the network tests getting 
 offline
 from time to time, it was dues to vdsm_functional_tests not cleaning up
 properly, as removing all the *dummy* interface configurations from
 /etc/sysconfig/network-scripts and restarting the network brings it back 
 again.
 
 Can you take a look on why it would not have been cleaning up properly?
 
 the last job that made the host go offline is:
 http://jenkins.ovirt.org/job/vdsm_network_functional_tests/1173/console
 
 The number of the dummy interface that was left over was 6 (not sure it's
 reelevant, but just in case).

It seems that Jenkins killed the job forcefully, before it had the
chance to clean after itself:

testIPv6ConfigNetwork   FATAL: Unable 
to delete script file /home/jenkins/workspace/tmp/hudson6940503892814963275.sh
hudson.util.IOException2: remote file operation failed: 
/home/jenkins/workspace/tmp/hudson6940503892814963275.sh at 
hudson.remoting.Channel@67a91450:fedora19-vm01
at hudson.FilePath.act(FilePath.java:906)
at hudson.FilePath.act(FilePath.java:883)
at hudson.FilePath.delete(FilePath.java:1268)
...

Does jenkins have the ability to add a job-specific cleanup script, to be
excuted on such occasions? I'm sure Toni could populate it witha couple
of lines to remove remainders?
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] [gerrit 23138] VM responsiveness and libvirt timeouts

2014-01-22 Thread Dan Kenigsberg
On Wed, Jan 22, 2014 at 05:49:11AM -0500, Francesco Romani wrote:
 Hi everyone,
 
 I'd like to gather some thoughts about this change 
 http://gerrit.ovirt.org/#/c/23138/5
 
 Sorry for the long mail, but the scenario is not trivial to explain and I 
 tried
 to achieve a decent compromise between terseness and completeness.
 
 
 Context
 ---
 
 Everything starts from the bugzilla https://bugzilla.redhat.com/866484 .
 
 Summary of the bug: when a blockInfo call to libvirt goes in timeout,
 the status of the affected VM start to bounce in the engine between
 'Non Responsive' and 'Up'.
 This is because the responsiveness of a VM is  tracked using a single flag,
 updated each after libvirt call, and the flag could be overwritten behind the 
 back of
 the engine (between two getStats). In a nutshell, the responsiveness of a VM
 could observed in a different way depending on the time sequence of the 
 updates.
 
 It is worth to mention that, I *cannot* yet reproduce in the terms described 
 in the bug
 description.
 Part of the reason why I failed to reproduce it is a particular event sequence
 is required to achieve the bouncing effect described in the bug.
 The event sequence is in turn determined by the timing of the block storage 
 timeout
 and by the time the engine queries the affected VDSM.

Pardon my skeptism, but a bug that was opened with How reproducible:
100% and now cannot be reproduced at all smells like a misunderstood
bug.

 
 However, my understanding is that the scenario which emerges from the bug 
 description
 and the analysis of the logs attached to the bug is still relevant, although 
 somewhat
 unlikely. This is the reason why I submitted the change 
 http://gerrit.ovirt.org/#/c/23138/5
 
 The issue with the new patch
 
 
 The new code determines the responsiveness of the monitor -and thus of the VM-
 in a `pessimistic` way, meaning that now as long as at least one call to 
 libvirt fails,
 the VM is deemed unresponsive.
 
 On the other hand, with the current code, if an attribute is unresponsive but 
 if the sequence
 of events is good enough, the not responsive time interval could be hidden 
 from the engine's sight.
 Or, of course, the engine could see the bouncing we're discussing.
 
 In a nutshell, the new code could be too pessimistic in reporting the state.
 A VM could be reported not responsive longer and more often (again it all 
 depends on the event
 sequence) than before.

Indeed, this issue is crippling. For example, if you fire up migration,
and due to a transient hickup, libvirt times out, we would get into a
dead end. The attribute for migrate would be set to -1, the VM would
be reported as non-responsive, and no future migrate command would ever
be sent to free it.


 It that is the case, the problem is: how to fix the reporting in a more 
 optimistic way?
 
 Possible solution
 -
 
 While reviewing the change with Michal, we found that partitioning the set of
 the calls to be monitored in critical/not critical could be a way:
 
 * critical libvirt calls: if at least one fails for timeout, and while it 
 fails for timeout,
   the VM is considered Unresponsive.
 * not critical: a timeout is logged but ignored to decide the responsiveness 
 of the VM
 
 Would that be an improvement with respect to the current state and, possible, 
 a viable
 long-term solution?

I like the idea of partitioning, but criticality is not a clear
indicator imo. monitorResponse was introduced as a heuristical answer to
the question would qemu accept an upcoming command and act upon it in a
timely manner?

A libvirt.virDomain method that succeeds despite its qemu process being
SIGSTOPped, should not be considered when setting monitorResponse. I we
undestand what has reset monitorResponse in Dafna's scenerio, we could
avoid wrapping it with _timeoutExperienced.

Dan.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] [Engine-devel] Copy reviewer scores on trivial rebase/commit msg changes

2014-01-18 Thread Dan Kenigsberg
On Sat, Jan 18, 2014 at 01:48:52AM +0200, Itamar Heim wrote:
 I'd like to enable these - comments welcome:
 
 1. label.Label-Name.copyAllScoresOnTrivialRebase
 
 If true, all scores for the label are copied forward when a new
 patch set is uploaded that is a trivial rebase. A new patch set is
 considered as trivial rebase if the commit message is the same as in
 the previous patch set and if it has the same code delta as the
 previous patch set. This is the case if the change was rebased onto
 a different parent. This can be used to enable sticky approvals,
 reducing turn-around for trivial rebases prior to submitting a
 change. Defaults to false.
 
 
 2. label.Label-Name.copyAllScoresIfNoCodeChange
 
 If true, all scores for the label are copied forward when a new
 patch set is uploaded that has the same parent commit as the
 previous patch set and the same code delta as the previous patch
 set. This means only the commit message is different. This can be
 used to enable sticky approvals on labels that only depend on the
 code, reducing turn-around if only the commit message is changed
 prior to submitting a change. Defaults to false.
 
 
 https://gerrit-review.googlesource.com/Documentation/config-labels.html

I think that the time saved by these copying is worth the dangers.

But is there a way to tell a human ack from an ack auto-copied by these
options? It's not so fair to blame X for X approved this patch when he
only approved a very similar version thereof.

Assuming that a clean rebase can do no wrong is sometimes wrong
(a recent example is detailed by Nir's http://gerrit.ovirt.org/21649/ )

Dan.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] [Engine-devel] Gerrit NEW Change Screen

2014-01-16 Thread Dan Kenigsberg
On Thu, Jan 16, 2014 at 08:16:17AM -0500, Adam Litke wrote:
 On 16/01/14 00:08 +0200, Itamar Heim wrote:
 with gerrit 2.8, there is a new change screen.
 its not enabled by default (yet), please use and see what you think.
 
 to enable, go to settings (click the top-right arrow next to your
 name, and choose settings).
 select preferences and set Change View: to New Screen.
 
 Thanks Itamar.  For me, this new screen is much better.  Things are
 more compact and for most patches all of the important information
 fits easily on one screen.  I also like the easy to find gitweb link
 and the ability to edit the commit message right from the interface.

On my narrow lcd, though, the new look makes comments in diff to be seen
only partially. I failed to find a configurable to make the comment
windows narrower, so my only solution is to reduce the font (which is
bad for my poor old eyes).
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] [Users] [QE] 3.4.0 Release tracker

2014-01-13 Thread Dan Kenigsberg
On Mon, Jan 13, 2014 at 11:05:05AM +0100, Sander Grendelman wrote:
 On Mon, Jan 13, 2014 at 12:32 AM, Dan Kenigsberg dan...@redhat.com wrote:
  On Fri, Jan 10, 2014 at 04:55:18PM +0200, Itamar Heim wrote:
  On 01/10/2014 01:52 PM, Sander Grendelman wrote:
  Can I propose BZ#1035314 for 3.3.3 or 3.4.0, simple, trivial fix to a 
  hook.
 
  Hi Sander,
 
  please use bug summary so folks won't have to go and look what the
  number means just to see if relevant to them.
 
  this is about:
  Bug 1035314 - vdsm-hook-nestedvt uses kvm_intel-only syntax
 
 OK, I'll keep that in mind!
 
  i see you posted a patch in the bug - can you post the patch to
  gerrit as well?
 
  I've done that alraedy: http://gerrit.ovirt.org/#/c/23164/
 
  Sander (and others), could you review and formally verify it?
 
 Looks fine to me, should I also do/confirm something in gerrit?

Yes, it would be great if you tick the patch as verified, as I did not
try it myself.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] [Users] [QE] 3.4.0 Release tracker

2014-01-12 Thread Dan Kenigsberg
On Fri, Jan 10, 2014 at 04:55:18PM +0200, Itamar Heim wrote:
 On 01/10/2014 01:52 PM, Sander Grendelman wrote:
 Can I propose BZ#1035314 for 3.3.3 or 3.4.0, simple, trivial fix to a hook.
 
 Hi Sander,
 
 please use bug summary so folks won't have to go and look what the
 number means just to see if relevant to them.
 
 this is about:
 Bug 1035314 - vdsm-hook-nestedvt uses kvm_intel-only syntax
 
 i see you posted a patch in the bug - can you post the patch to
 gerrit as well?

I've done that alraedy: http://gerrit.ovirt.org/#/c/23164/

Sander (and others), could you review and formally verify it?
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] Smarter network_setup hooks

2014-01-08 Thread Dan Kenigsberg
On Wed, Jan 08, 2014 at 03:34:24PM +0100, Miguel Angel wrote:
 Hi Assaf, Thank you very much for the summary,
 
 Just a few questions, there are things I don't understand yet:
 
 2014/1/8 Assaf Muller amul...@redhat.com
 
  Mid-thread summary:
 
  First some terminology so we're all on the same page:
  'request' - The setupNetworks that just arrived
  'running config' - If unified persistence is being used - All networking
  on the host as VDSM sees it
  'expected' - If unified persistence is being used - The running config
  after the effects of the request, but before the hook is ran
 
 
 correct
 
 
  I think we all agree that we have to expose the 'request' to the hook.
  However, there's two different approaches:
  1) Allow the hook to rewrite parts of the request (For example, if two
  networks were sent, and the hook handled one of them, it could
 then delete that network from the request, so VDSM will only continue
  to configure the 2nd network).
 
 
 In this case ,if the hook deleted the 2nd network, VDSM couldn't handle
 the network config persistence anymore, right?,

Right. If the before_network_setup hook decides to drop an item from the
'request', it means that neither following hooks, nor Vdsm proper, would
see it. I find it as a useful and simple sematics: the hook practically
says I'm taking it from here, and from then on, it is repsonsible for
everything. Implementation and persistence alike.


 I didn't expect this use case, I only expected tweaks, etc (before or after
 network setup), for example, setting special hardware
 capabilities using ethtool or those kind of things.

Neither have I expected this. But it's a powerful tool that's relatively
easy to implement.


  2) Toni's idea with marking devices as hook-configured. So if a network is
  over a bridge over a VLAN over a bond over three NICs, hooks
 could configure only the NICs (For example) and VDSM would configure
  the rest, whereas the first idea means that the hook would
 have to configure the entire network (Bridge, VLANs, bonds, and all 3
  NICs) and then remove that network from the request.
 The disadvantage of this method is that it would be harder to
  implement, and have references to the hooks flag throughout all of VDSM.
 
 
 You mean that if the hook marked a certain device as hook handled then,
 VDSM would just keep this information along with
 the network config, etc... and won't do any setup, just config-keeping,
 right?

That's my understanding, too. And I share the feeling that this is error
prone: Vdsm sees the config, but must remember that it should not touch
it or act upon it.

  Then there's the matter of IF to expose the 'running config' and
  'expected'.
 
  My main argument is that it's very easy to expose to the hooks (But then
  why not expose other 'random' data that is easy for VDSM to calculate?),
  and that we don't understand all usages for the setupNetworks hooks. I'd
  rather we expose too much information than not enough and screw
  over hook writers.
 
 
 We have something important here, the hooks don't need to be
 python-written, they could be bash scripts, or any other thing...
 that means that they wouldn't have access to get running config, but they
 could calculate expected.

 So, my vote here goes for request + running config.


The running config is accessible to any process, albeit not in its
Vdsm representation. All the information is available if you do `ip a`
and `virsh net-list`.

I do not imagine why a hook would need the running config; If it does
end up needing it, it could re-calculate it just as Vdsm does. Passing
this as argument to each hook script is a premature optimization imho.

If I end up wrong, it would not be hard to add the running config to
the Vdsm/hook API. Removing something from an API is a much harder
task.

My vote goes to only request.

 
  Either way, let's get some votes in a timely manner so we could manage to
  implement this for 3.4.
 
 
 Thanks Assaf! ;)
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] Smarter network_setup hooks

2014-01-06 Thread Dan Kenigsberg
On Sun, Jan 05, 2014 at 05:05:59AM -0500, Assaf Muller wrote:
 Whichever way we decide to do this, I think the important bit is 
 documentation - We have
 to make sure to update the oVirt wiki hooks pages. If users aren't aware of 
 how to retrieve
 the networking config then we might as well not implement it.
 
 That being said, I'd expose three dictionaries: What's currently configure,
 the current request, as well as the proposed end result.

I do not see why current is a good idea to pass from Vdsm to the hook
- if the hook needs it, it could compute the current state on its own.

What do you mean by the proposed end result? setupNetwork API works in
allows diffs relative to the current state. The end result, however,
may even be unexpressable within the scope of our API (that's exactly
what hooks are for!). And again, if Vdsm is able to calculate the end
result based on the current state + diff, so can the hook itself.

 It's easy to add
 and I see how it would be useful to hook writers. And just to state the 
 obvious,
 just like how traditional hooks can change the VM or device XML,
 the hook should be able to rewrite the current request contents.
 For example, if a user would like to take over host networking configuration,
 he could just write a before_setup_networks hook that would configure
 networking however he wants, then writes the empty dictionary to the current 
 request,
 meaning that VDSM wouldn't do anything further with the current setup 
 networks request.

This would be a very good additions. Cascaded scripts would then be able
to drop parts of the command, which they have implemented themselves.

Regards,
Dan.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] Smarter network_setup hooks

2014-01-06 Thread Dan Kenigsberg
On Sun, Jan 05, 2014 at 08:58:07PM -0500, Antoni Segura Puimedon wrote:
 
 
 - Original Message -
  From: Miguel Angel miguelan...@ajo.es
  To: Livnat Peer lp...@redhat.com
  Cc: dsull...@redhat.com, vdsm-de...@fedorahosted.org, a...@ovirt.org
  Sent: Sunday, January 5, 2014 11:41:44 PM
  Subject: Re: [vdsm] Smarter network_setup hooks
  
  
  2014/1/5 Livnat Peer  lp...@redhat.com 
  
  
  
  On 01/05/2014 12:05 PM, Assaf Muller wrote:
   Whichever way we decide to do this, I think the important bit is
   documentation - We have
   to make sure to update the oVirt wiki hooks pages. If users aren't aware 
   of
   how to retrieve
   the networking config then we might as well not implement it.
   
   That being said, I'd expose three dictionaries: What's currently 
   configure,
   the current request, as well as the proposed end result. It's easy to add
   and I see how it would be useful to hook writers. And just to state the
   obvious,
   just like how traditional hooks can change the VM or device XML,
   the hook should be able to rewrite the current request contents.
   For example, if a user would like to take over host networking
   configuration,
   he could just write a before_setup_networks hook that would configure
   networking however he wants, then writes the empty dictionary to the
   current request,
   meaning that VDSM wouldn't do anything further with the current setup
   networks request.
 
 My original thought on this was that the hook would be able to mark a device
 as configured by it adding a key value like 'hooked': X. This was in order
 that if, for example a bond is to be configured by the hook, it would still
 stay in the config passed to objectivize but the configurator configure method
 for it would be short circuited and X would have the data that is to be put
 in the running config by vdsm for that device.
 
 I didn't mature this thought much though... (holidays have kept me a bit busy)

Oh, few seconds ago I noted about the idea that hooks scripts would be
able to remove parts of the configuration which they have already
implemented, so that Vdsm proper is left unaware of them.

It's not as flexible as your hooked=True suggestion, as it does not
alow to implement only part of a network, but I think that it is
as-powerful and with clearer atomicity.

 
   
  
  
  I'm not sure if it's easy to get the final state without actually applying
  it, it's easy
  to get an approximate final state (just aggregating dictionaries to networks
  and bondings,
  and erasing the removed ones), but I suppose that'd be good enough :-)
  
  
  May be it's good if you can provide a use case for this third expected
  final state,
  I can't come up with one. :-)
 
 For this reason I thought of the 'hooked' value, to make the hook writer own
 the task of filling in the missing piece of state.

Sorry, Toni, I do not understaed how hooked=True is related to the
expected dictionary suggested by Assaf.

 One final consideration that I didn't see arise is the need for having caps
 hooks. As soon as we allow setupNetworks hooks it is very necessary that we
 enable vdsm caps. One very easy example.
 
 Let us say that I write a hook that makes setupNetworks specified bondings
 be configured in the system by libteam/teamd. In order for the changes to
 be made visible to the engine we need to fake that team link aggregation
 as a bond so that the engine can, in the future, change/delete/use it.

Correct. The same json-based framework devised by Miguel (as well as
Adam's getContext) would come up hand when implementing it.

Dan.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] what does custom option mean when vmcreate ?

2014-01-03 Thread Dan Kenigsberg
On Fri, Jan 03, 2014 at 11:25:04AM +0100, Michal Skrivanek wrote:
 
 On 26 Dec 2013, at 07:32, bigclouds wrote:
 
  the paramters received when vmCreate has a long part called custom,
  do we use custom infomation?  what does the long key 
  device_* means?
 
 custom properties for devices
 see the UI Custom Properties tab in Edit VM

I think that bigclouds meant the bizarre-looking properties that are
sent by Engine automatically, with no human interventions, per VmDevice that is
reported by Vdsm but not managed by Engine:

'device_82934f8e-7470-4c46-b9ed-d1c73fb17477device_87f6330a-278d-4b2a-9674-db58eb0e7023device_a19855be-70a9-49d0-9ea4-ee746b2d9739':
 'VmDevice {vmId=46914913-8bcf-4fa0-b1fb-3680ca8b6c16, 
deviceId=a19855be-70a9-49d0-9ea4-ee746b2d9739, device=virtio-serial, 
type=CONTROLLER, bootOrder=0, specParams={}, address={bus=0x00, domain=0x, 
type=pci, slot=0x06, function=0x0}, managed=false, plugged=true, 
readOnly=false, deviceAlias=virtio-serial0, customProperties={}, 
snapshotId=null}'

It was an attempt for future-compatibility, by (ab)using the custom properties
interface. Certainly Eli or Igor can explain the motivation better.

Dan.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] [Users] Trouble Upgrading (Fedora)

2013-12-08 Thread Dan Kenigsberg
On Fri, Dec 06, 2013 at 03:35:22PM -0500, Eyal Edri wrote:
 we only have 2 jobs (that i'm aware of) that installs vdsm:
 
 http://jenkins.ovirt.org/view/vdsm/job/vdsm_3.3_install_rpm_sanity_gerrit/
 http://jenkins.ovirt.org/view/vdsm/job/vdsm_master_install_rpm_sanity_gerrit/
 
 i changed 3.3 to remove python-cpopen at the start of the test. 

Thanks, but do you understand why
http://jenkins.ovirt.org/job/vdsm_3.3_install_rpm_sanity_gerrit/315/
(triggered after you've send this email) still fails?

Dan.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] [Users] Trouble Upgrading (Fedora)

2013-12-05 Thread Dan Kenigsberg
On Thu, Dec 05, 2013 at 09:13:02PM -0500, Douglas Schilling Landgraf wrote:
 On 12/05/2013 08:45 AM, Vinzenz Feenstra wrote:
 On 12/05/2013 01:49 PM, Dan Kenigsberg wrote:
 Joining two threads on that subject
 
 On Sun, Nov 24, 2013 at 11:32:55AM -0500, Yaniv Bronheim wrote:
 Hey,
 
 - Original Message -
 From: Nicholas Kesick cybertimber2...@hotmail.com
 To: oVirt Mailing List us...@ovirt.org
 Sent: Friday, November 22, 2013 6:55:23 PM
 Subject: [Users] Trouble Upgrading (Fedora)
 
 I am having trouble upgrading on Fedora 19 to 3.3.1.
 
 Two observations:
 1) I can't find ovirt-engine-3.3.1, even after a yum clear all; yum
 update.
 
 
 2) There is a dependency update circle involving
 python-cpopen-1.2.3-4.fc19.x86_64 and dsm-python-cpopen.x86_64
 0:4.13.0-11.fc19
 This issue is fine. Not so smoothly but we had to do that.
 cpopen package is published as part of vdsm code in ovirt-3.3 and
 before,
 On master branch we require python-cpopen formal package that is
 shipped with fedora\rhel currently .
 
 Each time you'll switch between newer version of current master (3.4)
 to ovirt-3.3 or older you will notice this report
 We should break this vicious cycle.
 
 I think it was intended to simplify upgrade from ovirt-3.4 to
 ovirt-3.3 on Jenkins slaves. But that's plain wrong on a wider context.
 The Jenkins issue should be solved explicitly by removing python-cpopen
 when vdsm-python-cpopen has to be installed.
 python-cpopen should obsolete vdsm-python-cpopen, solving OUR jenkins
 issues in the released rpms.
 
 I see this loop too.
 python-cpopen spec obsoletes vdsm-python-cpopen, so vdsm.spec don't
 need to obsolete python-cpopen. Otherwise, it will generate an
 infinity loop during the yum update between these packages.
 http://gerrit.ovirt.org/#/c/22113/
 
 
 Therefore +1 for solving that on jenkins
 +1

Douglas, before taking your patch to the ovirt-3.3 branch, would you
work with infra@ to get all their ovirt-3.3 vdsm jobs remove
python-cpopen before attempting to install vdsm?

Regards,
Dan.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] vdsm sync meeting December 2th

2013-12-03 Thread Dan Kenigsberg
On Mon, Dec 02, 2013 at 01:40:24PM -0500, Douglas Schilling Landgraf wrote:
 VDSM sync meeting Monday December 2th 2013
 =
 
 - Douglas going to make tests between VDSM and Fedora 20
 
 - Antoni will rework on the 'make dist check' patches.
 
 - Dan raised to have virt functional tests for checking vdsm
 failure/disconnect versus reconnect. Antoni going to check with
 Martin.
 
 - Federico going to review:
   tests: Add simple mocking library
   http://gerrit.ovirt.org/#/c/21155/
 
 - Antoni raised the thread:
   seobject vdsm master install
 
 https://lists.fedorahosted.org/pipermail/vdsm-devel/2013-November/002773.html
 
 - Yaniv going to check with Federico the ovirt-host-deploy/libvirt
 fallback problem.
 
 - Dan raised that we are going to increase the version of iscsi (rpm
 dependency)
 
 - Federico to review Douglas's doc about building new release:
 
 http://www.ovirt.org/Vdsm_Developers#Building_new_releases_for_Fedora_or_EPEL_with_fedpkg
 
 - tarball release - Yaniv is going to talk with infra team to check
 the best way to provide the vdsm tarball and patches. We never done
 it before, until now only provided the srpm/rpm builds.

I've filed https://fedorahosted.org/ovirt/ticket/99 to track this issue.

 
 - The below patch going to be verified:
   Removing hack for stopping libvirt-guests as conflict service
   http://gerrit.ovirt.org/#/c/21205/1
 
 - Fedora rpm branches and vdsm beta/stable builds.
   Yaniv and Douglas to push new beta 3.3.2 vdsm rpm into master branch.
 
 -- 
 Cheers
 Douglas
 ___
 vdsm-devel mailing list
 vdsm-devel@lists.fedorahosted.org
 https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] VDSM and Fedora 20 test

2013-12-03 Thread Dan Kenigsberg
On Tue, Dec 03, 2013 at 02:31:16PM +0100, Sandro Bonazzola wrote:
 Il 03/12/2013 17:20, Douglas Schilling Landgraf ha scritto:
  Hi,
  
  Below the initial tests that I have executed with vdsm version
  (4.13.0-11 and 4.13.0-12) on Fedora 20. Faced an 'iscsiadm: Error
  while adding record: no available memory' trying to connect to iscsi
  storage and opened a bugzilla [1] for it.

Thanks, Douglas!

  
  Engine and otopi
  =
  oVirt Engine Version: 3.4.0-0.2.master.20131129131413.git3f58634.fc19
  otopi-1.1.2-1.fc19

Testing this part of the project with f20 makes sense, too (but is less
likely to break).

  
  
  Hosts
  ===
  
  # cat /etc/fedora-release
  Fedora release 20 (Heisenbug)
  
  # getenforce
  Enforcing
  
  
  vdsm 4.13.0-11 (current stable)
  
  
  - Host deployed
  
  NFS:
  - Added NFS DATA and ISO [OK]
 
 have you tried to populate using iso-uploader / image-uploader?
 is log-collector able to collect logs for the traceback you've reported down 
 here?
 
 

snip

  ISCSI:
  
  Thread-498::ERROR::2013-12-03 
  00:33:29,199::hsm::2369::Storage.HSM::(connectStorageServer) Could not 
  connect to storageServer
  Traceback (most recent call last):
File /usr/share/vdsm/storage/hsm.py, line 2366, in connectStorageServer
  conObj.connect()
File /usr/share/vdsm/storage/storageServer.py, line 359, in connect
  iscsi.addIscsiNode(self._iface, self._target, self._cred)
File /usr/share/vdsm/storage/iscsi.py, line 150, in addIscsiNode
  iscsiadm.node_new(iface.name, portalStr, targetName)
File /usr/share/vdsm/storage/iscsiadm.py, line 239, in node_new
  raise IscsiNodeError(rc, out, err)
  IscsiNodeError: (3, [], ['iscsiadm: Error while adding record: no available 
  memory'])
  
  [1] F20: iscsiadm: Error while adding record: no available memory
  https://bugzilla.redhat.com/show_bug.cgi?id=1037602

To move ahead with block storage testing, could you try to rebuild
iscsiadm or run Vdsm with fibre channel?

Dan.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] VDSM and Fedora 20 test

2013-12-03 Thread Dan Kenigsberg
On Tue, Dec 03, 2013 at 05:03:26PM -0500, Douglas Schilling Landgraf wrote:
 
 I have found that iscsi-initiator-utils-6.2.0.873-11.fc20.x86_64 (or
 higher) introduces this error.
 
 I was able to create and run a virtual machine with the last build
 before the above one
 (iscsi-initiator-utils-6.2.0.873-9.fc20.x86_64).
 
 Also did manual tests with below versions and it worked as well:
 iscsi-initiator-utils-6.2.0.873-5.fc20.x86_64
 iscsi-initiator-utils-6.2.0.873-7.fc20.x86_64
 iscsi-initiator-utils-6.2.0.873-9.fc20.x86_64
 
 Finally, I have talked with iscsi-initiator-utils package maintainer
 and he is aware of the problem.

So, have you managed to fire up VMs with one of those installed?

A basic migration/hibernation test would be interesting to conduct, too.

Thanks Douglas!
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] Fedora 20 seobject vdsm master install

2013-12-02 Thread Dan Kenigsberg
On Tue, Nov 26, 2013 at 09:24:00AM -0500, Antoni Segura Puimedon wrote:
 Hi List,
 
 I just built master with some gerrit patch and when installing in Fedora20 I
 get the following:
 
   Installing : vdsm-python-4.13.0-179.gite51bbc3.fc20.x86_64  
   1/7 
   Installing : vdsm-xmlrpc-4.13.0-179.gite51bbc3.fc20.noarch  
   2/7 
   Installing : vdsm-python-zombiereaper-4.13.0-179.gite51bbc3.fc20.noarch 
   3/7 
   Installing : vdsm-4.13.0-179.gite51bbc3.fc20.x86_64 
   4/7 
 Traceback (most recent call last):
   File /usr/bin/vdsm-tool, line 145, in module
 sys.exit(main())
   File /usr/bin/vdsm-tool, line 142, in main
 return tool_command[cmd][command](*args[1:])
   File /usr/lib64/python2.7/site-packages/vdsm/tool/seboolsetup.py, 
 line 59, in sebool_config
 setup_booleans(True)
   File /usr/lib64/python2.7/site-packages/vdsm/tool/seboolsetup.py, 
 line 41, in setup_booleans
 sebool_obj = seobject.booleanRecords()
   File /usr/lib/python2.7/site-packages/seobject/__init__.py, line 
 2070, in __init__
 semanageRecords.__init__(self, store)
   File /usr/lib/python2.7/site-packages/seobject/__init__.py, line 205, 
 in __init__
 self.mylog = logger()
   File /usr/lib/python2.7/site-packages/seobject/__init__.py, line 90, 
 in __init__
 self.audit_fd = audit.audit_open()
 OSError: [Errno 93] Protocol not supported
 OSError: Protocol not supported
   Installing : vdsm-debug-plugin-4.13.0-179.gite51bbc3.fc20.noarch
   5/7 
   Installing : vdsm-reg-4.13.0-179.gite51bbc3.fc20.noarch 
   6/7 
   Installing : vdsm-cli-4.13.0-179.gite51bbc3.fc20.noarch 
   7/7 
   Verifying  : vdsm-debug-plugin-4.13.0-179.gite51bbc3.fc20.noarch
   1/7 
   Verifying  : vdsm-reg-4.13.0-179.gite51bbc3.fc20.noarch 
   2/7 
   Verifying  : vdsm-python-zombiereaper-4.13.0-179.gite51bbc3.fc20.noarch 
   3/7 
   Verifying  : vdsm-cli-4.13.0-179.gite51bbc3.fc20.noarch 
   4/7 
   Verifying  : vdsm-4.13.0-179.gite51bbc3.fc20.x86_64 
   5/7 
   Verifying  : vdsm-xmlrpc-4.13.0-179.gite51bbc3.fc20.noarch  
   6/7 
   Verifying  : vdsm-python-4.13.0-179.gite51bbc3.fc20.x86_64
 
 Shouldn't the auditing be optional (I have audit=0 in the kernel cmdline).

We've discussed this issue today in the vdsm sync meeting.
If make this optional, we must check for the flag existence on every
startup of vdsm. Presumably, this is not a long ardious task.

It would be good to check the flags even if we choose to keep the ugly
install-time errors: an admin should be warned that his VMs are not
going to start due to future selinux errors, and should be notified with
which vdsm-tool action (sebool-config) he should fix the issue.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] addNetwork to use existing nic conf

2013-11-27 Thread Dan Kenigsberg
On Wed, Nov 27, 2013 at 04:38:54AM -0500, Yedidyah Bar David wrote:
 Hi all,
 
 While looking at [1], and reading the source of bridge.py of 
 ovirt-hosted-engine-setup
 and of ovirt-host-deploy, I think the correct thing will be for addNetwork to 
 not
 require either bootproto=dhcp or ipaddr/netmask/etc., but to take this from 
 the existing
 conf of the nics to be put inside the bridge.
 
 Does this make sense?
 
 [1] https://bugzilla.redhat.com/show_bug.cgi?id=1013666

I do not think it does: addNetwork was designed to set up a network
accroding to a complete definition thereof. The current state and the
configuration of the underlying devices is intentionally overridden.

addNetwork's default for bootproto is not to set it in the ifcfg file
(which initscript translates to none, i.e. static). Other users of
addNetwork may already rely on that, and forcing them to change is not
realistic (particularly given our willingness to deprecate
addNetwork/delNetwork/editNetwork in favor of setupNetworks).

The suggested logic is particularly fuzzy if we consider the fact that
addNetwork may set multiple ifcfg files. Which bootproto should it
select as a default if it is building a bond on top of 3 nics?

I assume that hosted-engine-setup has an argument for selecting the
management interface. I see no way around having it receive
--ipaddr/--netmask/--gateway/--bootproto too, or doing the guestwork
itself.


___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] addNetwork to use existing nic conf

2013-11-27 Thread Dan Kenigsberg
On Wed, Nov 27, 2013 at 06:55:26AM -0500, Antoni Segura Puimedon wrote:
 
 
 - Original Message -
  From: Yedidyah Bar David d...@redhat.com
  To: Antoni Segura Puimedon asegu...@redhat.com
  Cc: vdsm-devel vdsm-de...@fedorahosted.org, Alon Bar-Lev 
  alo...@redhat.com, Sandro Bonazzola
  sbona...@redhat.com
  Sent: Wednesday, November 27, 2013 11:18:39 AM
  Subject: Re: addNetwork to use existing nic conf
  
  - Original Message -
   From: Antoni Segura Puimedon asegu...@redhat.com
   To: Yedidyah Bar David d...@redhat.com
   Cc: vdsm-devel vdsm-de...@fedorahosted.org, Alon Bar-Lev
   alo...@redhat.com, Sandro Bonazzola
   sbona...@redhat.com
   Sent: Wednesday, November 27, 2013 11:52:51 AM
   Subject: Re: addNetwork to use existing nic conf
   
   
   
   - Original Message -
From: Yedidyah Bar David d...@redhat.com
To: vdsm-devel vdsm-de...@fedorahosted.org
Cc: Alon Bar-Lev alo...@redhat.com, Sandro Bonazzola
sbona...@redhat.com, Antoni Segura Puimedon
asegu...@redhat.com
Sent: Wednesday, November 27, 2013 10:38:54 AM
Subject: addNetwork to use existing nic conf

Hi all,

While looking at [1], and reading the source of bridge.py of
ovirt-hosted-engine-setup
and of ovirt-host-deploy, I think the correct thing will be for
addNetwork
to
not
require either bootproto=dhcp or ipaddr/netmask/etc., but to take this
from
the existing
conf of the nics to be put inside the bridge.
   
   You mean the addNetwork in vdsm/vdsm/configNetwork.py?
  
  We currently use vdsClient - 'vdsClient addNetwork'.
  
   If that is the case I would very much prefer not to make such change, 
   since
   that
   would make addNetwork even farther from referential transparency.
   
   What I propose is that the calling code will use the new getBootProto
   introduced in:
   http://gerrit.ovirt.org/18484 (already merged).
  
  Does vdsClient support that?
 
 Well, vdsClient getVdsCaps returns the cfg information for all the devices,
 part of that will be the bootproto if it is defined on the ifcfg file of the
 device.

We'd very much like to drop the 'cfg' element reported per nic (it
breaks Vdsm's network abrstraction and would be anoying to maintain
with iproute2 when we move to it. BOOTPROTO is there, and would have
to stay there (for backward compatibility with Engine), but please avoid
using any other data.

Dan.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] addNetwork to use existing nic conf

2013-11-27 Thread Dan Kenigsberg
On Wed, Nov 27, 2013 at 07:28:23AM -0500, Antoni Segura Puimedon wrote:
 
 
 - Original Message -
  From: Dan Kenigsberg dan...@redhat.com
  To: Antoni Segura Puimedon asegu...@redhat.com
  Cc: Yedidyah Bar David d...@redhat.com, vdsm-devel 
  vdsm-de...@fedorahosted.org
  Sent: Wednesday, November 27, 2013 1:22:47 PM
  Subject: Re: [vdsm] addNetwork to use existing nic conf
  
  On Wed, Nov 27, 2013 at 06:55:26AM -0500, Antoni Segura Puimedon wrote:
   
   
   - Original Message -
From: Yedidyah Bar David d...@redhat.com
To: Antoni Segura Puimedon asegu...@redhat.com
Cc: vdsm-devel vdsm-de...@fedorahosted.org, Alon Bar-Lev
alo...@redhat.com, Sandro Bonazzola
sbona...@redhat.com
Sent: Wednesday, November 27, 2013 11:18:39 AM
Subject: Re: addNetwork to use existing nic conf

- Original Message -
 From: Antoni Segura Puimedon asegu...@redhat.com
 To: Yedidyah Bar David d...@redhat.com
 Cc: vdsm-devel vdsm-de...@fedorahosted.org, Alon Bar-Lev
 alo...@redhat.com, Sandro Bonazzola
 sbona...@redhat.com
 Sent: Wednesday, November 27, 2013 11:52:51 AM
 Subject: Re: addNetwork to use existing nic conf
 
 
 
 - Original Message -
  From: Yedidyah Bar David d...@redhat.com
  To: vdsm-devel vdsm-de...@fedorahosted.org
  Cc: Alon Bar-Lev alo...@redhat.com, Sandro Bonazzola
  sbona...@redhat.com, Antoni Segura Puimedon
  asegu...@redhat.com
  Sent: Wednesday, November 27, 2013 10:38:54 AM
  Subject: addNetwork to use existing nic conf
  
  Hi all,
  
  While looking at [1], and reading the source of bridge.py of
  ovirt-hosted-engine-setup
  and of ovirt-host-deploy, I think the correct thing will be for
  addNetwork
  to
  not
  require either bootproto=dhcp or ipaddr/netmask/etc., but to take
  this
  from
  the existing
  conf of the nics to be put inside the bridge.
 
 You mean the addNetwork in vdsm/vdsm/configNetwork.py?

We currently use vdsClient - 'vdsClient addNetwork'.

 If that is the case I would very much prefer not to make such change,
 since
 that
 would make addNetwork even farther from referential transparency.
 
 What I propose is that the calling code will use the new getBootProto
 introduced in:
 http://gerrit.ovirt.org/18484 (already merged).

Does vdsClient support that?
   
   Well, vdsClient getVdsCaps returns the cfg information for all the 
   devices,
   part of that will be the bootproto if it is defined on the ifcfg file of
   the
   device.
  
  We'd very much like to drop the 'cfg' element reported per nic (it
  breaks Vdsm's network abrstraction and would be anoying to maintain
  with iproute2 when we move to it. BOOTPROTO is there, and would have
  to stay there (for backward compatibility with Engine), but please avoid
  using any other data.
 
 Well, for installation time, there must be something (not necessarily part of
 vdsm proper that is able to get the ip configuration of the device we connect
 through.
 
 Since it could be ifcfg info (el6), NM (newer distros), 
 /etc/network/interfaces
 (debian derived) I would really make a separate solution to vdsm.

No doubt about that:
Bug 987813 - [RFE] report BOOTPROTO and BONDING_OPTS independent of
netdevice.cfg 
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] Pending Self-Hosted-Engine Patches in ovirt-3.3.2

2013-11-26 Thread Dan Kenigsberg
On Tue, Nov 26, 2013 at 05:18:49AM -0500, Yaniv Bronheim wrote:
 
 
 - Original Message -
  From: Sandro Bonazzola sbona...@redhat.com
  To: Dan Kenigsberg dan...@redhat.com, vdsm-de...@fedorahosted.org
  Cc: hat...@redhat.com
  Sent: Tuesday, November 26, 2013 8:50:38 AM
  Subject: Re: [vdsm] Pending Self-Hosted-Engine Patches in ovirt-3.3.2
  
  Il 25/11/2013 22:16, Dan Kenigsberg ha scritto:
   An important feature of ovirt-3.3 has not made it to the ovirt-3.3.0
   deadline: http://www.ovirt.org/Features/Self_Hosted_Engine.
   
   It would allow to run Engine as a VM on top one of the hosts controlled
   by itself, saving resources and allowing high availablity
   out-of-the-box.
   
   The feature was not ready on time for the ovirt-3.3.0 release but now
   its Engine-side patches are merged, and its Vdsm-side patches are
   pending aproval for entry into Vdsm's stable branch.
   
   http://gerrit.ovirt.org/20189
   http://gerrit.ovirt.org/20190
   http://gerrit.ovirt.org/20191
   http://gerrit.ovirt.org/20192
   http://gerrit.ovirt.org/20193
   http://gerrit.ovirt.org/20194
   http://gerrit.ovirt.org/20209
   http://gerrit.ovirt.org/20300
   http://gerrit.ovirt.org/21357
   
   I suggest to take the pending patches into the ovirt-3.3 branch, and
   ship ovirt-3.3.2 with this feature enabled.
  
  +1
  
   We shall not release
   ovirt-3.3.2 before we are assured by quality engineering that we have
   no regression to main Vdsm code. If we do not receive this green light,
   and there's urgency to release ovirt-3.3.2, we'd fork and ship 3.3.2 only
   with
   the urgent fixes.
   
   Any objections?
 
 not from my side. waiting for the green light

I'm suggesting not to wait with merging. A CR+2 V+1 should be enough for
those patches. The release of ovirt-3.3.2 should wait, though.

 are there more pending urgent fixes that I'm not aware of?

You may be aware of http://gerrit.ovirt.org/#/c/21700/ but
I'm sure there would be more.

Dan.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


[vdsm] Pending Self-Hosted-Engine Patches in ovirt-3.3.2

2013-11-25 Thread Dan Kenigsberg
An important feature of ovirt-3.3 has not made it to the ovirt-3.3.0
deadline: http://www.ovirt.org/Features/Self_Hosted_Engine.

It would allow to run Engine as a VM on top one of the hosts controlled
by itself, saving resources and allowing high availablity
out-of-the-box.

The feature was not ready on time for the ovirt-3.3.0 release but now
its Engine-side patches are merged, and its Vdsm-side patches are
pending aproval for entry into Vdsm's stable branch.

http://gerrit.ovirt.org/20189
http://gerrit.ovirt.org/20190
http://gerrit.ovirt.org/20191
http://gerrit.ovirt.org/20192
http://gerrit.ovirt.org/20193
http://gerrit.ovirt.org/20194
http://gerrit.ovirt.org/20209
http://gerrit.ovirt.org/20300
http://gerrit.ovirt.org/21357

I suggest to take the pending patches into the ovirt-3.3 branch, and
ship ovirt-3.3.2 with this feature enabled. We shall not release
ovirt-3.3.2 before we are assured by quality engineering that we have
no regression to main Vdsm code. If we do not receive this green light,
and there's urgency to release ovirt-3.3.2, we'd fork and ship 3.3.2 only with
the urgent fixes.

Any objections?

Dan.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] [Engine-devel] oVirt 3.3.2 beta status

2013-11-18 Thread Dan Kenigsberg
On Mon, Nov 18, 2013 at 10:12:02AM +0100, Sandro Bonazzola wrote:
 Hi,
 
 we're going to branch and build oVirt 3.3.2 beta on Nov 27th.
 A bug tracker is available at [1] and it shows only 2 bugs blocking the 
 release:
 
 Bug 1029792 - VDSM does not report the qemu version in capabilities, if 
 qemu-kvm-rhev is used

Backported
http://gerrit.ovirt.org/21363
http://gerrit.ovirt.org/21364
to ovirt-3.3 branch to address this request.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


[vdsm] vdsm sync meeting minutes - Nov 18

2013-11-18 Thread Dan Kenigsberg
Things we've discussed

- pending iscsi dependency (fixed master branch since meeting)
  http://gerrit.ovirt.org/21385

- We should strive to keep master usable by el6 (and other supported
  platforms).

  We forgot to discuss the issue blocking our spec-checking jenkins job:
  cpopen collision. Yaniv, are there any updates on that?

- function storage/virt tests needs a parent and extension. Ayal says
  that A storage developer was assigned to that task, expect improvement
  within weeks.

- Vdsm is asked to report information about various VM-related long
  processes: VM migration, live storage migration, live merge. There was
  a universal agreement that getAllVmStats is a reasonable place to add
  those. Michal, Saggi, would it be fine?

- failing `make distcheck`: Toni to try monkey-patching the constants
  module when the schema is built.

- Nir's mock module needs review, and later - a decision. There was a
  disagreement regarding its necessity/helpfulness.

See ya!
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] [Users] Low quality of el6 vdsm rpms

2013-11-12 Thread Dan Kenigsberg
On Tue, Nov 12, 2013 at 11:31:04AM +0100, Sandro Bonazzola wrote:
 Il 12/11/2013 10:34, Patrick Hurrelmann ha scritto:
  Hi all,
  
  sorry for this rant, but...

Thanks for ranting. Community testing and ranting are to be cherished.
We must improve in the points you have raised.

  
  I now tried several times to test the beta 3.3.1 rpms, but they can't
  even be installed in the most times.
 
 I'm glad to read you're testing 3.3.1. May I ask you to add yourself to
 http://www.ovirt.org/Testing/Ovirt_3.3.1_testing ?
 
  One time it required a future
  selinux-policy, although the needed selinux fix was delivered in a much
  lower version. Now the rpms have broken requirements. It requires
  hostname instead of /bin/hostname. This broken requirement is not
  included in the vdsm 3.3 branch, so I wonder where it comes from?
  Anyway. So I proceeded and tried to build vdsm myself once again.
  Currently the build fails with (but worked fine some days ago):
  
  /usr/bin/pep8 --exclude=config.py,constants.py --filename '*.py,*.py.in' \
  client lib/cpopen/*.py lib/vdsm/*.py lib/vdsm/*.py.in tests
  vds_bootstrap vdsm-tool vdsm/*.py vdsm/*.py.in vdsm/netconf
  vdsm/sos/vdsm.py.in vdsm/storage vdsm/vdsm vdsm_api vdsm_hooks vdsm_reg
  vdsm/storage/imageRepository/formatConverter.py:280:29: E128
  continuation line under-indented for visual indent
  
  
  - How can the quality of the vdsm builds be increased? It is frustrating
  to spend time on testing and then the hosts cannot even be installed to
  broken vdsm rpms.

I suspect you are not interested in excuses for each of the failures,
let us look forwards. My conclusions are:
- Do not require non-yet-existing rpms. If we require a feature that is
  not yet in Fedora/Centos, we must wait. This is already in effect, see
  for example http://gerrit.ovirt.org/#/c/20248/ and
  http://gerrit.ovirt.org/19545

- There's a Jenkins job to enforce the former requirement of spec
  requirement. David, Sandro, any idea why it is not running these days?

- Keep the docs updated. Our Jenkins slaves have pep8-1.4.6, so we
  should update
  http://www.ovirt.org/Vdsm_Developers#Installing_required_packages
  accordingly - and more importantly, make that version available.

  Sandro, who built the python-pep8-1.4.6 that sits on the el6 Jenkins
  slave? Could you make it publicly available? (I can volunteer
  http://danken.fedorapeople.org again)

  - How are the builds prepared? Is there a Jenkins job that prepares
  stable rpms in addition to the nightly job? Or is this totally
  handcrafted?
  - How can it be that the rpm spec differs between the 3.3 branch and
  released rpms? What is the source/branch for el6 vdsm rpms? Maybe I'm
  just tracking on the wrong source tree...

Based on your reports, you are tracking the correct tree; but please
describe which differences do you see (and between what releases
exactly).

Regards,
Dan.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] [Users] Low quality of el6 vdsm rpms

2013-11-12 Thread Dan Kenigsberg
On Tue, Nov 12, 2013 at 10:31:24AM -0500, Eyal Edri wrote:
 

snip

  
  - There's a Jenkins job to enforce the former requirement of spec
requirement. David, Sandro, any idea why it is not running these days?
 
 it does run, but we can't enable it since it's still failing:
 http://jenkins.ovirt.org/job/vdsm_3.3_install_rpm_sanity_gerrit/label=fedora19/282/console
 
 once that's fixed, the job will be enabled and run per patch.

Yaniv, any idea why we still have
file /usr/lib64/python2.7/site-packages/cpopen/__init__.py from install of 
vdsm-python-cpopen-4.12.1-4.fc19.x86_64 conflicts with file from package 
python-cpopen-1.2.3-2.fc19.x86_64
?

 
  
  - Keep the docs updated. Our Jenkins slaves have pep8-1.4.6, so we
should update
http://www.ovirt.org/Vdsm_Developers#Installing_required_packages
accordingly - and more importantly, make that version available.
  
Sandro, who built the python-pep8-1.4.6 that sits on the el6 Jenkins
slave? Could you make it publicly available? (I can volunteer
http://danken.fedorapeople.org again)
 
 i tend to agree here with patrick on using non released pep8 packages, that 
 are not available
 via rpms, but only via python-pip. 
 the jenkins slaves were updated via pyhon-pip and not via yum upgrade. 

Eyal, I do not understand your opinion here... In any case, I've added
http://danken.fedorapeople.org/python-pep8-1.4.5-2.el6.noarch.rpm if
anyone find this unsigned unsupported hardly tested package useful.

Dan.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] [Users] Low quality of el6 vdsm rpms

2013-11-12 Thread Dan Kenigsberg
On Tue, Nov 12, 2013 at 04:41:53PM +, Dan Kenigsberg wrote:
 On Tue, Nov 12, 2013 at 10:31:24AM -0500, Eyal Edri wrote:
  
 
 snip
 
   - Keep the docs updated. Our Jenkins slaves have pep8-1.4.6, so we
 should update
 http://www.ovirt.org/Vdsm_Developers#Installing_required_packages
 accordingly - and more importantly, make that version available.
   
 Sandro, who built the python-pep8-1.4.6 that sits on the el6 Jenkins
 slave? Could you make it publicly available? (I can volunteer
 http://danken.fedorapeople.org again)
  
  i tend to agree here with patrick on using non released pep8 packages, that 
  are not available
  via rpms, but only via python-pip. 
  the jenkins slaves were updated via pyhon-pip and not via yum upgrade. 
 
 Eyal, I do not understand your opinion here... In any case, I've added
 http://danken.fedorapeople.org/python-pep8-1.4.5-2.el6.noarch.rpm if
 anyone find this unsigned unsupported hardly tested package useful.

A concerned user just asked me how did I build this. I just took the
source from koji,
http://kojipkgs.fedoraproject.org//packages/python-pep8/1.4.5/2.fc20/src/python-pep8-1.4.5-2.fc20.src.rpm
and built it on an el6 host with
http://kojipkgs.fedoraproject.org//packages/python-sphinx/1.0.8/1.el6/noarch/python-sphinx-1.0.8-1.el6.noarch.rpm
 installed.

Dan.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] [Users] Low quality of el6 vdsm rpms

2013-11-12 Thread Dan Kenigsberg
On Tue, Nov 12, 2013 at 03:51:36PM -0500, Douglas Schilling Landgraf wrote:
 
 Indeed, that's bad. It has been included from a patch only on Fedora
 koji build rawhide. The others points here already have been
 answered by others developers. Anyway, we have updated the build.
 
 We would appreciate if you could continue the tests with the new test build:
 
 F19: http://koji.fedoraproject.org/koji/taskinfo?taskID=6172359
 F20: http://koji.fedoraproject.org/koji/taskinfo?taskID=6172521
 EL6: http://koji.fedoraproject.org/koji/taskinfo?taskID=6172612
 
 This last update includes the following patches:
 - The require hostname fix

(http://gerrit.ovirt.org/#/c/21194/ should have been sent to gerrit long
ago, I suppose)

___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] [Engine-devel] daily oVirt 3.3.1 blocker status

2013-11-11 Thread Dan Kenigsberg
On Mon, Nov 11, 2013 at 10:26:19AM +0100, Sandro Bonazzola wrote:
 Il 08/11/2013 01:28, Dan Kenigsberg ha scritto:
  On Wed, Nov 06, 2013 at 02:36:08AM -0500, Eduardo Warszawski wrote:
  Hi,
  only one bloker remains to be fixed:
  Bug 1022961 - Running a VM from a gluster domain uses mount instead of
  gluster URI
  I don't see any activity about it.
  Can anybody join Eduardo for fixing it ASAP?
 
  I'm working on it. 
  Hope soon we will have a fix.
  Don't panic! :)
  
  It's not the /real thing/ that Eduardo wanted, but it's something and
  it's working. Please review and verify
  
  http://gerrit.ovirt.org/#/c/21059/
  gluster prepareImage: return gluster-sepecific information
  
 
 Hi,
 Bug 1022961 - Running a VM from a gluster domain uses mount instead of 
 gluster URI
 is the only bug blocking 3.3.1 and is on POST.
 the related patch http://gerrit.ovirt.org/21059 is marked as verified and has 
 two +1.
 Can you merge the patch and build VDSM for having it under testing so we can 
 try to release 3.3.1 this week?

I'd love to merge this in, but given this being my own code, I prefer
another core developer ack it, even though no one really likes it.

Eduardo dislikes the fact that this patch maintains the awkward API of
prepareImage returning this 'info' dictionary on top of the simple
'path' it used to return before the glusterSD effort.

Federico dislikes the fact that this patch adds gluster-specific
code into our top-level class HSM. To relieve this, I am now suggesting
three other cleanup patches
   http://gerrit.ovirt.org/21122
   http://gerrit.ovirt.org/21123
   http://gerrit.ovirt.org/21124
but I would like to take the simple fix first. Unbreaking a major
feature in master and ovirt-3.3.1 should have precedence over
architectural debates.

Dan.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] Host local_host running without virtualization hardware acceleration

2013-11-05 Thread Dan Kenigsberg
On Tue, Nov 05, 2013 at 09:08:33AM +0100, Sandro Bonazzola wrote:
 Hi,
 I had to reinstall ovirt yesterday and now it seems that it doesn't work 
 anymore.
 I'm running nightly on Fedora 18.
  kernel-3.11.4-101.fc18.x86_64
  sanlock-2.8-1.fc18.x86_64
  libvirt-1.1.4-1.fc18.x86_64
  qemu-1.5.1-1.fc18.x86_64
  vdsm-4.13.0-93.gitea8c8f0.fc18.x86_64
  ovirt-engine-3.4.0-0.2.master.20131104192919.git3b65870.fc18.noarch
 
 engine-setup with all-in-one detects hardware virtualization and allow me to 
 configure the system.
 (it fails detecting engine health status due probably to recent changes in 
 its URL, I'm already looking into it)
 
 Once added localhost to the engine, it has been moved to non operational mode 
 saying
 I don't have virtualization hardware acceleration anymore.
 
 I've found that:
 
 # modinfo kvm
 filename:   /lib/modules/3.11.4-101.fc18.x86_64/kernel/arch/x86/kvm/kvm.ko
 license:GPL
 author: Qumranet
 depends:
 intree: Y
 vermagic:   3.11.4-101.fc18.x86_64 SMP mod_unload
 parm:   min_timer_period_us:uint
 parm:   ignore_msrs:bool
 parm:   tsc_tolerance_ppm:uint
 parm:   allow_unsafe_assigned_interrupts:Enable device assignment on 
 platforms without interrupt remapping support. (bool)
 
 # /usr/bin/qemu-kvm
 Could not access KVM kernel module: No such file or directory
 failed to initialize KVM: No such file or directory
 
 looking at strace:
 open(/dev/kvm, O_RDWR|O_CLOEXEC)  = -1 ENOENT (No such file or 
 directory)
 
 Any clue on what may be happened?

What's your `lsmod | grep kvm` ?
If empty, does `modprobe kvm_intel` trigger tail-telling lines on
/var/log/message?

Dan.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] [Engine-devel] daily oVirt 3.3.1 blocker status

2013-11-04 Thread Dan Kenigsberg
On Mon, Nov 04, 2013 at 09:05:19AM +0100, Sandro Bonazzola wrote:
 Il 31/10/2013 08:49, Sandro Bonazzola ha scritto:
  Hi,
  The following blockers are still not fixed:
  VDSM:
  Bug 1022961 - Running a VM from a gluster domain uses mount instead of 
  gluster URI
  Bug 1022975 - [vdsm] storage domain upgrade fails with attributeError
  
  Federico, Eduardo, can you provide an ETA for those?
 
 Still not fixed with no ETA.

WWIW, Bug 1022975 is solved by http://gerrit.ovirt.org/20721; waiting
only for a formal verification.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] Problem building master on rhel-6.4 according ovirt.org

2013-11-02 Thread Dan Kenigsberg
On Thu, Oct 31, 2013 at 07:07:08AM -0400, Mooli Tayer wrote:
 While building master on a rhel-6.4 according to instructions at ovirt.org [1]
 I encountered several issues:
 
 1.)This line:
 yum install --enablerepo=ovirt-beta x86_64/* noarch/vdsm-xml* noarch/vdsm-cli
 
 Is missing:  noarch/vdsm-python-zombiereaper*
 
 Is that correct? I'll fix it if so
 
 
 2.) we need the gluster 3.4.1 repos which we get, but we also get 3.4.0 (from 
 resources.ovirt.org)
 so after following instructions we end up with glusterfs-epel.repo.1 with 
 3.4.1 and glusterfs-epel.repo with 3.4.0
 
 Is this a problem? are newer versions considered?

Thanks for fixing the wiki for 1 and 2.

 
 3.)  selinux-policy-targeted = 3.7.19-195.el6.13 is missing (known issue I 
 believe)
 It also can be found on-line, can we still build somehow?

This package has been released by Red Hat recently, coming Centos way
anytime soon. You can take my http://gerrit.ovirt.org/#/c/20313/ hack if
you do not want to use --force when installing vdsm.

 
 4.) this issue, which I don't get:
 
 Error: Package: libvirt-lock-sanlock-0.10.2-18.el6.x86_64 (RHEL-6.4-optional)
Requires: libvirt = 0.10.2-18.el6
Available: libvirt-0.10.2-18.el6.x86_64 (RHEL-6.4)
libvirt = 0.10.2-18.el6
Installing: libvirt-0.10.2-18.el6_4.14.x86_64 (RHEL-6-updates)
libvirt = 0.10.2-18.el6_4.14

Red Hat ships libvirt-lock-sanlock in the EL optional channel, not in
the basic platform one. You have an old libvirt-lock-sanlock installed on your
host (from optional), and a newer libvirt is available on your
platform channel. However, their versions must match. So you can
either wait for a newer libvirt-lock-sanlock, or ignore the error and
keep the older libvirt.

Dan.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] vdsm Requires: selinux-policy-targeted = 3.7.19-213

2013-10-28 Thread Dan Kenigsberg
On Mon, Oct 28, 2013 at 10:36:17AM +0100, Fabian Deutsch wrote:
 Am Montag, den 28.10.2013, 01:23 + schrieb Dan Kenigsberg:
  On Sun, Oct 27, 2013 at 02:45:43PM +0100, Fabian Deutsch wrote:
   Hey,
   
   when I try to install vdsm, I get:
  
  Would you be so kind to verify my fix for this
  http://gerrit.ovirt.org/#/c/20313/
  on ovirt node?
 
 Dan,
 
 I've done a draft build and your basically works (the policy wihtin the
 image is modified), only restorceon needs to be run on /var/run/vdsm .

Oh, right. Thanks. Could you have another look?
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] vdsm sync meeting - October 21th 2013

2013-10-22 Thread Dan Kenigsberg
On Mon, Oct 21, 2013 at 11:48:45AM -0400, Yaniv Bronheim wrote:
 Vdsm Sync Meeting - October 21th 2013
 
 
  * Functional unit tests fail due to an old python-noise version on EL6:
   * Tony submitted clean up script before tests as workaround.
   * Old noise version is missing noise_xml, which is still a mystery how it 
 worked before. Seems that old python-noise ignores xunit file and does not 
 produce it at all (correct me if I missed or wrong on the details here)
   * Kiril should update python-noise on slaves.

Thanks Kiril and infra in general. The issue was that EPEL6's
python-nose does not support xml output, hence a newer version was
installed via pip on top of the rpm.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] Recent changes in vdsmd startup

2013-10-13 Thread Dan Kenigsberg
On Sat, Oct 12, 2013 at 05:26:53PM -0400, Yaniv Bronheim wrote:
 
 
 - Original Message -
  From: Dan Kenigsberg dan...@redhat.com
  To: Antoni Segura Puimedon asegu...@redhat.com
  Cc: arch a...@ovirt.org, VDSM Project Development 
  vdsm-devel@lists.fedorahosted.org
  Sent: Thursday, October 10, 2013 11:51:30 PM
  Subject: Re: [vdsm] Recent changes in vdsmd startup
  
  On Thu, Oct 10, 2013 at 04:40:36PM -0400, Antoni Segura Puimedon wrote:
   Is this what happened with the network functional testing? After these
   last changes, before the yum install of the new vdsm and the yum erase
   of the old I had to add the vdsm-tool libvirt reconfigure and start.
   Otherwise vdsm would not be able to reply to VdsGetCaps (on f19, on el6
   somehow it didn't happen).
  
  IMO what you've experienced is the intended change that Yaniv has
  reported in this thread. Fresh installation of vdsm now requires
  more-than-just `service vdsmd start`.
  
  I'm trying to understand our action on the unattended upgrade path.
 
 Interesting point. Thanks for raising that up, raising such issues was part 
 of this mail goals
 You are right that after an upgrade we will need to restart libvirt service 
 manually to start working with new configuration,
 But, AFAIK we don't force vdsmd restart after upgrading.

Actually, we have to restart (you cannot expect the old process to keep
on running porperly after its data/code files have changed underneath)
and we do, by virtue of
/sbin/service vdsmd condrestart
in the %post script.

 If the user updated the package and still wants the new configuration,
 manual vdsmd restart will be required anyway. So the print to restart also 
 related service is just enough.

I'm afraid we're ain't so lucky...

Dan.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] [Users] Resizing disks destroys contents

2013-10-12 Thread Dan Kenigsberg
On Sat, Oct 12, 2013 at 11:07:07PM +0330, Mohsen Saeedi wrote:
We should update both of vdsm and ovirt-engine or vdsm is enough? I check
ovirt el repo and it does not have new version for vdsm.
On 10/12/2013 02:11 AM, Dan Kenigsberg wrote:

Only a vdsm update is required for resolving this issue.
Apparently, vdsm-4.12.1-4 is still in testing stage
http://resources.ovirt.org/releases/updates-testing/rpm/EL/6/x86_64/

Please try it out!

 
  On Sat, Oct 12, 2013 at 12:27:02AM +0330, Mohsen Saeedi wrote:
 
 I have same problem too.   I have one node as ovirt engine and as
 hypervisor too. I successfully upgrade our engine from 3.1 to 3.2 and then
 from 3.2 to 3.3 and everything is ok.
 then i upgrade our cluster version to 3.3 and then our datacenter to 3.3.
 I increase 1GB on virtual machine with 20GB thin disk  and I receive
 success message on event and our disk increase to 21GB but after host
 reboot, i saw message that contains no bootable disk is available!!!
  then i try to install new os on that disk (Centos 6.4 minimal) and
 anaconda show the error on screen : No usable disks have found. I think
 disk corrupt after resizing.
 it was test VM but it's very if occur for real and functional VMs.
 
  Yes, this is indeed a bad bad bug. It was inadvertantly fixed on the
  master branch a while ago. Yesterday, we've backported the fix to
  3.3.0.1
 
  http://lists.ovirt.org/pipermail/users/2013-October/017018.html
 
  Please avoid trying ovirt-3.3.0's new disk resize feature on file-based 
 storage,
  unless you upgrade to vdsm vdsm-4.12.1-4 first.
 
  Dan.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] [Users] Resizing disks destroys contents

2013-10-11 Thread Dan Kenigsberg
On Sat, Oct 12, 2013 at 12:27:02AM +0330, Mohsen Saeedi wrote:
I have same problem too.   I have one node as ovirt engine and as
hypervisor too. I successfully upgrade our engine from 3.1 to 3.2 and then
from 3.2 to 3.3 and everything is ok.
then i upgrade our cluster version to 3.3 and then our datacenter to 3.3.
I increase 1GB on virtual machine with 20GB thin disk  and I receive
success message on event and our disk increase to 21GB but after host
reboot, i saw message that contains no bootable disk is available!!!
 then i try to install new os on that disk (Centos 6.4 minimal) and
anaconda show the error on screen : No usable disks have found. I think
disk corrupt after resizing.
it was test VM but it's very if occur for real and functional VMs.

Yes, this is indeed a bad bad bug. It was inadvertantly fixed on the
master branch a while ago. Yesterday, we've backported the fix to
3.3.0.1

http://lists.ovirt.org/pipermail/users/2013-October/017018.html

Please avoid trying ovirt-3.3.0's new disk resize feature on file-based storage,
unless you upgrade to vdsm vdsm-4.12.1-4 first.

Dan.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] vdsm rpm sanity test is failing

2013-10-10 Thread Dan Kenigsberg
On Thu, Oct 10, 2013 at 08:48:00AM -0400, Eyal Edri wrote:
 
 
 - Original Message -
  From: Dan Kenigsberg dan...@redhat.com
  To: Aravinda avish...@redhat.com
  Cc: vdsm-devel@lists.fedorahosted.org, ee...@redhat.com, ybron...@redhat.com
  Sent: Thursday, October 10, 2013 3:43:29 PM
  Subject: Re: [vdsm] vdsm rpm sanity test is failing
  
  On Thu, Oct 10, 2013 at 11:39:51AM +0530, Aravinda wrote:
   Hi,
   
   I sent a patch to vdsm in gerrit, but rpm sanity test is failing.
   Please suggest how to fix this.
   
   patch: http://gerrit.ovirt.org/#/c/17822/6
   sanity test:
   http://jenkins.ovirt.org/job/vdsm_install_rpm_sanity_gerrit/148/
  
  Can we install  glusterfs-3.4.1-2.el6 from
  https://koji.fedoraproject.org/koji/buildinfo?buildID=468826
  on our el6 slave?
 
 i rather have a stable repo to use constantly and not specific pkg in koji. 

I think that we should take the glusterfs straight from Fedora EPEL6, as
I would like to avoid shipping our own glusterfs in the ovirt repo.

My request for a manaual fix is only to unbreak current situation. It is
important to do that in order to ubreak the new rpm sanity test, which
is going to save us this trouble in the future.

Regards,
Dan.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] Recent changes in vdsmd startup

2013-10-10 Thread Dan Kenigsberg
On Thu, Oct 10, 2013 at 01:51:00PM -0400, Alon Bar-Lev wrote:
 
 
 - Original Message -
  From: Itamar Heim ih...@redhat.com
  To: Alon Bar-Lev alo...@redhat.com
  Cc: Yaniv Bronheim ybron...@redhat.com, arch a...@ovirt.org, VDSM 
  Project Development
  vdsm-devel@lists.fedorahosted.org
  Sent: Thursday, October 10, 2013 8:33:39 PM
  Subject: Re: [vdsm] Recent changes in vdsmd startup
  
  On 10/10/2013 07:47 PM, Alon Bar-Lev wrote:
  
  
   - Original Message -
   From: Itamar Heim ih...@redhat.com
   To: Alon Bar-Lev alo...@redhat.com
   Cc: Yaniv Bronheim ybron...@redhat.com, arch a...@ovirt.org, 
   VDSM
   Project Development
   vdsm-devel@lists.fedorahosted.org
   Sent: Thursday, October 10, 2013 7:45:36 PM
   Subject: Re: [vdsm] Recent changes in vdsmd startup
  
   On 10/10/2013 07:43 PM, Alon Bar-Lev wrote:
  
  
   - Original Message -
   From: Itamar Heim ih...@redhat.com
   To: Alon Bar-Lev alo...@redhat.com
   Cc: Yaniv Bronheim ybron...@redhat.com, arch a...@ovirt.org,
   VDSM
   Project Development
   vdsm-devel@lists.fedorahosted.org
   Sent: Thursday, October 10, 2013 7:39:35 PM
   Subject: Re: [vdsm] Recent changes in vdsmd startup
  
   On 10/10/2013 07:38 PM, Alon Bar-Lev wrote:
  
  
   - Original Message -
   From: Itamar Heim ih...@redhat.com
   To: Yaniv Bronheim ybron...@redhat.com
   Cc: arch a...@ovirt.org, VDSM Project Development
   vdsm-devel@lists.fedorahosted.org
   Sent: Thursday, October 10, 2013 7:37:14 PM
   Subject: Re: [vdsm] Recent changes in vdsmd startup
  
   On 10/10/2013 05:38 PM, Yaniv Bronheim wrote:
  
  
   - Original Message -
   From: Itamar Heim ih...@redhat.com
   To: Yaniv Bronheim ybron...@redhat.com
   Cc: VDSM Project Development vdsm-devel@lists.fedorahosted.org,
   arch
   a...@ovirt.org
   Sent: Thursday, October 10, 2013 5:24:40 PM
   Subject: Re: Recent changes in vdsmd startup
  
   On 10/10/2013 04:32 PM, Yaniv Bronheim wrote:
   Hey Everybody,
   FYI, Recently we merged a fix that changes the way vdsmd starts.
   Before
   that service vdsmd start command performed its main logic in
   addition
   to
   related services manipulation, as configure libvirt service and
   restart
   it
   for example.
   Now we are trying to avoid that and only alert the user if restart
   or
   other
   operations on related services are required.
  
   So now when you perform vdsmd start after clear installation you
   will
   see:
   ~$ sudo service vdsmd restart
   Shutting down vdsm daemon:
   vdsm watchdog stop [  OK  
   ]
   vdsm: not running  
   [FAILED]
   vdsm: Running run_final_hooks
   vdsm stop  [  OK  
   ]
   supervdsm start[  OK  
   ]
   vdsm: Running run_init_hooks
   vdsm: Running gencerts
   hostname: Unknown host
   vdsm: Running check_libvirt_configure
   libvirt is not configured for vdsm yet
   Perform 'vdsm-tool libvirt-configure' before starting vdsmd
   vdsm: failed to execute check_libvirt_configure, error code 1
   vdsm start 
   [FAILED]
  
   This asks you to run vdsm-tool libvirt-configure
   After running it you should see:
  
   ~$ sudo vdsm-tool libvirt-configure
   Stopping libvirtd daemon: [  OK  ]
   libvirt is not configured for vdsm yet
   Reconfiguration of libvirt is done.
  
   To start working with the new configuration, execute:
   'vdsm-tool libvirt-configure-services-restart'
   This will manage restarting of the following services:
   libvirtd, supervdsmd
  
   After performing: 'vdsm-tool libvirt-configure-services-restart'
   you
   are
   ready to start vdsmd again as usual.
  
   All those vdsm-tool commands require root privileges, otherwise
   it'll
   alert
   and stop the operation.
   Over systemd the errors\output can be watched in /var/log/messages
  
   Thanks,
   Yaniv Bronhaim.
   ___
   Arch mailing list
   a...@ovirt.org
   http://lists.ovirt.org/mailman/listinfo/arch
  
  
   how will this affect the following use cases:
  
   1. i added a new host to the system via the engine.
   at the end of the installation i expect the host to work without
   admin
   having to do any operation on the host.
  
   2. i update a host to latest vdsm version via the engine.
   at the end of the update i expect the host to be updated without
   admin
   having to do any operation on the host

I think we've quite neglected a third case, of unattended `yum upgrade
-y`.

If and when we need to update BY_VDSM_VERS, this would result in a
failure to restart vdsm. How should we handle that? I do not see a way
other than adding libvirt-configure and libvirt condrestart to vdsm
%postin.

Dan.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org

Re: [vdsm] vdsm-4.12.1-4

2013-10-10 Thread Dan Kenigsberg
On Thu, Oct 10, 2013 at 10:13:33AM -0400, Mike Burns wrote:
 On 10/10/2013 10:12 AM, Mike Burns wrote:
 On 10/10/2013 10:20 AM, Douglas Schilling Landgraf wrote:
 Changes:
 =
 - remoteFileHandler: Add create exclusive option for truncateFile
 (BZ#979193)
 - oop: improve safety for truncateFile
 
 F19: http://koji.fedoraproject.org/koji/taskinfo?taskID=6046682
 F20: http://koji.fedoraproject.org/koji/taskinfo?taskID=6046573
 EL6: http://koji.fedoraproject.org/koji/taskinfo?taskID=6046734
 
 
 Done
 
 Done as in they're uploaded to updates-testing on ovirt.org

Thanks! As Federico reported earlier, they are quite important.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] Recent changes in vdsmd startup

2013-10-10 Thread Dan Kenigsberg
On Thu, Oct 10, 2013 at 04:40:36PM -0400, Antoni Segura Puimedon wrote:
 Is this what happened with the network functional testing? After these
 last changes, before the yum install of the new vdsm and the yum erase
 of the old I had to add the vdsm-tool libvirt reconfigure and start.
 Otherwise vdsm would not be able to reply to VdsGetCaps (on f19, on el6
 somehow it didn't happen).

IMO what you've experienced is the intended change that Yaniv has
reported in this thread. Fresh installation of vdsm now requires
more-than-just `service vdsmd start`.

I'm trying to understand our action on the unattended upgrade path.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] vdsm sync meeting - October 7th 2013

2013-10-09 Thread Dan Kenigsberg
On Wed, Oct 09, 2013 at 08:27:58AM -0400, Mike Burns wrote:
 On 10/08/2013 04:52 PM, Douglas Schilling Landgraf wrote:
 On 10/08/2013 05:08 AM, Dan Kenigsberg wrote:
 On Mon, Oct 07, 2013 at 03:25:22PM +0100, Dan Kenigsberg wrote:
 We had an unpleasant talk, hampered by statics and disconnection on
 danken's side. Beyond the noises I've managed to recognize Yaniv, Toni,
 Douglas, Danken, Ayal, Timothy, Yeela and Mooli. We've managed to
 discuss:
 
 I've forgot to mention the resolution of
 
  Bug 1007980 - Failed to run VM with gluster drives: internal error
  unexpected address type for ide disk
 
 by http://gerrit.ovirt.org/19949 . Due to its being a serious painpoint
 to users. I've backport this fix to both ovirt-3.3 and ovirt-3.3.0
 branches.
 
 Douglas, would you issue a Fedora build from the ovirt-3.3.0 branch? I
 think this deserves an async build and an update of the ovirt-3.3
 reporsitory.
 
 
 Sure.
 
 F19: http://koji.fedoraproject.org/koji/taskinfo?taskID=6037496
 F20: http://koji.fedoraproject.org/koji/taskinfo?taskID=6037437
 EL6: http://koji.fedoraproject.org/koji/taskinfo?taskID=6037527
 
 Changes included into vdsm 4.12.1-3:
 - vm.Vm._getUnderlyingDriveInfo: extract path of gluster disks (BZ#1007980)
 - Require libvirt that allows vmUpdateDevice (BZ#1001001)
 - imageSharing: return proper size in httpGetSize
 - vdsmd.init: Add service-is-managed in shutdown_conflicting_srv
 (BZ#1006842)
 
 Adding in CC mburns for ovirt.org stuff.
 
 
 
 Builds posted to ovirt.org updates-testing repo.  They'll be pushed
 live with 3.3.0.1 release

Thanks!
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] vdsm sync meeting - October 7th 2013

2013-10-08 Thread Dan Kenigsberg
On Mon, Oct 07, 2013 at 03:25:22PM +0100, Dan Kenigsberg wrote:
 We had an unpleasant talk, hampered by statics and disconnection on
 danken's side. Beyond the noises I've managed to recognize Yaniv, Toni,
 Douglas, Danken, Ayal, Timothy, Yeela and Mooli. We've managed to discuss:

I've forgot to mention the resolution of

Bug 1007980 - Failed to run VM with gluster drives: internal error
unexpected address type for ide disk

by http://gerrit.ovirt.org/19949 . Due to its being a serious painpoint
to users. I've backport this fix to both ovirt-3.3 and ovirt-3.3.0
branches.

Douglas, would you issue a Fedora build from the ovirt-3.3.0 branch? I
think this deserves an async build and an update of the ovirt-3.3
reporsitory.

Dan.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


[vdsm] vdsm sync meeting - October 7th 2013

2013-10-07 Thread Dan Kenigsberg
We had an unpleasant talk, hampered by statics and disconnection on
danken's side. Beyond the noises I've managed to recognize Yaniv, Toni,
Douglas, Danken, Ayal, Timothy, Yeela and Mooli. We've managed to discuss:

- vdsm-4.13.0 is tagged, with a know selinux issue on el6. Expect a new
  seliux-policy solving it any time soon.

- All bugfixes should be backported to ovirt-3.3, so that we have a
  stable and comfortable vdsm in ovirt-3.3.1. Risky changes and new
  features should remain in master IMO.

- We incorporated a glusterfs requirement breaking rpm installaiton for
  people. We should avoid that by posters notifying reviewers more
  prominently and by having
  http://jenkins.ovirt.org/job/vdsm_install_rpm_sanity_gerrit/
  run on every patch that touches vdsm.spec.in.

  David, could you make the adjustment to the job?

- We discussed feature negotiation: Toni and Dan liked the idea of
  having vdsm expose a feature flags, to make it easier on Engine to
  check if a certain feature is supported.

  Ayal argues that this is useful only for capabilities that depend on
  existence on lower level components. Sees little value in fine
  feature granularity on vdsm side - versions is enough.

  So the disputed question is only how many feature flags we should
  have, and when to set them: statically or based on negotiation with
  kernel/libvirt/gluster/what not.

- Unified network persistence patches are being merged into master

- Timothy is working on fixing
  http://jenkins.ovirt.org/job/vdsm_verify_error_codes/lastBuild/console
  (hopefully by introducing the new error codes to Engine)

I was dropped from the call, so please append with stuff that I've
missed. Sorry for the noise!

Dan.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] vdsm sync meeting - September 23th 2013

2013-10-04 Thread Dan Kenigsberg
On Tue, Sep 24, 2013 at 10:23:21AM +0100, Dan Kenigsberg wrote:
 On Mon, Sep 23, 2013 at 04:34:29PM -0400, Douglas Schilling Landgraf wrote:
  - Heads up about 3.3.1/stable branch
http://lists.ovirt.org/pipermail/engine-devel/2013-September/005601.html
 
 Our intention is to rename the current ovirt-3.3 branch to ovirt-3.3.0,
 and create a new ovirt-3.3 branch for following ovirt-3.3.z releases,
 based on current master.
 
 Please help in testing the tip of master branch for bugs and other
 surprises!

After fighting several more surprises, v4.13.0 has been tagged.
Note that the ovirt-3.3 branch has been rebased as stated above.

We have a pending SELinux issue on el6: v4.13.0 requires a
yet-unpublished selinux-policy-targeted package that would allow it to
use /var/run/vdsm/storage for its storage repository.

Until it is published, you would need to something like

sudo semanage fcontext -a -s system_u -t virt_var_run_t -r s0 
'/var/run/vdsm(/.*)?'

and --force install your vdsm.rpm.


Regards,
Dan.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] supervdsmServer: Enable logging from multiprocessing library

2013-10-02 Thread Dan Kenigsberg
On Wed, Oct 02, 2013 at 03:59:58PM +1000, David Gibson wrote:
 On Tue, 01 Oct 2013 20:25:21 +0100
 Lee Yarwood lyarw...@redhat.com wrote:
 
  On 10/01/2013 06:35 PM, Dan Kenigsberg wrote:
   On Tue, Oct 01, 2013 at 02:33:00PM +0100, Lee Yarwood wrote:
   On 10/01/2013 09:00 AM, Dan Kenigsberg wrote:
   It is prefered to post patches to gerrit.ovirt.org.
  
   Apologies for jumping in David but I've pushed this here for now :
  
   http://gerrit.ovirt.org/19741
   
   Thanks!
 
 Thanks, I'm new to ovirt development and gerrit, so I'm going to need
 to work that out.
 
   On Tue, Oct 01, 2013 at 01:18:25PM +1000, David Gibson wrote:
   At present, if the super vdsm server dies with an exception inside
   Python's multiprocessing module, then it will not usually produce any
   useful debugging output.
  
   For our context - when do you notice such supervdsm deaths?
   Is it frequent? What is the cause?
  
   BZ#1011661  BZ#1010030 downstream.
   
   Ok, I can see them, dig into them and find an answer to my question. But
   it's not fair to the wider community of users and partner to cite
   private bugs.
   
   https://www.berrange.com/posts/2012/06/27/thoughts-on-improving-openstack-git-commit-practicehistory/
  
  Apologies Dan,
  
  I believe David was referring to the public BZ#1011661. I believe that
  has been attributed to the following change merged upstream in May :
  
  http://gerrit.ovirt.org/#/c/14998
 
 Uh, the problem's not attributed to that, rather that patch fixes it.
 The problem was that the process ctimes were changing, leading vdsm to
 erroneously think that supervdsm had died and restarting it.  That lead
 to several complications, including the supervdsm servers failing
 silently due to lack of logging from multiprocessing.

As much as I hate restarting supervdsmd service, I'm so glad that these
issues are solved in ovirt-3.3.

 
 We don't yet know why the ctimes were changing in this particular
 customer environment.

a longshot: maybe a system date change managed to confuse vdsm?

Dan.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] supervdsmServer: Enable logging from multiprocessing library

2013-10-01 Thread Dan Kenigsberg
It is prefered to post patches to gerrit.ovirt.org.

On Tue, Oct 01, 2013 at 01:18:25PM +1000, David Gibson wrote:
 At present, if the super vdsm server dies with an exception inside
 Python's multiprocessing module, then it will not usually produce any
 useful debugging output.

For our context - when do you notice such supervdsm deaths?
Is it frequent? What is the cause?

 
 The multiprocessing module includes error logging using Python's logging
 module, but it is not enabled by default.  This patch enables the
 logging and sets it to propagate to the supervdsmServer's root logger,
 so it now defaults to producing logging and can be configured easily
 from /etc/vdsm/svdsm.logger.conf.
 
 Signed-off-by: David Gibson da...@gibson.dropbear.id.au
 ---
  vdsm/supervdsmServer | 4 
  1 file changed, 4 insertions(+)
 
 diff --git a/vdsm/supervdsmServer b/vdsm/supervdsmServer
 index 40ec9df..43a408f 100755
 --- a/vdsm/supervdsmServer
 +++ b/vdsm/supervdsmServer
 @@ -42,6 +42,7 @@ except:
  log.warn(Could not init proper logging, exc_info=True)
  
  from storage import fuser
 +import multiprocessing
  from multiprocessing import Pipe, Process
  from gluster import listPublicFunctions
  import storage.misc as misc
 @@ -356,6 +357,9 @@ class _SuperVdsm(object):
  
  def main(sockfile, pidfile=None):
  log = logging.getLogger(SuperVdsm.Server)
 +# Wire up the multiprocessing logger
 +mlog = multiprocessing.get_logger()
 +mlog.propagate = 1

The code seems fine, though I'd avoid the variable

multiprocessing.get_logger().propagate = 1

Thanks.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] supervdsmServer: Enable logging from multiprocessing library

2013-10-01 Thread Dan Kenigsberg
On Tue, Oct 01, 2013 at 02:33:00PM +0100, Lee Yarwood wrote:
 On 10/01/2013 09:00 AM, Dan Kenigsberg wrote:
  It is prefered to post patches to gerrit.ovirt.org.
 
 Apologies for jumping in David but I've pushed this here for now :
 
 http://gerrit.ovirt.org/19741

Thanks!

 
  On Tue, Oct 01, 2013 at 01:18:25PM +1000, David Gibson wrote:
  At present, if the super vdsm server dies with an exception inside
  Python's multiprocessing module, then it will not usually produce any
  useful debugging output.
  
  For our context - when do you notice such supervdsm deaths?
  Is it frequent? What is the cause?
 
 BZ#1011661  BZ#1010030 downstream.

Ok, I can see them, dig into them and find an answer to my question. But
it's not fair to the wider community of users and partner to cite
private bugs.

https://www.berrange.com/posts/2012/06/27/thoughts-on-improving-openstack-git-commit-practicehistory/
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] [ATN] vdsm yum install is missing pkg for centos f18

2013-09-30 Thread Dan Kenigsberg
On Mon, Sep 30, 2013 at 04:11:56AM -0400, Eyal Edri wrote:
 fyi,
 
 as part of adding more verification jobs on vdsm, infra recently added a new 
 job [1] that installs vdsm.
 currently that job is working only on f19 and fails on f18 and centos64[2]:
 
 Are there any plans to fix this soon? how users are supposed to install vdsm 
 if those pkg are missing?
 
 
 Eyal.
 
 
 [1]http://jenkins.ovirt.org/job/vdsm_install_rpm_sanity_gerrit/label=fedora18/117/console
 
 [2]
 centos:
 
 Error: Package: vdsm-4.12.0-149.git44993c6.el6.x86_64 
 (/vdsm-4.12.0-149.git44993c6.el6.x86_64)
Requires: selinux-policy-targeted = 3.7.19-213
Installed: selinux-policy-targeted-3.7.19-195.el6_4.12.noarch 
 (@updates)

Thanks for reporting this issue. We must revert this part of
http://gerrit.ovirt.org/4220 (One shot prepare)
The selinux requirement came since that patch moves the storage
reprository to a new location (/var/run/vdsm/storage) instead of the
trademark-offensive non-standard /rhev/data-center.

EL6.4 would not be able to use the new location with selinux enabled.

We have a much bigger issue with this change, as it breaks migration to
and from ovirt-3.3; solving this is in the works.

 f18:
 
 -- Running transaction check
 --- Package dracut.x86_64 0:024-18.git20130102.fc18 will be updated
 --- Package dracut.x86_64 0:024-25.git20130205.fc18 will be an update
 --- Package vdsm.x86_64 0:4.12.0-149.git44993c6.fc18 will be installed
 -- Processing Dependency: libvirt-daemon = 1.0.2-1 for package: 
 vdsm-4.12.0-149.git44993c6.fc18.x86_64
 -- Finished Dependency Resolution
 Error: Package: vdsm-4.12.0-149.git44993c6.fc18.x86_64 
 (/vdsm-4.12.0-149.git44993c6.fc18.x86_64)
Requires: libvirt-daemon = 1.0.2-1
Installed: libvirt-daemon-0.10.2.7-1.fc18.x86_64 (@updates)
libvirt-daemon = 0.10.2.7-1.fc18
Available: libvirt-daemon-0.10.2.2-3.fc18.x86_64 (fedora)
libvirt-daemon = 0.10.2.2-3.fc18

This is a known and unavoidable issue (since
http://gerrit.ovirt.org/11709 of Mar 28 2013). oVirt users on F18 would
have to obtain libvirt from virt-preview or compile it themselves.

Dan.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] [ATN] vdsm yum install is missing pkg for centos f18

2013-09-30 Thread Dan Kenigsberg
On Mon, Sep 30, 2013 at 01:01:43PM -0400, Mike Burns wrote:
 On 09/30/2013 11:53 AM, Dan Kenigsberg wrote:
 On Mon, Sep 30, 2013 at 04:11:56AM -0400, Eyal Edri wrote:
 fyi,
 
 as part of adding more verification jobs on vdsm, infra recently added a 
 new job [1] that installs vdsm.
 currently that job is working only on f19 and fails on f18 and centos64[2]:
 
 Are there any plans to fix this soon? how users are supposed to install 
 vdsm if those pkg are missing?
 
 
 Eyal.
 
 
 [1]http://jenkins.ovirt.org/job/vdsm_install_rpm_sanity_gerrit/label=fedora18/117/console
 
 [2]
 centos:
 
 Error: Package: vdsm-4.12.0-149.git44993c6.el6.x86_64 
 (/vdsm-4.12.0-149.git44993c6.el6.x86_64)
 Requires: selinux-policy-targeted = 3.7.19-213
 Installed: selinux-policy-targeted-3.7.19-195.el6_4.12.noarch 
  (@updates)
 
 Thanks for reporting this issue. We must revert this part of
 http://gerrit.ovirt.org/4220 (One shot prepare)
 The selinux requirement came since that patch moves the storage
 reprository to a new location (/var/run/vdsm/storage) instead of the
 trademark-offensive non-standard /rhev/data-center.
 
 EL6.4 would not be able to use the new location with selinux enabled.
 
 We have a much bigger issue with this change, as it breaks migration to
 and from ovirt-3.3; solving this is in the works.
 
 f18:
 
 -- Running transaction check
 --- Package dracut.x86_64 0:024-18.git20130102.fc18 will be updated
 --- Package dracut.x86_64 0:024-25.git20130205.fc18 will be an update
 --- Package vdsm.x86_64 0:4.12.0-149.git44993c6.fc18 will be installed
 -- Processing Dependency: libvirt-daemon = 1.0.2-1 for package: 
 vdsm-4.12.0-149.git44993c6.fc18.x86_64
 -- Finished Dependency Resolution
 Error: Package: vdsm-4.12.0-149.git44993c6.fc18.x86_64 
 (/vdsm-4.12.0-149.git44993c6.fc18.x86_64)
 Requires: libvirt-daemon = 1.0.2-1
 Installed: libvirt-daemon-0.10.2.7-1.fc18.x86_64 (@updates)
 libvirt-daemon = 0.10.2.7-1.fc18
 Available: libvirt-daemon-0.10.2.2-3.fc18.x86_64 (fedora)
 libvirt-daemon = 0.10.2.2-3.fc18
 
 This is a known and unavoidable issue (since
 http://gerrit.ovirt.org/11709 of Mar 28 2013). oVirt users on F18 would
 have to obtain libvirt from virt-preview or compile it themselves.
 
 We should simply disable the job on Fedora 18.

Or install Fedora 20 alpha on the slave!
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] [Users] vdsm live migration errors in latest master

2013-09-27 Thread Dan Kenigsberg
On Thu, Sep 26, 2013 at 12:14:06PM -0400, Federico Simoncelli wrote:
 - Original Message -
  From: Dan Kenigsberg dan...@redhat.com
  To: Federico Simoncelli fsimo...@redhat.com
  Cc: Dead Horse deadhorseconsult...@gmail.com, users 
  us...@ovirt.org, vdsm-de...@fedorahosted.org,
  aba...@redhat.com
  Sent: Thursday, September 26, 2013 2:09:15 PM
  Subject: Re: [Users] vdsm live migration errors in latest master
  
  On Thu, Sep 26, 2013 at 05:35:46AM -0400, Federico Simoncelli wrote:
   - Original Message -
From: Dan Kenigsberg dan...@redhat.com
To: Federico Simoncelli fsimo...@redhat.com
Cc: Dead Horse deadhorseconsult...@gmail.com, users
us...@ovirt.org, vdsm-de...@fedorahosted.org,
aba...@redhat.com
Sent: Thursday, September 26, 2013 1:38:15 AM
Subject: Re: [Users] vdsm live migration errors in latest master

On Tue, Sep 24, 2013 at 12:04:14PM -0400, Federico Simoncelli wrote:
 - Original Message -
  From: Dan Kenigsberg dan...@redhat.com
  To: Dead Horse deadhorseconsult...@gmail.com
  Cc: us...@ovirt.org us...@ovirt.org,
  vdsm-de...@fedorahosted.org,
  fsimo...@redhat.com, aba...@redhat.com
  Sent: Tuesday, September 24, 2013 11:44:48 AM
  Subject: Re: [Users] vdsm live migration errors in latest master
  
  On Mon, Sep 23, 2013 at 04:05:34PM -0500, Dead Horse wrote:
   Seeing failed live migrations and these errors in the vdsm logs
   with
   latest
   VDSM/Engine master.
   Hosts are EL6.4
  
  Thanks for posting this report.
  
  The log is from the source of migration, right?
  Could you trace the history of the hosts of this VM? Could it be 
  that
  it
  was started on an older version of vdsm (say ovirt-3.3.0) and then
  (due
  to migration or vdsm upgrade) got into a host with a much newer 
  vdsm?
  
  Would you share the vmCreate (or vmMigrationCreate) line for this Vm
  in
  your log? I smells like an unintended regression of
  http://gerrit.ovirt.org/17714
  vm: extend shared property to support locking
  
  solving it may not be trivial, as we should not call
  _normalizeDriveSharedAttribute() automatically on migration
  destination,
  as it may well still be apart of a 3.3 clusterLevel.
  
  Also, migration from vdsm with extended shared property, to an ovirt
  3.3
  vdsm is going to explode (in a different way), since the destination
  does not expect the extended values.
  
  Federico, do we have a choice but to revert that patch, and use
  something like shared3 property instead?
 
 I filed a bug at:
 
 https://bugzilla.redhat.com/show_bug.cgi?id=1011608
 
 A possible fix could be:
 
 http://gerrit.ovirt.org/#/c/19509

Beyond this, we must make sure that on Engine side, the extended shared
values would be used only for clusterLevel 3.4 and above.

Are the extended shared values already used by Engine?
   
   Yes. That's the idea. Actually to be fair, the second case you mentioned
   (migrating from extended shared property to old vdsm) it wouldn't have 
   been
   possible I suppose (the issue here is that Dead Horse has one or more
   hosts running on master instead of 3.3). The extended shared property 
   would
   have appeared only in 3.4 and to allow the migration you would have had to
   upgrade all the nodes.
   
   But anyway since we were also talking about a new 3.3.1 barnch I just went
   ahead and covered all cases.
  
  I do not see how the 3.3.1 branch is relevant to the discussion, as its
  Vdsm is NOT going to support clusterLevel 3.4.
 
 That is what I was referring to.
 
 If 3.3.1 was 3.3.0 + backported patches then we just wouldn't backport the
 extended shared attributes patch and that's it. But from what I understood
 3.3.1 will be rebased on master (where instead we have the extended shared
 attributes) and that is why we have to cover both migration direction cases
 (instead of just the simple getattr one).
 
  Pardon my slowliness, but would you confirm that this feature is to be
  used only on clusterLevel 3.4 and above? If so, I'm +2ing your patch.
 
 Yes, the extended attributes will be used in the hosted engine and cluster
 level 3.4.
 But what the engine does is not relevant to +2ing correct vdsm patches.

IMHO the patch is correct only if all clients understned that this is a
clusterLevel 3.4 feature.

Now that that's clear to me. Ack by me.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] [Users] vdsm live migration errors in latest master

2013-09-26 Thread Dan Kenigsberg
On Thu, Sep 26, 2013 at 05:35:46AM -0400, Federico Simoncelli wrote:
 - Original Message -
  From: Dan Kenigsberg dan...@redhat.com
  To: Federico Simoncelli fsimo...@redhat.com
  Cc: Dead Horse deadhorseconsult...@gmail.com, users 
  us...@ovirt.org, vdsm-de...@fedorahosted.org,
  aba...@redhat.com
  Sent: Thursday, September 26, 2013 1:38:15 AM
  Subject: Re: [Users] vdsm live migration errors in latest master
  
  On Tue, Sep 24, 2013 at 12:04:14PM -0400, Federico Simoncelli wrote:
   - Original Message -
From: Dan Kenigsberg dan...@redhat.com
To: Dead Horse deadhorseconsult...@gmail.com
Cc: us...@ovirt.org us...@ovirt.org, vdsm-de...@fedorahosted.org,
fsimo...@redhat.com, aba...@redhat.com
Sent: Tuesday, September 24, 2013 11:44:48 AM
Subject: Re: [Users] vdsm live migration errors in latest master

On Mon, Sep 23, 2013 at 04:05:34PM -0500, Dead Horse wrote:
 Seeing failed live migrations and these errors in the vdsm logs with
 latest
 VDSM/Engine master.
 Hosts are EL6.4

Thanks for posting this report.

The log is from the source of migration, right?
Could you trace the history of the hosts of this VM? Could it be that it
was started on an older version of vdsm (say ovirt-3.3.0) and then (due
to migration or vdsm upgrade) got into a host with a much newer vdsm?

Would you share the vmCreate (or vmMigrationCreate) line for this Vm in
your log? I smells like an unintended regression of
http://gerrit.ovirt.org/17714
vm: extend shared property to support locking

solving it may not be trivial, as we should not call
_normalizeDriveSharedAttribute() automatically on migration destination,
as it may well still be apart of a 3.3 clusterLevel.

Also, migration from vdsm with extended shared property, to an ovirt 3.3
vdsm is going to explode (in a different way), since the destination
does not expect the extended values.

Federico, do we have a choice but to revert that patch, and use
something like shared3 property instead?
   
   I filed a bug at:
   
   https://bugzilla.redhat.com/show_bug.cgi?id=1011608
   
   A possible fix could be:
   
   http://gerrit.ovirt.org/#/c/19509
  
  Beyond this, we must make sure that on Engine side, the extended shared
  values would be used only for clusterLevel 3.4 and above.
  
  Are the extended shared values already used by Engine?
 
 Yes. That's the idea. Actually to be fair, the second case you mentioned
 (migrating from extended shared property to old vdsm) it wouldn't have been
 possible I suppose (the issue here is that Dead Horse has one or more
 hosts running on master instead of 3.3). The extended shared property would
 have appeared only in 3.4 and to allow the migration you would have had to
 upgrade all the nodes.
 
 But anyway since we were also talking about a new 3.3.1 barnch I just went
 ahead and covered all cases.

I do not see how the 3.3.1 branch is relevant to the discussion, as its
Vdsm is NOT going to support clusterLevel 3.4.

Pardon my slowliness, but would you confirm that this feature is to be
used only on clusterLevel 3.4 and above? If so, I'm +2ing your patch.

Dan.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] [Users] vdsm live migration errors in latest master

2013-09-25 Thread Dan Kenigsberg
On Tue, Sep 24, 2013 at 12:04:14PM -0400, Federico Simoncelli wrote:
 - Original Message -
  From: Dan Kenigsberg dan...@redhat.com
  To: Dead Horse deadhorseconsult...@gmail.com
  Cc: us...@ovirt.org us...@ovirt.org, vdsm-de...@fedorahosted.org, 
  fsimo...@redhat.com, aba...@redhat.com
  Sent: Tuesday, September 24, 2013 11:44:48 AM
  Subject: Re: [Users] vdsm live migration errors in latest master
  
  On Mon, Sep 23, 2013 at 04:05:34PM -0500, Dead Horse wrote:
   Seeing failed live migrations and these errors in the vdsm logs with 
   latest
   VDSM/Engine master.
   Hosts are EL6.4
  
  Thanks for posting this report.
  
  The log is from the source of migration, right?
  Could you trace the history of the hosts of this VM? Could it be that it
  was started on an older version of vdsm (say ovirt-3.3.0) and then (due
  to migration or vdsm upgrade) got into a host with a much newer vdsm?
  
  Would you share the vmCreate (or vmMigrationCreate) line for this Vm in
  your log? I smells like an unintended regression of
  http://gerrit.ovirt.org/17714
  vm: extend shared property to support locking
  
  solving it may not be trivial, as we should not call
  _normalizeDriveSharedAttribute() automatically on migration destination,
  as it may well still be apart of a 3.3 clusterLevel.
  
  Also, migration from vdsm with extended shared property, to an ovirt 3.3
  vdsm is going to explode (in a different way), since the destination
  does not expect the extended values.
  
  Federico, do we have a choice but to revert that patch, and use
  something like shared3 property instead?
 
 I filed a bug at:
 
 https://bugzilla.redhat.com/show_bug.cgi?id=1011608
 
 A possible fix could be:
 
 http://gerrit.ovirt.org/#/c/19509

Beyond this, we must make sure that on Engine side, the extended shared
values would be used only for clusterLevel 3.4 and above.

Are the extended shared values already used by Engine?
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


[vdsm] HELP NEEDED: vdsm sync meeting - September 23th 2013

2013-09-24 Thread Dan Kenigsberg
On Mon, Sep 23, 2013 at 04:34:29PM -0400, Douglas Schilling Landgraf wrote:
 - Heads up about 3.3.1/stable branch
   http://lists.ovirt.org/pipermail/engine-devel/2013-September/005601.html

Our intention is to rename the current ovirt-3.3 branch to ovirt-3.3.0,
and create a new ovirt-3.3 branch for following ovirt-3.3.z releases,
based on current master.

Please help in testing the tip of master branch for bugs and other
surprises!

 
 - New VDSM release for this week (please raise any urgent/blockers bugs)
   - Patch pending to be included (status: review):
glusterSD: BZ 988299: Deduce gluster volume name from domain metadata
http://gerrit.ovirt.org/#/c/19219/1
 
 
 
 -- 
 Cheers
 Douglas
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] [Users] vdsm live migration errors in latest master

2013-09-24 Thread Dan Kenigsberg
On Mon, Sep 23, 2013 at 04:05:34PM -0500, Dead Horse wrote:
 Seeing failed live migrations and these errors in the vdsm logs with latest
 VDSM/Engine master.
 Hosts are EL6.4

Thanks for posting this report.

The log is from the source of migration, right?
Could you trace the history of the hosts of this VM? Could it be that it
was started on an older version of vdsm (say ovirt-3.3.0) and then (due
to migration or vdsm upgrade) got into a host with a much newer vdsm?

Would you share the vmCreate (or vmMigrationCreate) line for this Vm in
your log? I smells like an unintended regression of
http://gerrit.ovirt.org/17714
vm: extend shared property to support locking

solving it may not be trivial, as we should not call
_normalizeDriveSharedAttribute() automatically on migration destination,
as it may well still be apart of a 3.3 clusterLevel.

Also, migration from vdsm with extended shared property, to an ovirt 3.3
vdsm is going to explode (in a different way), since the destination
does not expect the extended values.

Federico, do we have a choice but to revert that patch, and use
something like shared3 property instead?

 
 Thread-1306::ERROR::2013-09-23
 16:02:42,422::BindingXMLRPC::993::vds::(wrapper) unexpected error
 Traceback (most recent call last):
   File /usr/share/vdsm/BindingXMLRPC.py, line 979, in wrapper
 res = f(*args, **kwargs)
   File /usr/share/vdsm/BindingXMLRPC.py, line 211, in vmDestroy
 return vm.destroy()
   File /usr/share/vdsm/API.py, line 323, in destroy
 res = v.destroy()
   File /usr/share/vdsm/vm.py, line 4326, in destroy
 response = self.releaseVm()
   File /usr/share/vdsm/vm.py, line 4292, in releaseVm
 self._cleanup()
   File /usr/share/vdsm/vm.py, line 2750, in _cleanup
 self._cleanupDrives()
   File /usr/share/vdsm/vm.py, line 2482, in _cleanupDrives
 drive, exc_info=True)
   File /usr/lib64/python2.6/logging/__init__.py, line 1329, in error
 self.logger.error(msg, *args, **kwargs)
   File /usr/lib64/python2.6/logging/__init__.py, line 1082, in error
 self._log(ERROR, msg, args, **kwargs)
 File /usr/lib64/python2.6/logging/__init__.py, line 1082, in error
 self._log(ERROR, msg, args, **kwargs)
   File /usr/lib64/python2.6/logging/__init__.py, line 1173, in _log
 self.handle(record)
   File /usr/lib64/python2.6/logging/__init__.py, line 1183, in handle
 self.callHandlers(record)
   File /usr/lib64/python2.6/logging/__init__.py, line 1220, in
 callHandlers
 hdlr.handle(record)
   File /usr/lib64/python2.6/logging/__init__.py, line 679, in handle
 self.emit(record)
   File /usr/lib64/python2.6/logging/handlers.py, line 780, in emit
 msg = self.format(record)
   File /usr/lib64/python2.6/logging/__init__.py, line 654, in format
 return fmt.format(record)
   File /usr/lib64/python2.6/logging/__init__.py, line 436, in format
 record.message = record.getMessage()
   File /usr/lib64/python2.6/logging/__init__.py, line 306, in getMessage
 msg = msg % self.args
   File /usr/share/vdsm/vm.py, line 107, in __str__
 if not a.startswith('__')]
   File /usr/share/vdsm/vm.py, line 1344, in hasVolumeLeases
 if self.shared != DRIVE_SHARED_TYPE.EXCLUSIVE:
 AttributeError: 'Drive' object has no attribute 'shared'
 
 - DHC

 ___
 Users mailing list
 us...@ovirt.org
 http://lists.ovirt.org/mailman/listinfo/users

___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] VDSM RPM Build failed with error SecureXMLRPCServer.py:143:1: E303 too many blank lines

2013-09-20 Thread Dan Kenigsberg
On Fri, Sep 20, 2013 at 05:31:59AM -0400, Timothy Asir Jeyasingh wrote:
 Hi All,
 
 The vdsm (master branch) rpm build failed with the following error:
 
 find . -path './.git' -prune -type f -o \
   -name '*.py' -o -name '*.py.in'  | xargs /usr/bin/pyflakes | \
   grep -w -v \./vdsm/storage/lvm\.py.*: list comprehension redefines 
 'lv' from line .* | \
   while read LINE; do echo $LINE; false; done
 /usr/bin/pep8 --version
 1.4.6
 /usr/bin/pep8 --exclude=config.py,constants.py --filename '*.py,*.py.in' \
   client lib/cpopen/*.py lib/vdsm/*.py lib/vdsm/*.py.in tests 
 vds_bootstrap vdsm-tool vdsm/*.py vdsm/*.py.in vdsm/netconf 
 vdsm/sos/vdsm.py.in vdsm/storage vdsm/vdsm vdsm_api vdsm_hooks vdsm_reg 
 lib/vdsm/SecureXMLRPCServer.py:143:1: E303 too many blank lines (3)
 make[3]: *** [check-local] Error 1
 make[3]: Leaving directory `/home/timothy/rpmbuild/BUILD/vdsm-4.12.0'
 make[2]: *** [check-am] Error 2
 make[2]: Leaving directory `/home/timothy/rpmbuild/BUILD/vdsm-4.12.0'
 make[1]: *** [check-recursive] Error 1
 make[1]: Leaving directory `/home/timothy/rpmbuild/BUILD/vdsm-4.12.0'
 error: Bad exit status from /var/tmp/rpm-tmp.PcZNhb (%check)

Sorry about that. The pep8 failure was masked by another failure of
apiTests.JsonRawTest.testBadMethod
http://jenkins.ovirt.org/job/vdsm_unit_tests_gerrit/4499/testReport/junit/apiTests/JsonRawTest/testBadMethod/

Saggi, could you look into these recurring breakage?

As David Caro explains in a comment to
http://gerrit.ovirt.org/#/c/19332/ , I carelessly ignored the pep8
failure in Juan's patch.

Problem solved by Juan's http://gerrit.ovirt.org/19408 .

Thanks, and sorry for the noise.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] start vm which in pool failed

2013-09-18 Thread Dan Kenigsberg
On Wed, Sep 18, 2013 at 01:03:08PM +0800, bigclouds wrote:
 when starting a vm of a pool  fails.in the process of hooks, i modify 
 guestvm hostname, and modify the path of backing file of chain.  do nothing 
 else. 
 i can munually define a xml(after hooks), and start it without error.

Could you compare (or share) the manual domxml and the one produced by
your hook? It might be easier to spot the bug that way.

  
  
 env:
 libvirt-0.10.2-18.el6_4.5.x86_64
 2.6.32-358.6.2.el6.x86_64
 centos6.4
  
  
 1.where does this error message come from ?
 Storage.StorageDomain WARNING Could not find mapping for lv 
 d028d521-d4a9-4dd7-a0fe-3e9b60e7c4e4/cae1bb2d-0529-4287-95a3-13dcb14f082f
  
 2.
 Thread-6291::ERROR::2013-09-18 
 12:44:21,205::vm::683::vm.Vm::(_startUnderlyingVm) 
 vmId=`24f7e975-9aa5-4a14-b0f0-590add14c8b5`::The vm start process failed
 Traceback (most recent call last):
   File /usr/share/vdsm/vm.py, line 645, in _startUnderlyingVm
 self._run()
   File /usr/share/vdsm/libvirtvm.py, line 1529, in _run
 self._connection.createXML(domxml, flags),
   File /usr/lib64/python2.6/site-packages/vdsm/libvirtconnection.py, line 
 83, in wrapper
 ret = f(*args, **kwargs)
   File /usr/lib64/python2.6/site-packages/libvirt.py, line 2645, in 
 createXML
 if ret is None:raise libvirtError('virDomainCreateXML() failed', 
 conn=self)
 libvirtError: internal error Process exited while reading console log output: 
 char device redirected to /dev/pts/4
 qemu-kvm: -drive 
 file=/rhev/data-center/7828f2ae-955e-4e4b-a4bb-43807629dc52/d028d521-d4a9-4dd7-a0fe-3e9b60e7c4e4/images/ac025dc1-4e25-4b71-8c56-88dcb61b9f09/cae1bb2d-0529-4287-95a3-13dcb14f082f,if=none,id=drive-ide0-0-0,format=qcow2,serial=ac025dc1-4e25-4b71-8c56-88dcb61b9f09,cache=none,werror=stop,rerror=stop,aio=native:
  could not open disk image 
 /rhev/data-center/7828f2ae-955e-4e4b-a4bb-43807629dc52/d028d521-d4a9-4dd7-a0fe-3e9b60e7c4e4/images/ac025dc1-4e25-4b71-8c56-88dcb61b9f09/cae1bb2d-0529-4287-95a3-13dcb14f082f:
  Operation not permitted

 ___
 vdsm-devel mailing list
 vdsm-devel@lists.fedorahosted.org
 https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel

___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] rebasing for oVirt 3.3.1?

2013-09-18 Thread Dan Kenigsberg
On Mon, Sep 09, 2013 at 03:17:35PM +0300, Itamar Heim wrote:
 with 3.3.0 coming soon, one of the questions I heard is what about
 3.3.1 considering the number of patches fox bugs that went into
 master branch since since we branched to stabilize 3.3.0.
 i.e., most of the work in master branch has been focused on bug fixes)
 
 so my suggestion is for 3.3.1 that we rebase from master, then move
 to backporting patches to that branch for the rest of 3.3 time
 frame.
 
 while this poses a small risk, i believe its the best course forward
 to making ovirt 3.3 a more robust and stable version going forward.
 
 this is mostly about ovirt-engine, and probably vdsm. for the other
 projects, its up to the maintainer, based on risk/benefit.

To make this happen for Vdsm, we need to slow things down a bit,
stabilize what we have, and test it out.

Most of our work since ovirt-3.3 was bug fixing (23 patches), but some
of the 101 patches we've got are related to refactoring (19), cleanups
(27), test improvements (21), behind-the-scenes features (6), and
visible features (5). 

Refactoring included Zhou Zheng Sheng's Ubuntu-readiness patches, which
may still incur surprises to sysV/systemd/upstart service framework, and
changes to how network configurators are to be used.

Behind-the-scenes features include speedup to block-based storage:
- One shot teardown.
- Avoid Img and Vol produces in fileVolume.getV*Size
- Make lvm.listPVNames() be based on vgs information.
- One shot prepare.
- Introduce lvm short filters.

Visible features are few, and only one of them:
- clientIF: automatically unpause vms in EIO when SD becomes active
carries some kind of a risk to a timely release. The rest of them are:
- Support for multiple heads for Qxl display device
- Add support for direct setting of cpu_shares when creating a VM
- Introducing hidden_vlans configurable.
- macspoof hooks: new hook script to enable macspoof filtering per vnic.

I think we can release vdsm-4.13.0 within a week if we put a hold on new
features and big changes, and put enough effort into testing the
mostly-changed areas:
- service framework
- VM lifecycle over block storage (including auto unpause)
- network configuration

Then, we could release vdsm-4.13.z without risking the stability of
ovirt-3.3.1.

Let's do it!

Dan.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] VDSM sync meeting September 9th 2013

2013-09-10 Thread Dan Kenigsberg
On Tue, Sep 10, 2013 at 08:34:42AM +0200, Sandro Bonazzola wrote:
 Il 09/09/2013 16:09, Dan Kenigsberg ha scritto:
  - Master branch is broken for EL6 since we've taken
Id8e514df1ca88500f746242134ddb24c31588046 with its daemonAdapter
Yaniv Bronheim is working with Zhou and Alon Bar Lev to mend this.
Please review

  http://gerrit.ovirt.org/#/q/status:open+project:vdsm+branch:master+topic:error-handling-daemonAdapter,n,z
  
  - Another breakage of master is the libvirt requirement introduced by me
in http://gerrit.ovirt.org/18937. libvirt-0.10.2-18.el6_4.10 is not
yet public. Current workaround is to take a newer libvirt from
virt-preview, or ignore this dependency (which makes you unable to
re-wire networks for a running VM).
  
  - We now require gluster-3.4.0 which is not yet availabled from koji for
EL6. That's why we ship it in
http://resources.ovirt.org/releases/3.3/rpm/EL/6/x86_64/
However, it is missing from the nightly repos. Can we softlink to them
from the nightly repo, Eyal?
  
  - Timothy reports that the gluster team is working on geo-replication.
One of the patches under review is http://gerrit.ovirt.org/#/c/17766/
  
  Have I missed something? Please report it as a response!
 
 Master branch is also broken for Fedora 18, missing selinux-policy-targeted 
 = 3.12.1-71

Eduardo, this has been introduced by your one shot prepare
http://gerrit.ovirt.org/4220 .

Once Bug 1005950 - Current selinux policy prevents running a VM with
volumes under /var/run/vdsm/storage
is solved in Fedora 18, we should require the proper version of selinux
in our spec. Untill then, Vdsm cannot work with and enforcing F18's selinux.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


[vdsm] VDSM sync meeting September 9th 2013

2013-09-09 Thread Dan Kenigsberg
- Master branch is broken for EL6 since we've taken
  Id8e514df1ca88500f746242134ddb24c31588046 with its daemonAdapter
  Yaniv Bronheim is working with Zhou and Alon Bar Lev to mend this.
  Please review
  
http://gerrit.ovirt.org/#/q/status:open+project:vdsm+branch:master+topic:error-handling-daemonAdapter,n,z

- Another breakage of master is the libvirt requirement introduced by me
  in http://gerrit.ovirt.org/18937. libvirt-0.10.2-18.el6_4.10 is not
  yet public. Current workaround is to take a newer libvirt from
  virt-preview, or ignore this dependency (which makes you unable to
  re-wire networks for a running VM).

- We now require gluster-3.4.0 which is not yet availabled from koji for
  EL6. That's why we ship it in
  http://resources.ovirt.org/releases/3.3/rpm/EL/6/x86_64/
  However, it is missing from the nightly repos. Can we softlink to them
  from the nightly repo, Eyal?

- Timothy reports that the gluster team is working on geo-replication.
  One of the patches under review is http://gerrit.ovirt.org/#/c/17766/

Have I missed something? Please report it as a response!

Dan.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] glusterfs-client and el6

2013-08-27 Thread Dan Kenigsberg
They are here:
http://resources.ovirt.org/releases/3.3/rpm/EL/6/x86_64/glusterfs-cli-3.4.0-8.el6.x86_64.rpm

posting top we are why?

On Tue, Aug 27, 2013 at 02:41:54PM -0400, Yaniv Bronheim wrote:
 Any estimation when the packages will be added to 6.4?
 If it'll take long we should revert the dependencies on those packages till 
 then
 
 - Original Message -
  From: Antoni Segura Puimedon asegu...@redhat.com
  To: Yaniv Bronheim ybron...@redhat.com
  Cc: VDSM Project Development vdsm-devel@lists.fedorahosted.org
  Sent: Tuesday, August 27, 2013 10:35:13 AM
  Subject: Re: [vdsm] glusterfs-client and el6
  
  Gluster 3.4 pkgs are not available in epel for el6 (6.4, they will be, I
  think for 6.5). However, as I put yesterday in the notes, Federico said
  that those packages will be added to the ovirt.org repository for 6.4.
  At the moment I'm using the gluster repo for them.
  
  - Original Message -
   From: Yaniv Bronheim ybron...@redhat.com
   To: Antoni Segura Puimedon asegu...@redhat.com
   Cc: VDSM Project Development vdsm-devel@lists.fedorahosted.org
   Sent: Tuesday, August 27, 2013 8:43:01 AM
   Subject: Re: [vdsm] glusterfs-client and el6
   
   Before merging such patches that adds new requirements please verify that
   the
   required packages are available. Such issues raised more than few times
   before..
   As I understood in yesterday's vdsm call the gluster packages are already
   available, so this issue is not relevant anymore. right?
   
   Yaniv Bronhaim.
   
   
   - Original Message -
From: Antoni Segura Puimedon asegu...@redhat.com
To: VDSM Project Development vdsm-devel@lists.fedorahosted.org
Sent: Friday, August 23, 2013 2:15:33 PM
Subject: [vdsm] glusterfs-client and el6

Hi list,

This patch http://gerrit.ovirt.org/#/c/17994/ made vdsm on el6 and
Fedora18+
depend, among others, on glusterfs-cli, which is on the epel candidates
but
not yet on epel stable (at least this morning it was not
in
http://dl.fedoraproject.org/pub/epel/6/x86_64/repoview/letter_g.group.html
). To be able to install vdsm on el6, then, it is necessary to get the
packages in:

http://koji.fedoraproject.org/koji/buildinfo?buildID=454492

which are labeled at the moment as epel-testing-candidate.

In the future we should strive delay such changes until the packages hit
epel
stable, I'd propose.

Best regards,

Toni
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel

   
  
 ___
 vdsm-devel mailing list
 vdsm-devel@lists.fedorahosted.org
 https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] suggesting new Vdsm maintaniner(s)

2013-08-08 Thread Dan Kenigsberg
On Thu, Aug 01, 2013 at 04:55:07PM +0300, Dan Kenigsberg wrote:
 Nominating Toni Puimedon and/or Mark Wu as maintainers
 ==
 
 During their recent year-or-two of participation in Vdsm development,
 both Toni and Mark has demonstrated a genuine care for the health and
 evolution of the the project.
 
 They both carry this hard-to-quantify quality of serious-mindedness when
 reviewing patches.
 
 Toni's 73 committed patches touch everything that is related to network
 configuration and reporting
 
  27 vdsm/configNetwork.py
  11 lib/vdsm/netinfo.py
   8 vdsm/API.py
   8 tests/configNetworkTests.py
   7 vdsm/netinfo.py
   7 vdsm/netconf/ifcfg.py
   5 vdsm/libvirtvm.py
   4 vdsm/netmodels.py
 
 ... similar to Mark's 80:
 
  16 vdsm/configNetwork.py
  12 vdsm/netconf/ifcfg.py
  12 vdsm/libvirtvm.py
   7 vdsm/clientIF.py
   6 lib/vdsm/netinfo.py
   5 vdsm/caps.py
   5 vdsm/API.py
   4 vdsm/vm.py
   4 vdsm/netmodels.py
   4 vdsm/Makefile.am
 
 I suggest that both of them obtain +2/-2 rights, in the understanding
 that they are to be used on net-related parts of the code. This is
 particularly important given my nearing, relatively long, vacation.
 
 What do you say?

We've got no objectation, but neither an overwhelming support. Maybe
it's due to the summer season.

To me, it is important to tell whether Toni/Mark only reviewed a patch
(+1) or are completely standing behind it, and going to fix bugs in it
if the author is run over by a bus (God forbid).

Itamar and other board members: if there is no objection within, say,
one week, would you grant the duo +2 rights?

Dan.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] Error while rebooting VM

2013-08-03 Thread Dan Kenigsberg
On Fri, Aug 02, 2013 at 03:15:46PM +0200, Sandro Bonazzola wrote:
 Pushed a patch for the issue here: http://gerrit.ovirt.org/#/c/17599/
 
 Now I see the following traceback in the logs:
 
 
 Traceback (most recent call last):
   File /usr/share/vdsm/clientIF.py, line 382, in teardownVolumePath
 res = self.irs.teardownImage(drive['domainID'],
   File /usr/share/vdsm/vm.py, line 1343, in __getitem__
 raise KeyError(key)
 KeyError: 'domainID'

I do not understand what it is, but I've seen that before. It's
unrelated to Yeela's latest patch, and unlike it, it affects ovirt-3.3
branch)
https://bugzilla.redhat.com/show_bug.cgi?id=987965#c2
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] [Users] vdsm v4.12.0 dependencies for el6 hosts

2013-08-01 Thread Dan Kenigsberg
On Wed, Jul 31, 2013 at 04:36:00PM -0500, Dead Horse wrote:
 vdsm v4.12.0 for engine 3.3 has some unsatisfied dependencies on el6 hosts:
 
 -- Finished Dependency Resolution
 Error: Package: vdsm-hook-sriov-4.12.0-10.el6.noarch (el6v-test)
Requires: libvirt-daemon-driver-nodedev
 Error: Package: vdsm-4.12.0-10.el6.x86_64 (el6v-test)
Requires: mom = 0.3.2-3
Removing: mom-0.3.0-1.el6.noarch (@epel)
mom = 0.3.0-1.el6
Updated By: mom-0.3.2-2.el6.noarch (el6v-test)
mom = 0.3.2-2.el6
 
 mom can be built from http://gerrit.ovirt.org/gitweb?p=mom.git
 
 but libvirt-daemon-driver-nodedev would require a newer libvirt.
  (https://apps.fedoraproject.org/packages/libvirt-daemon-driver-nodedev)

That's my bad. Would you verify
http://gerrit.ovirt.org/#/c/17554/ ?

Regrads,
Dan.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


[vdsm] suggesting new Vdsm maintaniner(s)

2013-08-01 Thread Dan Kenigsberg
Nominating Toni Puimedon and/or Mark Wu as maintainers
==

During their recent year-or-two of participation in Vdsm development,
both Toni and Mark has demonstrated a genuine care for the health and
evolution of the the project.

They both carry this hard-to-quantify quality of serious-mindedness when
reviewing patches.

Toni's 73 committed patches touch everything that is related to network
configuration and reporting

 27 vdsm/configNetwork.py
 11 lib/vdsm/netinfo.py
  8 vdsm/API.py
  8 tests/configNetworkTests.py
  7 vdsm/netinfo.py
  7 vdsm/netconf/ifcfg.py
  5 vdsm/libvirtvm.py
  4 vdsm/netmodels.py

... similar to Mark's 80:

 16 vdsm/configNetwork.py
 12 vdsm/netconf/ifcfg.py
 12 vdsm/libvirtvm.py
  7 vdsm/clientIF.py
  6 lib/vdsm/netinfo.py
  5 vdsm/caps.py
  5 vdsm/API.py
  4 vdsm/vm.py
  4 vdsm/netmodels.py
  4 vdsm/Makefile.am

I suggest that both of them obtain +2/-2 rights, in the understanding
that they are to be used on net-related parts of the code. This is
particularly important given my nearing, relatively long, vacation.

What do you say?

Dan.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] [Users] [ATN] oVirt 3.3 release, moving to RC [3.3 branching, blockers review]

2013-07-30 Thread Dan Kenigsberg
On Tue, Jul 30, 2013 at 06:05:46PM +0300, Moran Goldboim wrote:
 We would like to go a head on oVirt 3.3 release process and issue an RC
 
 for tomorrow meeting:
 
 Maintainers/ Package owners
 ==
 -branch your git repo (master) with 3.3 branch (making master ready for 3.4)
 -for tomorrow meeting, overview [1], see if you have any changes to it
 -if you don't have any blockers under your component, please branch
 3.3 with RC1 branch
 
 |
 -master
 |
 -engine_3.3
 |
 -RC1
 |
 -engine_3.2
 ...
 
 Users
 
 with your experience from test day and with nightlies, if you feel
 there are additional candidates to block this version for please add
 it to the tracker bug [1].
 
 Suggested Schedule
 
 Wed Jul 31st - review of blockers for the version and component readiness
 Mon Aug 5th - RC1 release
 Wed Aug 7th - Release readiness review (in case of blockers an RC2
 will be issued)
 
 Thanks.
 
 [1]*Bug 918494* https://bugzilla.redhat.com/show_bug.cgi?id=918494
 -Tracker: oVirt 3.3 release

I've just tagged vdsm-4.12.0 and branched ovirt-3.3 branch for the vdsm
repository starting at git has 620343d6317c849fc985a5083cebb68c995f7c15.

Expect a deluge of non-ovirt-3.3 merges to the master branch soon.
Future ovirt-3.3 fixes would have to be backported and cherry-picked.

Dan.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


Re: [vdsm] migration progress feature

2013-07-22 Thread Dan Kenigsberg
On Sun, Jul 21, 2013 at 10:29:03PM +0300, Itamar Heim wrote:
 On 07/21/2013 12:14 PM, Dan Kenigsberg wrote:
 Introducing a new verb, getMigrationStatuses(), would allow Engine to
 collect this piece of statistics in the exact frequency that it needs
 it. Coupling this data with the frequently-polled state of VM begs for
 future complaints that getVMList(full=False) is too heavy, and that we
 must introduce getVMList(full=False, veryveryskinny=True).
 
 so the most likely outcome is engine would be calling this every
 time it calls getAllVmStats?

Having two different verbs allows the client to play with the frequency
of polling. Last word I've heard was that getAllVmStats is too heavy to
be called in the frequency that getMigrationStatuses is needed.

If Engine is happy to lock the frequency of getMigrationStatuses to that
of getAllVmStats, we can include migrateStatus within getAllVmStats. I
already conceded to this earlier in this thread.

Dan.
___
vdsm-devel mailing list
vdsm-devel@lists.fedorahosted.org
https://lists.fedorahosted.org/mailman/listinfo/vdsm-devel


  1   2   3   >