Re: [openstack-dev] [neutron] [third-party] Neutron 3rd Party CI status dashboard

2014-06-30 Thread Anita Kuno
On 06/29/2014 07:59 PM, Anita Kuno wrote:
 On 06/29/2014 07:43 PM, Anita Kuno wrote:
 On 06/29/2014 03:25 PM, Ilya Shakhat wrote:
 Hi!

 During last couple weeks there is an increasing demand on tracking
 3rd-party CI statuses. We at Stackalytics decided to be in trend and (with
 some inspiration from Salvatore's proposal) implemented report that shows
 summary on external CI status. The initial version is available for Neutron
 - http://stackalytics.com/report/ci/neutron/7

 The report shows summary of all CI jobs during specified period of time,
 including:
  * stats of runs on merged patch sets:
 - total number of runs
 - success rate (success to total ratio)
 - time of the latest run
 - last test result
   * stats for all patch sets (the same set as for merged)
   * last test results for every merged patch set grouped by days (useful to
 see how different CIs correlate with each other and how often they run)

 Under merged patch set report means the last patch set in the merged
 change request, thus it is almost the same as the trunk code. CI
 configuration is taken from DriverLog's default data
 https://git.openstack.org/cgit/stackforge/driverlog/tree/etc/default_data.json.
 Standard Stackalytics screen is also available for CIs -
 http://stackalytics.com/?metric=ci, including votes breakdown and activity
 log.

 Since this is the first version there are some open questions:
  * Currently report shows results per CI id, but there are CIs that run
 tests against multiple drivers and this case is not supported. What would
 be more useful: to get stats for a driver or for CI?
  * Most CIs run tests when patch set is posted. So even if change request
 is merged within selected time period corresponding CI results may be
 missing.
  * Patterns for non-voting CIs need to be verified. For example Cisco CI
 now runs 5 jobs, but DriverLog data still contains old data.

 Thanks,
 Ilya

 2014-06-16 17:52 GMT+04:00 Salvatore Orlando sorla...@nicira.com:


 However, it would be great if we could start devising a solution for
 having health reports from the various CI systems.
 This report should report the following kind of information:
 - timestamp of last run
 - timestamp of last vote (a system might start job which then get aborted
 for CI infra problems)
 - % of success vs failures (not sure how important is that one but
 provides a metric of comparison with upstream jenkins)
 - % of disagreements with jenkins (this might allow us to quickly spot
 those CI systems which are randomly -1'ing patches)

 The percentage metrics might be taken over a 48 hours or 7 days interval,
 or both.
 Does this idea sound reasonable?





 ___
 OpenStack-dev mailing list
 OpenStack-dev@lists.openstack.org
 http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev

 Hi Ilya:

 I look forward to hearing more about this dashboard and ensuring you or
 someone else associated with this dashboard are available for questions
 at the third party meeting tomorrow:
 https://wiki.openstack.org/wiki/Meetings/ThirdParty

 We missed you last week.

 Thanks Ilya,
 Anita.

 And one question I will have when we discuss this is regarding this
 statement: Green cell - tests ran successfully,
 
 Currently we don't have community consensus around the use of the
 statement tests ran successfully regarding third party ci systems.
 This is as statement, you recall, we had discussed at the third party
 meeting when we talked about driverlog.
 
 18:46:02 krtaylor but what does CI tested really mean? just running
 tests? or tested to pass some level of requirements?
 http://eavesdrop.openstack.org/meetings/third_party/2014/third_party.2014-06-16-18.00.log.html
 
 Using a statement to convey success prior to having a definition from
 the community about what the statement means will create confusion in
 the community and frustration from folks, including myself, who are
 willing to have the conversation about what it means, who feel
 circumvented by this second use of a phrase which implies a decided
 meaning where none yet exists. Please participate in conversations
 around the definition of phrases of success and failure regarding third
 party systems and point to the logs where consensus is reached prior it
 its use in future.
 
 In addition to attending the third party meeting, please attend the
 infra meeting or the qa meeting, and hopefully meetings that include
 programs that have interactions with third party ci systems including
 nova, neutron, and cinder (if there are other programs interacting with
 third party ci systems please attend the third party meeting so I know
 about you).
 
 Thanks Ilya, I look forward to our future discussions,
 Anita.
 
As an update, Ilya did attend the third party meeting today, thank you Ilya.

I am disappointed to realize that Ilya (or stackalytics, I don't know
where this is coming from) is unwilling to cease making up definitions
of success 

Re: [openstack-dev] [neutron] [third-party] Neutron 3rd Party CI status dashboard

2014-06-30 Thread Luke Gorrie
On 30 June 2014 21:08, Anita Kuno ante...@anteaya.info wrote:

 I am disappointed to realize that Ilya (or stackalytics, I don't know
 where this is coming from) is unwilling to cease making up definitions
 of success for third party ci systems to allow the openstack community
 to arrive at its own definition.


There is indeed a risk that the new dashboards won't give a meaningful view
of whether a 3rd party CI is voting correctly or not.

However, there is an elephant in the room and a much more important problem:

To measure how accurately a CI is voting says much more about a driver
author's Gerrit-fu ability to operate a CI system than it does about
whether the code they have contributed to OpenStack actually works, and the
latter is what is actually important.

To my mind the whole 3rd party testing discussion should refocus on helping
developers maintain good working code and less on waving you will be
kicked out of OpenStack if you don't keep your swarm of nebulous daemons
running 24/7.
___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [neutron] [third-party] Neutron 3rd Party CI status dashboard

2014-06-30 Thread Anita Kuno
On 06/30/2014 03:27 PM, Luke Gorrie wrote:
 On 30 June 2014 21:08, Anita Kuno ante...@anteaya.info wrote:
 
 I am disappointed to realize that Ilya (or stackalytics, I don't know
 where this is coming from) is unwilling to cease making up definitions
 of success for third party ci systems to allow the openstack community
 to arrive at its own definition.

 
 There is indeed a risk that the new dashboards won't give a meaningful view
 of whether a 3rd party CI is voting correctly or not.
 
 However, there is an elephant in the room and a much more important problem:
 
 To measure how accurately a CI is voting says much more about a driver
 author's Gerrit-fu ability to operate a CI system than it does about
 whether the code they have contributed to OpenStack actually works, and the
 latter is what is actually important.
 
 To my mind the whole 3rd party testing discussion should refocus on helping
 developers maintain good working code and less on waving you will be
 kicked out of OpenStack if you don't keep your swarm of nebulous daemons
 running 24/7.
 
 
 
 ___
 OpenStack-dev mailing list
 OpenStack-dev@lists.openstack.org
 http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
 
Well I am glad you have reached that conclusion, Luke.

Given that I gave you months of my time with that exact approach and you
squandered it away, claiming I wasn't doing enough to help you, I am
glad you have finally reached a productive place.

I hope your direction produces results.

Thanks Luke,
Anita.

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [neutron] [third-party] Neutron 3rd Party CI status dashboard

2014-06-30 Thread Sukhdev Kapur
Well, Luke, this is collaborative effort by everybody. By having these CI
systems in place ensures that one person's code does not break other
person's code and vice versa. Therefore, having these CI systems
operational and voting 24x7 is a critical step in achieving this goal.

However, the details as to how and what should be tested is definitely
debatable and the team has done excellent job in converging on that.

Now, as to the issue at hand which Anita is describing, I attended the
meeting this morning and was very pleased with the debate that took place
and the definition as to Sucess


On Mon, Jun 30, 2014 at 12:27 PM, Luke Gorrie l...@tail-f.com wrote:

 On 30 June 2014 21:08, Anita Kuno ante...@anteaya.info wrote:

 I am disappointed to realize that Ilya (or stackalytics, I don't know
 where this is coming from) is unwilling to cease making up definitions
 of success for third party ci systems to allow the openstack community
 to arrive at its own definition.


 There is indeed a risk that the new dashboards won't give a meaningful
 view of whether a 3rd party CI is voting correctly or not.

 However, there is an elephant in the room and a much more important
 problem:

 To measure how accurately a CI is voting says much more about a driver
 author's Gerrit-fu ability to operate a CI system than it does about
 whether the code they have contributed to OpenStack actually works, and the
 latter is what is actually important.

 To my mind the whole 3rd party testing discussion should refocus on
 helping developers maintain good working code and less on waving you will
 be kicked out of OpenStack if you don't keep your swarm of nebulous daemons
 running 24/7.



 ___
 OpenStack-dev mailing list
 OpenStack-dev@lists.openstack.org
 http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [neutron] [third-party] Neutron 3rd Party CI status dashboard

2014-06-30 Thread Sukhdev Kapur
Sorry, accidentally hit the wrong key and message went out.

Was making a mention about the definition of Success. I thought the debate
in the meeting was very productive - when a CI posts a +1 that is success,
and when a CI posts a -1 (or no vote with comment) is also a success - as
this reflects that the CI is doing what it is suppose to do.

So, when it comes to stackalytics, it is more critical to show if a given
CI is operational or not - and for how long?
Another thing we can debate is how to present the +1/-1 votes by a given CI
- unless we have some benchmark, it will be hard to consider
success/failure.

So, I am of the  opinion that, initially, we only report on the operational
status and duration of the CIs, and a counter of +1 and -1 votes over a
period of time. For example, looking at Arista CI, it has casted 7,958
votes so far and it has been operational for past 6 months. This
information is not available anywhere - hence, presenting this kind of
information on a dashboard created by Ilya would be very useful to the
community as well to the vendors..

thoughts?

-Sukhdev





On Mon, Jun 30, 2014 at 12:49 PM, Sukhdev Kapur sukhdevka...@gmail.com
wrote:

 Well, Luke, this is collaborative effort by everybody. By having these CI
 systems in place ensures that one person's code does not break other
 person's code and vice versa. Therefore, having these CI systems
 operational and voting 24x7 is a critical step in achieving this goal.

 However, the details as to how and what should be tested is definitely
 debatable and the team has done excellent job in converging on that.

 Now, as to the issue at hand which Anita is describing, I attended the
 meeting this morning and was very pleased with the debate that took place
 and the definition as to Sucess


 On Mon, Jun 30, 2014 at 12:27 PM, Luke Gorrie l...@tail-f.com wrote:

 On 30 June 2014 21:08, Anita Kuno ante...@anteaya.info wrote:

 I am disappointed to realize that Ilya (or stackalytics, I don't know
 where this is coming from) is unwilling to cease making up definitions
 of success for third party ci systems to allow the openstack community
 to arrive at its own definition.


 There is indeed a risk that the new dashboards won't give a meaningful
 view of whether a 3rd party CI is voting correctly or not.

  However, there is an elephant in the room and a much more important
 problem:

 To measure how accurately a CI is voting says much more about a driver
 author's Gerrit-fu ability to operate a CI system than it does about
 whether the code they have contributed to OpenStack actually works, and the
 latter is what is actually important.

 To my mind the whole 3rd party testing discussion should refocus on
 helping developers maintain good working code and less on waving you will
 be kicked out of OpenStack if you don't keep your swarm of nebulous daemons
 running 24/7.



 ___
 OpenStack-dev mailing list
 OpenStack-dev@lists.openstack.org
 http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev



___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


[openstack-dev] [neutron] [third-party] Neutron 3rd Party CI status dashboard

2014-06-29 Thread Ilya Shakhat
Hi!

During last couple weeks there is an increasing demand on tracking
3rd-party CI statuses. We at Stackalytics decided to be in trend and (with
some inspiration from Salvatore's proposal) implemented report that shows
summary on external CI status. The initial version is available for Neutron
- http://stackalytics.com/report/ci/neutron/7

The report shows summary of all CI jobs during specified period of time,
including:
 * stats of runs on merged patch sets:
- total number of runs
- success rate (success to total ratio)
- time of the latest run
- last test result
  * stats for all patch sets (the same set as for merged)
  * last test results for every merged patch set grouped by days (useful to
see how different CIs correlate with each other and how often they run)

Under merged patch set report means the last patch set in the merged
change request, thus it is almost the same as the trunk code. CI
configuration is taken from DriverLog's default data
https://git.openstack.org/cgit/stackforge/driverlog/tree/etc/default_data.json.
Standard Stackalytics screen is also available for CIs -
http://stackalytics.com/?metric=ci, including votes breakdown and activity
log.

Since this is the first version there are some open questions:
 * Currently report shows results per CI id, but there are CIs that run
tests against multiple drivers and this case is not supported. What would
be more useful: to get stats for a driver or for CI?
 * Most CIs run tests when patch set is posted. So even if change request
is merged within selected time period corresponding CI results may be
missing.
 * Patterns for non-voting CIs need to be verified. For example Cisco CI
now runs 5 jobs, but DriverLog data still contains old data.

Thanks,
Ilya

2014-06-16 17:52 GMT+04:00 Salvatore Orlando sorla...@nicira.com:


 However, it would be great if we could start devising a solution for
 having health reports from the various CI systems.
 This report should report the following kind of information:
 - timestamp of last run
 - timestamp of last vote (a system might start job which then get aborted
 for CI infra problems)
 - % of success vs failures (not sure how important is that one but
 provides a metric of comparison with upstream jenkins)
 - % of disagreements with jenkins (this might allow us to quickly spot
 those CI systems which are randomly -1'ing patches)

 The percentage metrics might be taken over a 48 hours or 7 days interval,
 or both.
 Does this idea sound reasonable?


___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [neutron] [third-party] Neutron 3rd Party CI status dashboard

2014-06-29 Thread Anita Kuno
On 06/29/2014 03:25 PM, Ilya Shakhat wrote:
 Hi!
 
 During last couple weeks there is an increasing demand on tracking
 3rd-party CI statuses. We at Stackalytics decided to be in trend and (with
 some inspiration from Salvatore's proposal) implemented report that shows
 summary on external CI status. The initial version is available for Neutron
 - http://stackalytics.com/report/ci/neutron/7
 
 The report shows summary of all CI jobs during specified period of time,
 including:
  * stats of runs on merged patch sets:
 - total number of runs
 - success rate (success to total ratio)
 - time of the latest run
 - last test result
   * stats for all patch sets (the same set as for merged)
   * last test results for every merged patch set grouped by days (useful to
 see how different CIs correlate with each other and how often they run)
 
 Under merged patch set report means the last patch set in the merged
 change request, thus it is almost the same as the trunk code. CI
 configuration is taken from DriverLog's default data
 https://git.openstack.org/cgit/stackforge/driverlog/tree/etc/default_data.json.
 Standard Stackalytics screen is also available for CIs -
 http://stackalytics.com/?metric=ci, including votes breakdown and activity
 log.
 
 Since this is the first version there are some open questions:
  * Currently report shows results per CI id, but there are CIs that run
 tests against multiple drivers and this case is not supported. What would
 be more useful: to get stats for a driver or for CI?
  * Most CIs run tests when patch set is posted. So even if change request
 is merged within selected time period corresponding CI results may be
 missing.
  * Patterns for non-voting CIs need to be verified. For example Cisco CI
 now runs 5 jobs, but DriverLog data still contains old data.
 
 Thanks,
 Ilya
 
 2014-06-16 17:52 GMT+04:00 Salvatore Orlando sorla...@nicira.com:
 

 However, it would be great if we could start devising a solution for
 having health reports from the various CI systems.
 This report should report the following kind of information:
 - timestamp of last run
 - timestamp of last vote (a system might start job which then get aborted
 for CI infra problems)
 - % of success vs failures (not sure how important is that one but
 provides a metric of comparison with upstream jenkins)
 - % of disagreements with jenkins (this might allow us to quickly spot
 those CI systems which are randomly -1'ing patches)

 The percentage metrics might be taken over a 48 hours or 7 days interval,
 or both.
 Does this idea sound reasonable?


 
 
 
 ___
 OpenStack-dev mailing list
 OpenStack-dev@lists.openstack.org
 http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
 
Hi Ilya:

I look forward to hearing more about this dashboard and ensuring you or
someone else associated with this dashboard are available for questions
at the third party meeting tomorrow:
https://wiki.openstack.org/wiki/Meetings/ThirdParty

We missed you last week.

Thanks Ilya,
Anita.

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [neutron] [third-party] Neutron 3rd Party CI status dashboard

2014-06-29 Thread Anita Kuno
On 06/29/2014 07:43 PM, Anita Kuno wrote:
 On 06/29/2014 03:25 PM, Ilya Shakhat wrote:
 Hi!

 During last couple weeks there is an increasing demand on tracking
 3rd-party CI statuses. We at Stackalytics decided to be in trend and (with
 some inspiration from Salvatore's proposal) implemented report that shows
 summary on external CI status. The initial version is available for Neutron
 - http://stackalytics.com/report/ci/neutron/7

 The report shows summary of all CI jobs during specified period of time,
 including:
  * stats of runs on merged patch sets:
 - total number of runs
 - success rate (success to total ratio)
 - time of the latest run
 - last test result
   * stats for all patch sets (the same set as for merged)
   * last test results for every merged patch set grouped by days (useful to
 see how different CIs correlate with each other and how often they run)

 Under merged patch set report means the last patch set in the merged
 change request, thus it is almost the same as the trunk code. CI
 configuration is taken from DriverLog's default data
 https://git.openstack.org/cgit/stackforge/driverlog/tree/etc/default_data.json.
 Standard Stackalytics screen is also available for CIs -
 http://stackalytics.com/?metric=ci, including votes breakdown and activity
 log.

 Since this is the first version there are some open questions:
  * Currently report shows results per CI id, but there are CIs that run
 tests against multiple drivers and this case is not supported. What would
 be more useful: to get stats for a driver or for CI?
  * Most CIs run tests when patch set is posted. So even if change request
 is merged within selected time period corresponding CI results may be
 missing.
  * Patterns for non-voting CIs need to be verified. For example Cisco CI
 now runs 5 jobs, but DriverLog data still contains old data.

 Thanks,
 Ilya

 2014-06-16 17:52 GMT+04:00 Salvatore Orlando sorla...@nicira.com:


 However, it would be great if we could start devising a solution for
 having health reports from the various CI systems.
 This report should report the following kind of information:
 - timestamp of last run
 - timestamp of last vote (a system might start job which then get aborted
 for CI infra problems)
 - % of success vs failures (not sure how important is that one but
 provides a metric of comparison with upstream jenkins)
 - % of disagreements with jenkins (this might allow us to quickly spot
 those CI systems which are randomly -1'ing patches)

 The percentage metrics might be taken over a 48 hours or 7 days interval,
 or both.
 Does this idea sound reasonable?





 ___
 OpenStack-dev mailing list
 OpenStack-dev@lists.openstack.org
 http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev

 Hi Ilya:
 
 I look forward to hearing more about this dashboard and ensuring you or
 someone else associated with this dashboard are available for questions
 at the third party meeting tomorrow:
 https://wiki.openstack.org/wiki/Meetings/ThirdParty
 
 We missed you last week.
 
 Thanks Ilya,
 Anita.
 
And one question I will have when we discuss this is regarding this
statement: Green cell - tests ran successfully,

Currently we don't have community consensus around the use of the
statement tests ran successfully regarding third party ci systems.
This is as statement, you recall, we had discussed at the third party
meeting when we talked about driverlog.

18:46:02 krtaylor but what does CI tested really mean? just running
tests? or tested to pass some level of requirements?
http://eavesdrop.openstack.org/meetings/third_party/2014/third_party.2014-06-16-18.00.log.html

Using a statement to convey success prior to having a definition from
the community about what the statement means will create confusion in
the community and frustration from folks, including myself, who are
willing to have the conversation about what it means, who feel
circumvented by this second use of a phrase which implies a decided
meaning where none yet exists. Please participate in conversations
around the definition of phrases of success and failure regarding third
party systems and point to the logs where consensus is reached prior it
its use in future.

In addition to attending the third party meeting, please attend the
infra meeting or the qa meeting, and hopefully meetings that include
programs that have interactions with third party ci systems including
nova, neutron, and cinder (if there are other programs interacting with
third party ci systems please attend the third party meeting so I know
about you).

Thanks Ilya, I look forward to our future discussions,
Anita.

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev