Re: [openstack-dev] [neutron] [third-party] Neutron 3rd Party CI status dashboard

2014-06-30 Thread Sukhdev Kapur
Sorry, accidentally hit the wrong key and message went out.

Was making a mention about the definition of Success. I thought the debate
in the meeting was very productive - when a CI posts a +1 that is success,
and when a CI posts a -1 (or no vote with comment) is also a success - as
this reflects that the CI is doing what it is suppose to do.

So, when it comes to stackalytics, it is more critical to show if a given
CI is operational or not - and for how long?
Another thing we can debate is how to present the +1/-1 votes by a given CI
- unless we have some benchmark, it will be hard to consider
success/failure.

So, I am of the  opinion that, initially, we only report on the operational
status and duration of the CIs, and a counter of +1 and -1 votes over a
period of time. For example, looking at Arista CI, it has casted 7,958
votes so far and it has been operational for past 6 months. This
information is not available anywhere - hence, presenting this kind of
information on a dashboard created by Ilya would be very useful to the
community as well to the vendors..

thoughts?

-Sukhdev





On Mon, Jun 30, 2014 at 12:49 PM, Sukhdev Kapur 
wrote:

> Well, Luke, this is collaborative effort by everybody. By having these CI
> systems in place ensures that one person's code does not break other
> person's code and vice versa. Therefore, having these CI systems
> operational and voting 24x7 is a critical step in achieving this goal.
>
> However, the details as to how and what should be tested is definitely
> debatable and the team has done excellent job in converging on that.
>
> Now, as to the issue at hand which Anita is describing, I attended the
> meeting this morning and was very pleased with the debate that took place
> and the definition as to Sucess
>
>
> On Mon, Jun 30, 2014 at 12:27 PM, Luke Gorrie  wrote:
>
>> On 30 June 2014 21:08, Anita Kuno  wrote:
>>
>>> I am disappointed to realize that Ilya (or stackalytics, I don't know
>>> where this is coming from) is unwilling to cease making up definitions
>>> of success for third party ci systems to allow the openstack community
>>> to arrive at its own definition.
>>>
>>
>> There is indeed a risk that the new dashboards won't give a meaningful
>> view of whether a 3rd party CI is voting correctly or not.
>>
>>  However, there is an elephant in the room and a much more important
>> problem:
>>
>> To measure how accurately a CI is voting says much more about a driver
>> author's "Gerrit-fu" ability to operate a CI system than it does about
>> whether the code they have contributed to OpenStack actually works, and the
>> latter is what is actually important.
>>
>> To my mind the whole 3rd party testing discussion should refocus on
>> helping developers maintain good working code and less on waving "you will
>> be kicked out of OpenStack if you don't keep your swarm of nebulous daemons
>> running 24/7".
>>
>>
>>
>> ___
>> OpenStack-dev mailing list
>> OpenStack-dev@lists.openstack.org
>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
>>
>>
>
___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [neutron] [third-party] Neutron 3rd Party CI status dashboard

2014-06-30 Thread Sukhdev Kapur
Well, Luke, this is collaborative effort by everybody. By having these CI
systems in place ensures that one person's code does not break other
person's code and vice versa. Therefore, having these CI systems
operational and voting 24x7 is a critical step in achieving this goal.

However, the details as to how and what should be tested is definitely
debatable and the team has done excellent job in converging on that.

Now, as to the issue at hand which Anita is describing, I attended the
meeting this morning and was very pleased with the debate that took place
and the definition as to Sucess


On Mon, Jun 30, 2014 at 12:27 PM, Luke Gorrie  wrote:

> On 30 June 2014 21:08, Anita Kuno  wrote:
>
>> I am disappointed to realize that Ilya (or stackalytics, I don't know
>> where this is coming from) is unwilling to cease making up definitions
>> of success for third party ci systems to allow the openstack community
>> to arrive at its own definition.
>>
>
> There is indeed a risk that the new dashboards won't give a meaningful
> view of whether a 3rd party CI is voting correctly or not.
>
> However, there is an elephant in the room and a much more important
> problem:
>
> To measure how accurately a CI is voting says much more about a driver
> author's "Gerrit-fu" ability to operate a CI system than it does about
> whether the code they have contributed to OpenStack actually works, and the
> latter is what is actually important.
>
> To my mind the whole 3rd party testing discussion should refocus on
> helping developers maintain good working code and less on waving "you will
> be kicked out of OpenStack if you don't keep your swarm of nebulous daemons
> running 24/7".
>
>
>
> ___
> OpenStack-dev mailing list
> OpenStack-dev@lists.openstack.org
> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
>
>
___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [neutron] [third-party] Neutron 3rd Party CI status dashboard

2014-06-30 Thread Anita Kuno
On 06/30/2014 03:27 PM, Luke Gorrie wrote:
> On 30 June 2014 21:08, Anita Kuno  wrote:
> 
>> I am disappointed to realize that Ilya (or stackalytics, I don't know
>> where this is coming from) is unwilling to cease making up definitions
>> of success for third party ci systems to allow the openstack community
>> to arrive at its own definition.
>>
> 
> There is indeed a risk that the new dashboards won't give a meaningful view
> of whether a 3rd party CI is voting correctly or not.
> 
> However, there is an elephant in the room and a much more important problem:
> 
> To measure how accurately a CI is voting says much more about a driver
> author's "Gerrit-fu" ability to operate a CI system than it does about
> whether the code they have contributed to OpenStack actually works, and the
> latter is what is actually important.
> 
> To my mind the whole 3rd party testing discussion should refocus on helping
> developers maintain good working code and less on waving "you will be
> kicked out of OpenStack if you don't keep your swarm of nebulous daemons
> running 24/7".
> 
> 
> 
> ___
> OpenStack-dev mailing list
> OpenStack-dev@lists.openstack.org
> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
> 
Well I am glad you have reached that conclusion, Luke.

Given that I gave you months of my time with that exact approach and you
squandered it away, claiming I wasn't doing enough to help you, I am
glad you have finally reached a productive place.

I hope your direction produces results.

Thanks Luke,
Anita.

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [neutron] [third-party] Neutron 3rd Party CI status dashboard

2014-06-30 Thread Luke Gorrie
On 30 June 2014 21:08, Anita Kuno  wrote:

> I am disappointed to realize that Ilya (or stackalytics, I don't know
> where this is coming from) is unwilling to cease making up definitions
> of success for third party ci systems to allow the openstack community
> to arrive at its own definition.
>

There is indeed a risk that the new dashboards won't give a meaningful view
of whether a 3rd party CI is voting correctly or not.

However, there is an elephant in the room and a much more important problem:

To measure how accurately a CI is voting says much more about a driver
author's "Gerrit-fu" ability to operate a CI system than it does about
whether the code they have contributed to OpenStack actually works, and the
latter is what is actually important.

To my mind the whole 3rd party testing discussion should refocus on helping
developers maintain good working code and less on waving "you will be
kicked out of OpenStack if you don't keep your swarm of nebulous daemons
running 24/7".
___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [neutron] [third-party] Neutron 3rd Party CI status dashboard

2014-06-30 Thread Anita Kuno
On 06/29/2014 07:59 PM, Anita Kuno wrote:
> On 06/29/2014 07:43 PM, Anita Kuno wrote:
>> On 06/29/2014 03:25 PM, Ilya Shakhat wrote:
>>> Hi!
>>>
>>> During last couple weeks there is an increasing demand on tracking
>>> 3rd-party CI statuses. We at Stackalytics decided to be in trend and (with
>>> some inspiration from Salvatore's proposal) implemented report that shows
>>> summary on external CI status. The initial version is available for Neutron
>>> - http://stackalytics.com/report/ci/neutron/7
>>>
>>> The report shows summary of all CI jobs during specified period of time,
>>> including:
>>>  * stats of runs on merged patch sets:
>>> - total number of runs
>>> - success rate (success to total ratio)
>>> - time of the latest run
>>> - last test result
>>>   * stats for all patch sets (the same set as for merged)
>>>   * last test results for every merged patch set grouped by days (useful to
>>> see how different CIs correlate with each other and how often they run)
>>>
>>> Under "merged patch set" report means "the last patch set in the merged
>>> change request", thus it is almost the same as the trunk code. CI
>>> configuration is taken from DriverLog's default data
>>> .
>>> Standard Stackalytics screen is also available for CIs -
>>> http://stackalytics.com/?metric=ci, including votes breakdown and activity
>>> log.
>>>
>>> Since this is the first version there are some open questions:
>>>  * Currently report shows results per CI id, but there are CIs that run
>>> tests against multiple drivers and this case is not supported. What would
>>> be more useful: to get stats for a driver or for CI?
>>>  * Most CIs run tests when patch set is posted. So even if change request
>>> is merged within selected time period corresponding CI results may be
>>> missing.
>>>  * Patterns for non-voting CIs need to be verified. For example Cisco CI
>>> now runs 5 jobs, but DriverLog data still contains old data.
>>>
>>> Thanks,
>>> Ilya
>>>
>>> 2014-06-16 17:52 GMT+04:00 Salvatore Orlando :
>>>

 However, it would be great if we could start devising a solution for
 having "health" reports from the various CI systems.
 This report should report the following kind of information:
 - timestamp of last run
 - timestamp of last vote (a system might start job which then get aborted
 for CI infra problems)
 - % of success vs failures (not sure how important is that one but
 provides a metric of comparison with upstream jenkins)
 - % of disagreements with jenkins (this might allow us to quickly spot
 those CI systems which are randomly -1'ing patches)

 The percentage metrics might be taken over a 48 hours or 7 days interval,
 or both.
 Does this idea sound reasonable?


>>>
>>>
>>>
>>> ___
>>> OpenStack-dev mailing list
>>> OpenStack-dev@lists.openstack.org
>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
>>>
>> Hi Ilya:
>>
>> I look forward to hearing more about this dashboard and ensuring you or
>> someone else associated with this dashboard are available for questions
>> at the third party meeting tomorrow:
>> https://wiki.openstack.org/wiki/Meetings/ThirdParty
>>
>> We missed you last week.
>>
>> Thanks Ilya,
>> Anita.
>>
> And one question I will have when we discuss this is regarding this
> statement: "Green cell - tests ran successfully,"
> 
> Currently we don't have community consensus around the use of the
> statement "tests ran successfully" regarding third party ci systems.
> This is as statement, you recall, we had discussed at the third party
> meeting when we talked about driverlog.
> 
> 18:46:02  but what does CI tested really mean? just running
> tests? or tested to pass some level of requirements?
> http://eavesdrop.openstack.org/meetings/third_party/2014/third_party.2014-06-16-18.00.log.html
> 
> Using a statement to convey success prior to having a definition from
> the community about what the statement means will create confusion in
> the community and frustration from folks, including myself, who are
> willing to have the conversation about what it means, who feel
> circumvented by this second use of a phrase which implies a decided
> meaning where none yet exists. Please participate in conversations
> around the definition of phrases of success and failure regarding third
> party systems and point to the logs where consensus is reached prior it
> its use in future.
> 
> In addition to attending the third party meeting, please attend the
> infra meeting or the qa meeting, and hopefully meetings that include
> programs that have interactions with third party ci systems including
> nova, neutron, and cinder (if there are other programs interacting with
> third party ci systems please attend the third party meeting so I know
> about you).
> 
> Thanks Ilya, I look forward to our fut

Re: [openstack-dev] [neutron] [third-party] Neutron 3rd Party CI status dashboard

2014-06-29 Thread Anita Kuno
On 06/29/2014 07:43 PM, Anita Kuno wrote:
> On 06/29/2014 03:25 PM, Ilya Shakhat wrote:
>> Hi!
>>
>> During last couple weeks there is an increasing demand on tracking
>> 3rd-party CI statuses. We at Stackalytics decided to be in trend and (with
>> some inspiration from Salvatore's proposal) implemented report that shows
>> summary on external CI status. The initial version is available for Neutron
>> - http://stackalytics.com/report/ci/neutron/7
>>
>> The report shows summary of all CI jobs during specified period of time,
>> including:
>>  * stats of runs on merged patch sets:
>> - total number of runs
>> - success rate (success to total ratio)
>> - time of the latest run
>> - last test result
>>   * stats for all patch sets (the same set as for merged)
>>   * last test results for every merged patch set grouped by days (useful to
>> see how different CIs correlate with each other and how often they run)
>>
>> Under "merged patch set" report means "the last patch set in the merged
>> change request", thus it is almost the same as the trunk code. CI
>> configuration is taken from DriverLog's default data
>> .
>> Standard Stackalytics screen is also available for CIs -
>> http://stackalytics.com/?metric=ci, including votes breakdown and activity
>> log.
>>
>> Since this is the first version there are some open questions:
>>  * Currently report shows results per CI id, but there are CIs that run
>> tests against multiple drivers and this case is not supported. What would
>> be more useful: to get stats for a driver or for CI?
>>  * Most CIs run tests when patch set is posted. So even if change request
>> is merged within selected time period corresponding CI results may be
>> missing.
>>  * Patterns for non-voting CIs need to be verified. For example Cisco CI
>> now runs 5 jobs, but DriverLog data still contains old data.
>>
>> Thanks,
>> Ilya
>>
>> 2014-06-16 17:52 GMT+04:00 Salvatore Orlando :
>>
>>>
>>> However, it would be great if we could start devising a solution for
>>> having "health" reports from the various CI systems.
>>> This report should report the following kind of information:
>>> - timestamp of last run
>>> - timestamp of last vote (a system might start job which then get aborted
>>> for CI infra problems)
>>> - % of success vs failures (not sure how important is that one but
>>> provides a metric of comparison with upstream jenkins)
>>> - % of disagreements with jenkins (this might allow us to quickly spot
>>> those CI systems which are randomly -1'ing patches)
>>>
>>> The percentage metrics might be taken over a 48 hours or 7 days interval,
>>> or both.
>>> Does this idea sound reasonable?
>>>
>>>
>>
>>
>>
>> ___
>> OpenStack-dev mailing list
>> OpenStack-dev@lists.openstack.org
>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
>>
> Hi Ilya:
> 
> I look forward to hearing more about this dashboard and ensuring you or
> someone else associated with this dashboard are available for questions
> at the third party meeting tomorrow:
> https://wiki.openstack.org/wiki/Meetings/ThirdParty
> 
> We missed you last week.
> 
> Thanks Ilya,
> Anita.
> 
And one question I will have when we discuss this is regarding this
statement: "Green cell - tests ran successfully,"

Currently we don't have community consensus around the use of the
statement "tests ran successfully" regarding third party ci systems.
This is as statement, you recall, we had discussed at the third party
meeting when we talked about driverlog.

18:46:02  but what does CI tested really mean? just running
tests? or tested to pass some level of requirements?
http://eavesdrop.openstack.org/meetings/third_party/2014/third_party.2014-06-16-18.00.log.html

Using a statement to convey success prior to having a definition from
the community about what the statement means will create confusion in
the community and frustration from folks, including myself, who are
willing to have the conversation about what it means, who feel
circumvented by this second use of a phrase which implies a decided
meaning where none yet exists. Please participate in conversations
around the definition of phrases of success and failure regarding third
party systems and point to the logs where consensus is reached prior it
its use in future.

In addition to attending the third party meeting, please attend the
infra meeting or the qa meeting, and hopefully meetings that include
programs that have interactions with third party ci systems including
nova, neutron, and cinder (if there are other programs interacting with
third party ci systems please attend the third party meeting so I know
about you).

Thanks Ilya, I look forward to our future discussions,
Anita.

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/lis

Re: [openstack-dev] [neutron] [third-party] Neutron 3rd Party CI status dashboard

2014-06-29 Thread Anita Kuno
On 06/29/2014 03:25 PM, Ilya Shakhat wrote:
> Hi!
> 
> During last couple weeks there is an increasing demand on tracking
> 3rd-party CI statuses. We at Stackalytics decided to be in trend and (with
> some inspiration from Salvatore's proposal) implemented report that shows
> summary on external CI status. The initial version is available for Neutron
> - http://stackalytics.com/report/ci/neutron/7
> 
> The report shows summary of all CI jobs during specified period of time,
> including:
>  * stats of runs on merged patch sets:
> - total number of runs
> - success rate (success to total ratio)
> - time of the latest run
> - last test result
>   * stats for all patch sets (the same set as for merged)
>   * last test results for every merged patch set grouped by days (useful to
> see how different CIs correlate with each other and how often they run)
> 
> Under "merged patch set" report means "the last patch set in the merged
> change request", thus it is almost the same as the trunk code. CI
> configuration is taken from DriverLog's default data
> .
> Standard Stackalytics screen is also available for CIs -
> http://stackalytics.com/?metric=ci, including votes breakdown and activity
> log.
> 
> Since this is the first version there are some open questions:
>  * Currently report shows results per CI id, but there are CIs that run
> tests against multiple drivers and this case is not supported. What would
> be more useful: to get stats for a driver or for CI?
>  * Most CIs run tests when patch set is posted. So even if change request
> is merged within selected time period corresponding CI results may be
> missing.
>  * Patterns for non-voting CIs need to be verified. For example Cisco CI
> now runs 5 jobs, but DriverLog data still contains old data.
> 
> Thanks,
> Ilya
> 
> 2014-06-16 17:52 GMT+04:00 Salvatore Orlando :
> 
>>
>> However, it would be great if we could start devising a solution for
>> having "health" reports from the various CI systems.
>> This report should report the following kind of information:
>> - timestamp of last run
>> - timestamp of last vote (a system might start job which then get aborted
>> for CI infra problems)
>> - % of success vs failures (not sure how important is that one but
>> provides a metric of comparison with upstream jenkins)
>> - % of disagreements with jenkins (this might allow us to quickly spot
>> those CI systems which are randomly -1'ing patches)
>>
>> The percentage metrics might be taken over a 48 hours or 7 days interval,
>> or both.
>> Does this idea sound reasonable?
>>
>>
> 
> 
> 
> ___
> OpenStack-dev mailing list
> OpenStack-dev@lists.openstack.org
> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
> 
Hi Ilya:

I look forward to hearing more about this dashboard and ensuring you or
someone else associated with this dashboard are available for questions
at the third party meeting tomorrow:
https://wiki.openstack.org/wiki/Meetings/ThirdParty

We missed you last week.

Thanks Ilya,
Anita.

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev