On 07/07/2016 08:58 PM, Pranith Kumar Karampuri wrote:


On Fri, Jul 8, 2016 at 8:40 AM, Jeff Darcy <jda...@redhat.com <mailto:jda...@redhat.com>> wrote:

    > What gets measured gets managed.

    Exactly.  Reviewing is part of everyone's job, but reviews aren't
    tracked
    in any way that matters.  Contrast that with the *enormous*
    pressure most
    of us are under to get our own patches in, and it's pretty predictable
    what will happen.  We need to change that calculation.


    > What I have seen at least is that it is easy to find
    > people who sent patches, how many patches someone sent in a
    month etc. There
    > is no easy way to get these numbers for reviews. 'Reviewed-by'
    tag in commit
    > only includes the people who did +1/+2 on the final revision of
    the patch,
    > which is bad.

    That's a very good point.  I think people people who comment also get
    Reviewed-by: lines, but it doesn't matter because there's still a
    whole
    world of things completely outside of Gerrit.  Reviews done by
    email won't
    get counted, nor will consultations in the hallway or on IRC.  I
    have some
    ideas who's most active in those ways.  Some (such as yourself)
    show up in
    the Reviewed-by: statistics.  Others do not.  In terms of making sure
    people get all the credit they deserve, those things need to be
    counted
    too.  However, in terms of *getting the review queue unstuck* I'm
    not so
    sure.  What matters for that is the reviews that Gerrit uses to
    determine
    merge eligibility, so I think encouraging that specific kind of review
    still moves us in a positive direction.


In my experience at least it was only adding 'reviewied-by' for the people who gave +1/+2 on the final version of the patch

I agree about encouraging specific kind of review. At the same time we need to make reviewing, helping users in the community as important as sending patches in the eyes of everyone. It is very important to know these statistics to move in the right direction. My main problem with this is, everyone knows that reviews are important, then why are they not happening? Is it really laziness? Are we sure if there are people in the team who are not sharing the burden because of which it is becoming too much for 1 or 2 people to handle the total load? All these things become very easy to reason about if we have this data. Then I am sure we can easily find how best to solve this issue. Same goes for spurious failures. These are not problems that are not faced by others in the world either. I remember watching a video where someone shared (I think it was in google) that they started putting giant TVs in the hall-way in all the offices and the people who don't attend to spurious-build-failure problems would show up on the screen for everyone in the world to see. Apparently the guy with the biggest picture(the one who was not attending to any build failures at all I guess) came to these folks and asked how should he get his picture removed from the screen, and it was solved in a day or two. We don't have to go to those lengths, but we do need data to nudge people in the right direction.



Perhaps it's imposter syndrome. I know that even when I do leave comments on a patch, I don't add a +-1 because I don't think that my vote counts. I know I'm not part of the core developers so maybe I'm right, I don't know. Maybe some sort of published guidelines or mentorship could help?
_______________________________________________
Gluster-devel mailing list
Gluster-devel@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-devel

Reply via email to