I have an improvement to the process. I just made a scrapy <https://scrapy.org/> script that can download the PR title and description from Github. (Yeah that should be accessible via the Github API but I couldn't figure out how.)
So I can make a spreadsheet or an HTML doc that has all the PR descriptions in it. Seems like this would be faster to than clicking on them but I don't know. I'll give that a shot and post it here. -adam On Thu, Apr 2, 2020 at 11:35 AM Nathan Hartman <hartman.nat...@gmail.com> wrote: > On Thu, Apr 2, 2020 at 1:01 PM Adam Feuer <a...@starcat.io> wrote: > > > > Bumping this up. It seems like we need a plan to tackle going through the > > 613 closed PRs and summarizing (only merged ones need to be summarized). > > This would be easier with a team of people... anyone want to help? > > > > One way we could do this is one group take bug fixes, and another take > > features, and then make a list for each. Then we put them together in a > > document. > > So that we don't duplicate work, I think we should also stick to a range of > PRs. > > For example I could go through PRs 1 through 200 looking for new features > -- architectures, drivers, boards, etc -- but *not* bug fixes per Adam's > suggestion and make a list of those. Another volunteer could look at PRs > 201 through 400, etc. > > So we need 3 volunteers to look for features, I'm 1 so we need 2 more. > > And we need the same thing for bug fixes, 3 volunteers. > > Any takers? > > Nathan > -- Adam Feuer <a...@starcat.io>