On Wed, May 5, 2010 at 16:09, Reinhard Tartler <siret...@tauware.de> wrote: > On Mon, May 03, 2010 at 00:55:43 (CEST), Felipe Sateler wrote: > >> Ideas on how to make it faster are welcome (somehow finding out at >> once all the up to date repositories would be good). The script is >> specially slow in the best-case scenario (when all repositories are up >> to date). Also, it currently does not abort when a pull was not a fast >> forward, since gbp-pull does not return non-zero on that event. A bug >> has been filed against git-buildpackage. >> Finally, if some package has been removed from the pkg-multimedia >> area, it will not be removed from the local computer. > > What I wonder is what's the correct workflow.
It depends on what do you want to achieve. > Your script clones ever > pkg-multimedia git repository. That's the whole purpose :). I want to be able to look up a package's source without having to clone it JIT. >Do you have numbers how much disk space > this takes? fel...@pcfelipe:pkg-multimedia% du -s . 1.1G . However, running git gc over all of them gives a bit of space savings (small, but probably because most disk space usage is in the checkout, which can't be compressed). Just for fun, lets get the 10 piggies out of here: fel...@pcfelipe:pkg-multimedia% \du -s * | sort -nr | head 120520 vlc 71268 snd 68580 mplayer 58024 audacity 56176 rosegarden 38156 hydrogen-drumkits 33720 ardour 27948 csound 27236 ffmpeg 25160 freewrl > How much data needs to be transferred, and how long does git > need for that? fsateler-gu...@alioth:~% du -sh /git/pkg-multimedia 1.1G /git/pkg-multimedia So there's an upper bound. I didn't measure the actual transfer, though. I wonder if git gc would make any space gains in there. > > Just a brainfart, but perhaps this functionality can be implemented > better in gbp-clone by adding a concept of group-cloneing? Probably. But that would require me having to actually think out how to do this elegantly, and then code it up in python, and I just don't have time now. > > BTW, the list of repos can be probably generated by a cronjob, that way > also non-team members can test and use your script. How so? A cron job on alioth with results exported somewhere? Non team members still get read access. Non alioth users are another matter, though. -- Saludos, Felipe Sateler _______________________________________________ pkg-multimedia-maintainers mailing list email@example.com http://lists.alioth.debian.org/mailman/listinfo/pkg-multimedia-maintainers