On Fri, Nov 9, 2012 at 11:02 AM, Hydriz Wikipedia <[email protected]>wrote:

>
> Can you provide me with your GitHub username? I did a git-svn conversion
> to a repository there, and I can add you as a member to contribute the code.
>
>
One of the main advantages of git/Github is that you don't need to do
precoordination like this.  He can just fork the repository and start
working on it, then  send you a pull request when it's done.


> The repo is at https://github.com/WikiTeam/WikiGrabber.
>
> Its just a few revisions out of date, got to figure out a way to pull
> changes from the main repository to this unofficial clone. (Note that I can
> push directly to the main repository on Google Code)
>

You should still have SVN as a remote that you can pull from to get
yourself synced.

Tom

On Fri, Nov 9, 2012 at 11:56 PM, John <[email protected]> wrote:
>
>> If someone wants to contact me off list I will gladly re-write this
>> clusterfuck I just need some help with the details, and a step-by step
>> of exactly what/how you want the process of dumping a wiki done, Ive
>> done it before quite a few times myself and getting you a functional
>> bot shouldnt be that hard (I am both a python and API programmer :P )
>>
>> On Fri, Nov 9, 2012 at 5:27 AM, Federico Leva (Nemo) <[email protected]>
>> wrote:
>> > It's completely broken:
>> > https://code.google.com/p/wikiteam/issues/detail?id=56
>> > It will download only a fraction of the wiki, 500 pages at most per
>> > namespace.
>> >
>> > Let me reiterate that
>> https://code.google.com/p/wikiteam/issues/detail?id=44
>> > is a very urgent bug and we've seen no work on it in many months. We
>> need an
>> > actual programmer with some knowledge of python to fix it and make the
>> > script work properly; I know there are several on this list (and
>> elsewhere),
>> > please please help. The last time I, as a non-coder, tried to fix a
>> bug, I
>> > made things worse (
>> https://code.google.com/p/wikiteam/issues/detail?id=26).
>> >
>> > Only after API is implemented/fixed, I'll be able to re-archive the 4-5
>> > thousands wikis we've recently archived on archive.org
>> > (https://archive.org/details/wikiteam) and possibly many more. Many of
>> those
>> > dumps contain errors and/or are just partial because of the script's
>> > unreliability, and wikis die on a daily basis. (So, quoting emijrp,
>> there IS
>> > a deadline.)
>> >
>> > Nemo
>> >
>> > P.s.: Cc'ing some lists out of desperation; sorry for cross-posting.
>> >
>> > _______________________________________________
>> > Pywikipedia-l mailing list
>> > [email protected]
>> > https://lists.wikimedia.org/mailman/listinfo/pywikipedia-l
>>
>> _______________________________________________
>> Pywikipedia-l mailing list
>> [email protected]
>> https://lists.wikimedia.org/mailman/listinfo/pywikipedia-l
>>
>
>
>
> --
> Regards,
> Hydriz
>
> We've created the greatest collection of shared knowledge in history. Help
> protect Wikipedia. Donate now: http://donate.wikimedia.org
>
> _______________________________________________
> Pywikipedia-l mailing list
> [email protected]
> https://lists.wikimedia.org/mailman/listinfo/pywikipedia-l
>
>
_______________________________________________
Pywikipedia-l mailing list
[email protected]
https://lists.wikimedia.org/mailman/listinfo/pywikipedia-l

Reply via email to