Either only wikitext or both, at least until I would get some wikitext
to html convertor lib

On Sun, Jan 25, 2015 at 7:41 PM, rupert THURNER
<[email protected]> wrote:
> The storage format is very efficient and there is a c library for it :
> https://libgit2.github.com
> It should be not necessary to create complex versioning around it.
>
> You plan to store html or wikitext?
>
> Rupert
> On Jan 25, 2015 6:37 PM, "Petr Bena" <[email protected]> wrote:
>
>> I don't really know, it is technically possible but probably not
>> suitable. I don't want to create offline wiki. Just a reader of a
>> wiki, so no complex versioning is required for that.
>>
>> On Sun, Jan 25, 2015 at 6:00 PM, rupert THURNER
>> <[email protected]> wrote:
>> > Petr, do you think it would be an option to   use git version control as
>> a
>> > storage format instead of openzim? Which would facilitate edit and merge
>> > back changes?
>> >
>> > Rupert
>> > On Jan 23, 2015 11:59 AM, "Petr Bena" <[email protected]> wrote:
>> >
>> >> Hi,
>> >>
>> >> I know most of you hate reinventing a wheel so I first send it here,
>> >> before I launch that project :)
>> >>
>> >> Some of you probably know kiwix - kiwix.org which is offline wikipedia
>> >> reader. I think the idea of this reader is cool, most of you probably
>> >> sometimes wanted to access wikipedia while being offline somewhere,
>> >> but couldn't. Kiwix can help with this, however it has one big problem
>> >> and solution for it is so complex that it would basically need a
>> >> rewrite of whole thing.
>> >>
>> >> That problem is that you need to download pretty huge file (40+GB) in
>> >> order to use it for en wikipedia for example. And if you wanted to
>> >> update those few wikipages you are interested in, to a latest
>> >> revision, then you again need to download that huge file.
>> >>
>> >> That suck. Especially with GPRS internet and similar connectivity and
>> >> it also suck because mobile phones don't even have space for so much
>> >> data. My idea is to create app similar to kiwix, that would use SQLite
>> >> DB and using wikipedia API it would (slowly, apache friendly) download
>> >> contents of any mediawiki installation based on user selection, so
>> >> that you could download just a 1 page for offline reading, or 1
>> >> category. Or 1000 categories. Or precompiled sets of pages created by
>> >> users (books). You could easily update these using API anytime to
>> >> latest version. You could get media files for these pages, etc, etc...
>> >> (You could probably even edit the pages offline, and then update them
>> >> when you are online, but that is just extra feature)
>> >>
>> >> I think this approach would work much better and it's sad kiwix
>> >> already doesn't support it. At some point, if it worked I think this
>> >> new code could be merged back into kiwix, I am going to use C++ in the
>> >> end, which kiwix uses as well.
>> >>
>> >> What do you think about it, is it worth of working on? Is there
>> >> actually a community of "offline wikipedia readers" that would
>> >> appreciate it?
>> >>
>> >> Thanks
>> >>
>> >> _______________________________________________
>> >> Wikitech-l mailing list
>> >> [email protected]
>> >> https://lists.wikimedia.org/mailman/listinfo/wikitech-l
>> > _______________________________________________
>> > Wikitech-l mailing list
>> > [email protected]
>> > https://lists.wikimedia.org/mailman/listinfo/wikitech-l
>>
>> _______________________________________________
>> Wikitech-l mailing list
>> [email protected]
>> https://lists.wikimedia.org/mailman/listinfo/wikitech-l
> _______________________________________________
> Wikitech-l mailing list
> [email protected]
> https://lists.wikimedia.org/mailman/listinfo/wikitech-l

_______________________________________________
Wikitech-l mailing list
[email protected]
https://lists.wikimedia.org/mailman/listinfo/wikitech-l

Reply via email to