On Mon, Oct 25, 2010 at 12:38 PM, Paul Houle <[email protected]> wrote:
>     I want Wikipedia converted into facts in a representation system
> that supports modal,  temporal,  and "microtheory" reasoning.  You
> know,  in the "real" world,  :James_T_Kirk is a :Fictional_Character,
> but in the Star Trek universe,  he's a :Person.

This sounds like it would take far more work to actually write the
program in the first place than to parallelize it.

On Mon, Oct 25, 2010 at 4:47 PM, Platonides <[email protected]> wrote:
> Make the best dump compressor ever? :)
>
> The page http://www.mediawiki.org/wiki/Dbzip2 is worth looking at just
> for the available options. Continuing dbzip2 is the first idea but not
> the only one. I'm sure many things can be dig from there.
> Also worth noting, Ariel has been doing the last en dumps in page batches.

Possible.  It looks like dbzip2 has had a lot of optimization put into
it already, so I don't know that there would be much low-hanging fruit
for me to get.  I'm not sure if that would be acceptable for a final
project ("do lots of benchmarking and probably not end up improving it
much at all").  Whereas rewriting it to use the GPU sounds like a
suspiciously large project . . . plus I'm not sure GPUs would even be
suited for it.

_______________________________________________
Wikitech-l mailing list
[email protected]
https://lists.wikimedia.org/mailman/listinfo/wikitech-l

Reply via email to