On Monday, 21 November 2016 at 17:39:40 UTC, Andrei Alexandrescu wrote:
Hey folks, I'm working on a paper for fast median computation and https://issues.dlang.org/show_bug.cgi?id=16517 came to mind. I see the Google ngram corpus has occurrences of n-grams per year. Is data aggregated for all years available somewhere? I'd like to compute e.g. "the word (1-gram) with the median frequency across all English books" so I don't need the frequencies per year, only totals.

Of course I can download the entire corpus and then do some processing, but that would take a long time.

Also, if you can think of any large corpus that would be pertinent for median computation, please let me know!


Thanks,

Andrei

You might following worthwhile.

http://opendata.stackexchange.com/questions/6114/dataset-for-english-words-of-dictionary-for-a-nlp-project

I would just generate a bunch of integers randomly and use that, but I don't know if you specifically need to work with strings.

Reply via email to