Re: [OT] All your medians are belong to me

2016-11-21 Thread Patrick Schluter via Digitalmars-d
On Monday, 21 November 2016 at 18:39:26 UTC, Andrei Alexandrescu wrote: On 11/21/2016 01:18 PM, jmh530 wrote: I would just generate a bunch of integers randomly and use that, but I don't know if you specifically need to work with strings. I have that, too, but was looking for some real data

Re: [OT] All your medians are belong to me

2016-11-21 Thread Andrei Alexandrescu via Digitalmars-d
On 11/21/2016 01:18 PM, jmh530 wrote: I would just generate a bunch of integers randomly and use that, but I don't know if you specifically need to work with strings. I have that, too, but was looking for some real data as well. It would be a nice addition. -- Andrei

Re: [OT] All your medians are belong to me

2016-11-21 Thread jmh530 via Digitalmars-d
On Monday, 21 November 2016 at 17:39:40 UTC, Andrei Alexandrescu wrote: Hey folks, I'm working on a paper for fast median computation and https://issues.dlang.org/show_bug.cgi?id=16517 came to mind. I see the Google ngram corpus has occurrences of n-grams per year. Is data aggregated for all

[OT] All your medians are belong to me

2016-11-21 Thread Andrei Alexandrescu via Digitalmars-d
Hey folks, I'm working on a paper for fast median computation and https://issues.dlang.org/show_bug.cgi?id=16517 came to mind. I see the Google ngram corpus has occurrences of n-grams per year. Is data aggregated for all years available somewhere? I'd like to compute e.g. "the word (1-gram)