Hi Peter,
thanks for the quick answer.
This dataset would (hopefully) motivate some design choices in a larger
project.
I'm working on this as my masters thesis. Therefore, involving you as a
co-author is not my decision to make.
Sharing our analysis and results with you is certainly possible
I have downloaded the 2014 logs from here:
https://archive.org/details/20140127MnemosynelogsAll.db
Now I would like to apply the parser as mentioned in this thread:
https://groups.google.com/forum/#!topic/mnemosyne-proj-users/QwmjLZLXtVY
The parser code is downloaded from here:
Hi,
Feel free to have your supervisor contact me off-list ☺
Cheers,
Peter
From: mnemosyne-proj-users@googlegroups.com
On Behalf Of lars klein
Sent: 28 November 2018 11:51
To: mnemosyne-proj-users
Subject: Re: [mnemosyne-proj-users] updated mnemosyne dataset
Hi Peter,
thanks for the quick
Hi,
I would like to do some research on spaced repetition and your dataset
looks amazing.
The last version comes from 2014 though, would it be possible to get an
updated version?
There was a thread about this in 2016, where Peter said he could upload to
FTP.
Does that offer still stand ?
I
Hi,
If you have e.g. an ftp site where I could drop a few gigs, I could do this,
provided that if you do serious academic research on this data leading to a
publication, I get involved as a co-author. The reason as that knowing how the
data is collected allows to prevent certain pitfalls,
Hi,
Let’s do this off-list, after sorting out the access rights to the data ☺
Cheers,
Peter
From: mnemosyne-proj-users@googlegroups.com
On Behalf Of lars klein
Sent: 28 November 2018 11:58
To: mnemosyne-proj-users
Subject: [mnemosyne-proj-users] running parse_logs.py
I have downloaded the
Are you sure you need an updated dataset? The 2014 one is enormous
already and it's difficult to see what you would do with another few
million rows that you couldn't do with the previous millions of rows.
--
gwern
https://www.gwern.net
--
You received this message because you are subscribed
Hi Gwern,
you're surely right.
My request was more a matter of principle.
In a previous project I made the mistake of changing the dataset while
fitting models.
Which invaluated previous results.
Since I don't know what preprocessing and cleaning is necessary, it made
sense to try and get the