Hello All,
My dataset has 93 features and just under 62,000 observations (61,878 to be
exact). I'm running out of memory right after the mean sigma value is
computed/displayed. I've tried using dimensionality reduction via TruncatedSVD
with n_components set at different levels (78, 50 and 2 respectively) prior to
sending the data to TSNE but I still run out of memory. For TSNE,
n_components=2 and perplexity=40 (I've also tried 20). I've got 24GB of RAM on
my 64-bit windows 7 machine. Should I try a subsample of the dataset and if so,
does anyone have a recommendation on the size? Thanks!
-Jason
------------------------------------------------------------------------------
BPM Camp - Free Virtual Workshop May 6th at 10am PDT/1PM EDT
Develop your own process in accordance with the BPMN 2 standard
Learn Process modeling best practices with Bonita BPM through live exercises
http://www.bonitasoft.com/be-part-of-it/events/bpm-camp-virtual- event?utm_
source=Sourceforge_BPM_Camp_5_6_15&utm_medium=email&utm_campaign=VA_SF
_______________________________________________
Scikit-learn-general mailing list
Scikit-learn-general@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/scikit-learn-general