"s gave GC overhead limit exceeded errors in ABt-job which indicates a..."
There was one of those patches that i did for Sebastien's experiments that tackles that quite a bit, along with some problems you cite with ABt job -d On Thu, Feb 23, 2012 at 4:56 PM, Dmitriy Lyubimov <dlie...@gmail.com> wrote: > Reading further... Yep. that's exactly how it is done there in > distributed QR solver. At least on top of it. > > On Thu, Feb 23, 2012 at 4:54 PM, Dmitriy Lyubimov <dlie...@gmail.com> wrote: >> Wow. Cantor patterns for Givens rotations. I wondered if it already >> had a name or somebody already figured to do something similar. It >> looks like you really got into that level of details there. That's >> extremely cool, sir ! >> >> On Thu, Feb 23, 2012 at 4:45 PM, Dmitriy Lyubimov <dlie...@gmail.com> wrote: >>> Thank you, Nathan. >>> >>> On Wed, Feb 22, 2012 at 7:01 PM, Nathan Halko <nat...@spotinfluence.com> >>> wrote: >>>> Hi Dmitriy, >>>> >>>> Just a few comments: >>>> >>>> --the computed factors are approximate A \approx U\SigmaV^{T} >>> >>> Thanks, agreed. >>> >>>> >>>> -- the projection steps seemed transposed to me but they are consistent >>>> throughout ie. >>>> (2) \tilde{u} = \tilde{c}_{r} V \Sigma^{-1} >>> >>> Yes this is probably an earlier error, but in section 3 fold in >>> expressions I beleive should be correct. I assume the convention is >>> that all vectors in equations have columnar orientation (i.e. x'x is >>> inner product, xx' is always outer product). I will check it >>> >>>> >>>> p. 3: transpose \xi to emphasize row vector >>>> >>>> - 'mean of all rows' is a bit misleading, \xi entries are the mean of each >>>> column (column-wise mean as you state below) >>>> >>> >>> Yeah this keeps coming up. means of rows is the same as column mean. >>> Column mean seems to sound more familiar to people, but mean of rows >>> seems to be more visual: if we have a bunch of data points in multiple >>> dimensions and compute their 'center' (mean) then we say "center of >>> points', or applying to pca situation it converts to 'mean of rows'. >>> But i think concensus is growing that we should always opt out for >>> 'column mean' or at least not mix the two to prevent confusion. >>> >>> >>>> - dimention -> dimension >>>> >>>> I haven't code dived into the new pca code to be familiar with it so the >>>> above comments are just picky notational stuff. I did however, do some >>>> extensive analysis on the standard decomposition part (as of 0.6 SNAPSHOT) >>>> which can be found here >>> >>> >>> Yeah i meant validation of PCA approach. There seems to be somewhat >>> different ways to do it. Some people run eigendecomposition on a >>> covariance matrix which i guess would be adjusted for 1/n. which >>> should be technically equivalent to running svd and then adjusting >>> singular values for n^-0.5 but since nobody really cares about >>> singular values after PCA is done, it seems to be moot. Also it >>> doesn't seem to affect the transformational equations in any way. >>> >>> I was also not sure if i could safely label U rows as original >>> datapoints converted into PCA space (is there is such a thing as PCA >>> space anyway? I saw this concept in some texts i think but i now not >>> sure what was meant by it back there). >>> >>>> >>>> http://amath.colorado.edu/faculty/martinss/Pubs/2012_halko_dissertation.pdf >>>> (starting page 139) >>> >>> This is all cool stuff. I will read it as soon as i get a spare time >>> window. Great! >>> >>> once again, thank you for doing this. >>> >>> -d