and everything before (e.g., I generate the matrix using another DML) was indeed run by Spark and shows up on the UI.
On Sun, Apr 22, 2018 at 5:05 PM, Qifan Pu <qifan...@gmail.com> wrote: > Thanks Jeremy and Matthias. When I run the script, the cholesky or the inv > is executed completely on the driver, and nothing shows up on Spark UI. > Is that the expected behavior? > > On Sun, Apr 22, 2018 at 3:34 PM, Jeremy Nilmeier <nilme...@us.ibm.com> > wrote: > >> Yes, I also spoke with Sasha about this some time last year. Thanks for >> following up. >> >> Cheers, J >> >> >> Jerome Nilmeier, PhD >> Data Scientist and Engineer >> IBM Spark Technology Center >> http://www.spark.tc/ >> >> >> >> ----- Original message ----- >> From: Matthias Boehm <mboe...@gmail.com> >> To: dev@systemml.apache.org >> Cc: Qifan Pu <qifan...@gmail.com>, Jeremy Nilmeier <nilme...@us.ibm.com> >> Subject: Re: distributed cholesky on systemml >> Date: Sun, Apr 22, 2018 2:41 PM >> >> thanks for the context Jeremy - that helps. I also had an offline >> conversion with Sasha and he pointed me to a script that does exactly >> that (iterative invert_lower_triangular) combined with a parfor over >> independent blocks. We'll merge these scripts soon and I'll reach out >> individually as necessary. Thanks everybody for now. >> >> Regards, >> Matthias >> >> On Sun, Apr 22, 2018 at 12:40 PM, Jeremy Nilmeier <nilme...@us.ibm.com> >> wrote: >> > This may be a duplicate...it was bounced from the dev list. >> > >> > I think that scalable triangular inverse will also have similar >> properties, >> > in that there is a sequential approach if it uses back substitution. >> > >> > For most of these algorithms (LU, Cholesky, QR), they are inherently >> > sequential, and the focus of the work is on minimizing interprocess >> > communication during the operations, which may explain why there was >> only >> > limited interest in pursuing this further. >> > >> > I had originally recommended that the recursive algorithms be rewritten >> as >> > iterative algorithms (and in fact provided an example of the LU in >> iterative >> > form), which would make the counting of operations more transparent, as >> well >> > as revealing possible parallelization points. >> > >> > Cheers, J >> > Jerome Nilmeier, PhD >> > Data Scientist and Engineer >> > IBM Spark Technology Center >> > https://urldefense.proofpoint.com/v2/url?u=http-3A__www.spar >> k.tc_&d=DwIFaQ&c=jf_iaSHvJObTbx-siA1ZOg&r=3mYOfURw_FSirAnoSv >> 2pWvLSi1psso4F9RdGjEWL6yc&m=VIdNVaIRvibBlaNVAOXLKmxXf7ma-EXr >> LWbjMd9Bmgo&s=YktpBBbqor3DKzS90Ah75BF6NBYtE4RauITF7QaL87g&e= >> >> > >> > >> > >> > ----- Original message ----- >> > From: Matthias Boehm <mboe...@gmail.com> >> > To: dev@systemml.apache.org >> > Cc: Qifan Pu <qifan...@gmail.com> >> > Subject: Re: distributed cholesky on systemml >> > Date: Sun, Apr 22, 2018 1:21 AM >> > >> > sure no problem - thanks again for catching this issue that was hidden >> > for a while. >> > >> > Yes, the same depth-first characteristic applies to the Cholesky >> > function as well. In contrast to U_triangular_inv, however, there are >> > data dependencies between the blocks per level (at least in the >> > current algorithm formulation), which means we cannot use the approach >> > I described for U_triangular_inv. >> > >> > L11 = Cholesky(A11, nb) >> > A22 = ... U_triangular_inv(t(L11)) >> > L22 = Cholesky(A22, nb) >> > >> > However, note that there are much fewer calls to Cholesky due to the >> > switch to the builtin cholesky according to the given min block size. >> > For example, in our new test for dimensions 1362 x 1362 and min size >> > of 200, we call Cholesky 15 times but U_triangular_inv 2539 times. >> > >> > For sufficiently large min block size this might be ok for Cholesky, >> > because each level also does a number of matrix multiplies that will >> > exploit the available parallelism of your cluster. In that regard. you >> > might want to experiment with different block sizes and driver memory >> > budgets. If I get a chance, I will also run a number of experiments >> > and see if we can rewrite these scripts. >> > >> > Regards, >> > Matthias >> > >> > On Sun, Apr 22, 2018 at 12:48 AM, Qifan Pu <qifan...@gmail.com> wrote: >> >> Matthias, >> >> >> >> Thanks so much for taking time to fix. Really appreciated it. >> >> Does the same reasoning apply to the cholesky script? The recursive >> >> approach >> >> also looks inherently sequential. >> >> >> >> Best, >> >> Qifan >> >> >> >> On Sat, Apr 21, 2018 at 11:39 PM, Matthias Boehm <mboe...@gmail.com> >> >> wrote: >> >>> >> >>> just as a quick update: this issue has now been fixed in SystemML >> >>> master - it was essentially a missing guard for recursive functions >> >>> when checking for unary size-preserving functions during >> >>> inter-procedural analysis (IPA). >> >>> >> >>> However, while working with this recursive cholesky function I came to >> >>> the conclusion that it may need some rework. The current top-down, >> >>> depth-first, approach is inherently sequential. This is partially >> >>> unnecessary because for the used recursive function U_triangular_inv >> >>> (which is called many more times than cholesky), blocks per level are >> >>> independent. Therefore, we should look into a bottom-up, breadth-first >> >>> approach to parallelize over the blocks in each level, which could be >> >>> done via parfor at script level. >> >>> >> >>> Regards, >> >>> Matthias >> >>> >> >>> On Sat, Apr 21, 2018 at 6:59 PM, Matthias Boehm <mboe...@gmail.com> >> >>> wrote: >> >>> > thanks for catching this - I just ran a toy example and this seems >> to >> >>> > be a rewrite issue (there are specific right indexing rewrites that >> >>> > collapse U[1:k,1:k] and U[1:k,k+1:n] into a single access to U which >> >>> > helps for large distributed matrices). As a workaround, you can set >> >>> > "sysml.optlevel" to 1 (instead of default 2, where 1 disables all >> >>> > rewrites), which worked fine for me. I'll fix this later today. Also >> >>> > I'll fix the naming from "Choleskey" to "Cholesky". Thanks again. >> >>> > >> >>> > Regards, >> >>> > Matthias >> >>> > >> >>> > >> >>> > On Sat, Apr 21, 2018 at 6:28 PM, Qifan Pu <qifan...@gmail.com> >> wrote: >> >>> >> Hi Matthias, >> >>> >> >> >>> >> Thanks for the fast response and detailed information. This is >> really >> >>> >> helpful. >> >>> >> >> >>> >> I just tried to run it, and was tracing down a indexing bug that >> can >> >>> >> be >> >>> >> repeated by simply running the test script of triangle solve[1] >> >>> >> Caused by: org.apache.sysml.runtime.DMLRuntimeException: Invalid >> >>> >> values >> >>> >> for >> >>> >> matrix indexing: [1667:3333,1:1666] must be within matrix >> dimensions >> >>> >> [1000,1000] >> >>> >> >> >>> >> >> >>> >> Am I missing some configuration here? >> >>> >> >> >>> >> >> >>> >> [1] >> >>> >> >> >>> >> >> >>> >> https://urldefense.proofpoint.com/v2/url?u=https-3A__github. >> com_apache_systemml_blob_master_scripts_staging_scalable- >> 5Flinalg_test_test-5Ftriangular-5Finv.dml&d=DwIBaQ&c=jf_ >> iaSHvJObTbx-siA1ZOg&r=3mYOfURw_FSirAnoSv2pWvLSi1psso >> 4F9RdGjEWL6yc&m=FvqDr_AKzY5EAD_GAXIJoot0Z09NtMUt8kLS >> hXcJxqQ&s=zIEgt74yeZzCTqvLCgV_0J8ECApG541uUlbaGMcK8bs&e= >> >>> >> >> >>> >> >> >>> >> Best, >> >>> >> Qifan >> >>> >> >> >>> >> >> >>> >> On Sat, Apr 21, 2018 at 4:06 PM, Matthias Boehm <mboe...@gmail.com >> > >> >>> >> wrote: >> >>> >>> >> >>> >>> Hi Qifan, >> >>> >>> >> >>> >>> thanks for your feedback. You're right, the builtin functions >> >>> >>> cholesky, inverse, eigen, solve, svd, qr, and lu are currently >> only >> >>> >>> supported as single-node operations because they're still >> implemented >> >>> >>> via Apache commons.math. >> >>> >>> >> >>> >>> However, there is an experimental script for distributed cholesky >> [1] >> >>> >>> which uses a recursive approach (with operations that allow for >> >>> >>> automatic distributed computation) for matrices larger than a >> >>> >>> user-defined block size. Once blocks become small enough, we use >> >>> >>> again >> >>> >>> the builtin cholesky. Graduating this script would require a >> broader >> >>> >>> set of experiments (and potential improvements) but it simply did >> not >> >>> >>> have the highest priority so far. You might want to give it a try >> >>> >>> though. >> >>> >>> >> >>> >>> Thanks again for your feedback - we'll consider a higher priority >> for >> >>> >>> these distributed operations when discussing the roadmap for the >> next >> >>> >>> releases. >> >>> >>> >> >>> >>> [1] >> >>> >>> >> >>> >>> >> >>> >>> https://urldefense.proofpoint.com/v2/url?u=https-3A__github. >> com_apache_systemml_blob_master_scripts_staging_scalable- >> 5Flinalg_cholesky.dml&d=DwIBaQ&c=jf_iaSHvJObTbx-siA1ZO >> g&r=3mYOfURw_FSirAnoSv2pWvLSi1psso4F9RdGjEWL6yc&m=FvqDr_ >> AKzY5EAD_GAXIJoot0Z09NtMUt8kLShXcJxqQ&s=Yrj4GGcTlpZGRw34RoON_oO6-xDUti >> IEUcO7-qIOyoc&e= >> >>> >>> >> >>> >>> Regards, >> >>> >>> Matthias >> >>> >>> >> >>> >>> On Sat, Apr 21, 2018 at 2:15 PM, Qifan Pu <qifan...@gmail.com> >> wrote: >> >>> >>> > Hi, >> >>> >>> > >> >>> >>> > I would love to do distributed cholesky on large matrix with >> >>> >>> > SystemML. I >> >>> >>> > found two related jiras (SYSTEMML-1213, SYSTEMML-1163), but >> AFAIK, >> >>> >>> > this >> >>> >>> > is >> >>> >>> > currently not implemented? I just wanted to check. >> >>> >>> > >> >>> >>> > Best, >> >>> >>> > Qifan >> >>> >> >> >>> >> >> >> >> >> >> > >> > >> > >> > >> >> >> >> >> >> >