Hi guys,

I finally reconsidered parallelization of 
Law2_ScGeom_CapillaryPhys_Capillarity, after our passed discussion (*).

I noticed only after re-starting this work that implementations of e.g. 
remove() and resize() in in-house class OpenMPVector would be necessary to do 
it completely right (using OpenMPVector objects in BodiesMenisciiList class). 
Then, I only coded yet parallel loops for the "addForce" block, at the end of 
the Law2....

Obviously, the results (in terms of capillary stresses as given by 
getCapillaryStress() function) are unaffected by the changes : see the four 
attached pictures, showing sXXcap or sYYcap for two triaxial loadings with 
20000 particles, repeted before and after changing code. One loading is with 10 
kPa capillary pressure : meniscii fusion occur in the sample, affecting the 
simulation (with fusionDetection = 1,binaryFusion=0 attributes of the Law2). 
For the other one, the suction value of 300 kPa prevents any meniscii fusion.

As for the speed up, *maybe* such changes in the code allow a significant speed 
up (of capillary Law2) to be obtained with parallel computations... See the 
data sheet "NBl2cp" (in ods or pdf format).
Anyway, I hope it is clear this little change in code does not harm. Since it 
might still be little useful, and exists already on my computer, I plann to 
commit it, after time for discussion.

Jerome

PS : these results might still be of interest for you if you are curious of 
results scattering through parallel computations, or capillary law2 and 
NewtonIntegrator (NI) timings... I obtain here e.g. a speed up of NI about 2.3 
from j1 to j4


(*) https://www.mail-archive.com/yade-dev@lists.launchpad.net/msg10842.html)

Attachment: NBl2cp.ods
Description: NBl2cp.ods

Attachment: NBl2cp.pdf
Description: NBl2cp.pdf

_______________________________________________
Mailing list: https://launchpad.net/~yade-dev
Post to     : yade-dev@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yade-dev
More help   : https://help.launchpad.net/ListHelp

Reply via email to