Hi David, Just ran a test over here on Red Hat 6. The case runs, but memory consumption seems to be in the order of 5GB. Are you sure your VM has enough memory?
Cheers, Alf From: Opm <[email protected]> On Behalf Of David Baxendale (Private) Sent: 27. september 2018 02:56 To: [email protected] Subject: Re: [Opm] Opm Digest, Vol 33, Issue 5 Markus, I tried running this deck in sequential mode in an Ubuntu VM and it just fails: david@EIPC02-VirtualBox:/media/sf_Linux/OPM-Flow/spe10model2$ flow SPE10_MODEL2.DATA ********************************************************************** * * * This is flow 2018.04 * * * * Flow is a simulator for fully implicit three-phase black-oil flow, * * including solvent and polymer capabilities. * * For more information, see https://opm-project.org * * * ********************************************************************** Killed david@EIPC02-VirtualBox:/media/sf_Linux/OPM-Flow/spe10model2$ I ran under strace and got david@EIPC02-VirtualBox:/media/sf_Linux/OPM-Flow/spe10model2$ tail SPE_MODEL2A.LOG mprotect(0x7f48f54a4000, 4096, PROT_READ|PROT_WRITE) = 0 mprotect(0x7f48f54a5000, 4096, PROT_READ|PROT_WRITE) = 0 mprotect(0x7f48f54a6000, 4096, PROT_READ|PROT_WRITE) = 0 mprotect(0x7f48f54a7000, 4096, PROT_READ|PROT_WRITE) = 0 mprotect(0x7f48f54a8000, 4096, PROT_READ|PROT_WRITE) = 0 mprotect(0x7f48f54a9000, 4096, PROT_READ|PROT_WRITE) = 0 mprotect(0x7f48f54aa000, 4096, PROT_READ|PROT_WRITE) = 0 mprotect(0x7f48f54ab000, 4096, PROT_READ|PROT_WRITE) = 0 mprotect(0x7f48f54ac000, 4096, PROT_READ|PROT_WRITE <unfinished ...> +++ exited with 65 +++ david@EIPC02-VirtualBox:/media/sf_Linux/OPM-Flow/spe10model2$ So something is amiss here. Regards, OPMUSER ________________________________ On 26-Sep-18 20:00, [email protected]<mailto:[email protected]> wrote: Send Opm mailing list submissions to [email protected]<mailto:[email protected]> To subscribe or unsubscribe via the World Wide Web, visit https://opm-project.org/cgi-bin/mailman/listinfo/opm or, via email, send a message with subject or body 'help' to [email protected]<mailto:[email protected]> You can reach the person managing the list at [email protected]<mailto:[email protected]> When replying, please edit your Subject line so it is more specific than "Re: Contents of Opm digest..." Today's Topics: 1. MPI Parallel OPM Flow - SPE 10 Model 2 Hangs (sindimo) 2. Re: MPI Parallel OPM Flow - SPE 10 Model 2 Hangs (Markus Blatt) ---------------------------------------------------------------------- Message: 1 Date: Tue, 25 Sep 2018 21:30:26 -0400 From: sindimo <[email protected]><mailto:[email protected]> To: [email protected]<mailto:[email protected]> Subject: [Opm] MPI Parallel OPM Flow - SPE 10 Model 2 Hangs Message-ID: <cacoqo+knjeeh_5of91c5uabkzbx4-hsf3boux+kxodggsos...@mail.gmail.com><mailto:cacoqo+knjeeh_5of91c5uabkzbx4-hsf3boux+kxodggsos...@mail.gmail.com> Content-Type: text/plain; charset="utf-8" Hi, I am running OPM Flow with MPICH MPI on RedHat 7 (installed via yum through OPM repo, version 2018.04). I am able to successfully run in parallel some of the test models (norne, spe5, spe9), however spe10model2 always hangs when I try to launch it. It seems it hangs during the cell partitioning as below. It just partially does the partitioning on a subset of the processes only (example below where I have 4 processes and it hangs after doing 2 partitions, I've also tried with 8 processors and it shows similar behavior). Any help with this is much appreciated as I need to run SPE10 for some work I am doing, many thanks! Sincerely, Mohamad mpirun -np 4 /usr/lib64/mpich/bin/flow SPE10_MODEL2.DATA output_dir=out_parallel ********************************************************************** * * * This is flow 2018.04 * * * * Flow is a simulator for fully implicit three-phase black-oil flow, * * including solvent and polymer capabilities. * * For more information, see https://opm-project.org * * * ********************************************************************** After loadbalancing process 0 has 322630 cells. After loadbalancing process 3 has 340338 cells. -------------- next part -------------- An HTML attachment was scrubbed... URL: <//opm-project.org/pipermail/opm/attachments/20180925/2aa17dc1/attachment-0001.html> ------------------------------ Message: 2 Date: Wed, 26 Sep 2018 10:10:26 +0200 From: Markus Blatt <[email protected]><mailto:[email protected]> To: [email protected]<mailto:[email protected]> Subject: Re: [Opm] MPI Parallel OPM Flow - SPE 10 Model 2 Hangs Message-ID: <20180926081026.GB3152@smaug<mailto:20180926081026.GB3152@smaug>> Content-Type: text/plain; charset=iso-8859-1 Hi On Tue, Sep 25, 2018 at 09:30:26PM -0400, sindimo wrote: I am running OPM Flow with MPICH MPI on RedHat 7 (installed via yum through OPM repo, version 2018.04). I am able to successfully run in parallel some of the test models (norne, spe5, spe9), however spe10model2 always hangs when I try to launch it. It seems it hangs during the cell partitioning as below. It just partially does the partitioning on a subset of the processes only (example below where I have 4 processes and it hangs after doing 2 partitions, I've also tried with 8 processors and it shows similar behavior). Any help with this is much appreciated as I need to run SPE10 for some work I am doing, many thanks! May I ask what work that is? So you are using a release (even of a target distribution). That is a bit weired. Unfortunately I do not have access to such a system and cannot be of much help here. Maybe somebody else can do a quick test? But I did a quick test with the current master on my system and it works with 4 processes. So if nobody else can help you, then you might want to checkout master and compile OPM yourself. Cheers, Markus [https://ipmcdn.avast.com/images/icons/icon-envelope-tick-round-orange-animated-no-repeat-v1.gif]<https://www.avast.com/sig-email?utm_medium=email&utm_source=link&utm_campaign=sig-email&utm_content=emailclient> Virus-free. www.avast.com<https://www.avast.com/sig-email?utm_medium=email&utm_source=link&utm_campaign=sig-email&utm_content=emailclient> ------------------------------------------------------------------- The information contained in this message may be CONFIDENTIAL and is intended for the addressee only. Any unauthorized use, dissemination of the information or copying of this message is prohibited. If you are not the addressee, please notify the sender immediately by return e-mail and delete this message. Thank you
_______________________________________________ Opm mailing list [email protected] https://opm-project.org/cgi-bin/mailman/listinfo/opm
