Dear Moe

Ok, I see. I have to explain in detail. We have a user on our cluster,
who wants to run for each MPI rank a special wrapper script, which setup
things on a per MPI rank basis. He has used this technique with an other
MPI implementation in combination with mpirun for some time.

The sbatch script is run by: "sbatch sbatch_script.sh"

My sbatch test script "sbatch_script.sh":
#!/bin/sh
#SBATCH --ntasks=12
# ... other sbatch options
srun -n 12 wrapper_script.sh


My "wrapper_script.sh":
#!/bin/sh
# ... some stuff the user wants to run
my_mpi_binary

So, sbatch run the sbatch_script, for each

If I start "my_mpi_binary" directly via the srun command I have no
problems. Anything is fine. But if I use the wrapper_script I get
defunct processes. Is this a bug, or is this beyond the MPI specification?

Best regards,
Tibor

Am 07.11.2011 17:12, schrieb Moe Jette:
> I suspect that he really wants to do something this:
>
> sbatch -n12 ../wrapper_skript.sh
>   OR
> salloc -n12 ../wrapper_skript.sh
>
> wrapper_skript.sh:
> #setup
> srun ./my_mpi_app
> #cleanup
>
> Quoting Tibor Pausz <[email protected]>:
>
>> Dear all,
>>
>> i want to start a programm via a wrapper script. If I try this, MPI does
>> not work correct. Is their any solution to this problem?
>>
>> example:
>> srun -n 12 -l ../wrapper_skript.sh
>>
>> wrapper_skript.sh:
>> #setup
>> ./my_mpi_app
>> #cleanup
>>
>> Best regards,
>> Tibor
>>
>
>
>

Reply via email to