Hello there, I am running a simple binary black hole simulation from the parameter file kept in the Cactus/par/arXiv-1111.3344 where I simulated high resolution. Furthermore, I assume simulation for lower resolution is less costly than higher, but I wonder it keeps giving me errors in the low-resolution run. (kernel dies for the MedResolution). I tried almost all the ways of adjusting the par file according to my local machine strength, but I failed. The error is below along with the par file, *.err, and *.out filesgenerators attached. Note: There is no storage issue at all.
Any help would be highly appreciated. Warning: Too many threads per process specified: specified num-threads=2 (ppn-used is 2) Warning: Total number of threads and number of threads per process are inconsistent: procs=1, num-threads=2 (procs*num-smt must be an integer multiple of num-threads) Warning: Total number of threads and number of cores per node are inconsistent: procs=1, ppn-used=2 (procs must be an integer multiple of ppn-used) + set -e + cd /home/karima/simulations/bbhL/output-0000-active + echo Checking: + pwd + hostname + date + echo Environment: + export CACTUS_NUM_PROCS=1 + export CACTUS_NUM_THREADS=2 + export GMON_OUT_PREFIX=gmon.out + export OMP_NUM_THREADS=2 + env + sort + echo Starting: + date +%s + export CACTUS_STARTTIME=1608653732 + [ 1 = 1 ] + [ 0 -eq 0 ] + /home/karima/simulations/bbhL/SIMFACTORY/exe/cactus_sim -L 3 /home/karima/simulations/bbhL/output-0000/BBHLowRes.par WARNING level 0 from host karima-Latitude-E5470 process 0 in thorn cactus, file BBHLowRes.par:1: -> ERROR IN PARAMETER FILE:Parse Error Expected one of the following characters: ':' CarpetMask::excluded_surface_factor[1] = 1.0 CarpetMask::excluded_surface [2] = 2 CarpetMask::excluded_surface_factor[2] = 1.0 Contents successfully written to /home/karima/Cactus/repos/simfactory2/etc/defs.local.ini ^ -------------------------------------------------------------------------- MPI_ABORT was invoked on rank 0 in communicator MPI_COMM_WORLD with errorcode 1. NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes. You may or may not see output from other processes, depending on exactly when Open MPI kills them. Thanks in advance best, Karima S
BBHLowRes.par
Description: Binary data
bbhL.err
Description: Binary data
bbhL.out
Description: Binary data
_______________________________________________ Users mailing list Users@einsteintoolkit.org http://lists.einsteintoolkit.org/mailman/listinfo/users