| ²é¿´: 4115 | »Ø¸´: 1 | |||
sic029Ìú³æ (³õÈëÎÄ̳)
|
[ÇóÖú]
qsubÌá½»²¢ÐÐsiesta²»³É¹¦£¬ÇóÖú
|
|
´ó¼ÒºÃ£¬½»Á÷ϼ¯Èº³ÌÐòʹÓÃÓöµ½µÄÎÊÌ⣬¶àл¡£ [node21:10714] *** An error occurred in MPI_Comm_rank [node21:10714] *** on communicator MPI_COMM_WORLD [node21:10714] *** MPI_ERR_COMM: invalid communicator [node21:10714] *** MPI_ERRORS_ARE_FATAL: your MPI job will now abort -------------------------------------------------------------------------- mpirun has exited due to process rank 3 with PID 10711 on node node21 exiting improperly. There are two reasons this could occur: 1. this process did not call "init" before exiting, but others in the job did. This can cause a job to hang indefinitely while it waits for all processes to call "init". By rule, if one process calls "init", then ALL processes must call "init" prior to termination. 2. this process called "init", but exited without calling "finalize". By rule, all processes that call "init" MUST call "finalize" prior to exiting or it will be considered an "abnormal termination" This may have caused other processes in the application to be terminated by signals sent by mpirun (as reported here). -------------------------------------------------------------------------- [node21:10707] 7 more processes have sent help message help-mpi-errors.txt / mpi_errors_are_fatal [node21:10707] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages бàÒëµÄ¼ÆËã³ÌÐòsiesta£¬ÓÃqsub jobÌá½»ÉÏÈ¥ºÜ¿ì½áÊøÌáʾµÄÐÅÏ¢£¬ÄÜ·ñ°ïæÕï¶ÏÒ»ÏÂÇé¿ö¡£ÔÚÁíÍâÒ»¸ö¼¯ÈºÉϱàÒëºóÖ±½ÓÓÃmpirun -np 4 siesta¿ÉÒÔ˳ÀûÖ´Ðе쬲»ÖªµÀΪºÎÔÚм¯ÈºÓÃqsub³öÏÖÕâ¸öÎÊÌ⣬Õâ¸öм¯Èº²»ÈýøÈëµ½×ӽڵ㣬ËùÒÔ±ØÐëÒª½â¾öÕâ¸öÎÊÌâ²ÅÐУ¬¶àлÁË¡£ ²»ÖªµÀÊÇÄÄÀïµÄÎÊÌ⣬֮ǰÔڸû·¾³²¢ÐбàÒëµÄlammpsºÍvasp¶¼Ê¹ÓúÜ˳Àû£¬¾ÍÊÇsiestaÓÃqsubÌá½»×÷Òµ×ÜÊÇÎÞ·¨Õý³£¼ÆË㣬µ«ÊDz¢ÐбàÒëµÄsiestaÔÚÁíÍâ»·¾³ÏµÄ×Ó½ÚµãÓÃmpirun -np 4 siestaÖ´ÐкÜ˳Àû£¬¾À½áÁË¡£ Ŷ£¬µÇ¼½ÚµãÉÏmpirunÎÒÊÔ¹ýµÄ£¬Çë°ïæ¿´¿´£¬¸Ð¾õ±»¹ÜÀíÔ±ÉèÖÃÁËÒ²ÎÞ·¨Ó㺠mpirun -np 4 siesta libibverbs: Warning: RLIMIT_MEMLOCK is 32768 bytes. This will severely limit memory registrations. libibverbs: Warning: RLIMIT_MEMLOCK is 32768 bytes. This will severely limit memory registrations. libibverbs: Warning: RLIMIT_MEMLOCK is 32768 bytes. This will severely limit memory registrations. libibverbs: Warning: RLIMIT_MEMLOCK is 32768 bytes. This will severely limit memory registrations. -------------------------------------------------------------------------- The OpenFabrics (openib) BTL failed to initialize while trying to allocate some locked memory. This typically can indicate that the memlock limits are set too low. For most HPC installations, the memlock limits should be set to "unlimited". The failure occured here: Local host: manage1 OMPI source: btl_openib_component.c:1115 Function: ompi_free_list_init_ex_new() Device: mlx4_0 Memlock limit: 32768 You may need to consult with your system administrator to get this problem fixed. This FAQ entry on the Open MPI web site may also be helpful: http://www.open-mpi.org/faq/?category=openfabrics#ib-locked-pages -------------------------------------------------------------------------- -------------------------------------------------------------------------- WARNING: There was an error initializing an OpenFabrics device. Local host: manage1 Local device: mlx4_0 -------------------------------------------------------------------------- [manage1:16214] *** An error occurred in MPI_Comm_rank [manage1:16214] *** on communicator MPI_COMM_WORLD [manage1:16214] *** MPI_ERR_COMM: invalid communicator [manage1:16214] *** MPI_ERRORS_ARE_FATAL: your MPI job will now abort -------------------------------------------------------------------------- mpirun has exited due to process rank 0 with PID 16212 on node manage1 exiting improperly. There are two reasons this could occur: 1. this process did not call "init" before exiting, but others in the job did. This can cause a job to hang indefinitely while it waits for all processes to call "init". By rule, if one process calls "init", then ALL processes must call "init" prior to termination. 2. this process called "init", but exited without calling "finalize". By rule, all processes that call "init" MUST call "finalize" prior to exiting or it will be considered an "abnormal termination" This may have caused other processes in the application to be terminated by signals sent by mpirun (as reported here). -------------------------------------------------------------------------- [manage1:16211] 3 more processes have sent help message help-mpi-btl-openib.txt / init-fail-no-mem [manage1:16211] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages [manage1:16211] 3 more processes have sent help message help-mpi-btl-openib.txt / error in device init [manage1:16211] 3 more processes have sent help message help-mpi-errors.txt / mpi_errors_are_fatal ¼ÆËã×Ó½ÚµãÎÞ·¨½øÈ룬±»ÏÞÖÆËÀÁ˵ġ£ Õâ±ßÓõÄpbs×÷Òµ¹ÜÀígridview£¬ÎÒÓõÄÌá½»½Å±¾ÊÇ£º ==================================== #PBS -N test #PBS -l nodes=1:ppn=8 #PBS -j oe #PBS -l walltime=24:00:00 cd $PBS_O_WORKDIR NP=`cat $PBS_NODEFILE|wc -l` source /public/software/mpi/openmpi1.5.4-intel.sh mpirun -machinefile $PBS_NODEFILE -np $NP \ /home/sw/siesta/siesta-3.1/Obj/siesta < fe.fdf | tee output ===================================== ¸Ðл³æÓѰïæ£¬¶àл¡£ |
» ²ÂÄãϲ»¶
ÊÛSCIÒ»ÇøÎÄÕ£¬ÎÒ:8 O5 51O 54,¿ÆÄ¿ÆëÈ«,¿É+¼±
ÒѾÓÐ8È˻ظ´
ÌåÖÆÄÚ³¤±²ËµÌåÖÆÄÚ¾ø´ó²¿·ÖÒ»±²×ÓÔڵײ㣬ÈçͬÄãÃÇÒ»Ñù´ó²¿·ÖÆÕͨ½ÌʦæÇÒÊÕÈëµÍ
ÒѾÓÐ13È˻ظ´
ÊÛSCIÒ»ÇøÎÄÕ£¬ÎÒ:8 O5 51O 54,¿ÆÄ¿ÆëÈ«,¿É+¼±
ÒѾÓÐ8È˻ظ´
ÊÛSCIÒ»ÇøÎÄÕ£¬ÎÒ:8 O5 51O 54,¿ÆÄ¿ÆëÈ«,¿É+¼±
ÒѾÓÐ8È˻ظ´
ÊÛSCIÒ»ÇøÎÄÕ£¬ÎÒ:8 O5 51O 54,¿ÆÄ¿ÆëÈ«,¿É+¼±
ÒѾÓÐ11È˻ظ´
ΪʲôÖйú´óѧ¹¤¿Æ½ÌÊÚÃÇË®ÁËÄÇô¶àËùνµÄ¶¥»á¶¥¿¯£¬µ«»¹ÊÇ×ö²»³öÓîÊ÷»úÆ÷ÈË£¿
ÒѾÓÐ8È˻ظ´
ÊÛSCIÒ»ÇøÎÄÕ£¬ÎÒ:8 O5 51O 54,¿ÆÄ¿ÆëÈ«,¿É+¼±
ÒѾÓÐ5È˻ظ´
ÊÛSCIÒ»ÇøÎÄÕ£¬ÎÒ:8 O5 51O 54,¿ÆÄ¿ÆëÈ«,¿É+¼±
ÒѾÓÐ6È˻ظ´
ÊÛSCIÒ»ÇøÎÄÕ£¬ÎÒ:8 O5 51O 54,¿ÆÄ¿ÆëÈ«,¿É+¼±
ÒѾÓÐ9È˻ظ´
ÊÛSCIÒ»ÇøÎÄÕ£¬ÎÒ:8 O5 51O 54,¿ÆÄ¿ÆëÈ«,¿É+¼±
ÒѾÓÐ4È˻ظ´
» ±¾Ö÷ÌâÏà¹Ø¼ÛÖµÌùÍÆ¼ö£¬¶ÔÄúͬÑùÓаïÖú:
¡¾ÇóÖú¡¿siesta °²×°³ö´í
ÒѾÓÐ18È˻ظ´
¡¾ÇóÖú¡¿ÇóÖúsiestaµÄOrdern·½·¨²»ÄÜÊÕÁ²µÄÎÊÌâ
ÒѾÓÐ4È˻ظ´
¡¾ÇóÖú¡¿siesta3.0²¢Ðгɹ¦°²×°ºó²»Äܲ¢ÐеÄÎÊÌâ
ÒѾÓÐ15È˻ظ´

redsnowolf
Òø³æ (СÓÐÃûÆø)
- Ó¦Öú: 7 (Ó×¶ùÔ°)
- ½ð±Ò: 416.5
- É¢½ð: 36
- ºì»¨: 5
- Ìû×Ó: 213
- ÔÚÏß: 689.8Сʱ
- ³æºÅ: 1332218
- ×¢²á: 2011-06-27
- ÐÔ±ð: GG
- רҵ: °ëµ¼Ìå²ÄÁÏ
¡¾´ð°¸¡¿Ó¦Öú»ØÌû
¡ï
liliangfang: ½ð±Ò+1, лл½»Á÷ 2012-09-15 15:12:35
liliangfang: ½ð±Ò+1, лл½»Á÷ 2012-09-15 15:12:35
|
ÎÒǰÁ½ÌìÓÃvaspÒ²³öÏÖÀàËÆÎÊÌ⣬¸Õ¸Õ½â¾ö¡« The OpenFabrics (openib) BTL failed to initialize while trying to allocate some locked memory. This typically can indicate that the memlock limits are set too low. For most HPC installations, the memlock limits should be set to "unlimited". The failure occured here: Local host: node21 OMPI source: btl_openib_component.c:1055 Function: ompi_free_list_init_ex_new() Device: mlx4_0 Memlock limit: 65536 You may need to consult with your system administrator to get this problem fixed. This FAQ entry on the Open MPI web site may also be helpful: http://www.open-mpi.org/faq/?category=openfabrics#ib-locked-pages -------------------------------------------------------------------------- -------------------------------------------------------------------------- WARNING: There was an error initializing an OpenFabrics device. ÉÏÃæÄǸöÍøÖ·Àï15¡¢16¡¢17˵µÄͦÇå³þµÄ£¬ÎÒµÄÇé¿öÊÇÔÚÿ¸ö½Úµãulimit -aÏÔʾlocked memory¶¼Õý³££¬¿É¾ÍÊdzö´í˵ÄÚ´æ·ÖÅä²»Õý³££¬ÄǸöÍøÖ·Àï˵¿ÉÄÜÊǵǼʱûÓÐÕý³£Ö´ÐÐϵͳËùÉèµÄlocked memory£¬»òÕß×÷Òµµ÷¶ÈϵͳûÓзÖÅ䏸ӦÓóÌÐò×ã¹»´óµÄÄڴ棬×îºóÖØÆôÁËÒ»ÏÂÿ¸ö½ÚµãµÄpbsµ÷¶ÈϵͳµÄÊØ»¤½ø³Ì£¬ÎÊÌâ½â¾öÁË¡« »òÕßÄã¿ÉÒÔÔÚmpirunǰ±ß¶ù¼ÓÉÏulimit -l unlimited£¬ÓÃqsubÌá½»ÏÂÊÔÊÔ Ï£ÍûÒÔÉÏÐÅÏ¢¶ÔÂ¥Ö÷ÓÐÓá« |
2Â¥2012-09-15 14:24:17













»Ø¸´´ËÂ¥