²é¿´: 4114  |  »Ø¸´: 1

sic029

Ìú³æ (³õÈëÎÄ̳)

[ÇóÖú] qsubÌá½»²¢ÐÐsiesta²»³É¹¦£¬ÇóÖú

´ó¼ÒºÃ£¬½»Á÷ϼ¯Èº³ÌÐòʹÓÃÓöµ½µÄÎÊÌ⣬¶àл¡£
[node21:10714] *** An error occurred in MPI_Comm_rank
[node21:10714] *** on communicator MPI_COMM_WORLD
[node21:10714] *** MPI_ERR_COMM: invalid communicator
[node21:10714] *** MPI_ERRORS_ARE_FATAL: your MPI job will now abort
--------------------------------------------------------------------------
mpirun has exited due to process rank 3 with PID 10711 on
node node21 exiting improperly. There are two reasons this could occur:

1. this process did not call "init" before exiting, but others in
the job did. This can cause a job to hang indefinitely while it waits
for all processes to call "init". By rule, if one process calls "init",
then ALL processes must call "init" prior to termination.

2. this process called "init", but exited without calling "finalize".
By rule, all processes that call "init" MUST call "finalize" prior to
exiting or it will be considered an "abnormal termination"

This may have caused other processes in the application to be
terminated by signals sent by mpirun (as reported here).
--------------------------------------------------------------------------
[node21:10707] 7 more processes have sent help message help-mpi-errors.txt / mpi_errors_are_fatal
[node21:10707] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages


бàÒëµÄ¼ÆËã³ÌÐòsiesta£¬ÓÃqsub jobÌá½»ÉÏÈ¥ºÜ¿ì½áÊøÌáʾµÄÐÅÏ¢£¬ÄÜ·ñ°ïæÕï¶ÏÒ»ÏÂÇé¿ö¡£ÔÚÁíÍâÒ»¸ö¼¯ÈºÉϱàÒëºóÖ±½ÓÓÃmpirun -np 4 siesta¿ÉÒÔ˳ÀûÖ´Ðе쬲»ÖªµÀΪºÎÔÚм¯ÈºÓÃqsub³öÏÖÕâ¸öÎÊÌ⣬Õâ¸öм¯Èº²»ÈýøÈëµ½×ӽڵ㣬ËùÒÔ±ØÐëÒª½â¾öÕâ¸öÎÊÌâ²ÅÐУ¬¶àлÁË¡£

²»ÖªµÀÊÇÄÄÀïµÄÎÊÌ⣬֮ǰÔڸû·¾³²¢ÐбàÒëµÄlammpsºÍvasp¶¼Ê¹ÓúÜ˳Àû£¬¾ÍÊÇsiestaÓÃqsubÌá½»×÷Òµ×ÜÊÇÎÞ·¨Õý³£¼ÆË㣬µ«ÊDz¢ÐбàÒëµÄsiestaÔÚÁíÍâ»·¾³ÏµÄ×Ó½ÚµãÓÃmpirun -np 4 siestaÖ´ÐкÜ˳Àû£¬¾À½áÁË¡£

Ŷ£¬µÇ¼½ÚµãÉÏmpirunÎÒÊÔ¹ýµÄ£¬Çë°ïæ¿´¿´£¬¸Ð¾õ±»¹ÜÀíÔ±ÉèÖÃÁËÒ²ÎÞ·¨Óãº
mpirun -np 4 siesta
libibverbs: Warning: RLIMIT_MEMLOCK is 32768 bytes.
    This will severely limit memory registrations.
libibverbs: Warning: RLIMIT_MEMLOCK is 32768 bytes.
    This will severely limit memory registrations.
libibverbs: Warning: RLIMIT_MEMLOCK is 32768 bytes.
    This will severely limit memory registrations.
libibverbs: Warning: RLIMIT_MEMLOCK is 32768 bytes.
    This will severely limit memory registrations.
--------------------------------------------------------------------------
The OpenFabrics (openib) BTL failed to initialize while trying to
allocate some locked memory.  This typically can indicate that the
memlock limits are set too low.  For most HPC installations, the
memlock limits should be set to "unlimited".  The failure occured
here:

  Local host:    manage1
  OMPI source:   btl_openib_component.c:1115
  Function:      ompi_free_list_init_ex_new()
  Device:        mlx4_0
  Memlock limit: 32768

You may need to consult with your system administrator to get this
problem fixed.  This FAQ entry on the Open MPI web site may also be
helpful:

    http://www.open-mpi.org/faq/?category=openfabrics#ib-locked-pages
--------------------------------------------------------------------------
--------------------------------------------------------------------------
WARNING: There was an error initializing an OpenFabrics device.

  Local host:   manage1
  Local device: mlx4_0
--------------------------------------------------------------------------
[manage1:16214] *** An error occurred in MPI_Comm_rank
[manage1:16214] *** on communicator MPI_COMM_WORLD
[manage1:16214] *** MPI_ERR_COMM: invalid communicator
[manage1:16214] *** MPI_ERRORS_ARE_FATAL: your MPI job will now abort
--------------------------------------------------------------------------
mpirun has exited due to process rank 0 with PID 16212 on
node manage1 exiting improperly. There are two reasons this could occur:

1. this process did not call "init" before exiting, but others in
the job did. This can cause a job to hang indefinitely while it waits
for all processes to call "init". By rule, if one process calls "init",
then ALL processes must call "init" prior to termination.

2. this process called "init", but exited without calling "finalize".
By rule, all processes that call "init" MUST call "finalize" prior to
exiting or it will be considered an "abnormal termination"

This may have caused other processes in the application to be
terminated by signals sent by mpirun (as reported here).
--------------------------------------------------------------------------
[manage1:16211] 3 more processes have sent help message help-mpi-btl-openib.txt / init-fail-no-mem
[manage1:16211] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages
[manage1:16211] 3 more processes have sent help message help-mpi-btl-openib.txt / error in device init
[manage1:16211] 3 more processes have sent help message help-mpi-errors.txt / mpi_errors_are_fatal

¼ÆËã×Ó½ÚµãÎÞ·¨½øÈ룬±»ÏÞÖÆËÀÁ˵ġ£

Õâ±ßÓõÄpbs×÷Òµ¹ÜÀígridview£¬ÎÒÓõÄÌá½»½Å±¾ÊÇ£º
====================================
#PBS -N test
#PBS -l nodes=1:ppn=8
#PBS -j oe
#PBS -l walltime=24:00:00  
                 
cd $PBS_O_WORKDIR
NP=`cat $PBS_NODEFILE|wc -l`
source /public/software/mpi/openmpi1.5.4-intel.sh                    
mpirun  -machinefile $PBS_NODEFILE -np $NP  \
/home/sw/siesta/siesta-3.1/Obj/siesta < fe.fdf | tee output  
=====================================

¸Ðл³æÓѰïæ£¬¶àл¡£
»Ø¸´´ËÂ¥
research
ÒÑÔÄ   »Ø¸´´ËÂ¥   ¹Ø×¢TA ¸øTA·¢ÏûÏ¢ ËÍTAºì»¨ TAµÄ»ØÌû

redsnowolf

Òø³æ (СÓÐÃûÆø)

¡¾´ð°¸¡¿Ó¦Öú»ØÌû

¡ï
liliangfang: ½ð±Ò+1, лл½»Á÷ 2012-09-15 15:12:35
ÎÒǰÁ½ÌìÓÃvaspÒ²³öÏÖÀàËÆÎÊÌ⣬¸Õ¸Õ½â¾ö¡«

The OpenFabrics (openib) BTL failed to initialize while trying to
allocate some locked memory.  This typically can indicate that the
memlock limits are set too low.  For most HPC installations, the
memlock limits should be set to "unlimited".  The failure occured
here:

  Local host:    node21
  OMPI source:   btl_openib_component.c:1055
  Function:      ompi_free_list_init_ex_new()
  Device:        mlx4_0
  Memlock limit: 65536

You may need to consult with your system administrator to get this
problem fixed.  This FAQ entry on the Open MPI web site may also be
helpful:

    http://www.open-mpi.org/faq/?category=openfabrics#ib-locked-pages
--------------------------------------------------------------------------
--------------------------------------------------------------------------
WARNING: There was an error initializing an OpenFabrics device.
ÉÏÃæÄǸöÍøÖ·Àï15¡¢16¡¢17˵µÄͦÇå³þµÄ£¬ÎÒµÄÇé¿öÊÇÔÚÿ¸ö½Úµãulimit -aÏÔʾlocked memory¶¼Õý³££¬¿É¾ÍÊdzö´í˵ÄÚ´æ·ÖÅä²»Õý³££¬ÄǸöÍøÖ·Àï˵¿ÉÄÜÊǵǼʱûÓÐÕý³£Ö´ÐÐϵͳËùÉèµÄlocked memory£¬»òÕß×÷Òµµ÷¶ÈϵͳûÓзÖÅ䏸ӦÓóÌÐò×ã¹»´óµÄÄڴ棬×îºóÖØÆôÁËÒ»ÏÂÿ¸ö½ÚµãµÄpbsµ÷¶ÈϵͳµÄÊØ»¤½ø³Ì£¬ÎÊÌâ½â¾öÁË¡«
»òÕßÄã¿ÉÒÔÔÚmpirunǰ±ß¶ù¼ÓÉÏulimit -l unlimited£¬ÓÃqsubÌá½»ÏÂÊÔÊÔ
Ï£ÍûÒÔÉÏÐÅÏ¢¶ÔÂ¥Ö÷ÓÐÓá«
2Â¥2012-09-15 14:24:17
ÒÑÔÄ   »Ø¸´´ËÂ¥   ¹Ø×¢TA ¸øTA·¢ÏûÏ¢ ËÍTAºì»¨ TAµÄ»ØÌû
Ïà¹Ø°æ¿éÌø×ª ÎÒÒª¶©ÔÄÂ¥Ö÷ sic029 µÄÖ÷Ìâ¸üÐÂ
×î¾ßÈËÆøÈÈÌûÍÆ¼ö [²é¿´È«²¿] ×÷Õß »Ø/¿´ ×îºó·¢±í
[ÂÛÎÄͶ¸å] ÊÛSCIÒ»ÇøÎÄÕ£¬ÎÒ:8 O5 51O 54,¿ÆÄ¿ÆëÈ«,¿É+¼± +3 w89i99eaeh 2026-02-22 5/250 2026-02-23 08:04 by w4l55oybr1
[²©ºóÖ®¼Ò] ÊÛSCIÒ»ÇøÎÄÕ£¬ÎÒ:8 O5 51O 54,¿ÆÄ¿ÆëÈ«,¿É+¼± +4 khieu8v8m0 2026-02-22 6/300 2026-02-23 07:59 by w4l55oybr1
[¿¼ÑÐ] ÊÛSCIÒ»ÇøÎÄÕ£¬ÎÒ:8 O5 51O 54,¿ÆÄ¿ÆëÈ«,¿É+¼± +3 khieu8v8m0 2026-02-22 7/350 2026-02-23 07:54 by w4l55oybr1
[ÂÛÎÄͶ¸å] ÊÛSCIÒ»ÇøÎÄÕ£¬ÎÒ:8 O5 51O 54,¿ÆÄ¿ÆëÈ«,¿É+¼± +4 khieu8v8m0 2026-02-22 7/350 2026-02-23 07:51 by w4l55oybr1
[²©ºóÖ®¼Ò] ÊÛSCIÒ»ÇøÎÄÕ£¬ÎÒ:8 O5 51O 54,¿ÆÄ¿ÆëÈ«,¿É+¼± +6 3dfhjxgsh7 2026-02-22 9/450 2026-02-23 07:49 by w4l55oybr1
[¿¼²©] ÊÛSCIÒ»ÇøÎÄÕ£¬ÎÒ:8 O5 51O 54,¿ÆÄ¿ÆëÈ«,¿É+¼± +4 khieu8v8m0 2026-02-22 4/200 2026-02-23 06:46 by jsjzfl
[˶²©¼ÒÔ°] ÊÛSCIÒ»ÇøÎÄÕ£¬ÎÒ:8 O5 51O 54,¿ÆÄ¿ÆëÈ«,¿É+¼± +3 8rmuugja8q 2026-02-22 6/300 2026-02-23 06:39 by w4l55oybr1
[˶²©¼ÒÔ°] ÊÛSCIÒ»ÇøÎÄÕ£¬ÎÒ:8 O5 51O 54,¿ÆÄ¿ÆëÈ«,¿É+¼± +4 khieu8v8m0 2026-02-22 8/400 2026-02-23 06:24 by w4l55oybr1
[¿¼ÑÐ] ÊÛSCIÒ»ÇøÎÄÕ£¬ÎÒ:8 O5 51O 54,¿ÆÄ¿ÆëÈ«,¿É+¼± +4 usprnugpzw 2026-02-21 10/500 2026-02-23 04:58 by 5jlh3qtdvx
[¿¼²©] ÊÛSCIÒ»ÇøÎÄÕ£¬ÎÒ:8 O5 51O 54,¿ÆÄ¿ÆëÈ«,¿É+¼± +5 3dfhjxgsh7 2026-02-22 6/300 2026-02-23 02:04 by 5jlh3qtdvx
[»ù½ðÉêÇë] ÃæÉÏ¿ÉÒÔ³¬¹ý30Ò³°É£¿ +4 °¢À­¹±aragon 2026-02-22 4/200 2026-02-22 21:22 by ɽÎ÷Ðü¿ÕË¿ÕÐüÎ
[½Ìʦ֮¼Ò] ΪʲôÖйú´óѧ½ÌÊÚÃÇË®ÁËÄÇô¶àËùνµÄ¶¥»á¶¥¿¯£¬µ«»¹ÊÇ×ö²»³öÓîÊ÷»úÆ÷ÈË£¿ +5 »¶ÀÖËÌÒ¶Ýè 2026-02-21 5/250 2026-02-22 21:15 by ɽÎ÷Ðü¿ÕË¿ÕÐüÎ
[ÂÛÎÄͶ¸å] ÊÛSCIÒ»ÇøÎÄÕ£¬ÎÒ:8 O5 51O 54,¿ÆÄ¿ÆëÈ«,¿É+¼± +4 usprnugpzw 2026-02-21 6/300 2026-02-22 19:48 by w89i99eaeh
[¿¼ÑÐ] ÊÛSCIÒ»ÇøÎÄÕ£¬ÎÒ:8 O5 51O 54,¿ÆÄ¿ÆëÈ«,¿É+¼± +3 3dfhjxgsh7 2026-02-22 4/200 2026-02-22 16:52 by khieu8v8m0
[ÕÒ¹¤×÷] ÊÛSCIÒ»ÇøÎÄÕ£¬ÎÒ:8 O5 51O 54,¿ÆÄ¿ÆëÈ«,¿É+¼± +3 usprnugpzw 2026-02-22 3/150 2026-02-22 16:37 by khieu8v8m0
[¹«Åɳö¹ú] ÊÛSCIÒ»ÇøÎÄÕ£¬ÎÒ:8 O5 51O 54,¿ÆÄ¿ÆëÈ«,¿É+¼± +3 usprnugpzw 2026-02-21 4/200 2026-02-22 16:27 by khieu8v8m0
[»ù½ðÉêÇë] ¡°ÈËÎÄÉç¿Æ¶øÂÛ£¬Ðí¶àѧÊõÑо¿»¹Ã»ÓдﵽÃñ¹úʱÆÚµÄˮƽ¡± +4 ËÕ¶«ÆÂ¶þÊÀ 2026-02-18 5/250 2026-02-22 16:07 by liangep1573
[»ù½ðÉêÇë] ʲôÊÇÈËÒ»Éú×îÖØÒªµÄ£¿ +4 ˲ϢÓîÖæ 2026-02-21 4/200 2026-02-22 11:44 by huagongfeihu
[»ù½ðÉêÇë] ½ñÄê´ºÍíÓм¸¸ö½ÚÄ¿ºÜ²»´í£¬µãÔÞ£¡ +11 ˲ϢÓîÖæ 2026-02-16 12/600 2026-02-21 21:14 by lq493392203
[»ù½ðÉêÇë] ÌåÖÆÄÚ³¤±²ËµÌåÖÆÄÚ¾ø´ó²¿·ÖÒ»±²×ÓÔڵײ㣬ÈçͬÄãÃÇÒ»Ñù´ó²¿·ÖÆÕͨ½ÌʦæÇÒÊÕÈëµÍ +9 ˲ϢÓîÖæ 2026-02-20 12/600 2026-02-21 10:39 by »¶ÀÖËÌÒ¶Ýè
ÐÅÏ¢Ìáʾ
ÇëÌî´¦ÀíÒâ¼û