Dear All, In the past (LSF) it was possible to submit parallel jobs on several cores for ABAQUS. However, now with SLURM i encounter the problem stated below: Abaqus Error: The number of cpus used is more than the total number of processors specified by the host list. mp_host_list: [['ncm0081.hpc.itc.rwth-aachen.de', 1]]. Abaqus/Analysis exited with error(s). I suspect it is due to a fautly setting of the enviromental variable $R_WLM_ABAQUSHOSTLIST, as the job submission script, found on the RWTH documentation website states: cat << EOF > $abaqus_v6.env #verbose = 3 #ask_delete = OFF mp_file_system = (SHARED, LOCAL) mp_host_list = $R_WLM_ABAQUSHOSTLIST EOF It seems like currently there is only one host in that list, which i suspect to be insufficient when running on, let say 16 cores. Best Regards, Karthik
Dear All, I am getting problem with SLURM job submission. like this: job_submission_slurmupdated.sh: line 44: module: command not found job_submission_slurmupdated.sh: line 45: module: command not found These lines have two modules: module load CHEMISTRY module load gromacs/2019 These are already sourced in .zshrc file and GROMACS is also working. This is the beginning of script also: #!/usr/local_rwth/bin/zsh #SBATCH -p c18m If anybody has any clue/idea to overcome this, please suggest me. Thanks in advance. Best regards, Subrata. On Wed, Aug 28, 2019 at 1:50 PM Krishna Murthy, Karthik Ravi < ravi.krishna@rwth-aachen.de> wrote:
Dear All, In the past (LSF) it was possible to submit parallel jobs on several cores for ABAQUS. However, now with SLURM i encounter the problem stated below: Abaqus Error: The number of cpus used is more than the total number of processors specified by the host list. mp_host_list: [[' ncm0081.hpc.itc.rwth-aachen.de', 1]]. Abaqus/Analysis exited with error(s). I suspect it is due to a fautly setting of the enviromental variable $R_WLM_ABAQUSHOSTLIST, as the job submission script, found on the RWTH documentation website states: cat << EOF > $abaqus_v6.env #verbose = 3 #ask_delete = OFF mp_file_system = (SHARED, LOCAL) mp_host_list = $R_WLM_ABAQUSHOSTLIST EOF It seems like currently there is only one host in that list, which i suspect to be insufficient when running on, let say 16 cores. Best Regards, Karthik _______________________________________________ claix18-slurm-pilot mailing list -- claix18-slurm-pilot@lists.rwth-aachen.de To unsubscribe send an email to claix18-slurm-pilot-leave@lists.rwth-aachen.de
Could you please attach your batch script used *as attachment file*? On 10/4/19 3:21 PM, Subrata Pramanik wrote:
Dear All,
I am getting problem with SLURM job submission. like this:
job_submission_slurmupdated.sh: line 44: module: command not found job_submission_slurmupdated.sh: line 45: module: command not found
These lines have two modules:
module load CHEMISTRY module load gromacs/2019
These are already sourced in .zshrc file and GROMACS is also working.
This is the beginning of script also:
#!/usr/local_rwth/bin/zsh
#SBATCH -p c18m
If anybody has any clue/idea to overcome this, please suggest me.
Thanks in advance. Best regards, Subrata.
On Wed, Aug 28, 2019 at 1:50 PM Krishna Murthy, Karthik Ravi <ravi.krishna@rwth-aachen.de <mailto:ravi.krishna@rwth-aachen.de>> wrote:
Dear All, In the past (LSF) it was possible to submit parallel jobs on several cores for ABAQUS. However, now with SLURM i encounter the problem stated below: Abaqus Error: The number of cpus used is more than the total number of processors specified by the host list. mp_host_list: [['ncm0081.hpc.itc.rwth-aachen.de <http://ncm0081.hpc.itc.rwth-aachen.de>', 1]]. Abaqus/Analysis exited with error(s). I suspect it is due to a fautly setting of the enviromental variable $R_WLM_ABAQUSHOSTLIST, as the job submission script, found on the RWTH documentation website states: cat << EOF > $abaqus_v6.env #verbose = 3 #ask_delete = OFF mp_file_system = (SHARED, LOCAL) mp_host_list = $R_WLM_ABAQUSHOSTLIST EOF It seems like currently there is only one host in that list, which i suspect to be insufficient when running on, let say 16 cores. Best Regards, Karthik
_______________________________________________ claix18-slurm-pilot mailing list -- claix18-slurm-pilot@lists.rwth-aachen.de <mailto:claix18-slurm-pilot@lists.rwth-aachen.de> To unsubscribe send an email to claix18-slurm-pilot-leave@lists.rwth-aachen.de <mailto:claix18-slurm-pilot-leave@lists.rwth-aachen.de>
_______________________________________________ claix18-slurm-pilot mailing list -- claix18-slurm-pilot@lists.rwth-aachen.de To unsubscribe send an email to claix18-slurm-pilot-leave@lists.rwth-aachen.de
-- Dipl.-Inform. Paul Kapinos - High Performance Computing, RWTH Aachen University, IT Center Seffenter Weg 23, D 52074 Aachen (Germany) Tel: +49 241/80-24915
Dear Subrata, since you have already filed a support ticket for this particular issue, please follow Paul's request and attach your current job script to that ticket. This mailing list was created to help people with issues in adopting the Claix18 environment, it is no replacement for our general support channels via email or our ticket tool website. I would not want to upset people by flooding their inboxes with unrelated support correspondence. Best, Sven On 10/4/19 3:21 PM, Subrata Pramanik wrote:
Dear All,
I am getting problem with SLURM job submission. like this:
job_submission_slurmupdated.sh: line 44: module: command not found job_submission_slurmupdated.sh: line 45: module: command not found
These lines have two modules:
module load CHEMISTRY module load gromacs/2019
These are already sourced in .zshrc file and GROMACS is also working.
This is the beginning of script also:
#!/usr/local_rwth/bin/zsh
#SBATCH -p c18m
If anybody has any clue/idea to overcome this, please suggest me.
Thanks in advance. Best regards, Subrata.
-- Sven Hansen IT Center Abteilung: Systeme und Betrieb RWTH Aachen University Seffenter Weg 23 52074 Aachen (Germany) Tel.: + 49 241 80-29114 s.hansen@itc.rwth-aachen.de www.itc.rwth-aachen.de
participants (4)
-
Krishna Murthy, Karthik Ravi
-
Paul Kapinos
-
Subrata Pramanik
-
Sven Hansen