Welcome, Guest
Username: Password: Remember me

TOPIC: Telemac parallel

Telemac parallel 5 years 8 months ago #33229

When i try to run in the parallel i am getting error like

C:\opentelemac-mascaret\v7p3>mpiexec
Microsoft MPI Startup Program [Version 10.0.12498.5]

Launches an application on multiple hosts.

Usage:

mpiexec [options] executable [args] [ : [options] exe [args] : ... ]
mpiexec -configfile <file name>

Common options:

-n <num_processes>
-env <env_var_name> <env_var_value>
-wdir <working_directory>
-hosts n host1 [m1] host2 [m2] ... hostn [mn]
-cores <num_cores_per_host>
-lines
-debug [0-3]
-logfile <log file>

Examples:

mpiexec -n 4 pi.exe
mpiexec -hosts 1 server1 master : -n 8 worker

For a complete list of options, run mpiexec -help2
For a list of environment variables, run mpiexec -help3

You can reach the Microsoft MPI team via email at This email address is being protected from spambots. You need JavaScript enabled to view it.

C:\opentelemac-mascaret\v7p3>cd C:\Users\Prof.Murali\Desktop\3D_en

C:\Users\Prof.Murali\Desktop\3D_en>telemac3d.py en.cas


Loading Options and Configurations
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~

_____ ____ __
|___ | |___ \ /_ |
__ __ _/ / _ __ __) | _ __ | |
\ \ / / |_ _| | '_ \ |__ < | '__| | |
\ V / / / | |_) | ___) || | | |
\_/ /_/ | .__/ |____/ |_| |_|
| |
|_|
_ _ __ __ ___ ___ __
_| || |_ /_ |/_ | / _ \ / _ \ /_ |
_ __ ___ __ __ |_ __ _| | | | || (_) || (_) | | |
| '__| / _ \\ \ / / _| || |_ | | | | \__, | \__, | | |
| | | __/ \ V / _ |_ __ _| | | | | / / / / | |
|_| \___| \_/ (_) |_||_| |_| |_| /_/ /_/ |_|


... parsing configuration file: C:\opentelemac-mascaret\v7p3\configs\systel_parallel_v7p3.cfg


Running your CAS file for:
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~

+> configuration: wing64mpi
+> root: C:\opentelemac-mascaret\v7p3


~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~


... reading the main module dictionary

... processing the main CAS file(s)
+> running in English

... handling temporary directories

... checking coupling between codes

... checking parallelisation

... first pass at updating all input files
copying: u_tpxo7.2 C:\Users\Prof.Murali\Desktop\3D_en\en.cas_2019-03-16-15h52min01s\T3DBB2
copying: en.cas C:\Users\Prof.Murali\Desktop\3D_en\en.cas_2019-03-16-15h52min01s\T3DCAS
copying: en.slf C:\Users\Prof.Murali\Desktop\3D_en\en.cas_2019-03-16-15h52min01s\T3DGEO
copying: h_tpxo7.2 C:\Users\Prof.Murali\Desktop\3D_en\en.cas_2019-03-16-15h52min01s\T3DBB1
copying: en.cli C:\Users\Prof.Murali\Desktop\3D_en\en.cas_2019-03-16-15h52min01s\T3DCLI
copying: telemac3d.dico C:\Users\Prof.Murali\Desktop\3D_en\en.cas_2019-03-16-15h52min01s\T3DDICO

... checking the executable
copying: telemac3d.exe C:\Users\Prof.Murali\Desktop\3D_en\en.cas_2019-03-16-15h52min01s\out_telemac3d.exe

... modifying run command to MPI instruction

... modifying run command to PARTEL instruction

... handling sortie file(s)


Running your simulation(s) :
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~



C:\opentelemac-mascaret\mpich2\bin\mpiexec.exe -wdir C:\Users\Prof.Murali\Desktop\3D_en\en.cas_2019-03-16-15h52min01s -n 1 C:\Users\Prof.Murali\Desktop\3D_en\en.cas_2019-03-16-15h52min01s\out_telemac3d.exe


[01:11144]..ERROR:Error while connecting to host, No connection could be made because the target machine actively refused it. (10061)Fatal error in MPI_Init: Other MPI error, error stack:

[01:11144]..ERROR:Connect on sock (host=localhost, port=1) failed, exhaused all end pointsMPIR_Init_thread(392):

SMPDU_Sock_post_connect failed.MPID_Init(107).......: channel initialization failed

[0] PMI_ConnectToHost failed: unable to post a connect to localhost:1, error: Undefined dynamic error codeMPID_Init(371).......: PMI_Init returned -1

uPMI_ConnectToHost returning PMI_FAIL
[0] PMI_Init failed.
_____________
runcode::main:
:
|runCode: Fail to run
|C:\opentelemac-mascaret\mpich2\bin\mpiexec.exe -wdir C:\Users\Prof.Murali\Desktop\3D_en\en.cas_2019-03-16-15h52min01s -n 1 C:\Users\Prof.Murali\Desktop\3D_en\en.cas_2019-03-16-15h52min01s\out_telemac3d.exe
|~~~~~~~~~~~~~~~~~~
|MPID_Init(371).......: PMI_Init returned -1
|~~~~~~~~~~~~~~~~~~
The administrator has disabled public write access.

Telemac parallel 5 years 8 months ago #33241

  • EnGlobe
  • EnGlobe's Avatar
Did you try the command «mpiexec -validate» ? To make sure your mpi is working fine?
The administrator has disabled public write access.

Telemac parallel 5 years 8 months ago #33248

  • cyamin
  • cyamin's Avatar
  • OFFLINE
  • openTELEMAC Guru
  • Posts: 997
  • Thank you received: 234
From MS-MPI documentation:
There are two new options to mpiexec (-pwd and -savecreds) that allow you to provide the necessary credentials for launching processes using the launch service. The -pwd option allows specifying the password for the submitting users in non-interactive scenarios (e.g., using a script or scheduled task). In this case the password is provided in clear text. The -savecreds option causes the provided credentials (if specified with -pwd) to be securely stored on all the hosts specified in the mpiexec command line.

As another example, the following command will authenticate the user using the provided password, store it on the specified machines host1 and host2, then launch three processes, one on host1 and two on host2:
mpiexec –hosts 2 host1 1 host2 2 –wdir C:\MpiTests -pwd <password> -savecreds mpiapp.exe [parameters]

After you run mpiexec with the -savecreds option on a set of nodes, you do not have to provide the password on subsequent runs for those nodes unless the password is changed. If the user is running in interactive mode, mpiexec prompts for the password if the launch service is running and the password has not been provided or previously saved with the -savecreds option.
Costas
The administrator has disabled public write access.

The open TELEMAC-MASCARET template for Joomla!2.5, the HTML 4 version.