[mpich-discuss] MPI ERROR

Pavan Balaji balaji at mcs.anl.gov
Wed Apr 4 16:59:11 CDT 2012


You seem to be using Open MPI, not MPICH2.  Please send an email to the 
appropriate mailing list.

  -- Pavan

On 04/04/2012 04:56 PM, moussa brahim wrote:
> brahim21322 at hotmail.fr
> mpich-discuss at mcs.anl.gov
> Hello
> I try to run clien/server program (on Ubuntu), and when i run the sever
> program this error appears :
> --------------------------------------------------------------------------
> At least one pair of MPI processes are unable to reach each other for
> MPI communications. This means that no Open MPI device has indicated
> that it can be used to communicate between these processes. This is
> an error; Open MPI requires that all MPI processes be able to reach
> each other. This error can sometimes be the result of forgetting to
> specify the "self" BTL.
>
> Process 1 ([[973,1],0]) is on host: TIGRE
> Process 2 ([[913,1],0]) is on host: TIGRE
> BTLs attempted: self sm tcp
>
> Your MPI job is now going to abort; sorry.
> --------------------------------------------------------------------------
> [TIGRE:2205] *** An error occurred in MPI_Send
> [TIGRE:2205] *** on communicator
> [TIGRE:2205] *** MPI_ERR_INTERN: internal error
> [TIGRE:2205] *** MPI_ERRORS_ARE_FATAL (your MPI job will now abort)
> --------------------------------------------------------------------------
> mpiexec has exited due to process rank 0 with PID 2205 on
> node TIGRE exiting without calling "finalize". This may
> have caused other processes in the application to be
> terminated by signals sent by mpiexec (as reported here).
> --------------------------------------------------------------------------
>
>
>
> *******and this error appears when i run the client program
> --------------------------------------------------------------------------
> At least one pair of MPI processes are unable to reach each other for
> MPI communications. This means that no Open MPI device has indicated
> that it can be used to communicate between these processes. This is
> an error; Open MPI requires that all MPI processes be able to reach
> each other. This error can sometimes be the result of forgetting to
> specify the "self" BTL.
>
> Process 1 ([[913,1],0]) is on host: TIGRE
> Process 2 ([[973,1],0]) is on host: TIGRE
> BTLs attempted: self sm tcp
>
> Your MPI job is now going to abort; sorry.
> --------------------------------------------------------------------------
>
>
>
> ***here we have the server program :
> #include <stdio.h>
>
> #include <mpi.h>
>
>
> main(int argc, char **argv)
> {
>
> int my_id;
>
> char port_name[MPI_MAX_PORT_NAME];
>
> MPI_Comm newcomm;
>
> int passed_num;
>
>
> MPI_Init(&argc, &argv);
>
> MPI_Comm_rank(MPI_COMM_WORLD, &my_id);
>
>
> passed_num = 111;
>
>
> if (my_id == 0)
>
> {
>
> MPI_Open_port(MPI_INFO_NULL, port_name);
>
> printf("%s\n\n", port_name); fflush(stdout);
>
> } /* endif */
>
>
> MPI_Comm_accept(port_name, MPI_INFO_NULL, 0, MPI_COMM_WORLD, &newcomm);
>
> if (my_id == 0)
> {
>
> MPI_Send(&passed_num, 1, MPI_INT, 0, 0, newcomm);
>
> printf("after sending passed_num %d\n", passed_num); fflush(stdout);
>
> MPI_Close_port(port_name);
>
> } /* endif */
>
>
> MPI_Finalize();
>
>
> exit(0);
>
>
> } /* end main() */
>
>
>
> ***and here the client program :
> #include <stdio.h>
>
> #include <mpi.h>
>
>
> main(int argc, char **argv)
>
> {
>
> int passed_num;
>
> int my_id;
>
> MPI_Comm newcomm;
>
>
> MPI_Init(&argc, &argv);
>
> MPI_Comm_rank(MPI_COMM_WORLD, &my_id);
>
>
> MPI_Comm_connect(argv[1], MPI_INFO_NULL, 0, MPI_COMM_WORLD, &newcomm);
>
> if (my_id == 0)
>
> {
>
> MPI_Status status;
>
> MPI_Recv(&passed_num, 1, MPI_INT, 0, 0, newcomm, &status);
>
> printf("after receiving passed_num %d\n", passed_num); fflush(stdout);
>
> } /* endif */
>
>
> MPI_Finalize();
>
>
> exit(0);
>
>
> } /* end main() */
>
> If anybody has sotlution, please help me
> thank you
>
>
> _______________________________________________
> mpich-discuss mailing list     mpich-discuss at mcs.anl.gov
> To manage subscription options or unsubscribe:
> https://lists.mcs.anl.gov/mailman/listinfo/mpich-discuss

-- 
Pavan Balaji
http://www.mcs.anl.gov/~balaji


More information about the mpich-discuss mailing list