[mpich-discuss] Facing problem while using MPI_File_set_view
Christina Patrick
christina.subscribes at gmail.com
Wed May 27 14:23:48 CDT 2009
I will try it with the latest releace rc1 and let you know.
Thanks,
Christina.
On Wed, May 27, 2009 at 3:04 PM, Christina Patrick
<christina.subscribes at gmail.com> wrote:
> I have a feeling that there is a bug in the MPI code when we use
> MPI_Type_create_subarray() to create row file view as specified below.
> If you change the buffer size below in my program to something smaller
> than the entire view, you will get an error. I checked the use of
> MPI_Type_create_subarray() API for creating column and block file
> views. They work fine.
>
> If somebody could please take a look at this problem, it would be
> really helpful.
>
> Thanks and Regards,
> Christina.
>
> On Mon, May 25, 2009 at 5:19 PM, Christina Patrick
> <christina.subscribes at gmail.com> wrote:
>> Hi Everybody,
>>
>> I am writing a program to read an array of 16384 x 16384 of type
>> double ~ 2GB file size
>> I am using the file view as follows:
>>
>> ------------------------------
>> | | P0
>> ------------------------------
>> | | P1
>> ------------------------------
>> | | P2
>> ------------------------------
>> | | P3
>> ------------------------------
>>
>> I create the file view using MPI_Type_create_subarray() and use
>> collective I/O MPI_File_read_all()
>>
>> When I read the entire file view of P0 (P1, P2, P3) into a buffer
>> (512MB) (in a single instance), the program works fine.
>> However, I do not want to use a buffer as big as 512 MB. So I use a
>> smaller buffer (such as 8MB) and iterate over the file view of P0 (P1,
>> P2, P3), my program starts throwing errors/get segmentation faults:
>> [E 16:49:28.377861] Error: payload_progress: Bad address
>> OR
>> segmentation fault in ADIOI_Calc_my_off_len() at line: while
>> (flat_file->type != fd->filetype) flat_file = flat_file->next;
>> because flat_file becomes 0x0 in one of the processes.
>>
>> If I use a column view and logically do the same thing, I do not face
>> this problem. (I hope that I have been able to explain the problem. In
>> case of doubt, please let me know.)
>>
>> Could somebody please help me,
>>
>> Thanks and Regards,
>> Christina.
>>
>> PS: I am pasting the program below:
>>
>> #include "mpi.h"
>> #include <stdio.h>
>> #include <string.h>
>> #include <stdlib.h>
>> #include <math.h>
>> #include <errno.h>
>>
>> #define ROWS (16384)
>> #define COLS (16384)
>> #define MPI_DATATYPE (MPI_DOUBLE)
>> #define C_DATATYPE double
>> #define DIMS (2)
>> #define COLL_BUFSIZE (536870912)
>>
>> int main(int argc, char **argv) {
>> char fname[] = "pvfs2:/home/mdl/patrick/pvfs2/testfile";
>> int i = 0, nprocs = 0, mynod = 0, provided = 0, c_size =
>> 0, mpi_size = 0, iterations = 0,
>> array_size[] = {0, 0}, array_subsize[] = {0, 0},
>> array_start[] = {0, 0};
>> long rows = 0l, cols = 0l, coll_bufsize = 0l, rows_view =
>> 0l, cols_view = 0l, rows_collbuf = 0l, cols_collbuf = 0l, elts_collbuf
>> = 0l;
>> unsigned long filesize = 0l;
>> double *buffer = NULL;
>> MPI_File fhandle;
>> MPI_Status status;
>> MPI_Datatype subarray;
>>
>> MPI_Init_thread(&argc, &argv, MPI_THREAD_MULTIPLE, &provided);
>> MPI_Comm_rank(MPI_COMM_WORLD, &mynod);
>> MPI_Comm_size(MPI_COMM_WORLD, &nprocs);
>>
>> MPI_Type_size(MPI_DATATYPE, &mpi_size);
>> c_size = sizeof(C_DATATYPE);
>> if(c_size != mpi_size) {
>> fprintf(stderr, "Datatypes in MPI and C do not match\n");
>> MPI_Abort(MPI_COMM_WORLD, EIO);
>> }
>>
>> rows = ROWS;
>> cols = COLS;
>> coll_bufsize = COLL_BUFSIZE;
>> elts_collbuf = coll_bufsize / mpi_size;
>> rows_view = rows / nprocs;
>> cols_view = cols;
>> cols_collbuf = cols_view;
>> rows_collbuf = elts_collbuf / cols_collbuf;
>> filesize = rows * cols * mpi_size;
>> array_size[0] = rows;
>> array_size[1] = cols;
>> array_subsize[0] = rows_view;
>> array_subsize[1] = cols_view;
>> array_start[0] = rows_view * mynod;
>> array_start[1] = 0;
>>
>> buffer = (C_DATATYPE *)malloc(coll_bufsize);
>> if(!buffer) {
>> fprintf(stderr, "calloc error\n");
>> MPI_Abort(MPI_COMM_WORLD, ENOMEM);
>> }
>>
>> MPI_File_open(MPI_COMM_WORLD, fname, MPI_MODE_RDONLY, MPI_INFO_NULL,
>> &fhandle);
>>
>> MPI_Type_create_subarray(DIMS, array_size, array_subsize,
>> array_start, MPI_ORDER_C, MPI_DATATYPE, &subarray);
>> MPI_Type_commit(&subarray);
>> MPI_File_set_view(fhandle, 0, MPI_DATATYPE, subarray, "native",
>> MPI_INFO_NULL);
>>
>> iterations = rows_view / rows_collbuf;
>>
>> for(i = 0; i < iterations; i++)
>> MPI_File_read_all(fhandle, buffer, elts_collbuf, MPI_DATATYPE, &status);
>>
>> MPI_File_close(&fhandle);
>> free(buffer);
>> MPI_Finalize();
>>
>> return 0;
>> }
>>
>
More information about the mpich-discuss
mailing list