C -*- Mode: Fortran; -*-
C
C (C) 2003 by Argonne National Laboratory.
C See COPYRIGHT in top-level directory.
C
program main
implicit none
include 'mpif.h'
include 'iooffset.h'
integer errs, ierr, size, rank
integer fh, comm, status(MPI_STATUS_SIZE)
integer buf(1024)
errs = 0
call MTest_Init( ierr )
C This test reads a header then sets the view to every "size" int,
C using set view and current displacement. The file is first written
C using a combination of collective and ordered writes
comm = MPI_COMM_WORLD
call MPI_File_open( comm, "test.ord", MPI_MODE_WRONLY +
$ MPI_MODE_CREATE, MPI_INFO_NULL, fh, ierr )
if (ierr .ne. MPI_SUCCESS) then
errs = errs + 1
call MTestPrintErrorMsg( "Open(1)", ierr )
endif
call MPI_Comm_size( comm, size, ierr )
call MPI_Comm_rank( comm, rank, ierr )
if (size .gt. 1024) then
if (rank .eq. 0) then
print *,
$"This program must be run with no more than 1024 processes"
call MPI_Abort( MPI_COMM_WORLD, 1, ierr )
endif
endif
buf(1) = size
call MPI_File_write_all( fh, buf, 1, MPI_INTEGER, status, ierr )
if (ierr .ne. MPI_SUCCESS) then
errs = errs + 1
call MTestPrintErrorMsg( "Write_all", ierr )
endif
call MPI_File_get_position( fh, offset, ierr )
if (ierr .ne. MPI_SUCCESS) then
errs = errs + 1
call MTestPrintErrorMsg( "Get_position", ierr )
endif
call MPI_File_seek_shared( fh, offset, MPI_SEEK_SET, ierr )
if (ierr .ne. MPI_SUCCESS) then
errs = errs + 1
call MTestPrintErrorMsg( "Seek_shared", ierr )
endif
buf(1) = rank
call MPI_File_write_ordered( fh, buf, 1, MPI_INTEGER, status,ierr)
if (ierr .ne. MPI_SUCCESS) then
errs = errs + 1
call MTestPrintErrorMsg( "Write_ordered", ierr )
endif
call MPI_File_close( fh, ierr )
if (ierr .ne. MPI_SUCCESS) then
errs = errs + 1
call MTestPrintErrorMsg( "Close(1)", ierr )
endif
C Reopen the file as sequential
call MPI_File_open( comm, "test.ord", MPI_MODE_RDONLY +
$ MPI_MODE_SEQUENTIAL + MPI_MODE_DELETE_ON_CLOSE,
$ MPI_INFO_NULL, fh, ierr )
if (ierr .ne. MPI_SUCCESS) then
errs = errs + 1
call MTestPrintErrorMsg( "Open(Read)", ierr )
endif
if (rank .eq. 0) then
call MPI_File_read_shared( fh, buf, 1, MPI_INTEGER, status,
$ ierr )
if (ierr .ne. MPI_SUCCESS) then
errs = errs + 1
call MTestPrintErrorMsg( "Read_all", ierr )
endif
if (buf(1) .ne. size) then
errs = errs + 1
print *, "Unexpected value for the header = ", buf(1),
$ ", should be ", size
endif
endif
call MPI_Barrier( comm, ierr )
C All processes must provide the same file view for MODE_SEQUENTIAL
call MPI_File_set_view( fh, MPI_DISPLACEMENT_CURRENT, MPI_INTEGER
$ ,MPI_INTEGER, "native", MPI_INFO_NULL, ierr )
if (ierr .ne. MPI_SUCCESS) then
errs = errs + 1
call MTestPrintErrorMsg( "Set_view", ierr )
endif
buf(1) = -1
call MPI_File_read_ordered( fh, buf, 1, MPI_INTEGER, status, ierr
$ )
if (ierr .ne. MPI_SUCCESS) then
errs = errs + 1
call MTestPrintErrorMsg( "Read_all", ierr )
endif
if (buf(1) .ne. rank) then
errs = errs + 1
print *, rank, ": buf(1) = ", buf(1)
endif
call MPI_File_close( fh, ierr )
if (ierr .ne. MPI_SUCCESS) then
errs = errs + 1
call MTestPrintErrorMsg( "Close(2)", ierr )
endif
call MTest_Finalize( errs )
call MPI_Finalize( ierr )
end