problem writing BSEFATBAND

Queries about input and output files, running specific calculations, etc.


Moderators: Global Moderator, Moderator

Locked
Message
Author
xiaoming_wang
Jr. Member
Jr. Member
Posts: 58
Joined: Tue Nov 12, 2019 4:34 am

problem writing BSEFATBAND

#1 Post by xiaoming_wang » Tue Dec 14, 2021 3:12 am

Hello,

I'm trying to print out the BSE eigenvectors by setting NBSEEIG. However, when the NBSEEIG is getting larger, I experienced the following error

Code: Select all

Abort(336139522) on node 4 (rank 4 in comm 0): Fatal error in PMPI_Allreduce: Invalid count, error stack:
PMPI_Allreduce(432): MPI_Allreduce(sbuf=0x2b75cdcf3880, rbuf=0x2cc9780, count=-2147483648, datatype=MPI_DOUBLE_PRECISION, op=MPI_SUM, comm=comm=0xc4000004) failed
PMPI_Allreduce(374): Negative count, value is -2147483648
It seems this was caused by the command line

Code: Select all

CALLMPI( M_sum_z(WHF%WDES%COMM_INTER, EIGVECT, NBSEEIG*NCV))
in the bse.F file. However, it seems that the dimension of the bse eigenvector (NCV*NBSEEIG) is not larger than the dimension of WAVEDER (NK*NBANDS*NBANDS*3) for my system. Why can the latter be printed out successfully while the former cannot? For my applications, I do need almost the full BSE eigenvectors. How can I remedy this?

Best,
Xiaoming Wang

fabien_tran1
Global Moderator
Global Moderator
Posts: 418
Joined: Mon Sep 13, 2021 11:02 am

Re: problem writing BSEFATBAND

#2 Post by fabien_tran1 » Tue Dec 14, 2021 8:47 am

Hi,

Could you please provide more details about your system and the values of NBSEEIG that you have tried. Does the calculation systematically stop or it may depends on the number of cores you are using?

xiaoming_wang
Jr. Member
Jr. Member
Posts: 58
Joined: Tue Nov 12, 2019 4:34 am

Re: problem writing BSEFATBAND

#3 Post by xiaoming_wang » Tue Dec 14, 2021 1:31 pm

Hi,

My system has 188 atoms, 1200 bands, 256 kpoints and ncl calculation. For BSE, I have NBANDSO=64, NBANDSV=8. When I set NBSEEIG=12000, it produced the mentioned error. Reduce NBSEEIG to 8000, it works fine. The abortion seems to depends also on the nodes (memory) I used. From the log file, both the BSE matrix setup and diagonalization run successfully. It is the last step about collecting the matrix from different nodes that causing this error.

Best,
Xiaoming

alexey.tal
Global Moderator
Global Moderator
Posts: 314
Joined: Mon Sep 13, 2021 12:45 pm

Re: problem writing BSEFATBAND

#4 Post by alexey.tal » Wed Dec 15, 2021 10:02 am

Hi,

It looks like it could be an integer overflow in the MPI_Allreduce counter. Could you try to change the counter variable to 8 byte integer. For that you will need to set the type of the variable n to INTEGER(KIND=8) in the subroutine M_sum_z in mpi.F

xiaoming_wang
Jr. Member
Jr. Member
Posts: 58
Joined: Tue Nov 12, 2019 4:34 am

Re: problem writing BSEFATBAND

#5 Post by xiaoming_wang » Wed Dec 15, 2021 2:31 pm

Hi,

I tried as your suggestion. But seems cannot compile.

Code: Select all

mpi.F(1908): error #6284: There is no matching specific function for this generic function reference.   [STR]
         CALL vtutor%bug("M_sumb_z: invalid vector size n " // str(n), "mpi.F", 1908)
---------------------------------------------------------------^
mpi.F(1908): error #6054: A CHARACTER data type is required in this context.   [STR]
         CALL vtutor%bug("M_sumb_z: invalid vector size n " // str(n), "mpi.F", 1908)
The size of EIGVECT matrix is 64*8*256*12000=1572864000 which still does not exceed the limit 2147483648. And for the WAVEDER, the size is 1200*1200*256=368640000 which has already exceed the limit but can still be printed out. So I'm confused here.

Best,
Xiaoming

xiaoming_wang
Jr. Member
Jr. Member
Posts: 58
Joined: Tue Nov 12, 2019 4:34 am

Re: problem writing BSEFATBAND

#6 Post by xiaoming_wang » Wed Dec 15, 2021 3:21 pm

xiaoming_wang wrote: Wed Dec 15, 2021 2:31 pm

Code: Select all

mpi.F(1908): error #6284: There is no matching specific function for this generic function reference.   [STR]
         CALL vtutor%bug("M_sumb_z: invalid vector size n " // str(n), "mpi.F", 1908)
---------------------------------------------------------------^
mpi.F(1908): error #6054: A CHARACTER data type is required in this context.   [STR]
         CALL vtutor%bug("M_sumb_z: invalid vector size n " // str(n), "mpi.F", 1908)
I managed to get rid of this by removing str(n) and recompile and rerun BSE with NBSEEIG=12000, but still got the save error about the negative count.

Best,
Xiaoming

alexey.tal
Global Moderator
Global Moderator
Posts: 314
Joined: Mon Sep 13, 2021 12:45 pm

Re: problem writing BSEFATBAND

#7 Post by alexey.tal » Wed Dec 15, 2021 5:02 pm

Are you performing a spin-polarzied calculation? If yes, the matrix size should be 64*8*256*12000*2. Can you provide the rank of the BSE matrix?
Regarding the error, can you change the counter type in M_sumb_z to 8 byte? One should make sure that subroutines called from M_sum_z use compatible arguments.

xiaoming_wang
Jr. Member
Jr. Member
Posts: 58
Joined: Tue Nov 12, 2019 4:34 am

Re: problem writing BSEFATBAND

#8 Post by xiaoming_wang » Thu Dec 16, 2021 1:00 am

It is not spin polarized. The rank of the BSE matrix is 32768. I also changed the counter type to 8 byte in M_sumb_z. However, the error persists.

Best,
Xiaoming

alexey.tal
Global Moderator
Global Moderator
Posts: 314
Joined: Mon Sep 13, 2021 12:45 pm

Re: problem writing BSEFATBAND

#9 Post by alexey.tal » Thu Dec 16, 2021 8:36 am

I will try to reproduce this issue. Can you please provide the input files for your calculation and the version of the code that you are using?

xiaoming_wang
Jr. Member
Jr. Member
Posts: 58
Joined: Tue Nov 12, 2019 4:34 am

Re: problem writing BSEFATBAND

#10 Post by xiaoming_wang » Thu Dec 16, 2021 2:16 pm

Please find the attached input files.
You do not have the required permissions to view the files attached to this post.

xiaoming_wang
Jr. Member
Jr. Member
Posts: 58
Joined: Tue Nov 12, 2019 4:34 am

Re: problem writing BSEFATBAND

#11 Post by xiaoming_wang » Thu Dec 16, 2021 3:50 pm

I'm using the version 6.2.0.

alexey.tal
Global Moderator
Global Moderator
Posts: 314
Joined: Mon Sep 13, 2021 12:45 pm

Re: problem writing BSEFATBAND

#12 Post by alexey.tal » Mon Dec 20, 2021 5:22 pm

Thank you for sending the input files. Indeed, it seems like the problem is in the MPI_Allreduce counter overflow but what I was suggesting above would not be
sufficient. The eigenvectors array size for your system should be 64*8*256*12000*2, where the last factor 2 is due to the fact that here complex
numbers are communicated between cores as two doubles.

I created a patch that changes the relevant counters to 8 byte integers. Unfortunately, I cannot test it for your system specifically due to its size, so
could you please apply this patch to your 6.2.0 code and try to repeat the calculation?

Code: Select all

diff --git a/src/bse.F b/src/bse.F
index 83de6ab9..682ce859 100644
--- a/src/bse.F
+++ b/src/bse.F
@@ -5926,7 +5926,7 @@ cp: DO N=1,WHF%WDES%COMM_INTER%NCPU
 
 !Gather all parts of the eigenvectors stored on the different nodes
 !    CALLMPI( M_sum_z(WHF%WDES%COMM, EIGVECT, NBSEEIG*NCV))
-    CALLMPI( M_sum_z(WHF%WDES%COMM_INTER, EIGVECT, NBSEEIG*NCV))
+    CALLMPI( M_sum_z8(WHF%WDES%COMM_INTER, EIGVECT, INT(NBSEEIG*NCV,8)))
     IF (IU6>=0) CALL WRITE_BSE(WHF,NCV,BSE_INDEX,EIGVECT,R,NBSEEIG)
 
     DEALLOCATE(EIGVECT)
diff --git a/src/mpi.F b/src/mpi.F
index 23e78596..8a63cc65 100644
--- a/src/mpi.F
+++ b/src/mpi.F
@@ -1792,6 +1791,82 @@
 
       END SUBROUTINE M_sumb_d
 
+!----------------------------------------------------------------------
+!
+! M_sumb_d8: performs a global sum on n doubles in vector vec
+!  uses MPI_allreduce which is usually very inefficient
+!  faster alternative routines can be found below 
+!  copy of M_sumb_d with int8 counter
+!
+!----------------------------------------------------------------------
+
+      SUBROUTINE M_sumb_d8(COMM, vec, n )
+#ifdef _OPENACC
+      USE mopenacc_struct_def
+#endif
+      USE mpimy
+      USE string, ONLY: str
+      USE tutor, ONLY: vtutor
+      IMPLICIT NONE
+      INCLUDE "pm.inc"
+
+      TYPE(communic) COMM
+      INTEGER(KIND=8) n,j, ichunk
+      REAL(q) vec(n)
+
+      INTEGER  ierror, status(MPI_status_size)
+
+! quick return if possible
+      IF (COMM%NCPU == 1) RETURN
+
+! check whether n is sensible
+      !IF (n==0) THEN
+      !   RETURN
+      !ELSE IF (n<0) THEN
+      !   CALL vtutor%bug("M_sumb_d: invalid vector size n " // str(n), __FILE__, __LINE__)
+      !END IF
+
+      PROFILING_START('m_sumb_d')
+
+#ifdef _OPENACC
+      IF (ACC_IS_PRESENT(vec).AND.ACC_EXEC_ON) THEN
+! invoke CUDA-aware MPI_allreduce
+!$ACC ENTER DATA CREATE(DTMP_m) ASYNC(ACC_ASYNC_Q)
+      DO j = 1, n, NDTMP
+         ichunk = MIN( n-j+1 , NDTMP)
+!$ACC WAIT(ACC_ASYNC_Q)
+!$ACC HOST_DATA USE_DEVICE(vec,DTMP_m)
+         CALL MPI_allreduce( vec(j), DTMP_m(1), ichunk, &
+                             MPI_double_precision, MPI_sum, &
+                             COMM%MPI_COMM, ierror )
+!$ACC END HOST_DATA
+         CALL __DCOPY__(ichunk , DTMP_m(1), 1 ,  vec(j) , 1)
+      ENDDO
+!$ACC EXIT DATA DELETE(DTMP_m) ASYNC(ACC_ASYNC_Q)
+      PROFILING_STOP('m_sumb_d')
+      RETURN
+      ENDIF
+#endif
+!  there is no inplace global sum in MPI, thus we have to use
+!  a work array
+
+      DO j = 1, n, NDTMP
+         ichunk = MIN( n-j+1 , NDTMP)
+
+         CALL MPI_allreduce( vec(j), DTMP_m(1), ichunk, &
+                             MPI_double_precision, MPI_sum, &
+                             COMM%MPI_COMM, ierror )
+
+         IF ( ierror /= MPI_success ) &
+            CALL vtutor%error('M_sumb_d: MPI_allreduce returns: ' // str(ierror))
+
+         CALL DCOPY(ichunk , DTMP_m(1), 1 ,  vec(j) , 1)
+      ENDDO
+
+      PROFILING_STOP('m_sumb_d')
+
+      END SUBROUTINE M_sumb_d8
+
 !----------------------------------------------------------------------
 !
 ! M_sumb_d: performs a global sum on n singles in vector vec
@@ -4307,6 +4382,109 @@
 #endif
       END SUBROUTINE M_sumf_d
 
+
+!----------------------------------------------------------------------
+!
+! M_sumf_d8: performs a fast global sum on n doubles in
+! vector vec (algorithm by Kresse Georg)
+!
+! uses complete interchange algorithm (my own invention, but I guess
+!  some people must know it)
+! exchange data between nodes, sum locally and
+! interchange back, this algorithm is faster than typical MPI based
+! algorithms (on 8 nodes under MPICH a factor 4)
+! copy of M_sumf_d with int8 counter
+!
+!----------------------------------------------------------------------
+
+      SUBROUTINE M_sumf_d8(COMM, vec, n)
+      USE mpimy
+      USE string, ONLY: str
+      USE tutor, ONLY: vtutor
+      IMPLICIT NONE
+
+      TYPE(communic) COMM
+      INTEGER(KIND=8) n,ncount,nsummed,ndo, n_,mmax
+      INTEGER(KIND=8) i,j, info
+      REAL(q) vec(n)
+!----------------------------------------------------------------------
+#if defined(T3D_SMA)
+!----------------------------------------------------------------------
+      INTEGER MALLOC_DONE
+      INTEGER, EXTERNAL :: ISHM_CHECK
+      COMMON /SHM/ MALLOC_DONE, PBUF
+      POINTER ( PBUF, vec_inter )
+      REAL(q) :: vec_inter(n/COMM%NCPU*COMM%NCPU)
+      INTEGER(kind=8) :: max_=n/COMM%NCPU
+
+    ! quick return if possible
+      IF (COMM%NCPU == 1) RETURN
+      ! do we have sufficient shm workspace to use fast interchange algorithm
+      ! no use conventional M_sumb_d
+      IF (ISHM_CHECK(n) == 0) THEN
+         CALL M_sumb_d8(COMM, vec, n)
+         RETURN
+      ENDIF
+!----------------------------------------------------------------------
+#else
+!----------------------------------------------------------------------
+      REAL(q), ALLOCATABLE :: vec_inter(:)
+    ! maximum work space for quick sum
+!
+! maximum communication blocks
+! too large blocks are slower on the Pentium architecture
+! probably due to caching
+!
+      INTEGER(kind=8), PARAMETER :: max_=MPI_BLOCK
+
+    ! quick return if possible
+      IF (COMM%NCPU == 1) RETURN
+      
+      mmax=MIN(n/COMM%NCPU,max_)
+      ALLOCATE(vec_inter(mmax*COMM%NCPU))
+!----------------------------------------------------------------------
+#endif
+!----------------------------------------------------------------------
+      nsummed=0
+      n_=n/COMM%NCPU
+
+      DO ndo=0,n_-1,mmax
+     ! forward exchange
+         ncount =MIN(mmax,n_-ndo)
+         nsummed=nsummed+ncount*COMM%NCPU
+
+         CALL M_alltoall_d(COMM, ncount*COMM%NCPU, vec(ndo*COMM%NCPU+1), vec_inter(1))
+        ! sum localy
+         DO i=2, COMM%NCPU
+           CALL DAXPY(ncount, 1.0_q, vec_inter(1+(i-1)*ncount), 1, vec_inter(1), 1)
+         ENDDO
+     ! replicate data (will be send to each proc)
+         DO i=1, COMM%NCPU
+            DO j=1,ncount
+               vec(ndo*COMM%NCPU+j+(i-1)*ncount) = vec_inter(j)
+            ENDDO
+         ENDDO
+     ! backward exchange
+         CALL M_alltoall_d(COMM, ncount*COMM%NCPU, vec(ndo*COMM%NCPU+1), vec_inter(1))
+         CALL DCOPY( ncount*COMM%NCPU, vec_inter(1), 1, vec(ndo*COMM%NCPU+1), 1 )
+      ENDDO
+      
+     ! that should be it
+      !IF (n_*COMM%NCPU /= nsummed) THEN
+      !   CALL vtutor%bug("M_sumf_d: " // str(n_) // " " // str(nsummed), __FILE__, __LINE__)
+      !ENDIF
+
+      IF (n-nsummed /= 0 ) &
+        CALL M_sumb_d8(COMM, vec(nsummed+1), n-nsummed)
+
+#if defined(T3D_SMA)
+     ! nup nothing to do here
+#else
+      DEALLOCATE(vec_inter)
+#endif
+      END SUBROUTINE M_sumf_d8
+
+
 !----------------------------------------------------------------------
 !
 ! M_sumf_s: performs a fast global sum on n singles in
@@ -4426,6 +4604,33 @@
 #endif
       END SUBROUTINE M_sum_z
 
+!----------------------------------------------------------------------
+!
+! M_sum_z8: performs a sum on n double complex numbers
+!  it uses either sumb_d8 or sumf_d
+!  copy of the M_sum_z with int8 counter
+!
+!----------------------------------------------------------------------
+
+      SUBROUTINE M_sum_z8(COMM, vec, n)
+      USE mpimy
+      IMPLICIT NONE
+
+      TYPE(communic) COMM
+      INTEGER(KIND=8) n
+      REAL(q) vec(n)
+
+#ifdef use_collective_sum
+      CALL M_sumb_d8(COMM, vec, 2*n)
+#else
+      IF ( 2*n>MPI_BLOCK) THEN
+         CALL M_sumf_d8(COMM, vec, 2*n)
+      ELSE
+         CALL M_sumb_d8(COMM, vec, 2*n)
+      ENDIF
+#endif
+      END SUBROUTINE M_sum_z8
+
 !----------------------------------------------------------------------
 !
 ! M_sum_qd: performs a sum on n quadruple complex numbers

xiaoming_wang
Jr. Member
Jr. Member
Posts: 58
Joined: Tue Nov 12, 2019 4:34 am

Re: problem writing BSEFATBAND

#13 Post by xiaoming_wang » Mon Dec 20, 2021 11:59 pm

Thanks! I tried the patch and it works.

Best,
Xiaoming Wang

Locked