Hi Fabien, Mg2Si is a small structure with 2 irreducible positions in the IBZ. Here is the struct file: Mg2Si cubic 225 Fm-3m F LATTICE,NONEQUIV.ATOMS 2 225 Fm-3m MODE OF CALC=RELA unit=bohr 11.999761 11.999761 11.999761 90.000000 90.000000 90.000000 ATOM 1: X=0.00000000 Y=0.00000000 Z=0.00000000 MULT= 1 ISPLIT= 2 Si NPT= 781 R0=.000050000 RMT= 2.39 Z: 14.00000 LOCAL ROT MATRIX: 1.0000000 0.0000000 0.0000000 0.0000000 1.0000000 0.0000000 0.0000000 0.0000000 1.0000000 ATOM 2: X=0.25000000 Y=0.25000000 Z=0.25000000 MULT= 2 ISPLIT= 2 2: X=0.25000000 Y=0.25000000 Z=0.75000000 Mg NPT= 781 R0=.000050000 RMT= 2.50000 Z: 12.00000 LOCAL ROT MATRIX: 1.0000000 0.0000000 0.0000000 0.0000000 1.0000000 0.0000000 0.0000000 0.0000000 1.0000000 48 NUMBER OF SYMMETRY OPERATIONS
The number of k-points is 72 (12x12x12), RmtKmax is 7, GMAX=12 (for test purpose, usually I set it to 24). Best, Pascal Pascal Boulet — Professor in computational materials - DEPARTMENT OF CHEMISTRY University of Aix-Marseille - Avenue Escadrille Normandie Niemen - F-13013 Marseille - FRANCE Tél: +33(0)4 13 55 18 10 - Fax : +33(0)4 13 55 18 50 Email : pascal.bou...@univ-amu.fr <mailto:pascal.bou...@univ-amu.fr> > Le 15 août 2020 à 11:40, Tran, Fabien <fabien.t...@tuwien.ac.at> a écrit : > > Hi, > > For calculations on small cells with many k-points it is more preferable (for > speediness) to use k-point paralllelization instead of MPI parallelization. > And, as mentioned by PB, MPI applied to small matrices may not work. > > Of course, if the number of cores that you have at disposal is larger (twice > for instance) than the number of k-points in the IBZ, then you can combine > the k-point and MPI parallelizations (two cores for each k-point). > > An example of .machines file for k-point parallelization is (supposing that > you have 6 k-points in the IBZ and want to use one machine having 6 cores): > > lapw0: n1071 n1071 n1071 n1071 n1071 n1071 > dstart: n1071 n1071 n1071 n1071 n1071 n1071 > nlvdw: n1071 n1071 n1071 n1071 n1071 n1071 > 1:1071 > 1:1071 > 1:1071 > 1:1071 > 1:1071 > 1:1071 > granularity:1 > extrafine:1 > > The six lines "1:1071" mean that lapw1, lapw2 and hf are k-point (no MPI) > parallelized (one line fore each core). lapw0, dstart and nlvdw are MPI > parallelized. In this example, the omp parallelization is ignored by > supposing that OMP_NUM_THREADS is set to 1. > > Besides, I find your computational time with HF as very large. What is the > number of atoms in the cell, the number of k-points (plz, specify th n1xn2nx3 > k-mesh), RKmax, etc.? > > Best, > FT > > From: Wien <wien-boun...@zeus.theochem.tuwien.ac.at> on behalf of pboulet > <pascal.bou...@univ-amu.fr> > Sent: Saturday, August 15, 2020 10:35 AM > To: A Mailing list for WIEN2k users > Subject: Re: [Wien] Wien post from pascal.bou...@univ-amu.fr (Errors with > DGEMM and hybrid calculations) > > Dear Peter, > > Thank you for your response. It clarifies some points for me. > > I have run another calculation for Mg2Si, which is a small system, on 12 > cores (no MKL errors!). The job ran for 12 hours (CPU time limit I set) and > made only 3 SCF cycles without converging. > > The .machines file I use looks like this: > > 1:1071:12 > lapw0: n1071 n1071 > dstart: n1071 n1071 > nlvdw: n1071 n1071 > granularity:1 > extrafine:1 > > I guess I am not optimising the number of cores w.r.t. the size of the > problem (72 k-points, 14 HF bands, 12 occupied +2 unoccupied). > > I changed the number of processors for 72, hoping for a 1 k-point/core > parallelisation and commenting all the lines of .machines except granularity > and extrafine. I got less than 1 cycle in 12 hours. > > What should I do to run the HF part on k-points parallelisation only (no > mpi)? This point that is not clear for me from the manual. > > Thank you > Best regards > Pascal > > Pascal Boulet > — > Professor in computational materials - DEPARTMENT OF CHEMISTRY > > University of Aix-Marseille - Avenue Escadrille Normandie Niemen - F-13013 > Marseille - FRANCE > Tél: +33(0)4 13 55 18 10 - Fax : +33(0)4 13 55 18 50 > Email : pascal.bou...@univ-amu.fr > > Le 12 août 2020 à 14:12, Peter Blaha <pbl...@theochem.tuwien.ac.at> a écrit : > > Your message is too big to be accepted. > > Anyway, the DGEMM messages seem to be a relict of the mkl you are using, and > most likely is related to the use of too many mpi-cores for such a small > matrix. At least when I continue your Mg2Si calculations (in k-parallel mode) > the :DIS and :ENE are continuous, which means that the previous results are > ok. > > Concerning hf, I don't know. Again, running this in sequential (k-point > parallel) mode is no problems and converges quickly. > > I suggest that you change your setup to a k-parallel run for such small > systems. > > Best regards > Peter Blaha > > --------------------------------------------------------------------- > Subject: > Errors with DGEMM and hybrid calculations > From: > pboulet <pascal.bou...@univ-amu.fr> > Date: > 8/11/20, 6:31 PM > To: > A Mailing list for WIEN2k users <wien@zeus.theochem.tuwien.ac.at> > > Dear all, > > I have a strange problem with LAPACK. I get an error message with wrong > parameters sent to DGEMM, but still wien2k (19.2) seems to converge the scf. > Is that possible? What could be the "problem"? > > I have attached an archive containing the summary of the SCF + compilation > options + SLURM output file. The error message is in the dayfile. > The same error shows up with Wien2k 18.1. > > Actually this case is a test case for testing hybrid calculations as I have > problems with my real case, which is found to be metallic with PBE. At least > Mg2Si is a small band gap semiconductor. > > When I go ahead with the HSE06 functional and Mg2Si I get a different error: > segmentation fault during the hf run. As Mg2Si is a small system I guess this > is not a memory problem: the node is 128GB. > Note that the same problem occurs for my real case file, but the LAPACK > problem does not occur. > > The second archive contains some files related to the hybrid calculation. > > Some hints would be welcome as I am completely lost in these (unrelated) > errors! > > Thank you. > Best regards, > Pascal > > > Pascal Boulet > -- > > P.Blaha > -------------------------------------------------------------------------- > Peter BLAHA, Inst.f. Materials Chemistry, TU Vienna, A-1060 Vienna > Phone: +43-1-58801-165300 FAX: +43-1-58801-165982 > Email: bl...@theochem.tuwien.ac.at WIEN2k: http://www.wien2k.at > WWW: http://www.imc.tuwien.ac.at/TC_Blaha > -------------------------------------------------------------------------- > _______________________________________________ > Wien mailing list > Wien@zeus.theochem.tuwien.ac.at > http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien > SEARCH the MAILING-LIST at: > http://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/index.html > _______________________________________________ > Wien mailing list > Wien@zeus.theochem.tuwien.ac.at > http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien > SEARCH the MAILING-LIST at: > http://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/index.html
smime.p7s
Description: S/MIME cryptographic signature
_______________________________________________ Wien mailing list Wien@zeus.theochem.tuwien.ac.at http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien SEARCH the MAILING-LIST at: http://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/index.html