[Wien] Hybrid Test on MgO

2015-03-18 Thread Paul Fons
As suggested I have calculated a hybrid functional for the MgO primitive cell.

My .machines file was

# 
lapw0:localhost:2
8:localhost:2
8:draco-ib:2
granularity
extrafine:1

and my MgO.inhf file looks like

0.25 alpha
Tscreened (T) or unscreened (F)
0.165lambda
9nband
6gmax
3lmaxe
3lmaxv
1d-3 tolu



For the regular PBE calculation (just run_lapw -p -in1new 2), a SCF loop took 
about 15 seconds.  For the hybrid calculation with options “run_lapw -hf -p 
-in1new 2”, with the above case.inhf file (screening off), the calculation on 
the same structure took about 90 seconds/SCF loop, making the hybrid 
calculations about a factor of six slower.

I encountered no errors with this run unlike my attempts with the 96 atom 
system.  In response to the question asked about the size of aCGT.weighf, the 
size of the file was 117,020 bytes.


Best wishes,
Paul Fons





 On Mar 13, 2015, at 11:37 PM, t...@theochem.tuwien.ac.at wrote:
 
 What is the size of aCGT.weighhf? Is it empty?
 
 Also, before continuing further with your big system, it would be
 interesting to know if the same problem occurs with a very small system
 like MgO (on same machine and in MPI mode).
 
 Anyway, I still think that it is hopeless to apply hybrid functionals on
 such a big system.
 
 F. Tran
 
 
 On Fri, 13 Mar 2015, Paul Fons wrote:
 
 I attempted to run a SCF loop using a hybrid functional and have run into 
 some problems.  In my earlier try I had a incorrectly specified .machines 
 file now I have addressed this problem. I also changed the SCRATCH 
 environment variable to “./“ so that it points to the main directory for the 
 calculation.  I have run a PBE SCF loop to normal terminal for an amorphous 
 cluster of 96 atoms of Cu-Ge-Te.  I then ran the init_hf script and after 
 setting the number of bands to 770 for my 1526 electron system, I set the MP 
 grid to 2x2x2 for a total of four k-points.  I then ran the command 
 run_lapw -hf -p -in1new 2”.  The SCF loop ran through lapw0, lapw1, lapw2, 
 core, and then crashed at the program hf.  The MPI processes upon crashing 
 reported the following error:
 
 forrtl: severe (67): input statement requires too much data, unit 26, file 
 /usr/local/share/wien2k/Fons/aCGT/aCGT.weighhf
 
 I have no idea why the program failed this time.  I am using the Intel 
 compiler (15) and the Intel MPI environment (e.g. mpiexec.hydra) to launch 
 parallel programs as can be seen in the “parallel_options” file.  The only 
 *.error files are those like “hf_1.error” to “hf_4.error”  which contain the 
 not particularly useful information “error in hf”. So the error occurred in 
 the routine “hf”.   I would be most grateful for any advice as to what to 
 try next.  I have included what I hope is relevant debugging information 
 below.
 
 
 My parallel_options files (in all nodes) are
 
 
 mats...@gemini.a04.aist.go.jp:~/Wien2Kcat parallel_options setenv TASKSET 
 no
 setenv USE_REMOTE 0
 setenv MPI_REMOTE 0
 setenv WIEN_GRANULARITY 1
 setenv WIEN_MPIRUN mpiexec.hydra -n _NP_ -machinefile _HOSTS_ _EXEC_
 
 
 My .machines files is as follows:
 
 
 lapw0:localhost:12
 1:localhost:12
 1:localhost:12
 1:draco-ib:12
 1:draco-ib:12
 granularity:1
 extrafine:1
 
 
 
 CONTENTS of :parallel
 
 -
 starting parallel lapw1 at Thu Mar 12 14:25:13 JST 2015
localhost localhost localhost localhost localhost localhost localhost 
 localhost localhost localhost localhost localhost(1) 25254.539u 519.601s 
 35:44.50 1201.8%   0+0k 8+882304io 0pf+0w
localhost localhost localhost localhost localhost localhost localhost 
 localhost localhost localhost localhost localhost(1) 24889.112u 585.238s 
 35:41.95 1189.3%   0+0k 0+719488io 0pf+0w
draco-ib draco-ib draco-ib draco-ib draco-ib draco-ib draco-ib draco-ib 
 draco-ib draco-ib draco-ib draco-ib(1) 0.034u 0.021s 32:40.68 0.0%
 0+0k 0+0io 0pf+0w
draco-ib draco-ib draco-ib draco-ib draco-ib draco-ib draco-ib draco-ib 
 draco-ib draco-ib draco-ib draco-ib(1) 0.035u 0.017s 32:39.14 0.0%
 0+0k 0+0io 0pf+0w
  Summary of lapw1para:
  localhostk=0 user=0  wallclock=0
  draco-ib k=0 user=0  wallclock=0
 -  done at Thu Mar 12 15:01:00 JST 2015
 -
 -  starting Fermi on gemini.a04.aist.go.jp at Thu Mar 12 15:28:19 JST 2015
 -  starting parallel lapw2c at Thu Mar 12 15:28:20 JST 2015
 localhost 389.940u 7.565s 0:36.04 1102.9% 0+0k 718416+253704io 0pf+0w
 localhost 347.944u 5.749s 0:31.67 1116.7% 0+0k 718528+199776io 0pf+0w
 draco-ib 0.029u 0.026s 0:33.86 0.1% 0+0k 8+0io 0pf+0w
 draco-ib 0.032u 0.020s 0:33.80 0.1% 0+0k 8+0io 0pf+0w
  Summary of lapw2para:
  localhostuser=737.884wallclock=67.71
  draco-ib user=0.061  wallclock=67.66
 -  done at Thu Mar 12 15:28:57 JST 2015

[Wien] Update on hybrid troubles

2015-03-13 Thread Paul Fons
I attempted to run a SCF loop using a hybrid functional and have run into some 
problems.  In my earlier try I had a incorrectly specified .machines file now I 
have addressed this problem. I also changed the SCRATCH environment variable to 
“./“ so that it points to the main directory for the calculation.  I have run a 
PBE SCF loop to normal terminal for an amorphous cluster of 96 atoms of 
Cu-Ge-Te.  I then ran the init_hf script and after setting the number of bands 
to 770 for my 1526 electron system, I set the MP grid to 2x2x2 for a total of 
four k-points.  I then ran the command run_lapw -hf -p -in1new 2”.  The SCF 
loop ran through lapw0, lapw1, lapw2, core, and then crashed at the program hf. 
 The MPI processes upon crashing reported the following error:

forrtl: severe (67): input statement requires too much data, unit 26, file 
/usr/local/share/wien2k/Fons/aCGT/aCGT.weighhf

I have no idea why the program failed this time.  I am using the Intel compiler 
(15) and the Intel MPI environment (e.g. mpiexec.hydra) to launch parallel 
programs as can be seen in the “parallel_options” file.  The only *.error files 
are those like “hf_1.error” to “hf_4.error”  which contain the not particularly 
useful information “error in hf”. So the error occurred in the routine “hf”.   
I would be most grateful for any advice as to what to try next.  I have 
included what I hope is relevant debugging information below.


My parallel_options files (in all nodes) are


mats...@gemini.a04.aist.go.jp:~/Wien2Kcat parallel_options 
setenv TASKSET no
setenv USE_REMOTE 0
setenv MPI_REMOTE 0
setenv WIEN_GRANULARITY 1
setenv WIEN_MPIRUN mpiexec.hydra -n _NP_ -machinefile _HOSTS_ _EXEC_


My .machines files is as follows:


lapw0:localhost:12
1:localhost:12
1:localhost:12
1:draco-ib:12
1:draco-ib:12
granularity:1
extrafine:1



CONTENTS of :parallel

-
starting parallel lapw1 at Thu Mar 12 14:25:13 JST 2015
 localhost localhost localhost localhost localhost localhost localhost 
localhost localhost localhost localhost localhost(1) 25254.539u 519.601s 
35:44.50 1201.8%0+0k 8+882304io 0pf+0w
 localhost localhost localhost localhost localhost localhost localhost 
localhost localhost localhost localhost localhost(1) 24889.112u 585.238s 
35:41.95 1189.3%0+0k 0+719488io 0pf+0w
 draco-ib draco-ib draco-ib draco-ib draco-ib draco-ib draco-ib draco-ib 
draco-ib draco-ib draco-ib draco-ib(1) 0.034u 0.021s 32:40.68 0.0% 0+0k 0+0io 
0pf+0w
 draco-ib draco-ib draco-ib draco-ib draco-ib draco-ib draco-ib draco-ib 
draco-ib draco-ib draco-ib draco-ib(1) 0.035u 0.017s 32:39.14 0.0% 0+0k 0+0io 
0pf+0w
   Summary of lapw1para:
   localhost k=0 user=0  wallclock=0
   draco-ib  k=0 user=0  wallclock=0
-  done at Thu Mar 12 15:01:00 JST 2015
-
-  starting Fermi on gemini.a04.aist.go.jp at Thu Mar 12 15:28:19 JST 2015
-  starting parallel lapw2c at Thu Mar 12 15:28:20 JST 2015
  localhost 389.940u 7.565s 0:36.04 1102.9% 0+0k 718416+253704io 0pf+0w
  localhost 347.944u 5.749s 0:31.67 1116.7% 0+0k 718528+199776io 0pf+0w
  draco-ib 0.029u 0.026s 0:33.86 0.1% 0+0k 8+0io 0pf+0w
  draco-ib 0.032u 0.020s 0:33.80 0.1% 0+0k 8+0io 0pf+0w
   Summary of lapw2para:
   localhost user=737.884wallclock=67.71
   draco-ib  user=0.061  wallclock=67.66
-  done at Thu Mar 12 15:28:57 JST 2015
-  starting sumpara 4 on gemini.a04.aist.go.jp at Thu Mar 12 15:28:58 JST 2015
-  done at Thu Mar 12 15:29:27 JST 2015
-
-  starting parallel hfc at Thu Mar 12 15:29:35 JST 2015
**  HF crashed at Thu Mar 12 15:29:40 JST 2015
**  check ERROR FILES!




DAYFILE

mats...@libra.a04.aist.go.jp:/usr/local/share/wien2k/Fons/aCGTcat aCGT.dayfile

Calculating aCGT in /usr/local/share/wien2k/Fons/aCGT
on gemini.a04.aist.go.jp with PID 46746
using WIEN2k_14.2 (Release 15/10/2014) in /home/matstud/Wien2K


start   (Thu Mar 12 11:22:09 JST 2015) with lapw0 (40/99 to go)

cycle 1 (Thu Mar 12 11:22:09 JST 2015)  (40/99 to go)

   lapw0 -grr -p   (11:22:09) starting parallel lapw0 at Thu Mar 12 
 11:22:09 JST 2015
 .machine0 : 12 processors
745.365u 3.350s 1:05.09 1150.2% 0+0k 144+796936io 0pf+0w
   lapw0 -p(11:23:14) starting parallel lapw0 at Thu Mar 12 11:23:15 JST 
 2015
 .machine0 : 12 processors
620.682u 2.669s 0:54.15 1151.1% 0+0k 40+203264io 0pf+0w
   lapw1-c (11:24:09) 20736.270u 146.444s 3:01:03.72 192.2%
 0+0k 11992+5913840io 0pf+0w
   lapw1  -p   -c  (14:25:13) starting parallel lapw1 at Thu Mar 12 
 14:25:13 JST 2015
-  starting parallel LAPW1 jobs at Thu Mar 12 14:25:13 JST 2015
running LAPW1 in parallel mode (using .machines)
4 number_of_parallel_jobs
 localhost localhost localhost localhost localhost localhost localhost 

[Wien] Update on hybrid troubles

2015-03-13 Thread Paul Fons
I attempted to run a SCF loop using a hybrid functional and have run into some 
problems.  In my earlier try I had a incorrectly specified .machines file now I 
have addressed this problem. I also changed the SCRATCH environment variable to 
“./“ so that it points to the main directory for the calculation.  I have run a 
PBE SCF loop to normal terminal for an amorphous cluster of 96 atoms of 
Cu-Ge-Te.  I then ran the init_hf script and after setting the number of bands 
to 770 for my 1526 electron system, I set the MP grid to 2x2x2 for a total of 
four k-points.  I then ran the command run_lapw -hf -p -in1new 2”.  The SCF 
loop ran through lapw0, lapw1, lapw2, core, and then crashed at the program hf. 
 The MPI processes upon crashing reported the following error:

forrtl: severe (67): input statement requires too much data, unit 26, file 
/usr/local/share/wien2k/Fons/aCGT/aCGT.weighhf

I have no idea why the program failed this time.  I am using the Intel compiler 
(15) and the Intel MPI environment (e.g. mpiexec.hydra) to launch parallel 
programs as can be seen in the “parallel_options” file.  The only *.error files 
are those like “hf_1.error” to “hf_4.error”  which contain the not particularly 
useful information “error in hf”. So the error occurred in the routine “hf”.   
I would be most grateful for any advice as to what to try next.  I have 
included what I hope is relevant debugging information below.


My parallel_options files (in all nodes) are


mats...@gemini.a04.aist.go.jp:~/Wien2Kcat parallel_options 
setenv TASKSET no
setenv USE_REMOTE 0
setenv MPI_REMOTE 0
setenv WIEN_GRANULARITY 1
setenv WIEN_MPIRUN mpiexec.hydra -n _NP_ -machinefile _HOSTS_ _EXEC_


My .machines files is as follows:


lapw0:localhost:12
1:localhost:12
1:localhost:12
1:draco-ib:12
1:draco-ib:12
granularity:1
extrafine:1



CONTENTS of :parallel

-
starting parallel lapw1 at Thu Mar 12 14:25:13 JST 2015
 localhost localhost localhost localhost localhost localhost localhost 
localhost localhost localhost localhost localhost(1) 25254.539u 519.601s 
35:44.50 1201.8%0+0k 8+882304io 0pf+0w
 localhost localhost localhost localhost localhost localhost localhost 
localhost localhost localhost localhost localhost(1) 24889.112u 585.238s 
35:41.95 1189.3%0+0k 0+719488io 0pf+0w
 draco-ib draco-ib draco-ib draco-ib draco-ib draco-ib draco-ib draco-ib 
draco-ib draco-ib draco-ib draco-ib(1) 0.034u 0.021s 32:40.68 0.0% 0+0k 0+0io 
0pf+0w
 draco-ib draco-ib draco-ib draco-ib draco-ib draco-ib draco-ib draco-ib 
draco-ib draco-ib draco-ib draco-ib(1) 0.035u 0.017s 32:39.14 0.0% 0+0k 0+0io 
0pf+0w
   Summary of lapw1para:
   localhost k=0 user=0  wallclock=0
   draco-ib  k=0 user=0  wallclock=0
-  done at Thu Mar 12 15:01:00 JST 2015
-
-  starting Fermi on gemini.a04.aist.go.jp at Thu Mar 12 15:28:19 JST 2015
-  starting parallel lapw2c at Thu Mar 12 15:28:20 JST 2015
  localhost 389.940u 7.565s 0:36.04 1102.9% 0+0k 718416+253704io 0pf+0w
  localhost 347.944u 5.749s 0:31.67 1116.7% 0+0k 718528+199776io 0pf+0w
  draco-ib 0.029u 0.026s 0:33.86 0.1% 0+0k 8+0io 0pf+0w
  draco-ib 0.032u 0.020s 0:33.80 0.1% 0+0k 8+0io 0pf+0w
   Summary of lapw2para:
   localhost user=737.884wallclock=67.71
   draco-ib  user=0.061  wallclock=67.66
-  done at Thu Mar 12 15:28:57 JST 2015
-  starting sumpara 4 on gemini.a04.aist.go.jp at Thu Mar 12 15:28:58 JST 2015
-  done at Thu Mar 12 15:29:27 JST 2015
-
-  starting parallel hfc at Thu Mar 12 15:29:35 JST 2015
**  HF crashed at Thu Mar 12 15:29:40 JST 2015
**  check ERROR FILES!




DAYFILE

mats...@libra.a04.aist.go.jp:/usr/local/share/wien2k/Fons/aCGTcat aCGT.dayfile

Calculating aCGT in /usr/local/share/wien2k/Fons/aCGT
on gemini.a04.aist.go.jp with PID 46746
using WIEN2k_14.2 (Release 15/10/2014) in /home/matstud/Wien2K


start   (Thu Mar 12 11:22:09 JST 2015) with lapw0 (40/99 to go)

cycle 1 (Thu Mar 12 11:22:09 JST 2015)  (40/99 to go)

   lapw0 -grr -p   (11:22:09) starting parallel lapw0 at Thu Mar 12 
 11:22:09 JST 2015
 .machine0 : 12 processors
745.365u 3.350s 1:05.09 1150.2% 0+0k 144+796936io 0pf+0w
   lapw0 -p(11:23:14) starting parallel lapw0 at Thu Mar 12 11:23:15 JST 
 2015
 .machine0 : 12 processors
620.682u 2.669s 0:54.15 1151.1% 0+0k 40+203264io 0pf+0w
   lapw1-c (11:24:09) 20736.270u 146.444s 3:01:03.72 192.2%
 0+0k 11992+5913840io 0pf+0w
   lapw1  -p   -c  (14:25:13) starting parallel lapw1 at Thu Mar 12 
 14:25:13 JST 2015
-  starting parallel LAPW1 jobs at Thu Mar 12 14:25:13 JST 2015
running LAPW1 in parallel mode (using .machines)
4 number_of_parallel_jobs
 localhost localhost localhost localhost localhost localhost localhost 

[Wien] Problems with hybrid calculations: case.vectorhf_old file missing

2015-03-10 Thread Paul Fons
I have an update and some questions on hybrid calculations on a 96 atom 
cluster.  I am running my initial tests with two 24 core machines connected by 
Infiniband.  I have included 4 k-points using a 2x2x2 MP grid.  My .machines 
file is as below.

lapw0:localhost:12
1:localhost:12
1:localhost:12
1:draco-ib:12
1:draco-ib:12
granularity:1
extrafine:1


 I have done a conventional PBE calculation on the same cluster using the above 
.machines file and the calculation finished without errors in a few hours.  I 
then initialized a hf calculation using lapw_hf_lapw and specified the same 
2x2x2 grid.  I specified 770 bands in my case.inhf as I have 1526 electrons.  
The initialize ran without errors.  I then invoked the scf loop using “run_lapw 
-hf -p” using the same machines file.  The lapw0 and the initial part of the 
scf loop appears to have run without errors, but the calculation stopped on the 
second iteration of the SCF loop.  In particular, the second loop failed due to 
a missing file /home/matstud/WIENSCRATCH/aCGT.vectorhf_old STATUS: old”.  
Before I continue, I should add that the WIENSCRATCH environmental variable is 
correctly set and the directories on both machines exist.  I should add that of 
course the regular parallel PBE run ran without errors as well and I assumed it 
used the same scratch directories without error.  The file in question 
“aCGT.vectorhf_old” does not exist in either of the WIENSCRATCH directories nor 
does it exist in the home directory of the calculation.  The two nodes are 
gemini (localhost) and draco-ib (the infini-band connected second node).  The 
contents of the scratch directories on both nodes are listed below as well as 
the files with vector within the files on the project directory.   The current 
calculation only involves four k-points and I have been careful not to limit 
the number of MPI jobs to four.  I have done the calculation twice now with the 
same errors.  The first time, I tried it immediately after a successful PBE 
calculation while the second time I tried it after deleting all of the files in 
the WIENSCRATCH directory thinking it the file error could have been caused by 
filename confusion.  The end result was the same (reprinted below).  The run 
stops because the file aCGT.vectorhf_old cannot be found.  Are there any 
suggestions as to what I might try next to solve the problem?  Thanks in 
advance for any help.



The actual run output went as follows:

run_lapw -hf -p -in1new 2
 LAPW0 END
 LAPW0 END
 LAPW1 END
mv: cannot stat `aCGT.vector': No such file or directory
 LAPW1 END
 LAPW1 END
 LAPW1 END
 LAPW1 END
mv: cannot stat `aCGT.vectorhf_old': No such file or directory
 LAPW2 END
mv: cannot stat `aCGT.vector': No such file or directory
LAPW2 - FERMI; weighs written
 LAPW2 END
 LAPW2 END
 LAPW2 END
 LAPW2 END
 SUMPARA END
 CORE  END
OPEN FAILED

  Above message repeats for a total of 48 times
   
error with vector files

   stop error





vector files in the different directories.

On gemini (locahost) working directory

ls -l aCGT*vector*
-rw-rw-r-- 1 matstud matstud 0 Mar  8 19:18 aCGT.vectorhf


On gemini (locahost) WIENSCRATCH

mats...@gemini.a04.aist.go.jp:/usr/local/share/wien2k/Fons/aCGTls -l 
$HOME/WIENSCRATCH
total 3592696
-rw-rw-r-- 1 matstud matstud 2943235040 Mar  9 13:31 aCGT.vector
-rw-rw-r-- 1 matstud matstud  367792558 Mar  9 14:11 aCGT.vector_1
-rw-rw-r-- 1 matstud matstud  367877082 Mar  9 14:11 aCGT.vector_2


On draco-ib (remote host) WIENSCRATCH directory

mats...@gemini.a04.aist.go.jp:/usr/local/share/wien2k/Fons/aCGTssh draco-ib ls 
-l WIENSCRATCH
total 718780
-rw-r--r-- 1 matstud matstud 367646498 Mar  9 14:04 aCGT.vector_3
-rw-r--r-- 1 matstud matstud 368373958 Mar  9 14:04 aCGT.vector_4




DAYFILE

cat aCGT.dayfile 

Calculating aCGT in /usr/local/share/wien2k/Fons/aCGT
on gemini.a04.aist.go.jp with PID 45216
using WIEN2k_14.2 (Release 15/10/2014) in /home/matstud/Wien2K


start   (Mon Mar  9 10:27:54 JST 2015) with lapw0 (40/99 to go)

cycle 1 (Mon Mar  9 10:27:55 JST 2015)  (40/99 to go)

   lapw0 -grr -p   (10:27:55) starting parallel lapw0 at Mon Mar  9 
 10:27:55 JST 2015
 .machine0 : 12 processors
755.913u 3.546s 1:06.22 1146.8% 0+0k 184+796936io 0pf+0w
   lapw0 -p(10:29:01) starting parallel lapw0 at Mon Mar  9 10:29:01 JST 
 2015
 .machine0 : 12 processors
622.223u 2.856s 0:54.57 1145.4% 0+0k 48+203264io 0pf+0w
   lapw1-c (10:29:56) 20873.217u 161.505s 3:01:39.31 192.9%
 0+0k 14448+5913840io 0pf+0w
   lapw1  -p   -c  (13:31:36) starting parallel lapw1 at Mon Mar  9 
 13:31:36 JST 2015
-  starting parallel LAPW1 jobs at Mon Mar  9 13:31:36 JST 2015
running LAPW1 in parallel mode (using .machines)
4 number_of_parallel_jobs
 localhost localhost localhost localhost localhost localhost localhost 
localhost localhost localhost localhost localhost(1) 27361.552u 625.141s 
39:53.54 1169.2%0+0k 8+882304io 0pf+0w
 localhost localhost 

[Wien] Problems with hybrid calculations: case.vectorhf_old file missing

2015-03-10 Thread Paul Fons
I have an update and some questions on hybrid calculations on a 96 atom 
cluster.  I am running my initial tests with two 24 core machines connected by 
Infiniband.  I have included 4 k-points using a 2x2x2 MP grid.  My .machines 
file is as below.

lapw0:localhost:12
1:localhost:12
1:localhost:12
1:draco-ib:12
1:draco-ib:12
granularity:1
extrafine:1


 I have done a conventional PBE calculation on the same cluster using the above 
.machines file and the calculation finished without errors in a few hours.  I 
then initialized a hf calculation using lapw_hf_lapw and specified the same 
2x2x2 grid.  I specified 770 bands in my case.inhf as I have 1526 electrons.  
The initialize ran without errors.  I then invoked the scf loop using “run_lapw 
-hf -p” using the same machines file.  The lapw0 and the initial part of the 
scf loop appears to have run without errors, but the calculation stopped on the 
second iteration of the SCF loop.  In particular, the second loop failed due to 
a missing file /home/matstud/WIENSCRATCH/aCGT.vectorhf_old STATUS: old”.  
Before I continue, I should add that the WIENSCRATCH environmental variable is 
correctly set and the directories on both machines exist.  I should add that of 
course the regular parallel PBE run ran without errors as well and I assumed it 
used the same scratch directories without error.  The file in question 
“aCGT.vectorhf_old” does not exist in either of the WIENSCRATCH directories nor 
does it exist in the home directory of the calculation.  The two nodes are 
gemini (localhost) and draco-ib (the infini-band connected second node).  The 
contents of the scratch directories on both nodes are listed below as well as 
the files with vector within the files on the project directory.   The current 
calculation only involves four k-points.



The actual run output went as follows:

run_lapw -hf -p -in1new 2
 LAPW0 END
 LAPW0 END
 LAPW1 END
mv: cannot stat `aCGT.vector': No such file or directory
 LAPW1 END
 LAPW1 END
 LAPW1 END
 LAPW1 END
mv: cannot stat `aCGT.vectorhf_old': No such file or directory
 LAPW2 END
mv: cannot stat `aCGT.vector': No such file or directory
LAPW2 - FERMI; weighs written
 LAPW2 END
 LAPW2 END
 LAPW2 END
 LAPW2 END
 SUMPARA END
 CORE  END
OPEN FAILED
OPEN FAILED
OPEN FAILED
OPEN FAILED
OPEN FAILED
OPEN FAILED
OPEN FAILED
OPEN FAILED
OPEN FAILED
OPEN FAILED
OPEN FAILED
OPEN FAILED
OPEN FAILED
OPEN FAILED
OPEN FAILED
OPEN FAILED
OPEN FAILED
OPEN FAILED
OPEN FAILED
OPEN FAILED
OPEN FAILED
OPEN FAILED
OPEN FAILED
OPEN FAILED
OPEN FAILED
OPEN FAILED
OPEN FAILED
OPEN FAILED
OPEN FAILED
OPEN FAILED
OPEN FAILED
OPEN FAILED
OPEN FAILED
OPEN FAILED
OPEN FAILED
OPEN FAILED
OPEN FAILED
OPEN FAILED
OPEN FAILED
OPEN FAILED
OPEN FAILED
OPEN FAILED
OPEN FAILED
OPEN FAILED
OPEN FAILED
OPEN FAILED
OPEN FAILED
OPEN FAILED
error with vector files

   stop error





vector files in the different directories.

On gemini (locahost) working directory

ls -l aCGT*vector*
-rw-rw-r-- 1 matstud matstud 0 Mar  8 19:18 aCGT.vectorhf


On gemini (locahost) WIENSCRATCH

mats...@gemini.a04.aist.go.jp:/usr/local/share/wien2k/Fons/aCGTls -l 
$HOME/WIENSCRATCH
total 3592696
-rw-rw-r-- 1 matstud matstud 2943235040 Mar  9 13:31 aCGT.vector
-rw-rw-r-- 1 matstud matstud  367792558 Mar  9 14:11 aCGT.vector_1
-rw-rw-r-- 1 matstud matstud  367877082 Mar  9 14:11 aCGT.vector_2


On draco-ib (remote host) WIENSCRATCH directory

mats...@gemini.a04.aist.go.jp:/usr/local/share/wien2k/Fons/aCGTssh draco-ib ls 
-l WIENSCRATCH
total 718780
-rw-r--r-- 1 matstud matstud 367646498 Mar  9 14:04 aCGT.vector_3
-rw-r--r-- 1 matstud matstud 368373958 Mar  9 14:04 aCGT.vector_4




DAYFILE

cat aCGT.dayfile 

Calculating aCGT in /usr/local/share/wien2k/Fons/aCGT
on gemini.a04.aist.go.jp with PID 45216
using WIEN2k_14.2 (Release 15/10/2014) in /home/matstud/Wien2K


start   (Mon Mar  9 10:27:54 JST 2015) with lapw0 (40/99 to go)

cycle 1 (Mon Mar  9 10:27:55 JST 2015)  (40/99 to go)

   lapw0 -grr -p   (10:27:55) starting parallel lapw0 at Mon Mar  9 
 10:27:55 JST 2015
 .machine0 : 12 processors
755.913u 3.546s 1:06.22 1146.8% 0+0k 184+796936io 0pf+0w
   lapw0 -p(10:29:01) starting parallel lapw0 at Mon Mar  9 10:29:01 JST 
 2015
 .machine0 : 12 processors
622.223u 2.856s 0:54.57 1145.4% 0+0k 48+203264io 0pf+0w
   lapw1-c (10:29:56) 20873.217u 161.505s 3:01:39.31 192.9%
 0+0k 14448+5913840io 0pf+0w
   lapw1  -p   -c  (13:31:36) starting parallel lapw1 at Mon Mar  9 
 13:31:36 JST 2015
-  starting parallel LAPW1 jobs at Mon Mar  9 13:31:36 JST 2015
running LAPW1 in parallel mode (using .machines)
4 number_of_parallel_jobs
 localhost localhost localhost localhost localhost localhost localhost 
localhost localhost localhost localhost localhost(1) 27361.552u 625.141s 
39:53.54 1169.2%0+0k 8+882304io 0pf+0w
 localhost localhost localhost localhost localhost localhost localhost 
localhost localhost localhost localhost 

Re: [Wien] Problems with hybrid calculation

2015-03-03 Thread Paul Fons
Hi I did run init_hf_lapw and I saw no sign of errors upon running it, however, 
I ran it from the web interface the first time.  I just restarted the 
calculation using bash and have confirmed the ibz and fbz files exist.
The reason that I am interested in hybrid mode is that I actually have 96 
processors available with infiniband interconnects.  Is a 96 atom system too 
expensive to run with PBE0 even with 96 cores? How much time should I expect 
such a calculation to take?

Paul Fons 


 On Mar 3, 2015, at 4:56 PM, t...@theochem.tuwien.ac.at wrote:
 
 Hi,
 
 init_hf_lapw executes run_kgenhf_lapw which creates the files
 case.klist_ibz, case.klist_fbz, case.kgen_ibz and case.kgen_fbz.
 So, I don't understand why they are not present in your directory.
 Are you sure that init_hf_lapw ran without problems?
 
 Beside this, I really think that you should forget about running hybrid
 functionals for 96 atoms, in particular with this small number of
 processors (unless you are willing to wait until retirement).
 Depending on the system/properties that you are considering,
 the use of a GGA functional maybe give results which are fairly reliable.
 
 F. Tran
 
 On Tue, 3 Mar 2015, Paul Fons wrote:
 
 Hi All,
   I am trying to calculate the density of states of a small cluster of 96 
 atoms.  I am using the hybrid mpi/k-point mode.  The
 calculation ran to completion without problems for PBE.  My .machines file 
 is listed below.  As the cluster is amorphous and
 roughly cubic in shape, I used a 2x2x2 MP mesh.  I then attempted to use the 
 PBE0 hybrid functional.  After using the
 init_hf_lapw script, I started the SCF loop.  The calculation ran for 
 several hours and then stopped.  I have attached the STDOUT
 output below.  I am using intel ifort (15) as well as the intel mpi 
 environment.  I do have a scratch directory set up on this
 same node.
  These calculations were run on a single 24 core machine.  I have five of 
 these machines connected by Infiband and would like to
 scale up the calculation to use several nodes if I can find and solve the 
 hybrid problem as I am aware of the cost of the PBE0
 calculations.
 Any advice?
 Best wishes,
 Paul Fons
 .machines
 lapw0:localhost:24
 1:localhost:2
 1:localhost:2
 1:localhost:2
 1:localhost:2
 1:localhost:2
 1:localhost:2
 1:localhost:2
 1:localhost:2
 1:localhost:2
 1:localhost:2
 1:localhost:2
 1:localhost:2
 granularity:1
 extrafine:1
 STDOUT
  LAPW0 END
  LAPW0 END
 cp: cannot stat `aCGT.kgen_fbz': No such file or directory
 cp: cannot stat `aCGT.klist_fbz': No such file or directory
  LAPW1 END
 mv: cannot stat `aCGT.vector': No such file or directory
 cp: cannot stat `aCGT.kgen_ibz': No such file or directory
 cp: cannot stat `aCGT.klist_ibz': No such file or directory
  LAPW1 END
  LAPW1 END
  LAPW1 END
  LAPW1 END
  LAPW1 END
  LAPW1 END
  LAPW1 END
  LAPW1 END
  LAPW1 END
  LAPW1 END
  LAPW1 END
  LAPW1 END
  LAPW1 END
  LAPW1 END
  LAPW1 END
  LAPW1 END
  LAPW1 END
  LAPW1 END
  LAPW1 END
  LAPW1 END
 cp: cannot stat `aCGT.kgen_fbz': No such file or directory
 cp: cannot stat `aCGT.klist_fbz': No such file or directory
 mv: cannot stat `aCGT.vectorhf_old': No such file or directory
 forrtl: severe (24): end-of-file during read, unit 1001, file 
 /usr/local/share/wien2k/Fons/aCGT/031
 Image  PCRoutineLineSource   
   
 lapw2c 00648027  Unknown   Unknown  Unknown
 lapw2c 00669FF3  Unknown   Unknown  Unknown
 lapw2c 00482FEE  outp_ 180  outp.f
 lapw2c 004701EC  l2main_  2125  
 l2main_tmp_.F
 lapw2c 0047CEA5  MAIN__607  
 lapw2_tmp_.F
 lapw2c 0040407E  Unknown   Unknown  Unknown
 libc.so.6  00343F61ED5D  Unknown   Unknown  Unknown
 lapw2c 00403F89  Unknown   Unknown  Unknown
stop error
 ___
 Wien mailing list
 Wien@zeus.theochem.tuwien.ac.at
 http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien
 SEARCH the MAILING-LIST at:  
 http://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/index.html

Dr. Paul Fons
Senior Research Scientist
Functional Nano-phase-change Research Team
Nanoelectronics Research Institute
National Institute for Advanced Industrial Science  TechnologyMETI

AIST Central 4, Higashi 1-1-1Tsukuba, Ibaraki JAPAN 305-8568

tel. +81-298-61-5636fax. +81-298-61-2939

email: paul-f...@aist.go.jp mailto:paul-f...@aist.go.jp

The following lines are in a Japanese font

〒305-8562 茨城県つくば市つくば中央東 1-1-1
産業技術総合研究所
ナノエレクトロニクス研究部門
相変化新規機能デバイス研究チーム
上級主任研究員
ポール・フォンス






___
Wien mailing list
Wien@zeus.theochem.tuwien.ac.at
http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien
SEARCH the MAILING-LIST at:  
http://www.mail-archive.com

[Wien] Problems with parallel hybrid functional calculation

2015-03-03 Thread Paul Fons

Hi All,
  I am trying to run a hybrid functional calculation on a 96 atom system using 
PBE0.  I successfully used a PBE functional using the test machines file below 
that is a hybrid k-point/MPI parallel job.  While I posted a day ago about a 
failure due to missing ibz fbz files when I attempted to use the web interface, 
this time, I strictly used the terminal.  The output is shown below.  There 
were no problems with the fbz/ibz file generation, however, the run still 
terminated with an error.  The error was that case.vectorhf_old was not found.  
I am not sure where this file show have been generated or why it was not.  I 
expect this calculation to take some time to run and am planning on using 
96-cores for the real PBE0 calculation, but at the moment, I am in 
troubleshooting mode.  Does anyone have any idea of what might have gone wrong. 
 I used a shifted 2x2x2 MP grid and 770 bands as I have 1536 electrons.

Thanks,
 Paul Fons


(.machines file)

lapw0:localhost:24
1:localhost:2
1:localhost:2
1:localhost:2
1:localhost:2
1:localhost:2
1:localhost:2
1:localhost:2
1:localhost:2
1:localhost:2
1:localhost:2
1:localhost:2
1:localhost:2
#1:draco-ib:12
granularity:1
extrafine:1


TERMINAL OUTPUT

mats...@gemini.a04.aist.go.jp:/usr/local/share/wien2k/Fons/aCGTinit_hf_lapw
:GAP  : -9.   Ry = -.eV  ( metallic )
 Bandranges (emin - emax) and occupancy:
:BAN00759: 7590.3530280.368374  2.
:BAN00760: 7600.3540990.369025  2.
:BAN00761: 7610.3579440.370696  2.
:BAN00762: 7620.3607320.374369  2.
:BAN00763: 7630.3628900.377049  2.
:BAN00764: 7640.3668570.378864  2.
:BAN00765: 7650.3706710.382529  2.
:BAN00766: 7660.3743100.386296  2.
:BAN00767: 7670.3790370.388138  2.
:BAN00768: 7680.3830690.392822  1.61356106
:BAN00769: 7690.3875150.396096  0.38643894
:BAN00770: 7700.3926270.399987  0.
:BAN00771: 7710.3960490.406404  0.
:BAN00772: 7720.3996870.414211  0.
:BAN00773: 7730.4105640.422211  0.
:BAN00774: 7740.4191010.430560  0.
Energy to separate low and high energystates:   -0.09012


:NOE  : NUMBER OF ELECTRONS  =1536.000

:FER  : F E R M I - ENERGY(TETRAH.M.)=   0.3906577224
You must set NBAND to at least  NB_occ + 1  and you have 1536.00 electrons
edit aCGT.inhf …   (SET NUMBER OF BANDS TO 770)
PS: For very accurate calc. and large NBAND you may have to increase EMAX in 
aCGT.in1 by hand

Prepare k-mesh for HF. Use identical mesh and shift for IBZ and FBZ
Do you want to use a REDUCED k-mesh for HF (saving cpu-time) (Y/n) ? 
n
 This script runs   x kgen   twice and generates equivalent meshes in the
 IBZ and FBZ.



KGEN ENDS
KGEN ENDS
How many k-points in full BZ?
If you type 0 you can give 3 numbers for nx,ny,nz
0
How many in x?
2
How many in y?
2
How many in z?
2
Do you want to shift? (yes=1, no=0)
1
   1  symmetry operations without inversion
 inversion added (non-spinpolarized non-so calculation)
  NUMBER OF K-POINTS IN WHOLE CELL: (0 allows to specify 3 divisions of G)
 length of reciprocal lattice vectors:   0.246   0.245   0.247   0.000   0.000  
 0.000
  Specify 3 mesh-divisions (n1,n2,n3):
  Shift of k-mesh allowed. Do you want to shift: (0=no, 1=shift)
   4  k-points generated, ndiv=   2   2   2
KGEN ENDS
0.002u 0.008s 0:00.07 0.0%  0+0k 24+128io 0pf+0w
   1  symmetry operations without inversion
  NUMBER OF K-POINTS IN WHOLE CELL: (0 allows to specify 3 divisions of G)
 length of reciprocal lattice vectors:   0.246   0.245   0.247   0.000   0.000  
 0.000
  Specify 3 mesh-divisions (n1,n2,n3):
  Shift of k-mesh allowed. Do you want to shift: (0=no, 1=shift)
   8  k-points generated, ndiv=   2   2   2
KGEN ENDS
0.000u 0.008s 0:00.39 0.0%  0+0k 72+40io 0pf+0w
aCGT.in0_grr and aCGT.inhf and hf-kmesh prepared
Now do the hybrid calculation:   run_lapw -hf  ...  
mats...@gemini.a04.aist.go.jp:/usr/local/share/wien2k/Fons/aCGTrun_lapw -hf -p
 LAPW0 END
 LAPW0 END
 LAPW1 END
mv: cannot stat `aCGT.vector': No such file or directory
 LAPW1 END
 LAPW1 END
 LAPW1 END
 LAPW1 END
mv: cannot stat `aCGT.vectorhf_old': No such file or directory
forrtl: severe (24): end-of-file during read, unit 1001, file 
/usr/local/share/wien2k/Fons/aCGT/031
Image  PCRoutineLineSource  
   
lapw2c 00648027  Unknown   Unknown  Unknown
lapw2c 00669FF3  Unknown   Unknown  Unknown
lapw2c 00482FEE  outp_ 180  outp.f
lapw2c 004701EC  l2main_  2125  
l2main_tmp_.F
lapw2c 0047CEA5  MAIN__607  lapw2_tmp_.F
lapw2c

[Wien] Problems with hybrid calculation

2015-03-02 Thread Paul Fons
Hi All,

I am trying to calculate the density of states of a small 
cluster of 96 atoms.  I am using the hybrid mpi/k-point mode.  The calculation 
ran to completion without problems for PBE.  My .machines file is listed below. 
 As the cluster is amorphous and roughly cubic in shape, I used a 2x2x2 MP 
mesh.  I then attempted to use the PBE0 hybrid functional.  After using the 
init_hf_lapw script, I started the SCF loop.  The calculation ran for several 
hours and then stopped.  I have attached the STDOUT output below.  I am using 
intel ifort (15) as well as the intel mpi environment.  I do have a scratch 
directory set up on this same node.

 These calculations were run on a single 24 core machine.  I have five of these 
machines connected by Infiband and would like to scale up the calculation to 
use several nodes if I can find and solve the hybrid problem as I am aware of 
the cost of the PBE0 calculations.

Any advice?


Best wishes,
Paul Fons



.machines

lapw0:localhost:24
1:localhost:2
1:localhost:2
1:localhost:2
1:localhost:2
1:localhost:2
1:localhost:2
1:localhost:2
1:localhost:2
1:localhost:2
1:localhost:2
1:localhost:2
1:localhost:2
granularity:1
extrafine:1


STDOUT

 LAPW0 END
 LAPW0 END
cp: cannot stat `aCGT.kgen_fbz': No such file or directory
cp: cannot stat `aCGT.klist_fbz': No such file or directory
 LAPW1 END
mv: cannot stat `aCGT.vector': No such file or directory
cp: cannot stat `aCGT.kgen_ibz': No such file or directory
cp: cannot stat `aCGT.klist_ibz': No such file or directory
 LAPW1 END
 LAPW1 END
 LAPW1 END
 LAPW1 END
 LAPW1 END
 LAPW1 END
 LAPW1 END
 LAPW1 END
 LAPW1 END
 LAPW1 END
 LAPW1 END
 LAPW1 END
 LAPW1 END
 LAPW1 END
 LAPW1 END
 LAPW1 END
 LAPW1 END
 LAPW1 END
 LAPW1 END
 LAPW1 END
cp: cannot stat `aCGT.kgen_fbz': No such file or directory
cp: cannot stat `aCGT.klist_fbz': No such file or directory
mv: cannot stat `aCGT.vectorhf_old': No such file or directory
forrtl: severe (24): end-of-file during read, unit 1001, file 
/usr/local/share/wien2k/Fons/aCGT/031
Image  PCRoutineLineSource  
   
lapw2c 00648027  Unknown   Unknown  Unknown
lapw2c 00669FF3  Unknown   Unknown  Unknown
lapw2c 00482FEE  outp_ 180  outp.f
lapw2c 004701EC  l2main_  2125  
l2main_tmp_.F
lapw2c 0047CEA5  MAIN__607  lapw2_tmp_.F
lapw2c 0040407E  Unknown   Unknown  Unknown
libc.so.6  00343F61ED5D  Unknown   Unknown  Unknown
lapw2c 00403F89  Unknown   Unknown  Unknown

   stop error___
Wien mailing list
Wien@zeus.theochem.tuwien.ac.at
http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien
SEARCH the MAILING-LIST at:  
http://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/index.html


Re: [Wien] QTL-B error for Zn K-edge ELNES

2014-08-27 Thread Paul Fons
Dear Peter,

Hello from Japan.
Thank you very much for your comment.  I have always employed supercells, but 
the sign of the compensating charge was more subtle.  In truth, for the MgO 
supercell example, I tried both -1 and +1 in case.inm, but I didn't see much 
different in the XSPEC spectra.


Paul
On Aug 27, 2014, at 16:11, Peter Blaha pbl...@theochem.tuwien.ac.at wrote:

 About core-hole calculations for XSPEC or TELNES:
 
 1) You need to make a sizable supercell !!! Maybe start with 16 atoms/cell, 
 but 64 or more are even better. It is WRONG to make a core-hole calculation 
 with just 2 Zn atoms/cell. Why: Consider reality: You excite only VERY few 
 atoms in such a measurements (actually VERY few even means 1 out of 10^10 or 
 less!!!), the majority of atoms remain in the ground state and can contribute 
 to the screening. This is, what we want to simulate.
 
 2) We need to have charge neutrality, i.e. the sum of nuclear charges and 
 electronic charges must cancel. By removing 1 e- from one atom in the cell, 
 you need to add this charge somewhere. 2 possibilities:
 a) add it as valence electron (increase NE in case.in2; but don't forget to 
 remove this e- AFTER the scf cycle and before calculation the spectra)
 b) add a background charge in case.inm.
 MSR1  -1.0  YES  (BROYD/PRATT, extra charge (+1 for additional e), norm)
 As the comment says, you should set this value to +1, when you have added an 
 e-. For core holes, we have REMOVED an electron, thus we must set this to   
 -1.0   !!!
 You SHOULD ALWAYS CHECK your charge neutrality using
 grep :NEC01 case.scf.
 
 Now which method is better, a) or b) ???
 In many cases there is NOT much difference, though in some cases there could 
 be a difference.
 Again, remember we want to simulate experiment as close as possible:
 the Zn 1s electron will be excited (dipole rule) into Zn-4p electrons. The 
 electronic structure of ZnO has mainly Zn-s and p character at the conduction 
 band minimum, thus when we use method b), we actually add this e- into Zn-s+p 
 states, which is approximately correct. However, since the Zn-4sp states 
 are very delocalized states with little weight inside the Zn-sphere, I don't 
 think it matters too much when you put the charge into the background.
 An opposite example would be eg. CeO2, which has very localized Ce-4f states 
 just above EF. When simulating O-K edges, it is NOT a good idea to add the e- 
 into the valence, because it will go into LOCALIZED Ce-4f states and not 
 contribute significantly to screen the core-hole on oxygen. The method b) 
 would be better in this case. Of course, for Ce-M edges obviously method a) 
 should be much better.
 
 Hope this clarifies the unclear discussion on the mailing list.
 
 Peter Blaha
 
 ---
 Subject: Re: [Wien] QTL-B error for Zn K-edge ELNES
 From: Paul Fons paul-f...@aist.go.jp
 Date: 08/27/2014 06:37 AM
 To: A Mailing list for WIEN2k users wien@zeus.theochem.tuwien.ac.at
 Hi Kevin,
Are you sure you are correct for using -1 in case.inm.  I can reproduce 
 the example from the Wien2K course (MgO supercell) just fine when I use a 
 +1 in case.inm.  The documentation states that 1 adds another electron.  
 From what I understand this is to compensate for the electron that is removed 
 when a core hole is generated by changing the occupancy in case.inc.  For the 
 case of MgO.inc
 
 1 0.00  0  NUMBER OF ORBITALS (EXCLUDING SPIN), SHIFT, IPRINT
 1,-1,2   ( N,KAPPA,OCCUP)
 1 0.00  0  NUMBER OF ORBITALS (EXCLUDING SPIN), SHIFT, IPRINT
 1,-1,2   ( N,KAPPA,OCCUP)
 1 0.00  0  NUMBER OF ORBITALS (EXCLUDING SPIN), SHIFT, IPRINT
 1,-1,2   ( N,KAPPA,OCCUP)
 1 0.00  0  NUMBER OF ORBITALS (EXCLUDING SPIN), SHIFT, IPRINT
 1,-1,2   ( N,KAPPA,OCCUP)
 1 0.00  0  NUMBER OF ORBITALS (EXCLUDING SPIN), SHIFT, IPRINT
 1,-1,2   ( N,KAPPA,OCCUP)
 0
 
 
 I changed the second line to 1,-1,1 to make a core hole.  I then compensated 
 for the missing charge with case.inm to add a +1 background charge via
 
 MSR1  1.0   YES  (BROYD/PRATT, extra charge (+1 for additional e), norm)
 0.20mixing FACTOR for BROYD/PRATT scheme
 1.00  1.00  PW and CLM-scaling factors
   8 idum, HISTORY
 
 After this, I run a SCF calculation.  When it is done, I then remove the 
 charge from case.inm and carry out a XSPEC calculation.  This is what I 
 learned in the Wien2K course a couple of years ago when the developers had a 
 seminar here in Japan.  As I mentioned, the above procedure reproduces the 
 MgO core hole present plot in the exercise notes.  I would be very much 
 interested in hearing if I am doing something incorrect.
 
 Cheers,
 Paul
 
 
 
 
 -- 
 
  P.Blaha
 --
 Peter BLAHA, Inst.f. Materials Chemistry, TU Vienna, A-1060 Vienna
 Phone: +43-1-58801-165300 FAX: +43-1

Re: [Wien] Parity determination by group theory analysis question

2014-06-05 Thread Paul Fons
Hi,
  Thank you for your note.  I am most grateful for some help on the group 
theory question as there seem to be very few people in general who answer such 
questions.  I should have mentioned I realize in retrospect that the 
calculation was carried out with spin-orbit coupling.  The degeneracy in the 
spin is presumably due to the presence of inversion symmetry in the crystal 
structure (the crystal structure does have inversion symmetry).  Would it be 
fair to conclude that this is a case of accidental degeneracy and that there 
are two sets of states (GM5+ + GM6+) and (GM5- + GM6-) and that the two pairs 
of states have parity 1 and -1 respectively?  


Best wishes,
  Paul
On Jun 4, 2014, at 18:08, Juan Manuel Perez Mato wmppe...@lg.ehu.es wrote:

 Hi,
 
 (GM5+ + GM6+ ) and (GM5- + GM6-) form two different physically irreducible 
 representations of even and odd parity of D3d (as double group). Therefore 
 the degeneracy of the bands corresponding to these two physically irreps is 
 not forced by symmetry. Consequently the four-fold degeneracy of the band 
 most probably comes from not considering the spin. I would bet that it will 
 disappear if spin orbit coupling is introduced, and they will then split into 
 two separate two-fold degenerate bands associated with the two different 
 physically irreducible representations of different parity.
 
 regards,
 
 J. Manuel Perez-Mato
 Fac. Ciencia y Tecnologia,
 Universidad del Pais Vasco, UPV
 48080 BILBAO,
 Spain
 
 tel. +34 946012473
 fax. +34 946013500
 ***
 
 
 
 El 04/06/2014, a las 09:29, Paul Fons escribió:
 
 Hi,
  I have been using the irreducible representations calculated by irrep 
 to determine the parity of the wavefunction for particular k-points for each 
 band.  A typical case is shown below for the gamma point.As I am solving 
 the system without spin, each band is at least two-fold degenerate.  For 
 most bands, for example the first band, the situation is clear as the irrep 
 is G4+ that the parity is +.  This also follows from the character of the 
 (I) class being the same dimension as the irrep.  I am a little more 
 confused by what to do a band such as band 9 where there are two irreps 
 spanning the same band, but presumably as the character of the (I) class is 
 the same as that of the dimension of the irrep (and that both irreps are of 
 the same parity, +) that the parity is again +.  The situation I am writing 
 to ask about, however is that of band 63 which is four-fold degenerate and 
 is composed of irreps of both parities.  How does one interpret the parity 
 from the information below for band 63?  Thanks for any advice.
 
 
 
 
The point group is D3d
12 symmetry operations in  6 classes
Table 55   on page  58 in Koster  et al [7]
Table 42.4 on page 371 in Altmann et al [8]
 
E   2C3   3C2 I  2IC3  3IC2   

G1+   A1g   1 1 1 1 1 1  
G2+   A2g   1 1-1 1 1-1  
G3+   Eg2-1 0 2-1 0  
G1-   A1u   1 1 1-1-1-1  
G2-   A2u   1 1-1-1-1 1  
G3-   Eu2-1 0-2 1 0  

G4+   E1/2g 2 1 0 2 1 0  
G5+  1E3/2g 1-1 i 1-1 i  
G6+  2E3/2g 1-1-i 1-1-i  
G4-   E1/2u 2 1 0-2-1 0  
G5-  1E3/2u 1-1 i-1 1-i  
G6-  2E3/2u 1-1-i-1 1 i  
 
 
 class, symmetry ops, exp(-i*k*taui)
E   10 (+1.00+0.00i)
  2C32  7  (+1.00+0.00i)
  3C21  8  9   (+1.00+0.00i)
I3 (+1.00+0.00i)
 2IC36 11  (+1.00+0.00i)
 3IC24  5 12   (+1.00+0.00i)
 
 bnd ndg  eigval E 2C3 3C2   I2IC3
 3IC2   
   1  2 -7.239676 2.00+0.00i  1.00+0.00i  0.00-0.00i  2.00+0.00i  1.00-0.00i  
 0.00-0.00i =G4+ 
   3  2 -7.239533 2.00+0.00i  1.00+0.00i -0.00+0.00i -2.00+0.00i -1.00-0.00i  
 0.00-0.00i =G4- 
   5  2 -6.641446 2.00+0.00i  1.00+0.00i  0.00+0.00i  2.00-0.00i  1.00+0.00i  
 0.00+0.00i =G4+ 
   7  2 -6.641393 2.00-0.00i  1.00-0.00i -0.00-0.00i -2.00-0.00i -1.00+0.00i  
 0.00-0.00i =G4- 
   9  2 -6.641277 2.00+0.00i -2.00-0.00i -0.00+0.00i  2.00+0.00i -2.00-0.00i 
 -0.00-0.00i =G5+ + G6+
 
 
 
  63  4 -1.880979 4.00-0.00i -4.00+0.00i  0.00-0.00i  0.00-0.00i -0.00+0.00i 
 -0.00+0.00i =G5+ + G6+ + G5- + G6-  
  
 ___
 Wien mailing list
 Wien@zeus.theochem.tuwien.ac.at
 http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien
 SEARCH the MAILING-LIST at:  
 http://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/index.html

[Wien] Parity determination by group theory analysis question

2014-06-04 Thread Paul Fons
Hi,
I have been using the irreducible representations calculated by irrep 
to determine the parity of the wavefunction for particular k-points for each 
band.  A typical case is shown below for the gamma point.As I am solving 
the system without spin, each band is at least two-fold degenerate.  For most 
bands, for example the first band, the situation is clear as the irrep is G4+ 
that the parity is +.  This also follows from the character of the (I) class 
being the same dimension as the irrep.  I am a little more confused by what to 
do a band such as band 9 where there are two irreps spanning the same band, but 
presumably as the character of the (I) class is the same as that of the 
dimension of the irrep (and that both irreps are of the same parity, +) that 
the parity is again +.  The situation I am writing to ask about, however is 
that of band 63 which is four-fold degenerate and is composed of irreps of both 
parities.  How does one interpret the parity from the information below for 
band 63?  Thanks for any advice.




   The point group is D3d
   12 symmetry operations in  6 classes
   Table 55   on page  58 in Koster  et al [7]
   Table 42.4 on page 371 in Altmann et al [8]

   E   2C3   3C2 I  2IC3  3IC2  

   G1+   A1g   1 1 1 1 1 1  
   G2+   A2g   1 1-1 1 1-1  
   G3+   Eg2-1 0 2-1 0  
   G1-   A1u   1 1 1-1-1-1  
   G2-   A2u   1 1-1-1-1 1  
   G3-   Eu2-1 0-2 1 0  
   
   G4+   E1/2g 2 1 0 2 1 0  
   G5+  1E3/2g 1-1 i 1-1 i  
   G6+  2E3/2g 1-1-i 1-1-i  
   G4-   E1/2u 2 1 0-2-1 0  
   G5-  1E3/2u 1-1 i-1 1-i  
   G6-  2E3/2u 1-1-i-1 1 i  


class, symmetry ops, exp(-i*k*taui)
   E   10 (+1.00+0.00i)
 2C32  7  (+1.00+0.00i)
 3C21  8  9   (+1.00+0.00i)
   I3 (+1.00+0.00i)
2IC36 11  (+1.00+0.00i)
3IC24  5 12   (+1.00+0.00i)

bnd ndg  eigval E 2C3 3C2   I2IC3
3IC2   
  1  2 -7.239676 2.00+0.00i  1.00+0.00i  0.00-0.00i  2.00+0.00i  1.00-0.00i  
0.00-0.00i =G4+ 
  3  2 -7.239533 2.00+0.00i  1.00+0.00i -0.00+0.00i -2.00+0.00i -1.00-0.00i  
0.00-0.00i =G4- 
  5  2 -6.641446 2.00+0.00i  1.00+0.00i  0.00+0.00i  2.00-0.00i  1.00+0.00i  
0.00+0.00i =G4+ 
  7  2 -6.641393 2.00-0.00i  1.00-0.00i -0.00-0.00i -2.00-0.00i -1.00+0.00i  
0.00-0.00i =G4- 
  9  2 -6.641277 2.00+0.00i -2.00-0.00i -0.00+0.00i  2.00+0.00i -2.00-0.00i 
-0.00-0.00i =G5+ + G6+



 63  4 -1.880979 4.00-0.00i -4.00+0.00i  0.00-0.00i  0.00-0.00i -0.00+0.00i 
-0.00+0.00i =G5+ + G6+ + G5- + G6-  
 ___
Wien mailing list
Wien@zeus.theochem.tuwien.ac.at
http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien
SEARCH the MAILING-LIST at:  
http://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/index.html


[Wien] Problems with mpi for Wien12.1

2012-08-29 Thread Paul Fons
I compiled fftw3 using the Intel suite as well.  The appropriate line from 
config.log reads

./configure CC=icc F77=ifort MPICC=mpiicc --prefix=/opt/local --enable-mpi 
--enable-threads --prefix=/opt/local/fftw3

I note that the configuration file only calls for a mpicc compiler (and I used 
the Intel compiler) and not a fortran compiler.   The compiled code (mpi-bench 
does work fine with the Intel mpirun).


After commenting out the call W2kinit subroutine and recompiling lapw0 (via the 
siteconfig script), I attempted to run run_lapw in both serial and parallel 
forms as you can see below.  The serial form worked fine

Paul

matstud at ursa:~/WienDisk/Fons/GaAs run_lapw
 LAPW0 END
 LAPW1 END
 LAPW2 END
 CORE  END
 MIXER END
ec cc and fc_conv 1 1 1

   stop


matstud at ursa:~/WienDisk/Fons/GaAs run_lapw -p
forrtl: severe (174): SIGSEGV, segmentation fault occurred
forrtl: severe (174): SIGSEGV, segmentation fault occurred
forrtl: severe (174): SIGSEGV, segmentation fault occurred
forrtl: severe (174): SIGSEGV, segmentation fault occurred
forrtl: severe (174): SIGSEGV, segmentation fault occurred
forrtl: severe (174): SIGSEGV, segmentation fault occurred
forrtl: severe (174): SIGSEGV, segmentation fault occurred
forrtl: severe (174): SIGSEGV, segmentation fault occurred
forrtl: severe (174): SIGSEGV, segmentation fault occurred
forrtl: severe (174): SIGSEGV, segmentation fault occurred
forrtl: severe (174): SIGSEGV, segmentation fault occurred
forrtl: severe (174): SIGSEGV, segmentation fault occurred

   stop error



On Aug 28, 2012, at 9:16 AM, Laurence Marks wrote:

 One suggestion: comment out the line towards the top of lapw0.F
 
  call W2kinit
 
 You should get a more human readable error message.
 
 As an addendum, was fftw3 compiled with mpiifort? I assume from your email 
 that it was, just checking.
 
 N.B., there is a small chance that this will hang your computer.
 
 ___
 Wien mailing list
 Wien at zeus.theochem.tuwien.ac.at
 http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien

Dr. Paul Fons
Senior Research Scientist
Functional Nano-phase-change Research Team
Nanoelectronics Research Institute
National Institute for Advanced Industrial Science  Technology
METI

AIST Central 4, Higashi 1-1-1
Tsukuba, Ibaraki JAPAN 305-8568

tel. +81-298-61-5636
fax. +81-298-61-2939

email: paul-fons at aist.go.jp

The following lines are in a Japanese font

?305-8562 ? 1-1-1
?
??

?






-- next part --
An HTML attachment was scrubbed...
URL: 
http://zeus.theochem.tuwien.ac.at/pipermail/wien/attachments/20120829/98a836c2/attachment.htm


[Wien] Problems with mpi for Wien12.1

2012-08-28 Thread Paul Fons
   if it points to the proper 
 directory/version of mpi,
 
 Am 24.08.2012 15:35, schrieb Laurence Marks:
 In my experience the SIGSEV normally comes from mixing different flavors of 
 mpif90 and mpirun. Openmpi, mpich2 and Intels mpi all need different 
 versions of blacs. You can also
 have problems if you choose the wrong model for integers in the linking 
 advisor page. I would check using ldd that lapw0_mpi is linked to the right 
 version, and that the default
 versions are correct (e.g. which mpirun). Often you can minimize problems by 
 using static linking for mpi.
 
 N.B. The contact developers message is a relic of when some code was added 
 for fault handlers and to eliminate issues with limits that used to be 
 pervasive. It should probably
 be removed.
 
 ---
 Professor Laurence Marks
 Department of Materials Science and Engineering
 Northwestern University
 www.numis.northwestern.edu http://www.numis.northwestern.edu 1-847-491-3996
 Research is to see what everybody else has seen, and to think what nobody 
 else has thought
 Albert Szent-Gyorgi
 
 On Aug 24, 2012 8:22 AM, Paul Fons paul-fons at aist.go.jp 
 mailto:paul-fons at aist.go.jp wrote:
 
Dear Prof. Blaha,
Thank you for your earlier email.  Running the command manually gives the 
 following output (for a GaAs structure that works fine in serial or k-point 
 parallel form).  I am
still not sure what to try next.  Any suggestions?
 
matstud at ursa:~/WienDisk/Fons/GaAs mpirun -np 4 ${WIENROOT}/lapw0_mpi 
 lapw0.def
w2k_dispatch_signal(): received: Segmentation fault
w2k_dispatch_signal(): received: Segmentation fault
w2k_dispatch_signal(): received: Segmentation fault
w2k_dispatch_signal(): received: Segmentation fault
  Child id   0 SIGSEGV, contact developers
  Child id   1 SIGSEGV, contact developers
  Child id   3 SIGSEGV, contact developers
  Child id   2 SIGSEGV, contact developers
application called MPI_Abort(MPI_COMM_WORLD, 1) - process 3
application called MPI_Abort(MPI_COMM_WORLD, 1) - process 2
application called MPI_Abort(MPI_COMM_WORLD, 1) - process 1
application called MPI_Abort(MPI_COMM_WORLD, 1) - process 0
APPLICATION TERMINATED WITH THE EXIT STRING: Hangup (signal 1)
 
 
The MPI compilation options from siteconfig are as follows: (the settings 
 are from the Intel MKL link advisor plus the fftw3 library)
 
  Current settings:
  RP  RP_LIB(SCALAPACK+PBLAS): -L$(MKLROOT)/lib/intel64 
 $(MKLROOT)/lib/intel64/libmkl_blas95_lp64.a 
 $(MKLROOT)/lib/intel64/libmkl_lapack95_lp64.a -lmkl_scalapack_lp64
-lmkl_cdft_core -lmkl_intel_lp64 -lmkl_intel_thread -lmkl_core 
 -lmkl_blacs_intelmpi_lp64 -openmp -lpthread -lm -L/opt/local/fftw3/lib/ 
 -lfftw3_mpi -lfftw3 $(R_LIBS)
  FP  FPOPT(par.comp.options): -I$(MKLROOT)/include/intel64/lp64 
 -I$(MKLROOT)/include -FR -mp1 -w -prec_div -pc80 -pad -ip -DINTEL_VML 
 -DFFTW3 -traceback
  MP  MPIRUN commando: mpirun -np _NP_ -machinefile _HOSTS_ 
 _EXEC_
 
The file parallel_options now reads
setenv USE_REMOTE 1
setenv MPI_REMOTE 0
setenv WIEN_GRANULARITY 1
setenv WIEN_MPIRUN mpirun -np _NP_ -machinefile _HOSTS_ _EXEC_
 
 
I changed the MPI_REMOTE to 0 as suggested (I was not sure this applied 
 to the Intel MPI environment as the siteconfig prompt only mentioned mich2.
 
As I mentioned the mpirun command seems to work fine.  For example, the 
 fftw3 benchmark program gives with 24 processes
 
mpirun -np 24 ./mpi-bench 1024x1024
Problem: 1024x1024, setup: 126.32 ms, time: 15.98 ms, ``mflops'': 6562.2
 
 
 
On Aug 24, 2012, at 3:05 PM, Peter Blaha wrote:
 
Hard to say.
 
What is in $WIENROOT/parallel_options ?
MPI_REMOTE should be 0 !
 
Otherwise run lapw0_mpi by hand:
 
mpirun -np 4 $WIENROOT/lapw0_mpi lapw0.def   (or including  .machinefile 
 .machine0)
 
 
Am 24.08.2012 02:24, schrieb Paul Fons:
Greetings all,
  I have compiled Wien2K 12.1 under OpenSuse 11.4 (and OpenSuse 12.1)
and the latest Intel compilers with identical mpi launch problems and I
am hoping for some suggestions as to where to look to fix things.  Note
that the serial and k-point parallel versions of the code run fine (I
have optimized GaAs a lot in my troubleshooting!).
 
Environment.
 
I am using the latest intel fort, icc, and impi libraries for linux.
 
matstud at pyxis:~/Wien2K ifort --version
ifort (IFORT) 12.1.5 20120612
Copyright (C) 1985-2012 Intel Corporation.  All rights reserved.
 
matstud at pyxis:~/Wien2K mpirun --version
Intel(R) MPI Library for Linux* OS, Version 4.0 Update 3 Build 20110824
Copyright (C) 2003-2011, Intel Corporation. All rights reserved.
 
matstud at pyxis:~/Wien2K icc --version
icc (ICC) 12.1.5 20120612
Copyright (C) 1985-2012 Intel Corporation.  All rights reserved.
 
 
My OPTIONS files

[Wien] Problems with mpi for Wien12.1

2012-08-24 Thread Paul Fons
Greetings all,

  I have compiled Wien2K 12.1 under OpenSuse 11.4 (and OpenSuse 12.1) and the 
latest Intel compilers with identical mpi launch problems and I am hoping for 
some suggestions as to where to look to fix things.  Note that the serial and 
k-point parallel versions of the code run fine (I have optimized GaAs a lot in 
my troubleshooting!).

Environment.

I am using the latest intel fort, icc, and impi libraries for linux.

matstud at pyxis:~/Wien2K ifort --version
ifort (IFORT) 12.1.5 20120612
Copyright (C) 1985-2012 Intel Corporation.  All rights reserved.

matstud at pyxis:~/Wien2K mpirun --version
Intel(R) MPI Library for Linux* OS, Version 4.0 Update 3 Build 20110824
Copyright (C) 2003-2011, Intel Corporation. All rights reserved.

matstud at pyxis:~/Wien2K icc --version
icc (ICC) 12.1.5 20120612
Copyright (C) 1985-2012 Intel Corporation.  All rights reserved.


My OPTIONS files from /siteconfig_lapw

current:FOPT:-FR -mp1 -w -prec_div -pc80 -pad -ip -DINTEL_VML -traceback
current:FPOPT:-I$(MKLROOT)/include/intel64/lp64 -I$(MKLROOT)/include -FR -mp1 
-w -prec_div -pc80 -pad -ip -DINTEL_VML -DFFTW3 -traceback
current:LDFLAGS:$(FOPT) -L$(MKLROOT)/lib/$(MKL_TARGET_ARCH) -pthread
current:DPARALLEL:'-DParallel'
current:R_LIBS:-lmkl_lapack95_lp64 -lmkl_intel_lp64 -lmkl_intel_thread 
-lmkl_core -openmp -lpthread
current:RP_LIBS:-L$(MKLROOT)/lib/intel64 
$(MKLROOT)/lib/intel64/libmkl_blas95_lp64.a 
$(MKLROOT)/lib/intel64/libmkl_lapack95_lp64.a -lmkl_scalapack_lp64 
-lmkl_cdft_core -lmkl_intel_lp64 -lmkl_intel_thread -lmkl_core 
-lmkl_blacs_intelmpi_lp64 -openmp -lpthread -lm -L/opt/local/fftw3/lib/ 
-lfftw3_mpi -lfftw3 $(R_LIBS)
current:MPIRUN:mpirun -np _NP_ -machinefile _HOSTS_ _EXEC_




The code compiles and links without error.  It runs fine in serial mode and in 
k-point parallel mode, e.g.

.machines with

1:localhost
1:localhost
1:localhost
granularity:1
extrafine:1

This runs fine.  When I attempt to run a mpi process with 12 processes (on a 12 
core machine), I crash and burn (see below) with a SIGSEV error with 
instructions to contact the developers.

The linking options were derived from Intel's mkl link advisor (the version on 
the intel site.  I should add that the mpi-bench in fftw3 works fine using the 
intel mpi as do commands like hostname or even abinit so it would appear that 
that the Intel MPI environment itself is fine.  I have wasted a lot of time 
trying to figure out how to fix this before writing to the list, but at this 
point, I feel like a monkey at a keyboard attempting to duplicate Shakesphere 
-- if you know what I mean.  Thanks in advance for any heads up that you can 
offer.



.machines

lapw0:localhost:12
1:localhost:12
granularity:1
extrafine:1

   stop error

error: command   /home/matstud/Wien2K/lapw0para -c lapw0.def   failed
0.029u 0.046s 0:00.93 6.4%  0+0k 0+176io 0pf+0w
 Child id   2 SIGSEGV, contact developers
 Child id   8 SIGSEGV, contact developers
 Child id   7 SIGSEGV, contact developers
 Child id  11 SIGSEGV, contact developers
 Child id  10 SIGSEGV, contact developers
 Child id   9 SIGSEGV, contact developers
 Child id   6 SIGSEGV, contact developers
 Child id   5 SIGSEGV, contact developers
 Child id   4 SIGSEGV, contact developers
 Child id   3 SIGSEGV, contact developers
 Child id   1 SIGSEGV, contact developers
 Child id   0 SIGSEGV, contact developers
 .machine0 : 12 processors
   lapw0 -p(09:04:45) starting parallel lapw0 at Fri Aug 24 09:04:45 JST 
 2012

cycle 1 (Fri Aug 24 09:04:45 JST 2012)  (40/99 to go)

start   (Fri Aug 24 09:04:45 JST 2012) with lapw0 (40/99 to go)


using WIEN2k_12.1 (Release 22/7/2012) in /home/matstud/Wien2K
on pyxis with PID 15375
Calculating GaAs in /usr/local/share/Wien2K/Fons/GaAs

-- next part --
An HTML attachment was scrubbed...
URL: 
http://zeus.theochem.tuwien.ac.at/pipermail/wien/attachments/20120824/1d183071/attachment.htm


[Wien] Problems with mpi for Wien12.1

2012-08-24 Thread Paul Fons
Dear Prof. Blaha,
Thank you for your earlier email.  Running the command manually gives the 
following output (for a GaAs structure that works fine in serial or k-point 
parallel form).  I am still not sure what to try next.  Any suggestions?
 

matstud at ursa:~/WienDisk/Fons/GaAs mpirun -np 4 ${WIENROOT}/lapw0_mpi 
lapw0.def
w2k_dispatch_signal(): received: Segmentation fault
w2k_dispatch_signal(): received: Segmentation fault
w2k_dispatch_signal(): received: Segmentation fault
w2k_dispatch_signal(): received: Segmentation fault
 Child id   0 SIGSEGV, contact developers
 Child id   1 SIGSEGV, contact developers
 Child id   3 SIGSEGV, contact developers
 Child id   2 SIGSEGV, contact developers
application called MPI_Abort(MPI_COMM_WORLD, 1) - process 3
application called MPI_Abort(MPI_COMM_WORLD, 1) - process 2
application called MPI_Abort(MPI_COMM_WORLD, 1) - process 1
application called MPI_Abort(MPI_COMM_WORLD, 1) - process 0
APPLICATION TERMINATED WITH THE EXIT STRING: Hangup (signal 1)


The MPI compilation options from siteconfig are as follows: (the settings are 
from the Intel MKL link advisor plus the fftw3 library)

 Current settings:
 RP  RP_LIB(SCALAPACK+PBLAS): -L$(MKLROOT)/lib/intel64 
$(MKLROOT)/lib/intel64/libmkl_blas95_lp64.a 
$(MKLROOT)/lib/intel64/libmkl_lapack95_lp64.a -lmkl_scalapack_lp64 
-lmkl_cdft_core -lmkl_intel_lp64 -lmkl_intel_thread -lmkl_core 
-lmkl_blacs_intelmpi_lp64 -openmp -lpthread -lm -L/opt/local/fftw3/lib/ 
-lfftw3_mpi -lfftw3 $(R_LIBS)
 FP  FPOPT(par.comp.options): -I$(MKLROOT)/include/intel64/lp64 
-I$(MKLROOT)/include -FR -mp1 -w -prec_div -pc80 -pad -ip -DINTEL_VML -DFFTW3 
-traceback
 MP  MPIRUN commando: mpirun -np _NP_ -machinefile _HOSTS_ _EXEC_

The file parallel_options now reads
setenv USE_REMOTE 1
setenv MPI_REMOTE 0
setenv WIEN_GRANULARITY 1
setenv WIEN_MPIRUN mpirun -np _NP_ -machinefile _HOSTS_ _EXEC_


I changed the MPI_REMOTE to 0 as suggested (I was not sure this applied to the 
Intel MPI environment as the siteconfig prompt only mentioned mich2.

As I mentioned the mpirun command seems to work fine.  For example, the fftw3 
benchmark program gives with 24 processes

mpirun -np 24 ./mpi-bench 1024x1024
Problem: 1024x1024, setup: 126.32 ms, time: 15.98 ms, ``mflops'': 6562.2



On Aug 24, 2012, at 3:05 PM, Peter Blaha wrote:

 Hard to say.
 
 What is in $WIENROOT/parallel_options ?
 MPI_REMOTE should be 0 !
 
 Otherwise run lapw0_mpi by hand:
 
 mpirun -np 4 $WIENROOT/lapw0_mpi lapw0.def   (or including  .machinefile 
 .machine0)
 
 
 Am 24.08.2012 02:24, schrieb Paul Fons:
 Greetings all,
   I have compiled Wien2K 12.1 under OpenSuse 11.4 (and OpenSuse 12.1)
 and the latest Intel compilers with identical mpi launch problems and I
 am hoping for some suggestions as to where to look to fix things.  Note
 that the serial and k-point parallel versions of the code run fine (I
 have optimized GaAs a lot in my troubleshooting!).
 
 Environment.
 
 I am using the latest intel fort, icc, and impi libraries for linux.
 
 matstud at pyxis:~/Wien2K ifort --version
 ifort (IFORT) 12.1.5 20120612
 Copyright (C) 1985-2012 Intel Corporation.  All rights reserved.
 
 matstud at pyxis:~/Wien2K mpirun --version
 Intel(R) MPI Library for Linux* OS, Version 4.0 Update 3 Build 20110824
 Copyright (C) 2003-2011, Intel Corporation. All rights reserved.
 
 matstud at pyxis:~/Wien2K icc --version
 icc (ICC) 12.1.5 20120612
 Copyright (C) 1985-2012 Intel Corporation.  All rights reserved.
 
 
 My OPTIONS files from /siteconfig_lapw
 
 current:FOPT:-FR -mp1 -w -prec_div -pc80 -pad -ip -DINTEL_VML -traceback
 current:FPOPT:-I$(MKLROOT)/include/intel64/lp64 -I$(MKLROOT)/include -FR
 -mp1 -w -prec_div -pc80 -pad -ip -DINTEL_VML -DFFTW3 -traceback
 current:LDFLAGS:$(FOPT) -L$(MKLROOT)/lib/$(MKL_TARGET_ARCH) -pthread
 current:DPARALLEL:'-DParallel'
 current:R_LIBS:-lmkl_lapack95_lp64 -lmkl_intel_lp64 -lmkl_intel_thread
 -lmkl_core -openmp -lpthread
 current:RP_LIBS:-L$(MKLROOT)/lib/intel64
 $(MKLROOT)/lib/intel64/libmkl_blas95_lp64.a
 $(MKLROOT)/lib/intel64/libmkl_lapack95_lp64.a -lmkl_scalapack_lp64
 -lmkl_cdft_core -lmkl_intel_lp64 -lmkl_intel_thread -lmkl_core
 -lmkl_blacs_intelmpi_lp64 -openmp -lpthread -lm -L/opt/local/fftw3/lib/
 -lfftw3_mpi -lfftw3 $(R_LIBS)
 current:MPIRUN:mpirun -np _NP_ -machinefile _HOSTS_ _EXEC_
 
 
 
 
 The code compiles and links without error.  It runs fine in serial mode
 and in k-point parallel mode, e.g.
 
 .machines with
 
 1:localhost
 1:localhost
 1:localhost
 granularity:1
 extrafine:1
 
 This runs fine.  When I attempt to run a mpi process with 12 processes
 (on a 12 core machine), I crash and burn (see below) with a SIGSEV error
 with instructions to contact the developers.
 
 The linking options were derived from Intel's mkl link advisor (the
 version on the intel site.  I should add that the mpi-bench in fftw3
 works fine using the intel mpi as do commands like hostname or even
 abinit

[Wien] Electric Field Effects

2012-08-08 Thread Paul Fons
Thank you for your quick help with my electric field question.  I have a couple 
of other related questions that I am still puzzled about.

The first question is the applicability of the Berry phase approach.  From what 
I understand, the Berry phase approach is usually couched within the density 
functional perturbation theory paradigm, e.g. in the limit of small E fields.  
In fact, in the paper you co-authored with Stahn [PRB v63, 165205 (2001)] you 
actually mentioned the Berry Phase. 

   For a quantum-mechanical description of 'polarization in a crystal' a 
geometrical Berry's phase approach as been introduced in recent years.  
Unfortunately, this approach is only valid if there is no external electric 
field in the crystal, which means that the surfaces of the crystal are short 
circuited.

As the fields we are interested in are fairly large (a factor of ten below the 
dielectric breakdown voltage), I was not confident about the applicability of a 
perturbative Berry's phase approach.   It was for this reason, I thought that 
your supercell approach (with the introduction of a vacuum slab at the kink and 
the fixing of the end atoms to prevent surface relaxation) would be a good 
approach that could be used even for moderately strong fields (0.01-0.1 
V/Angstrom = 0.2-2 mRy).  I also noted that in the same paper above, it was 
stated that calculations at artificially larger fields were necessary to ensure 
enough numerical precision.  In the paper values of 200 times the experimental 
field values were used (e.g. 1400 kV/mm = 14 V/Angstrom = 2 mRy) to ensure that 
there was enough numerical precision in the answer as for numerical reasons, 
the accuracy of the calculated potential Vint is of the order of the maximum 
potential applied.  Is this still the case for the latest revision of the Wien 
code (Wien12)?

Let me close, by thanking you again for your time.  I hope my questions are not 
too naive and can help others as well.  I hope to see you in Tokyo next month.

Best Wishes,
Paul Fons



On Aug 6, 2012, at 11:59 PM, Peter Blaha wrote:

 It depends a bit on what you are looking for.
 We were interested in GaAs and the relative change of positions of Ga and As 
 atoms
 under the E-filed.
 For this question such large cells are necessary, because in this setup we 
 had 2 kinks
 and the relaxations near the kink are very much influenced by them.
 
 In fact, nowadays I would do it differently, namely exactly as you have in 
 mind:
 
 In a slab geometry (surface) one can make this kink in the vacuum region and 
 we have now
 even asymmetric fields, so that you can put the strong variation into the 
 vacuum.
 
 In any case, as far as I remember it was a tedious work.
 
 Better approaches are now done using a Berry-phase method and as far as I 
 know, Oleg Rubel
 (rubelo at tbh.net) has such an approach programmed into WIEN2k.
 
 Am 06.08.2012 16:38, schrieb Paul Fons:
 Hi All,
  I have been studying the work of Stahn, Pietsch, Blaha, and Schwartz on 
 electric field induced charge density variations (in GaAs) [PRB v63, 165205 
 (2001).  I had a couple of quick questions.  In the paper a large (22 
 formula units) of GaAs is used and the field is inverted halfway across the 
 cell in the z-direction (e.g. to maintain the periodic boundary conditions). 
  I am curious to know if such a large number of repeats is necessary or is 
 it simply necessary to have the field applied over a minimum spatial length. 
  First, what is the case for GaAs, can the number of repeats be varied 
 without large changes in the result?  The second question has more to do 
 with the sort of systems I wish to calculate E-field-induced effects upon.  
 They are already quite large in size with about 20 layers in a supercell and 
 I would like to avoid making the cell unnecessarily large.  Would it make 
 sense to make a vacuum layer between two units cells (and leave the 
 electrostatic inflection!
 po
  int there).  I assume I would have to fix the positions of the atoms at the 
 end of the cell next to the vacuum layer if I did this.  Does this make 
 sense?  Any advice on what cell sizes would be appropriate would be 
 gratefully received.
 
  Paul Fons
 ___
 Wien mailing list
 Wien at zeus.theochem.tuwien.ac.at
 http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien
 
 
 -- 
 
  P.Blaha
 --
 Peter BLAHA, Inst.f. Materials Chemistry, TU Vienna, A-1060 Vienna
 Phone: +43-1-58801-165300 FAX: +43-1-58801-165982
 Email: blaha at theochem.tuwien.ac.atWWW: 
 http://info.tuwien.ac.at/theochem/
 --
 ___
 Wien mailing list
 Wien at zeus.theochem.tuwien.ac.at
 http://zeus.theochem.tuwien.ac.at

[Wien] Spin Texture

2012-08-07 Thread Paul Fons
Dear All,
I am trying to reproduce the results of some spin texture calculations 
on topological insulators in the literature, specifically the work of Basak et 
al in PRB84, 121401 (2011).  They calculated the spin texture (the helical 
nature) of the in-plane spin components in reciprocal space at the Fermi level 
(surface) using Wien2K.  I have reproduced the slab calculations and can see 
the surface bands, however, I am unsure how to go about calculating the 
expectation values for the spin in reciprocal space.  I did note the presence 
of a density matrix routine LAPWDM  in section 7.7 of the user guide which 
allows for the calculation of expectation values including spin, but only 
apparently in real space within the atomic spheres.  Any guidance as to how to 
go about calculating a spin texture map (e.g. the projected spin direction on 
the 2D fermi surface in reciprocal space of the above topologically insulating 
structure) would be greatly appreciated.  I have surveyed the literature, but 
there are no details of how the spin texture maps were calculated in any of the 
papers I have read.  Thanks for any help in advance.


Paul Fons
-- next part --
An HTML attachment was scrubbed...
URL: 
http://zeus.theochem.tuwien.ac.at/pipermail/wien/attachments/20120806/4e5f5e43/attachment.htm


[Wien] Electric Field Effects

2012-08-07 Thread Paul Fons
Hi All,
I have been studying the work of Stahn, Pietsch, Blaha, and Schwartz on 
electric field induced charge density variations (in GaAs) [PRB v63, 165205 
(2001).  I had a couple of quick questions.  In the paper a large (22 formula 
units) of GaAs is used and the field is inverted halfway across the cell in the 
z-direction (e.g. to maintain the periodic boundary conditions).  I am curious 
to know if such a large number of repeats is necessary or is it simply 
necessary to have the field applied over a minimum spatial length.  First, what 
is the case for GaAs, can the number of repeats be varied without large changes 
in the result?  The second question has more to do with the sort of systems I 
wish to calculate E-field-induced effects upon.  They are already quite large 
in size with about 20 layers in a supercell and I would like to avoid making 
the cell unnecessarily large.  Would it make sense to make a vacuum layer 
between two units cells (and leave the electrostatic inflection point there).  
I assume I would have to fix the positions of the atoms at the end of the cell 
next to the vacuum layer if I did this.  Does this make sense?  Any advice on 
what cell sizes would be appropriate would be gratefully received.

Paul Fons


[Wien] Spin texture in reciprocal space

2012-08-03 Thread Paul Fons
Dear All,
I am trying to reproduce the results of some spin texture calculations 
on topological insulators in the literature, specifically the work of Basak et 
al in PRB84, 121401 (2011) on Bi2T3.  They calculated the spin texture (the 
helical nature) of the in-plane spin components in reciprocal space at the 
Fermi level (surface) using Wien2K.  I have reproduced the slab calculations 
and can see the surface bands so so-far so good. However, I am unsure how to go 
about calculating the expectation values for the spin in reciprocal space, e.g. 
what is called spin texture (I must admit I have some additional questions 
about quantization axes as well due to the lack of commutation among Sx, Sy, 
and Sz.  The plot of the spin texture in the paper showed for the 2D projected 
surface state, a ring of arrows following a locus of points described by a 
circle -- e.g. it demonstrated that the spin was helical.
  I did note the presence of a density matrix routine LAPWDM  in section 7.7 of 
the user guide which allows for the calculation of expectation values including 
spin, but only apparently in real space and only within the atomic spheres.  
Any guidance as to how to go about calculating a spin texture map (e.g. the 
projected spin direction on the 2D fermi surface in reciprocal space of the 
above topologically insulating structure) would be greatly appreciated.  I have 
surveyed the literature, but there are no details of how the spin texture maps 
were calculated in any of the papers I have read.  Thanks for any help in 
advance.
-- next part --
An HTML attachment was scrubbed...
URL: 
http://zeus.theochem.tuwien.ac.at/pipermail/wien/attachments/20120803/3da166f7/attachment.htm


[Wien] Spin texture in reciprocal space

2012-08-03 Thread Paul Fons
Dear All,
I am trying to reproduce the results of some spin texture calculations 
on topological insulators in the literature, specifically the work of Basak et 
al in PRB84, 121401 (2011).  They calculated the spin texture (the helical 
nature) of the in-plane spin components in reciprocal space at the Fermi level 
(surface) using Wien2K.  I have reproduced the slab calculations and can see 
the surface bands, however, I am unsure how to go about calculating the 
expectation values for the spin in reciprocal space.  I did note the presence 
of a density matrix routine LAPWDM  in section 7.7 of the user guide which 
allows for the calculation of expectation values including spin, but only 
apparently in real space within the atomic spheres.  Any guidance as to how to 
go about calculating a spin texture map (e.g. the projected spin direction on 
the 2D fermi surface in reciprocal space of the above topologically insulating 
structure) would be greatly appreciated.  I have surveyed the literature, but 
there are no details of how the spin texture maps were calculated in any of the 
papers I have read.  Thanks for any help in advance.
-- next part --
An HTML attachment was scrubbed...
URL: 
http://zeus.theochem.tuwien.ac.at/pipermail/wien/attachments/20120803/2c89230b/attachment.htm


[Wien] Spin texture in reciprocal space

2012-08-02 Thread Paul Fons
Dear All,
I am trying to reproduce the results of some spin texture calculations 
on topological insulators in the literature, specifically the work of Basak et 
al in PRB84, 121401 (2011).  They calculated the spin texture (the helical 
nature) of the in-plane spin components in reciprocal space at the Fermi level 
(surface) using Wien2K.  I have reproduced the slab calculations and can see 
the surface bands, however, I am unsure how to go about calculating the 
expectation values for the spin in reciprocal space.  I did note the presence 
of a density matrix routine LAPWDM  in section 7.7 of the user guide which 
allows for the calculation of expectation values including spin, but only 
apparently in real space within the atomic spheres.  Any guidance as to how to 
go about calculating a spin texture map (e.g. the projected spin direction on 
the 2D fermi surface in reciprocal space of the above topologically insulating 
structure) would be greatly appreciated.  I have surveyed the literature, but 
there are no details of how the spin texture maps were calculated in any of the 
papers I have read.  Thanks for any help in advance.

Paul Fons
-- next part --
An HTML attachment was scrubbed...
URL: 
http://zeus.theochem.tuwien.ac.at/pipermail/wien/attachments/20120802/ea4b0d94/attachment.htm


[Wien] Spin texture in reciprocal space

2012-08-01 Thread Paul Fons
Dear All,
I am trying to reproduce the results of some spin texture calculations 
on topological insulators in the literature, specifically the work of Basak et 
al in PRB84, 121401 (2011).  They calculated the spin texture (the helical 
nature) of the in-plane spin components in reciprocal space at the Fermi level 
(surface) using Wien2K.  I have reproduced the slab calculations and can see 
the surface bands, however, I am unsure how to go about calculating the 
expectation values for the spin in reciprocal space.  I did note the presence 
of a density matrix routine LAPWDM  in section 7.7 of the user guide which 
allows for the calculation of expectation values including spin, but only 
apparently in real space within the atomic spheres.  Any guidance as to how to 
go about calculating a spin texture map (e.g. the projected spin direction on 
the 2D fermi surface in reciprocal space of the above topologically insulating 
structure) would be greatly appreciated.  I have surveyed the literature, but 
there are no details of how the spin texture maps were calculated in any of the 
papers I have read.  Thanks for any help in advance.
-- next part --
An HTML attachment was scrubbed...
URL: 
http://zeus.theochem.tuwien.ac.at/pipermail/wien/attachments/20120801/54359b85/attachment.htm


[Wien] Run Wien2k on Mac OS X Lion

2012-06-04 Thread Paul Fons
The HFS+ file system is historically case insensitive.  There is an option in 
Disk Utility to format a disk with a case sensitive file system (e.g. a 
traditional unix-like style).While this will not cause any problems for the 
operating system, it could in principle cause problems for other Mac software 
in that some developers may assume that case differences are indistinguishable, 
however, they will be distinguishable for the case sensitive system.

Best Wishes,
Paul Fons



On Jun 2, 2012, at 11:28 PM, Marcelo Barbosa wrote:

 Hi Gilles
 
 Sorry for such a late late response.
 
 How can i format the disk with case sensitive option? Mine doesn't seem to 
 be, because doing cd Desktop and cd desktop is exactly the same thing.
 
 As far as i know, Macs usually use .profile instead of .bashrc (at least the 
 later ones), however i created a .bashrc file and i call it from .profile.
 Can this complicate anything?
 
 Cheers,
 Marcelo
 
 On Apr 4, 2012, at 6:20 AM, Gilles Hug wrote:
 
 Dear Marcelo,
 I have Wien2k running on the Mac although I don't use it very often.
 I have not installed Lion.
 You should first check that the disk is formatted with case sensitive 
 (Unix-like) option.
 Then try the Wien2k programs and scripts from the Terminal.
 One problem maybe to find the runtime libraries which are located in a 
 different place compared to other Linux styles, i. e. you need to check that 
 the environment variables in the .bashrc are correct. This changes from 
 versions of OSX and Intel.
 Bon courage,
 Gllles
 
 Le 4 avr. 2012 ? 06:58, Peter Blaha a ?crit :
 
 I've no experience with Macs myself, but I know that some Mac-experts have 
 wien2k running on Macs.
 
 When you say, the scripts are not working, you may not have a   csh (or 
 tcsh).
 
 
 Am 03.04.2012 13:10, schrieb Marcelo Barbosa:
 Hello
 
 I'm currently using Wien2k on an Ubuntu machine and it's working fine, but 
 i would like to run it on Mac OS X 10.7 Lion.
 Is there anyone who could help me to compile and run Wien2k on a Mac OS X 
 machine?
 
 I have installed the Intel compilers for mac and the compilation runs 
 without any errors.
 I run w2web on terminal and it starts a session with no problem, but then 
 on the browser i'm only able to create the .struct file, none of the 
 scripts are working (including the RMT calculation when creating the 
 .struct file).
 
 Can anyone help me?
 
 Best regards,
 Marcelo Barbosa
 ___
 Wien mailing list
 Wien at zeus.theochem.tuwien.ac.at
 http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien
 
 -- 
 -
 Peter Blaha
 Inst. Materials Chemistry, TU Vienna
 Getreidemarkt 9, A-1060 Vienna, Austria
 Tel: +43-1-5880115671
 Fax: +43-1-5880115698
 email: pblaha at theochem.tuwien.ac.at
 -
 ___
 Wien mailing list
 Wien at zeus.theochem.tuwien.ac.at
 http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien
 
 ___
 Wien mailing list
 Wien at zeus.theochem.tuwien.ac.at
 http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien
 
 ___
 Wien mailing list
 Wien at zeus.theochem.tuwien.ac.at
 http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien

Dr. Paul Fons
Senior Research Scientist
Functional Nano-phase-change Research Team
Nanoelectronics Research Institute
National Institute for Advanced Industrial Science  Technology
METI

AIST Central 4, Higashi 1-1-1
Tsukuba, Ibaraki JAPAN 305-8568

tel. +81-298-61-5636
fax. +81-298-61-2939

email: paul-fons at aist.go.jp

The following lines are in a Japanese font

?305-8562 ? 1-1-1
?
??

?






-- next part --
An HTML attachment was scrubbed...
URL: 
http://zeus.theochem.tuwien.ac.at/pipermail/wien/attachments/20120604/8e338f46/attachment.htm


[Wien] MPI Problem

2012-01-23 Thread Paul Fons

Hi,
I have Wien2K running on a cluster of linux boxes each with 32 cores 
and connected by 10Gb ethernet.  I have compiled Wien2K by the 3.174 version of 
Wien2K (I learned the hard way that bugs in the newer versions of the Intel 
compiler lead to crashes in Wien2K).  I have also installed Intel's MPI.  
First, the single process Wien2K, let's say for the TiC case, works fine.  It 
also works fine when I use a .machines file like

granulaity:1
localhost:1
localhost:1
?  (24 times).

This file leads to parallel execution without error.  I can vary the number of 
processes by increasing the number of localhost:1 and the number of localhost:1 
lines in the file and still everything works fine.  When I try to use mpi to 
communicate with one process, it works as well.

1:localhost:1  

 lstarting parallel lapw1 at Mon Jan 23 06:49:16 JST 2012
 -  starting parallel LAPW1 jobs at Mon Jan 23 06:49:16 JST 2012
 running LAPW1 in parallel mode (using .machines)
 1 number_of_parallel_jobs
 [1] 22417
  LAPW1 END
 [1]  + Done  ( cd $PWD; $t $exe ${def}_$loop.def; rm 
 -f .lock_$lockfile[$p] )  .time1_$loop
  localhost(111) 179.004u 4.635s 0:32.73 561.0%0+0k 0+26392io 0pf+0w
Summary of lapw1para:
localhost   k=111   user=179.004wallclock=32.73
 179.167u 4.791s 0:35.61 516.5%0+0k 0+26624io 0pf+0w
 

Changing the machine file to use more than one process  (the same form of error 
occurs for more than 2)

1:localhost:2

lead to a run time error in the MPI subsystem.

 starting parallel lapw1 at Mon Jan 23 06:51:04 JST 2012
 -  starting parallel LAPW1 jobs at Mon Jan 23 06:51:04 JST 2012
 running LAPW1 in parallel mode (using .machines)
 1 number_of_parallel_jobs
 [1] 22673
 Fatal error in MPI_Comm_size: Invalid communicator, error stack:
 MPI_Comm_size(123): MPI_Comm_size(comm=0x5b, size=0x7ed20c) failed
 MPI_Comm_size(76).: Invalid communicator
 Fatal error in MPI_Comm_size: Invalid communicator, error stack:
 MPI_Comm_size(123): MPI_Comm_size(comm=0x5b, size=0x7ed20c) failed
 MPI_Comm_size(76).: Invalid communicator
 [1]  + Done  ( cd $PWD; $t $ttt; rm -f 
 .lock_$lockfile[$p] )  .time1_$loop
  localhost localhost(111) APPLICATION TERMINATED WITH THE EXIT STRING: 
 Hangup (signal 1)
 0.037u 0.036s 0:00.06 100.0%  0+0k 0+0io 0pf+0w
 TiC.scf1_1: No such file or directory.
Summary of lapw1para:
localhost   k=0 user=111wallclock=0
 0.105u 0.168s 0:03.21 8.0%0+0k 0+216io 0pf+0w

I have properly sourced the appropriate runtime environment for the Intel 
system.  For example, compiling (mpiifort) and running the f90 mpi test program 
from intel produces:



 mpirun -np 32 /home/paulfons/mpitest/testf90
  Hello world: rank0  of   32  running on 
  asccmp177
   
  
  Hello world: rank1  of   32  running on(32 times)

Does anyone have any suggestions as to what to try next?  I am not sure how to 
debug things from here.  I have about 512 nodes that I can use for larger 
calculations that only can be accessed by mpi (the ssh setup works fine as well 
by the way).  It would be great to figure out what is wrong.

Thanks.

















Dr. Paul Fons
Functional Nano-phase-change Research Team
Team Leader
Nanodevice Innovation Research Center (NIRC)
National Institute for Advanced Industrial Science  Technology
METI

AIST Central 4, Higashi 1-1-1
Tsukuba, Ibaraki JAPAN 305-8568

tel. +81-298-61-5636
fax. +81-298-61-2939

email: paul-fons at aist.go.jp

The following lines are in a Japanese font

?305-8562 ? 1-1-1
?
??






-- next part --
An HTML attachment was scrubbed...
URL: 
http://zeus.theochem.tuwien.ac.at/pipermail/wien/attachments/20120123/a608980e/attachment.htm


[Wien] MPI Problem

2012-01-23 Thread Paul Fons
Thank you very much for your suggestion.  I actually managed to figure this out 
by myself an hour or so ago.  At the same time (usually not a good idea) I also 
compiled the mkl interface for fftw2 rather than using the standalone version I 
had compiled by myself earlier.  Thus the RP library line looks like

-lmkl_scalapack_lp64 -lmkl_solver_lp64 -lmkl_blacs_intelmpi_lp64 
/opt/intel/mkl/lib/intel64/libfftw2x_cdft_DOUBLE.a $(R_LIBS)

All compiles 

In the past ten minutes I have been running the TiC example under mpi and I 
must say the mpi version seems tricker to use than I first though as each mpi 
process seems to use multiple threads and it is easy to over saturate the 
number of cores one has.  I have been testing the mpi version on a single node 
(32 cores) of a 500 core cluster to get a feel for things.

Running the TiC calculation (with 4000 k-points) in serial mode (the mkl 
routine uses multiple threads) took about 75 seconds

Running the sample calculation in parallel mode (with 5 mpi processes each 
which spawn multiple threads for an average cpu load of 28 in a 32 core 
machine) takes longer, specifically 300 seconds.

Carrying out the same calculation with 1:localhost* on six separate lines 
(e.g. lapw runs six processes locally and then aggregates without using mpi) 
finishes quickest of all in about 60 seconds.


I don't really understand what is going on.  Obviously I don't really need to 
use the mpi/or parallel calculation for small problems as in the TiC example, 
but I would like to use it for larger supercells.  Is there some sort of 
guiding principle which will help me decide which approach is best?  I assume 
the parallelization by running multiple jobs is also limited to a single 
machine so this won't scale very well.

Thanks for any advice.

Paul

On Jan 23, 2012, at 8:08 AM, Laurence Marks wrote:

 A guess: you are using the wrong version of blacs. You need a
 -lmkl_blacs_intelmpi_XX
 where XX is the one for your system. I have seen this give the same error.
 
 Use http://software.intel.com/en-us/articles/intel-mkl-link-line-advisor/
 
 For reference, with openmpi it is _openmpi_ instead of _intelmpi_, and
 similarly for sgi.
 
 2012/1/22 Paul Fons paul-fons at aist.go.jp:
 
 Hi,
 I have Wien2K running on a cluster of linux boxes each with 32 cores and
 connected by 10Gb ethernet.  I have compiled Wien2K by the 3.174 version of
 Wien2K (I learned the hard way that bugs in the newer versions of the Intel
 compiler lead to crashes in Wien2K).  I have also installed Intel's MPI.
  First, the single process Wien2K, let's say for the TiC case, works fine.
  It also works fine when I use a .machines file like
 
 granulaity:1
 localhost:1
 localhost:1
 ?  (24 times).
 
 This file leads to parallel execution without error.  I can vary the number
 of processes by increasing the number of localhost:1 and the number of
 localhost:1 lines in the file and still everything works fine.  When I try
 to use mpi to communicate with one process, it works as well.
 
 1:localhost:1
 
 lstarting parallel lapw1 at Mon Jan 23 06:49:16 JST 2012
 
 -  starting parallel LAPW1 jobs at Mon Jan 23 06:49:16 JST 2012
 running LAPW1 in parallel mode (using .machines)
 1 number_of_parallel_jobs
 [1] 22417
 LAPW1 END
 [1]  + Done  ( cd $PWD; $t $exe ${def}_$loop.def; rm
 -f .lock_$lockfile[$p] )  .time1_$loop
 localhost(111) 179.004u 4.635s 0:32.73 561.0%0+0k 0+26392io 0pf+0w
   Summary of lapw1para:
   localhost   k=111   user=179.004wallclock=32.73
 179.167u 4.791s 0:35.61 516.5%   0+0k 0+26624io 0pf+0w
 
 
 Changing the machine file to use more than one process  (the same form of
 error occurs for more than 2)
 
 1:localhost:2
 
 lead to a run time error in the MPI subsystem.
 
 starting parallel lapw1 at Mon Jan 23 06:51:04 JST 2012
 -  starting parallel LAPW1 jobs at Mon Jan 23 06:51:04 JST 2012
 running LAPW1 in parallel mode (using .machines)
 1 number_of_parallel_jobs
 [1] 22673
 Fatal error in MPI_Comm_size: Invalid communicator, error stack:
 MPI_Comm_size(123): MPI_Comm_size(comm=0x5b, size=0x7ed20c) failed
 MPI_Comm_size(76).: Invalid communicator
 Fatal error in MPI_Comm_size: Invalid communicator, error stack:
 MPI_Comm_size(123): MPI_Comm_size(comm=0x5b, size=0x7ed20c) failed
 MPI_Comm_size(76).: Invalid communicator
 [1]  + Done  ( cd $PWD; $t $ttt; rm -f
 .lock_$lockfile[$p] )  .time1_$loop
 localhost localhost(111) APPLICATION TERMINATED WITH THE EXIT STRING:
 Hangup (signal 1)
 0.037u 0.036s 0:00.06 100.0% 0+0k 0+0io 0pf+0w
 TiC.scf1_1: No such file or directory.
   Summary of lapw1para:
   localhost   k=0 user=111wallclock=0
 0.105u 0.168s 0:03.21 8.0%   0+0k 0+216io 0pf+0w
 
 
 I have properly sourced the appropriate runtime environment for the Intel
 system.  For example, compiling (mpiifort) and running the f90 mpi test
 program from intel produces:
 
 
 
 mpirun -np 32 /home/paulfons/mpitest

[Wien] WIEN2k on Xeon 5500 Series

2011-11-24 Thread Paul Fons
I am trying to compile Wien2K as well with the latest intel compilers.  At 
first I thought it might be a stack overflow problem (of course limit stacksize 
was set to unlimited) and I tried the suggestion -heap-arrays from intel 
(http://software.intel.com/en-us/articles/determining-root-cause-of-sigsegv-or-sigbus-errors/)
 but it did not help either.  

Has any progress been made in tracking down the source of this error?  The 
traceback is essentially worthless.  Has anyone corresponded with intel with 
regards to the problem?


I am now in the process of downloading an older version of the compiler.  I am 
a little worried about downloading having different versions of the compiler as 
there is also C code in Wien2K.  Exactly which versions of the C, Fortran, and 
MKL libraries should be used.  I have a copy of a Intel Parallel Studio XE for 
Linux and don't typically download the individual packages hence my confusion.