Re: [Wien] lapw0 hang with mbj calculation for slab

2013-10-18 Thread Peter Blaha

Switch to PBE for one iteration.
Increase  IFFTfactor from 2 to 4 in case.in0 (and, most likely, reduce RKMAX)

Am 18.10.2013 05:55, schrieb alpa dashora:

Dear Prof. Blaha and Wien2k users,

I am trying to run mbj calculation for MoS2 10 layer slab. After few cycles 
lapw0 hanged. I have seen the earlier posts discussed on the same topics. I 
have used the
vxclm2.f file as provided by Prof. Blaha and removed case.in0_grr file. I have 
also increased the Rmt*Kmax up to 9 but still I am not able to solve my problem.

Please suggest any other solution for this error or any other method to 
calculate the correct band gap.

Thanks in advance.

With kind regards,

--
Alpa Dashora


___
Wien mailing list
Wien@zeus.theochem.tuwien.ac.at
http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien
SEARCH the MAILING-LIST at:  
http://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/index.html



--
-
Peter Blaha
Inst. Materials Chemistry, TU Vienna
Getreidemarkt 9, A-1060 Vienna, Austria
Tel: +43-1-5880115671
Fax: +43-1-5880115698
email: pbl...@theochem.tuwien.ac.at
-
___
Wien mailing list
Wien@zeus.theochem.tuwien.ac.at
http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien
SEARCH the MAILING-LIST at:  
http://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/index.html


Re: [Wien] lapw0 hang with mbj calculation for slab

2013-10-18 Thread tran

Hi,

Most likely the problem is due to the Newton iteration procedure in brj.f
which does not stop. Replace this subroutine by the one that I attached
(it is based on an analytical representation of the solution).

By the way, it is nonsense to calculate the average of grad rho/rho when
there is vacuum. So, fix the value of c to the one obtained for the bulk.

F. Tran

On Fri, 18 Oct 2013, alpa dashora wrote:


Dear Prof. Blaha and Wien2k users,
I am trying to run mbj calculation for MoS2 10 layer slab. After few cycles 
lapw0 hanged. I have seen the earlier posts discussed on the same topics. I 
have used the vxclm2.f file as
provided by Prof. Blaha and removed case.in0_grr file. I have also increased 
the Rmt*Kmax up to 9 but still I am not able to solve my problem.

Please suggest any other solution for this error or any other method to 
calculate the correct band gap.

Thanks in advance.

With kind regards,
 
--
Alpa Dashora

  subroutine brj(rho,grho,g2rho,tau,vxbrj,ir)

!A. D. Becke and M. R. Roussel, Phys. Rev. A 39, 3761 (1989).
!A. D. Becke and E. R. Johnson, J. Chem. Phys. 124, 221101 (2006).
!E. Proynov, Z. Gan, and J. Kong, Chem. Phys. Lett. 455, 103 (2008).

  use xcparam 

  implicit real*8(a-h,o-z)

  real*8 :: a(1:3), b(0:5), c(0:5), d(0:5), e(0:5), yp(0:5)

  save iint,isphere
  data iint/0/,isphere/0/
  pi = 4d0*atan(1d0)

  vxbrj = 0d0

  if (rho .gt. 1d-18) then

 tautf = (3d0/10d0)*(3d0*pi**2)**(2d0/3d0)*(2d0*rho)**(5d0/3d0)
 tauw = 0.125d0*grho*grho*2.d0/rho

 if (tau.lt.tauw) then
tau_falsch=tau
tau=tauw
 endif

 if (tau.eq.tauw .and. rho.lt.10.d0.and.ir.lt.900.and.isphere.eq.0) then
print*,'sphere:rho,tauw,grho,g2rho',rho,tau,grho,g2rho,'tauwrong=',tau_falsch
isphere=1
 endif

 if (tau.eq.tauw .and. ir.gt.900.and.iint.lt.10) then
print*,'int:rho,tauw,grho,g2rho',rho,tau,grho,g2rho,'tauwrong=',tau_falsch
iint=iint+1
 endif

 a(1) = 1.5255251812009530d0
 a(2) = 0.4576575543602858d0
 a(3) = 0.4292036732051034d0

 c(0) =   0.7566445420735584d0
 c(1) =  -2.6363977871370960d0
 c(2) =   5.4745159964232880d0
 c(3) = -12.657308127108290d0
 c(4) =   4.1250584725121360d0
 c(5) = -30.425133957163840d0

 b(0) =   0.4771976183772063d0
 b(1) =  -1.7799813494556270d0
 b(2) =   3.8433841862302150d0
 b(3) =  -9.5912050880518490d0
 b(4) =   2.1730180285916720d0
 b(5) = -30.425133851603660d0

 d(0) =0.4435009886795587d0
 d(1) =0.58128653604457910d0
 d(2) =   66.742764515940610d0
 d(3) =  434.26780897229770d0
 d(4) =  824.7765766052239000d0
 d(5) = 1657.9652731582120d0

 e(0) =0.3347285060926091d0
 e(1) =0.47917931023971350d0
 e(2) =   62.392268338574240d0
 e(3) =  463.14816427938120d0
 e(4) =  785.2360350104029000d0
 e(5) = 1657.96296822327300d0

 dd = tau - 0.25d0*grho**2/rho
 q = (g2rho - 1.6d0*dd)/6d0

 if (abs(q) .gt. 1d-18) then

y = (2d0/3d0)*pi**(2d0/3d0)*rho**(5d0/3d0)/q
do i=0, 5
   yp(i) = y**i
enddo

if (y .le. 0d0) then
   g = -atan(a(1)*y + a(2)) + a(3)
   p1 = sum(c(0:5)*yp(0:5))
   p2 = sum(b(0:5)*yp(0:5))
elseif (y .gt. 0d0) then
   z = 2.085749716493756d0*y
   g = log(sqrt(1d0 + 1d0/z**2) + 1d0/z) + 2d0
   p1 = sum(d(0:5)*yp(0:5))
   p2 = sum(e(0:5)*yp(0:5))
endif

if (abs(p2) .gt. 1d-18) then

   x = g*p1/p2

   if (abs(x) .gt. 1d-18) then

  vxbrj = -2d0*pi**(1d0/3d0)*rho**(1d0/3d0)*exp(x/3d0)/x*(1d0 - exp(-x) - 0.5d0*x*exp(-x))

  if (tau .ge. 0d0) then
 vxbrj = xcconst*vxbrj + (3d0*xcconst-2d0)*sqrt(5d0/12d0)/pi*sqrt(tau/rho)
  else
 vxbrj = xcconst*vxbrj - (3d0*xcconst-2d0)*sqrt(5d0/12d0)/pi*sqrt(abs(tau/rho))
  endif

   endif
endif
 endif
  endif

  return
  end
___
Wien mailing list
Wien@zeus.theochem.tuwien.ac.at
http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien
SEARCH the MAILING-LIST at:  
http://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/index.html


[Wien] installation issues

2013-10-18 Thread Lawal Mohammed
Dear users and developers 


We are having problems with two of our systems, and couldn't figure out what is 
wrong. Even after checking the mailing list.
 1- We couldn't plot our output using rhoplot, For example, when we try 
 electron density plots for TiC, there was no visual. However, it works 
in some of our systems under the same condition.
 2- When I try to copy/upload the original case struct file into a new 
directory for supercell calculations I got this msg

 Can't locate CGI.pm in @INC (@INC contains: /usr/local/lib64/perl5 
/usr/local/share/perl5 /usr/lib64/perl5/vendor_perl 
/usr/share/perl5/vendor_perl /usr/lib64/perl5 /usr/share/perl5 .) at 
/home/mohammed/WIEN2K/SRC_w2web/htdocs/util/upload2.pl line 3. BEGIN 
failed--compilation aborted at 
/home/mohammed/WIEN2K/SRC_w2web/htdocs/util/upload2.pl line 3.


However, it works on one of our system. If there are any info that anybody may 
need to assist we will be grateful to provide.

With best regards

Mohammed lawal
universiti Technologi Malaysia___
Wien mailing list
Wien@zeus.theochem.tuwien.ac.at
http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien
SEARCH the MAILING-LIST at:  
http://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/index.html


Re: [Wien] installation issues

2013-10-18 Thread Gavin Abo
1- Is gnuplot installed on the non-working systems? Of note, it could be 
installed with sudo apt-get install gnuplot on older Ubuntu versions.  
However, I believe you have to use now sudo apt-get install gnuplot 
gnuplot-x11 on recent Ubuntu versions, since it seems that they decided 
to remove gnuplot-x11 as a dependency of gnuplot.


2- You likely need to install a missing Perl module.  For example, for 
Fedora:


https://ask.fedoraproject.org/question/8330/where-is-perl-module-cgipm-installed/

Since we don't know what operating systems you are using, you can search 
the Internet or your OS support (manual, forum, etc) on how to install 
these for your systems.


On 10/18/2013 12:27 AM, Lawal Mohammed wrote:

Dear users and developers

We are having problems with two of our systems, and couldn't figure 
out what is wrong. Even after checking the mailing list.
 1- We couldn't plot our output using rhoplot, For example, when we 
try  electron density plots for TiC, there was no visual. However, it 
works in some of our systems under the same condition.
 2- When I try to copy/upload the original case struct file into a new 
directory for supercell calculations I got this msg


 Can't locate CGI.pm in @INC (@INC contains: /usr/local/lib64/perl5 
/usr/local/share/perl5 /usr/lib64/perl5/vendor_perl 
/usr/share/perl5/vendor_perl /usr/lib64/perl5 /usr/share/perl5 .) at 
/home/mohammed/WIEN2K/SRC_w2web/htdocs/util/upload2.pl line 3. BEGIN 
failed--compilation aborted at 
/home/mohammed/WIEN2K/SRC_w2web/htdocs/util/upload2.pl line 3.


However, it works on one of our system. If there are any info that 
anybody may need to assist we will be grateful to provide.


With best regards
Mohammed lawal
universiti Technologi Malaysia


___
Wien mailing list
Wien@zeus.theochem.tuwien.ac.at
http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien
SEARCH the MAILING-LIST at:  
http://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/index.html


Re: [Wien] Intel(R) Xeon(R) CPU X5550 @ 2.67GHz vs Intel(R) Xeon(R) CPU E5620 @ 2.40GHz

2013-10-18 Thread Yundi Quan
First, thank Peter. I should have described my problem thoroughly.

:RKM  : MATRIX SIZE 9190LOs:1944  RKM= 4.88  WEIGHT= 2.00  PGR

The reduced RKM is 4.88. The reduced matrix size is 9190 which is about 2/5 of 
the full matrix. So that explains a lot. I'm using P1 symmetry. Therefore, the 
complex veion of lapw1, lapw2 are used. Compared with LDA calculations, LSDA 
almost doubles lapw1 and lapw2. 

I'm using P1 symmetry. Therefore, symmetry cannot reduce the number of stars 
(i.e. planes waves) in the interstitial region or the number of spherical 
harmonics inside the muffin-tin sphere. I guess that's why my job takes so 
long. And moreover, I'm only using k-point parallel without mpi.

Oxygen is the smallest atom in the unit cell. Reducing RKMAX to 6.5 is what I'm 
going to do first.

One of the clusters to which I have access has 8 cores per node and 8GB memory 
per node. Given the constraint of memory, I wonder how to improve the core 
usage when calculating compounds with large unit cells. For the compound I'm 
currently working on, I request one core per node and 8 nodes(=nkp) per job. So 
7*8=56 cores are wasted while running my job. I'm in dire need of help.


Yundi


On Oct 17, 2013, at 10:58 PM, Peter Blaha pbl...@theochem.tuwien.ac.at wrote:

 You still did not tell us the matrix size for the truncated RKmax, but yes,
 the scaling is probably ok.   (scaling goes with n^3; i.e. in case of of
 matrix size 12000 and 24000 we expect almost a factor of 8 !!! in cpu time.
 It also explaines the memory 
 
 You also did not tell us if you have inversion or not.
 
 One of my real cases with  NMAT= 21500   takes 400 sec on 64 cores (mpi), so 
 one
 could estimate something like 2 sec on a single core, which comes to the 
 right order
 of magnitude compared to your case.
 
 And: you may have 72 inequivalent atoms, but you did not tell us how many 
 atoms in total you have.
 The total number of atoms is the important info !!
 
 Probably you can reduce RKMAX (you did not tell us which atom has RMT=1.65 
 (probably O ??)
 and most likely you should use mpi AND iterative diagonalization.
 
 As I said, a case with 72 atoms (or whatever you have) can run in minutes on 
 a reasonable cluster
 and with a proper optimized setup (not just the defaults).
 
 
 Am 17.10.2013 18:05, schrieb Yundi Quan:
 Thanks a lot.
 On cluster A, RKM was automatically reduced to 4.88 while on cluster B RKM 
 was kept at 7. I didn't expect this, though I was aware that WIEN2k would 
 automatically reduce
 RKM in some cases. But is it reasonable for an iteration to run for eight 
 hours with the following parameters?
 Minimum sphere size: 1.65000 Bohr.
 Total k-mesh : 8
 Gmax : 12
 
 :RKM  : MATRIX SIZE23486LOs:1944  RKM= 7.00  WEIGHT= 2.00  PGR:
 :RKM  : MATRIX SIZE23486LOs:1944  RKM= 7.00  WEIGHT= 2.00  PGR:
 
 
 On Thu, Oct 17, 2013 at 8:54 AM, Peter Blaha pbl...@theochem.tuwien.ac.at 
 mailto:pbl...@theochem.tuwien.ac.at wrote:
 
The Xeon X5550 processor is a 4 core processor and your cluster may have 
 combined a few of them on one node (2-4 ?) Anyway, 14 cores are not really 
 possible ??
 
Have you done more than just looking on the total time ?
 
Is the machines file the same on both clusters ? Such a machines file 
 does NOT use  mpi.
 
One guess in case you really use mpi on cluster B (with a different 
 .machines file): In the sequential run (A) the basis set is limited by 
 NMATMAX, in the mpi-parallel
run it is not (or it is scaled up by sqrt(N-core)).
So it could be that run A has a MUCH smaller RKMAX than run (B).
Check grep :RKM case.scf   of the two runs.
What are the real matrix sizes 
 
Alternatively, NMATMAX could be chosen differently on the two machines 
 since somebody else installed WIEN2k.
 
Please compare carefully the resulting case.output1_1 files and 
 eventually send the RELEVANT PARTS OF THEM.
 
 
In any case, a 72 atom cell should NOT take 2 h / iteration (or even 8 
 ??).
 
What are your sphere sizes ???, what gives :RKM in case.scf ???
 
At least one can set   OMP_NUM_THREAD=2 or 4   and speed up the code by a 
 factor of almost 2. (You should see in the dayfile something close to 200 % 
 instead of ~100%
 
c1208-ib(1) 26558.265u 17.956s 7:34:14.39 97.5%0+0k 0+0io 0pf+0w
 
In essence:  A matrix size of 1 (real, with inversion) lapw1 should 
 take in the order of 10 min  (no mpi, maybe with OMP_NUM_THREAD=2)
 
 
 
On 10/17/2013 04:33 PM, Yundi Quan wrote:
 
Thanks for your reply.
a). both machines are set up in a way that once a node is assigned to 
 a
job, it cannot be assigned to another.
b). The .machines file looks like this
1:node1
1:node2
1:node3
1:node4
1:node5
1:node6
1:node7
1:node8
granularity:1
extrafine:1
lapw2_vector_split:1
 
I've been trying to avoid using mpi 

Re: [Wien] Intel(R) Xeon(R) CPU X5550 @ 2.67GHz vs Intel(R) Xeon(R) CPU E5620 @ 2.40GHz

2013-10-18 Thread Peter Blaha
As was mentioned before, such a big case needs   mpi  in order to run 
efficiently.


As a quick small improvement  set the OMP_NUM_THREAD variable to 2 or 
4.   This should give a speedup of about 2 and in the dayfile you should 
see that not 905% of the cpu was used, but 180% or so.



On 10/18/2013 10:51 AM, Yundi Quan wrote:

First, thank Peter. I should have described my problem thoroughly.

:RKM  : MATRIX SIZE 9190LOs:1944  RKM= 4.88  WEIGHT= 2.00  PGR

The reduced RKM is 4.88. The reduced matrix size is 9190 which is about 2/5 of 
the full matrix. So that explains a lot. I'm using P1 symmetry. Therefore, the 
complex veion of lapw1, lapw2 are used. Compared with LDA calculations, LSDA 
almost doubles lapw1 and lapw2.

I'm using P1 symmetry. Therefore, symmetry cannot reduce the number of stars 
(i.e. planes waves) in the interstitial region or the number of spherical 
harmonics inside the muffin-tin sphere. I guess that's why my job takes so 
long. And moreover, I'm only using k-point parallel without mpi.

Oxygen is the smallest atom in the unit cell. Reducing RKMAX to 6.5 is what I'm 
going to do first.

One of the clusters to which I have access has 8 cores per node and 8GB memory 
per node. Given the constraint of memory, I wonder how to improve the core 
usage when calculating compounds with large unit cells. For the compound I'm 
currently working on, I request one core per node and 8 nodes(=nkp) per job. So 
7*8=56 cores are wasted while running my job. I'm in dire need of help.


Yundi


On Oct 17, 2013, at 10:58 PM, Peter Blaha pbl...@theochem.tuwien.ac.at wrote:


You still did not tell us the matrix size for the truncated RKmax, but yes,
the scaling is probably ok.   (scaling goes with n^3; i.e. in case of of
matrix size 12000 and 24000 we expect almost a factor of 8 !!! in cpu time.
It also explaines the memory 

You also did not tell us if you have inversion or not.

One of my real cases with  NMAT= 21500   takes 400 sec on 64 cores (mpi), so one
could estimate something like 2 sec on a single core, which comes to the 
right order
of magnitude compared to your case.

And: you may have 72 inequivalent atoms, but you did not tell us how many atoms 
in total you have.
The total number of atoms is the important info !!

Probably you can reduce RKMAX (you did not tell us which atom has RMT=1.65 
(probably O ??)
and most likely you should use mpi AND iterative diagonalization.

As I said, a case with 72 atoms (or whatever you have) can run in minutes on a 
reasonable cluster
and with a proper optimized setup (not just the defaults).


Am 17.10.2013 18:05, schrieb Yundi Quan:

Thanks a lot.
On cluster A, RKM was automatically reduced to 4.88 while on cluster B RKM was 
kept at 7. I didn't expect this, though I was aware that WIEN2k would 
automatically reduce
RKM in some cases. But is it reasonable for an iteration to run for eight hours 
with the following parameters?
Minimum sphere size: 1.65000 Bohr.
Total k-mesh : 8
Gmax : 12

:RKM  : MATRIX SIZE23486LOs:1944  RKM= 7.00  WEIGHT= 2.00  PGR:
:RKM  : MATRIX SIZE23486LOs:1944  RKM= 7.00  WEIGHT= 2.00  PGR:


On Thu, Oct 17, 2013 at 8:54 AM, Peter Blaha pbl...@theochem.tuwien.ac.at 
mailto:pbl...@theochem.tuwien.ac.at wrote:

The Xeon X5550 processor is a 4 core processor and your cluster may have 
combined a few of them on one node (2-4 ?) Anyway, 14 cores are not really 
possible ??

Have you done more than just looking on the total time ?

Is the machines file the same on both clusters ? Such a machines file does 
NOT use  mpi.

One guess in case you really use mpi on cluster B (with a different 
.machines file): In the sequential run (A) the basis set is limited by NMATMAX, 
in the mpi-parallel
run it is not (or it is scaled up by sqrt(N-core)).
So it could be that run A has a MUCH smaller RKMAX than run (B).
Check grep :RKM case.scf   of the two runs.
What are the real matrix sizes 

Alternatively, NMATMAX could be chosen differently on the two machines 
since somebody else installed WIEN2k.

Please compare carefully the resulting case.output1_1 files and eventually 
send the RELEVANT PARTS OF THEM.


In any case, a 72 atom cell should NOT take 2 h / iteration (or even 8 ??).

What are your sphere sizes ???, what gives :RKM in case.scf ???

At least one can set   OMP_NUM_THREAD=2 or 4   and speed up the code by a 
factor of almost 2. (You should see in the dayfile something close to 200 % 
instead of ~100%

c1208-ib(1) 26558.265u 17.956s 7:34:14.39 97.5%0+0k 0+0io 0pf+0w

In essence:  A matrix size of 1 (real, with inversion) lapw1 should 
take in the order of 10 min  (no mpi, maybe with OMP_NUM_THREAD=2)



On 10/17/2013 04:33 PM, Yundi Quan wrote:

Thanks for your reply.
a). both machines are set up in a way that once a node is assigned to a
job, it cannot be assigned to another.
b). The .machines file 

Re: [Wien] lapw0 hang with mbj calculation for slab

2013-10-18 Thread alpa dashora
Dear Prof. Blaha,

Thanks for reply. I am unable to understand the reply. In only one
iteration FFT factor should be 4 or for complete mbj cycle it should be
same or again changed to 2.

Should I remove case.in0_grr file?

Please clear it.
 With regards,



On Fri, Oct 18, 2013 at 11:30 AM, Peter Blaha
pbl...@theochem.tuwien.ac.atwrote:

 Switch to PBE for one iteration.
 Increase  IFFTfactor from 2 to 4 in case.in0 (and, most likely, reduce
 RKMAX)

 Am 18.10.2013 05:55, schrieb alpa dashora:

 Dear Prof. Blaha and Wien2k users,

 I am trying to run mbj calculation for MoS2 10 layer slab. After few
 cycles lapw0 hanged. I have seen the earlier posts discussed on the same
 topics. I have used the
 vxclm2.f file as provided by Prof. Blaha and removed case.in0_grr file. I
 have also increased the Rmt*Kmax up to 9 but still I am not able to solve
 my problem.

 Please suggest any other solution for this error or any other method to
 calculate the correct band gap.

 Thanks in advance.

 With kind regards,

 --
 Alpa Dashora


 __**_
 Wien mailing list
 w...@zeus.theochem.tuwien.ac.**at Wien@zeus.theochem.tuwien.ac.at
 http://zeus.theochem.tuwien.**ac.at/mailman/listinfo/wienhttp://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien
 SEARCH the MAILING-LIST at:  http://www.mail-archive.com/**
 w...@zeus.theochem.tuwien.ac.**at/index.htmlhttp://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/index.html


 --
 --**---
 Peter Blaha
 Inst. Materials Chemistry, TU Vienna
 Getreidemarkt 9, A-1060 Vienna, Austria
 Tel: +43-1-5880115671
 Fax: +43-1-5880115698
 email: pbl...@theochem.tuwien.ac.at
 --**---
 __**_
 Wien mailing list
 w...@zeus.theochem.tuwien.ac.**at Wien@zeus.theochem.tuwien.ac.at
 http://zeus.theochem.tuwien.**ac.at/mailman/listinfo/wienhttp://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien
 SEARCH the MAILING-LIST at:  http://www.mail-archive.com/**
 w...@zeus.theochem.tuwien.ac.**at/index.htmlhttp://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/index.html




-- 
Alpa Dashora
___
Wien mailing list
Wien@zeus.theochem.tuwien.ac.at
http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien
SEARCH the MAILING-LIST at:  
http://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/index.html


Re: [Wien] lapw0 hang with mbj calculation for slab

2013-10-18 Thread Peter Blaha
For mBJ it is ALWAYS a good idea to use a larger IFFTfactor (complete 
cycle).


For a surface, case.in0_grr should not be present.

However, you need a file case.grr (normally prepared when case.in0_grr 
is present). This file (or the corresponding   avrg [(grad rho)/rho] 
value) should be taken from a corresponding bulk calculation.


On 10/18/2013 12:37 PM, alpa dashora wrote:

Dear Prof. Blaha,

Thanks for reply. I am unable to understand the reply. In only one
iteration FFT factor should be 4 or for complete mbj cycle it should be
same or again changed to 2.

Should I remove case.in0_grr file?

Please clear it.
  With regards,



On Fri, Oct 18, 2013 at 11:30 AM, Peter Blaha
pbl...@theochem.tuwien.ac.at mailto:pbl...@theochem.tuwien.ac.at wrote:

Switch to PBE for one iteration.
Increase  IFFTfactor from 2 to 4 in case.in0 (and, most likely,
reduce RKMAX)

Am 18.10.2013 05:55, schrieb alpa dashora:

Dear Prof. Blaha and Wien2k users,

I am trying to run mbj calculation for MoS2 10 layer slab. After
few cycles lapw0 hanged. I have seen the earlier posts discussed
on the same topics. I have used the
vxclm2.f file as provided by Prof. Blaha and removed
case.in0_grr file. I have also increased the Rmt*Kmax up to 9
but still I am not able to solve my problem.

Please suggest any other solution for this error or any other
method to calculate the correct band gap.

Thanks in advance.

With kind regards,

--
Alpa Dashora


_
Wien mailing list
w...@zeus.theochem.tuwien.ac.__at
mailto:Wien@zeus.theochem.tuwien.ac.at
http://zeus.theochem.tuwien.__ac.at/mailman/listinfo/wien
http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien
SEARCH the MAILING-LIST at:

http://www.mail-archive.com/__wien@zeus.theochem.tuwien.ac.__at/index.html
http://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/index.html


--
--__---
Peter Blaha
Inst. Materials Chemistry, TU Vienna
Getreidemarkt 9, A-1060 Vienna, Austria
Tel: +43-1-5880115671
Fax: +43-1-5880115698
email: pbl...@theochem.tuwien.ac.at
mailto:pbl...@theochem.tuwien.ac.at
--__---
_
Wien mailing list
w...@zeus.theochem.tuwien.ac.__at
mailto:Wien@zeus.theochem.tuwien.ac.at
http://zeus.theochem.tuwien.__ac.at/mailman/listinfo/wien
http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien
SEARCH the MAILING-LIST at:
http://www.mail-archive.com/__wien@zeus.theochem.tuwien.ac.__at/index.html
http://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/index.html




--
Alpa Dashora


___
Wien mailing list
Wien@zeus.theochem.tuwien.ac.at
http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien
SEARCH the MAILING-LIST at:  
http://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/index.html



--

  P.Blaha
--
Peter BLAHA, Inst.f. Materials Chemistry, TU Vienna, A-1060 Vienna
Phone: +43-1-58801-165300 FAX: +43-1-58801-165982
Email: bl...@theochem.tuwien.ac.atWWW: 
http://info.tuwien.ac.at/theochem/

--
___
Wien mailing list
Wien@zeus.theochem.tuwien.ac.at
http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien
SEARCH the MAILING-LIST at:  
http://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/index.html


Re: [Wien] installation issues

2013-10-18 Thread Lawal Mohammed
Thank you so much Prof. Peter Blaha, Abo Gavin, ...
The link you provided help. The issues are now resolved using two commands
1) yum install gnutplot gnuplot-x11   
2) yum install perl-CGI
in fedora 16 and18.

Another issue, in this regard, I got this error 

/home/mohammed/WIEN2K/lapw0: error while loading shared libraries: 
libmkl_intel_lp64.so: cannot open shared object file: No such file or directory 
   stop error
while running scf, Chakrabarti Shamik suggested that it is because the path for 
the file (libmkl_intel_lp64.so) is not properly given during compilation of 
wien2k (during userconfig_lapw). If that is the case, can anybody please tell 
me how to set the path for that file properly.

Thanking you all in advance for your response.

Best regards.








On Friday, October 18, 2013 3:30 PM, Gavin Abo gs...@crimson.ua.edu wrote:
 
1- Is gnuplot installed on the non-working systems? Of note, it could be 
installed with sudo apt-get install gnuplot on older Ubuntu versions.  
However, I believe you have to use now sudo apt-get install gnuplot 
gnuplot-x11 on recent Ubuntu versions, since it seems that they decided to 
remove gnuplot-x11 as a dependency of gnuplot.

2- You likely need to install a missing Perl module.  For example,
  for Fedora:

https://ask.fedoraproject.org/question/8330/where-is-perl-module-cgipm-installed/

Since we don't know what operating systems you are using, you can
  search the Internet or your OS support (manual, forum, etc) on how
  to install these for your systems.

On 10/18/2013 12:27 AM, Lawal Mohammed wrote:

Dear users and developers 



We are having problems with two of our systems, and couldn't figure out what 
is wrong. Even after checking the mailing list.
 1- We couldn't plot our output using rhoplot, For example,
  when we try  electron density plots for TiC, there was no
  visual. However, it works in some of our systems under the
  same condition.
 2- When I try to copy/upload the original case struct file
  into a new directory for supercell calculations I got this msg

 Can't locate CGI.pm in @INC (@INC contains:
  /usr/local/lib64/perl5 /usr/local/share/perl5
  /usr/lib64/perl5/vendor_perl /usr/share/perl5/vendor_perl
  /usr/lib64/perl5 /usr/share/perl5 .) at
  /home/mohammed/WIEN2K/SRC_w2web/htdocs/util/upload2.pl line 3.
  BEGIN failed--compilation aborted at
  /home/mohammed/WIEN2K/SRC_w2web/htdocs/util/upload2.pl line 3.



However, it works on one of our system. If there are any info that anybody may 
need to assist we will be grateful to provide.

With best regards

Mohammed lawal
universiti Technologi Malaysia



___
Wien mailing list
Wien@zeus.theochem.tuwien.ac.at
http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien
SEARCH the MAILING-LIST at:  
http://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/index.html___
Wien mailing list
Wien@zeus.theochem.tuwien.ac.at
http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien
SEARCH the MAILING-LIST at:  
http://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/index.html


Re: [Wien] lapw0 hang with mbj calculation for slab

2013-10-18 Thread NACIR M. TIT
Can you please remove my name (my email address= n...@uaeu.ac.ae) from your 
mailing List. Please do. I would appreciate that very much. Thanks!

**
Dr. Nacir Tit
Professor of Computational Physics
Physics Department
UAE University
Al-Ain
United Arab Emirates
Tel: +971 3 7136 326
Fax: +971 3 7671 291
Email: n...@uaeu.ac.ae
**

From: wien-boun...@zeus.theochem.tuwien.ac.at 
[wien-boun...@zeus.theochem.tuwien.ac.at] on behalf of Peter Blaha 
[pbl...@theochem.tuwien.ac.at]
Sent: Friday, October 18, 2013 3:36 PM
To: A Mailing list for WIEN2k users
Subject: Re: [Wien] lapw0 hang with mbj calculation for slab

For mBJ it is ALWAYS a good idea to use a larger IFFTfactor (complete
cycle).

For a surface, case.in0_grr should not be present.

However, you need a file case.grr (normally prepared when case.in0_grr
is present). This file (or the corresponding   avrg [(grad rho)/rho]
value) should be taken from a corresponding bulk calculation.

On 10/18/2013 12:37 PM, alpa dashora wrote:
 Dear Prof. Blaha,

 Thanks for reply. I am unable to understand the reply. In only one
 iteration FFT factor should be 4 or for complete mbj cycle it should be
 same or again changed to 2.

 Should I remove case.in0_grr file?

 Please clear it.
   With regards,



 On Fri, Oct 18, 2013 at 11:30 AM, Peter Blaha
 pbl...@theochem.tuwien.ac.at mailto:pbl...@theochem.tuwien.ac.at wrote:

 Switch to PBE for one iteration.
 Increase  IFFTfactor from 2 to 4 in case.in0 (and, most likely,
 reduce RKMAX)

 Am 18.10.2013 05:55, schrieb alpa dashora:

 Dear Prof. Blaha and Wien2k users,

 I am trying to run mbj calculation for MoS2 10 layer slab. After
 few cycles lapw0 hanged. I have seen the earlier posts discussed
 on the same topics. I have used the
 vxclm2.f file as provided by Prof. Blaha and removed
 case.in0_grr file. I have also increased the Rmt*Kmax up to 9
 but still I am not able to solve my problem.

 Please suggest any other solution for this error or any other
 method to calculate the correct band gap.

 Thanks in advance.

 With kind regards,

 --
 Alpa Dashora


 _
 Wien mailing list
 w...@zeus.theochem.tuwien.ac.__at
 mailto:Wien@zeus.theochem.tuwien.ac.at
 http://zeus.theochem.tuwien.__ac.at/mailman/listinfo/wien
 http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien
 SEARCH the MAILING-LIST at:
 
 http://www.mail-archive.com/__wien@zeus.theochem.tuwien.ac.__at/index.html
 
 http://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/index.html


 --
 --__---
 Peter Blaha
 Inst. Materials Chemistry, TU Vienna
 Getreidemarkt 9, A-1060 Vienna, Austria
 Tel: +43-1-5880115671
 Fax: +43-1-5880115698
 email: pbl...@theochem.tuwien.ac.at
 mailto:pbl...@theochem.tuwien.ac.at
 --__---
 _
 Wien mailing list
 w...@zeus.theochem.tuwien.ac.__at
 mailto:Wien@zeus.theochem.tuwien.ac.at
 http://zeus.theochem.tuwien.__ac.at/mailman/listinfo/wien
 http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien
 SEARCH the MAILING-LIST at:
 http://www.mail-archive.com/__wien@zeus.theochem.tuwien.ac.__at/index.html
 http://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/index.html




 --
 Alpa Dashora


 ___
 Wien mailing list
 Wien@zeus.theochem.tuwien.ac.at
 http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien
 SEARCH the MAILING-LIST at:  
 http://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/index.html


--

   P.Blaha
--
Peter BLAHA, Inst.f. Materials Chemistry, TU Vienna, A-1060 Vienna
Phone: +43-1-58801-165300 FAX: +43-1-58801-165982
Email: bl...@theochem.tuwien.ac.atWWW:
http://info.tuwien.ac.at/theochem/
--
___
Wien mailing list
Wien@zeus.theochem.tuwien.ac.at
http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien
SEARCH the MAILING-LIST at:  
http://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/index.html
___
Wien mailing list
Wien@zeus.theochem.tuwien.ac.at
http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien
SEARCH the MAILING-LIST at:  
http://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/index.html


Re: [Wien] hyperfine field question in ORB package for the aluminum

2013-10-18 Thread Jing-Han Chen
Dear Prof. Blaha

  It works very well after TEMP broadening is turned on. Thanks for your
suggestion.


2013/10/15 Peter Blaha pbl...@theochem.tuwien.ac.at

 Hi,

 I guess I never suggested B=1 T, but anyway, what you should check is
 if the calculated HFF vary linear with the applied field.

 I could imagine that with such calculations where you should have some
 artificial degeneracy of the 4 Al atoms, the TETRA method makes some
 small problem. In any case, it looks already fairly similar.

 Have you ever tried TEMP (with a small broadening ??, so that you do not
 destroy the magnetic shift).
 In addition, I suggest to increase the IFFT factor in case.in0 to 4 or 6,
 so that aliasing problems are reduced.

 Otherwise I would need to check this out myself.


 On 10/15/2013 06:25 PM, Jing-Han Chen wrote:

 Dear Prof. Blaha and other wien2k users:

 (I posted a similar message yesterday, apologies in case this appears as
 a repeat; the first message has not appeared on the list, perhaps
 reflected due to included images.)

 Regarding tests of the hyperfine fields in aluminum metal, we had
 thought about the issue of insufficient k-points, however we thought we
 had a handle on this issue. In a 9 T field, a rough calculation shows
 that the thin spin-polarized shell at Ef represents about 1/3000 of the
 BZ volume for fcc-Al. We ran a script gradually increasing the number of
 k-points, with a result (shown in
 http://people.physics.tamu.**edu/jhchen/points.pnghttp://people.physics.tamu.edu/jhchen/points.png)
 that the HFF settles
 down within about 20% of the expected value for 10,000 k-points in B=9T,
 with fluctuations dying down to the order of 10% and less in the range
 30,000 - 80,000 k-points. We also ran a test for linearity in B at a
 setting of 10,000 k-points, and the results appeared to be quite linear
 up to 100 T (shown in 
 http://people.physics.tamu.**edu/jhchen/field.pnghttp://people.physics.tamu.edu/jhchen/field.png
 ).

 We ran the test treating fcc-Al as simple cubic with 4 sites in order to
 be sure we understood how the field is applied in ORB, and expected if
 anything better convergence since the expanded cell gives a greater
 k-point density. However the results seem strange: with several k-point
 settings we found that in general, the HFF approached the expected value
 for fcc-Al after a relatively small number of iterations, yet without
 quite converging, and finally the HFF values diverged, with one or more
 going large and negative. We had not tried as many variations as for fcc
 since the results are much slower to obtain converged HFF.

 Following the suggestion of Prof. Blaha after our last posting we tried
 increasing to very large field and k-point values, and did finally get
 convergence (more than 10 last iterations of HFF is the same) for a
 setting of 10 k-points and 1  T, yielding 4 reasonably close
 positive values as in the following:

 --
 :HFF001:  143.345   0.000   0.572
 143.917 (KGAUSS)
 :HFF002:  143.344   0.000   0.572
 143.916 (KGAUSS)
 :HFF003:  144.427   0.000   0.583
 145.010 (KGAUSS)
 :HFF004:  143.344   0.000   0.572
 143.916 (KGAUSS)
 --

 However we are concerned that the HFF values are still not identical,
 whereas at 10,000 T the spin-polarized shell at Ef represents a
 significant fraction of the BZ, and the spin energy is quite large. We
 expected this to be more than enough k-points for random sampling of the
 shell at Ef.  For this reason, and in particular in light of the strange
 behavior in which the HFF values almost converge before diverging to
 widely separated values, is it possible that there might be some other
 issue that we are overlooking?

 Any suggestions would be appreciated.


 2013/10/7 Peter Blaha pbl...@theochem.tuwien.ac.at
 mailto:pblaha@theochem.**tuwien.ac.at pbl...@theochem.tuwien.ac.at


 The hyperfine field for a metal is coming mainly from the contact
 term due to the induced spin-polarization by the magnetic field.

 You should notice, that a field of 9 T is (for theoretical
 calculations) an extremely small field, causing a very small
 spin-splitting of the states near EF, which causes the HFF.
 I suppose all you see is numerical noise.

 Since only the states at EF are of interest (the field can only
 reoccupy states within a few mRy (or less) around EF), you need to
 converge your calculation with respect to:

 a) the k-mesh   (test MUCH larger meshes (1, 5 10 k or
 more)
 b) the magnetic field (increase it and test fields up to 1000 T),
 You are not interested in the absolute number, but in ppm, i.e. the
 relative induced field.

 c) The angular momentum component of the hFF introduced by
 case.vorbup/dn is NOT correct. I would even suggest that you put l=0
 to
 minimize the effect (or use-orbc  with case.vorbup/dn ,