Re: namd-l: namd on nvidia 302.17

2012-09-27 Thread Aron Broom
glad to hear you fixed it.  I guess that means you can use the normal
"non-development" driver to run NAMD?  That is good news.

On Thu, Sep 27, 2012 at 1:39 PM, Francesco Pietra wrote:

> SOLVED. Although not told by the the tests"dpkg -l |grep nvia| amd
> "modinfo nvidia", there was a mismatch between the runtime and the
> driver. When this became clear, on a new "apt-get upgrade" a mixture
> of versions 302 and 304 was installed, creating a mess. I had to
> correct manually by installing the specific version 304 for all
> packages with "apt-get install=version".
>
> In face of so much time lost for trivial problems - and posted on NAMD
> for NAMD non existing problems (I apologize for that)- I think now
> that it would be better (at least for people using the OS for
> scientific purposes) to install the driver the "nvidia way" rather
> than the "Debian way". In order to have something fixed when upgrading
> Debian. As I am presently short of time, I decided for no more
> upgrading Debian until I have free time enough to change the "way".
>
> Thanks
> francesco Pietra
>
> On Thu, Sep 27, 2012 at 3:54 PM, Francesco Pietra 
> wrote:
> > There are for me two ways of getting cuda at work: (a) install the
> > driver according to nvidia (as probably is implied in what you
> > suggested); (b) rely on Debian amd64, which furnishes precompiled
> > nvidia driver. I adopted (b) because upgrading is automatic and Debian
> > is notoriously highly reliable.
> >
> > I did not take notice of the cuda driver I had just before the "fatal"
> > upgrading, but it were months that I did not upgrade. The version noed
> > on my amd64 notebook is 295.53; probably I upgraded from that version.
> >
> > Now, on amd64,version 304.48.1 is available, while in my system
> > version 302.17-3 is installed, along with the basic
> > nvidia-kernel-dkms, as I posted initially. All under cuda-toolkit
> > version 4 (although this is not used in the "Debian way" of my
> > installation).
> >
> > The output of
> >
> > dpkg -l |grep nvidia
> >
> > modinfo nvidia
> >
> > which I posted initially, indicate, in my experience, that everything
> > is working correctly. On these basis, I suspected that 302.17-3 is too
> > advanced for current namd builds, although everything is under toolkit
> > 4 (or equivalent way).
> >
> > I could try to install 295 driver in place of 302 but probably someone
> > knows better than me what could be expected. Moving forward is easy,
> > going back, with any OS, is matter for experts.
> >
> > I am not sure that all I said is correct. I am a biochemist, not a
> > software expert.
> >
> > Thanks for your kind attention.
> >
> > francesco pietra
> >
> > On Thu, Sep 27, 2012 at 2:58 PM, Aron Broom  wrote:
> >> So one potential problem here: is 302.17 a development driver, or just
> the
> >> one Linux installs itself from the proprietary drivers?  It looks to me
> like
> >> the absolutely newest development driver is ver 295.41.  I'm not
> confident
> >> that you'd be able to run NAMD without the development driver installed.
> >> The installation is manual, and it should overwrite whatever driver you
> have
> >> there.  I recommend a trip to the CUDA development zone webpage.
> >>
> >> ~Aron
> >>
> >> On Thu, Sep 27, 2012 at 3:52 AM, Francesco Pietra <
> chiendar...@gmail.com>
> >> wrote:
> >>>
> >>> Hello:
> >>> I have tried the NAMD_CVS-2012-09-26_Linux-x86_64-multicore-CUDA with
> >>> nvidia version 302.17:
> >>>
> >>> Running command: namd2 heat-01.conf +p6 +idlepoll
> >>>
> >>> Charm++: standalone mode (not using charmrun)
> >>> Converse/Charm++ Commit ID: v6.4.0-beta1-0-g5776d21
> >>> CharmLB> Load balancer assumes all CPUs are same.
> >>> Charm++> Running on 1 unique compute nodes (12-way SMP).
> >>> Charm++> cpu topology info is gathered in 0.001 seconds.
> >>> Info: NAMD CVS-2012-09-26 for Linux-x86_64-multicore-CUDA
> >>> Info:
> >>> Info: Please visit http://www.ks.uiuc.edu/Research/namd/
> >>> Info: for updates, documentation, and support information.
> >>> Info:
> >>> Info: Please cite Phillips et al., J. Comp. Chem. 26:1781-1802 (2005)
> >>> Info: in all publications reporting results obtained with NAMD.
> >>> Info:
> >>> Info: Based on Charm++/Converse 60400 for multicore-linux64-iccstatic
> >>> Info: Built Wed Sep 26 02:25:08 CDT 2012 by jim on lisboa.ks.uiuc.edu
> >>> Info: 1 NAMD  CVS-2012-09-26  Linux-x86_64-multicore-CUDA  6gig64
> >>> francesco
> >>> Info: Running on 6 processors, 1 nodes, 1 physical nodes.
> >>> Info: CPU topology information available.
> >>> Info: Charm++/Converse parallel runtime startup completed at 0.085423 s
> >>> FATAL ERROR: CUDA error in cudaGetDeviceCount on Pe 3 (gig64):
> >>> initialization error
> >>> FATAL ERROR: CUDA error in cudaGetDeviceCount on Pe 1 (gig64):
> >>> initialization error
> >>> - Processor 3 Exiting: Called CmiAbort 
> >>> Reason: FATAL ERROR: CUDA error in cudaGetDeviceCount on Pe 3 (gig64):
> >>> initialization error
> >>>
> >

Re: Issues with nvidia 302.17

2012-09-27 Thread Francesco Pietra
SOLVED. Although not told by the the tests"dpkg -l |grep nvia| amd
"modinfo nvidia", there was a mismatch between the runtime and the
driver. When this became clear, on a new "apt-get upgrade" a mixture
of versions 302 and 304 was installed, creating a mess. I had to
correct manually by installing the specific version 304 for all
packages with "apt-get install=version".

In face of so much time lost for trivial problems - and posted on NAMD
for NAMD non existing problems (I apologize for that)- I think now
that it would be better (at least for people using the OS for
scientific purposes) to install the driver the "nvidia way" rather
than the "Debian way". In order to have something fixed when upgrading
Debian. As I am presently short of time, I decided for no more
upgrading Debian until I have free time enough to change the "way".

Thanks
francesco Pietra

On Thu, Sep 27, 2012 at 4:30 PM, Francesco Pietra  wrote:
> Looking better, I noticed from the list below
>
> nvidia-xconfig 302.17-2
>
> while all others are 302.17-3
>
> Might that be relevant?
>
>
> The alternative to have all packages exactly the same version is
> upgrading to v 304.48-1
>
> francesco pietra
>
>
> -- Forwarded message --
> From: Francesco Pietra 
> Date: Thu, Sep 27, 2012 at 12:48 PM
> Subject: Re: Issues with nvidia 302.17
> To: Darac Marjal , amd64 Debian
> 
>
>
> Thanks, I am examining the archive, although, so far, it is not clear
> to me what to do with the nvidia installed version 302.17-3, and how
> DKMS will behave if I intall previous nvidia packages.
>
> At any event, I forgot to indicate that: xserver and nvidia are the
> same version:
>
> francesco@gig64:~$ dpkg -l |grep nvidia
> ii  glx-alternative-nvidia0.2.2
>   amd64allows the selection of NVIDIA as GLX provider
> ii  libgl1-nvidia-alternatives302.17-3
>   amd64transition libGL.so* diversions to
> glx-alternative-nvidia
> ii  libgl1-nvidia-glx:amd64   302.17-3
>   amd64NVIDIA binary OpenGL libraries
> ii  libglx-nvidia-alternatives302.17-3
>   amd64transition libgl.so diversions to
> glx-alternative-nvidia
> ii  libnvidia-ml1:amd64   302.17-3
>   amd64NVIDIA management library (NVML) runtime library
> ii  nvidia-alternative302.17-3
>   amd64allows the selection of NVIDIA as GLX provider
> ii  nvidia-glx302.17-3
>   amd64NVIDIA metapackage
> ii  nvidia-installer-cleanup  20120630+3
>   amd64Cleanup after driver installation with the
> nvidia-installer
> ii  nvidia-kernel-common  20120630+3
>   amd64NVIDIA binary kernel module support files
> ii  nvidia-kernel-dkms302.17-3
>   amd64NVIDIA binary kernel module DKMS source
> ii  nvidia-smi302.17-3
>   amd64NVIDIA System Management Interface
> ii  nvidia-support20120630+3
>   amd64NVIDIA binary graphics driver support files
> ii  nvidia-vdpau-driver:amd64 302.17-3
>   amd64NVIDIA vdpau driver
> ii  nvidia-xconfig302.17-2
>   amd64X configuration tool for non-free NVIDIA drivers
> ii  xserver-xorg-video-nvidia 302.17-3
>   amd64NVIDIA binary Xorg driver
> francesco@gig64:~$
>
>
> root@gig64:/home/francesco# modinfo nvidia
> filename:   /lib/modules/3.2.0-2-amd64/updates/dkms/nvidia.ko
> alias:  char-major-195-*
> version:302.17
> supported:  external
> license:NVIDIA
> alias:  pci:v10DEd0E00sv*sd*bc04sc80i00*
> alias:  pci:v10DEd0AA3sv*sd*bc0Bsc40i00*
> alias:  pci:v10DEd*sv*sd*bc03sc02i00*
> alias:  pci:v10DEd*sv*sd*bc03sc00i00*
> depends:i2c-core
> vermagic:   3.2.0-2-amd64 SMP mod_unload modversions
> parm:   NVreg_EnableVia4x:int
> parm:   NVreg_EnableALiAGP:int
> parm:   NVreg_ReqAGPRate:int
> parm:   NVreg_EnableAGPSBA:int
> parm:   NVreg_EnableAGPFW:int
> parm:   NVreg_Mobile:int
> parm:   NVreg_ResmanDebugLevel:int
> parm:   NVreg_RmLogonRC:int
> parm:   NVreg_ModifyDeviceFiles:int
> parm:   NVreg_DeviceFileUID:int
> parm:   NVreg_DeviceFileGID:int
> parm:   NVreg_DeviceFileMode:int
> parm:   NVreg_RemapLimit:int
> parm:   NVreg_UpdateMemoryTypes:int
> parm:   NVreg_InitializeSystemMemoryAllocations:int
> parm:   NVreg_UseVBios:int
> parm:   NVreg_RMEdgeIntrCheck:int
> parm:   NVreg_UsePageAttributeTable:int
> parm:   NVreg_EnableMSI:int
> parm:   NVreg_MapRegistersEarly:int
> parm:   NVreg_RegisterForACPIEvents:int
> parm:   NVreg_RegistryDwords:charp
> parm:   NVreg_RmMsg:charp
> parm

Re: namd-l: namd on nvidia 302.17

2012-09-27 Thread Francesco Pietra
SOLVED. Although not told by the the tests"dpkg -l |grep nvia| amd
"modinfo nvidia", there was a mismatch between the runtime and the
driver. When this became clear, on a new "apt-get upgrade" a mixture
of versions 302 and 304 was installed, creating a mess. I had to
correct manually by installing the specific version 304 for all
packages with "apt-get install=version".

In face of so much time lost for trivial problems - and posted on NAMD
for NAMD non existing problems (I apologize for that)- I think now
that it would be better (at least for people using the OS for
scientific purposes) to install the driver the "nvidia way" rather
than the "Debian way". In order to have something fixed when upgrading
Debian. As I am presently short of time, I decided for no more
upgrading Debian until I have free time enough to change the "way".

Thanks
francesco Pietra

On Thu, Sep 27, 2012 at 3:54 PM, Francesco Pietra  wrote:
> There are for me two ways of getting cuda at work: (a) install the
> driver according to nvidia (as probably is implied in what you
> suggested); (b) rely on Debian amd64, which furnishes precompiled
> nvidia driver. I adopted (b) because upgrading is automatic and Debian
> is notoriously highly reliable.
>
> I did not take notice of the cuda driver I had just before the "fatal"
> upgrading, but it were months that I did not upgrade. The version noed
> on my amd64 notebook is 295.53; probably I upgraded from that version.
>
> Now, on amd64,version 304.48.1 is available, while in my system
> version 302.17-3 is installed, along with the basic
> nvidia-kernel-dkms, as I posted initially. All under cuda-toolkit
> version 4 (although this is not used in the "Debian way" of my
> installation).
>
> The output of
>
> dpkg -l |grep nvidia
>
> modinfo nvidia
>
> which I posted initially, indicate, in my experience, that everything
> is working correctly. On these basis, I suspected that 302.17-3 is too
> advanced for current namd builds, although everything is under toolkit
> 4 (or equivalent way).
>
> I could try to install 295 driver in place of 302 but probably someone
> knows better than me what could be expected. Moving forward is easy,
> going back, with any OS, is matter for experts.
>
> I am not sure that all I said is correct. I am a biochemist, not a
> software expert.
>
> Thanks for your kind attention.
>
> francesco pietra
>
> On Thu, Sep 27, 2012 at 2:58 PM, Aron Broom  wrote:
>> So one potential problem here: is 302.17 a development driver, or just the
>> one Linux installs itself from the proprietary drivers?  It looks to me like
>> the absolutely newest development driver is ver 295.41.  I'm not confident
>> that you'd be able to run NAMD without the development driver installed.
>> The installation is manual, and it should overwrite whatever driver you have
>> there.  I recommend a trip to the CUDA development zone webpage.
>>
>> ~Aron
>>
>> On Thu, Sep 27, 2012 at 3:52 AM, Francesco Pietra 
>> wrote:
>>>
>>> Hello:
>>> I have tried the NAMD_CVS-2012-09-26_Linux-x86_64-multicore-CUDA with
>>> nvidia version 302.17:
>>>
>>> Running command: namd2 heat-01.conf +p6 +idlepoll
>>>
>>> Charm++: standalone mode (not using charmrun)
>>> Converse/Charm++ Commit ID: v6.4.0-beta1-0-g5776d21
>>> CharmLB> Load balancer assumes all CPUs are same.
>>> Charm++> Running on 1 unique compute nodes (12-way SMP).
>>> Charm++> cpu topology info is gathered in 0.001 seconds.
>>> Info: NAMD CVS-2012-09-26 for Linux-x86_64-multicore-CUDA
>>> Info:
>>> Info: Please visit http://www.ks.uiuc.edu/Research/namd/
>>> Info: for updates, documentation, and support information.
>>> Info:
>>> Info: Please cite Phillips et al., J. Comp. Chem. 26:1781-1802 (2005)
>>> Info: in all publications reporting results obtained with NAMD.
>>> Info:
>>> Info: Based on Charm++/Converse 60400 for multicore-linux64-iccstatic
>>> Info: Built Wed Sep 26 02:25:08 CDT 2012 by jim on lisboa.ks.uiuc.edu
>>> Info: 1 NAMD  CVS-2012-09-26  Linux-x86_64-multicore-CUDA  6gig64
>>> francesco
>>> Info: Running on 6 processors, 1 nodes, 1 physical nodes.
>>> Info: CPU topology information available.
>>> Info: Charm++/Converse parallel runtime startup completed at 0.085423 s
>>> FATAL ERROR: CUDA error in cudaGetDeviceCount on Pe 3 (gig64):
>>> initialization error
>>> FATAL ERROR: CUDA error in cudaGetDeviceCount on Pe 1 (gig64):
>>> initialization error
>>> - Processor 3 Exiting: Called CmiAbort 
>>> Reason: FATAL ERROR: CUDA error in cudaGetDeviceCount on Pe 3 (gig64):
>>> initialization error
>>>
>>> Program finished.
>>> FATAL ERROR: CUDA error in cudaGetDeviceCount on Pe 4 (gig64):
>>> initialization error
>>> FATAL ERROR: CUDA error in cudaGetDeviceCount on Pe 2 (gig64):
>>> initialization error
>>>
>>>
>>> As I had (nearly) no comment to such failures, I can only imagine that
>>> either (i) my question - disregarding obvious issues - was too silly
>>> to merit attention; (ii) it is well known that nvidia version 302

Fwd: Issues with nvidia 302.17

2012-09-27 Thread Francesco Pietra
Looking better, I noticed from the list below

nvidia-xconfig 302.17-2

while all others are 302.17-3

Might that be relevant?


The alternative to have all packages exactly the same version is
upgrading to v 304.48-1

francesco pietra


-- Forwarded message --
From: Francesco Pietra 
Date: Thu, Sep 27, 2012 at 12:48 PM
Subject: Re: Issues with nvidia 302.17
To: Darac Marjal , amd64 Debian



Thanks, I am examining the archive, although, so far, it is not clear
to me what to do with the nvidia installed version 302.17-3, and how
DKMS will behave if I intall previous nvidia packages.

At any event, I forgot to indicate that: xserver and nvidia are the
same version:

francesco@gig64:~$ dpkg -l |grep nvidia
ii  glx-alternative-nvidia0.2.2
  amd64allows the selection of NVIDIA as GLX provider
ii  libgl1-nvidia-alternatives302.17-3
  amd64transition libGL.so* diversions to
glx-alternative-nvidia
ii  libgl1-nvidia-glx:amd64   302.17-3
  amd64NVIDIA binary OpenGL libraries
ii  libglx-nvidia-alternatives302.17-3
  amd64transition libgl.so diversions to
glx-alternative-nvidia
ii  libnvidia-ml1:amd64   302.17-3
  amd64NVIDIA management library (NVML) runtime library
ii  nvidia-alternative302.17-3
  amd64allows the selection of NVIDIA as GLX provider
ii  nvidia-glx302.17-3
  amd64NVIDIA metapackage
ii  nvidia-installer-cleanup  20120630+3
  amd64Cleanup after driver installation with the
nvidia-installer
ii  nvidia-kernel-common  20120630+3
  amd64NVIDIA binary kernel module support files
ii  nvidia-kernel-dkms302.17-3
  amd64NVIDIA binary kernel module DKMS source
ii  nvidia-smi302.17-3
  amd64NVIDIA System Management Interface
ii  nvidia-support20120630+3
  amd64NVIDIA binary graphics driver support files
ii  nvidia-vdpau-driver:amd64 302.17-3
  amd64NVIDIA vdpau driver
ii  nvidia-xconfig302.17-2
  amd64X configuration tool for non-free NVIDIA drivers
ii  xserver-xorg-video-nvidia 302.17-3
  amd64NVIDIA binary Xorg driver
francesco@gig64:~$


root@gig64:/home/francesco# modinfo nvidia
filename:   /lib/modules/3.2.0-2-amd64/updates/dkms/nvidia.ko
alias:  char-major-195-*
version:302.17
supported:  external
license:NVIDIA
alias:  pci:v10DEd0E00sv*sd*bc04sc80i00*
alias:  pci:v10DEd0AA3sv*sd*bc0Bsc40i00*
alias:  pci:v10DEd*sv*sd*bc03sc02i00*
alias:  pci:v10DEd*sv*sd*bc03sc00i00*
depends:i2c-core
vermagic:   3.2.0-2-amd64 SMP mod_unload modversions
parm:   NVreg_EnableVia4x:int
parm:   NVreg_EnableALiAGP:int
parm:   NVreg_ReqAGPRate:int
parm:   NVreg_EnableAGPSBA:int
parm:   NVreg_EnableAGPFW:int
parm:   NVreg_Mobile:int
parm:   NVreg_ResmanDebugLevel:int
parm:   NVreg_RmLogonRC:int
parm:   NVreg_ModifyDeviceFiles:int
parm:   NVreg_DeviceFileUID:int
parm:   NVreg_DeviceFileGID:int
parm:   NVreg_DeviceFileMode:int
parm:   NVreg_RemapLimit:int
parm:   NVreg_UpdateMemoryTypes:int
parm:   NVreg_InitializeSystemMemoryAllocations:int
parm:   NVreg_UseVBios:int
parm:   NVreg_RMEdgeIntrCheck:int
parm:   NVreg_UsePageAttributeTable:int
parm:   NVreg_EnableMSI:int
parm:   NVreg_MapRegistersEarly:int
parm:   NVreg_RegisterForACPIEvents:int
parm:   NVreg_RegistryDwords:charp
parm:   NVreg_RmMsg:charp
parm:   NVreg_NvAGP:int
root@gig64:/home/francesco#



francesco
***

On Thu, Sep 27, 2012 at 11:25 AM, Darac Marjal  wrote:
> On Thu, Sep 27, 2012 at 09:58:40AM +0200, Francesco Pietra wrote:
>> Hello:
>>
>> Following upgrading of amd64 wheezy, version 302.17 of nvidia is
>> incompatible with software I was using for molecular dynamics on
>> CPU-GPU.
>>
>> On the other hand, amd64 stable offers a version of nvidia that is too
>> old for the above application.
>>
>> Question: is it possible within amd64 wheezy to have a previous
>> version of nvidia? If yes, a generic guideline on how-to would be much
>> appreciated.
>
> It certainly is. Have a read of the instructions at
> http://snapshot.debian.org/
>
>
> -BEGIN PGP SIGNATURE-
> Version: GnuPG v1.4.10 (GNU/Linux)
>
> iQIcBAEBCgAGBQJQZBuPAAoJEKB7YbRsd8TGCgYP/19FPqF28+PHpVZq42fQuveK
> zYZVj6Eus/wvLawTN8BSUwDSuL+JLk/RhUAEwvygzcW5bZuvQ09UkFPLU5V6oZEP
> YjghVH2fCBbOA1UpidNfYwfLlXiiTNSU55zDVmhNkaRlb1YsRqBQCg5m4X8M1uzy
> ueJNGiAbkPpMJ80wPYhvKsFdH5qbPceAvL0imdfYdNsLLUrJyF6F9bhmkGeQ9yWV
> sqr+lgvPPPmv3o3jn1KlsvgbPbQ

Re: namd-l: namd on nvidia 302.17

2012-09-27 Thread Francesco Pietra
There are for me two ways of getting cuda at work: (a) install the
driver according to nvidia (as probably is implied in what you
suggested); (b) rely on Debian amd64, which furnishes precompiled
nvidia driver. I adopted (b) because upgrading is automatic and Debian
is notoriously highly reliable.

I did not take notice of the cuda driver I had just before the "fatal"
upgrading, but it were months that I did not upgrade. The version noed
on my amd64 notebook is 295.53; probably I upgraded from that version.

Now, on amd64,version 304.48.1 is available, while in my system
version 302.17-3 is installed, along with the basic
nvidia-kernel-dkms, as I posted initially. All under cuda-toolkit
version 4 (although this is not used in the "Debian way" of my
installation).

The output of

dpkg -l |grep nvidia

modinfo nvidia

which I posted initially, indicate, in my experience, that everything
is working correctly. On these basis, I suspected that 302.17-3 is too
advanced for current namd builds, although everything is under toolkit
4 (or equivalent way).

I could try to install 295 driver in place of 302 but probably someone
knows better than me what could be expected. Moving forward is easy,
going back, with any OS, is matter for experts.

I am not sure that all I said is correct. I am a biochemist, not a
software expert.

Thanks for your kind attention.

francesco pietra

On Thu, Sep 27, 2012 at 2:58 PM, Aron Broom  wrote:
> So one potential problem here: is 302.17 a development driver, or just the
> one Linux installs itself from the proprietary drivers?  It looks to me like
> the absolutely newest development driver is ver 295.41.  I'm not confident
> that you'd be able to run NAMD without the development driver installed.
> The installation is manual, and it should overwrite whatever driver you have
> there.  I recommend a trip to the CUDA development zone webpage.
>
> ~Aron
>
> On Thu, Sep 27, 2012 at 3:52 AM, Francesco Pietra 
> wrote:
>>
>> Hello:
>> I have tried the NAMD_CVS-2012-09-26_Linux-x86_64-multicore-CUDA with
>> nvidia version 302.17:
>>
>> Running command: namd2 heat-01.conf +p6 +idlepoll
>>
>> Charm++: standalone mode (not using charmrun)
>> Converse/Charm++ Commit ID: v6.4.0-beta1-0-g5776d21
>> CharmLB> Load balancer assumes all CPUs are same.
>> Charm++> Running on 1 unique compute nodes (12-way SMP).
>> Charm++> cpu topology info is gathered in 0.001 seconds.
>> Info: NAMD CVS-2012-09-26 for Linux-x86_64-multicore-CUDA
>> Info:
>> Info: Please visit http://www.ks.uiuc.edu/Research/namd/
>> Info: for updates, documentation, and support information.
>> Info:
>> Info: Please cite Phillips et al., J. Comp. Chem. 26:1781-1802 (2005)
>> Info: in all publications reporting results obtained with NAMD.
>> Info:
>> Info: Based on Charm++/Converse 60400 for multicore-linux64-iccstatic
>> Info: Built Wed Sep 26 02:25:08 CDT 2012 by jim on lisboa.ks.uiuc.edu
>> Info: 1 NAMD  CVS-2012-09-26  Linux-x86_64-multicore-CUDA  6gig64
>> francesco
>> Info: Running on 6 processors, 1 nodes, 1 physical nodes.
>> Info: CPU topology information available.
>> Info: Charm++/Converse parallel runtime startup completed at 0.085423 s
>> FATAL ERROR: CUDA error in cudaGetDeviceCount on Pe 3 (gig64):
>> initialization error
>> FATAL ERROR: CUDA error in cudaGetDeviceCount on Pe 1 (gig64):
>> initialization error
>> - Processor 3 Exiting: Called CmiAbort 
>> Reason: FATAL ERROR: CUDA error in cudaGetDeviceCount on Pe 3 (gig64):
>> initialization error
>>
>> Program finished.
>> FATAL ERROR: CUDA error in cudaGetDeviceCount on Pe 4 (gig64):
>> initialization error
>> FATAL ERROR: CUDA error in cudaGetDeviceCount on Pe 2 (gig64):
>> initialization error
>>
>>
>> As I had (nearly) no comment to such failures, I can only imagine that
>> either (i) my question - disregarding obvious issues - was too silly
>> to merit attention; (ii) it is well known that nvidia version 302.17
>> is incompatible with current namd builds for Linux-GNU.
>>
>> At any event, in the frame of metapackages, it is probably impossible
>> within Debian GNU-Linux wheezy to go back to a previous version of
>> nvidia. On the other hand, the stable version of the OS furnishes a
>> much too old version of nvidia. Therefore, my question is:
>>
>> Any chance to compile namd in front of installed nvidia version 302.17?
>>
>> Thanks for advice. Without access to namd-cuda I am currently hindered
>> to answer a question raised by the reviewers of a manuscript (the CPU
>> cluster has long ago been shut down, as it became too expensive for
>> our budget)
>>
>> francesco pietra
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> On Wed, Sep 26, 2012 at 4:08 PM, Francesco Pietra 
>> wrote:
>> > I forgot to mention that I am at final version 2.9 of namd.
>> > f.
>> >
>> > On Wed, Sep 26, 2012 at 4:05 PM, Aron Broom  wrote:
>> >> I'm not certain, but I think the driver version needs to match the CUDA
>> >> toolkit version that NAMD uses, and I think the library fi

Re: Issues with nvidia 302.17

2012-09-27 Thread Francesco Pietra
Thanks, I am examining the archive, although, so far, it is not clear
to me what to do with the nvidia installed version 302.17-3, and how
DKMS will behave if I intall previous nvidia packages.

At any event, I forgot to indicate that: xserver and nvidia are the
same version:

francesco@gig64:~$ dpkg -l |grep nvidia
ii  glx-alternative-nvidia0.2.2
  amd64allows the selection of NVIDIA as GLX provider
ii  libgl1-nvidia-alternatives302.17-3
  amd64transition libGL.so* diversions to
glx-alternative-nvidia
ii  libgl1-nvidia-glx:amd64   302.17-3
  amd64NVIDIA binary OpenGL libraries
ii  libglx-nvidia-alternatives302.17-3
  amd64transition libgl.so diversions to
glx-alternative-nvidia
ii  libnvidia-ml1:amd64   302.17-3
  amd64NVIDIA management library (NVML) runtime library
ii  nvidia-alternative302.17-3
  amd64allows the selection of NVIDIA as GLX provider
ii  nvidia-glx302.17-3
  amd64NVIDIA metapackage
ii  nvidia-installer-cleanup  20120630+3
  amd64Cleanup after driver installation with the
nvidia-installer
ii  nvidia-kernel-common  20120630+3
  amd64NVIDIA binary kernel module support files
ii  nvidia-kernel-dkms302.17-3
  amd64NVIDIA binary kernel module DKMS source
ii  nvidia-smi302.17-3
  amd64NVIDIA System Management Interface
ii  nvidia-support20120630+3
  amd64NVIDIA binary graphics driver support files
ii  nvidia-vdpau-driver:amd64 302.17-3
  amd64NVIDIA vdpau driver
ii  nvidia-xconfig302.17-2
  amd64X configuration tool for non-free NVIDIA drivers
ii  xserver-xorg-video-nvidia 302.17-3
  amd64NVIDIA binary Xorg driver
francesco@gig64:~$


root@gig64:/home/francesco# modinfo nvidia
filename:   /lib/modules/3.2.0-2-amd64/updates/dkms/nvidia.ko
alias:  char-major-195-*
version:302.17
supported:  external
license:NVIDIA
alias:  pci:v10DEd0E00sv*sd*bc04sc80i00*
alias:  pci:v10DEd0AA3sv*sd*bc0Bsc40i00*
alias:  pci:v10DEd*sv*sd*bc03sc02i00*
alias:  pci:v10DEd*sv*sd*bc03sc00i00*
depends:i2c-core
vermagic:   3.2.0-2-amd64 SMP mod_unload modversions
parm:   NVreg_EnableVia4x:int
parm:   NVreg_EnableALiAGP:int
parm:   NVreg_ReqAGPRate:int
parm:   NVreg_EnableAGPSBA:int
parm:   NVreg_EnableAGPFW:int
parm:   NVreg_Mobile:int
parm:   NVreg_ResmanDebugLevel:int
parm:   NVreg_RmLogonRC:int
parm:   NVreg_ModifyDeviceFiles:int
parm:   NVreg_DeviceFileUID:int
parm:   NVreg_DeviceFileGID:int
parm:   NVreg_DeviceFileMode:int
parm:   NVreg_RemapLimit:int
parm:   NVreg_UpdateMemoryTypes:int
parm:   NVreg_InitializeSystemMemoryAllocations:int
parm:   NVreg_UseVBios:int
parm:   NVreg_RMEdgeIntrCheck:int
parm:   NVreg_UsePageAttributeTable:int
parm:   NVreg_EnableMSI:int
parm:   NVreg_MapRegistersEarly:int
parm:   NVreg_RegisterForACPIEvents:int
parm:   NVreg_RegistryDwords:charp
parm:   NVreg_RmMsg:charp
parm:   NVreg_NvAGP:int
root@gig64:/home/francesco#



francesco
***

On Thu, Sep 27, 2012 at 11:25 AM, Darac Marjal  wrote:
> On Thu, Sep 27, 2012 at 09:58:40AM +0200, Francesco Pietra wrote:
>> Hello:
>>
>> Following upgrading of amd64 wheezy, version 302.17 of nvidia is
>> incompatible with software I was using for molecular dynamics on
>> CPU-GPU.
>>
>> On the other hand, amd64 stable offers a version of nvidia that is too
>> old for the above application.
>>
>> Question: is it possible within amd64 wheezy to have a previous
>> version of nvidia? If yes, a generic guideline on how-to would be much
>> appreciated.
>
> It certainly is. Have a read of the instructions at
> http://snapshot.debian.org/
>
>
> -BEGIN PGP SIGNATURE-
> Version: GnuPG v1.4.10 (GNU/Linux)
>
> iQIcBAEBCgAGBQJQZBuPAAoJEKB7YbRsd8TGCgYP/19FPqF28+PHpVZq42fQuveK
> zYZVj6Eus/wvLawTN8BSUwDSuL+JLk/RhUAEwvygzcW5bZuvQ09UkFPLU5V6oZEP
> YjghVH2fCBbOA1UpidNfYwfLlXiiTNSU55zDVmhNkaRlb1YsRqBQCg5m4X8M1uzy
> ueJNGiAbkPpMJ80wPYhvKsFdH5qbPceAvL0imdfYdNsLLUrJyF6F9bhmkGeQ9yWV
> sqr+lgvPPPmv3o3jn1KlsvgbPbQfrj0x64JQWPbfkV1N59d8iwuJYeNspvRSmnnn
> EzgusvUbDmt1egR6Jm5bUBGnOlu/Qb68TJyEy9RP1dkflKSBBfy+FBEp+dcvpBwq
> l5cFMlBVJcmcXEVarQ/xgBVtIC1CNRhah4rDoaOTZghADWZ8U9r4x7QnNAZ++8jj
> u7ONLbrSc+c0ITBjYjzvSRmqYnUg5eZj2W9mKtvj5YDLqx+p3CJPSENbLsCEUATT
> /M8jOv1OzUjt5A0RGKlTL8KDBCmh5MCItHE8qfNqTH2UpqoeWL8JZWekhNJJacIv
> ypHk8TXxLukZcD79ol6gB8307vbMLRO87ia0gn0PX9+h9RBk3D2q1wplYOlW6BzM
> HKbvsHKTsVm2XUjb7KySkbJzuIW5v2X5rOCMC

Re: Issues with nvidia 302.17

2012-09-27 Thread Darac Marjal
On Thu, Sep 27, 2012 at 09:58:40AM +0200, Francesco Pietra wrote:
> Hello:
> 
> Following upgrading of amd64 wheezy, version 302.17 of nvidia is
> incompatible with software I was using for molecular dynamics on
> CPU-GPU.
> 
> On the other hand, amd64 stable offers a version of nvidia that is too
> old for the above application.
> 
> Question: is it possible within amd64 wheezy to have a previous
> version of nvidia? If yes, a generic guideline on how-to would be much
> appreciated.

It certainly is. Have a read of the instructions at
http://snapshot.debian.org/



signature.asc
Description: Digital signature