Re: namd-l: namd on nvidia 302.17
glad to hear you fixed it. I guess that means you can use the normal "non-development" driver to run NAMD? That is good news. On Thu, Sep 27, 2012 at 1:39 PM, Francesco Pietra wrote: > SOLVED. Although not told by the the tests"dpkg -l |grep nvia| amd > "modinfo nvidia", there was a mismatch between the runtime and the > driver. When this became clear, on a new "apt-get upgrade" a mixture > of versions 302 and 304 was installed, creating a mess. I had to > correct manually by installing the specific version 304 for all > packages with "apt-get install=version". > > In face of so much time lost for trivial problems - and posted on NAMD > for NAMD non existing problems (I apologize for that)- I think now > that it would be better (at least for people using the OS for > scientific purposes) to install the driver the "nvidia way" rather > than the "Debian way". In order to have something fixed when upgrading > Debian. As I am presently short of time, I decided for no more > upgrading Debian until I have free time enough to change the "way". > > Thanks > francesco Pietra > > On Thu, Sep 27, 2012 at 3:54 PM, Francesco Pietra > wrote: > > There are for me two ways of getting cuda at work: (a) install the > > driver according to nvidia (as probably is implied in what you > > suggested); (b) rely on Debian amd64, which furnishes precompiled > > nvidia driver. I adopted (b) because upgrading is automatic and Debian > > is notoriously highly reliable. > > > > I did not take notice of the cuda driver I had just before the "fatal" > > upgrading, but it were months that I did not upgrade. The version noed > > on my amd64 notebook is 295.53; probably I upgraded from that version. > > > > Now, on amd64,version 304.48.1 is available, while in my system > > version 302.17-3 is installed, along with the basic > > nvidia-kernel-dkms, as I posted initially. All under cuda-toolkit > > version 4 (although this is not used in the "Debian way" of my > > installation). > > > > The output of > > > > dpkg -l |grep nvidia > > > > modinfo nvidia > > > > which I posted initially, indicate, in my experience, that everything > > is working correctly. On these basis, I suspected that 302.17-3 is too > > advanced for current namd builds, although everything is under toolkit > > 4 (or equivalent way). > > > > I could try to install 295 driver in place of 302 but probably someone > > knows better than me what could be expected. Moving forward is easy, > > going back, with any OS, is matter for experts. > > > > I am not sure that all I said is correct. I am a biochemist, not a > > software expert. > > > > Thanks for your kind attention. > > > > francesco pietra > > > > On Thu, Sep 27, 2012 at 2:58 PM, Aron Broom wrote: > >> So one potential problem here: is 302.17 a development driver, or just > the > >> one Linux installs itself from the proprietary drivers? It looks to me > like > >> the absolutely newest development driver is ver 295.41. I'm not > confident > >> that you'd be able to run NAMD without the development driver installed. > >> The installation is manual, and it should overwrite whatever driver you > have > >> there. I recommend a trip to the CUDA development zone webpage. > >> > >> ~Aron > >> > >> On Thu, Sep 27, 2012 at 3:52 AM, Francesco Pietra < > chiendar...@gmail.com> > >> wrote: > >>> > >>> Hello: > >>> I have tried the NAMD_CVS-2012-09-26_Linux-x86_64-multicore-CUDA with > >>> nvidia version 302.17: > >>> > >>> Running command: namd2 heat-01.conf +p6 +idlepoll > >>> > >>> Charm++: standalone mode (not using charmrun) > >>> Converse/Charm++ Commit ID: v6.4.0-beta1-0-g5776d21 > >>> CharmLB> Load balancer assumes all CPUs are same. > >>> Charm++> Running on 1 unique compute nodes (12-way SMP). > >>> Charm++> cpu topology info is gathered in 0.001 seconds. > >>> Info: NAMD CVS-2012-09-26 for Linux-x86_64-multicore-CUDA > >>> Info: > >>> Info: Please visit http://www.ks.uiuc.edu/Research/namd/ > >>> Info: for updates, documentation, and support information. > >>> Info: > >>> Info: Please cite Phillips et al., J. Comp. Chem. 26:1781-1802 (2005) > >>> Info: in all publications reporting results obtained with NAMD. > >>> Info: > >>> Info: Based on Charm++/Converse 60400 for multicore-linux64-iccstatic > >>> Info: Built Wed Sep 26 02:25:08 CDT 2012 by jim on lisboa.ks.uiuc.edu > >>> Info: 1 NAMD CVS-2012-09-26 Linux-x86_64-multicore-CUDA 6gig64 > >>> francesco > >>> Info: Running on 6 processors, 1 nodes, 1 physical nodes. > >>> Info: CPU topology information available. > >>> Info: Charm++/Converse parallel runtime startup completed at 0.085423 s > >>> FATAL ERROR: CUDA error in cudaGetDeviceCount on Pe 3 (gig64): > >>> initialization error > >>> FATAL ERROR: CUDA error in cudaGetDeviceCount on Pe 1 (gig64): > >>> initialization error > >>> - Processor 3 Exiting: Called CmiAbort > >>> Reason: FATAL ERROR: CUDA error in cudaGetDeviceCount on Pe 3 (gig64): > >>> initialization error > >>> > >
Re: Issues with nvidia 302.17
SOLVED. Although not told by the the tests"dpkg -l |grep nvia| amd "modinfo nvidia", there was a mismatch between the runtime and the driver. When this became clear, on a new "apt-get upgrade" a mixture of versions 302 and 304 was installed, creating a mess. I had to correct manually by installing the specific version 304 for all packages with "apt-get install=version". In face of so much time lost for trivial problems - and posted on NAMD for NAMD non existing problems (I apologize for that)- I think now that it would be better (at least for people using the OS for scientific purposes) to install the driver the "nvidia way" rather than the "Debian way". In order to have something fixed when upgrading Debian. As I am presently short of time, I decided for no more upgrading Debian until I have free time enough to change the "way". Thanks francesco Pietra On Thu, Sep 27, 2012 at 4:30 PM, Francesco Pietra wrote: > Looking better, I noticed from the list below > > nvidia-xconfig 302.17-2 > > while all others are 302.17-3 > > Might that be relevant? > > > The alternative to have all packages exactly the same version is > upgrading to v 304.48-1 > > francesco pietra > > > -- Forwarded message -- > From: Francesco Pietra > Date: Thu, Sep 27, 2012 at 12:48 PM > Subject: Re: Issues with nvidia 302.17 > To: Darac Marjal , amd64 Debian > > > > Thanks, I am examining the archive, although, so far, it is not clear > to me what to do with the nvidia installed version 302.17-3, and how > DKMS will behave if I intall previous nvidia packages. > > At any event, I forgot to indicate that: xserver and nvidia are the > same version: > > francesco@gig64:~$ dpkg -l |grep nvidia > ii glx-alternative-nvidia0.2.2 > amd64allows the selection of NVIDIA as GLX provider > ii libgl1-nvidia-alternatives302.17-3 > amd64transition libGL.so* diversions to > glx-alternative-nvidia > ii libgl1-nvidia-glx:amd64 302.17-3 > amd64NVIDIA binary OpenGL libraries > ii libglx-nvidia-alternatives302.17-3 > amd64transition libgl.so diversions to > glx-alternative-nvidia > ii libnvidia-ml1:amd64 302.17-3 > amd64NVIDIA management library (NVML) runtime library > ii nvidia-alternative302.17-3 > amd64allows the selection of NVIDIA as GLX provider > ii nvidia-glx302.17-3 > amd64NVIDIA metapackage > ii nvidia-installer-cleanup 20120630+3 > amd64Cleanup after driver installation with the > nvidia-installer > ii nvidia-kernel-common 20120630+3 > amd64NVIDIA binary kernel module support files > ii nvidia-kernel-dkms302.17-3 > amd64NVIDIA binary kernel module DKMS source > ii nvidia-smi302.17-3 > amd64NVIDIA System Management Interface > ii nvidia-support20120630+3 > amd64NVIDIA binary graphics driver support files > ii nvidia-vdpau-driver:amd64 302.17-3 > amd64NVIDIA vdpau driver > ii nvidia-xconfig302.17-2 > amd64X configuration tool for non-free NVIDIA drivers > ii xserver-xorg-video-nvidia 302.17-3 > amd64NVIDIA binary Xorg driver > francesco@gig64:~$ > > > root@gig64:/home/francesco# modinfo nvidia > filename: /lib/modules/3.2.0-2-amd64/updates/dkms/nvidia.ko > alias: char-major-195-* > version:302.17 > supported: external > license:NVIDIA > alias: pci:v10DEd0E00sv*sd*bc04sc80i00* > alias: pci:v10DEd0AA3sv*sd*bc0Bsc40i00* > alias: pci:v10DEd*sv*sd*bc03sc02i00* > alias: pci:v10DEd*sv*sd*bc03sc00i00* > depends:i2c-core > vermagic: 3.2.0-2-amd64 SMP mod_unload modversions > parm: NVreg_EnableVia4x:int > parm: NVreg_EnableALiAGP:int > parm: NVreg_ReqAGPRate:int > parm: NVreg_EnableAGPSBA:int > parm: NVreg_EnableAGPFW:int > parm: NVreg_Mobile:int > parm: NVreg_ResmanDebugLevel:int > parm: NVreg_RmLogonRC:int > parm: NVreg_ModifyDeviceFiles:int > parm: NVreg_DeviceFileUID:int > parm: NVreg_DeviceFileGID:int > parm: NVreg_DeviceFileMode:int > parm: NVreg_RemapLimit:int > parm: NVreg_UpdateMemoryTypes:int > parm: NVreg_InitializeSystemMemoryAllocations:int > parm: NVreg_UseVBios:int > parm: NVreg_RMEdgeIntrCheck:int > parm: NVreg_UsePageAttributeTable:int > parm: NVreg_EnableMSI:int > parm: NVreg_MapRegistersEarly:int > parm: NVreg_RegisterForACPIEvents:int > parm: NVreg_RegistryDwords:charp > parm: NVreg_RmMsg:charp > parm
Re: namd-l: namd on nvidia 302.17
SOLVED. Although not told by the the tests"dpkg -l |grep nvia| amd "modinfo nvidia", there was a mismatch between the runtime and the driver. When this became clear, on a new "apt-get upgrade" a mixture of versions 302 and 304 was installed, creating a mess. I had to correct manually by installing the specific version 304 for all packages with "apt-get install=version". In face of so much time lost for trivial problems - and posted on NAMD for NAMD non existing problems (I apologize for that)- I think now that it would be better (at least for people using the OS for scientific purposes) to install the driver the "nvidia way" rather than the "Debian way". In order to have something fixed when upgrading Debian. As I am presently short of time, I decided for no more upgrading Debian until I have free time enough to change the "way". Thanks francesco Pietra On Thu, Sep 27, 2012 at 3:54 PM, Francesco Pietra wrote: > There are for me two ways of getting cuda at work: (a) install the > driver according to nvidia (as probably is implied in what you > suggested); (b) rely on Debian amd64, which furnishes precompiled > nvidia driver. I adopted (b) because upgrading is automatic and Debian > is notoriously highly reliable. > > I did not take notice of the cuda driver I had just before the "fatal" > upgrading, but it were months that I did not upgrade. The version noed > on my amd64 notebook is 295.53; probably I upgraded from that version. > > Now, on amd64,version 304.48.1 is available, while in my system > version 302.17-3 is installed, along with the basic > nvidia-kernel-dkms, as I posted initially. All under cuda-toolkit > version 4 (although this is not used in the "Debian way" of my > installation). > > The output of > > dpkg -l |grep nvidia > > modinfo nvidia > > which I posted initially, indicate, in my experience, that everything > is working correctly. On these basis, I suspected that 302.17-3 is too > advanced for current namd builds, although everything is under toolkit > 4 (or equivalent way). > > I could try to install 295 driver in place of 302 but probably someone > knows better than me what could be expected. Moving forward is easy, > going back, with any OS, is matter for experts. > > I am not sure that all I said is correct. I am a biochemist, not a > software expert. > > Thanks for your kind attention. > > francesco pietra > > On Thu, Sep 27, 2012 at 2:58 PM, Aron Broom wrote: >> So one potential problem here: is 302.17 a development driver, or just the >> one Linux installs itself from the proprietary drivers? It looks to me like >> the absolutely newest development driver is ver 295.41. I'm not confident >> that you'd be able to run NAMD without the development driver installed. >> The installation is manual, and it should overwrite whatever driver you have >> there. I recommend a trip to the CUDA development zone webpage. >> >> ~Aron >> >> On Thu, Sep 27, 2012 at 3:52 AM, Francesco Pietra >> wrote: >>> >>> Hello: >>> I have tried the NAMD_CVS-2012-09-26_Linux-x86_64-multicore-CUDA with >>> nvidia version 302.17: >>> >>> Running command: namd2 heat-01.conf +p6 +idlepoll >>> >>> Charm++: standalone mode (not using charmrun) >>> Converse/Charm++ Commit ID: v6.4.0-beta1-0-g5776d21 >>> CharmLB> Load balancer assumes all CPUs are same. >>> Charm++> Running on 1 unique compute nodes (12-way SMP). >>> Charm++> cpu topology info is gathered in 0.001 seconds. >>> Info: NAMD CVS-2012-09-26 for Linux-x86_64-multicore-CUDA >>> Info: >>> Info: Please visit http://www.ks.uiuc.edu/Research/namd/ >>> Info: for updates, documentation, and support information. >>> Info: >>> Info: Please cite Phillips et al., J. Comp. Chem. 26:1781-1802 (2005) >>> Info: in all publications reporting results obtained with NAMD. >>> Info: >>> Info: Based on Charm++/Converse 60400 for multicore-linux64-iccstatic >>> Info: Built Wed Sep 26 02:25:08 CDT 2012 by jim on lisboa.ks.uiuc.edu >>> Info: 1 NAMD CVS-2012-09-26 Linux-x86_64-multicore-CUDA 6gig64 >>> francesco >>> Info: Running on 6 processors, 1 nodes, 1 physical nodes. >>> Info: CPU topology information available. >>> Info: Charm++/Converse parallel runtime startup completed at 0.085423 s >>> FATAL ERROR: CUDA error in cudaGetDeviceCount on Pe 3 (gig64): >>> initialization error >>> FATAL ERROR: CUDA error in cudaGetDeviceCount on Pe 1 (gig64): >>> initialization error >>> - Processor 3 Exiting: Called CmiAbort >>> Reason: FATAL ERROR: CUDA error in cudaGetDeviceCount on Pe 3 (gig64): >>> initialization error >>> >>> Program finished. >>> FATAL ERROR: CUDA error in cudaGetDeviceCount on Pe 4 (gig64): >>> initialization error >>> FATAL ERROR: CUDA error in cudaGetDeviceCount on Pe 2 (gig64): >>> initialization error >>> >>> >>> As I had (nearly) no comment to such failures, I can only imagine that >>> either (i) my question - disregarding obvious issues - was too silly >>> to merit attention; (ii) it is well known that nvidia version 302
Fwd: Issues with nvidia 302.17
Looking better, I noticed from the list below nvidia-xconfig 302.17-2 while all others are 302.17-3 Might that be relevant? The alternative to have all packages exactly the same version is upgrading to v 304.48-1 francesco pietra -- Forwarded message -- From: Francesco Pietra Date: Thu, Sep 27, 2012 at 12:48 PM Subject: Re: Issues with nvidia 302.17 To: Darac Marjal , amd64 Debian Thanks, I am examining the archive, although, so far, it is not clear to me what to do with the nvidia installed version 302.17-3, and how DKMS will behave if I intall previous nvidia packages. At any event, I forgot to indicate that: xserver and nvidia are the same version: francesco@gig64:~$ dpkg -l |grep nvidia ii glx-alternative-nvidia0.2.2 amd64allows the selection of NVIDIA as GLX provider ii libgl1-nvidia-alternatives302.17-3 amd64transition libGL.so* diversions to glx-alternative-nvidia ii libgl1-nvidia-glx:amd64 302.17-3 amd64NVIDIA binary OpenGL libraries ii libglx-nvidia-alternatives302.17-3 amd64transition libgl.so diversions to glx-alternative-nvidia ii libnvidia-ml1:amd64 302.17-3 amd64NVIDIA management library (NVML) runtime library ii nvidia-alternative302.17-3 amd64allows the selection of NVIDIA as GLX provider ii nvidia-glx302.17-3 amd64NVIDIA metapackage ii nvidia-installer-cleanup 20120630+3 amd64Cleanup after driver installation with the nvidia-installer ii nvidia-kernel-common 20120630+3 amd64NVIDIA binary kernel module support files ii nvidia-kernel-dkms302.17-3 amd64NVIDIA binary kernel module DKMS source ii nvidia-smi302.17-3 amd64NVIDIA System Management Interface ii nvidia-support20120630+3 amd64NVIDIA binary graphics driver support files ii nvidia-vdpau-driver:amd64 302.17-3 amd64NVIDIA vdpau driver ii nvidia-xconfig302.17-2 amd64X configuration tool for non-free NVIDIA drivers ii xserver-xorg-video-nvidia 302.17-3 amd64NVIDIA binary Xorg driver francesco@gig64:~$ root@gig64:/home/francesco# modinfo nvidia filename: /lib/modules/3.2.0-2-amd64/updates/dkms/nvidia.ko alias: char-major-195-* version:302.17 supported: external license:NVIDIA alias: pci:v10DEd0E00sv*sd*bc04sc80i00* alias: pci:v10DEd0AA3sv*sd*bc0Bsc40i00* alias: pci:v10DEd*sv*sd*bc03sc02i00* alias: pci:v10DEd*sv*sd*bc03sc00i00* depends:i2c-core vermagic: 3.2.0-2-amd64 SMP mod_unload modversions parm: NVreg_EnableVia4x:int parm: NVreg_EnableALiAGP:int parm: NVreg_ReqAGPRate:int parm: NVreg_EnableAGPSBA:int parm: NVreg_EnableAGPFW:int parm: NVreg_Mobile:int parm: NVreg_ResmanDebugLevel:int parm: NVreg_RmLogonRC:int parm: NVreg_ModifyDeviceFiles:int parm: NVreg_DeviceFileUID:int parm: NVreg_DeviceFileGID:int parm: NVreg_DeviceFileMode:int parm: NVreg_RemapLimit:int parm: NVreg_UpdateMemoryTypes:int parm: NVreg_InitializeSystemMemoryAllocations:int parm: NVreg_UseVBios:int parm: NVreg_RMEdgeIntrCheck:int parm: NVreg_UsePageAttributeTable:int parm: NVreg_EnableMSI:int parm: NVreg_MapRegistersEarly:int parm: NVreg_RegisterForACPIEvents:int parm: NVreg_RegistryDwords:charp parm: NVreg_RmMsg:charp parm: NVreg_NvAGP:int root@gig64:/home/francesco# francesco *** On Thu, Sep 27, 2012 at 11:25 AM, Darac Marjal wrote: > On Thu, Sep 27, 2012 at 09:58:40AM +0200, Francesco Pietra wrote: >> Hello: >> >> Following upgrading of amd64 wheezy, version 302.17 of nvidia is >> incompatible with software I was using for molecular dynamics on >> CPU-GPU. >> >> On the other hand, amd64 stable offers a version of nvidia that is too >> old for the above application. >> >> Question: is it possible within amd64 wheezy to have a previous >> version of nvidia? If yes, a generic guideline on how-to would be much >> appreciated. > > It certainly is. Have a read of the instructions at > http://snapshot.debian.org/ > > > -BEGIN PGP SIGNATURE- > Version: GnuPG v1.4.10 (GNU/Linux) > > iQIcBAEBCgAGBQJQZBuPAAoJEKB7YbRsd8TGCgYP/19FPqF28+PHpVZq42fQuveK > zYZVj6Eus/wvLawTN8BSUwDSuL+JLk/RhUAEwvygzcW5bZuvQ09UkFPLU5V6oZEP > YjghVH2fCBbOA1UpidNfYwfLlXiiTNSU55zDVmhNkaRlb1YsRqBQCg5m4X8M1uzy > ueJNGiAbkPpMJ80wPYhvKsFdH5qbPceAvL0imdfYdNsLLUrJyF6F9bhmkGeQ9yWV > sqr+lgvPPPmv3o3jn1KlsvgbPbQ
Re: namd-l: namd on nvidia 302.17
There are for me two ways of getting cuda at work: (a) install the driver according to nvidia (as probably is implied in what you suggested); (b) rely on Debian amd64, which furnishes precompiled nvidia driver. I adopted (b) because upgrading is automatic and Debian is notoriously highly reliable. I did not take notice of the cuda driver I had just before the "fatal" upgrading, but it were months that I did not upgrade. The version noed on my amd64 notebook is 295.53; probably I upgraded from that version. Now, on amd64,version 304.48.1 is available, while in my system version 302.17-3 is installed, along with the basic nvidia-kernel-dkms, as I posted initially. All under cuda-toolkit version 4 (although this is not used in the "Debian way" of my installation). The output of dpkg -l |grep nvidia modinfo nvidia which I posted initially, indicate, in my experience, that everything is working correctly. On these basis, I suspected that 302.17-3 is too advanced for current namd builds, although everything is under toolkit 4 (or equivalent way). I could try to install 295 driver in place of 302 but probably someone knows better than me what could be expected. Moving forward is easy, going back, with any OS, is matter for experts. I am not sure that all I said is correct. I am a biochemist, not a software expert. Thanks for your kind attention. francesco pietra On Thu, Sep 27, 2012 at 2:58 PM, Aron Broom wrote: > So one potential problem here: is 302.17 a development driver, or just the > one Linux installs itself from the proprietary drivers? It looks to me like > the absolutely newest development driver is ver 295.41. I'm not confident > that you'd be able to run NAMD without the development driver installed. > The installation is manual, and it should overwrite whatever driver you have > there. I recommend a trip to the CUDA development zone webpage. > > ~Aron > > On Thu, Sep 27, 2012 at 3:52 AM, Francesco Pietra > wrote: >> >> Hello: >> I have tried the NAMD_CVS-2012-09-26_Linux-x86_64-multicore-CUDA with >> nvidia version 302.17: >> >> Running command: namd2 heat-01.conf +p6 +idlepoll >> >> Charm++: standalone mode (not using charmrun) >> Converse/Charm++ Commit ID: v6.4.0-beta1-0-g5776d21 >> CharmLB> Load balancer assumes all CPUs are same. >> Charm++> Running on 1 unique compute nodes (12-way SMP). >> Charm++> cpu topology info is gathered in 0.001 seconds. >> Info: NAMD CVS-2012-09-26 for Linux-x86_64-multicore-CUDA >> Info: >> Info: Please visit http://www.ks.uiuc.edu/Research/namd/ >> Info: for updates, documentation, and support information. >> Info: >> Info: Please cite Phillips et al., J. Comp. Chem. 26:1781-1802 (2005) >> Info: in all publications reporting results obtained with NAMD. >> Info: >> Info: Based on Charm++/Converse 60400 for multicore-linux64-iccstatic >> Info: Built Wed Sep 26 02:25:08 CDT 2012 by jim on lisboa.ks.uiuc.edu >> Info: 1 NAMD CVS-2012-09-26 Linux-x86_64-multicore-CUDA 6gig64 >> francesco >> Info: Running on 6 processors, 1 nodes, 1 physical nodes. >> Info: CPU topology information available. >> Info: Charm++/Converse parallel runtime startup completed at 0.085423 s >> FATAL ERROR: CUDA error in cudaGetDeviceCount on Pe 3 (gig64): >> initialization error >> FATAL ERROR: CUDA error in cudaGetDeviceCount on Pe 1 (gig64): >> initialization error >> - Processor 3 Exiting: Called CmiAbort >> Reason: FATAL ERROR: CUDA error in cudaGetDeviceCount on Pe 3 (gig64): >> initialization error >> >> Program finished. >> FATAL ERROR: CUDA error in cudaGetDeviceCount on Pe 4 (gig64): >> initialization error >> FATAL ERROR: CUDA error in cudaGetDeviceCount on Pe 2 (gig64): >> initialization error >> >> >> As I had (nearly) no comment to such failures, I can only imagine that >> either (i) my question - disregarding obvious issues - was too silly >> to merit attention; (ii) it is well known that nvidia version 302.17 >> is incompatible with current namd builds for Linux-GNU. >> >> At any event, in the frame of metapackages, it is probably impossible >> within Debian GNU-Linux wheezy to go back to a previous version of >> nvidia. On the other hand, the stable version of the OS furnishes a >> much too old version of nvidia. Therefore, my question is: >> >> Any chance to compile namd in front of installed nvidia version 302.17? >> >> Thanks for advice. Without access to namd-cuda I am currently hindered >> to answer a question raised by the reviewers of a manuscript (the CPU >> cluster has long ago been shut down, as it became too expensive for >> our budget) >> >> francesco pietra >> >> >> >> >> >> >> >> >> >> >> On Wed, Sep 26, 2012 at 4:08 PM, Francesco Pietra >> wrote: >> > I forgot to mention that I am at final version 2.9 of namd. >> > f. >> > >> > On Wed, Sep 26, 2012 at 4:05 PM, Aron Broom wrote: >> >> I'm not certain, but I think the driver version needs to match the CUDA >> >> toolkit version that NAMD uses, and I think the library fi
Re: Issues with nvidia 302.17
Thanks, I am examining the archive, although, so far, it is not clear to me what to do with the nvidia installed version 302.17-3, and how DKMS will behave if I intall previous nvidia packages. At any event, I forgot to indicate that: xserver and nvidia are the same version: francesco@gig64:~$ dpkg -l |grep nvidia ii glx-alternative-nvidia0.2.2 amd64allows the selection of NVIDIA as GLX provider ii libgl1-nvidia-alternatives302.17-3 amd64transition libGL.so* diversions to glx-alternative-nvidia ii libgl1-nvidia-glx:amd64 302.17-3 amd64NVIDIA binary OpenGL libraries ii libglx-nvidia-alternatives302.17-3 amd64transition libgl.so diversions to glx-alternative-nvidia ii libnvidia-ml1:amd64 302.17-3 amd64NVIDIA management library (NVML) runtime library ii nvidia-alternative302.17-3 amd64allows the selection of NVIDIA as GLX provider ii nvidia-glx302.17-3 amd64NVIDIA metapackage ii nvidia-installer-cleanup 20120630+3 amd64Cleanup after driver installation with the nvidia-installer ii nvidia-kernel-common 20120630+3 amd64NVIDIA binary kernel module support files ii nvidia-kernel-dkms302.17-3 amd64NVIDIA binary kernel module DKMS source ii nvidia-smi302.17-3 amd64NVIDIA System Management Interface ii nvidia-support20120630+3 amd64NVIDIA binary graphics driver support files ii nvidia-vdpau-driver:amd64 302.17-3 amd64NVIDIA vdpau driver ii nvidia-xconfig302.17-2 amd64X configuration tool for non-free NVIDIA drivers ii xserver-xorg-video-nvidia 302.17-3 amd64NVIDIA binary Xorg driver francesco@gig64:~$ root@gig64:/home/francesco# modinfo nvidia filename: /lib/modules/3.2.0-2-amd64/updates/dkms/nvidia.ko alias: char-major-195-* version:302.17 supported: external license:NVIDIA alias: pci:v10DEd0E00sv*sd*bc04sc80i00* alias: pci:v10DEd0AA3sv*sd*bc0Bsc40i00* alias: pci:v10DEd*sv*sd*bc03sc02i00* alias: pci:v10DEd*sv*sd*bc03sc00i00* depends:i2c-core vermagic: 3.2.0-2-amd64 SMP mod_unload modversions parm: NVreg_EnableVia4x:int parm: NVreg_EnableALiAGP:int parm: NVreg_ReqAGPRate:int parm: NVreg_EnableAGPSBA:int parm: NVreg_EnableAGPFW:int parm: NVreg_Mobile:int parm: NVreg_ResmanDebugLevel:int parm: NVreg_RmLogonRC:int parm: NVreg_ModifyDeviceFiles:int parm: NVreg_DeviceFileUID:int parm: NVreg_DeviceFileGID:int parm: NVreg_DeviceFileMode:int parm: NVreg_RemapLimit:int parm: NVreg_UpdateMemoryTypes:int parm: NVreg_InitializeSystemMemoryAllocations:int parm: NVreg_UseVBios:int parm: NVreg_RMEdgeIntrCheck:int parm: NVreg_UsePageAttributeTable:int parm: NVreg_EnableMSI:int parm: NVreg_MapRegistersEarly:int parm: NVreg_RegisterForACPIEvents:int parm: NVreg_RegistryDwords:charp parm: NVreg_RmMsg:charp parm: NVreg_NvAGP:int root@gig64:/home/francesco# francesco *** On Thu, Sep 27, 2012 at 11:25 AM, Darac Marjal wrote: > On Thu, Sep 27, 2012 at 09:58:40AM +0200, Francesco Pietra wrote: >> Hello: >> >> Following upgrading of amd64 wheezy, version 302.17 of nvidia is >> incompatible with software I was using for molecular dynamics on >> CPU-GPU. >> >> On the other hand, amd64 stable offers a version of nvidia that is too >> old for the above application. >> >> Question: is it possible within amd64 wheezy to have a previous >> version of nvidia? If yes, a generic guideline on how-to would be much >> appreciated. > > It certainly is. Have a read of the instructions at > http://snapshot.debian.org/ > > > -BEGIN PGP SIGNATURE- > Version: GnuPG v1.4.10 (GNU/Linux) > > iQIcBAEBCgAGBQJQZBuPAAoJEKB7YbRsd8TGCgYP/19FPqF28+PHpVZq42fQuveK > zYZVj6Eus/wvLawTN8BSUwDSuL+JLk/RhUAEwvygzcW5bZuvQ09UkFPLU5V6oZEP > YjghVH2fCBbOA1UpidNfYwfLlXiiTNSU55zDVmhNkaRlb1YsRqBQCg5m4X8M1uzy > ueJNGiAbkPpMJ80wPYhvKsFdH5qbPceAvL0imdfYdNsLLUrJyF6F9bhmkGeQ9yWV > sqr+lgvPPPmv3o3jn1KlsvgbPbQfrj0x64JQWPbfkV1N59d8iwuJYeNspvRSmnnn > EzgusvUbDmt1egR6Jm5bUBGnOlu/Qb68TJyEy9RP1dkflKSBBfy+FBEp+dcvpBwq > l5cFMlBVJcmcXEVarQ/xgBVtIC1CNRhah4rDoaOTZghADWZ8U9r4x7QnNAZ++8jj > u7ONLbrSc+c0ITBjYjzvSRmqYnUg5eZj2W9mKtvj5YDLqx+p3CJPSENbLsCEUATT > /M8jOv1OzUjt5A0RGKlTL8KDBCmh5MCItHE8qfNqTH2UpqoeWL8JZWekhNJJacIv > ypHk8TXxLukZcD79ol6gB8307vbMLRO87ia0gn0PX9+h9RBk3D2q1wplYOlW6BzM > HKbvsHKTsVm2XUjb7KySkbJzuIW5v2X5rOCMC
Re: Issues with nvidia 302.17
On Thu, Sep 27, 2012 at 09:58:40AM +0200, Francesco Pietra wrote: > Hello: > > Following upgrading of amd64 wheezy, version 302.17 of nvidia is > incompatible with software I was using for molecular dynamics on > CPU-GPU. > > On the other hand, amd64 stable offers a version of nvidia that is too > old for the above application. > > Question: is it possible within amd64 wheezy to have a previous > version of nvidia? If yes, a generic guideline on how-to would be much > appreciated. It certainly is. Have a read of the instructions at http://snapshot.debian.org/ signature.asc Description: Digital signature