Re: [gmx-users] residue SASA

2011-07-08 Thread Diego Enry Gomes
This is what I mean by quick and dirty. #!/bin/bash # Reads the output of g_sas per residue (-or) and writes only the residues you want. # Diego E.B. Gomes, Roberto D. Lins, Pedro G. Pascutti, Thereza A. Soares # mailto: di...@biof.ufrj.br #1 ) To perform multiple g_sas (will write out all

Re: [gmx-users] residue SASA

2011-07-07 Thread Diego Enry Gomes
There's a quick dirty workaround. You can write a script to g_sas each frame individually, writing many output files. Then grab the results from there. Sent from my iPhone On Jul 7, 2011, at 9:17 PM, Justin A. Lemkul jalem...@vt.edu wrote: ahmet yıldırım wrote: There are hydrophilic

Re: [gmx-users] Isn't dt=0.0001 very small?

2009-03-10 Thread Diego Enry Gomes
Hi Andrew, I have just repeated this tutorial. There is nothing wrong with that. Download all the files again and repeat the tutorial from the beginning. It can be a file that was corrupted during your previous download or you did something wrong in one of the steps. You can copypaste the

Re: [gmx-users] problem in pdb file after SA

2009-03-10 Thread Diego Enry Gomes
You mean that half the protein is in one side of your cubic box and the other half is on the other side ? Try removing PBC or fitting to a reference frame (like the one before SA). Search the list archives, this is a very common question. On Mar 9, 2009, at 10:17 PM, Homa Azizian wrote:

Re: [gmx-users] do_dssp fatal error

2009-03-10 Thread Diego Enry Gomes
Did you download the dssp software from here ? http://swift.cmbi.ru.nl/gv/dssp/HTML/dsspcmbi On Mar 10, 2009, at 1:12 AM, Homa Azizian wrote: Hi when I use do_dssp the following error come: Program do_dssp, VERSION 4.0.3 Source code file: do_dssp.c, line: 471 Fatal error: Failed to

Re: [gmx-users] gromacs in parallel version

2009-03-09 Thread Diego Enry Gomes
Looks like you are using MPICH2 as mpi software. Try including mpirun before mdrun_mpi. mpirun -n 4 mdun_mpi -v -s topol.tpr If that doesn't work you shoud run start the MPI DAEMON ( MPD ) before mpirun: mpdboot mpirun -n 2 mdun_mpi -v -s topol.tpr after you job finishes you might want to

Re: [gmx-users] question about parallel runing of gromacs

2009-03-09 Thread Diego Enry Gomes
On Mar 9, 2009, at 5:12 PM, Mark Abraham wrote: Xiang Mao wrote: My PC has a intel core2 cpu. I compiled gromacs under cygwin, and using MPICH2. After mpiboot, I use mpirun -n 2 mdrun_mpi to run EM, MD. OK that might be working right. Now you need to look in the log file for an

Re: [gmx-users] running GROMACS-MPI on Rocks cluster- strange results

2009-03-06 Thread Diego Enry Gomes
These results are not strange. Performance results really depend on the size/setup of your system. Next time use gmxbench so we can have a better reference. Are you using gromacs 4.0.4 ? It scales much better than gromacs-3.x.x versions. Anyway, this very bad scaling is normal over gigabit

Re: [gmx-users] Porcupine Plots

2009-02-02 Thread Diego Enry Gomes
Check the VMD list, you should find the script there. On Feb 1, 2009, at 9:28 PM, Mark Abraham wrote: nahren manuel wrote: Dear Gromacs Users, I have done PCA of my MD , I want to visually represent the motions in terms of porcupine plots. I came across Dynamite (web server) for this