Martin Feller wrote:
If the job is the one from below (globusrun-ws -submit -s -F gavosrv1
-c /bin/true)
then no file-cleanup with RFT is involved. I found a similar behavior
sometimes on one of
the ISSGC08 machines, where simple jobs seemed to take a nap between
cleanup and done.
Unfortunately i didn't follow up on this and cannot reproduce it right
now, but i
remember that the file-system on the headnode was very slow at that time.
The only thing i can think of is the cache cleanup step, where certain
files are removed.
Arthur:
* Is the filesystem on the headnode that contains the user homes very
slow (busy NFS or so)?
I don't know any reason the file system should be slow. (I also don't
know how to test it.) Note that several different machines are affected
in a similar way.
* Do you get a better performance if you run
globusrun-ws -submit -s -F gavosrv1 -f job.xml
with the following job description stored in job.xml ?
<job>
<executable>/bin/true</executable>
<stdout>/tmp/stdout</stdout>
<stderr>/tmp/stderr</stderr>
</job>
Yes, that helps noticeably, another factor of 2:
from gavosrv1.mpe.mpg.de to gavosrv1.mpe.mpg.de: 0m5.815s
from gavosrv1.mpe.mpg.de to udo-gt03.grid.tu-dortmund.de: 0m4.712s
from udo-gt03 to gavosrv1.mpe.mpg.de: 0m5.692s
from udo-gt03 to udo-gt03.grid.tu-dortmund.de: 0m3.661s
This is no longer terribly far removed from numbers other people were
reporting, and is getting to a range where I probably wouldn't bother
complaining. What is happening here? I get output back through streaming
in either case (with /bin/date instead of /bin/true). And I still don't
undertand why gsissh, which also has to do authentication, is ten times
faster.
--Art Carlson
Martin
Charles Bacon wrote:
Interesting - I remember some discussion like that on this list, I
think, but what does that rule achieve?
For the original user: delays like that are not normal. The activity
in the Cleanup->Done phase is an RFT job that deletes the files
associated with the job. Is your GRAM server configured to use a
local RFT server? Is the GridFTP server local to the machine running
the container? Do you notice slow results using globus-url-copy from
the machine to itself?
Charles
On Jul 21, 2008, at 4:39 AM, Steve White wrote:
Art,
As I understand it, your application runs a single process in the
"fork"
job manager. So you are referring to the latency in running a single
simple process, rather than to that in submission to a batch system.
I now remember that last September, Thomas Brüsemeister pointed out to
us a work-around for a similar problem, at least regarding file
transfers.
It was to add the following 'iptables' rule:
iptables -A OUTPUT -p tcp --syn --dport 113 -j REJECT --reject-with
tcp-reset
We implemented this on many of our systems at AIP, and observed a big
improvement in some kinds of latencey. Now I see that on some of them
the setting has been lost (after system upgrades, etc.)
Would this improve things for your application?
Cheers!
On 20.07.08, Arthur Carlson wrote:
In the thread "Globus not for real-time application?", a number of
users
discuss whether it is realistic or not to get latencies below 1
second.
Sounds like paradise. I am seeing latencies of up to a minute!
My workstation, gavosrv1.mpe.mpg.de, not the newest anymore, has GTK
4.0.5 installed. When I use globusrun-ws to go from this machine
back to
itself, ... but just look:
[EMAIL PROTECTED] ~]$ time globusrun-ws -submit -s -F gavosrv1 -c
/bin/true
Delegating user credentials...Done.
Submitting job...Done.
Job ID: uuid:52f0f962-54e1-11dd-a56f-0007e914d571
Termination time: 07/19/2008 15:51 GMT
Current job state: Active
Current job state: CleanUp-Hold
Current job state: CleanUp
Current job state: Done
Destroying job...Done.
Cleaning up any delegated credentials...Done.
real 0m24.327s
user 0m1.242s
sys 0m0.113s
Note that "user" and "sys" times are reasonable. Almost all of this
time
passes between "CleanUp" and "Done". It can't just be checking
credentials because gsissh is done in a jiffy:
[EMAIL PROTECTED] ~]$ time gsissh -p 2222 gavosrv1
/bin/true
real 0m0.649s
user 0m0.134s
sys 0m0.020s
Maybe that is already enough for someone to see where the problem
lies.
I can also point out that all (at least many) of the machines in our
grid (AstroGrid-D) seem to be affected, but to varying degrees.
Here is
a little matrix of tests:
from gavosrv1.mpe.mpg.de to gavosrv1.mpe.mpg.de: 0m27.235s
from gavosrv1.mpe.mpg.de to titan.ari.uni-heidelberg.de: 0m14.324s
from gavosrv1.mpe.mpg.de to udo-gt03.grid.tu-dortmund.de: 0m8.823s
from titan to gavosrv1.mpe.mpg.de: 0m57.208s
from titan to titan.ari.uni-heidelberg.de: 0m16.875s
from titan to udo-gt03.grid.tu-dortmund.de: 0m27.225s
from udo-gt03 to gavosrv1.mpe.mpg.de: 1m5.221s
from udo-gt03 to titan.ari.uni-heidelberg.de: 0m12.905s
from udo-gt03 to udo-gt03.grid.tu-dortmund.de: 0m6.952s
Please tell me I am doing something really stupid. For production
of my
application even a minute of latency is not a big deal, but it's a
pain
during development and debugging. Right now I am using gsissh
instead of
globusrun-ws just to work around this.
Thank for the lift,
Art Carlson
AstroGrid-D Project
Max-Planck-Institute für extraterrestrische Physik, Garching, Germany
--
| - - - - - - - - - - - - - - - - - - - - - -
- - -
| Steve White
+49(331)7499-202
| e-Science / AstroGrid-D Zi. 35
Bg. 20
| - - - - - - - - - - - - - - - - - - - - - -
- - -
| Astrophysikalisches Institut Potsdam (AIP)
| An der Sternwarte 16, D-14482 Potsdam
|
| Vorstand: Prof. Dr. Matthias Steinmetz, Peter A. Stolz
|
| Stiftung privaten Rechts, Stiftungsverzeichnis Brandenburg:
III/7-71-026
| - - - - - - - - - - - - - - - - - - - - - -
- - -
--
Dr. Arthur Carlson
Max-Planck-Institut fuer extraterrestrische Physik
Giessenbachstrasse, 85748 Garching, Germany
Phone: (+49 89) 30000-3357
E-Mail: [EMAIL PROTECTED]