I don't remember all the mods we made at NCSS, but one change that made
a BIG difference on the simplex and duplex 360/67's was this: in the CP
kernel, ALL SVC instructions were modified to a BAL to a specific
address in the first 4K of storage, where a "vector table" rerouted the
call to a specific CP "subroutine". All those interrupts and PSW swaps
took FOREVER on the 360/67, whereas a BAL to low storage SEEMED to fly
almost instantaineously. The change also seemed to be beneficial when we
switched to 370/168 platforms as well. The CMS kernel used a HVC (in
actual fact, a DIAGNOSE) to request services from the CP kernel,
including I/O services. We also modified MVT to run in a virtual machine
using DIAGNOSE, rather than SIO/TIO/HIO, for I/O services. Made MVT run
MUCH FASTER in the virtual machine and freed us from all the related
emulation of these I/O instructions. One thing I miss: Grant wrote a
program, called IMAGE, that created a complete image of the CP kernel,
which would load in record time when bringing up the system. I wish I
had a copy of that program now, because of its rather unique processing
of the RLD data from the object code. I've never quite understood how
RLD data is processed by either the linkage editor or the loader. :-(
---------------------------------<snip>------------------------------------
Anne & Lynn Wheeler wrote:
The following message is a courtesy copy of an article
that has been posted to bit.listserv.ibm-main,alt.folklore.computers as well.
[email protected] (Rick Fochtman) writes:
We called them "Gap Records" at NCSS and they worked very well for
paging on 2305 devices. We'd use the first exposure for the first
page, the second exposure for the third page and so on. They were
connected to a 370/168 via 2860 Selector Channels. Grant Tegtmeier
designed the code and computed the sizes of the Gap Records. He also
installed 3330 and 2305 support code, of his own design, in our
modified CP67/CMS system. In a long weekend! He was noted for long
periods of seeming boredom, punctuated by spurts of sheer genius.
re:
http://www.garlic.com/~lynn/2008s.html#51 Computer History Museum
http://www.garlic.com/~lynn/2008s.html#52 Computer History Museum
there were 3 people that came out from the science center to the
univ. to install cp67 the last week in jan68. One of these people left
the science center june68 to be part of ncss. He was suppose to teach a
cp67 class the following week (after he gave notice) for customers
... and the science center had to really scamble to find people to fill
in for him.
the initial cp67 code had fifo single operation processing for 2311,
2314s, and 2301 (drums). It would get about 80page transfers/sec on
2301. I redid the 2301 to do chained processing which increased peak
2301 page transfers to 300/sec. 2301 didn't have multiple request
exposure. i also redid the 2311 & 2314 code to implement ordered seek
operation (for all queued requests) ... both cp requests and cms
requests ... as well as chained request for page operations. On heavily
loaded CMS systems, the ordered seek queueing made big difference
... both graceful degradation as load increased ... as well as peak
throughput.
i also redid a whole bunch of the kernel pathlengths. This old
post:
http://www.garlic.com/~lynn/94.html#18 CP/67 & OS MFT14
contains part of presentation that I made at the '68 SHARE meeting in
Boston.
I had been doing heavy optimization of OS MFT system generations ...
carefully reordering all the STAGE2 output (from STAGE1) so that the
result would optimally place os/360 system files and PDS members on disk
(in order to minimize avg. arm seek distance). For the univ. student
work load, I would get a factor of about three times thruput
improvement. This would degrade over time as PTF maintenance was applied
... affecting high use system components. I would then have to
periodically rebuild the system in order to restore the carefully order
placement of files and PDS members.
I also got to do some work rewritting cp67 kernel ... besides redoing
the i/o stuff ... i also reworked a lot of the pathlengths ... in some
cases getting factor of 100 times improvement for some of the stuff.
As mentioned in the presentation, the original unmodified cp67 kernel
had 534 cpu seconds overhead for running MFT14 workload that took 322
seconds elapsed time. In the period between Jan68 and Aug68, I was able
to get that cp67 kernel virtual machine overhead down from 534 cpu
seconds to 113 cpu seconds (by rewritting several parts of the cp67
kernel).
I normally had classes during the week ... so much of my maintenance and
support work for OS/360 MFT and work on cp67 occurred on weekends. The
univ. typically shutdown the datacenter from 8am Sat. until 8am Monday
... during which time I could have the whole place for my personal use.
Monday classes were sometimes a problem after having been up for 48hrs
straight.
I had also done a dynamic adaptive resource manager and my own page
replacement algorithm and thrashing controls for cp67 ... lots of the
stuff IBM picked up and shipped in product while I was still
undergraduate at the univ (including the TTY/ASCII terminal support
mentioned in earlier post in this thread).
This recent post:
http://www.garlic.com/~lynn/2008s.html#17 IBM PC competitors
mentioned that I continued to do various cp67 things ... but much of it
was dropped in the product morph from cp67 to vm370. The above has
references/pointers to some old email regarding migrating various of the
pieces from cp67 to vm370 (after the science center finally replaced
their 360/67 with a 370/155-II).
some of the old email
http://www.garlic.com/~lynn/2006v.html#email731212
http://www.garlic.com/~lynn/2006w.html#email750102
http://www.garlic.com/~lynn/2006w.html#email750430
... included these posts from a couple years ago:
http://www.garlic.com/~lynn/2006v.html#36
http://www.garlic.com/~lynn/2006w.html#7
http://www.garlic.com/~lynn/2006w.html#8
Before the decision was made to release some of it in the standard vm370
product ... they let me build, distribute, and support highly modified
vm370 (aka csc/vm) systems for large number of internal systems. at one
point I would joke with the people on the 5th flr that the number peaked
about the same as the total number that they were supporting ... recent
reference here:
http://www.garlic.com/~lynn/2008s.html#48 New machines code
----------------------------------<unsnip>----------------------------------
--
Rick
--
Remember that if you're not the lead dog, the view never changes.
----------------------------------------------------------------------
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to [email protected] with the message: GET IBM-MAIN INFO
Search the archives at http://bama.ua.edu/archives/ibm-main.html