This log contains three months worth of sanitized accounting records for the 128-node iPSC/860 located in the Numerical Aerodynamic Simulation (NAS) Systems Division at NASA Ames Research Center. The NAS facility supports industry, acadamia, and government labs all across the country. The workload on the iPSC/860 is a mix of interactive and batch jobs (development and production) mainly consisting of computational aeroscience applications. For more information about NAS, see URL http://www.nas.nasa.gov/. This somewhat aged log has the distinction of being the first to be analyzed in detail. The results are described in a paper cited below. It includes basic information about the number of nodes, runtime, start time, user, and command. The number of nodes is limited to powers of two due to the architecture. Note that the log does not include arrival information, only start times. The workload log from the NASA Ames iPSC/860 was graciously provided by Bill Nitzberg, who also helped with background information and interpretation. If you use this log in your work, please use a similar acknowledgment.
You can also reference the following:
Downloads:
|
|
The following summarizes the resource usage rules in effect during the time covered by the log.
Batch jobs were handled by NQS, which was configured with the following queues:
Time limit | number of nodes | |||
16 | 32 | 64 | 128 | |
0:20 | q16s*# | q32s*# | q64s# | q128s# |
1:00 | q16m* | q32m* | q64m | q128m |
3:00 | q16l | q32l | q64l | q128l |
Prime time is defined as Monday to Friday 6:00 to 20:00 PST. During this time, the running queues are q16s, q16m, q32s, and q32m. NQS jobs can use no more than 64 nodes (the size of the batch partition), and NQS will not kill interactive jobs.
The rest of the time is non-prime time. At such times all queues are runnable, and NQS jobs can use the entire cube. Moreover, NQS will kill interactive jobs to make room for NQS jobs.
root | root |
sysadmin | Operations account |
intel | Intel analysts |
develop | NAS system development staff |
support | NAS system support staff |
user | Scientific users (including NAS researchers) |
The log also contains special entries about system status. Again there is one line per entry:
"special" System Type Duration Start-Date Start-time Comments...These entries are distinguished by the first word in the line, which is "special".
"System" is nearly always "CUBE", referring to the iPSC/860.
"Type" is one of:
D | Dedicated Time (reserved for exclusive use by a user or sysadmin) |
P | Preventative Maintenence |
M | Scheduled Facility Outage |
S | Software Failure |
H | Hardware Failure |
F | Unscheduled Facility Outage |
O | Other |
To be consistent with job entries, the special entries gives "Duration" and "Start-time" to the nearest second. However, all times were reported in minutes, and are only accurate within a few minutes.
The conversion was done by a log-specific parser in conjunction with a more general converter module.
The difference between conversion 3 (reflected in NASA-iPSC-1993-3.swf) and and conversion 2 (NASA-iPSC-1993-2.swf) is that in the older conversion wait times were listed as 0. In the new one this was changed to -1, as we actually do not know what the wait times were (and what the original submit times were).
The differences between conversion 2 (reflected in NASA-iPSC-1993-2.swf) and conversion 1 (NASA-iPSC-1993-1.swf) is that in the original conversion timegm was used to convert dates and ti mes into UTC. This is wrong in case daylight saving time is used. Conversion 2 used timelocal with the correct timezone setting, which is hopefully the right thing to do.
To aid in this, a cleaned version of the log is provided as NASA-iPSC-1993-3.1-cln.swf. The filter used to remove the spurious pwd jobs was
user=3 and application=1 and processors=1Note that this filter was applied to the original log, and unfiltered jobs remain untouched. As a result, in the filtered log job numbering is not consecutive.
File NASA-iPSC-1993-3.1-cln.swf