Sysop: | Amessyroom |
---|---|
Location: | Fayetteville, NC |
Users: | 42 |
Nodes: | 6 (0 / 6) |
Uptime: | 01:16:56 |
Calls: | 220 |
Calls today: | 1 |
Files: | 824 |
Messages: | 121,522 |
Posted today: | 6 |
Monitor during a long run shows average and peak I/O rates to the disks
with busy files at about 1/2 of what they do for normal runs.
That is exactly what happens when the cache battery on a RAID controller dies.
On Fri, 18 Oct 2024 20:22:23 -0400, Arne Vajhøj wrote:
On 10/18/2024 8:07 PM, Lawrence D'Oliveiro wrote:
Has VMS still not got any equivalent to mdraid?
VMS got volume shadowing in 1986 I believe.
Relevance being?
On 10/18/2024 8:35 PM, Lawrence D'Oliveiro wrote:
On Fri, 18 Oct 2024 20:22:23 -0400, Arne Vajhøj wrote:
On 10/18/2024 8:07 PM, Lawrence D'Oliveiro wrote:
Has VMS still not got any equivalent to mdraid?
VMS got volume shadowing in 1986 I believe.
Relevance being?
It is OS provided software RAID.
On Fri, 18 Oct 2024 20:39:33 -0400, Arne Vajhøj wrote:
On 10/18/2024 8:35 PM, Lawrence D'Oliveiro wrote:
On Fri, 18 Oct 2024 20:22:23 -0400, Arne Vajhøj wrote:
On 10/18/2024 8:07 PM, Lawrence D'Oliveiro wrote:
Has VMS still not got any equivalent to mdraid?
VMS got volume shadowing in 1986 I believe.
Relevance being?
It is OS provided software RAID.
Does “volume shadowing” mean just RAID 1?
On 10/18/2024 8:07 PM, Lawrence D'Oliveiro wrote:
Has VMS still not got any equivalent to mdraid?
VMS got volume shadowing in 1986 I believe.
On Fri, 18 Oct 2024 17:09:44 -0500, Craig A. Berry wrote:
That is exactly what happens when the cache battery on a RAID controller
dies.
I hate hardware RAID. Has VMS still not got any equivalent to mdraid?
We periodically (sometimes steady once a week, but sometimes more
frequent) one overnight batch job take much longer than normal to run.
Normal runtime about 30-35 minutes will take 4.5 - 6.5 hours. Several images called by that job all run much slower than normal. At the end
the overall CPU and I/O counts are very close between a normal and a
long job.
Note: Global buffers can be an advantage, but are not used when dealing with duplicate secondary keys. Those are handled in local buffers. I
have seen drastic differences in performance when changing bucket sizes
more with secondary keys that have many duplicates than with primary
keyed access. Hein has some tools that analyze the statistics of
indexed files that report the number of I/Os per operation. High values
here can indicate inefficient use of buckets or buckets that are too
small forcing the use of more I/Os to retrieve buckets. Increasing the bucket size can significantly reduce I/Os resulting in better overall
stats.
This won't directly address the reported slowdown, but might be a
trigger for it depending upon data locality.
Dan
Am 18.10.2024 um 20:26 schrieb Richard Jordan:
We periodically (sometimes steady once a week, but sometimes more
frequent) one overnight batch job take much longer than normal to run.
Normal runtime about 30-35 minutes will take 4.5 - 6.5 hours. Several
images called by that job all run much slower than normal. At the end
the overall CPU and I/O counts are very close between a normal and a
long job.
If 'overall CPU and I/O counts' are about the same, please re-consider
my advice to run T4. Look at the disk response times and I/O queue
length and compare a 'good' and a 'slow' run.
If 'the problem' would be somewhere in the disk IO sub-system, changing
RMS buffers will only 'muddy the waters'.
Volker.
On 11/5/24 10:32 AM, Volker Halle wrote:
Am 18.10.2024 um 20:26 schrieb Richard Jordan:...
I have monitor running and have been checking the I/O rates and queue lengths during the 30+ minute runs and the multi-hour runs, and
the only diffs there are the overall I/O rates to the two disks are much lower on the long runs than on the normal short ones.