We recently switched from EXT4 to XFS in an upgrade from rhel7 to 9. Both were using LVM and RAID-0.
We noticed that IO utilization stays near 100%
Before it was near 30%
That being said, disk performance seems to be improved:
xfs
Jobs: 1 (f=1): [m(1)][100.0%][r=179MiB/s,w=60.0MiB/s][r=45.8k,w=15.4k IOPS][eta 00m:00s]
test: (groupid=0, jobs=1): err= 0: pid=137015: Thu Feb 15 16:35:38 2024
read: IOPS=53.8k, BW=210MiB/s (220MB/s)(37.5GiB/182764msec)
bw ( KiB/s): min=162680, max=233032, per=100.00%, avg=215332.00, stdev=9183.35, samples=365
iops : min=40670, max=58258, avg=53832.96, stdev=2295.89, samples=365
write: IOPS=17.9k, BW=70.0MiB/s (73.4MB/s)(12.5GiB/182764msec); 0 zone resets
bw ( KiB/s): min=54584, max=77864, per=100.00%, avg=71764.69, stdev=3038.69, samples=365
iops : min=13646, max=19466, avg=17941.14, stdev=759.70, samples=365
cpu : usr=13.34%, sys=35.70%, ctx=1301039, majf=0, minf=466
IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.1%, >=64=100.0%
submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0%
complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.1%, >=64=0.0%
issued rwts: total=9830837,3276363,0,0 short=0,0,0,0 dropped=0,0,0,0
latency : target=0, window=0, percentile=100.00%, depth=64
Run status group 0 (all jobs):
READ: bw=210MiB/s (220MB/s), 210MiB/s-210MiB/s (220MB/s-220MB/s), io=37.5GiB (40.3GB), run=182764-182764msec
WRITE: bw=70.0MiB/s (73.4MB/s), 70.0MiB/s-70.0MiB/s (73.4MB/s-73.4MB/s), io=12.5GiB (13.4GB), run=182764-182764msec
Disk stats (read/write):
dm-2: ios=9892502/3815818, merge=0/0, ticks=8119978/3427405, in_queue=11547383, util=100.00%, aggrios=9900533/3832297, aggrmerge=3/1918, aggrticks=8119893/3475893, aggrin_queue=11595786, aggrutil=100.00%
sda: ios=9900533/3832297, merge=3/1918, ticks=8119893/3475893, in_queue=11595786, util=100.00%
ext4
Jobs: 1 (f=1): [m(1)][100.0%][r=138MiB/s,w=46.5MiB/s][r=35.2k,w=11.9k IOPS][eta 00m:00s]
test: (groupid=0, jobs=1): err= 0: pid=22582: Thu Feb 15 11:31:51 2024
read: IOPS=40.1k, BW=157MiB/s (164MB/s)(37.5GiB/245228msec)
bw ( KiB/s): min=69820, max=214576, per=100.00%, avg=160348.09, stdev=18431.93, samples=490
iops : min=17455, max=53644, avg=40086.98, stdev=4607.97, samples=490
write: IOPS=13.4k, BW=52.2MiB/s (54.7MB/s)(12.5GiB/245228msec)
bw ( KiB/s): min=22586, max=70568, per=100.00%, avg=53440.14, stdev=6173.39, samples=490
iops : min= 5646, max=17642, avg=13360.01, stdev=1543.36, samples=490
cpu : usr=16.55%, sys=75.96%, ctx=130793, majf=0, minf=10804
IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.1%, >=64=100.0%
submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0%
complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.1%, >=64=0.0%
issued rwts: total=9830837,3276363,0,0 short=0,0,0,0 dropped=0,0,0,0
latency : target=0, window=0, percentile=100.00%, depth=64
Run status group 0 (all jobs):
READ: bw=157MiB/s (164MB/s), 157MiB/s-157MiB/s (164MB/s-164MB/s), io=37.5GiB (40.3GB), run=245228-245228msec
WRITE: bw=52.2MiB/s (54.7MB/s), 52.2MiB/s-52.2MiB/s (54.7MB/s-54.7MB/s), io=12.5GiB (13.4GB), run=245228-245228msec
Disk stats (read/write):
dm-3: ios=10027319/6362087, merge=0/0, ticks=4203766/1995710, in_queue=6209508, util=100.00%, aggrios=10033482/4346423, aggrmerge=1/2020208, aggrticks=4201752/1496265, aggrin_queue=5697005, aggrutil=100.00%
sda: ios=10033482/4346423, merge=1/2020208, ticks=4201752/1496265, in_queue=5697005, util=100.00%
Test were run with fio
Any thoughts on what is causing utilization and iops to stay so high, and should we be concerned that they are?