SGI: Hardware

LSI SAS/SATA HBA's and the Fuel (or other IP35 Systems) - Page 5

This is NetApp a SAN filer and not a JBOD or an FC RAID shelf, right?

First of all, SAN servers are usually heavily optimized for IOPs and not streaming speed.

I have an IBM TotalStorage DS4300. It's a rebadged LSI unit also sold as the SGI TP9300. I think I clocked it at ~ 120MB/s streaming speed with 6x 300GB 10K disks, single 2Gb attachment. The deal with the DS4300 is that when the controller cache battery dies, it disables write caching altogether and this has a pretty disastrous effect on performance. These batteries are uncommon, quite expensive, and last only 2 or 3 years, so by now they are usually dead. Fortunately, I can overrule this behavior (and put the entire DS4300 behind a UPS if I care about integrity).

I have an SGI TP9100 as well. It's loaded with a grab bag of old disks, some even 1Gb/s models (Cheetah 73LP). But with some striping I got this:

Code: Select all

tezro 28# diskperf -W -D -n "Scratch All" -t 5 -m 8m -c2g /vol/junk
#---------------------------------------------------------
# Disk Performance Test Results Generated By Diskperf V1.2
#
# Test name     : Scratch All
# Test date     : Wed May 30 23:43:29 2012
# Test machine  : IRIX64 tezro 6.5 07202013 IP35

# Test type     : XFS data subvolume
# Test path     : /vol/junk
# Request sizes : min=16384 max=8388608
# Parameters    : direct=1 time=5 scale=1.000 delay=0.000
# XFS file size : 2147483648 bytes
#---------------------------------------------------------
# req_size  fwd_wt  fwd_rd  bwd_wt  bwd_rd  rnd_wt  rnd_rd
#  (bytes)  (MB/s)  (MB/s)  (MB/s)  (MB/s)  (MB/s)  (MB/s)
#---------------------------------------------------------
16384   53.17   59.62   34.28    3.03   32.44    3.61
32768   85.89   89.44   61.64    6.43   55.09    6.99
65536  117.55  121.65   80.78   13.81   83.84   12.42
131072  166.67  188.38  129.31   24.32  125.20   20.55
262144  191.93  216.40  175.22   38.56  157.68   35.47
524288  291.29  327.95  230.43   61.52  180.57   61.50
1048576  312.74  337.81  154.67  123.96   96.59  114.65
2097152  326.42  366.79  177.17  200.72  169.23  193.71
4194304  342.27  363.44  220.99  266.21  207.55  250.90
8388608  347.64  383.67  267.42  307.71  259.48  311.20

I plan to load the TP9100 with 146GB 15K.4 disks. That should fly :D

Two things I like about the DS4300:

1) It supports IRIX, so failover and mutipathing 'just' works.
2) It's relatively quiet.

My experience with the QL2342 is quite good. It will deliver the expected performance. My experience with the LS driver (SAS/SATA, U320 SCSI and 4Gb FC) is more spotty:
* I use it with a pair of SATA disks in the O350. Performance is limited by the 2 SATA disks (streaming speeds ~ 240MB/s)
* I've used it with dual channel U320 SCSI and a DELL PowerVault 220S loaded two 7-way stripes of 72GB 15K.3 or 15K.4 disks and got disappointing results (200 or 250MB/s). The PowerVault was the U320 model. That was in the Tezro.
* I've got two dual 4Gb FC boards (LSI 7204XP) in the Tezro as well. That's what I used to obtain the TP9100 results above. And it didn't matter whether I attached both channels to a single LS card, or one each.

If I can get 380MB/s from a dual-2Gb attached FC array, and 200-250MB/s from a dual 320MB/s attached SCSI array on the same system there's something fishy going on. We've seen that SATA performs well earlier in this (?) thread. I have the feeling there's not much wrong with FC performance either, but that the U320 SCSI performance of the LS driver or hardware just isn't delivering. For now this is just a feeling, based on my own experience, and posts by Ian and others. I should bench some of this storage with a modern Linux/x64 server as well.
Now this is a deep dark secret, so everybody keep it quiet :)
It turns out that when reset, the WD33C93 defaults to a SCSI ID of 0, and it was simpler to leave it that way... -- Dave Olson, in comp.sys.sgi

Currently in commercial service: Image :Onyx2: (2x) :O3x02L:
In the museum : almost every MIPS/IRIX system.
Wanted : GM1 board for Professional Series GT graphics (030-0076-003, 030-0076-004)
jan-jaap wrote: ... but that the U320 SCSI performance of the LS driver or hardware just isn't delivering. ...


I demonstrated this a long time ago in an earlier thread , though nobody posted a reply. The LSI and QLA cards are about the same speed with two disks,
but the QLA leaps far ahead with more than two.

Ian.
Hmm. I referred to your findings while drafting this message but looks like I removed that before posting :roll:
Credit where credit is due :)

Did you ever test the same array with an LS1030 U320 host adapter in a non-IRIX system to find out whether the lack of performance was due to the hardware or the driver support?
Now this is a deep dark secret, so everybody keep it quiet :)
It turns out that when reset, the WD33C93 defaults to a SCSI ID of 0, and it was simpler to leave it that way... -- Dave Olson, in comp.sys.sgi

Currently in commercial service: Image :Onyx2: (2x) :O3x02L:
In the museum : almost every MIPS/IRIX system.
Wanted : GM1 board for Professional Series GT graphics (030-0076-003, 030-0076-004)
JJ,
> This is NetApp a SAN filer and not a JBOD or an FC RAID shelf, right?


Yup.
> I have an SGI TP9100 as well. It's loaded with a grab bag of old disks, some even 1Gb/s models (Cheetah 73LP). But with some striping I got this: [...]


Neat! XVM stripe I assume?
> My experience with the QL2342 is quite good. It will deliver the expected performance. My experience with the LS driver (SAS/SATA, U320 SCSI and 4Gb FC) is more spotty: [...]


I couldn't agree more; there's something fishy with the LS driver; I found that under serious loads (500MB/s and more) the LS driver seems to drop data, likely a bug, either in the firmware or driver; I've had XFS corruptions with very fast 15k drives, regardless if I enable/disable read+write cache, CFQ, etc. Of course, maybe I did hit a XFS bug, that's a possibility as well; or simply some interaction between XFS and LS doesn't quite work right. Amazingly, I've looked into this, and those same errors are being seen with the linux implementation of XFS (and I'm referring to "current" (10-2012) issues; no patch yet, that said, only one person currently supports XFS, so I'm not sure he has all the time necessary to look into this.
:Onyx2:
jan-jaap writes:
> Hmm. I referred to your findings while drafting this message but looks like I removed that before posting :roll: Credit where credit is due :)

Hehe, thanks. :D


> Did you ever test the same array with an LS1030 U320 host adapter in a non-IRIX system to find out whether the lack of performance was
> due to the hardware or the driver support?

Hmm, good point, no I didn't.

I need to sort out an issue with my AMD PC (it has proper PCIX slots) but once I do, if I remember, yes I'll test the LSI card, see what happens.

Ian.
hi
could eventually also work with Octane + XIO-PCI ?
IP30/Octane2, linux kernel development, Irix Scientific Apps (I'd like to use Ansys and Catia, I need more ram)
AFAIK Octane does not support the relevant LSI chipset. It's for O3K-type systems only.

Ian.
(07/Mar/2015) FREE! (collection only) 16x Sagitta 12-bay dual-channel U160 SCSI JBOD units.
Email, phone or PM for details, or see my forum post .
[email protected]
+44 (0)131 476 0796
It's strange because my old D9 Stone with single channel FC-AL JBOD config (no striping) gives me those numbers :


Code: Select all

Cobalt 1# diskperf -W -D -t10 -n"D9 Stone IBM HUS153030VLF400" -c500 /stor/stor4/test
---------------------------------------------------------
# Disk Performance Test Results Generated By Diskperf V1.2
#
# Test name     : D9 Stone IBM HUS153030VLF400
# Test date     : Thu Apr 17 13:56:31 2014
# Test machine  : IRIX64 Cobalt 6.5 10070055 IP30
# Test type     : XFS data subvolume
# Test path     : /stor/stor4/test
# Request sizes : min=16384 max=4194304
# Parameters    : direct=1 time=10 scale=1.000 delay=0.000
# XFS file size : 4194304 bytes
#---------------------------------------------------------
# req_size  fwd_wt  fwd_rd  bwd_wt  bwd_rd  rnd_wt  rnd_rd
#  (bytes)  (MB/s)  (MB/s)  (MB/s)  (MB/s)  (MB/s)  (MB/s)
#---------------------------------------------------------
16384   17.94   49.04   11.17    4.16    9.67    7.39
32768   38.08   65.29   16.03    8.64   15.06   20.34
65536   58.71   80.96   20.11   18.32   21.76   66.04
131072   76.33   89.51   44.85   25.04   35.60   63.83
262144   88.06   94.74   41.77   38.87   53.89   86.92
524288   93.78   97.12   66.22   42.22   68.88   73.76
1048576   93.97   97.40   84.42   51.85   86.43   67.91
2097152   96.05   99.56   56.90   61.94   85.07   74.62
4194304   96.20   78.81    0.00    0.00   80.74   87.23



Not amazing by any stretch of the word but still about double what you got for top writing speed.


mia wrote: Another one for fun: (this one is work in progress)

I've replicated FC jedi's Chris Kalisiak's configuration from http://www.futuretech.blinkenlights.nl/fc.html (Ian's website):

Took a "relatively large Origin server", a netapp 14-disks (Seagate 300GB FC) lun array, and got those laughable results:

Code: Select all

root@plum:/mnt# diskperf -W -D -r 4k -m 4m testfile
#---------------------------------------------------------
# Disk Performance Test Results Generated By Diskperf V1.2
#
# Test name     : Unspecified
# Test date     : Thu Oct 18 15:44:29 2012
# Test machine  : IRIX64 plum 6.5 07202013 IP35
# Test type     : XFS data subvolume
# Test path     : testfile
# Request sizes : min=4096 max=4194304
# Parameters    : direct=1 time=10 scale=1.000 delay=0.000
# XFS file size : 4294967296 bytes
#---------------------------------------------------------
# req_size  fwd_wt  fwd_rd  bwd_wt  bwd_rd  rnd_wt  rnd_rd
#  (bytes)  (MB/s)  (MB/s)  (MB/s)  (MB/s)  (MB/s)  (MB/s)
#---------------------------------------------------------
4096    8.31   14.80    8.85   14.75    5.03    0.92
8192   15.24   26.80   15.12   25.85   12.91    1.78
16384   24.20   41.61   23.51   37.39   22.47    3.50
32768   32.82   52.38   33.18   44.34   23.89    5.86
65536   42.26   59.25   42.64   21.11   31.91    9.75
131072   44.83   63.53   46.15   25.68   32.12   15.08
262144   47.30   69.03   46.97   32.19   28.54   20.39
524288   46.78   69.99   39.64   41.26   37.37   22.61
1048576   45.64   70.74   41.75   43.68   39.97   34.29
2097152   41.44   74.54   42.69   49.77   39.57   40.78
4194304   45.52   72.21   38.86   59.10   40.88   48.07


At the time of the tests the load was roughly ~40% on the netapp; so I'm trying to understand where is the bottleneck, plausible causes:

- qla2342 driver?
- single channel (2Gbps) FC, not appropriate? (This netapp, while using a 2Gbps GBic is pushing only 1Gbps at most). I should trunk more ports.
- gremlins
:Octane2: :Octane: :Octane: :O2: