performance with LSI SAS 1064

Eric Anderson anderson at freebsd.org
Thu Aug 30 18:26:53 PDT 2007


Scott Long wrote:
> Eric Anderson wrote:
>> Scott Long wrote:
>>> Lutieri G. wrote:
>>>> 2007/8/30, Eric Anderson <anderson at freebsd.org>:
>>>>> I'm confused - you said in your first post you were getting 3MB/s, 
>>>>> where
>>>>>   above you show something like 55MB/s.
>>>> Sorry! using blogbench i got 3MB/s and 100% busy. Once is 100% busy i
>>>> thinked that 3MB/s is the maximum speed. But i was wrong...
>>>
>>> %busy is a completely useless number for a anything but untagged,
>>> uncached disk subsystems.  It's only an indirect measure of latency, and
>>> there are better tools for measuring latency (gstat).
>>>
>>>>> You didn't say what kind of disks, or how many, the configuration, 
>>>>> etc -
>>>>> so it's hard to answer much.  The 55MB/s seems pretty decent for many
>>>>> hard drives in a sequential use state (which is what dd tests really).
>>>>>
>>>> SAS disks. Seagate, i don't know what is the right model of disks.
>>>>
>>>> Ok. If 55Mb/s is a decent speed i'm happy. I'm getting problems with
>>>> squid cache and maybe should be a problem related with disks. But...
>>>> i'm investigating and discharging problems.
>>>>
>>>>
>>>>> Your errors before were probably caused because your queue depth is 
>>>>> set
>>>>> to 255 (or 256?) and the adapter can't do that many.  You should use
>>>>> camcontrol to reduce it, to maybe 32.  See the camcontrol man page for
>>>>> the right usage.  It's something that needs setting on every boot, 
>>>>> so a
>>>>> startup file is a good place for it maybe.
>>>>>
>>>> Is there any way of get the right number to reduce?!
>>>>
>>>
>>> If you're seeing erratic performance in production _AND_ you're seeing
>>> lots of accompanying messages on the console about tag depth jumping
>>> around, you can use camcontrol to force the depth to a lower number of
>>> you're choosing.  This kind of problem is pretty rare, though.
>>
>> Scott, you are far more of a SCSI guru than I, so please correct me if 
>> this is incorrect.  Can't you get a good estimate, by knowing the 
>> queue depth of the target(s), and dividing it by the number of 
>> initiators?  So in his case, he has one initiator, and (let's say) one 
>> target.  If the queue depth of the target (being the Seagate SAS 
>> drive) is 128 (see Seagate's paper here: 
>> http://www.seagate.com/staticfiles/support/disc/manuals/enterprise/savvio/Savvio%2015K.1/SAS/100407739b.pdf 
>> ), then he should have to reduce it down from 25[56] to 128, correct?
>>
>> With QLogic cards connected to a fabric, I saw queue depth issues 
>> under heavy load.
>>
> 
> I understand what you're saying, but you're a bit confused on 
> terminology =-)

Figured as much :)

> There are two factors in the calculation.  One is how many transactions
> the controller (the initiator) can have in progress as once.  This is
> really independent of what the disks are capable of or how many disks 
> are on the bus.  This is normally known to the driver in some 
> chip-specific way.  Second is how many tagged transactions a disk can
> handle.  This actually isn't something that can be discovered in a
> generic way, so the SCSI layer in FreeBSD guesses, and then revises that
> guess over time based on feedback from the drive.
> 
> Manually setting the queue depth is not something that he "should have 
> to [do]".  It perfectly normal to get console messages on occasion about
> the OS re-adjusting the depth.  Where it becomes a problem is in high
> latency topologies (like FC fabrics) and buggy drive firmware where the 
> algorithm winds up thrashing a bit.  For direct attached SAS disks, I
> highly doubt that it is needed.  Playing a guessing game with this will
> almost certainly result in lower performance.

Ok, that makes sense - my experience was in a heavily loaded fabric 
environment.

Thanks for the great info!

Eric




More information about the freebsd-scsi mailing list