[OmniOS-discuss] iSCSI poor write performance
Stephan Budach
stephan.budach at JVM.DE
Wed Sep 23 15:23:24 UTC 2015
Am 23.09.15 um 10:51 schrieb Martin Truhlář:
> Tests revealed, that problem is somewhere in disk array itself. Write performance of disk connected directly (via iSCSI) to KVM is poor as well, even write performance measured on Omnios is very poor. So loop is tightened, but there still remains lot of possible hacks.
> I strived to use professional hw (disks included), so I would try to seek the error in a software setup first. Do you have any ideas where to search first (and second, third...)?
>
> FYI mirror 5 was added lately to the running pool.
>
> pool: dpool
> state: ONLINE
> scan: scrub repaired 0 in 5h33m with 0 errors on Sun Sep 20 00:33:15 2015
> config:
>
> NAME STATE READ WRITE CKSUM CAP Product /napp-it IOstat mess
> dpool ONLINE 0 0 0
> mirror-0 ONLINE 0 0 0
> c1t50014EE00400FA16d0 ONLINE 0 0 0 1 TB WDC WD1002F9YZ-0 S:0 H:0 T:0
> c1t50014EE2B40F14DBd0 ONLINE 0 0 0 1 TB WDC WD1003FBYX-0 S:0 H:0 T:0
> mirror-1 ONLINE 0 0 0
> c1t50014EE05950B131d0 ONLINE 0 0 0 1 TB WDC WD1002F9YZ-0 S:0 H:0 T:0
> c1t50014EE2B5E5A6B8d0 ONLINE 0 0 0 1 TB WDC WD1003FBYZ-0 S:0 H:0 T:0
> mirror-2 ONLINE 0 0 0
> c1t50014EE05958C51Bd0 ONLINE 0 0 0 1 TB WDC WD1002F9YZ-0 S:0 H:0 T:0
> c1t50014EE0595617ACd0 ONLINE 0 0 0 1 TB WDC WD1002F9YZ-0 S:0 H:0 T:0
> mirror-3 ONLINE 0 0 0
> c1t50014EE0AEAE7540d0 ONLINE 0 0 0 1 TB WDC WD1002F9YZ-0 S:0 H:0 T:0
> c1t50014EE0AEAE9B65d0 ONLINE 0 0 0 1 TB WDC WD1002F9YZ-0 S:0 H:0 T:0
> mirror-5 ONLINE 0 0 0
> c1t50014EE0AEABB8E7d0 ONLINE 0 0 0 1 TB WDC WD1002F9YZ-0 S:0 H:0 T:0
> c1t50014EE0AEB44327d0 ONLINE 0 0 0 1 TB WDC WD1002F9YZ-0 S:0 H:0 T:0
> logs
> mirror-4 ONLINE 0 0 0
> c1t55CD2E404B88ABE1d0 ONLINE 0 0 0 120 GB INTEL SSDSC2BW12 S:0 H:0 T:0
> c1t55CD2E404B88E4CFd0 ONLINE 0 0 0 120 GB INTEL SSDSC2BW12 S:0 H:0 T:0
> cache
> c1t55CD2E4000339A59d0 ONLINE 0 0 0 180 GB INTEL SSDSC2BW18 S:0 H:0 T:0
> spares
> c2t2d0 AVAIL 1 TB WDC WD10EFRX-68F S:0 H:0 T:0
>
> errors: No known data errors
>
> Martin
>
>
> -----Original Message-----
> From: Dan McDonald [mailto:danmcd at omniti.com]
> Sent: Wednesday, September 16, 2015 1:51 PM
> To: Martin Truhlář
> Cc: omnios-discuss at lists.omniti.com; Dan McDonald
> Subject: Re: [OmniOS-discuss] iSCSI poor write performance
>
>
>> On Sep 16, 2015, at 4:04 AM, Martin Truhlář <martin.truhlar at archcon.cz> wrote:
>>
>> Yes, I'm aware, that problem can be hidden in many places.
>> MTU is 1500. All nics and their setup are included at this email.
> Start by making your 10GigE network use 9000 MTU. You'll need to configure this on both ends (is this directly-attached 10GigE? Or over a switch?).
>
> Dan
>
To understand what might be going on with our zpool, I'd monitor the
disks using iostat -xme 5 and keep an eye on the errors and svc_t. Just
today I had an issue, where the zpools on one of my OmniOS boxes showed
incredible svc_t for all my zpools, although the drives themselves
showed only moderate ones. The impact was a very high load on the
initiators, which were connected to the targets exported from those zpools.
As I couldn't figure out what was going on, I decided to boot that box
and afterwards things returned to normal again. Luckily, this was only
one side of an ASM mirror, so bouncing the box didn't matter.
Also, when you say, that mirror-5 has been added recently, how is the
data spread across thre vdevs? If the other vdevs have already been
quite full, than that could also lead to significant performance issues.
At any way, you will need to get the performance of your zpools straight
first, before even beginning to think on how to tweak the performance
over the network.
Cheers,
stephan
More information about the OmniOS-discuss
mailing list