[OmniOS-discuss] zfs pool 100% busy, disks less than 10%

Rune Tipsmark rt at steait.net
Mon Nov 3 02:24:29 UTC 2014


Looking a bit more at these numbers, am I seeing twice the actual rate due to mirroring?
How does compression affect the numbers?

Say I have 1 vdev mirrored and I see the pool writing 100 MB/sec, its what? 50 MB each disk? But from the client side only 50 MB/sec total? What if I compress it at the same time at say 1.50 ratio, will the pool show 100 MB/sec and the client write 75 MB/sec actual?

Br,
Rune

-----Original Message-----
From: Richard Elling [mailto:richard.elling at richardelling.com] 
Sent: Sunday, November 02, 2014 6:07 PM
To: Rune Tipsmark
Cc: Eric Sproul; omnios-discuss at lists.omniti.com
Subject: Re: [OmniOS-discuss] zfs pool 100% busy, disks less than 10%


On Oct 31, 2014, at 6:07 PM, Rune Tipsmark <rt at steait.net> wrote:

> So actually started storage vmotions on 3 host, 6 concurrent and am 
> getting about 1GB/sec Guess I need more hosts to really push this, the disk are not more than 20-25% busy, so in theory I could push a bit more.
> 
> I think this is resolved for now.... cpu sitting at 30-40% usage while 
> moving 1GB/sec

Yes, that seems about right.
 -- richard

> 
> Iostat -xn 1
> pool04       396G  39.5T      9  15.9K   325K  1.01G
> pool04       396G  39.5T      7  17.0K   270K  1.03G
> pool04       396G  39.5T     12  17.4K   558K  1.10G
> pool04       396G  39.5T     10  16.9K   442K  1.03G
> pool04       397G  39.5T      6  16.9K   332K  1021M
> pool04       397G  39.5T      1  16.3K  74.9K  1.01G
> pool04       397G  39.5T      8  17.0K   433K  1.05G
> pool04       397G  39.5T     20  17.1K   716K  1023M
> pool04       397G  39.5T     11  18.3K   425K  1.14G
> pool04       398G  39.5T      0  18.3K  65.9K  1.11G
> pool04       398G  39.5T     16  17.9K   551K  1.06G
> pool04       398G  39.5T      0  16.8K   105K  1.03G
> pool04       398G  39.5T      1  18.2K   124K  1.11G
> pool04       398G  39.5T      0  17.1K  45.9K  1.05G
> pool04       399G  39.5T      6  17.3K   454K  1.08G
> pool04       399G  39.5T      0  17.9K      0  1.06G
> pool04       399G  39.5T      2  16.9K   116K  1.04G
> pool04       399G  39.5T      2  18.8K   130K  1.09G
> pool04       399G  39.5T      0  17.6K      0  1.03G
> pool04       400G  39.5T      3  17.5K   155K  1.04G
> pool04       400G  39.5T      0  17.6K  31.5K  1.03G
> 
> -----Original Message-----
> From: OmniOS-discuss [mailto:omnios-discuss-bounces at lists.omniti.com] 
> On Behalf Of Rune Tipsmark
> Sent: Friday, October 31, 2014 12:38 PM
> To: Richard Elling; Eric Sproul
> Cc: omnios-discuss at lists.omniti.com
> Subject: Re: [OmniOS-discuss] zfs pool 100% busy, disks less than 10%
> 
> Ok, makes sense.
> What other kind of  indicators can I look at?
> 
> I get decent results from DD but still feels a bit slow...
> 
> Compression lz4 should not slow it down right? Cpu is not doing much when copying data over, maybe 15% busy or so... 
> 
> Sync=always, block size 1M
> 204800000000 bytes (205 GB) copied, 296.379 s, 691 MB/s
> real    4m56.382s
> user    0m0.461s
> sys     3m12.662s
> 
> Sync=disabled, block size 1M
> 204800000000 bytes (205 GB) copied, 117.774 s, 1.7 GB/s
> real    1m57.777s
> user    0m0.237s
> sys     1m57.466s
> 
> ... while doing this I was looking at my FIO cards, I think the reason is that the SLC's need more power to deliver higher performance, they are supposed to deliver 1.5GB/sec but only delivers around 350MB/sec each....
> 
> Now looking for aux power cables and will retest...
> 
> Br,
> Rune
> 
> -----Original Message-----
> From: Richard Elling [mailto:richard.elling at richardelling.com]
> Sent: Friday, October 31, 2014 9:03 AM
> To: Eric Sproul
> Cc: Rune Tipsmark; omnios-discuss at lists.omniti.com
> Subject: Re: [OmniOS-discuss] zfs pool 100% busy, disks less than 10%
> 
> 
> On Oct 31, 2014, at 7:14 AM, Eric Sproul <eric.sproul at circonus.com> wrote:
> 
>> On Fri, Oct 31, 2014 at 2:33 AM, Rune Tipsmark <rt at steait.net> wrote:
>> 
>>> Why is this pool showing near 100% busy when the underlying disks 
>>> are doing nothing at all....
>> 
>> Simply put, it's just how the accounting works in iostat.  It treats 
>> the pool like any other device, so if there is even one outstanding 
>> request to the pool, it counts towards the busy%.  Keith W. from 
>> Joyent explained this recently on the illumos-zfs list:
>> http://www.listbox.com/member/archive/182191/2014/10/sort/time_rev/pa
>> g 
>> e/3/entry/18:93/20141017161955:F3E11AB2-563A-11E4-8EDC-D0C677981E2F/
>> 
>> The TL;DR is: if your pool has more than one disk in it, the 
>> pool-wide busy% is useless.
> 
> FWIW, we use %busy as an indicator that we can ignore a device/subsystem when looking for performance problems. We don't use it as an indicator of problems. In other words, if the device isn't > 10% busy, forgetabouddit. If it is more busy, look in more detail at the meaningful performance indicators.
> -- richard
> 
> _______________________________________________
> OmniOS-discuss mailing list
> OmniOS-discuss at lists.omniti.com
> http://lists.omniti.com/mailman/listinfo/omnios-discuss



More information about the OmniOS-discuss mailing list