[OmniOS-discuss] announcement znapzend

Hafiz Rafibeyli rafibeyli at gmail.com
Mon Aug 11 08:06:42 UTC 2014


Tobias thank you for great job,it was missing backup  part for zfs on omnios,

I think ssh will slow for bigger datasets,as you mention znapzend 0.11 supporting use of mbuffer.

I could not find mbuffer package for omnios,could you explain how to setup/use mbuffer on omnios please?

regards



----- Original Message -----
From: omnios-discuss-request at lists.omniti.com
To: omnios-discuss at lists.omniti.com
Sent: Tuesday, 29 July, 2014 10:29:42 PM
Subject: OmniOS-discuss Digest, Vol 28, Issue 8

Send OmniOS-discuss mailing list submissions to
	omnios-discuss at lists.omniti.com

To subscribe or unsubscribe via the World Wide Web, visit
	http://lists.omniti.com/mailman/listinfo/omnios-discuss
or, via email, send a message with subject or body 'help' to
	omnios-discuss-request at lists.omniti.com

You can reach the person managing the list at
	omnios-discuss-owner at lists.omniti.com

When replying, please edit your Subject line so it is more specific
than "Re: Contents of OmniOS-discuss digest..."


Today's Topics:

   1. announcement znapzend a new zfs backup tool (Tobias Oetiker)
   2. Re: announcement znapzend a new zfs backup tool
      (Theo Schlossnagle)
   3. Re: announcement znapzend a new zfs backup tool (Saso Kiselkov)
   4. Re: Slow scrub performance (wuffers)


----------------------------------------------------------------------

Message: 1
Date: Tue, 29 Jul 2014 17:50:02 +0200 (CEST)
From: Tobias Oetiker <tobi at oetiker.ch>
To: omnios-discuss at lists.omniti.com
Subject: [OmniOS-discuss] announcement znapzend a new zfs backup tool
Message-ID: <alpine.DEB.2.02.1407291748500.6752 at froburg.oetiker.ch>
Content-Type: TEXT/PLAIN; charset=US-ASCII

Just out:

 ZnapZend a Multilevel Backuptool for ZFS

It is on Github. Check out

 http://www.znapzend.org

cheers
tobi

-- 
Tobi Oetiker, OETIKER+PARTNER AG, Aarweg 15 CH-4600 Olten, Switzerland
www.oetiker.ch tobi at oetiker.ch +41 62 775 9902



------------------------------

Message: 2
Date: Tue, 29 Jul 2014 11:54:07 -0400
From: Theo Schlossnagle <jesus at omniti.com>
To: "OmniOS-discuss at lists.omniti.com"
	<omnios-discuss at lists.omniti.com>
Subject: Re: [OmniOS-discuss] announcement znapzend a new zfs backup
	tool
Message-ID:
	<CACLsAptC_wDb+Stkw2-jZkgp7oQZ4OwEUWG_Nnrm_xkaoOkGRg at mail.gmail.com>
Content-Type: text/plain; charset="utf-8"

Awesome!


On Tue, Jul 29, 2014 at 11:50 AM, Tobias Oetiker <tobi at oetiker.ch> wrote:

> Just out:
>
>  ZnapZend a Multilevel Backuptool for ZFS
>
> It is on Github. Check out
>
>  http://www.znapzend.org
>
> cheers
> tobi
>
> --
> Tobi Oetiker, OETIKER+PARTNER AG, Aarweg 15 CH-4600 Olten, Switzerland
> www.oetiker.ch tobi at oetiker.ch +41 62 775 9902
>
> _______________________________________________
> OmniOS-discuss mailing list
> OmniOS-discuss at lists.omniti.com
> http://lists.omniti.com/mailman/listinfo/omnios-discuss
>



-- 

Theo Schlossnagle

http://omniti.com/is/theo-schlossnagle
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://omniosce.org/ml-archive/attachments/20140729/f8adbbf5/attachment-0001.html>

------------------------------

Message: 3
Date: Tue, 29 Jul 2014 17:59:18 +0200
From: Saso Kiselkov <skiselkov.ml at gmail.com>
To: omnios-discuss at lists.omniti.com
Subject: Re: [OmniOS-discuss] announcement znapzend a new zfs backup
	tool
Message-ID: <53D7C4D6.5060308 at gmail.com>
Content-Type: text/plain; charset=ISO-8859-1

On 7/29/14, 5:50 PM, Tobias Oetiker wrote:
> Just out:
> 
>  ZnapZend a Multilevel Backuptool for ZFS
> 
> It is on Github. Check out
> 
>  http://www.znapzend.org

Neat, especially the feature that the backup config is part of a
dataset's properties. Very cool.

-- 
Saso



------------------------------

Message: 4
Date: Tue, 29 Jul 2014 15:29:38 -0400
From: wuffers <moo at wuffers.net>
To: Richard Elling <richard.elling at richardelling.com>
Cc: omnios-discuss <omnios-discuss at lists.omniti.com>
Subject: Re: [OmniOS-discuss] Slow scrub performance
Message-ID:
	<CA+tR_KwX_1HN4tVa+-ZOFJk2mN7RE-nFh31sMcTNo7TJJjfyLg at mail.gmail.com>
Content-Type: text/plain; charset="utf-8"

Going to try to answer both responses in one message..

Short answer, yes. ? Keep in mind that
>
> 1. a scrub runs in the background (so as not to impact production I/O,
> this was not always the case and caused serious issues in the past with a
> pool being unresponsive due to a scrub)
>
> 2. a scrub essentially walks the zpool examining every transaction in
> order (as does a resilver)
>
> So the time to complete a scrub depends on how many write transactions
> since the pool was created (which is generally related to the amount of
> data but not always). You are limited by the random I/O capability of the
> disks involved. With VMs I assume this is a file server, so the I/O size
> will also affect performance.


I haven't noticed any slowdowns in our virtual environments, so I guess
that's a good thing it's so low priority that it doesn't impact workloads.

Run the numbers? you are scanning 24.2TB at about 5.5MB/sec ? 4,613,734
> seconds or 54 days. And that assumes the same rate for all of the scan. The
> rate will change as other I/O competes for resources.
>

The number was fluctuating when I started the scrub, and I had seen it go
as high as 35MB/s at one point. I am certain that our Hyper-V workload has
increased since the last scrub, so this does make sense.


> Looks like you have a fair bit of activity going on (almost 1MB/sec of
> writes per spindle).
>

As Richard correctly states below, this is the aggregate since boot (uptime
~56 days). I have another output from iostat as per his instructions below.


> Since this is storage for VMs, I assume this is the storage server for
> separate compute servers? Have you tuned the block size for the file share
> you are using? That can make a huge difference in performance.
>

Both the Hyper-V and VMware LUNs are created with 64K block sizes. From
what I've read of other performance and tuning articles, that is the
optimal block size (I did some limited testing when first configuring the
SAN, but results were somewhat inconclusive). Hyper-V hosts our testing
environment (we integrate with TFS, a MS product, so we have no choice
here) and probably make up the bulk of the workload (~300+ test VMs with
various OSes). VMware hosts our production servers (Exchange, file servers,
SQL, AD, etc - ~50+ VMs).

I also noted that you only have a single LOG device. Best Practice is to
> mirror log devices so you do not lose any data in flight if hit by a power
> outage (of course, if this server has more UPS runtime that all the clients
> that may not matter).
>

Actually, I do have a mirror ZIL device, it's just disabled at this time
(my ZIL devices are ZeusRAMs). At some point, I was troubleshooting some
kernel panics (turned out to be a faulty SSD on the rpool), and hadn't
re-enabled it yet. Thanks for the reminder (and yes, we do have a UPS as
well).

And oops.. re-attaching the ZIL as a mirror triggered a resilver now,
suspending or canceling the scrub? Will monitor this and restart the scrub
if it doesn't by itself.

  pool: tank
 state: ONLINE
status: One or more devices is currently being resilvered.  The pool will
        continue to function, possibly in a degraded state.
action: Wait for the resilver to complete.
  scan: resilver in progress since Tue Jul 29 14:48:48 2014
    3.89T scanned out of 24.5T at 3.06G/s, 1h55m to go
    0 resilvered, 15.84% done

At least it's going very fast. EDIT: Now about 67% done as I finish writing
this, speed dropping to ~1.3G/s.

maybe, maybe not
>>
>> this is slower than most, surely slower than desired
>>
>
Unfortunately reattaching the mirror to my log device triggered a resilver.
Not sure if this is desired behavior, but yes, 5.5MB/s seems quite slow.
Hopefully after the resilver the scrub will progress where it left off.


> The estimate is often very wrong, especially for busy systems.
>> If this is an older ZFS implementation, this pool is likely getting
>> pounded by the
>> ZFS write throttle. There are some tunings that can be applied, but the
>> old write
>> throttle is not a stable control system, so it will always be a little
>> bit unpredictable.
>>
>
The system is on r151008 (my BE states that I upgraded back in February,
putting me in r151008j or so), with all the pools upgraded for the new
enhancements as well as activating the new L2ARC compression feature.
Reading the release notes, the ZFS write throttle enhancements were in
since r151008e so I should be good there.


> # iostat -xnze
>>
>>
>> Unfortunately, this is the performance since boot and is not suitable for
>> performance
>> analysis unless the system has been rebooted in the past 10 minutes or
>> so. You'll need
>> to post the second batch from "iostat -zxCn 60 2"
>>
>
Ah yes, that was my mistake. Output from second count (before re-attaching
log mirror):

# iostat -zxCn 60 2

                    extended device statistics
    r/s    w/s   kr/s   kw/s wait actv wsvc_t asvc_t  %w  %b device
  255.7 1077.7 6294.0 41335.1  0.0  1.9    0.0    1.4   0 153 c1
    5.3   23.9  118.5  811.9  0.0  0.0    0.0    1.1   0   3
c1t5000C50055F8723Bd0
    5.9   14.5  110.0  834.3  0.0  0.0    0.0    1.3   0   2
c1t5000C50055E66B63d0
    5.6   16.6  123.8  822.7  0.0  0.0    0.0    1.3   0   2
c1t5000C50055F87E73d0
    4.7   27.8  118.6  796.6  0.0  0.0    0.0    1.3   0   3
c1t5000C50055F8BFA3d0
    5.6   14.5  139.7  833.8  0.0  0.0    0.0    1.6   0   3
c1t5000C50055F9E123d0
    4.4   27.1  112.3  825.2  0.0  0.0    0.0    0.8   0   2
c1t5000C50055F9F0B3d0
    5.0   20.2  121.7  803.4  0.0  0.0    0.0    1.2   0   3
c1t5000C50055F9D3B3d0
    5.4   26.4  137.0  857.3  0.0  0.0    0.0    1.4   0   4
c1t5000C50055E4FDE7d0
    4.7   12.3  123.7  832.7  0.0  0.0    0.0    2.0   0   3
c1t5000C50055F9A607d0
    5.0   23.9  125.9  830.9  0.0  0.0    0.0    1.3   0   3
c1t5000C50055F8CDA7d0
    4.5   31.4  112.2  814.6  0.0  0.0    0.0    1.1   0   3
c1t5000C50055E65877d0
    5.2   24.4  130.6  872.5  0.0  0.0    0.0    1.2   0   3
c1t5000C50055F9E7D7d0
    4.1   21.8  103.7  797.2  0.0  0.0    0.0    1.1   0   3
c1t5000C50055FA0AF7d0
    5.5   24.8  129.8  802.8  0.0  0.0    0.0    1.5   0   4
c1t5000C50055F9FE87d0
    5.7   17.7  137.2  797.6  0.0  0.0    0.0    1.4   0   3
c1t5000C50055F9F91Bd0
    6.0   30.6  139.1  852.0  0.0  0.1    0.0    1.5   0   4
c1t5000C50055F9FEABd0
    6.1   34.1  137.8  929.2  0.0  0.1    0.0    1.9   0   6
c1t5000C50055F9F63Bd0
    4.1   15.9  101.8  791.4  0.0  0.0    0.0    1.6   0   3
c1t5000C50055F9F3EBd0
    6.4   23.2  155.2  878.6  0.0  0.0    0.0    1.1   0   3
c1t5000C50055F9F80Bd0
    4.5   23.5  106.2  825.4  0.0  0.0    0.0    1.1   0   3
c1t5000C50055F9FB8Bd0
    4.0   23.2  101.1  788.9  0.0  0.0    0.0    1.3   0   3
c1t5000C50055F9F92Bd0
    4.4   11.3  125.7  782.3  0.0  0.0    0.0    1.9   0   3
c1t5000C50055F8905Fd0
    4.6   20.4  129.2  823.0  0.0  0.0    0.0    1.5   0   3
c1t5000C50055F8D48Fd0
    5.1   19.7  142.9  887.2  0.0  0.0    0.0    1.7   0   3
c1t5000C50055F9F89Fd0
    5.6   11.4  129.1  776.0  0.0  0.0    0.0    1.9   0   3
c1t5000C50055F9EF2Fd0
    5.6   23.7  137.4  811.9  0.0  0.0    0.0    1.2   0   3
c1t5000C50055F8C3ABd0
    6.8   13.9  132.4  834.3  0.0  0.0    0.0    1.8   0   3
c1t5000C50055E66053d0
    5.2   26.7  126.9  857.3  0.0  0.0    0.0    1.2   0   3
c1t5000C50055E66503d0
    4.2   27.1  104.6  825.2  0.0  0.0    0.0    1.0   0   3
c1t5000C50055F9D3E3d0
    5.2   30.7  140.9  852.0  0.0  0.1    0.0    1.5   0   4
c1t5000C50055F84FB7d0
    5.4   16.1  124.3  791.4  0.0  0.0    0.0    1.7   0   3
c1t5000C50055F8E017d0
    3.8   31.4   89.7  814.6  0.0  0.0    0.0    1.1   0   4
c1t5000C50055E579F7d0
    4.6   27.5  116.0  796.6  0.0  0.1    0.0    1.6   0   4
c1t5000C50055E65807d0
    4.0   21.5   99.7  797.2  0.0  0.0    0.0    1.1   0   3
c1t5000C50055F84A97d0
    4.7   20.2  116.3  803.4  0.0  0.0    0.0    1.4   0   3
c1t5000C50055F87D97d0
    5.0   11.5  121.5  776.0  0.0  0.0    0.0    2.0   0   3
c1t5000C50055F9F637d0
    4.9   11.3  112.4  782.3  0.0  0.0    0.0    2.3   0   3
c1t5000C50055E65ABBd0
    5.3   11.8  142.5  832.7  0.0  0.0    0.0    2.4   0   3
c1t5000C50055F8BF9Bd0
    5.0   20.3  121.4  823.0  0.0  0.0    0.0    1.7   0   3
c1t5000C50055F8A22Bd0
    6.6   24.3  170.3  872.5  0.0  0.0    0.0    1.3   0   3
c1t5000C50055F9379Bd0
    5.8   16.3  121.7  822.7  0.0  0.0    0.0    1.3   0   2
c1t5000C50055E57A5Fd0
    5.3   17.7  146.5  797.6  0.0  0.0    0.0    1.4   0   3
c1t5000C50055F8CCAFd0
    5.7   34.1  141.5  929.2  0.0  0.1    0.0    1.7   0   5
c1t5000C50055F8B80Fd0
    5.5   23.8  125.7  830.9  0.0  0.0    0.0    1.2   0   3
c1t5000C50055F9FA1Fd0
    5.0   23.2  127.9  878.6  0.0  0.0    0.0    1.1   0   3
c1t5000C50055E65F0Fd0
    5.2   14.0  163.7  833.8  0.0  0.0    0.0    2.0   0   3
c1t5000C50055F8BE3Fd0
    4.6   18.9  122.8  887.2  0.0  0.0    0.0    1.6   0   3
c1t5000C50055F8B21Fd0
    5.5   23.6  137.4  825.4  0.0  0.0    0.0    1.5   0   3
c1t5000C50055F8A46Fd0
    4.9   24.6  116.7  802.8  0.0  0.0    0.0    1.4   0   4
c1t5000C50055F856CFd0
    4.9   23.4  120.8  788.9  0.0  0.0    0.0    1.4   0   3
c1t5000C50055E6606Fd0
  234.9  170.1 4079.9 11127.8  0.0  0.2    0.0    0.5   0   9 c2
  119.0   28.9 2083.8  670.8  0.0  0.0    0.0    0.3   0   3
c2t500117310015D579d0
  115.9   27.4 1996.1  634.2  0.0  0.0    0.0    0.3   0   3
c2t50011731001631FDd0
    0.0  113.8    0.0 9822.8  0.0  0.1    0.0    1.0   0   2
c2t5000A72A3007811Dd0
    0.1   18.5    0.0   64.8  0.0  0.0    0.0    0.0   0   0 c4
    0.1    9.2    0.0   32.4  0.0  0.0    0.0    0.0   0   0 c4t0d0
    0.0    9.2    0.0   32.4  0.0  0.0    0.0    0.0   0   0 c4t1d0
  229.8   58.1 3987.4 1308.0  0.0  0.1    0.0    0.3   0   6 c12
  114.2   27.7 1994.8  626.0  0.0  0.0    0.0    0.3   0   3
c12t500117310015D59Ed0
  115.5   30.4 1992.6  682.0  0.0  0.0    0.0    0.3   0   3
c12t500117310015D54Ed0
    0.1   17.1    0.0   64.8  0.0  0.0    0.6    0.1   0   0 rpool
  720.3 1298.4 14361.2 53770.8 18.7  2.3    9.3    1.1   6  68 tank

Is 153% busy correct on c1? Seems to me that disks are quite "busy", but
are handling the workload just fine (wait at 6% and asvc_t at 1.1ms)

Interestingly, this is the same output now that the resilver is running:

                    extended device statistics
    r/s    w/s   kr/s   kw/s wait actv wsvc_t asvc_t  %w  %b device
 2876.9 1041.1 25400.7 38189.1  0.0 37.9    0.0    9.7   0 2011 c1
   60.8   26.1  540.1  845.2  0.0  0.7    0.0    8.3   0  39
c1t5000C50055F8723Bd0
   58.4   14.2  511.6  740.7  0.0  0.7    0.0   10.1   0  39
c1t5000C50055E66B63d0
   60.2   16.3  529.3  756.1  0.0  0.8    0.0   10.1   0  41
c1t5000C50055F87E73d0
   57.5   24.9  527.6  841.7  0.0  0.7    0.0    9.0   0  40
c1t5000C50055F8BFA3d0
   57.9   14.5  543.5  765.1  0.0  0.7    0.0    9.8   0  38
c1t5000C50055F9E123d0
   57.9   23.9  516.6  806.9  0.0  0.8    0.0    9.3   0  40
c1t5000C50055F9F0B3d0
   59.8   24.6  554.1  857.5  0.0  0.8    0.0    9.6   0  42
c1t5000C50055F9D3B3d0
   56.5   21.0  480.4  715.7  0.0  0.7    0.0    8.9   0  37
c1t5000C50055E4FDE7d0
   54.8    9.7  473.5  737.9  0.0  0.7    0.0   11.2   0  39
c1t5000C50055F9A607d0
   55.8   20.2  457.3  708.7  0.0  0.7    0.0    9.9   0  40
c1t5000C50055F8CDA7d0
   57.8   28.6  487.0  796.1  0.0  0.9    0.0    9.9   0  45
c1t5000C50055E65877d0
   60.8   27.1  572.6  823.7  0.0  0.8    0.0    8.8   0  41
c1t5000C50055F9E7D7d0
   55.8   21.1  478.2  766.6  0.0  0.7    0.0    9.7   0  40
c1t5000C50055FA0AF7d0
   57.0   22.8  528.3  724.5  0.0  0.8    0.0    9.6   0  41
c1t5000C50055F9FE87d0
   56.2   10.8  465.2  715.6  0.0  0.7    0.0   10.4   0  38
c1t5000C50055F9F91Bd0
   59.2   29.4  524.6  740.9  0.0  0.8    0.0    8.9   0  41
c1t5000C50055F9FEABd0
   57.3   30.7  496.7  788.3  0.0  0.8    0.0    9.1   0  42
c1t5000C50055F9F63Bd0
   55.5   16.3  461.9  652.9  0.0  0.7    0.0   10.1   0  39
c1t5000C50055F9F3EBd0
   57.2   22.1  495.1  701.1  0.0  0.8    0.0    9.8   0  41
c1t5000C50055F9F80Bd0
   59.5   30.2  543.1  741.8  0.0  0.9    0.0    9.6   0  45
c1t5000C50055F9FB8Bd0
   56.5   25.1  515.4  786.9  0.0  0.7    0.0    8.6   0  38
c1t5000C50055F9F92Bd0
   61.8   12.5  540.6  790.9  0.0  0.8    0.0   10.3   0  41
c1t5000C50055F8905Fd0
   57.0   19.8  521.0  774.3  0.0  0.7    0.0    9.6   0  39
c1t5000C50055F8D48Fd0
   56.3   16.3  517.7  724.7  0.0  0.7    0.0    9.9   0  38
c1t5000C50055F9F89Fd0
   57.0   13.4  504.5  790.5  0.0  0.8    0.0   10.7   0  40
c1t5000C50055F9EF2Fd0
   55.0   26.1  477.6  845.2  0.0  0.7    0.0    8.3   0  36
c1t5000C50055F8C3ABd0
   57.8   14.1  518.7  740.7  0.0  0.8    0.0   10.8   0  41
c1t5000C50055E66053d0
   55.9   20.8  490.2  715.7  0.0  0.7    0.0    9.0   0  37
c1t5000C50055E66503d0
   57.0   24.1  509.7  806.9  0.0  0.8    0.0   10.0   0  41
c1t5000C50055F9D3E3d0
   59.1   29.2  504.1  740.9  0.0  0.8    0.0    9.3   0  44
c1t5000C50055F84FB7d0
   54.4   16.3  449.5  652.9  0.0  0.7    0.0   10.4   0  39
c1t5000C50055F8E017d0
   57.8   28.4  503.3  796.1  0.0  0.9    0.0   10.1   0  45
c1t5000C50055E579F7d0
   58.2   24.9  502.0  841.7  0.0  0.8    0.0    9.2   0  40
c1t5000C50055E65807d0
   58.2   20.7  513.4  766.6  0.0  0.8    0.0    9.8   0  41
c1t5000C50055F84A97d0
   56.5   24.9  508.0  857.5  0.0  0.8    0.0    9.2   0  40
c1t5000C50055F87D97d0
   53.4   13.5  449.9  790.5  0.0  0.7    0.0   10.7   0  38
c1t5000C50055F9F637d0
   57.0   11.8  503.0  790.9  0.0  0.7    0.0   10.6   0  39
c1t5000C50055E65ABBd0
   55.4    9.6  461.1  737.9  0.0  0.8    0.0   11.6   0  40
c1t5000C50055F8BF9Bd0
   55.7   19.7  484.6  774.3  0.0  0.7    0.0    9.9   0  40
c1t5000C50055F8A22Bd0
   57.6   27.1  518.2  823.7  0.0  0.8    0.0    8.9   0  40
c1t5000C50055F9379Bd0
   59.6   17.0  528.0  756.1  0.0  0.8    0.0   10.1   0  41
c1t5000C50055E57A5Fd0
   61.2   10.8  530.0  715.6  0.0  0.8    0.0   10.7   0  40
c1t5000C50055F8CCAFd0
   58.0   30.8  493.3  788.3  0.0  0.8    0.0    9.4   0  43
c1t5000C50055F8B80Fd0
   56.5   19.9  490.7  708.7  0.0  0.8    0.0   10.0   0  40
c1t5000C50055F9FA1Fd0
   56.1   22.4  484.2  701.1  0.0  0.7    0.0    9.5   0  39
c1t5000C50055E65F0Fd0
   59.2   14.6  560.9  765.1  0.0  0.7    0.0    9.8   0  39
c1t5000C50055F8BE3Fd0
   57.9   16.2  546.0  724.7  0.0  0.7    0.0   10.1   0  40
c1t5000C50055F8B21Fd0
   59.5   30.0  553.2  741.8  0.0  0.9    0.0    9.8   0  45
c1t5000C50055F8A46Fd0
   57.4   22.5  504.0  724.5  0.0  0.8    0.0    9.6   0  41
c1t5000C50055F856CFd0
   58.4   24.6  531.4  786.9  0.0  0.7    0.0    8.4   0  38
c1t5000C50055E6606Fd0
  511.0  161.4 7572.1 11260.1  0.0  0.3    0.0    0.4   0  14 c2
  252.3   20.1 3776.3  458.9  0.0  0.1    0.0    0.2   0   6
c2t500117310015D579d0
  258.8   18.0 3795.7  350.0  0.0  0.1    0.0    0.2   0   6
c2t50011731001631FDd0
    0.0  123.4    0.0 10451.1  0.0  0.1    0.0    1.0   0   3
c2t5000A72A3007811Dd0
    0.2   16.1    1.9   56.7  0.0  0.0    0.0    0.0   0   0 c4
    0.2    8.1    1.6   28.3  0.0  0.0    0.0    0.0   0   0 c4t0d0
    0.0    8.1    0.3   28.3  0.0  0.0    0.0    0.0   0   0 c4t1d0
  495.6  163.6 7168.9 11290.3  0.0  0.2    0.0    0.4   0  14 c12
    0.0  123.4    0.0 10451.1  0.0  0.1    0.0    1.0   0   3
c12t5000A72B300780FFd0
  248.2   18.1 3645.8  323.0  0.0  0.1    0.0    0.2   0   5
c12t500117310015D59Ed0
  247.4   22.1 3523.1  516.2  0.0  0.1    0.0    0.2   0   6
c12t500117310015D54Ed0
    0.2   14.8    1.9   56.7  0.0  0.0    0.6    0.1   0   0 rpool
 3883.5 1357.7 40141.6 60739.5 22.8 38.6    4.4    7.4  54 100 tank

It is very busy with alot of wait % and higher asvc_t (2011% busy on c1?!).
I'm assuming resilvers are alot more aggressive than scrubs.

There are many variables here, the biggest of which is the current
>> non-scrub load.
>>
>
I might have lost 2 weeks of scrub time, depending on whether the scrub
will resume where it left off. I'll update when I can.
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://omniosce.org/ml-archive/attachments/20140729/1b53a492/attachment.html>

------------------------------

Subject: Digest Footer

_______________________________________________
OmniOS-discuss mailing list
OmniOS-discuss at lists.omniti.com
http://lists.omniti.com/mailman/listinfo/omnios-discuss


------------------------------

End of OmniOS-discuss Digest, Vol 28, Issue 8
*********************************************

-- 
This message has been scanned for viruses and
dangerous content by MailScanner, and is
believed to be clean.



More information about the OmniOS-discuss mailing list