FC and iSCSI does mean
FAT VMDK, unless you create them manually and specify thin provisioned (not
typical). The Storage VMotion info is good to know – I hope they get that
fixed soon.
Thanks for the
additional info – it’s something for us to watch out for. We went NFS from
the start (and performed P2V and V2V into the NFS-based datastores), but I know
that SVMotion has been used and templates as well. I’ll try to check our
use of templates a bit later today…
Glenn
From: Darren
Sykes [mailto:Darren.Sykes@csr.com]
Sent: Tuesday, October 14, 2008 3:00
AM
To: Glenn Walker; Page,
Jeremy; toasters@mathworks.com
Subject: RE: Snapvault slow on one specific
volume?
Glenn,
That's true; by default all new VM's
created on an NFS volume would be thin provisioned. I'm not sure if that's the
case for templates though (I thought they were created fat on purpose for
performance reasons when deploying them).
Also, we migrated from FC and iSCSI
LUNS (which is basically a file copy) so most of our VM's are fat anyway. From
what I understand using SVMOTION also results in a fat filed VM, though that's
not officially supported on NFS in ESX3.5.
So, in summary there are a few
reasons why you might end up with non-thin provisioned VM's on NFS and may
therefore hit this bug.
Darren
From: Glenn
Walker [mailto:ggwalker@mindspring.com]
Sent: Tue 10/14/2008 03:23
To: Darren Sykes; Page, Jeremy;
toasters@mathworks.com
Subject:
RE: Snapvault slow on one specific volume?
I was under the
impression that ESX over NFS used thin-provisioned VMDKs by default (that’s how
it is in our environment, and all of the files are appearing as
thin-provisioned). Would this then be not the same bug?
Thin-provisioned VMDKs means that the portion of the VMDK not allocated to the
guest would be treated as a sparse file, not a file filled with zeros.
(unless someone decided to perform a full format on the ‘disk’,
perhaps?)
From:
owner-toasters@mathworks.com [mailto:owner-toasters@mathworks.com] On Behalf Of Darren Sykes
Sent: Monday, October 13, 2008 3:51
PM
To: Page, Jeremy;
toasters@mathworks.com
Subject:
RE: Snapvault slow on one specific volume?
Jeremy/All,
Following on from our conversation
offline:
It would seem you (and I) have been
suffering from the bug described here: http://now.netapp.com/NOW/cgi-bin/bol?Type=Detail&Display=281669
We saw it on template volumes. I'm
planning to disable ASIS on that volume to attempt to speed up access.
Obviously, that solution may be less
than useful in your environment where it's live data volumes which benefit from
ASIS.
Darren
From:
owner-toasters@mathworks.com on behalf of Page, Jeremy
Sent: Mon 10/13/2008 13:14
To: toasters@mathworks.com
Subject: Snapvault slow on one specific
volume?
I have an aggr with
two volumes on it. One of them is a 3.5 TB CIFS/NFS share that is reasonably
fast to snapvault and a 1 TB NFS share (ESX VMs) that is exceptionally
slow. As in it’s been doing it’s initial copy for over a week and still
has not finished. NDMP backups of this volume are also quite slow, does anyone
know why it would be so much slower then the other volume using the same
spindles? The filer is not under extreme load, although occasionally it’s pretty
busy. Here is a “normal” sysstat:
CPU
Total Net kB/s Disk kB/s
Tape kB/s Cache Cache CP CP Disk
ops/s in out read write
read write age hit time ty
util
12%
1007 2970 8996 15769 9117
0 0 8 93%
49% : 41%
18%
920 2792 6510 11715 6924
0 0 8 99%
84% T 39%
15%
1276 3580 10469 15942 8041
0 0 10 92% 33%
T 36%
13%
1487 3416 11347 15632 4907
0 0 11 89% 42%
: 43%
17%
1417 3180 9890 14000 9444
0 0 9 98%
79% T 41%
13%
972 3704 9705 15427 9934
0 0 7 92%
46% T 51%
18%
1087 2947 11911 17717 4640
0 0 9 98%
33% T 47%
11%
1204 3358 11219 14090 5159
0 0 7 88% 100%
: 50%
12%
1161 2808 9085 12640 5936
0 0 9 90%
33% T 44%
13%
981 4735 11919 16125 7097
0 0 9 92%
45% : 43%
15%
1158 5780 12480 17565 8266
0 0 10 92% 88%
T 41%
I’m just having
difficulty trying to determine why two volumes on the same spindles would
be so different in the time it takes to do their initial transfer. Also,
the VM’s do not seem slower then those hosted on other aggregates (this one is 3
RG of 11 disks each, Ontap 7.2.4 on a 3070A IBM
rebranded).
To report this email as spam click here.
This
message (including any attachments) contains confidential
and/or proprietary
information intended only for the addressee.
Any unauthorized disclosure,
copying, distribution or reliance on
the contents of this information is
strictly prohibited and may
constitute a violation of law. If you are not
the intended
recipient, please notify the sender immediately by responding
to
this e-mail, and delete the message from your system. If you
have any
questions about this e-mail please notify the sender
immediately.