One final thing I notice is that snapshots are taking sometime to delete which I know is an intensive operations in DOT
Example (at the bottom)
snapid status date ownblks release fsRev name
------ ------ ------------ ------- ------- ----- --------
61 complete Jun 16 15:49 334938 8.1 22331 2014-06-16_1548+0100_daily
60 complete Jun 16 15:48 765 8.1 22331 uk-su-ss007(1575077051)_SV_L_LUN_VOL0199-base.788
44 complete Jun 16 14:08 5415287 8.1 22331 2014-06-16_1408+0100_daily
37 complete Jun 16 10:45 11877364 8.1 22331 2014-06-16_1044+0100_daily
12 complete Jun 15 23:38 16471069 8.1 22331 2014-06-15_2338+0100_daily
250 complete Jun 15 23:36 6949 8.1 22331 2014-06-15_2334+0100_daily
90 complete Jun 15 23:30 16455 8.1 22331 uk-su-ss009(1575007753)_SV_M_LUN_VOL0199.120
58 complete Jun 15 23:29 1044 8.1 22331 2014-06-15_2328+0100_daily
226 complete Jun 15 23:23 40544 8.1 22331 2014-06-15_2322+0100_daily
216 complete Jun 15 15:10 4498948 8.1 22331 2014-06-15_1509+0100_daily
208 complete Jun 15 12:09 2605999 8.1 22331 2014-06-15_1208+0100_daily
196 complete Jun 15 09:24 2187941 8.1 22331 2014-06-15_0924+0100_daily
184 complete Jun 15 01:20 6696250 8.1 22331 2014-06-15_0120+0100_daily
176 complete Jun 15 01:18 5461 8.1 22331 2014-06-15_0117+0100_daily
161 complete Jun 15 01:14 32047 8.1 22331 2014-06-15_0113+0100_daily
148 complete Jun 15 01:11 56057 8.1 22331 2014-06-15_0111+0100_daily
124 complete Jun 14 15:09 4817241 8.1 22331 2014-06-14_1508+0100_daily
112 complete Jun 14 12:10 3349333 8.1 22331 2014-06-14_1210+0100_daily
94 complete Jun 14 10:03 3527119 8.1 22331 2014-06-14_1002+0100_daily
75 complete Jun 13 22:31 20042240 8.1 22331 2014-06-13_2229+0100_daily
69 complete Jun 13 22:27 4954 8.1 22331 2014-06-13_2226+0100_daily
52 complete Jun 13 21:36 343477 8.1 22331 2014-06-13_2135+0100_daily
41 complete Jun 13 20:54 8632101 8.1 22331 2014-06-13_2054+0100_daily
33 complete Jun 13 20:35 4298937 8.1 22331 2014-06-13_2034+0100_daily
19 complete Jun 13 20:32 142680 8.1 22331 2014-06-13_2031+0100_daily
1 complete Jun 13 11:09 6691044 8.1 22331 2014-06-13_1108+0100_daily
238 complete Jun 12 21:46 19191567 8.1 22331 2014-06-12_2145+0100_daily
224 complete Jun 12 21:05 9997690 8.1 22331 2014-06-12_2104+0100_daily
126 complete Jun 08 23:02 109922787 8.1 22331 2014-06-08_2301+0100_daily
130 complete Jun 08 22:58 4925 8.1 22331 2014-06-08_2257+0100_daily
179 complete Jun 08 22:52 21545 8.1 22331 2014-06-08_2251+0100_daily
167 complete Jun 08 22:44 41255 8.1 22331 2014-06-08_2243+0100_daily
85 complete Jun 08 01:05 11868850 8.1 22331 2014-06-08_0104+0100_daily
63 complete Jun 08 01:00 5647 8.1 22331 2014-06-08_0059+0100_daily
144 complete Jun 08 00:55 32549 8.1 22331 2014-06-08_0054+0100_daily
99 complete Jun 08 00:47 38928 8.1 22331 2014-06-08_0045+0100_daily
152 complete Jun 07 00:58 18053974 8.1 22331 2014-06-07_0057+0100_daily
228 complete Jun 07 00:54 4288 8.1 22331 2014-06-07_0052+0100_daily
54 complete Jun 07 00:46 61614 8.1 22331 2014-06-07_0045+0100_daily
15 complete Jun 07 00:39 73153 8.1 22331 2014-06-07_0038+0100_daily
139 complete Jun 06 01:38 34035727 8.1 22331 2014-06-06_0137+0100_daily
49 complete Jun 06 01:35 5386 8.1 22331 2014-06-06_0134+0100_daily
241 complete Jun 06 01:31 43381 8.1 22331 2014-06-06_0131+0100_daily
34 complete Jun 06 01:28 60157 8.1 22331 2014-06-06_0127+0100_daily
25 complete Jun 05 01:57 32400189 8.1 22331 2014-06-05_0156+0100_daily
77 complete Jun 05 01:55 4076 8.1 22331 2014-06-05_0154+0100_daily
200 complete Jun 05 01:51 75258 8.1 22331 2014-06-05_0151+0100_daily
132 complete Jun 05 01:48 85317 8.1 22331 2014-06-05_0147+0100_daily
9 complete Jun 04 00:00 39722504 8.1 22331 2014-06-03_2359+0100_daily
198 complete Jun 03 23:57 5421 8.1 22331 2014-06-03_2355+0100_daily
253 complete Jun 03 23:51 71954 8.1 22331 2014-06-03_2350+0100_daily
13 complete Jun 03 23:45 102485 8.1 22331 2014-06-03_2343+0100_daily
146 complete Jun 03 15:55 9263616 8.1 22331 2014-06-03_1555+0100_daily
128 complete Jun 03 15:14 4950630 8.1 22331 2014-06-03_1514+0100_daily
97 complete Jun 03 00:46 22248178 8.1 22331 2014-06-03_0045+0100_daily
166 complete Jun 03 00:43 4495 8.1 22331 2014-06-03_0042+0100_daily
104 complete Jun 03 00:37 68909 8.1 22331 2014-06-03_0036+0100_daily
70 complete Jun 03 00:31 145138 8.1 22331 2014-06-03_0030+0100_daily
254 complete Jun 02 21:20 35346082 8.1 22331 2014-06-02_2120+0100_daily
202 complete Jun 02 21:19 4114 8.1 22331 2014-06-02_2118+0100_daily
53 complete Jun 02 05:09 528753 8.1 22331 2014-06-02_0509+0100_daily
249 complete Jun 02 05:06 50924 8.1 22331 2014-06-02_0506+0100_daily
50 complete May 28 01:54 64891840 8.1 22331 2014-05-28_0154+0100_daily
5 complete May 28 01:51 6479 8.1 22331 2014-05-28_0150+0100_daily
151 complete May 28 01:49 70924 8.1 22331 2014-05-28_0148+0100_daily
28 complete May 28 01:46 79695 8.1 22331 2014-05-28_0146+0100_daily
129 complete May 27 15:55 10422515 8.1 22331 2014-05-27_1554+0100_daily
16 complete May 27 13:02 9773150 8.1 22331 2014-05-27_1301+0100_daily
157 complete May 27 10:27 10143739 8.1 22331 2014-05-27_1027+0100_daily
105 complete May 26 22:44 14249999 8.1 22331 2014-05-26_2243+0100_daily
252 complete May 26 22:41 7559 8.1 22331 2014-05-26_2240+0100_daily
160 complete May 26 22:37 27078 8.1 22331 2014-05-26_2236+0100_daily
23 complete May 26 22:31 44864 8.1 22331 2014-05-26_2230+0100_daily
232 complete May 26 15:37 4273421 8.1 22331 2014-05-26_1537+0100_daily
136 complete May 26 12:40 2211938 8.1 22331 2014-05-26_1240+0100_daily
91 complete May 26 09:47 3502863 8.1 22331 2014-05-26_0947+0100_daily
21 complete May 25 22:29 4984748 8.1 22331 2014-05-25_2228+0100_daily
4 complete May 25 22:26 6579 8.1 22331 2014-05-25_2225+0100_daily
133 complete May 25 22:21 19720 8.1 22331 2014-05-25_2220+0100_daily
51 complete May 25 22:17 29593 8.1 22331 2014-05-25_2216+0100_daily
194 complete May 25 15:38 2568742 8.1 22331 2014-05-25_1537+0100_daily
127 complete May 25 12:37 2007218 8.1 22331 2014-05-25_1237+0100_daily
30 complete May 25 09:45 1702320 8.1 22331 2014-05-25_0944+0100_daily
172 complete May 24 22:19 4880430 8.1 22331 2014-05-24_2218+0100_daily
143 complete May 24 22:16 4764 8.1 22331 2014-05-24_2215+0100_daily
171 complete May 24 22:08 26268 8.1 22331 2014-05-24_2207+0100_daily
115 complete May 24 22:02 37012 8.1 22331 2014-05-24_2201+0100_daily
3 complete May 24 15:37 3208176 8.1 22331 2014-05-24_1536+0100_daily
163 complete May 24 12:40 2092982 8.1 22331 2014-05-24_1239+0100_daily
246 complete May 24 09:54 2258149 8.1 22331 2014-05-24_0954+0100_daily
164 complete May 24 00:44 7307421 8.1 22331 2014-05-24_0042+0100_daily
173 complete May 24 00:39 8171 8.1 22331 2014-05-24_0037+0100_daily
65 complete May 24 00:34 63931 8.1 22331 2014-05-24_0034+0100_daily
188 complete May 24 00:27 58960 8.1 22331 2014-05-24_0026+0100_daily
153 complete May 23 16:02 8251951 8.1 22331 2014-05-23_1602+0100_daily
110 complete May 23 13:00 8945685 8.1 22331 2014-05-23_1259+0100_daily
87 complete May 23 10:32 9701026 8.1 22331 2014-05-23_1031+0100_daily
14 complete May 23 01:19 17123971 8.1 22331 2014-05-23_0118+0100_daily
235 complete May 23 01:17 10509 8.1 22331 2014-05-23_0117+0100_daily
111 complete May 23 01:15 13058 8.1 22331 2014-05-23_0115+0100_daily
46 complete May 23 01:13 239438 8.1 22331 2014-05-23_0112+0100_daily
230 complete May 22 02:20 39300529 8.1 22331 2014-05-22_0220+0100_daily
89 complete May 22 02:18 5642 8.1 22331 2014-05-22_0217+0100_daily
32 complete May 22 02:14 41995 8.1 22331 2014-05-22_0213+0100_daily
22 complete May 22 02:11 174295 8.1 22331 2014-05-22_0210+0100_daily
93 complete May 21 01:36 42877364 8.1 22331 2014-05-21_0136+0100_daily
255 complete May 21 01:33 7137 8.1 22331 2014-05-21_0132+0100_daily
165 complete May 21 01:29 72219 8.1 22331 2014-05-21_0128+0100_daily
95 complete May 21 01:25 97763 8.1 22331 2014-05-21_0124+0100_daily
212 complete May 20 02:24 36695419 8.1 22331 2014-05-20_0223+0100_daily
177 complete May 20 02:22 2988 8.1 22331 2014-05-20_0222+0100_daily
131 complete May 20 02:18 42998 8.1 22331 2014-05-20_0218+0100_daily
227 complete May 20 02:15 78827 8.1 22331 2014-05-20_0214+0100_daily
155 complete May 18 23:07 39970911 8.1 22331 2014-05-18_2306+0100_daily
123 complete May 18 23:04 5246 8.1 22331 2014-05-18_2302+0100_daily
102 complete May 18 22:57 16970 8.1 22331 2014-05-18_2257+0100_daily
81 complete May 18 22:53 31928 8.1 22331 2014-05-18_2252+0100_daily
251 complete May 17 23:34 14464864 8.1 22331 2014-05-17_2333+0100_daily
147 complete May 17 23:30 8776 8.1 22331 2014-05-17_2328+0100_daily
45 complete May 17 23:25 16371 8.1 22331 2014-05-17_2324+0100_daily
223 complete May 17 23:19 36615 8.1 22331 2014-05-17_2318+0100_daily
98 complete May 16 23:21 20601283 8.1 22331 2014-05-16_2320+0100_daily
66 complete May 16 23:17 6112 8.1 22331 2014-05-16_2316+0100_daily
27 complete May 16 23:10 24164 8.1 22331 2014-05-16_2310+0100_daily
17 complete May 16 23:04 102933 8.1 22331 2014-05-16_2302+0100_daily
225 complete May 15 23:15 34033701 8.1 22331 2014-05-15_2314+0100_daily
187 complete May 15 23:12 5472 8.1 22331 2014-05-15_2311+0100_daily
96 complete May 15 23:07 22786 8.1 22331 2014-05-15_2306+0100_daily
10 complete May 15 23:01 71324 8.1 22331 2014-05-15_2300+0100_daily
191 complete May 14 23:28 33533416 8.1 22331 2014-05-14_2327+0100_daily
80 complete May 14 23:23 6759 8.1 22331 2014-05-14_2322+0100_daily
231 complete May 14 23:17 25124 8.1 22331 2014-05-14_2316+0100_daily
183 complete May 14 23:10 64807 8.1 22331 2014-05-14_2309+0100_daily
118 complete May 13 23:26 32621854 8.1 22331 2014-05-13_2325+0100_daily
203 complete May 13 23:21 5396 8.1 22331 2014-05-13_2320+0100_daily
154 complete May 13 23:14 20931 8.1 22331 2014-05-13_2313+0100_daily
134 complete May 13 23:08 51678 8.1 22331 2014-05-13_2307+0100_daily
101 deleting May 13 01:43 33375746 134.0 (319143/319143 remaining)
67 deleting May 13 01:41 3210 134.0 (319143/319143 remaining)
245 deleting May 13 01:37 0 67.0 (143401/319143 remaining)
From: Jordan Slingerland [mailto:Jordan.Slingerland@independenthealth.com]
Sent: 16 June 2014 21:45
To: Burchell, Will (ITSD); Toasters@teaparty.net
Subject: RE: High CPU VM misalignment confusion
I suppose it could depend but If I had to throw out a number, 10m would probably be a good start in most situations.
--Jordan
From:
Will.Burchell@skanska.co.uk [mailto:Will.Burchell@skanska.co.uk]
Sent: Monday, June 16, 2014 4:38 PM
To: Jordan Slingerland; marcus.nilsson@atea.se;
Toasters@teaparty.net
Subject: RE: High CPU VM misalignment confusion
Thanks
Happy to run wafltop and dump the ouput for comment. Just the standard volume and process?
How long is it a good idea to leave it running to get a useful output
William
From: Jordan Slingerland [mailto:Jordan.Slingerland@independenthealth.com]
Sent: 16 June 2014 21:24
To: Burchell, Will (ITSD); marcus.nilsson@atea.se;
Toasters@teaparty.net
Subject: RE: High CPU VM misalignment confusion
I am not actually sure that is an issue...but I was told it is by IBM n-series support. I do have an open case currently escalated from IBM to Netapp regarding the same issue with stale metadata.
IBM told me to run sis start –s once, and then sis start manually 2x on each volume.(they specifically said sis start (no switches) volume ) had to be run 2x on each volume and a scheduled run would not suffice) Still have some over 100% stale on several
volumes.
Maybe it was just a way to keep me busy for a week. “um yeah, go dedup all your volumes 3 times and come back if it does’t help”
More info here, but it sounds like you already got that.
https://library.netapp.com/ecmdocs/ECMP1368838/html/GUID-5B6B2A2E-FAFD-4A92-B8FB-01E7CD876FE3.html
Perhaps post your wafltop too and someone might be able to point something out in that.
--Jordan
From:
Will.Burchell@skanska.co.uk [mailto:Will.Burchell@skanska.co.uk]
Sent: Monday, June 16, 2014 3:56 PM
To: Jordan Slingerland; marcus.nilsson@atea.se;
Toasters@teaparty.net
Subject: RE: High CPU VM misalignment confusion
No deduplications are running
So I have run sis status –l and can confirm the following SIS jobs and their stale fingerprints. This looks pretty bad. What are we doing wrong here?
We have upgraded from various versions of ONTAP last year and believe we ran into SIS issue but thought they had been cleared by running the sis start –S command to clean them out
We are on 8.1.3P2 and came from 8.0.x into many versions of 8.1.x over the last 18 months
Will
8%
1%
79%
14%
0%
52%
130%
108%
126%
117%
181%
112%
121%
81%
7%
0%
0%
0%
0%
25%
0%
26%
From: Jordan Slingerland [mailto:Jordan.Slingerland@independenthealth.com]
Sent: 16 June 2014 20:47
To: Marcus Nilsson; Burchell, Will (ITSD);
Toasters@teaparty.net
Subject: RE: High CPU VM misalignment confusion
It sounds like you ruled out the obvious, but I will say it anyway. but no deduplciations running, right?
And not so obvious, If none running, look at sis status –l and check if any of the volumes are over 20% in the Stale Fingerprints: column.
--Jordan
From: Marcus Nilsson [mailto:marcus.nilsson@atea.se]
Sent: Monday, June 16, 2014 3:44 PM
To: Will.Burchell@skanska.co.uk; Jordan Slingerland;
Toasters@teaparty.net
Subject: RE: High CPU VM misalignment confusion
Hi,
Might be worth checking out the article at this link
http://www.jk-47.com/2014/02/attack-of-old-bugs-netapp-high-cpu/
We ran into this exact issue after upgrading a system from 8.0.3P2 to 8.1.4P1. A process looping in wafl scan blk_reclaim.
BR Marcus
From:
toasters-bounces@teaparty.net [mailto:toasters-bounces@teaparty.net]
On Behalf Of Will.Burchell@skanska.co.uk
Sent: den 16 juni 2014 21:19
To: Jordan.Slingerland@independenthealth.com;
Toasters@teaparty.net
Subject: RE: High CPU VM misalignment confusion
I am checking all cpus and they are pretty busy
We are in the UK so it’s out of hours (and our nightly process are mostly stopped right now)
We have the issue I mentioned where our exchange LUNs are on the same aggregate together and we have a high IO workload with 6000 mailboxes.
This is the sysstat –M 1 right now as an example. It seems high considering there is no de-dupe and only a single snapmirror running (to do a vol move for our exchange separation problem)
Any other thoughts and I am most interested
William
ANY1+ ANY2+ ANY3+ ANY4+ AVG CPU0 CPU1 CPU2 CPU3 Network Protocol Cluster Storage Raid Target Kahuna WAFL_Ex(Kahu) WAFL_XClean SM_Exempt Cifs Exempt Intr Host Ops/s CP
92% 76% 56% 32% 68% 67% 73% 75% 59% 47% 0% 0% 22% 26% 0% 12% 105%( 65%) 0% 5% 0% 34% 9% 13% 9152 0%
98% 90% 80% 60% 85% 87% 89% 88% 75% 31% 0% 0% 26% 42% 0% 15% 122%( 72%) 21% 11% 0% 52% 9% 9% 5637 50%
99% 92% 80% 55% 84% 86% 90% 91% 69% 31% 0% 0% 29% 58% 0% 16% 103%( 66%) 3% 11% 0% 67% 9% 8% 5962 100%
98% 91% 79% 54% 84% 84% 88% 90% 72% 45% 0% 0% 27% 45% 0% 9% 122%( 78%) 0% 12% 0% 54% 8% 10% 7222 100%
99% 94% 88% 67% 89% 91% 94% 94% 77% 25% 0% 0% 28% 63% 0% 21% 100%( 65%) 25% 9% 0% 71% 8% 7% 4452 100%
97% 91% 79% 52% 83% 84% 89% 91% 67% 39% 0% 0% 30% 51% 0% 8% 113%( 73%) 0% 10% 0% 62% 8% 9% 8253 100%
98% 87% 71% 44% 78% 79% 84% 83% 67% 46% 0% 0% 25% 33% 0% 14% 121%( 74%) 0% 9% 0% 42% 12% 11% 9237 66%
97% 93% 86% 65% 88% 87% 92% 93% 80% 29% 0% 0% 27% 50% 0% 22% 116%( 69%) 24% 9% 0% 59% 9% 8% 5213 63%
97% 85% 69% 42% 76% 77% 83% 85% 60% 37% 0% 0% 28% 42% 0% 9% 109%( 69%) 1% 10% 0% 48% 11% 10% 6795 100%
98% 91% 77% 50% 82% 83% 88% 91% 66% 39% 0% 0% 30% 50% 0% 8% 116%( 73%) 0% 10% 0% 58% 7% 10% 6993 100%
98% 92% 82% 62% 86% 85% 90% 91% 78% 29% 0% 0% 28% 51% 0% 20% 108%( 69%) 21% 14% 0% 59% 6% 8% 5308 90%
100% 97% 91% 65% 90% 92% 94% 96% 80% 30% 0% 0% 30% 59% 0% 16% 120%( 76%) 3% 20% 0% 68% 9% 7% 5593 100%
98% 85% 70% 47% 78% 76% 82% 81% 71% 33% 0% 0% 26% 41% 0% 16% 110%( 70%) 4% 12% 0% 48% 10% 10% 5907 79%
100% 98% 89% 61% 89% 91% 94% 96% 75% 28% 0% 0% 32% 62% 0% 20% 98%( 65%) 17% 10% 0% 73% 8% 7% 5290 100%
98% 91% 77% 50% 82% 80% 85% 89% 72% 33% 0% 0% 30% 48% 0% 21% 108%( 64%) 0% 12% 0% 59% 6% 9% 6047 100%
99% 91% 75% 49% 82% 80% 84% 85% 77% 36% 0% 0% 26% 29% 0% 15% 144%( 80%) 1% 12% 0% 44% 10% 10% 6412 67%
100% 95% 88% 68% 90% 88% 94% 97% 80% 26% 0% 0% 29% 59% 0% 26% 100%( 66%) 23% 16% 0% 65% 8% 7% 4602 100%
98% 87% 74% 48% 79% 78% 86% 90% 63% 30% 0% 0% 29% 52% 0% 9% 105%( 68%) 0% 14% 0% 60% 9% 8% 5533 100%
98% 88% 77% 58% 83% 81% 87% 90% 73% 30% 0% 0% 27% 47% 0% 19% 106%( 66%) 21% 10% 0% 54% 9% 8% 5691 98%
ANY1+ ANY2+ ANY3+ ANY4+ AVG CPU0 CPU1 CPU2 CPU3 Network Protocol Cluster Storage Raid Target Kahuna WAFL_Ex(Kahu) WAFL_XClean SM_Exempt Cifs Exempt Intr Host Ops/s CP
97% 86% 70% 43% 77% 77% 84% 87% 61% 39% 0% 0% 27% 39% 0% 7% 116%( 73%) 0% 11% 0% 49% 11% 10% 7526 100%
97% 86% 70% 44% 78% 80% 85% 87% 61% 34% 0% 0% 28% 44% 0% 9% 108%( 68%) 0% 14% 0% 53% 11% 13% 6308 100%
98% 87% 77% 59% 82% 80% 86% 88% 76% 28% 0% 0% 24% 44% 0% 23% 106%( 66%) 21% 14% 0% 53% 9% 8% 5200 82%
100% 96% 86% 57% 87% 88% 92% 95% 73% 30% 0% 0% 30% 56% 0% 18% 111%( 69%) 3% 18% 0% 68% 6% 8% 5163 100%
98% 90% 78% 55% 83% 82% 88% 91% 69% 32% 0% 0% 28% 44% 0% 11% 119%( 74%) 6% 19% 0% 54% 9% 9% 6148 99%
100% 97% 89% 64% 89% 91% 93% 96% 75% 34% 0% 0% 30% 62% 0% 22% 99%( 65%) 17% 9% 0% 70% 6% 8% 6496 100%
From: Jordan Slingerland [mailto:Jordan.Slingerland@independenthealth.com]
Sent: 16 June 2014 20:14
To: Burchell, Will (ITSD); Toasters@teaparty.net
Subject: RE: High CPU VM misalignment confusion
Even if it is 10k ops after 5 minutes...that is only 33 ops per second. I doubt 33 unaligned ops per second is your cpu issue.
Maybe you can fix that one top talker just to show support that is not the issue? …depending how critical that 1 system is that may or may not be worth fighting over support with.
Now, on to the cpu issue. Are using “sysstat –m 1” to look at all cpus and not only the “ANY” cpu metric right?
If you do , for example, “sysstat –x 1” you are looking at the % of time that ANY of your cpus are busy. Seems to me this metric is nearly completely useless.
--Jordan
From:
Will.Burchell@skanska.co.uk [mailto:Will.Burchell@skanska.co.uk]
Sent: Monday, June 16, 2014 3:07 PM
To: Jordan Slingerland; Toasters@teaparty.net
Subject: RE: High CPU VM misalignment confusion
Thanks
I reset with the –z switch
I then run –d again a 5 minutes later. Many of the counters are in the 10’s so I am happy with this. However 1 server is in the thousands already. This is a windows 2000 server (don’t ask please!) which has a
misaligned C drive but I have used the “functional aligned” datastore in VSC to get around this. I assume nfsstat –d won’t understand that hence the counters in the thousands
William
From: Jordan Slingerland [mailto:Jordan.Slingerland@independenthealth.com]
Sent: 16 June 2014 19:57
To: Burchell, Will (ITSD); Toasters@teaparty.net
Subject: RE: High CPU VM misalignment confusion
First off, make sure the values in nfsstat –d are actually incrementing significantly by running nfsstat –z to clear the counters and then wait a while and looking at nfsstat –d again.
You may find that you are only doing a handful of unaligned ops and not hundreds or thousands per second.
--Jordan
From:
toasters-bounces@teaparty.net [mailto:toasters-bounces@teaparty.net]
On Behalf Of Will.Burchell@skanska.co.uk
Sent: Monday, June 16, 2014 2:50 PM
To: Toasters@teaparty.net
Subject: High CPU VM misalignment confusion
Hello. I am hoping you can guide me in the right direction
We have been experiencing very high CPU load on a 7-mode HA pair of 3270 controllers run 8.1.3P2
We have worked with netapp support on these issues and they note our workload is very high on one controller (where we run our VMware setup from)
We also have so called “bad practice” where we are running our exchange ISCSI LUNs on SATA with logs and dbs on the same aggregate (currently separating this out as I type)
I have been told by support we have VMDK misalignment, however I spent a long time a few months ago resolving this firstly by using the VSC tool to confirm the problem and then fixing it with a combination of
MBRALIGN and VMware converter as a V2V process
The support guy tells me he seems misalignment when he runs nfsstat –d but MBRSCAN shows these are aligned. What is going on here?
Trying to reduce our CPU and IO burden but getting conflicting information.
Finally I think we should look to upgrade to 8.1.4P2 to remove some bugs? We would consider 8.2.x but I don’t think we can as we run Exchange 2010 (using SME 6.x etc)
Thanks in advance
William