Val Bercovici valb@netapp.com writes:
Unfortunately, one of the things we don't automatically trigger a failover on in our first Clustered Failover release will be a "normal" interface link failure (OTOH, if a NIC fries its PCI slot, we will failover). Using a Cisco Fast EtherChannel across at least two interfaces would be a nice way to protect yourself from a single interface failure on a filer.
We were hit by an interface link failures once last year. What options are there or will there be (either with or without Clustered Failover) to offer failover or redundancy for failed interfaces?
I'm especially concerned about this as we plan to transition from one or two quad fast ethernet cards on each filer to one or two gigabit ethernet cards per filer.
Thanks.
- Dan
Dan,
We were hit by an interface link failures once last year. What options are there or will there be (either with or without Clustered Failover) to offer failover or redundancy for failed interfaces?
I'm especially concerned about this as we plan to transition from one or two quad fast ethernet cards on each filer to one or two gigabit Ethernet cards per filer.
If interface link failures are your biggest concern, I would stick with a Fast EtherChannel (or similar port trunking technology) configuration which unfortunately cannot be part of a NetApp cluster today.
If Gigabit must be in your short-term future, the good news is that NetApp clusters do support it and it sure is fast! The bad news (you knew it was coming <g>) is that we don't support multiple MAC addresses per NIC with our Gigabit drivers today and consequently you'll need at least two Gigabit interfaces (one active, one passive) per filer in a cluster config. Again, no word on when that will change.
Seems like 6 in one, half dozen in the other but at least you have some options.
-Val.
============================================== Val Bercovici (613)724-8674 Systems Engineer valb@netapp.com Network Appliance www.netapp.com Ottawa, Canada FAST,SIMPLE,RELIABLE ==============================================
-----Original Message----- From: Daniel Quinlan [mailto:quinlan@transmeta.com] Sent: Sunday, September 20, 1998 7:44 PM To: valb@netapp.com Cc: 'Brian Tao'; toasters@mathworks.com Subject: Re: Quad Fast Ethernet vs. Single Fast Ethernets
Val Bercovici valb@netapp.com writes:
Unfortunately, one of the things we don't automatically trigger a failover on in our first Clustered Failover release will be a "normal" interface link failure (OTOH, if a NIC fries its PCI slot, we will failover). Using a Cisco Fast EtherChannel across at least two interfaces would be a nice way to protect yourself from a single interface failure on a filer.
We were hit by an interface link failures once last year. What options are there or will there be (either with or without Clustered Failover) to offer failover or redundancy for failed interfaces?
I'm especially concerned about this as we plan to transition from one or two quad fast ethernet cards on each filer to one or two gigabit ethernet cards per filer.
Thanks.
- Dan
If Gigabit must be in your short-term future, the good news is that NetApp clusters do support it and it sure is fast! The bad news (you knew it was coming <g>) is that we don't support multiple MAC addresses per NIC with our Gigabit drivers today and consequently you'll need at least two Gigabit interfaces (one active, one passive) per filer in a cluster config. Again, no word on when that will change.
It won't ever change, because the drivers can't do what the hardware doesn't support. We've looked at support for the AceNIC 2, which I believe can handle multiple MAC addresses in addition to having much higher performance, but the engineer working on that has been focused on higher-priority projects.
-- Karl
+--- In a previous state of mind, kls@netapp.com (Karl Swartz) wrote: | | It won't ever change, because the drivers can't do what the hardware | doesn't support. We've looked at support for the AceNIC 2, which I | believe can handle multiple MAC addresses in addition to having much | higher performance, but the engineer working on that has been focused | on higher-priority projects.
I for one would rather have a single gigabit nic per filer as opposed to 2 per filer. In large clusters this gets real expensive quick. At $3k per NIC, seems like a better supported card is in order... Plus, what about filers with only single 64bit slots?
Alexei
+--- In a previous state of mind, kls@netapp.com (Karl Swartz) wrote: | | It won't ever change, because the drivers can't do what the hardware | doesn't support. We've looked at support for the AceNIC 2, which I | believe can handle multiple MAC addresses in addition to having much | higher performance, but the engineer working on that has been focused | on higher-priority projects.
I for one would rather have a single gigabit nic per filer as opposed to 2 per filer. In large clusters this gets real expensive quick. At $3k per NIC, seems like a better supported card is in order...
Sure. And our customers who can actually push an F760 hard enough to saturate a single AceNIC 1 would probably rather not have to buy two just to get the full bandwidth.
On the other hand, there probably are a lot of customers who would rather have failover *now* even if it only works with 100base-TX or FDDI (or requires extra GbE NICs). If we told a FDDI customer that we'd be able to give them failover today but we won't because we've not finished a driver for a GbE NIC that can handle multiple MAC addresses, they'd probably be a bit peeved.
Still other customers are happy with the current GbE (or don't care) and don't care about failover, but want other features. Presumably we've decided that it makes more business sense to give some of those wishes higher priority than failover *and* GbE *and* a single NIC.
So, your options at the moment are (a) use GbE but not have failover (b) use failover but not GbE (c) buy a second GbE NIC and have both failover and and GbE (d) wait until we have a GbE NIC and driver which will allow you to have failover with a single GbE
You'll be able to have your cake and eat it, too, but you'll just have to wait a bit.
Plus, what about filers with only single 64bit slots?
That would be the F630. On that filer, we only use the 64-bit slots (there is a second, but it must contain the NVRAM) in 32-bit mode because of some chipset problems. Still, that's the only available primary PCI slot, so the second card would have to go in a secondary slot. On the other hand, if you've taken over for a failed partner, you'll have degraded performance anyway.
-- Karl Swartz - Technical Marketing Engineer Network Appliance Work: kls@netapp.com http://www.netapp.com/ Home: kls@chicago.com http://www.chicago.com/~kls/
On Mon, 21 Sep 1998, Karl Swartz wrote:
Plus, what about filers with only single 64bit slots?
That would be the F630. On that filer, we only use the 64-bit slots (there is a second, but it must contain the NVRAM) in 32-bit mode because of some chipset problems. Still, that's the only available primary PCI slot, so the second card would have to go in a secondary slot. On the other hand, if you've taken over for a failed partner, you'll have degraded performance anyway.
This is an interesting tidbit.
Karl Swartz - Technical Marketing Engineer Network Appliance
It is worth noting that this hardware bug is mentioned nowhere in your 'Technical Marketing' material. Thanks for saving the juicy ones.
matto
--matt@snark.net----------------------------------------------------- Matt Ghali MG406/GM023JP - System Administrator, interQ, Inc AS7506 "Sub-optimal is a state of mind." -Dave Rand, dlr@bungi.com
It is worth noting that this hardware bug is mentioned nowhere in your 'Technical Marketing' material. Thanks for saving the juicy ones.
Nor do we say anywhere that we make use of a 64-bit slot as such. I know we also didn't publish the LADDIS run on the clock-doubled F330 which showed essentially zero performance gain over the stock model, nor do we explicitly say that we don't use the hardware's ability to fault in a paged-out hunk of memory even though we all know that even a 486 could do that. I'm not sure what the point would be of trying to mention every little hardware nugget that we don't choose to take advantage of, so long as we don't misrepresent the product as gaining from such features.
-- Karl Swartz - Technical Marketing Engineer Network Appliance Work: kls@netapp.com http://www.netapp.com/ Home: kls@chicago.com http://www.chicago.com/~kls/
Karl Swartz wrote:
+--- In a previous state of mind, kls@netapp.com (Karl Swartz) wrote:
Plus, what about filers with only single 64bit slots?
That would be the F630. On that filer, we only use the 64-bit slots (there is a second, but it must contain the NVRAM) in 32-bit mode because of some chipset problems. Still, that's the only available primary PCI slot, so the second card would have to go in a secondary slot. On the other hand, if you've taken over for a failed partner, you'll have degraded performance anyway.
Err, well in order to have a Cluster, you need to have a Cluster Interconnect card and yes, you guessed it, it has to go into the unbridged slot - so there really aren't any available for high speed NICs.
-- Karl Swartz - Technical Marketing Engineer Network Appliance Work: kls@netapp.com http://www.netapp.com/ Home: kls@chicago.com http://www.chicago.com/~kls/
+--- In a previous state of mind, Paul Norman pnorman@netapp.com wrote: | | Err, well in order to have a Cluster, you need to have a Cluster Interconnect | card and yes, you guessed it, it has to go into the unbridged slot - so there | really aren't any available for high speed NICs.
So, are you saying that with GbE and a f630, you cannot do failover?
Failover to another interface would be a good feature to have (esp that all the 7xx have 10/100 built in). It is probably not simple to setup, but it would at least make it so I would not have to get dual GbE nics. Then again, if I am dropping an order of magnitude in speed, it may not be useable anyway.
Alexei
Alexei Rodriguez wrote:
+--- In a previous state of mind, Paul Norman pnorman@netapp.com wrote: | | Err, well in order to have a Cluster, you need to have a Cluster Interconnect | card and yes, you guessed it, it has to go into the unbridged slot - so there | really aren't any available for high speed NICs.
So, are you saying that with GbE and a f630, you cannot do failover?
No, you can do Cluster and GbE. You can have up to 4 GbE NICs in each F630. These can be a combination of active and stand-by. All I am saying is that you can't put them in the unbridged slots in the F630.
Failover to another interface would be a good feature to have (esp that all the 7xx have 10/100 built in). It is probably not simple to setup, but it would at least make it so I would not have to get dual GbE nics. Then again, if I am dropping an order of magnitude in speed, it may not be useable anyway.
Current product only allows failover to like interfaces. 100BaseT to 100BaseT or GbE to GbE or FDDI to FDDI. We do not have any plans to implement a GbE to 100BaseT failover, but I'll keep this in mind when we talk about future releases.
Alexei
+--- In a previous state of mind, kls@netapp.com (Karl Swartz) wrote: | | FDDI (or requires extra GbE NICs). If we told a FDDI customer that | we'd be able to give them failover today but we won't because we've | not finished a driver for a GbE NIC that can handle multiple MAC | addresses, they'd probably be a bit peeved.
Well, I am a FDDI customer and I would rather have the GbE driver. I am migrating from FDDI to GbE now. So, for clustering, you have just increased my costs by 50%. How? Well:
sofware for filers & interconnect cards: $30k
additional gigabit cards: $6k additional ports on GbE switch: $8k additional cabling $1k
So, do I want you to complete the drivers for a single card? Damn right.
| (a) use GbE but not have failover | (b) use failover but not GbE
I wish I had known about the delay of GbE and failover *before* I bought it.
| (c) buy a second GbE NIC and have both failover and and GbE
See the above financials fo the reason not to do this.
| (d) wait until we have a GbE NIC and driver which will allow you | to have failover with a single GbE
When will this be?
| You'll be able to have your cake and eat it, too, but you'll just | have to wait a bit.
As long as the windows folks had to wait for acls? I certainly hope not..
| slot. On the other hand, if you've taken over for a failed partner, | you'll have degraded performance anyway.
It depends. If my load is spread such that neither filer is foing more than 50%, then I should not be in degraded mode, right?
Alexei