Cheers Tom,
We get slightly different errors on the nodes -
* rpc error - mailbx1 * security package error - mailbx2
We did have a suggestion to completely disable the firewall service (not just switch it off in the control panel) - this worked until we rebooted both nodes and the problem came back.
Why I asked about the NIC's was because I wondered if the RPC traffic was going out over the data network rather than the iSCSI network. The data network potentially has two addresses per host - its own and the failover cluster IP.
I don't think we're doing anything particularly fancy - the cluster setup is obviously the complicating factor. There are a few NOW articles concerning cluster setup but they tend to relate to traditional shared disk rather than independent disk setups.
Cheers, Raj.
On Wed, Dec 3, 2008 at 11:00 PM, De Wit Tom (Consultant) tom.de.wit@consultant.volvo.com wrote:
Hey Raj,
Setting the Preferred filer address is what will tie those RPC calls to a specific destination IP address. Normally you do this to make that traffic go over a specific source interface. The source IP address woudln't matter that much if the traffic can get back to the same host using that IP address.
Are you experiencing these SD errors on both of your cluster nodes ?
Tom
-----Original Message----- From: Raj Patel [mailto:phigmov@gmail.com] Sent: dinsdag 2 december 2008 22:47 To: De Wit Tom (Consultant) Cc: toasters@mathworks.com Subject: Re: SD 6.01 / W2K8 x64 / iSCSI and NLB Cluster
Cheers Tom
We've done some more troubleshooting -
- if I do a clean w2k8 server install with two nics (one data and the
other iscsi) - SD 6.01 installs and works fine - the initiator even prompts to open the FW ports on the server
This made me think it was either something fishy with my previous install or something funny with the cluster (which is purely failover
- no shared storage).
- if we evict one of the cluster members (there are two mailbox
servers) the SD 6.01 RPC problem goes away immediately
So now I'm running up some test VM's to replicate the setup and run up SD first and then put the cluster on afterwards.
What I'm wondering now is what part of the SD / SAN interaction is making the RPC call - presumably over the clustered address ? Is there any way to tie it to a particular IP or NIC ?
We'll run another test with a couple of vm's, creating the luns and then creating the cluster - see at what point the RPC error occurs.
Cheers,
Raj.
On Tue, Dec 2, 2008 at 10:04 PM, De Wit Tom (Consultant) tom.de.wit@consultant.volvo.com wrote:
I know of two things that can cause such errors:
Fill in the "Preferred filer IP addresses". I don't have SD 6.x, on 4.2.1 it was located in MMC under Storage/Snapdrive/Disks, rightclick Disks, select properties. Then fill in the filer DNS/Netbios name and the corresponding IP address and restart Snapdrive services
Make sure the Snapdrive services is started with a domain user that has both local admin rights on the local server and admin rights on the Netapp filer.
Grtz, Tom