Cheers Tom,
We get slightly different errors on the nodes -
* rpc error - mailbx1
* security package error - mailbx2
We did have a suggestion to completely disable the firewall service
(not just switch it off in the control panel) - this worked until we
rebooted both nodes and the problem came back.
Why I asked about the NIC's was because I wondered if the RPC traffic
was going out over the data network rather than the iSCSI network. The
data network potentially has two addresses per host - its own and the
failover cluster IP.
I don't think we're doing anything particularly fancy - the cluster
setup is obviously the complicating factor. There are a few NOW
articles concerning cluster setup but they tend to relate to
traditional shared disk rather than independent disk setups.
Cheers,
Raj.
On Wed, Dec 3, 2008 at 11:00 PM, De Wit Tom (Consultant)
<tom.de.wit(a)consultant.volvo.com> wrote:
> Hey Raj,
>
> Setting the Preferred filer address is what will tie those RPC calls to a specific destination IP address. Normally you do this to make that traffic go over a specific source interface. The source IP address woudln't matter that much if the traffic can get back to the same host using that IP address.
>
> Are you experiencing these SD errors on both of your cluster nodes ?
>
> Tom
>
> -----Original Message-----
> From: Raj Patel [mailto:phigmov@gmail.com]
> Sent: dinsdag 2 december 2008 22:47
> To: De Wit Tom (Consultant)
> Cc: toasters(a)mathworks.com
> Subject: Re: SD 6.01 / W2K8 x64 / iSCSI and NLB Cluster
>
> Cheers Tom
>
> We've done some more troubleshooting -
>
> * if I do a clean w2k8 server install with two nics (one data and the
> other iscsi) - SD 6.01 installs and works fine - the initiator even
> prompts to open the FW ports on the server
>
> This made me think it was either something fishy with my previous
> install or something funny with the cluster (which is purely failover
> - no shared storage).
>
> * if we evict one of the cluster members (there are two mailbox
> servers) the SD 6.01 RPC problem goes away immediately
>
> So now I'm running up some test VM's to replicate the setup and run up
> SD first and then put the cluster on afterwards.
>
> What I'm wondering now is what part of the SD / SAN interaction is
> making the RPC call - presumably over the clustered address ? Is
> there any way to tie it to a particular IP or NIC ?
>
> We'll run another test with a couple of vm's, creating the luns and
> then creating the cluster - see at what point the RPC error occurs.
>
>
> Cheers,
>
> Raj.
>
>
> On Tue, Dec 2, 2008 at 10:04 PM, De Wit Tom (Consultant)
> <tom.de.wit(a)consultant.volvo.com> wrote:
>> I know of two things that can cause such errors:
>>
>> - Fill in the "Preferred filer IP addresses". I don't have SD 6.x, on 4.2.1 it was located in MMC under Storage/Snapdrive/Disks, rightclick Disks, select properties. Then fill in the filer DNS/Netbios name and the corresponding IP address and restart Snapdrive services
>>
>> - Make sure the Snapdrive services is started with a domain user that has both local admin rights on the local server and admin rights on the Netapp filer.
>>
>> Grtz,
>> Tom
>>
>