hi,

I am migrating 2 7mode iscsi luns to an iscsi svm on a cdot cluster (8.3.2). The migration is fine, I am now testing the luns from the client os (oracle linux 5.6, I do not have any say on the client os,).

So I have followed the instructions in here: https://library.netapp.com/ecmdocs/ECMP1654943/html/index.html

I see the luns, but now I see 2x 4 luns, because the cdot cluster has 4 iscsi lifs.

So I need to configure multipathd.

I have this /etc/multipath.conf

defaults {
    user_friendly_names yes
}

blacklist {
           devnode "^(ram|raw|loop|fd|md|dm-|sr|scd|st)[0-9]*"
           devnode "^sd[a-c]"
}

multipaths {
    multipath {
        wwid 3600a098038303370562b4946426b6742
        alias netapp_oracle_recovery
    }
    multipath {
        wwid 3600a098038303370562b4946426b6743
        alias netapp_oracle_datafile
    }

devices {
    device {
               vendor                  "NETAPP"
               product                 "LUN"
               path_grouping_policy    group_by_prio
               features                "1 queue_if_no_path"
               prio_callout            "/sbin/mpath_prio_alua /dev/%n"
               path_checker            directio
               path_selector           "round-robin 0"
               failback                immediate
               hardware_handler        "1 alua"
               rr_weight               uniform
               rr_min_io               128
               getuid_callout          "/sbin/scsi_id -g -u -s /block/%n"

    }
}

I blacklist sda, sdb and sdc (local vmware disks).

This is the output of multipath -ll:

#  multipath -ll
netapp_oracle_datafile (3600a098038303370562b4946426b6743) dm-3 NETAPP,LUN C-Mode
size=350G features='1 queue_if_no_path' hwhandler='0' wp=rw
|-+- policy='round-robin 0' prio=4 status=active
| |- 6:0:0:1 sdg 8:96  active ready running
| `- 3:0:0:1 sdf 8:80  active ready running
`-+- policy='round-robin 0' prio=1 status=enabled
  |- 4:0:0:1 sdj 8:144 active ready running
  `- 5:0:0:1 sdk 8:160 active ready running
netapp_oracle_recovery (3600a098038303370562b4946426b6742) dm-2 NETAPP,LUN C-Mode
size=300G features='1 queue_if_no_path' hwhandler='0' wp=rw
|-+- policy='round-robin 0' prio=4 status=active
| |- 3:0:0:0 sdd 8:48  active ready running
| `- 6:0:0:0 sde 8:64  active ready running
`-+- policy='round-robin 0' prio=1 status=enabled
  |- 5:0:0:0 sdh 8:112 active ready running
  `- 4:0:0:0 sdi 8:128 active ready running

And after installing the netapp host tools, this is the output of sanlun lun show -p

# sanlun lun show -p

                    ONTAP Path: ALR-SVM-iSCSI:/vol/vol_oracle_datafile/q_oracle_datafile/lun_oracle_datafile
                           LUN: 1
                      LUN Size: 350.1g
                       Product: cDOT
                   Host Device: netapp_oracle_datafile(3600a098038303370562b4946426b6743)
              Multipath Policy: round-robin 0
            Multipath Provider: Native
--------- ---------- ------- ------------ ----------------------------------------------
host      vserver                  
path      path       /dev/   host         vserver
state     type       node    adapter      LIF
--------- ---------- ------- ------------ ----------------------------------------------
up        primary    sdg     host6        iscsi_lif02b                                 
up        primary    sdf     host3        iscsi_lif02a                                 
up        secondary  sdj     host4        iscsi_lif01a                                 
up        secondary  sdk     host5        iscsi_lif01b                                 

                    ONTAP Path: ALR-SVM-iSCSI:/vol/vol_oracle_recovery/q_oracle_recovery/lun_oracle_recovery
                           LUN: 0
                      LUN Size: 300g
                       Product: cDOT
                   Host Device: netapp_oracle_recovery(3600a098038303370562b4946426b6742)
              Multipath Policy: round-robin 0
            Multipath Provider: Native
--------- ---------- ------- ------------ ----------------------------------------------
host      vserver                  
path      path       /dev/   host         vserver
state     type       node    adapter      LIF
--------- ---------- ------- ------------ ----------------------------------------------
up        primary    sdd     host3        iscsi_lif02a                                 
up        primary    sde     host6        iscsi_lif02b                                 
up        secondary  sdh     host5        iscsi_lif01b                                 
up        secondary  sdi     host4        iscsi_lif01a


and this is sanlun lun show:
sanlun lun show
controller(7mode/E-Series)/                                                                  device          host                  lun   
vserver(cDOT/FlashRay)        lun-pathname                                                   filename        adapter    protocol   size    product
-----------------------------------------------------------------------------------------------------------------------------------------------
ALR-SVM-iSCSI                 /vol/vol_oracle_datafile/q_oracle_datafile/lun_oracle_datafile /dev/sdk        host5      iSCSI      350.1g  cDOT
ALR-SVM-iSCSI                 /vol/vol_oracle_recovery/q_oracle_recovery/lun_oracle_recovery /dev/sdi        host4      iSCSI      300g    cDOT
ALR-SVM-iSCSI                 /vol/vol_oracle_datafile/q_oracle_datafile/lun_oracle_datafile /dev/sdj        host4      iSCSI      350.1g  cDOT
ALR-SVM-iSCSI                 /vol/vol_oracle_recovery/q_oracle_recovery/lun_oracle_recovery /dev/sdh        host5      iSCSI      300g    cDOT
ALR-SVM-iSCSI                 /vol/vol_oracle_datafile/q_oracle_datafile/lun_oracle_datafile /dev/sdf        host3      iSCSI      350.1g  cDOT
ALR-SVM-iSCSI                 /vol/vol_oracle_datafile/q_oracle_datafile/lun_oracle_datafile /dev/sdg        host6      iSCSI      350.1g  cDOT
ALR-SVM-iSCSI                 /vol/vol_oracle_recovery/q_oracle_recovery/lun_oracle_recovery /dev/sde        host6      iSCSI      300g    cDOT
ALR-SVM-iSCSI                 /vol/vol_oracle_recovery/q_oracle_recovery/lun_oracle_recovery /dev/sdd        host3      iSCSI      300g    cDOT


In /dev/mapper I see this:

 ls -l /dev/mapper/
total 0
crw------- 1 root root  10, 62 Jul 13 14:00 control
brw-rw---- 1 root disk 253,  3 Jul 13 14:01 netapp_oracle_datafile
brw-rw---- 1 root disk 253,  4 Jul 13 14:01 netapp_oracle_datafilep1
brw-rw---- 1 root disk 253,  2 Jul 13 14:01 netapp_oracle_recovery
brw-rw---- 1 root disk 253,  5 Jul 13 14:01 netapp_oracle_recoveryp1
brw-rw---- 1 root disk 253,  0 Jul 13 14:00 VolGroup00-LogVol00
brw-rw---- 1 root disk 253,  1 Jul 13 14:00 VolGroup00-LogVol01

Is this ok? What do I have to use in /etc/fstab? /dev/mapper/netapp_oracle_datafile en /dev/mapper/netapp_oracle_recovery? Or am I doing something terribly wrong. I called with support but as this is no break or fix stuff, they recommend calling professional services ...

Apologies for the wall of text.

Thanks in advance.

--
Groeten,
natxo