You may need to put the node where it is physically attached into maintenance mode and do it from there. You might also have luck using the nodeshell commands. I recovered this for a customer one, I forgot which method I had to use.
Sent from Mobile Outlook
On Thu, Jun 25, 2015 at 1:56 PM -0700, "Momonth" momonth@gmail.com wrote:
Hi, I've been configuring aggregates on 6 nodes cluster today and managed to assign a disk to a node that the disk is not "cabled" to, ie it sits in a disk shelf that is managed by a different HA-pair: na201::> disk show -container-type unknown Usable Disk Container Container Disk Size Shelf Bay Type Type Name Owner---------------- ---------- ----- --- ------- ----------- --------- --------6.45.16 - 45 16 SAS unknown - na201node-3a The disk is "cabled" to na201node-1a, so I'm trying to change the ownership: na201::> storage disk assign -disk 6.45.16 -owner na201node-1a -force
Error: command failed: Failed to assign disks. Reason: Disk 6.45.16 does not exist. OK, this time trying to remove the ownership: na201::> storage disk removeowner -disk 6.45.16 -force Warning: Disks may be automatically assigned to the node because the disk's auto-assign option is enabled. If the affected volumes are not offline, the disks may be auto-assigned during the remove owner operation, which will cause unexpected results. To verify that the volumes are offline, abort this command and use "volume show".Do you want to continue? {y|n}: y Error: command failed: Disk 6.45.16 is not connected to na201node-3a I'm out of ideas at this point .. any hints? Cheers,