the other nodes see it, but say it's not healthy ;)
filer1a::cluster*> show Node Health Eligibility Epsilon -------------------- ------- ------------ ------------ filer1a true true true filer1b true true false filer2a true true false filer2b true true false filer3a false true false filer4a true true false filer4b true true false filer5a true true false filer5b true true false
Not sure if a node that has aggregates (even if they are empty) can be unjoined or not. I can rig up a test for that, but it'll take a while.
I'm thinking there is some command(s) to tell it to pull over the rdb/whatnot it is upset about from the other nodes in the cluster, which I'm sure support would clue me in on (I see mention in the NetApp communities forums of someone going through the same thing, and support sorted them out) but unfortunately I can't call them about this particular cluster.
On Tue, Oct 15, 2013 at 1:59 PM, tmac tmacmd@gmail.com wrote:
Ouch... weird state
If there is truly nothing on that one node...I would try to unjoin it somehow.
What do the other nodes think?
--tmac
*Tim McCarthy* *Principal Consultant*
Clustered ONTAP Clustered ONTAP
NCDA ID: XK7R3GEKC1QQ2LVD RHCE6 110-107-141https://www.redhat.com/wapps/training/certification/verify.html?certNumber=110-107-141&isSearch=False&verify=Verify NCSIE ID: C14QPHE21FR4YWD4 Expires: 08 November 2014 Current until Aug 02, 2016 Expires: 08 November 2014
On Tue, Oct 15, 2013 at 3:55 PM, Mike Thompson mike.thompson@gmail.comwrote:
nodes get added one at a time
this one has a partner, it just hadn't gotten added to the cluster yet.
On Tue, Oct 15, 2013 at 1:53 PM, tmac tmacmd@gmail.com wrote:
Adding a single node? Um... Unsupported to start with.
8.2 only supports a single node or up to 24 nodes in pairs of same nodes.
--tmac
*Tim McCarthy* *Principal Consultant*
Clustered ONTAP Clustered ONTAP
NCDA ID: XK7R3GEKC1QQ2LVD RHCE6 110-107-141https://www.redhat.com/wapps/training/certification/verify.html?certNumber=110-107-141&isSearch=False&verify=Verify NCSIE ID: C14QPHE21FR4YWD4 Expires: 08 November 2014 Current until Aug 02, 2016 Expires: 08 November 2014
On Tue, Oct 15, 2013 at 3:47 PM, Mike Thompson mike.thompson@gmail.comwrote:
Hey all,
I've got a 8.0.2 c-mode cluster that recently had a single node joined to it, and a few empty aggregates created on it. We had a extended power outtage that required a lot of gear in the data center to get shut down, and since this node in the cluster didn't have any live data or VIFs on it, it got shut down.
A few days later, and we are now powering it up, but get this upon login:
"The contents of the root volume may have changed and the local management databases may be out of sync with the replicated databases due to corruption of NVLOG data during takeover. This node is not fully operational. Contact support personnel for the root volume recovery procedures."
The node comes up fine, can see all it's aggregates, and the other nodes in the cluster can see it via the cluster network, but the node is indeed not fully functional and part of the cluster again. It's aggregates and other info are not visible from the other nodes in the cluster.
Did a wafl_check of the root aggr and vol0 and that came back clean. I seem to recall having been through this before in the past, but can't find anything in my notes.
This particular cluster is not under support, due to some genius decisions by management, so I'm on my own with this.
There are a few empty aggregates on this node, no volumes other than the root vol. Maybe I can force unjoin it from the cluster and rebuild it? Would rather not try to do that. If there is a way to sync up the dbs on the rootvol so it will come back into the cluster, that would be ideal.
Any ideas?
Toasters mailing list Toasters@teaparty.net http://www.teaparty.net/mailman/listinfo/toasters