andrew.leecw wrote:
Hi Mamma,
After i connect with 3paradm directly to the 3PAR to run the command, the reply is as follow:
cli% servicemag resume 0 5
Are you sure you want to run servicemag?
select q=quit y=yes n=no: y
servicemag resume 0 5
... onlooping mag 0 5
... firmware is current on pd WWN [5000C5005F7884F8]
... firmware is current on pd WWN [5000C5007EF4FB04] Id [ 5]
... checking for valid disks...
... checking for valid disks...
... disks not normal yet..trying admit/onloop again
... onlooping mag 0 5
... checking for valid disks...
... checking for valid disks...
... disks not normal yet..trying admit/onloop again
... onlooping mag 0 5
... checking for valid disks...
... checking for valid disks...
... disks not normal yet..trying admit/onloop again
... onlooping mag 0 5
... checking for valid disks...
Failed --
disk WWN [5000C5005F7884F8] not admitted
Failed --
disk WWN [5000C5005F7884F8] is not normal. Please use showpd -s to see details of disk state
servicemag resume 0 5 -- Failed
Command failed
Thanks,
Andrew
Now the permissions issue is gone.
Now on to your next issue .... New drive will never go "normal" with your failed node because it will never onloop the missing A-port.
Reviewing "help servicemag" I don't see an option to force the admitting of a drive with a failed controller node.
I'm just throwing it out there.....
You have 1 of 2 nodes failed
You have 3 failed drives (5, 7, 31) and 2 of those have been tried to be replaced (5 and 31).
You have 2 drives (0, 6) throwing "over_temperature_alert" with "disabled_A_port" and "disabled_B_port" which should mean that they are out of the system.
You have 3 drives (16, 22, 28) throwing pre-failure warning that have not been acted on due to the number of failed drives in the system.
So what you do have, is 1 of 2 nodes online and 40 of 48 drives "Okay-ish". I think this system will probably cost more time and money to get back in a healthy state than it would replacing it.