12 Posts

2842

August 30th, 2021 13:00

PS6000 shows degraded but GUI shows healthy.

I have a PS6000 that shows a warning of degraded. When I look at the Group manager, all disks show online and status healthy.  I logged into the CLI and ran the diskview -j to see the status of the disks.

Capturex1.PNG

Capturex.PNG

which drive do you think would be causing the degraded status?  Could it be drive 9 ?  GUI shows all are working and spares are fine.  Could it be a false warning?

4 Operator

 • 

1.5K Posts

September 9th, 2021 02:00

There's no command to force a spare into the raidset.   Before you attempted to move it out of the pool will will involve a lot of IO activity on that member, do you have a backup?   

 If so, or when you do.  You  could during a quiet IO period try to fail the controller over and see if the passive controller when it picks up the drives will see the need to rebuild. 

 If you do decide to migrate the array out  of the pool, and if you have critical volumes that will fit on ONE member you can BIND volumes to individual members.  it's slower but would prioritize your volumes vs. moving the member to another pool will not allow that. 

 Also do not DELETE that member.  Create a new pool, and move that member to that new pool.  Once it's there, THEN you can delete it.   When you add it back into the group, please don't use RAID5 anymore.  Use RAID6.  

 Regards, 

Don 

 

 

12 Posts

August 30th, 2021 13:00

Here is what raidtool shows.

Driver Status: Ok
RAID LUN 0 Degraded.
15 Drives (0,10,12,11,8,13,6,1,3,5,7,9,15,f,14)
RAID 5 (64KB sectPerSU)
Capacity 8,065,660,223,488 bytes
Available Drives List: 2,4

 

 

Moderator

 • 

9.4K Posts

August 30th, 2021 17:00

Hi,

What is the status of the controller batteries? It could be drive 9, but it doesn’t seem like it is that many errors. Is the drive and array firmware up to date?

Let us know if you have any additional questions.

4 Operator

 • 

1.5K Posts

August 31st, 2021 07:00

Hello, 

 There should be 16 drives, 0 to 15.  Only 15 drives show in the RAIDset.  The "f" indicates one drive has failed. By process of elimination, it's drive 2 which now shows as a spare has failed out.  What version of firmware is that array running?  I suspect it's older.   The other spare should have kicked in, unless it too at some point failed and has a history of failure tag on it. 

 I would pull drive 2 and 4.  Wait 3 minutes and put drive 4 back in. See if a rebuild starts. 

 Drive 9 has a couple errors but nothing horrible. 

 Regards, 

Don

12 Posts

September 6th, 2021 05:00

Thanks Don and Josh for the responses.  The battery shows good.  We are not running the latest on the firmware. FW.PNG

The last operation we did was to replace the failed drive 4.   Drive 4 had failed a short time ago.  We put in a refurbished drive and added it to the array as a spare.  Does it change the situation and steps recommended.   

4 Operator

 • 

1.5K Posts

September 7th, 2021 04:00

Hello, 

 Yes that firmware was released Sept 2015.  I would suggest removing the two spares 2 and 4.  Wait two minutes and then reinstall drive FOUR (4).   See if a rebuild starts.  Then I would replace drive 2 with another qualified EQL drive. 

  I would also not use this array heavily and make sure you have a backup.  Another drive failure and the array will be offline. 

 Regards, 

Don 

12 Posts

September 7th, 2021 11:00

I remove Disk 2 and then Disk 4 .  Waited over 3 minutes and inserted 4.    Rebuild did not start. What now? 

12 Posts

September 7th, 2021 11:00

I read another post that said that adding a foreign  drive to the SAN will only allow to add as a spare.  They had to force the drive back into the raidset.   Do you know how to  this?

 

https://community.spiceworks.com/topic/1184392-equallogic-reporting-replacement-drive-as-spare

 

 

Moderator

 • 

9.4K Posts

September 7th, 2021 12:00

The drive shows foreign now? Can you set it as a spare?

12 Posts

September 7th, 2021 12:00

Is there a way to force one of the spares back into the raidset?

12 Posts

September 7th, 2021 12:00

Yes.  I can set it as spare.   But I am back at Two spares and no rebuild.

12 Posts

September 7th, 2021 13:00

No. I have not tried that command.  The volumes are up but the raidset is in degraded with too many spares. 

 

Driver Status: Ok
RAID LUN 0 Degraded.
15 Drives (0,10,12,11,8,13,6,1,3,5,7,9,15,f,14)
RAID 5 (64KB sectPerSU)
Capacity 8,065,660,223,488 bytes
Available Drives List: 2,4

 

Moderator

 • 

9.4K Posts

September 7th, 2021 13:00

You can try it and see if it will let you do it.

12 Posts

September 7th, 2021 13:00

I was checking and the other members in the pool have adequate space to accommodate the data on the member with the issue.  Could I move the degraded member out of the Pool?  Is  that a possibility or fix the spare issue and rebuild the only option.

4 Operator

 • 

2.9K Posts

September 7th, 2021 13:00

Looking through the documentation, I'm just finding volume select volume_name online, but not a force option. Have you tried this command already?

No Events found!

Top