Unsolved
This post is more than 5 years old
4 Posts
0
2377
June 29th, 2006 16:00
testing RAID disk failure?
Hello,
I'm running a Dell PE 2850 with the following RAID configuration:
RAID 5
------
Channel/Disk/Size
0/0/300GB
0/1/300GB
0/2/300GB
------
Channel/Disk/Size
0/0/300GB
0/1/300GB
0/2/300GB
RAID 1
------
Channel/Disk/Size
0/3/70GB
0/4/70GB
------
Channel/Disk/Size
0/3/70GB
0/4/70GB
The OS is running on the RAID1 array. Slot 0/5 is empty.
Just to test how the server acts, I stopped the server, pulled out disk 0/0 and then restarted the server. After checking the SCSI drives, it squawked about a missing drive and stopped booting.
I shutdown the server, put disk 0/0 back in, pulled out disk 0/3, restarted the server. After checking the SCSI drives, again it squawked about a missing drive and stopped booting.
Should the server not have continued booting, and warned me via beep codes or LED indicators?
I expexcted that it would have at least done that when a disk was missing from the RAID1 config?
If I had a spare disk in slot 0/5, would it have made a difference?
Any advice appreciated,
/guser
/guser
Message Edited by guser on 06-29-200612:18 PM


simpleme
24 Posts
0
July 8th, 2006 14:00
If you have installed OpenManage and would monitor you system you would getting error message.
jeffy210
33 Posts
0
July 21st, 2006 17:00
As for your system not wanting to boot, I can only think because it was in a power off state it didn't know what the correct configuration was supposed to be and was wanting some sort of user input in the setup.
Reason I say that is the PERC cards store the configuration in both NVRAM and on each drive. So if the card fails, it can rebuild from the drive configuration.