Unsolved

This post is more than 5 years old

36 Posts

24603

December 26th, 2005 02:00

How to upgrade system without reinstall Linux

Dear ALL,
i have some trouble with my system. I have PowerEdge 8450, which i want to upgrade from SCSI to RAID PERC3/DC. On Microsoft based O/S, i didn't have any difficulty. Just install the driver, and clone (symantec ghost or acronis true image) to new RAID disk storage. Everything done well.
 
The problem is in Linux. ES 3 and Linux 7.2. I have simulate in Desktop PC. it work well when i clone the original disk to new RAID disk. But, when i trying to do that on PowerEdge 8450, i have message GRUB Read Error, Kernel Panic, and any other error message.
 
Can somebody help me? How to upgrade system from SCSI to RAID in PowerEdge 8450, without reinstall O/S??
 
Thanks a lot

2 Intern

 • 

815 Posts

December 28th, 2005 13:00

There is a good chance that it has to due with your grub.conf configuation.  What I would do is get both the SCSI and PERC installed and seeing file systems when booting from SCSI.  Then copy the contents of your root volume from the SCSI file system to the RAID file system.  Do the same for all other file systems that are on the SCSI device.  Determine the new root volume device name, ie /dev/sdb2.  Then add a grub.conf stanza entry to match that new device name and reboot.  I would keep a copy of the old grub stanza just to ensure you have a bootable system.

Once your system boots, you can remove the old SCSI controller, and update your grub.conf kernel SCSI device again if needed. 

36 Posts

December 29th, 2005 01:00

Dear Eric,

Thanks a lot for your reply.

I have tried the way you said, but didn't success. Do you think it is possibly because of something protection from the server such as ESCD(Extended System Configuration Data Area)?? Because when i tried to cloning the entire Linux system in another PC, i have no problem with it. All the problem appear when i cloning the system in DELL PowerEdge 8450.

Which one is better, Cloning Drive (Norton Ghost, Acronis True Image) or manual copy the content?

May let me tell you the procedure i have done, so you can get clearly :

1. Cloning the original disk to new drive, with the same capacity.(to make sure the original disk is safe. i don't want to take any risk). The clone success, but when i start the new drive that has cloned, the GRUB cannot appear. So i get the rescue disk to start the system and overwrite the GRUB loader. The system start normally.

FYI, i didn't find this kind of error when i simulate on another system (non-branded server).

2. Install MegaRAID PERC3/DC to the server. Linux can detect the driver automatically.

3. Make a RAID-1 drive from 2 new HDD, and the copy the Drive(on step 2) to the new RAID partition. I got message "GRUB READ ERROR". I also didn't find this error when i simulate on another server.

For this error, i also have rescue disk and edit grub.conf to start for RAID disk. But didn't success. Some times i got message "kernel panic"

So do you think what should i do?

 

Thanks a lot

2 Intern

 • 

815 Posts

December 29th, 2005 12:00

I still believe the problem is with grub.  Either by cloning or manually coping the data should result in successful migration to the RAID controller.  The trick is the procedure of doing so.

If you can successfully boot the system using a rescue floppy disk and recognize the boot and root partitions, then you can probably bring up the system quickly.

I would first start by recreating the MBR using grub-install.  This will remake the stage 1 and stage 2 boot loaders that Grub will use on the new disk.

I would then ensure that the kernel stanza has the correct root partition declared.  If you are using a Label in grub, I would change it to reflect the actual SCSI device the root partition is on.  i.e.  /dev/sda2

Lastly I would double check the modules.conf to ensure the RAID controller module is listed as the first SCSI controller, and then remake the initrd.

36 Posts

January 9th, 2006 03:00

for detail i just resend my case
This is my case

1. I have three harddrive . The first harddrive is source disk and the second and the third as destination disk cause i want to upgrade my system to raid
I use AMI MEGARAID CARD. The old system use sym53c as scsi card. I want upgrade my server without reinstall linux so
I create mirror from bios megaraid card. and then I clone the source disk to megarid disk that have set as mirror disk .

My Old system use Redhat AS 3 with kernel 2.4.21-4.ELsmp
My machine Dell PowerEdge 8450

2. All hardisk have the same size .
3. After I clone I remove the source disk and try to boot from megaraid harddrive
4. The grub only appear sign " GRUB _"
5. So I try rewrite grub
6. first I reboot
7. And then input the cdrom disk I redhat installation
8. I type linux rescue noprobe and then enter
9. And then I give the disket driver for megaraid2
10. The console bash appear after cdrom recognize and found thE REDHAT installation partition.
11. I try rewrite grub three times

The first time
1. I type chroot /mnt/sysimage
2. I type grub-install --root-directory=/ '(hd0)'
3. reboot


The second time
1. I type chroot /mnt/sysimage
2. Edit /boot/grub/grub.conf
I unremark #boot=/dev/sda
I save
3. I type
grub --device-map=/boot/grub/device-map --config-file=/boot/grub/grub.conf --no-floppy

4. grub > root (hd0,0)
5. grub > setup (hd0)
6. grub > quit
7. reboot


The third time
1. I boot from floopy disk bootloader that I have created before
2. I boot and chose the kernel load to the hardisk
3. After load into the system normally
4. I type grub
5. grub > device (hd0) /dev/sda
6. grub > root (hd0,0)
7. grub > install (hd0,0)/boot/grub/stage1 d (hd0) (hd0,0)/boot/grub/stage2 0x8000 p (hd0,0)/boot/grub/menu.lst
8. grub > quit
9. reboot


All of that get result "GRUB READ ERROR" when boot from hardisk after I tried three attempt above.

Please can you help me . How to solve it
 
Can you give the way to rewrite grub that could succeed solve it...
 


Please.... reply soon.. I need that.

2 Intern

 • 

815 Posts

January 9th, 2006 12:00

Since you are able to boot using a floppy disk, the problem most likely is grub recognizing the boot device properly.

1. ensure bios is enabled

2. then post what you have in fdisk /l, /proc/scsi/scsi, and modules.conf

2 Intern

 • 

815 Posts

January 11th, 2006 01:00

Right now I'm only seeing 1 hard disk in the system, and it looks like it is connected to the SCSI controller, which is SCSI Bus 1. The MegaRAID driver is configured so that it would be SCSI Bus 0, but I don't see any drives listed in /proc/scsi/scsi from the MegaRAID Controller.
To verify if the SCSI BIOS is enabled on the MegaRAID, monitor the POST of your server, and when the MegaRAID BIOS appears, the last thing it should report is the BIOS is Enabled.
You can also go into the RAID Controller itself and verify. Did you configure your container on the RAID controller?

36 Posts

January 11th, 2006 01:00

dear Sir,
thanks a lot for your reply.
 
may you tell me detail, how to ensure bios is enabled?
 
The result of :
 
fdisk -l
Disk /dev/sda: 73.4 GB, 73407820800 bytes
255 heads, 63 sectors/track, 8924 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes

Device Boot    Start       End    Blocks   Id  System
/dev/sda1   *         1      1323  10626966   83  Linux
/dev/sda2          1324      8669  59006745   83  Linux
/dev/sda3          8670      8924   2048287+  82  Linux swap

proc/scsi/scsi
Attached devices:
Host: scsi0 Channel: 00 Id: 01 Lun: 00
  Vendor: SEAGATE  Model: ST373405LC       Rev: 2203
  Type:   Direct-Access                    ANSI SCSI revision: 03
 
 
/etc/modules.conf
alias eth0 e100
alias scsi_hostadapter megaraid2
alias scsi_hostadapter1 sym53c8xx
alias scsi_hostadapter2 qla2200
alias usb-controller usb-uhci
Do you see any problem with these? i'm very confused, i can't find any solution. As i try in another PC, i never have any problem. The problem only appear with PowerEdge server.
 
I'm so thanksfully for your kindness to help me solve the problem. Thanks.  

36 Posts

January 11th, 2006 02:00

Dear Eric,
i'm so sorry for this mistake. the last report i sent to you is from the original disk, not the megaraid.
I've make sure that RAID Bios is enabled. What do you mean by configure the container in RAID controller?
 
Here's the correct :
 
fdisk -l
Disk /dev/sda: 36.3 GB, 36364615680 bytes
255 heads, 63 sectors/track, 4421 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes

Device Boot    Start       End    Blocks   Id  System
/dev/sda1   *         1       637   5116671   83  Linux
/dev/sda2           638      4167  28354725   83  Linux
/dev/sda3          4168      4421   2040255   82  Linux swap
 
/etc/modules.conf

alias eth0 e100
alias scsi_hostadapter megaraid2
alias scsi_hostadapter1 sym53c8xx
alias scsi_hostadapter2 qla2200
alias usb-controller usb-uhci
 
cat /proc/scsi/scsi
Attached devices:
Host: scsi0 Channel: 00 Id: 00 Lun: 00
  Vendor: MegaRAID Model:LD0 RAID1 34680R Rev: 1148
  Type:   Direct-Access                    ANSI SCSI revision: 02
 
I can loading GRUB using this drive in another PC, but it show "GRUB READ ERROR" when attached to PowerEdge server. Do you have any idea?
 Btw, do you have any messeger like YM, MSN, or ICQ? May i reach you by messeger? And may i know the TimeZone of your location? Because i don't want to disturb you in the wrong time. Thanks a lot.

2 Intern

 • 

815 Posts

January 11th, 2006 11:00

Very your grub.conf looks similar to this:  Kernel versions may differ.
 
 
# grub.conf generated by anaconda
#
# Note that you do not have to rerun grub after making changes to this file
# NOTICE:  You have a /boot partition.  This means that
#          all kernel and initrd paths are relative to /boot/, eg.
#          root (hd0,0)
#          kernel /vmlinuz-version ro root=/dev/sda2
#          initrd /initrd-version.img
#boot=/dev/sda
default=0
timeout=10
splashimage=(hd0,0)/grub/splash.xpm.gz
title Red Hat Enterprise Linux AS (2.4.21-37.ELsmp)
        root (hd0,0)
        kernel /vmlinuz-2.4.21-37.ELsmp ro root=/dev/sda2
        initrd /initrd-2.4.21-37.ELsmp.img
title Red Hat Enterprise Linux AS-up (2.4.21-37.EL)
        root (hd0,0)
        kernel /vmlinuz-2.4.21-37.EL ro root=/dev/sda2
        initrd /initrd-2.4.21-37.EL.img
 
Then type
 
grub-install /dev/sda
 
It should respond back with something like:
Installation finished. No error reported.
This is the contents of the device map /boot/grub/device.map.
Check if this is correct or not. If any of the lines is incorrect,
fix it and re-run the script `grub-install'.
# this device map was generated by anaconda
(fd0)     /dev/fd0
(hd0)     /dev/sda
 
Reboot your system and see if it works.

36 Posts

January 17th, 2006 01:00

dear Eric,

i've always tried the way as u told me. But it really didn't works in PowerEdge8450. After reboot, system only displaying "GRUB READ ERROR".

Then i try another way, by removing GRUB from MBR, and then reinstall GRUB in SDA1 sector 0. And... it works. But then i wanna know, can you tell me the difference for the two way? Does it have any pros and cons to install GRUB in MBR or sector 0? Do you prefer in MBR or sector 0?

Other question, can you give me suggestion where to find the PERC3/DC driver for RedHat 7.2 kernel 2.4.7-10smp and FreeBSD 4.11?   i need it for other PowerEdge 8450 server.

Thanks a lot.

Regards,
 
Sumardi Huang

2 Intern

 • 

815 Posts

January 17th, 2006 12:00

Well I'm glad to here you resolve the boot loader problem.  The boot loader can be installed in either the Master Boot Record, or in the boot record of the partition.  By installing Grub into /dev/sda1, you loaded grub into the partition.  Typically that shouldn't work w/o something in the boot order pointing to that partition instead of the MBR.  The BIOS typically passes control to the first 512 bytes of the hard drive, which in this case should be the 1st stage of grub. 

I guess it really doesn't matter, as long as your system is booting.

As far as the PERC driver for RH 7.2, you could try here.

http://linux.dell.com/storage.shtml

I'm not sure where to find the perc drivers for FreeBSD, however the Linux source is available from LSI, and a crafty developer should be able to compile a binary driver for FreeBSD using that code.

3 Posts

January 27th, 2006 03:00

Ok, PG&E bounced the juice a few times and it looks like it left two of my three drives in a RAID 5 plex offline.

I struggled with them but was able to open WebBios and 'activate' the two failed drives. I then ran a consistency check on the RAID volumn and everything is OK. However, the system won't boot. It gets to the 1 Logical drive found, 1 Logical drive handled by BIOS, GRUB Stage2... and that's all she wrote. The system is hung.

I've tried to boot from floppy (6.2 RH Linux) but it fails. I have booted from the install disks for RH Linux 6.2 but it only gives me the install options. When I click on fdisk it says no drives detected.

I've read on the grub.conf edits but I can't get to the filesystem yet.

Question, how do I get to the file system to verify the grub.conf?

How do I get this system back to be bootable as it was on Tuesday?

Please help and give details,
Marc

36 Posts

January 27th, 2006 05:00

Heloo,
i didn't it could help you solve or not,
What kind of RAID Controller you used? Do you using Software RAID or Hardware RAID?
As you said web bios and consistency check, i thought it is a hardware RAID.
You have data on the array-disk? So don't ever doing initialize because it will remove all your data.
 
Usually step I do :
- restart system
- press Ctrl + M (or any other key as stated in the POST to entering configuration menu)
- View Configuration
- You should see any detected drive. Online (drive is OK), Ready (unconfigured), Offline( disabled)
- You can bring the Offline drive by menu : Object || Physical Drive
Select the drive which you want to make it online and bring it online. It should didn't affect the data.
 
Note : be carefull, when you wanna view configuration, "make sure" that you select COD (configuration on Disk), never-ever select NVRAM.
 
May be you need to check the driver. Can you check the log file?
 

2 Intern

 • 

815 Posts

January 27th, 2006 11:00

Marc, under your situation, I would first boot into a rescue environment and ensure the files systems are consistent first before trying to debug an MBR problem with Grub.  Let us know if your files are mountable in the Red Hat rescue environment.  If they are, there are many things we can do to fix a boot problem from within that environment.

You may need to use a newer version of Red Hat / Fedora Core to verify disks. 

3 Posts

January 27th, 2006 13:00

Yep, I did the WebBios configuration to bring the disk back online. I then ran the consistency check which came out ok. It is a PERC controller which came with the Del Poweredge 1550 (three Seagate 33 Gb drives). I viewed the configuration from the disks and saved it (did not use NVRAM config since it was suspect after power loss).

Eric, can you or somebody point me to where I can figure out how to make a rescue CD or disk. Prefer a step by step version if possible.

I tried to use the diskette and/or CD's that came with the Poweredge (RH Linux 6.2) but can't get to a command line to so fsck or view files.

Thanks for the input,
Marc

Top