Followed the guide here: ubuntuforums.org/showthread.php?t=713936
Once again, LVM is awesome.
Reproduced:
1. install your disk into the system.
2. Get the name of the disk you want to add:Quote:
[root@e6400:/home/moon/ 519]$lshw -C disk
*-disk:0
description: SCSI Disk
product: ST3320620AS
vendor: ATA
physical id: 0
bus info: scsi@0:0.0.0
logical name: /dev/sda
version: 3.AA
serial: 3QF0AL47
size: 298GB
capabilities: partitioned partitioned:dos
configuration: ansiversion=5
*-disk:1
description: SCSI Disk
product: ST3320620AS
vendor: ATA
physical id: 1
bus info: scsi@1:0.0.0
logical name: /dev/sdb
version: 3.AA
serial: 5QF0PEWW
size: 298GB
capabilities: partitioned partitioned:dos
configuration: ansiversion=5
*-disk:2
description: SCSI Disk
product: ST3320620AS
vendor: ATA
physical id: 2
bus info: scsi@4:0.0.0
logical name: /dev/sdc
version: 3.AA
serial: 5QF0DMH5
size: 298GB
capabilities: partitioned partitioned:dos
configuration: ansiversion=5
*-disk:3
description: SCSI Disk
product: ST3320620AS
vendor: ATA
physical id: 3
bus info: scsi@5:0.0.0
logical name: /dev/sdd
version: 3.AA
serial: 5QF0S6XE
size: 298GB
capabilities: partitioned partitioned:dos
configuration: ansiversion=5So the disk i want to add in is /dev/sdd
2. Partion this disk based on an existing disk in the array assuming /dev/sdb is already in the array:
Copy the partition from the existing drive in the raid.
Quote:
sfdisk -d /dev/sdb > partitions.sdb edit the partitions.sdb file and change all the references to sdb to sdd
Now we partition the new drive
Quote:
sfdisk /dev/sdd < partitions.sdb note:
I have 2 partition on my drive
/dev/sdd1 – Raid 0 on /dev/md0
/dev/sdd2 – Raid 5 on /dev/md13. Clean the superblock on the new drive partitions (In case it was in another Raid) either way it won’t hurt.
Quote:
mdadm –zero-superblock /dev/sdd1
mdadm –zero-superblock /dev/sdd24. Add the new drives to the arry, i have a Raid 0 for partition /dev/sdd1 and a Raid 5 for partition /dev/sdd2
Set raid-devices to the new number of devices in your array, i previously had 3.Quote:
mdadm –add /dev/md0 /dev/sdd1
mdadm –grow /dev/md0 –raid-devices=4mdadm –add /dev/md1 /dev/sdd2
mdadm –grow /dev/md1 –raid-devices=45. My Raid 0 (/dev/md0) is done, this is my /boot.
6. WAIT a long time for the Raid to rebuild itself. To check the progress run this command:
Quote:
cat /proc/mdstat 7. When it is finally done you should be able to see the new size of the Raid arry:
Quote:
[root@e6400:/home/moon/ 503]$mdadm -D /dev/md1
/dev/md1:
Version : 00.90.03
Creation Time : Sat Mar 1 06:36:40 2008
Raid Level : raid5
Array Size : 937488768 (894.06 GiB 959.99 GB)
Used Dev Size : 312496256 (298.02 GiB 320.00 GB)
Raid Devices : 4
Total Devices : 4
Preferred Minor : 1
Persistence : Superblock is persistentUpdate Time : Mon Mar 3 09:39:06 2008
State : active
Active Devices : 4
Working Devices : 4
Failed Devices : 0
Spare Devices : 0Layout : left-symmetric
Chunk Size : 64KUUID : 6edd19d2:6bc3e59d:ce21a63d:f12f7b33
Events : 0.208309Number Major Minor RaidDevice State
0 8 2 0 active sync /dev/sda2
1 8 18 1 active sync /dev/sdb2
2 8 34 2 active sync /dev/sdc2
3 8 50 3 active sync /dev/sdd28. Now we have to make changes to our LVM.
Resize our physical volume
Quote:
[root@e6400:~/ 499]$pvresize /dev/md1
Physical volume “/dev/md1” changed
1 physical volume(s) resized / 0 physical volume(s) not resized
[root@e6400:~/ 500]$pvdisplay
— Physical volume —
PV Name /dev/md1
VG Name VG_E6400
PV Size 894.06 GB / not usable 192.00 KB
Allocatable yes
PE Size (KByte) 4096
Total PE 228879
Free PE 76293
Allocated PE 152586
PV UUID dw0IGv-j1WA-hd63-mY3K-i6mJ-4yX6-MQftH89. Resize our logical volume:
Quote:
[root@e6400:/home/moon/ 503]$lvresize -l 95%VG /dev/VG_E6400/LV_E6400 For some reason it would not let me use 100%, so i had to add the rest of the free disk (38.7GB) space using the resize command again:
Quote:
[root@e6400:/home/moon/ 503]$lvresize -L +38.7G /dev/VG_E6400/LV_E6400 10. Check that your LV was resized correctly:
Quote:
[root@e6400:/home/moon/ 503]$lvdisplay
— Logical volume —
LV Name /dev/VG_E6400/LV_E6400
VG Name VG_E6400
LV UUID JIdQXg-C6Rv-WLq9-I61F-Foss-N8jE-wG13IQ
LV Write Access read/write
LV Status available
# open 1
LV Size 888.06 GB
Current LE 227343
Segments 1
Allocation inherit
Read ahead sectors 0
Block device 254:111. If everything looks good, resize your file system (i use ext 3 you may have to run a different cmd if you run a differnt FS). My volume was mouted as the root filesystem, so it did an online resize. This is probably dangerous.
Quote:
[root@e6400:/home/moon/ 503]$resize2fs /dev/VG_E6400/LV_E6400 12. CHeck your disk space:
Quote:
[root@e6400:/home/moon/ 503]$df -h /
Filesystem Size Used Avail Use% Mounted on
/dev/mapper/VG_E6400-LV_E6400
875G 267G 564G 33% /