[Gllug] RAID on LVM - heracy?
Nix
nix at esperi.org.uk
Fri Sep 28 21:00:49 UTC 2007
On 28 Sep 2007, Ken Smith told this:
> But consider this. If I had a collection of disks, none really suitably
> large or equal-ish in size for raid by themselves, would it be valid to
> group some of them together as for example - LV1 and some others
> together as LV2 and so on and then run soft raid over the top?
>
> Could you even run LVM over the resultant volume to retain the
> flexibility of resizing file systems....
I do exactly the opposite of this.
I have four disks, two IDE (an old 10Gb, a larger 45Gb) and two SCSI
(70Gb each).
These are rather inconveniently sized, so I split them into three RAID
arrays and a lump of un-RAIDed space for stuff I don't much care about
or can regenerate easily.
Each disk starts with one component of a small, non-LVMed RAID-1 array
containing /boot (in a primary partition on the IDE drives because of
limitations of my machine's BIOS: it locks up before boot if all IDE
disks don't have at least one primary partition, *sigh*):
Array Size : 56064 (54.76 MiB 57.41 MB)
Used Dev Size : 56064 (54.76 MiB 57.41 MB)
Number Major Minor RaidDevice State
0 8 5 0 active sync /dev/sda5
1 8 21 1 active sync /dev/sdb5
2 3 1 2 active sync /dev/hda1
3 22 1 3 active sync /dev/hdc1
(It's RAID-1 both for robustness against failure of any disk and because
LILO can't boot from anything more sophisticated: it's also the only
array using a v0.90 superblock for exactly the same reason.)
Then there are two RAID-5 arrays. There's one large one on both SCSI
disks and the large IDE disk, sized to fill the IDE disk completely
except for a gigabyte for swap:
Array Size : 76807296 (73.25 GiB 78.65 GB)
Used Dev Size : 76807296 (36.62 GiB 39.33 GB)
Number Major Minor RaidDevice State
0 8 6 0 active sync /dev/sda6
1 8 22 1 active sync /dev/sdb6
3 22 5 2 active sync /dev/hdc5
The other RAID-5 array is smaller and on the old 10Gb drive and the
two SCSI drives (as they still have room left);
Array Size : 19631104 (18.72 GiB 20.10 GB)
Used Dev Size : 19631104 (9.36 GiB 10.05 GB)
Number Major Minor RaidDevice State
0 8 23 0 active sync /dev/sdb7
1 8 7 1 active sync /dev/sda7
3 3 5 2 active sync /dev/hda5
The remaining space (20Gb or so on each SCSI disk) is not RAIDed, but
just a single LVM VG with two PVs.
Atop this we have two LVM volume groups, one covering the non-RAIDed
space, one the RAIDed space. We cut the space into two regions for an
important reason: if you lose a single PV, you're likely to lose the
whole VG unless you're very lucky or careful with LV assignment within
the VG. Both RAID arrays are combined into a single VG because even if
we lose one of the SCSI disks, all that will happen is that both RAID-5
arrays will go degraded: the VG is exactly as robust against failure as
RAID-5 natively is.
(The second VG is nonrobust; losing either SCSI disk will kill it. I
don't care: it contains things like my news spool, video crud, and stuff
that's so boring I don't even back it up.)
Booting from this requires an initramfs, of course: auto-assembly won't
find the root filesystem when it's on a v1 superblock, especially when
it's also inside LVM and thus a `vgscan' away from being mountable. The
linux-raid wiki used to contain instructions for building this
initramfs, but it seems to have gone down in a welter of wikispam (at
least it had last I checked, a few days ago).
--
`Some people don't think performance issues are "real bugs", and I think
such people shouldn't be allowed to program.' --- Linus Torvalds
--
Gllug mailing list - Gllug at gllug.org.uk
http://lists.gllug.org.uk/mailman/listinfo/gllug
More information about the GLLUG
mailing list