RAID upgrade
From ThorxWiki
				
								
				(Difference between revisions)
				
																
				
				
								
				|  (first writeup) |  (seso) | ||
| Line 1: | Line 1: | ||
| I have filesystems on LVMs on RAIDs on partitions. Updating a disk is then, arguably, rather non-trivial. These are my notes-to-self. Maybe they'll help you too? | I have filesystems on LVMs on RAIDs on partitions. Updating a disk is then, arguably, rather non-trivial. These are my notes-to-self. Maybe they'll help you too? | ||
| − | My setup:  | + | == My system == | 
| − |  # cat /etc/debian_version  | + | <pre> | 
| − |  lenny/sid | + | # cat /etc/debian_version  | 
| − |  # uname -a | + | lenny/sid | 
| − |  Linux falcon 2.6.18-6-amd64 #1 SMP Mon Jun 16 22:30:01 UTC 2008 x86_64 GNU/Linux | + | # uname -a | 
| − |  # mdadm --version | + | Linux falcon 2.6.18-6-amd64 #1 SMP Mon Jun 16 22:30:01 UTC 2008 x86_64 GNU/Linux | 
| − |  mdadm - v2.5.6 - 9 November 2006 | + | # mdadm --version | 
| − |  # lvm version | + | mdadm - v2.5.6 - 9 November 2006 | 
| − |    LVM version:     2.02.07 (2006-07-17) | + | # lvm version | 
| − |    Library version: 1.02.08 (2006-07-17) | + |   LVM version:     2.02.07 (2006-07-17) | 
| − |    Driver version:  4.7.0 | + |   Library version: 1.02.08 (2006-07-17) | 
| + |   Driver version:  4.7.0 | ||
| + | </pre> | ||
| + | == My setup == | ||
| + | <pre> | ||
| + | |||
| + | # cat /proc/mdstat | ||
| + | Personalities : [raid1] [raid6] [raid5] [raid4] | ||
| + | md3 : active raid5 sda4[0] sdd4[3] sdc4[2] sdb4[1] | ||
| + |       2637302400 blocks level 5, 64k chunk, algorithm 2 [4/4] [UUUU] | ||
| + | |||
| + | md2 : active raid1 sdc3[0] sdd3[1] | ||
| + |       87891520 blocks [2/2] [UU] | ||
| + | |||
| + | md0 : active raid1 sda2[0] sdb2[2](S) sdd2[3](S) sdc2[1] | ||
| + |       8787456 blocks [2/2] [UU] | ||
| + | |||
| + | md1 : active raid1 sda3[0] sdb3[1] | ||
| + |       87891520 blocks [2/2] [UU] | ||
| + | |||
| + | unused devices: <none> | ||
| + | |||
| + | # pvs | ||
| + |   PV         VG            Fmt  Attr PSize  PFree  | ||
| + |   /dev/md1   volgrp_home   lvm2 a-   83.82G     0  | ||
| + |   /dev/md2   volgrp_home   lvm2 a-   83.82G 27.63G | ||
| + |   /dev/md3   volgrp_shared lvm2 a-    2.46T     0  | ||
| + | |||
| + | # vgs | ||
| + |   VG            #PV #LV #SN Attr   VSize   VFree  | ||
| + |   volgrp_home     2   1   0 wz--n- 167.63G 27.63G | ||
| + |   volgrp_shared   1   1   0 wz--n-   2.46T     0 | ||
| + | |||
| + | # lvs | ||
| + |   LV        VG            Attr   LSize   Origin Snap%  Move Log Copy%  | ||
| + |   lv_home   volgrp_home   -wi-ao 140.00G | ||
| + |   lv_shared volgrp_shared -wi-ao   2.46T | ||
| + | |||
| + | </pre> | ||
| == Hot removing a disc from a raid1 ==  | == Hot removing a disc from a raid1 ==  | ||
| (spares already available) | (spares already available) | ||
| − | + | <pre> | |
| − |  mdadm --fail /dev/md0 /dev/sda2 | + | # mdadm --fail /dev/md0 /dev/sda2 | 
| − |  mdadm --remove /dev/md0 /dev/sda2 | + | # mdadm --remove /dev/md0 /dev/sda2 | 
| + | </pre> | ||
| You can watch the spares take over with | You can watch the spares take over with | ||
| − |  cat /proc/mdstat | + | <pre> | 
| + | # cat /proc/mdstat | ||
| + | </pre> | ||
| This process starts immediately after the --fail. | This process starts immediately after the --fail. | ||
Revision as of 12:29, 17 February 2010
I have filesystems on LVMs on RAIDs on partitions. Updating a disk is then, arguably, rather non-trivial. These are my notes-to-self. Maybe they'll help you too?
My system
# cat /etc/debian_version lenny/sid # uname -a Linux falcon 2.6.18-6-amd64 #1 SMP Mon Jun 16 22:30:01 UTC 2008 x86_64 GNU/Linux # mdadm --version mdadm - v2.5.6 - 9 November 2006 # lvm version LVM version: 2.02.07 (2006-07-17) Library version: 1.02.08 (2006-07-17) Driver version: 4.7.0
My setup
# cat /proc/mdstat
Personalities : [raid1] [raid6] [raid5] [raid4]
md3 : active raid5 sda4[0] sdd4[3] sdc4[2] sdb4[1]
      2637302400 blocks level 5, 64k chunk, algorithm 2 [4/4] [UUUU]
md2 : active raid1 sdc3[0] sdd3[1]
      87891520 blocks [2/2] [UU]
md0 : active raid1 sda2[0] sdb2[2](S) sdd2[3](S) sdc2[1]
      8787456 blocks [2/2] [UU]
md1 : active raid1 sda3[0] sdb3[1]
      87891520 blocks [2/2] [UU]
unused devices: <none>
# pvs
  PV         VG            Fmt  Attr PSize  PFree 
  /dev/md1   volgrp_home   lvm2 a-   83.82G     0 
  /dev/md2   volgrp_home   lvm2 a-   83.82G 27.63G
  /dev/md3   volgrp_shared lvm2 a-    2.46T     0 
# vgs
  VG            #PV #LV #SN Attr   VSize   VFree 
  volgrp_home     2   1   0 wz--n- 167.63G 27.63G
  volgrp_shared   1   1   0 wz--n-   2.46T     0
# lvs
  LV        VG            Attr   LSize   Origin Snap%  Move Log Copy% 
  lv_home   volgrp_home   -wi-ao 140.00G
  lv_shared volgrp_shared -wi-ao   2.46T
Hot removing a disc from a raid1
(spares already available)
# mdadm --fail /dev/md0 /dev/sda2 # mdadm --remove /dev/md0 /dev/sda2
You can watch the spares take over with
# cat /proc/mdstat
This process starts immediately after the --fail.

