[rlug] Cum reactivez un raid 5 cu mdadm ?
Paul Lacatus (Personal)
paul at paul-lacatus.ro
Sat Dec 30 12:47:00 EET 2017
La serverul meu de fisiere de acasa mi-a picat sursa . Pana sa ma prind
ca e sursa de vina , facea probleme cand erau toate hdd alimentate ,
s-a dus si centos 6.9 care dadea kernel panic not tainted. Cu ocazia
asta am reinstalat cu un centos 7.4
treaba e ca acum raid-ul fost /dev/md127 a devenit inactiv :
> [root at datavault ~]# mdadm --detail /dev/md127
> /dev/md127:
> Version : 0.90
> Raid Level : raid0
> Total Devices : 4
> Preferred Minor : 0
> Persistence : Superblock is persistent
>
> State : inactive
>
> UUID : 188f7506:c03bd2ac:cdc3d78b:f6534f77
> Events : 0.14838
>
> Number Major Minor RaidDevice
>
> - 8 1 - /dev/sda1
> - 8 17 - /dev/sdb1
> - 8 33 - /dev/sdc1
> - 8 49 - /dev/sdd1
la examinare pe fiecare disc in parte datele par a fi ok dar aria nu
pleaca din cauza testelor facute . Trebuie sa recreez aria sau se poate
reporni ?
> cat raid.status
> /dev/sda1:
> Magic : a92b4efc
> Version : 0.90.00
> UUID : 188f7506:c03bd2ac:cdc3d78b:f6534f77
> Creation Time : Sat Dec 18 20:07:41 2010
> Raid Level : raid5
> Used Dev Size : 976759936 (931.51 GiB 1000.20 GB)
> Array Size : 2930279808 (2794.53 GiB 3000.61 GB)
> Raid Devices : 4
> Total Devices : 4
> Preferred Minor : 127
>
> Update Time : Thu Dec 28 11:38:37 2017
> State : clean
> Active Devices : 4
> Working Devices : 4
> Failed Devices : 0
> Spare Devices : 0
> Checksum : 279913d8 - correct
> Events : 14838
>
> Layout : left-symmetric
> Chunk Size : 64K
>
> Number Major Minor RaidDevice State
> this 3 8 1 3 active sync /dev/sda1
>
> 0 0 8 17 0 active sync /dev/sdb1
> 1 1 8 33 1 active sync /dev/sdc1
> 2 2 8 49 2 active sync /dev/sdd1
> 3 3 8 1 3 active sync /dev/sda1
> /dev/sdb1:
> Magic : a92b4efc
> Version : 0.90.00
> UUID : 188f7506:c03bd2ac:cdc3d78b:f6534f77
> Creation Time : Sat Dec 18 20:07:41 2010
> Raid Level : raid5
> Used Dev Size : 976759936 (931.51 GiB 1000.20 GB)
> Array Size : 2930279808 (2794.53 GiB 3000.61 GB)
> Raid Devices : 4
> Total Devices : 3
> Preferred Minor : 127
>
> Update Time : Sat Dec 30 10:19:54 2017
> State : clean
> Active Devices : 3
> Working Devices : 3
> Failed Devices : 1
> Spare Devices : 0
> Checksum : 279ba477 - correct
> Events : 14846
>
> Layout : left-symmetric
> Chunk Size : 64K
>
> Number Major Minor RaidDevice State
> this 0 8 17 0 active sync /dev/sdb1
>
> 0 0 8 17 0 active sync /dev/sdb1
> 1 1 8 33 1 active sync /dev/sdc1
> 2 2 8 49 2 active sync /dev/sdd1
> 3 3 0 0 3 faulty removed
> /dev/sdc1:
> Magic : a92b4efc
> Version : 0.90.00
> UUID : 188f7506:c03bd2ac:cdc3d78b:f6534f77
> Creation Time : Sat Dec 18 20:07:41 2010
> Raid Level : raid5
> Used Dev Size : 976759936 (931.51 GiB 1000.20 GB)
> Array Size : 2930279808 (2794.53 GiB 3000.61 GB)
> Raid Devices : 4
> Total Devices : 3
> Preferred Minor : 127
>
> Update Time : Sat Dec 30 10:20:32 2017
> State : active
> Active Devices : 3
> Working Devices : 3
> Failed Devices : 1
> Spare Devices : 0
> Checksum : 279b6ab1 - correct
> Events : 14847
>
> Layout : left-symmetric
> Chunk Size : 64K
>
> Number Major Minor RaidDevice State
> this 1 8 33 1 active sync /dev/sdc1
>
> 0 0 8 17 0 active sync /dev/sdb1
> 1 1 8 33 1 active sync /dev/sdc1
> 2 2 8 49 2 active sync /dev/sdd1
> 3 3 0 0 3 faulty removed
> /dev/sdd1:
> Magic : a92b4efc
> Version : 0.90.00
> UUID : 188f7506:c03bd2ac:cdc3d78b:f6534f77
> Creation Time : Sat Dec 18 20:07:41 2010
> Raid Level : raid5
> Used Dev Size : 976759936 (931.51 GiB 1000.20 GB)
> Array Size : 2930279808 (2794.53 GiB 3000.61 GB)
> Raid Devices : 4
> Total Devices : 3
> Preferred Minor : 127
>
> Update Time : Sat Dec 30 10:20:32 2017
> State : active
> Active Devices : 3
> Working Devices : 3
> Failed Devices : 1
> Spare Devices : 0
> Checksum : 279b6ac3 - correct
> Events : 14847
>
> Layout : left-symmetric
> Chunk Size : 64K
>
> Number Major Minor RaidDevice State
> this 2 8 49 2 active sync /dev/sdd1
>
> 0 0 8 17 0 active sync /dev/sdb1
> 1 1 8 33 1 active sync /dev/sdc1
> 2 2 8 49 2 active sync /dev/sdd1
> 3 3 0 0 3 faulty removed
> [root at datavault ~]#
More information about the RLUG
mailing list