[rlug] Cum reactivez un raid 5 cu mdadm ?
Bogdan-Stefan Rotariu
bogdan at rotariu.ro
Sat Dec 30 13:10:34 EET 2017
Pare ok acum, Consistency Policy : resync
Nu uita sa faci mdadm.conf si mkinitrd astfel incat la reboot sa pastreze acelasi device
mdadm --detail --scan > /etc/mdadm/mdadm.conf
sau /etc/mdadm.conf
Sent from mobile
> On Dec 30, 2017, at 13:05, Paul Lacatus (Personal) <paul at paul-lacatus.ro> wrote:
>
> Am reasamblat aria ca /dev/md0 . La inceput nu a vrut sa adauge si /dev/sda1 , cu --readd nu a vrut cu --add l-a adaugat si acum ii face rebuilding
>
> [root at datavault ~]# mdadm /dev/md0 --add /dev/sda1
> mdadm: added /dev/sda1
> [root at datavault ~]# mdadm -D /dev/md0
> /dev/md0:
> Version : 0.90
> Creation Time : Sat Dec 18 20:07:41 2010
> Raid Level : raid5
> Array Size : 2930279808 (2794.53 GiB 3000.61 GB)
> Used Dev Size : 976759936 (931.51 GiB 1000.20 GB)
> Raid Devices : 4
> Total Devices : 4
> Preferred Minor : 0
> Persistence : Superblock is persistent
>
> Update Time : Sat Dec 30 13:00:12 2017
> State : clean, degraded, recovering
> Active Devices : 3
> Working Devices : 4
> Failed Devices : 0
> Spare Devices : 1
>
> Layout : left-symmetric
> Chunk Size : 64K
>
> Consistency Policy : resync
>
> Rebuild Status : 0% complete
>
> UUID : 188f7506:c03bd2ac:cdc3d78b:f6534f77
> Events : 0.14849
>
> Number Major Minor RaidDevice State
> 0 8 17 0 active sync /dev/sdb1
> 1 8 33 1 active sync /dev/sdc1
> 2 8 49 2 active sync /dev/sdd1
> 4 8 1 3 spare rebuilding /dev/sda1
>
>
>
>> On 30-Dec-17 12:47, Paul Lacatus (Personal) wrote:
>> La serverul meu de fisiere de acasa mi-a picat sursa . Pana sa ma prind ca e sursa de vina , facea probleme cand erau toate hdd alimentate , s-a dus si centos 6.9 care dadea kernel panic not tainted. Cu ocazia asta am reinstalat cu un centos 7.4
>>
>> treaba e ca acum raid-ul fost /dev/md127 a devenit inactiv :
>>
>>> [root at datavault ~]# mdadm --detail /dev/md127
>>> /dev/md127:
>>> Version : 0.90
>>> Raid Level : raid0
>>> Total Devices : 4
>>> Preferred Minor : 0
>>> Persistence : Superblock is persistent
>>>
>>> State : inactive
>>>
>>> UUID : 188f7506:c03bd2ac:cdc3d78b:f6534f77
>>> Events : 0.14838
>>>
>>> Number Major Minor RaidDevice
>>>
>>> - 8 1 - /dev/sda1
>>> - 8 17 - /dev/sdb1
>>> - 8 33 - /dev/sdc1
>>> - 8 49 - /dev/sdd1
>>
>> la examinare pe fiecare disc in parte datele par a fi ok dar aria nu pleaca din cauza testelor facute . Trebuie sa recreez aria sau se poate reporni ?
>>
>>> cat raid.status
>>> /dev/sda1:
>>> Magic : a92b4efc
>>> Version : 0.90.00
>>> UUID : 188f7506:c03bd2ac:cdc3d78b:f6534f77
>>> Creation Time : Sat Dec 18 20:07:41 2010
>>> Raid Level : raid5
>>> Used Dev Size : 976759936 (931.51 GiB 1000.20 GB)
>>> Array Size : 2930279808 (2794.53 GiB 3000.61 GB)
>>> Raid Devices : 4
>>> Total Devices : 4
>>> Preferred Minor : 127
>>>
>>> Update Time : Thu Dec 28 11:38:37 2017
>>> State : clean
>>> Active Devices : 4
>>> Working Devices : 4
>>> Failed Devices : 0
>>> Spare Devices : 0
>>> Checksum : 279913d8 - correct
>>> Events : 14838
>>>
>>> Layout : left-symmetric
>>> Chunk Size : 64K
>>>
>>> Number Major Minor RaidDevice State
>>> this 3 8 1 3 active sync /dev/sda1
>>>
>>> 0 0 8 17 0 active sync /dev/sdb1
>>> 1 1 8 33 1 active sync /dev/sdc1
>>> 2 2 8 49 2 active sync /dev/sdd1
>>> 3 3 8 1 3 active sync /dev/sda1
>>> /dev/sdb1:
>>> Magic : a92b4efc
>>> Version : 0.90.00
>>> UUID : 188f7506:c03bd2ac:cdc3d78b:f6534f77
>>> Creation Time : Sat Dec 18 20:07:41 2010
>>> Raid Level : raid5
>>> Used Dev Size : 976759936 (931.51 GiB 1000.20 GB)
>>> Array Size : 2930279808 (2794.53 GiB 3000.61 GB)
>>> Raid Devices : 4
>>> Total Devices : 3
>>> Preferred Minor : 127
>>>
>>> Update Time : Sat Dec 30 10:19:54 2017
>>> State : clean
>>> Active Devices : 3
>>> Working Devices : 3
>>> Failed Devices : 1
>>> Spare Devices : 0
>>> Checksum : 279ba477 - correct
>>> Events : 14846
>>>
>>> Layout : left-symmetric
>>> Chunk Size : 64K
>>>
>>> Number Major Minor RaidDevice State
>>> this 0 8 17 0 active sync /dev/sdb1
>>>
>>> 0 0 8 17 0 active sync /dev/sdb1
>>> 1 1 8 33 1 active sync /dev/sdc1
>>> 2 2 8 49 2 active sync /dev/sdd1
>>> 3 3 0 0 3 faulty removed
>>> /dev/sdc1:
>>> Magic : a92b4efc
>>> Version : 0.90.00
>>> UUID : 188f7506:c03bd2ac:cdc3d78b:f6534f77
>>> Creation Time : Sat Dec 18 20:07:41 2010
>>> Raid Level : raid5
>>> Used Dev Size : 976759936 (931.51 GiB 1000.20 GB)
>>> Array Size : 2930279808 (2794.53 GiB 3000.61 GB)
>>> Raid Devices : 4
>>> Total Devices : 3
>>> Preferred Minor : 127
>>>
>>> Update Time : Sat Dec 30 10:20:32 2017
>>> State : active
>>> Active Devices : 3
>>> Working Devices : 3
>>> Failed Devices : 1
>>> Spare Devices : 0
>>> Checksum : 279b6ab1 - correct
>>> Events : 14847
>>>
>>> Layout : left-symmetric
>>> Chunk Size : 64K
>>>
>>> Number Major Minor RaidDevice State
>>> this 1 8 33 1 active sync /dev/sdc1
>>>
>>> 0 0 8 17 0 active sync /dev/sdb1
>>> 1 1 8 33 1 active sync /dev/sdc1
>>> 2 2 8 49 2 active sync /dev/sdd1
>>> 3 3 0 0 3 faulty removed
>>> /dev/sdd1:
>>> Magic : a92b4efc
>>> Version : 0.90.00
>>> UUID : 188f7506:c03bd2ac:cdc3d78b:f6534f77
>>> Creation Time : Sat Dec 18 20:07:41 2010
>>> Raid Level : raid5
>>> Used Dev Size : 976759936 (931.51 GiB 1000.20 GB)
>>> Array Size : 2930279808 (2794.53 GiB 3000.61 GB)
>>> Raid Devices : 4
>>> Total Devices : 3
>>> Preferred Minor : 127
>>>
>>> Update Time : Sat Dec 30 10:20:32 2017
>>> State : active
>>> Active Devices : 3
>>> Working Devices : 3
>>> Failed Devices : 1
>>> Spare Devices : 0
>>> Checksum : 279b6ac3 - correct
>>> Events : 14847
>>>
>>> Layout : left-symmetric
>>> Chunk Size : 64K
>>>
>>> Number Major Minor RaidDevice State
>>> this 2 8 49 2 active sync /dev/sdd1
>>>
>>> 0 0 8 17 0 active sync /dev/sdb1
>>> 1 1 8 33 1 active sync /dev/sdc1
>>> 2 2 8 49 2 active sync /dev/sdd1
>>> 3 3 0 0 3 faulty removed
>>> [root at datavault ~]#
>>
>>
>>
>> _______________________________________________
>> RLUG mailing list
>> RLUG at lists.lug.ro
>> http://lists.lug.ro/mailman/listinfo/rlug_lists.lug.ro
>
>
> _______________________________________________
> RLUG mailing list
> RLUG at lists.lug.ro
> http://lists.lug.ro/mailman/listinfo/rlug_lists.lug.ro
More information about the RLUG
mailing list