Hallo Liste / Leute,
Ich glaub ich hab Mist gebaut ... Nach einem Stromausfall war mein RAID5 "weg". (6 Disks, kein Spare) Der Versuch, das wieder zusammenzusetzen schlug fehl.
root@openmediavault:~# mdadm --stop /dev/md127 mdadm: stopped /dev/md127 root@openmediavault:~# mdadm --assemble --force --verbose /dev/md127 /dev/sd[bcdefg] mdadm: looking for devices for /dev/md127 mdadm: /dev/sdb is identified as a member of /dev/md127, slot 0. mdadm: /dev/sdc is identified as a member of /dev/md127, slot 5. mdadm: /dev/sdd is identified as a member of /dev/md127, slot 4. mdadm: /dev/sde is identified as a member of /dev/md127, slot 3. mdadm: /dev/sdf is identified as a member of /dev/md127, slot 1. mdadm: /dev/sdg is identified as a member of /dev/md127, slot 2. mdadm: added /dev/sdf to /dev/md127 as 1 mdadm: added /dev/sdg to /dev/md127 as 2 mdadm: added /dev/sde to /dev/md127 as 3 mdadm: added /dev/sdd to /dev/md127 as 4 mdadm: added /dev/sdc to /dev/md127 as 5 mdadm: added /dev/sdb to /dev/md127 as 0 mdadm: /dev/md127 has been started with 5 drives (out of 6) and 1 rebuilding. root@openmediavault:~# cat /proc/mdstat Personalities : [raid6] [raid5] [raid4] [linear] [multipath] [raid0] [raid1] [raid10] md127 : active (auto-read-only) raid5 sdb[0] sdc[5] sdd[4] sde[3] sdg[6] sdf[1] 14650693120 blocks super 1.2 level 5, 512k chunk, algorithm 2 [6/5] [UU_UUU] bitmap: 0/22 pages [0KB], 65536KB chunk
unused devices: <none> root@openmediavault:~# cat /proc/mdstat Personalities : [raid6] [raid5] [raid4] [linear] [multipath] [raid0] [raid1] [raid10] md127 : active (auto-read-only) raid5 sdb[0] sdc[5] sdd[4] sde[3] sdg[6] sdf[1] 14650693120 blocks super 1.2 level 5, 512k chunk, algorithm 2 [6/5] [UU_UUU] bitmap: 0/22 pages [0KB], 65536KB chunk
root@openmediavault:~# mdadm -D /dev/md127 /dev/md127: Version : 1.2 Creation Time : Sat Oct 8 12:19:19 2016 Raid Level : raid5 Array Size : 14650693120 (13971.99 GiB 15002.31 GB) Used Dev Size : 2930138624 (2794.40 GiB 3000.46 GB) Raid Devices : 6 Total Devices : 6 Persistence : Superblock is persistent
Intent Bitmap : Internal
Update Time : Sat Oct 29 12:14:59 2022 State : clean, degraded Active Devices : 5 Working Devices : 6 Failed Devices : 0 Spare Devices : 1
Layout : left-symmetric Chunk Size : 512K
Consistency Policy : bitmap
Name : openmediavault:NAS (local to host openmediavault) UUID : 012a11bc:bce8c44a:f036c177:460a832e Events : 34773
Number Major Minor RaidDevice State 0 8 16 0 active sync /dev/sdb 1 8 80 1 active sync /dev/sdf 6 8 96 2 spare rebuilding /dev/sdg 3 8 64 3 active sync /dev/sde 4 8 48 4 active sync /dev/sdd 5 8 32 5 active sync /dev/sdc root@openmediavault:~#
Wie bekomme ich das RAID wieder zum laufen ? Der rebuild Prozess startet nicht ? Muss ich /dev/sdg noch mal entfernen und erneut hinzufügen ) (--remove / -add) Danke für Eure Hilfe.
Gruss Pritt