orb:/etc # mdadm --detail --scan -vvv /dev/md0: Version : 00.90.03 Creation Time : Sun Dec 2 17:27:38 2007 Raid Level : raid5 Array Size : 2441912320 (2328.79 GiB 2500.52 GB) Used Dev Size : 488382464 (465.76 GiB 500.10 GB) Raid Devices : 6 Total Devices : 6 Preferred Minor : 0 Persistence : Superblock is persistent
Update Time : Sun Dec 2 17:27:38 2007 State : clean, degraded Active Devices : 5 Working Devices : 6 Failed Devices : 0 Spare Devices : 1
Layout : left-symmetric Chunk Size : 4096K
UUID : 837eecb4:28f004fb:0e4f8c91:f584a4b1 Events : 0.1
Number Major Minor RaidDevice State 0 8 1 0 active sync /dev/sda1 1 8 17 1 active sync /dev/sdb1 2 8 49 2 active sync /dev/sdd1 3 8 65 3 active sync /dev/sde1 4 8 81 4 active sync /dev/sdf1 5 0 0 5 removed
6 8 97 - spare /dev/sdg1
All your disks are there, you can see the disk numbering starts at 0 and goes to 4, that's 5 disks + your spare. I'm not exactly sure why its showing the 5th entry, it looks like you've removed a disk from the array. I'm sure there is a way to remove the number 5 drive from the array using mdadm (I'm not familiar with the tool myself) and that should get you out of the degraded state. BTW using RAID 5 with +1 parity on 6 500G drives will result in about 66-70% storage. In your case it will be just over 2TB. Cheers Todd -- To unsubscribe, e-mail: opensuse+unsubscribe@opensuse.org For additional commands, e-mail: opensuse+help@opensuse.org