My workstation has two disks(/dev/sd[ab]), both with similar partitioning. /dev/sdb failed, and cat /proc/mdstat stopped showing the second sdb partition.
I ran mdadm --fail and mdadm --remove for all partitions from the failed disk on the arrays that use them, although all such commands failed with
mdadm: set device faulty failed for /dev/sdb2: No such device
mdadm: hot remove failed for /dev/sdb2: No such device or address
Then I hot swapped the failed disk, partitioned the new disk and added the partitions to the respective arrays. All arrays got rebuilt properly except one, because in /dev/md2, the failed disk doesn't seem to have been removed from the array properly. Because of this, the new partition keeps getting added as a spare to the partition, and its status remains degraded.
Here's what mdadm --detail /dev/md2 shows:
[root@ldmohanr ~]# mdadm --detail /dev/md2
/dev/md2:
Version : 1.1
Creation Time : Tue Dec 27 22:55:14 2011
Raid Level : raid1
Array Size : 52427708 (50.00 GiB 53.69 GB)
Used Dev Size : 52427708 (50.00 GiB 53.69 GB)
Raid Devices : 2
Total Devices : 2
Persistence : Superblock is persistent
Intent Bitmap : Internal
Update Time : Fri Nov 23 14:59:56 2012
State : active, degraded
Active Devices : 1
Working Devices : 2
Failed Devices : 0
Spare Devices : 1
Name : ldmohanr.net:2 (local to host ldmohanr.net)
UUID : 4483f95d:e485207a:b43c9af2:c37c6df1
Events : 5912611
Number Major Minor RaidDevice State
0 8 2 0 active sync /dev/sda2
1 0 0 1 removed
2 8 18 - spare /dev/sdb2
To remove a disk, mdadm needs a device filename, which was /dev/sdb2 originally, but that no longer refers to device number 1. I need help with removing device number 1 with 'removed' status and making /dev/sdb2 active.