About a month ago, one of the 4 drives in my array went clicky. I was able to identify the 2nd drive (sdb of sd[a-d]) was the bad drive and thought I failed/removed it.
After getting a replacement drive and getting as much data as possible, I don't think I'll get anywhere without editing imsm metadata.
cat /proc/mdstat
Personalities : [raid0] [raid1] [raid6] [raid5] [raid4] [raid10]
md0 : inactive sda(S) sdd(S) sdc(S)
15603 blocks super external:imsm
unused devices:
I took a closer look at "mdadm -E /dev/md0"
/dev/md0:
Magic : Intel Raid ISM Cfg Sig.
Version : 1.2.01
Orig Family : 6453e91b
Family : 6453e91b
Generation : 0003c070
Creation Time : Unknown
Attributes : All supported
UUID : bb0620fd:8274f3f7:498f64a4:dacbd25f
Checksum : 8e303583 correct
MPB Sectors : 2
Disks : 4
RAID Devices : 1
Disk00 Serial : WD-WCC3F3ARCLKS
State : active
Id : 00000000
Usable Size : 1953514766 (931.51 GiB 1000.20 GB)
UUID : 5a9f14bb:a252fd06:f08cf7cf:b920b29e
RAID Level : 10
Members : 4
Slots : [__UU]
Failed disk : 0
This Slot : 0 (out-of-sync)
Sector Size : 512
Array Size : 3906994176 (1863.00 GiB 2000.38 GB)
Per Dev Size : 1953499136 (931.50 GiB 1000.19 GB)
Sector Offset : 0
Num Stripes : 7630848
Chunk Size : 64 KiB
Reserved : 0
Migrate State : idle
Map State : failed
Dirty State : clean
RWH Policy : off
Volume ID : 1
Disk01 Serial : 57E07H1MS:0
State : active
Id : ffffffff
Usable Size : 1953514766 (931.51 GiB 1000.20 GB)
Disk02 Serial : WD-WCC6Y0RL73N1
State : active
Id : 00000002
Usable Size : 1953514766 (931.51 GiB 1000.20 GB)
Disk03 Serial : WD-WCC6Y0LPDK7T
State : active
Id : 00000003
Usable Size : 1953514766 (931.51 GiB 1000.20 GB)
Interesting! So It looks like the metadata is coming from /dev/sda, Disk00. This is the first time I've noticed 4 drives, and that "Disk01 Serial : 57E07H1MS:0" doesn't look "WD" like the other "WD" drives. I wouldn't be surprised if I missed something in the --remove process.
I've had access to the "Intel virtual raid on cpu intel vroc for linux user guide" pdf for lots of insight. The "update-subarray" stuff looks like what I'll need to research next to figure out [Volume0]'s Failed disk and Disk01's metadata. I'd expect something like
# mdadm --update-subarray=0 --update-id=ffffffff --state=failed /dev/md0
but that was from piecing together some examples I found in the "Linux-Intel-VROC-TPS-335893.pdf" I found.
Any help getting the array --run-ing again would be appreciated.
Asked by Jeff Archambeault
(1 rep)
May 7, 2025, 02:26 AM