On Sat, Mar 25, 2017 at 06:09:34PM -0300, Renaud OLGIATI wrote:
...
...
>  
> 3 Removing The Failed Disk
> 
> To remove /dev/sdb, we will mark /dev/sdb1 and /dev/sdb2 as failed and remove them from their respective RAID arrays (/dev/md0 and /dev/md1).
Possibly too late.  My defective RAID partitions are on a drive that 
is already not physically connected to the machine.  That's the 
problem I'm trying to solve *safely*.
> 
> First we mark /dev/sdb1 as failed:
>     mdadm --manage /dev/md0 --fail /dev/sdb1
mdadm is already aware that its other disk has failed.  It's 
physically absent.
...
...
> 
> Then power down the system:
>     shutdown -h now
> and replace the old /dev/sdb hard drive with a new one
I deeply believe it's the connection -- the interface and the cabling 
-- that is defective and not the hard drive.  I have alredy provided a 
new drive in this slot only to have problems recur with the new drive.
So I figure that it's probably not the drive, so the new drive will 
*be* the old one.
I plan to conect it by rearranging the cablng so the drive is 
connected with a different cable to a different interface and a 
different power connector.
  
Once I get it connected again,  the worry will be that 
the file systems on the now-disconected drive will fool the boot 
process into booting from it instead of the up-to-date drive.
Maybe I have to wipe the disk clean before reinstalling it.  Which 
I'll probably have to do on a different machine.
Or is it enough tto scramble the UUIDs of the file systems and to 
erase the RAI signatures?  Or use DD to empty the partitions?  Will dd 
to a partition erase the UUID of its file system?  And the RAID 
signature?
> (it must have at least the same size as the old one 
> if it's only a few MB smaller than the old one then rebuilding the 
> arrays will fail).
> 
>  
> 
> 4 Adding The New Hard Disk
> 
> After you have changed the hard disk /dev/sdb, boot the system.
Once I's at this stage, with a clean second drive, I know how to 
proceed from here.
-- hendrik