First some history. This is an Intel MB and processor some 6 years old, initially running CentOS 6. It has 4 x 1TB sata drives set up in two mdraid 1 mirrors. It has performed really well in a rural setting with frequent power cuts which the UPS has dealt with and auto shuts down the server after a few minutes and then auto restarts when power is restored. The clients needed a Windoze server for a proprietary accounting package they use, thus I have recently installed two SSD drives (500GB each) also in a raid 1 mirror and installed CentOS 7 as the host and also VirtualBox running Windoze 10. The hard drives continue to hold their data files. This appeared to work just fine until a few days ago. After a power cut the server would not reboot. It takes a while to get in front of the machine, add a monitor, keyboard and mouse only to find: Warning: /dev/disk/by-id/md-uuid-xxxxxxxx:xxxxxxxx:xxxxxxxx:xxxxxxxx does not exist repeated three times - one for each of the /, /boot, and swap raid member sets along with a Warning: /dev/disk/by-uuid/xxxxxxxx:xxxxxxxx:xxxxxxxx:xxxxxxxx does not exist for the /dev/md125 which is the actual raid 1 / device. The system is in a root shell of some sort as it has not made the transition from initramfs to the mdraid root drive. there are some other lines of info and a txt file with hundreds of lines of boot info, ending with the above info (as I recall). I tried a reboot - same result, rebooted and tried an earlier kernel - same result, tried a reboot to the recovery kernel and all went well. System comes up, all raids sets are up and in sync - no errors. So, no apparent H/W issues, no mdraid issues apparently, but none of the regular kernels will now boot. a blkid shows all the expected mdraid devices with the uuids from the error message all in place as expected. I did a yum reinstall of the most recent kernel as I thought that may repair any /boot file system problems - particularly initramfs, but no difference, will not boot, same exact error messages. Thus I now have it running on the recovery kernel, with all the required server functions being performed, albeit on an out of date kernel. Google has one solved problem similar to mine but the solution was change the BIOS from AHCI to IDE - that does not seem correct as I have not changed BIOS, although I have not checked it at this time. Another solution talks about a race condition and the md raid not being ready when required during the boot process and thus to add delay in the kernel boot line in grub2. Although no one indicated this actually worked. Another proposed solution is to mount the failed devices from a recovery boot and rebuild initramfs. Before I do this I would like to ask those that know a little more about the boot process, what is going wrong? I can believe the most recent initramfs being a problem, but all three other kernels too?? Yet the recovery kernel works just fine. As the system is remote, I would like some understanding of what's up before I do any changes - if a reboot occurs and fails, it will mean another trip. Oh, one other thing, it seems the UPS is not working correctly, thus it may not have shut down cleanly. Working to replace batteries in the UPS. TIA for your insight.