I would like to get your feedback about a controversial situation I found myself in some days ago. I was tasked to use a HP Proliant G6 for development purposes with 2 new (< 2 months, never used before) non-HP SSDs used in RAID1 configuration. They are being used for intensive development-oriented tasks (about 500GB written per day); there are also regular HDDs in RAID5, but we will discuss the RAID1 array here.
- Samsung SSD 840 PRO Series
- PLEXTOR PX-256M5Pro
smartctl output for both is available here: https://gist.github.com/anonymous/cf8a5208a7315440f796
Relevant past issues
The Plextor drive has always been affected by a reported overheated condition, that I assume is due to the fact it's not an original part
I've seen once the RAID1 being rebuilt after an occasional server reboot, and couldn't explain the reason for that.
Failure event
A few days ago the Plextor disk was reported in a plain "Failed" status:
physicaldrive 1I:1:1 (port 1I:box 1:bay 1, Solid State SATA, 256.0 GB, Failed)
False positive?
Thus I unplugged it, checked the SMART output and ran a full test (see smartctl outputs above). The test passed, and even worse seating back the drive gives a perfectly functional RAID1 array.
This is awkward.
Alternative monitoring?
I do not know how to let the P410i tell me what is the specific reason for the "Failed" status (I think it's not possible), and I know these are non-original HP parts (thus invalidating my paid HP support), but for this non mission-critical server I'd like to see if it's still possible to keep using non-HP disks and still have some kind of monitoring of their health status.
What is your opinion? I have 3 questions:
- should HP controller monitoring status be trusted only when used with original parts? (this is easy)
- are these (totally non high quality) SSDs objectively in good health?
- shall I give 100% trust to the results of the SMART tests?
Thanks in advance