ElJimador

Members
  • Posts

    234
  • Joined

  • Last visited

Converted

  • Gender
    Male
  • Location
    Tallahassee, FL

Recent Profile Visitors

The recent visitors block is disabled and is not being shown to other users.

ElJimador's Achievements

Explorer

Explorer (4/14)

0

Reputation

  1. Okay, thanks for confirming. The parity check is resumed and I'll keep this open until it completes just in case any other issues arise but I'll be sure to remember to come back and mark it solved then. Thank you!
  2. Thanks Jorge. So just to be sure on this, corrections are only written from data drives to parity and never the other way around? And since the current parity check I have paused is a correcting check (showing 735 corrections already, whatever I intended), I should just go ahead and resume it? And final question, if I do that and the error count continues to rise, at what point would you be concerned and pause it again? I've had data loss in the past from drives dropping off during a parity check so that's my biggest concern here. And though it'd be slower to do a new config and let parity be completely rewritten again, if it's in any way safer I wouldn't mind. Appreciate your help as always.
  3. Hello. Thanks to a faulty UPS I had an unclean shutdown on my HOME server in the middle of a big file transfer directly to a disk share and running the parity check required after I unchecked the box to write corrections to parity. Despite this I wake up this morning to 735 sync errors corrected. (The check was barely halfway through since with my massively underpowered CPU and now dual 12TB parity, checks take around 6 days now.) Anyway I've paused the check and am attaching diagnostics here if needed but what I really need to know is, do sync errors corrected only overwrite parity, or are corrections also (or only?) being written to data drives from parity? If it's only parity being re-written should I resume the check and let it complete no matter how many more errors may register (assuming the result would be the same as doing a new config with all drives in their current assignments and letting parity be completely re-written)? Or if data drives are being overwritten (with possibly corrupted data given the circumstances) is there any way to pinpoint where those errors might be before I stop the check, and then what would next steps be from there? Thanks. home-diagnostics-20240327-0441.zip
  4. Hey Jorge (or anyone else who can help). After running ddrescue for days, the desktop where I was running the terminal session got accidentally unplugged. Is there a command string I can use in a new session to see where the process is at this point, or just to end it safely? I assume it's still running on the HOME server even though I no longer have any visibility to it. From what I was seeing throughout the operation right up until a day or 2 ago, it was just counting up endless read errors and not saving anything anyway, so at this point I'm just resigned to having lost D6 entirely and recovering in the new config as best I can from offsite backups. Thanks.
  5. Thanks Jorge. Doing it now with the array stopped and it appears to be working fine so far.
  6. I did, thanks. I tried the option to copy the failed D6 to an array disk (D7) in maintenance mode and got the following. Is my command string incorrect? Since I had D7 backed up, I tried again only this time I started the array first and deleted everything off of D7 before re-starting in maintenance mode, with the same result. Since the destination disk is empty I'm not sure why it's saying there's no space left on it? root@HOME:~# ddrescue -f /dev/sdj1 /dev/md7 /boot/ddrescue.log GNU ddrescue 1.27 Press Ctrl-C to interrupt Initial status (read from mapfile) rescued: 8388 kB, tried: 0 B, bad-sector: 0 B, bad areas: 0 Current status ipos: 8388 kB, non-trimmed: 0 B, current rate: 0 B/s opos: 8388 kB, non-scraped: 0 B, average rate: 0 B/s non-tried: 6001 GB, bad-sector: 0 B, error rate: 0 B/s rescued: 8388 kB, bad areas: 0, run time: 0s pct rescued: 0.00%, read errors: 0, remaining time: n/a time since last successful read: n/a Copying non-tried blocks... Pass 1 (forwards) ddrescue: /dev/md7: Write error: No space left on device root@HOME:~#
  7. Everything that was on D11 is backed up locally so it won't be any problem copying it back to the new config once it's up and parity has been re-written. D6 is the one I want to clone because most of it is backed up remotely to a server I have at my parents 2,000 miles away, so recovery would be much more of a chore. So my question is how to clone D6 and actually recover that data (as much as can be) before I do the new config? I don't want to try to recover anything from parity or do anything that requires a parity sync before then. I just want to understand if I clone D6 to a same size drive already in the array (backing that drive up first to free up the space), should I be able to start the array after exactly as it is now, since that drive is already mounted, to see how much successfully copied to it? If so that seems likely to be the best option and I can figure out exactly how I want to map everything the new config and then write back from local backups after that.
  8. No. Old D11 had a SMART warning after a ton of errors to it on the last parity sync before I made any changes, which is why I was replacing it in the first place. I don't want to put it back in the array, especially since everything that was on it is already backed up locally and can be easily copied back once the new config is up and running. And actually it turns out the other 6TB I had in mind for the D6 clone isn't ideal either. I'm scanning it with StableBit on my desktop and 98% in it's showing sectors it can't read to, even though as yet there is no SMART warning. How about if I back up another 6TB drive already in the array (D7) and let ddrescue clone the old D6 to that instead? Wouldn't that allow me to re-start the array after to see what copied successfully, since D7 is already mounted?
  9. Hi Jorge. Just coming back to this since I've been tied up with work this week. Want to make sure I understand and am doing this correctly before I attempt it. Because having thought it through my preference now is to do a new config dropping the new D11 and adding it as a 2nd parity instead, and replacing the now failed and unmountable D6 with it's clone. So check me on this please. I shut down and replace D11 with a good 6TB drive (same size as D6) at which point I won't be able to restart the array (I assume even in maintenance mode?) due to too many failed/wrong disks. But I can run ddrescue to clone D6 to the new drive. Then once that completes, I shut down again, pull the failed D6 and replace it with the clone and put back the new D11 which I assign in the new config as the 2nd parity. Whatever did or did not copy successfully to the new D6 I'll see when I start the array and parity is re-written for the new config. Correct? Or does the new D6 get formatted first so that whatever was cloned is just wiped right away? If it's the latter then how would I actually back up the data off the clone so I can copy it back to the array after? I hadn't mentioned before, at least in this thread, but another part of my reason for wanting to do a new config is that there is another drive (D8) which I'd prefer to replace now too because although SMART is still showing it as healthy, there were some errors to it also during the last parity sync (before the snafu replacing D11 and starting this thread). So I'd prefer to just swap it for a new 12TB now and use it in my desktop instead before it gets worse. D8 and D11 I have full local backups for that I'll copy back to the new config after. Anyway, not sure if that makes any difference to what I'm trying to do cloning/recovering D6 or with the new config in general but just thought I'd mention for you and anyone else who might have input it in case it brings to mind any additional issues I'm not thinking of? Thanks again.
  10. Thanks Jorge, I'll look into that. Meantime I'm also starting to think if I go the New Config route about just leaving off D11 as a last data drive and adding it as a 2nd parity instead. My only worry is that this is a low power server with an extremely weak CPU and parity checks on it already take 3 days while I can't do anything else with it. So would a second parity drive make that take even twice as long? I like the idea of not losing data in a scenario like this again but I wonder if it's just not advisable with the hardware I have. (This is my HOME server if you want to see the specs in my sig below).
  11. Started the array with D6 enabled and it shows on the Main page as "Unmountable: Unsupported or no file system" and there is no disk6 visible at all when I try to open any emulated data through Windows. New diagnostics attached. Also D11 has folders but they are all empty even tough Main still shows 5+ TB used on the drive. So what do I do with that now too? home-diagnostics-20231210-1853.zip
  12. Thanks. Okay I powered down, checked all connections and when I booted up I had D5 read as missing as well as D6 (they sit next to each other and share the same sata power connector via a Y-splitter). So I powered down again and this time I took both drives out and cleaned all the contact points on all the cables and the drives themselves with an eraser + rubbing alcohol (though nothing really appeared to be dirty) and now both drives are there again but D6 is still showing as disabled and an extended SMART test ended with "Completed: Read Failure". New Diagnostics and SMART report for D6 are attached. So what now? On MAIN it looks like I can start the array either with D6 disabled or changed to "no device" without starting a new parity operation either way. But is there any risk to that vs. assuming D6 is not going to have any data emulated that can still be backed up and just replacing it right away? And if I do replace it, would it be better to do it as a New Config to completely re-write parity from scratch? Or does doing it as a regular drive replacement / data rebuild (even though there's probably nothing there) just do exactly the same thing? Not being 100% confident now with the physical connections to the drives, I'm most concerned with the possibility of another drive falling off during a new parity operation and suffering even further data loss. But I suppose there'd be the same risk of that either way? Even though D6 is in error now I can't help thinking there was nothing actually wrong with it until the last parity op ran through with the bad connection and as much as I can help it I just want to make sure it doesn't happen again. home-diagnostics-20231210-1229.zip home D6-smart-20231210-1229.zip
  13. Hello. I was replacing D11 on my HOME server and during the Data Rebuild / Parity Sync a different drive (D6) which had shown no issues before decided to take a crap. Now at the end, D6 shows as disabled with contents emulated but there's no actual data visible (just a few empty folders) when opening either D6 or the new D11 though my Windows desktop, even though the UnRaid Main page shows 5+ TB used on both. My hope is that there is still data on both drives and that maybe D6 just had a sata or power connector come loose early in the op? It was a green ball like all the rest when I started and when the errors to it started piling up I did notice that there was no temperature read for the drive even though it stayed green until the end. Anyway, what to do next? Everything is backed up, though some on D6 is offsite and would be a little harder to recover. I expect I'll have to do a New Config to completely rebuild parity once I assess what's really still there, but how do I even find out what's actually on the drives now without risking (further) data loss? Diagnostics attached and any help appreciated. home-diagnostics-20231209-1109.zip
  14. That's a lot easier actually. Thanks! I'll do that.
  15. Hmm, okay. Well I do have everything backed up to other servers. Not all of it is easy to get to, but assuming I can pull those backups together is there a way to use New Config to rebuild parity entirely with new replacement Disks 11 and 8 restored from those backups and make sure there is no corruption going forward that way? I hate the idea of having corruption that I don't even know where it is until I'm halfway into a movie or a TV season when it pops up. So whatever has the best chance of avoiding that please let me know.