I recently decided to replace the SD card in my Raspberry Pi and reinstall the system. Without any special backups in place, I turned to rsync to duplicate /var/lib/docker with all my containers, including Nextcloud.
Step #1: I mounted an external hard drive to /mnt/temp.
Step #2: I used rsync to copy the data to /mnt/tmp. See the difference?
Step #3: I reformatted the SD card.
Step #4: I realized my mistake.
Moral: no one is immune to their own stupidity 😂
If you have one backup, you have no backup. That's a hard lesson to learn, but if you care about those photos it's possible to recover them if you haven't written stuff on that sdcard yet.
Fuck up #2: using SD cards for data storage. SD cards and USB drives are ephemeral storage devices, not to be relied on. Most of the time they use file systems like FAT32 which are far less safe than NTFS or ext4. Use reliable storage media, like hard drives.
In my experience, flash drives are way more reliable than SD cards and I'd put SSD and HDD above both of those.
I wish they'd just ditch the SD card on the Pi already as it's always the most likely reason why your stuff stops working. For my Pi running Home Assistant, I've swapped to an SDD as the boot drive. For the others, I still use SD cards but they're just doing basic stuff like running Klipper on my 3d printer or a (WIP) live photo frame that can be easily swapped with a replacement SD later.
Much better. SSDs and HDDs do monitor the health of the drives (and you can see many parameters through SMART), while pen drives and SD cards don't.
Of course, they have their limits which is why raid exists. File systems like ZFS are built on the premise that drives are unreliable. It's up to you if you want that redundancy. The most important thing to not lose data is to have backups. Ideally at least 3 copies, 1 off site (e.g. on a cloud, or on a disk at some place other than your home).
SD cards and pen drives are (usally) made from lower quality, cheaper nand (the little memory chips that store the data) and also lack health monitoring, that being said ssds can and do die so it's important to have backups
No luck with extundelete (segfault) and testdisk (sees some deleted files, but not /var/lib/docker). At least I can always throws it away and not worry about safety of my data! :)
If you haven't done much writing to the SD card, you may be able to recover the data. Data isn't really "deleted", it is just labeled as deleted. There is software that can comb through the raw data and try to make sense of what files were there. I don't know of any specific software, so if anyone knows, please reply
Edit: Another commenter mentioned some success with DMDE
Edit 2: Worth mentioning that this is true of formats. As long as it doesn't zero out the entire media, it just edits the file system metadata to say there are no files.
Everyone else is gonna be like "if you don't have at least 3 backups of something blahblah" but you know, not everyone has the finances for that, so advice from a cheapskate computer nerd: when going through critical transfers/reformats/deletions like you were doing, ALWAYS try actually recovering stuff from the backup before you cross the point of no return. E.g. if the backup is a .zip, extract a few individual files from it and open them in their respective programs.
Testdisk and photorec, use them, they even saved my data from bricked Chinese usb flash drive, so it'll save yours unless you wrote dd if /dev/zero of /*/microsd. Also here's the tip, don't attempt to rebuild partition firstly, first step try to copy all files from microsd to another device with these programs and after that try other ways, edit: I've seen from your other comments that your data already was overwritten, my condolences
I know I'm going to get down voted for this but this would be almost impossible to fuck up with a gui. Yet people insist that writing commands manually is superior. I'm sorry for your loss.
This is on me for sure that I've never seen anyone be faster using a CLI compared to a GUI especially for basic operations which is what most of us do 95% of the time. I know there are specific cases where a command just does it better/easier but for me that's not the case for everyday stuff.
There is something to be said about CLI applications being risky by default ("rm" doesn't prompt to ask, rsync --delete will do just that). But I've definitely slipped on the mouse button while "drag & dropping" files in a GUI before. And it can be a right mess if you move a bunch of individual files rather than a sub-folder...
At least for windows, you can ctrl-z that away and it'll handle your mouse fumble. Explorer also highlights the files after a copy so if that doesn't work (and it was a copy action), just delete them immediately.
I haven't used *nix for daily stuff in years but I'm sure the same abilities are there, surely.
To play devil's advocate, tab completion would have also likely caught this. OP could have typed /mnt/t<Tab> and it would autofill temp, or <Tab><Tab> would show the matching options if it's ambiguous.
The first one would have deleted nothing as it needs to match the whole name. I recommend running find with an explicit -print before replacing it in place with -delete or -exec. It's good to remember that find has a complex order dependent language with -or and -and, but not maybe the best idea to try to use those features.
Unlesa you did a full zeroing format the info might still be available. There was an applicarion that attempts to rebuild the partition / Filesystem from left over meta data or inode info. I forget the name unfortunately.
Normall the strings command will get your photos but probably not if they were in a docker image database.
Those are good for sure. And maybe it was testdisk. There was one that just undeleted the partition table delete. as long as new data had not been written everthing would be intact
Rclone is superior IMHO, you have to explicitly name the output folder. Used to think it was a hassle but in hindsight being explicit about the destination reduces mistakes.
I think everyone has done this. I know I have. I believe I used dmda to recover all my photos back. Unfortunately I lost all the metadata for about 3000 photos. It took years to manually retap and redate them all but at least I didn’t lose them forever
This. I made a similarly boneheaded mistake a few months back and lost a fair amount of stuff. Thankfully 95+% of it was ancient program files and such that gave me a good excuse to go through and make some much needed cleanup actually happen for once.