From Newsgroup: alt.os.linux.mint
On Tue, 2/3/2026 7:25 AM, pinnerite wrote:
On Mon, 2 Feb 2026 22:32:35 +0000
pinnerite <pinnerite@gmail.com> wrote:
I rsync my main drive to backups.
However I am a bit messy at filing documents, just keeping them in
temporary places, like "Downloads" for example. Eventually I Will
decide to create one or more directories and move stuff into them.
When I backup of course, the new directories (and contents) will be
copied but the originals will remain on the backup drive. It would
time-consuming to go through the backup drive, work out what had been
duplicated and delete the original files.
In the past, because I have multiple backups, I would delete all the
contents of a backup drive before backing up to it again.
That though is time consuming and seems like using a hammer to crack a
nut.
The advice given (thank you) appears to assume that the file locations
will be in the same place. They won't. The source file will have been
moved to a more suitable directory. The existing backup copy will be
matching the original location of the file (which will no longer be
there of course).
That is why I now get duplicates.
If you don't like this scheme, because it doesn't happen to
hardlink moved duplicates...
# Intro and inode number demo
https://www.admin-magazine.com/Articles/Using-rsync-for-Backups
# More meaty
http://www.mikerubel.org/computers/rsync_snapshots/
you could ask the AI how to modify the script to achieve that end.
The AI seems to be able to read material if you provide a URL
(which means the free ones have limited agentic capability). The
site of the URL though, may not like the visit of Agentic AI,
and may repulse the thing.
Or you can show it the recommended scripted sequence. The first
article, tries to pick the best of what it found on the second
web page.
rm -rf backup.3
mv backup.2 backup.3
mv backup.1 backup.2
cp -al backup.0 backup.1 <=== I think this hard links the entire backup
rsync -a --delete source_directory/ backup.0/
If I was personally "nervous" about what was going on,
then I would "track" the source directory and the destination
directories (the point-in-time backups with their hardlinks
for unchanged files), and manually deduplicate the latest backup
made (backup.0) so that additional things in there which had
been assigned brand new inodes, were replaced with hardlinks
to existing files. Perhaps the rsync can generate a log of sufficient
quality, to track everything that has happened from generation
to generation.
Summary: An insistence on complexity, eventually leads to disaster.
At some point, it's OK to suffer a little inefficiency, if it
means "the thing will never break". You have demonstrated in this
group already, your problems with getting altered NFS mounts to
work properly, as an example of a complexity that required quite
a bit of boot-kicking to fix.
Why do I like Macrium backups ? It's definitely not the efficiency.
It's the "hard to blow up" behavior. I know that the integrity of
backups can be ruined by... bad RAM in the computer! It already happened!
And, by using Verify capability, I caught it in time! So while it wastes
gobs of space for the "free version", it gets the job done in such a
way that I never have to worry about details. It's a point-in-time.
It's "complete". Is it ideal ? Does it cure cancer ? Nope.
If you want it completely deduplicated, you're probably pretty close
already. The AI may figure out a way to finish the job. Your job then,
is to assemble a testbench case, of moved duplicates and so on, in the
same style as the author of the first link above.
Paul
--- Synchronet 3.21b-Linux NewsLink 1.2