I’m reading through this and realizing it seems no one has a good solution for this much data.
OK, what follows is basically a long, unfiltered rant/story. Feel free to skip it. 
@guille2306 is “lucky” with “just” 500 GB.
It’s much more managable than the multiple TB archives many of us have. (I have > 3 TB of photos, mostly raw, since 1999. First few years were still JPEGs.) Dealing with this much data (and more, as I have more than just photos) is a super-unfun, expensive, time-consuming chore.
My “solution” is to have a huge working hard drive over USB 3. I periodically rsync that to my 4-disk NAS. And that’s already too technical for most people, really. (And I’ve been using Linux since 1996. And develop software. So that’s saying something.)
Meanwhile, I haven’t figured out cloud storage (due to the massive volume of data and cost and having a slow network until a month ago) and I don’t have an office to drop off a spare drive at (as I work from home).
And I have had data loss — I’ve had several hard drives crash and die over the years. I’ve been lucky enough to be able to scrape some data off of the spinning disks and piece it together with (mostly otherwise complete) backups to get my whole collection back together. (Once, when this happened, I was lucky enough to not have deleted SD cards.) And this is with backing up.
And it always feels like I’m building new hard drives all the time (even for my NAS, which went through a phase of having one hard drive die after another — even one during a rebuild after replacing another; good thing it’s 2 disk failover).
For what it’s worth, my directory structure is:
[Storage location]/Photos/YYYY/YYYY-MM-DD/
so they’re all organized, I can have photos in some consistent manner, I can view them fine in a sorted manner from a file manager (in addition to darktable or digiKam), and there aren’t way too many photos in a single folder. The stuff like “So and so’s Wedding” or “Random park” or “Sister’s birthday” is all in the metadata. I use tags for this. Both darktable and digikam (and basically everything else) can quickly find photos through metadata, especially if you bookmark it (for building a collection, like adding tags for “Calendar 2021” and you’re slowly adding photos to pick from to narrow down).
…Except, really, the weddings. I keep that as [Storage location]/Photos/Weddings/[name of couple]
because they’re (generally) timeless, have lots of people I don’t know, aren’t photos I revisit (that often, unless someone requests it), and I’m not really a wedding photographer.
I used to do the random name affixes to directories and I also used to break down the directories by month, but then I found it’s easy to lose context, hard to search for an exact date, and the browsing should be done in the photo manager, not file manager (generally). (Plus, darktable has an extension to open in a file manager and I think digiKam can do this by default.) So you can still show the containing folder for any photo quite easily, if you want to do “file management” instead.
Yes, it’s redundant to have year folders and include years in the date, but it makes the folders self-contained and more obvious. It’s also redundant with the date/time browsing in darktable, etc., but at least there’s an easy automated way to store files on disk that I don’t need to think about.
Yes, some events span over a day (especially trips) and some days have multiple — but, again, that’s what metadata is for and why darktable, digiKam, and others let you browse by that.
The only thing that really fails with a scheme like this? If you don’t add metadata. But I make it a point to add at least a minimum of the location of where I took the photos. And I have done some archeology to figure out where some photos were taken by looking at spikes of a high # of photos taken together (while date browsing) and looking through events I’ve been to and old emails to figure out where I took something based on when I took it.
I still wish someone made a nice, simple backup solution with a “best practices” guide that:
- doesn’t require use of a terminal
- doesn’t require someone to compile something
- makes backing up somewhere (especially remote) very easy
- has the remote in a file structure that’s comprehensible instead of some hashed or encrypted thing (I care about encryption, but not for my photos), so if I don’t have access to the tool (or if it gets confused for whatever reason), I can just copy them over (via rsync or even a drag-and-drop in a filemanager)
(I can compile something and use a terminal, but when I’m photo mode on personal time, I don’t really want to have to. Plus, others shouldn’t have to anyway.)
All the solutions currently seem to be like “here’s some random program with lots of options; compile it and set it up with a huge config file… it’s super flexible so choose AWS or Backblaze or something… have fun!” (How do Windows and Mac folks deal with this problem of too-much-data? Surely they have it too?)
I guess this all kind of works so far for me? I still have all my photos, despite hardware failures. But it could be a lot better, I know.
TL;DR: Directory format is a solved problem. Metadata is good to add and have, especially for browsing. Storage is a burden. Backups are even more of an chore (especially figuring out what to do about off-site backups).