Do you use NAS?

Well… I guess I disagree with joeyh there. But then again, maybe the point is rather that you don’t need backup when you got proper archival policies in place…

Who needs backups when you have a first-class archiving system like git-annex anyways? :smile:

This is drifting off-topic (in git-annex land, can we split this thread out?), but I have had very good experience with git-annex in general. I lost data only once: it was a single movie that I hadn’t copied to my other locations which I forcibly removed by mistake. Definitely foot-shooting, and git-annex allows it, but I don’t know of any backup system that could have kept me from pulling that trigger, apart maybe for bup because it doesn’t support removing older snapshots. :stuck_out_tongue:

I’d be curious to hear what software you use for that and where it’s hosted. I host my own gallery at home with sigal but my poor uplink is showing its weakness for large images…

Funny you should ask because if I’m not mistaken I’m using some code you’ve touched for the geolocation/map stuff. I’m hosting with hosthatch OpenVS, because cheapskate, in a location close to where I am. I’m generating my sites/galleries with… https://ikiwiki.info. hehe not superfast with 3k photos but I upload once a week on the private site and 20min generation time don’t bother me as it’s fire and forget. My mother however find the commenting a bit slow…

edit: I’ve looked into sigal as well realistically it does most of what I need. I do intend to expand complicate, elaborate on my sites which prevent me from switching to such simple software atm.

1 Like

About raid, I’m not well read on the topic but I thought it’s a uptime/reliability thing. Only useful for hardware failure? So essentially about decreasing time lost at failure. Not that useful compared to backups when restore time is a non issue. I consider myself and other users an equal or greater greater risk to data than hardware.

1 Like

Correct. The only reason I’m thinking of RAID here is because if you want to collate multiple drives to make a larger one, the failure of any of those drives will take down the entire array. Therefore it becomes more important to have reliability there. The idea would be that instead of having (say) one large 4TB drive, you’d have (say) 4x2TB drives in RAID-10. Performance and reliability would be better, and you can reuse older drives you might otherwise throw away.

Synology RS814 NAS using four 1 TB WD ‘Y’ type server hard drives since about 2014. RS814 communicates via link aggregation to smart switch and thence to household PCs using NFS. NAS data are stored using Synology’s Hybrid RAID. Digital video can be played in real time from this device. This reminds me that I should test (for the fun of it) whether it can support two PCs playing video. Stored photography files are not yet enough file space to consider upgrading to larger disks or adding a piggy-back NAS extension.

Thanks to this topic for reminding me it’s time to back-up my back-up.

1 Like

Do you guys keep your NAS running 24/7 ?
Just wondering how long would HDDs live in these conditions.

I will shut my NAS down, but its under my desk, so not a pain to turn it back on. I’ve had some WD red disks for like 4 years.

Absolutely. Turning things on and off is annoying and error-prone. In particular, my SSHFS setup doesn’t deal well with disconnects, and NFS is not much better either. If you do do that, wake-on-LAN is a good avenue to consider…

HDDs are designed to keep running. It’s counter-intuitive, but spin-up/down can impose more stress on the drive than just the regular spinning. As for the actual numbers, this is usually part of the specs, and also includes how much data is written to the drive.

For example, the Seagate IronWolf 8TB has the following specs, according to Newegg:

  • Workload rate of 180TB/year.
  • Always-on, always-accessible 24×7 performance.
  • 1M hours MTBF, 3-year limited warranty.

If we would be to believe those specs (and if my math is right), it would mean the drive could run continuously for 114 years. Notice how that stands in stark contrast with the 3-year warranty. :wink:

Those drives are designed to stay on and keep a load. You do pay a little more, but if you’re going to be building a NAS anyways, I think it’s worth the cost.

The real problem with drive is random, out of spec failures of all sorts. Some drives will just die earlier than those specs for no reasons: you can return them, get a RMA and a new drive, but your data is still lost. This is why checksums and/or backups and/or RAID or so critical to data integrity and/or reliability and/or high availability. Ultimately, things fail, if only because cosmic radiations will flip a bit on the disk or the controller and damage your data.

Keeping your drives running at least allows you to monitor their health continuously. The real trade-off, for me, is not reliability, but power usage and the associated environmental impact. Then you need to make other calculations regarding the power usage at spin-up/down time and continuous use, which are much harder because less clearly specified.

2 Likes

Yes, turning such stuff on and off reduces its operating life. There’s inrush current that stresses the electronics, and more significant, there are friction and mechanical stress dynamics associated with getting masses such as disk platters spun up.

All of this points to having a well-considered redundancy strategy. And it’s not mirrored RAID, especially with same make/model/purchasedate HDDs…

I think RAID is about availability not redundancy.

Depends on the mode. RAID 1 is “mirroring”, where the identical data is written to at least two separate drives. That is redundancy supporting availability…

Certain RAID modes are also about performance, spreading the head movement to access data across multiple drives. This was actually the original incentive behind the concept.

RAID ( Redundant Array of Independent Disks , originally Redundant Array of Inexpensive Disks )

:wink:

I have never used a NAS but I want to try it. I have a Raspberry Pi kicking around and was thinking to start with something very simple like an RPi with two external HDDs over Samba (shared folder). I will be backing up the HDDs to another drive via rsync once in a while. The main application will be hosting the image collection for digiKam. I also develop the images I like the most using darktable.
The host (RPi) will be connected over 5.8 GHz wi-fi but I can hard wire it if required. The clients are two laptops (one with Linux, the other one with Windows 10) and possibly a desktop PC (Linux). Both laptops support 5.8 GHz wi-fi.
I don’t think I will ever need to access the collection from outside of my home network.

Do you think the setup is going to work well? I suspect it to be a bit slow but I hope the speed will be acceptable what what I do.

wifi can be a bit slow, and 5GHz can especially be problematic if you have walls to cross. That said, Darktable has a “local copy” feature which you can use to keep a local copy of photos that you are currently working on, so that you don’t hit the network all the time.

It should be fine. If it’s not, try wired, and if it’s not, replace the raspi with something more powerful. :slight_smile:

I hard-wire anything that doesn’t move, save the wifi for the tablets and cellphones. On a given channel, you’re competing not only amongst your devices, but anyone else in the neighborhood on that same channel.

Even with a wired connection, I think the bottleneck on your RPi will be the USB interface. I have six RPis around the house (sprinkler control, experimenting with camera control right now), but I do my “NAS” on PC builds with SATA HDDs, wired ethernet. rsync works great; I just run a script with a rsync command every time I upload images from my camera and everything gets put on two machines.

My oldest son has a few.
I have sent his comments as PM to you.

Have fun!
Claes in Lund, Sweden

I agree 100%. I will play around with the Pi and will upgrade to something else if needs be. I have seen some single board computers with SATA interface. I just hope I can keep mainstream and won’t have to deal with systems with custom linux kernel, etc.

Right now, I’m configuring a RPi as a camera image collector/controller. Connected to the camera with USB, running QDSLRDashboard’s ddserver to make the camera controls available on the WiFi to my tablet running QDSLRDashboard. Works, as far as I have tested it, and a whole bunch cheaper than Nikon’s product to do this. Next to try is using gphoto2 alongside this software lashup to capture raws, process them to small JPEGs and post them to some album.

Later, if I decide to pursue some sort of astrophotography, the same box would be a candidate to control the mount. All just good fun…

1 Like

On my desktop pc I have a RAID1 for my home partition. I’m using restic to backup to my NAS system (RAID1) and to a SSD every 6 month with the most important data. The SSD has a friend of mine.

I don’t think there is a better tool than restic for backups! A backup is only a backup if you can verify that the data can be restored!

https://restic.net/

Interesting that I didn’t read any responses the mentioned Amazon Prime backup.

My set up is as follows:

  • SD transfer via Rapid Photo Downloader to
  • Synology 2-bay NAS
    • Mirrored 3TB drives
    • NFS wired networking
    • Synology synchronization with Amazon Prime backup for images

What’s interesting to me, is that JPG and raw (CR2) files do not count against any storage limits in Amazon Prime photo backup. Video does count.