Dar Documentation

Dar used within the Max Planck Institute

Who
Max Planck Institute for Gravitational Physics
Context
The Atlas computing cluster (HPC)
Start date
Year 2021
Backup/archive size
from a few 100GB up to 45TByte (compressed) so far - all in 1 TByte chunks with isolated catalogue
Media used
locally HDD (NAS/ZFS), remotely tape LTO8/9 I think at the moment
Why dar
before using regular old tar/star but seqential only access is prohibitively expensive when only need some files from very few chunks in the middle
other remarks
only downside right now is that it seems to be very slow when generating archive/compressing very large directory structures with a LOT of files. But so far, no real time to investigate on my part, just running something like: dar --create PROJECT --verbose=messages \ --fs-root /srv/00transfer/PROJECT --alter=binary\ --slice=1T --min-digits=4 \ --compression=zstd:20:1M --multi-thread=12 \ --slice-mode=0600 --hash=sha512
Source
Carsten Aulbert in December 2025

Back to page of references