Maintenance

  • Home
  • Log in
  • « Lab build process notes for next-gen machines
  • Prepping for new research machines »

CLI: parallel compression FTW

Posted by gregster on 28 Jul 2019 in Announcements

I have a directory with 350,000 xml files and I want to archive it in a compressed format. I tried the usual methods, but they're slooooww - apparently because they are not optimized for parallel operations. I found a thread on superuser.com that recommends pigz (parallel implementation of gzip). It was so much faster I assumed it had cocked-up, but apparently not. It turned my 4GB dir in to a gz file less than 1GB in just a few minutes. Very impressive.

You run it like this:

tar -c --use-compress-program=pigz -f tar.file dir_to_zip
This entry was posted by Greg and filed under Announcements.

Maintenance

This blog is the location for all work involving software and hardware maintenance, updates, installs, etc., both routine and urgent.
  • Home
  • Recently
  • Archives
  • Categories

Search

Categories

  • All
  • Announcements
  • Hit by a bus
  • Labs
    • Activity log
    • Documentation
  • Notes
  • R & D
    • Activity log
    • Documentation
  • Servers
    • Activity log
    • Documentation
  • Tasks

All blogs

  • Academic
  • AdaptiveDB
  • Admin
  • Announcements
  • CanMys
  • Cascade
  • CGWP
  • ColDesp
  • Depts
  • DVPP
  • Endings
  • HCMC Blogs
  • Landscapes
  • LEMDO
  • Linguistics
  • Maint
  • LondonMap
  • Mariage
  • MoM
  • Moses
  • Pro-D
  • Projects
  • ScanCan
  • HumsSites
  • Wendat

This collection ©2023 by admin • Help • Multiple blogs done right!