Filesystem packer slows down after 30k files
After the filesystem packer has hashed all 255k files then the DB operations starts to slow down the entire application.
The DB writes get to 30k files before the TAR writer catches up and slows down to the DB writers speed. then it uses hours maybe days to finish...
Any way I can speed this up?
https://github.com/OptoCloud/OptoPacker
Current status:
GitHub
GitHub - OptoCloud/OptoPacker: Pre-packs huge filesystems containin...
Pre-packs huge filesystems containing repositories or other projects for compression, parsing gitignore files to exclude unnessecary files from packing - OptoCloud/OptoPacker
3 Replies
tar writer has to wait for DB job before doing its thing because if a BLOB with a matching hash has already been written to the TAR file then there is no use writing it again
Application workflow:
dumping a ton of binary blobs into a database (especially sqlite) sounds slow
do you really need the file contents in there?
im not dumping the binary blobs into the database
only their hashes
the blobs im writing to a tar stream thats streamed into 7zip
sorry for the miswrite
I tried to explain it better now
I updated my code a bit more to try to maybe optimize it?
Idk how much more I can do
Bump?
@Jimmacle 🤔