-
Notifications
You must be signed in to change notification settings - Fork 340
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
kind duplicacy and other tools benchmark report #635
Comments
Hi @deajan, I took a look at your script and I have a few suggestions: 1-For backup and restore commands, please use the -threads option with 8 threads for your setup. It will significantly increase speed. Increase -threads from 8 until you saturate the network link or see a decrease in speed. 2-During init please play with chunk size: -chunk-size, -c the average size of chunks (default is 4M) With homogeneous data, you should see smaller backups and better deduplication. see Chunk size details 3-Some clarifications for your shopping list on Duplicacy: 1-.Redundant index copies : duplicacy doesn't use indexes. Important: 13- Duplicacy is serverless: Less cost, less maintenance, less attack surface. Hope this helps a bit. Feel free to join the Forum. Keep up the good work. |
Thanks for your time. Table was updated with Also, would you have a link to something explaining why CRIME/BREACH style attacks are not feasable perhaps ? |
Thinking of it, it seems that duplicacy has bigger repository sizes as it's contenders. |
Could I suggest trying out something like Backblaze's B2 with Duplicacy? I just today experimented with restore times on SSH vs. B2 and B2 was 10x faster than SSH (and that was SSH to multiple remote hosts just to confirm) |
Hello,
I'm currently doing benchmarks for deduplication backup tools, including duplicacy.
I decided to write a script that would:
The idea of the script would be to have reproductible results, the only changing factor being the machine specs & network link between sources and targets.
So far, I've run two sets of benchmarks, each done locally and remotely.
You can find the results at https://github.com/deajan/backup-bench
I'd love you to review the recipe I used for duplicacy, and perhaps guide me on what parameters to use to get maximum performance.
Any remarks / ideas / PRs are welcome.
I've also made a comparaison table of some features of the backup solutions I'm benchmarking.
I still miss some informations for some of the backup programs.
Would you mind having a look at the comparaison table and fill the question marks related to the features of duplicacy ?
Also, if duplicacy has an interesting feature I didn't list, I'll be happy to extend the comparaison.
PS: I'm trying to be as unbiased as possible when it comes to those benchmarks, please forgive me if I didn't treat your program with the parameters it deserves.
Also, I've created the same issue in every git repo of the backup tools I'm testing, so every author / team / community member can judge / improve the instructions for better benchmarking.
The text was updated successfully, but these errors were encountered: