-
Notifications
You must be signed in to change notification settings - Fork 0
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
GNU parallel vs xargs #6
Comments
He observed a copy rate of about 1.2GBps within NCI which is about what we'd expect from our |
Hi all. I had a bit more of a think about this, and I came up with the following:
This launches a different The |
Thanks @hot007 & @dsroberts for documenting this here. It's been a little while since I really tested the |
Just resurrecting this post, I've been using this to move lots of data around and I've found that in the case of files of varying size, you can wind up with a 'long tail' problem whereby a large file ends up towards the end of the file list which means the whole command takes much longer to run. I propose the following:
Which sorts the output of |
Thanks for that advice / experience @dsroberts. I have not needed to move lots of data around recently but it's great that you are using and now "tuning" this. Q: In your opinion does this still beat the "new" offerings via Globus? |
I'm moving data between file systems on Gadi, so not really in a place where I can compare it with Globus. |
That is some utterly arcane bash!! That said, that's a good idea, thank you. |
Splitting the files and transferring the chunks in parallel would negate the need to sort by file size. But well beyond anything that can be done sensibly in bash. |
I've never been an
xargs
user, it confuses me, but here's an example of doing anrsync
withxargs
insead ofparallel
, just documenting this here for reference (h/t @dsroberts).The following copies the contents of the current directory in 8 parallel streams, using xargs as a sort of metascheduler.
The text was updated successfully, but these errors were encountered: