[ale] best way to copy 3Tb of data

Brian Mathis brian.mathis+ale at betteradmin.com
Wed Oct 28 13:15:08 EDT 2015


The default SSH implementation does have intrinsic performance limits, as
outlined here:
    http://www.psc.edu/index.php/hpn-ssh
so not using SSH might help with the speed, as long as the loss of security
can be tolerated.


~ Brian Mathis
@orev


On Tue, Oct 27, 2015 at 11:06 AM, Jim Kinney <jim.kinney at gmail.com> wrote:

> The slowdown is the encypt/decrypt. The non-server method relies on ssh
> for transport. You can also use rsh for no security at all and it will be
> faster. By using the rsync server, you drop the ssh security so if the user
> must enter credentials to access the NAS, you might want to double check if
> the rsync credentials are sent as plain text in the same way as ftp.
> On Oct 27, 2015 10:49 AM, "Todor Fassl" <fassl.tod at gmail.com> wrote:
>
>> I know you don't need to have an rsync server to copy files via rsync but
>> from what I've read, rsync protocol is way fater than ssh. And you have to
>> have an rsync server to use the rsync protocol, right?
>>
>> On 10/27/2015 08:46 AM, Jim Kinney wrote:
>>
>>> Rsync doesn't require an rsync server. It provides a solid backup. Rsync
>>> it back and it's all golden.
>>>
>>> Tarball will need enough space to be built or will need to be built
>>> 'over the wire' using a tar|<transport method>|tar process. Second
>>> optional.
>>>
>>> Tar is faster but rsync is easier.
>>>
>>> A 4TB external hard drive and sneaker net also works and provides
>>> verifiable copies. Rsync to a local drive is fast especially with an
>>> external sata port.
>>>
>>> On Oct 27, 2015 9:37 AM, "Todor Fassl" <fassl.tod at gmail.com
>>> <mailto:fassl.tod at gmail.com>> wrote:
>>>
>>>     One of the researchers I support wants to backup 3T of data to his
>>>     space on our NAS. The data is on an HPC cluster on another network.
>>>     It's not an on-going backup. He just needs to save it to our NAS
>>>     while the HPC cluster is rebuilt. Then he'll need to copy it right
>>> back.
>>>
>>>     There is a very stable 1G connection between the 2 networks. We have
>>>     plenty of space on our NAS. What is the best way to do the caopy?
>>>     Ideally, it seems we'd want to have boththe ability to restart the
>>>     copy if it fails part way through and to end up with a compressed
>>>     archive like a tarball. Googling around tends to suggest that it's
>>>     eitehr rsync or tar. But with rsync, you wouldn't end up with a
>>>     tarball. And with tar, you can't restart it in the middle. Any other
>>>     ideas?
>>>     Since the network connection is very stable, I am thinking of
>>>     suggesting tar.
>>>
>>>     tar zcvf - /datadirectory | ssh user at backup.server "cat >
>>>     backupfile.tgz"
>>>
>>>     If the researcher would prefer his data to be copied to our NAS as
>>>     regular files, just use rsync with compression. We don't have an
>>>     rsync server that is accessible to the outside world. He could use
>>>     ssh with rsync but I could set up rsync if it would be worthwhile.
>>>
>>>     Ideas? Suggestions?
>>>
>>>
>>>
>>>     on at the far end.
>>>
>>>     He is going to need to copy the data back in a few weeks. It might
>>>     even be worthwhile to send it via tar without
>>>     uncompressing/unarchiving it on receiving end.
>>>
>>>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://mail.ale.org/pipermail/ale/attachments/20151028/4bfbd4bd/attachment.html>


More information about the Ale mailing list