Skip to main content

Cloud Data Transfer Using RCLONE - Knowledgebase / Research Systems - SOMTech (Technology Services) - VCU School of Medicine

Cloud Data Transfer Using RCLONE

Authors list

Data Transfer from Decode Genetics Iceland using rclone

1.     from the email link download liquidfiles ( excel which contains secret ID and code)

  1. update it under /home/huan/.config/rclone/rclone.conf  ##rclone-1.55.1-1.el7.x86_64 is installed in system as client

  2. use command line to list the cloud concept

  3. rclone ls s3-decode:

rclone lsd s3-decode:

rclone tree s3-decode:

rclone copy s3-decode:directory/ /local/destination/directory/ --progress

rclone move /local/directory/ s3-decode:/directory-on-decode-storage/

Liquidfiles is not an s3 client. Those are totally different protocols.

Liquidfiles was just to transfer the credentials to you :)

You need to use an s3 client to connect to s3-ext.decode.is:10443. That‘s where you use the credentials.

There are multiple clients available:

  • S3 client from amazon (aws cli)

  • S3 client from google (gsutil)

  • S3 client from microsoft (azure cli)

  • 3rd party s3 clients (my favorite is rclone)

  • S3 GUI browsers for smaller transfers that can run on a desktop/laptop computer like windows or mac

You can probably ask your IT helpdesk for assistance with getting an s3 client up and running. Maybe your company has a preferred client.

An example rclone config would look like this:

[s3-decode]

type = s3

provider = other

env_auth = false

endpoint = s3-ext.decode.is:10443

access_key_id = xxx

secret_access_key = xxx


After that I can use commands like:

rclone ls s3-decode:

rclone lsd s3-decode:

rclone tree s3-decode:

rclone copy s3-decode:directory/ /local/destination/directory/ --progress

rclone move /local/directory/ s3-decode:/directory-on-decode-storage/



Helpful Unhelpful