2014-03-29 06:34:13 +08:00
---
title: "Amazon S3"
description: "Rclone docs for Amazon S3"
2016-07-11 19:12:28 +08:00
date: "2016-07-11"
2014-03-29 06:34:13 +08:00
---
2018-04-13 00:05:53 +08:00
< i class = "fa fa-amazon" > < / i > Amazon S3 Storage Providers
--------------------------------------------------------
2018-04-13 23:08:00 +08:00
The S3 backend can be used with a number of different providers:
* {{< provider name = "AWS S3" home = "https://aws.amazon.com/s3/" config = "/s3/#amazon-s3" > }}
2018-04-13 00:05:53 +08:00
* {{< provider name = "Ceph" home = "http://ceph.com/" config = "/s3/#ceph" > }}
* {{< provider name = "DigitalOcean Spaces" home = "https://www.digitalocean.com/products/object-storage/" config = "/s3/#digitalocean-spaces" > }}
2018-04-13 23:08:00 +08:00
* {{< provider name = "Dreamhost" home = "https://www.dreamhost.com/cloud/storage/" config = "/s3/#dreamhost" > }}
2018-04-13 00:05:53 +08:00
* {{< provider name = "IBM COS S3" home = "http://www.ibm.com/cloud/object-storage" config = "/s3/#ibm-cos-s3" > }}
* {{< provider name = "Minio" home = "https://www.minio.io/" config = "/s3/#minio" > }}
* {{< provider name = "Wasabi" home = "https://wasabi.com/" config = "/s3/#wasabi" > }}
2014-07-18 03:03:11 +08:00
Paths are specified as `remote:bucket` (or `remote:` for the `lsd`
command.) You may put subdirectories in too, eg `remote:bucket/path/to/dir` .
2014-03-29 06:34:13 +08:00
2018-04-13 23:08:00 +08:00
Once you have made a remote (see the provider specific section above)
you can use it like this:
See all buckets
rclone lsd remote:
Make a new bucket
rclone mkdir remote:bucket
List the contents of a bucket
rclone ls remote:bucket
Sync `/home/local/directory` to the remote bucket, deleting any excess
files in the bucket.
rclone sync /home/local/directory remote:bucket
## AWS S3 {#amazon-s3}
2014-03-29 06:34:13 +08:00
Here is an example of making an s3 configuration. First run
rclone config
This will guide you through an interactive setup process.
```
No remotes found - make a new one
n) New remote
2016-02-21 21:39:04 +08:00
s) Set configuration password
2018-03-14 04:47:29 +08:00
q) Quit config
n/s/q> n
2014-03-29 06:34:13 +08:00
name> remote
2016-02-21 21:39:04 +08:00
Type of storage to configure.
Choose a number from below, or type in your own value
2018-03-14 04:47:29 +08:00
1 / Alias for a existing remote
\ "alias"
2 / Amazon Drive
2016-02-21 21:39:04 +08:00
\ "amazon cloud drive"
2018-04-13 23:08:00 +08:00
3 / Amazon S3 Compliant Storage Providers (AWS, Ceph, Dreamhost, IBM COS, Minio)
2016-02-21 21:39:04 +08:00
\ "s3"
2018-03-14 04:47:29 +08:00
4 / Backblaze B2
2016-02-21 21:39:04 +08:00
\ "b2"
2018-03-14 04:47:29 +08:00
[snip]
23 / http Connection
\ "http"
2018-04-13 23:08:00 +08:00
Storage> s3
Choose your S3 provider.
2018-04-13 00:05:53 +08:00
Choose a number from below, or type in your own value
2018-04-13 23:08:00 +08:00
1 / Amazon Web Services (AWS) S3
2018-04-13 00:05:53 +08:00
\ "AWS"
2018-04-13 23:08:00 +08:00
2 / Ceph Object Storage
2018-04-13 00:05:53 +08:00
\ "Ceph"
2018-04-13 23:08:00 +08:00
3 / Digital Ocean Spaces
\ "DigitalOcean"
4 / Dreamhost DreamObjects
2018-04-13 00:05:53 +08:00
\ "Dreamhost"
2018-04-13 23:08:00 +08:00
5 / IBM COS S3
2018-04-13 00:05:53 +08:00
\ "IBMCOS"
2018-04-13 23:08:00 +08:00
6 / Minio Object Storage
2018-04-13 00:05:53 +08:00
\ "Minio"
2018-04-13 23:08:00 +08:00
7 / Wasabi Object Storage
\ "Wasabi"
8 / Any other S3 compatible provider
\ "Other"
provider> 1
2017-11-23 05:21:36 +08:00
Get AWS credentials from runtime (environment variables or EC2/ECS meta data if no env vars). Only applies if access_key_id and secret_access_key is blank.
2016-02-01 21:11:27 +08:00
Choose a number from below, or type in your own value
2016-02-21 21:39:04 +08:00
1 / Enter AWS credentials in the next step
\ "false"
2 / Get AWS credentials from the environment (env vars or IAM)
\ "true"
env_auth> 1
2016-02-01 21:11:27 +08:00
AWS Access Key ID - leave blank for anonymous access or runtime credentials.
2018-03-14 04:47:29 +08:00
access_key_id> XXX
2016-02-01 21:11:27 +08:00
AWS Secret Access Key (password) - leave blank for anonymous access or runtime credentials.
2018-03-14 04:47:29 +08:00
secret_access_key> YYY
2018-04-13 23:08:00 +08:00
Region to connect to.
2014-03-29 06:34:13 +08:00
Choose a number from below, or type in your own value
2016-02-21 21:39:04 +08:00
/ The default endpoint - a good choice if you are unsure.
1 | US Region, Northern Virginia or Pacific Northwest.
| Leave location constraint empty.
\ "us-east-1"
2018-03-14 04:47:29 +08:00
/ US East (Ohio) Region
2 | Needs location constraint us-east-2.
\ "us-east-2"
2016-02-21 21:39:04 +08:00
/ US West (Oregon) Region
2018-03-14 04:47:29 +08:00
3 | Needs location constraint us-west-2.
2016-02-21 21:39:04 +08:00
\ "us-west-2"
/ US West (Northern California) Region
2018-03-14 04:47:29 +08:00
4 | Needs location constraint us-west-1.
2016-02-21 21:39:04 +08:00
\ "us-west-1"
2018-03-14 04:47:29 +08:00
/ Canada (Central) Region
5 | Needs location constraint ca-central-1.
\ "ca-central-1"
/ EU (Ireland) Region
6 | Needs location constraint EU or eu-west-1.
2016-02-21 21:39:04 +08:00
\ "eu-west-1"
2018-03-14 04:47:29 +08:00
/ EU (London) Region
7 | Needs location constraint eu-west-2.
\ "eu-west-2"
2016-02-21 21:39:04 +08:00
/ EU (Frankfurt) Region
2018-03-14 04:47:29 +08:00
8 | Needs location constraint eu-central-1.
2016-02-21 21:39:04 +08:00
\ "eu-central-1"
/ Asia Pacific (Singapore) Region
2018-03-14 04:47:29 +08:00
9 | Needs location constraint ap-southeast-1.
2016-02-21 21:39:04 +08:00
\ "ap-southeast-1"
/ Asia Pacific (Sydney) Region
2018-03-14 04:47:29 +08:00
10 | Needs location constraint ap-southeast-2.
2016-02-21 21:39:04 +08:00
\ "ap-southeast-2"
/ Asia Pacific (Tokyo) Region
2018-03-14 04:47:29 +08:00
11 | Needs location constraint ap-northeast-1.
2016-02-21 21:39:04 +08:00
\ "ap-northeast-1"
2017-01-09 13:09:19 +08:00
/ Asia Pacific (Seoul)
2018-03-14 04:47:29 +08:00
12 | Needs location constraint ap-northeast-2.
2017-01-09 13:09:19 +08:00
\ "ap-northeast-2"
/ Asia Pacific (Mumbai)
2018-03-14 04:47:29 +08:00
13 | Needs location constraint ap-south-1.
2017-01-09 13:09:19 +08:00
\ "ap-south-1"
2016-02-21 21:39:04 +08:00
/ South America (Sao Paulo) Region
2018-03-14 04:47:29 +08:00
14 | Needs location constraint sa-east-1.
2016-02-21 21:39:04 +08:00
\ "sa-east-1"
region> 1
2015-08-16 01:44:45 +08:00
Endpoint for S3 API.
Leave blank if using AWS to use the default endpoint for the region.
2018-04-13 23:08:00 +08:00
endpoint>
2015-08-16 01:44:45 +08:00
Location constraint - must be set to match the Region. Used when creating buckets only.
2014-03-29 06:34:13 +08:00
Choose a number from below, or type in your own value
2016-02-21 21:39:04 +08:00
1 / Empty for US Region, Northern Virginia or Pacific Northwest.
\ ""
2018-03-14 04:47:29 +08:00
2 / US East (Ohio) Region.
\ "us-east-2"
3 / US West (Oregon) Region.
2016-02-21 21:39:04 +08:00
\ "us-west-2"
2018-03-14 04:47:29 +08:00
4 / US West (Northern California) Region.
2016-02-21 21:39:04 +08:00
\ "us-west-1"
2018-03-14 04:47:29 +08:00
5 / Canada (Central) Region.
\ "ca-central-1"
6 / EU (Ireland) Region.
2016-02-21 21:39:04 +08:00
\ "eu-west-1"
2018-03-14 04:47:29 +08:00
7 / EU (London) Region.
\ "eu-west-2"
8 / EU Region.
2016-02-21 21:39:04 +08:00
\ "EU"
2018-03-14 04:47:29 +08:00
9 / Asia Pacific (Singapore) Region.
2016-02-21 21:39:04 +08:00
\ "ap-southeast-1"
2018-03-14 04:47:29 +08:00
10 / Asia Pacific (Sydney) Region.
2016-02-21 21:39:04 +08:00
\ "ap-southeast-2"
2018-03-14 04:47:29 +08:00
11 / Asia Pacific (Tokyo) Region.
2016-02-21 21:39:04 +08:00
\ "ap-northeast-1"
2018-03-14 04:47:29 +08:00
12 / Asia Pacific (Seoul)
2017-01-09 13:09:19 +08:00
\ "ap-northeast-2"
2018-03-14 04:47:29 +08:00
13 / Asia Pacific (Mumbai)
2017-01-09 13:09:19 +08:00
\ "ap-south-1"
2018-03-14 04:47:29 +08:00
14 / South America (Sao Paulo) Region.
2016-02-21 21:39:04 +08:00
\ "sa-east-1"
location_constraint> 1
2016-08-22 20:59:03 +08:00
Canned ACL used when creating buckets and/or storing objects in S3.
2017-03-29 20:38:34 +08:00
For more info visit https://docs.aws.amazon.com/AmazonS3/latest/dev/acl-overview.html#canned-acl
2016-08-22 20:59:03 +08:00
Choose a number from below, or type in your own value
1 / Owner gets FULL_CONTROL. No one else has access rights (default).
\ "private"
2 / Owner gets FULL_CONTROL. The AllUsers group gets READ access.
\ "public-read"
/ Owner gets FULL_CONTROL. The AllUsers group gets READ and WRITE access.
3 | Granting this on a bucket is generally not recommended.
\ "public-read-write"
4 / Owner gets FULL_CONTROL. The AuthenticatedUsers group gets READ access.
\ "authenticated-read"
/ Object owner gets FULL_CONTROL. Bucket owner gets READ access.
5 | If you specify this canned ACL when creating a bucket, Amazon S3 ignores it.
\ "bucket-owner-read"
/ Both the object owner and the bucket owner get FULL_CONTROL over the object.
6 | If you specify this canned ACL when creating a bucket, Amazon S3 ignores it.
\ "bucket-owner-full-control"
2018-03-14 04:47:29 +08:00
acl> 1
2016-06-15 04:22:54 +08:00
The server-side encryption algorithm used when storing this object in S3.
Choose a number from below, or type in your own value
1 / None
\ ""
2 / AES256
\ "AES256"
2018-03-14 04:47:29 +08:00
server_side_encryption> 1
2016-09-02 05:27:50 +08:00
The storage class to use when storing objects in S3.
Choose a number from below, or type in your own value
1 / Default
\ ""
2 / Standard storage class
\ "STANDARD"
3 / Reduced redundancy storage class
\ "REDUCED_REDUNDANCY"
4 / Standard Infrequent Access storage class
\ "STANDARD_IA"
2018-04-13 20:36:25 +08:00
5 / One Zone Infrequent Access storage class
\ "ONEZONE_IA"
2018-12-06 22:35:29 +08:00
6 / Glacier storage class
\ "GLACIER"
2018-03-14 04:47:29 +08:00
storage_class> 1
2015-08-16 01:44:45 +08:00
Remote config
2014-03-29 06:34:13 +08:00
--------------------
[remote]
2018-04-13 23:08:00 +08:00
type = s3
provider = AWS
2016-02-21 21:39:04 +08:00
env_auth = false
2018-03-14 04:47:29 +08:00
access_key_id = XXX
secret_access_key = YYY
2016-02-21 21:39:04 +08:00
region = us-east-1
2018-04-13 23:08:00 +08:00
endpoint =
location_constraint =
2017-01-09 13:09:19 +08:00
acl = private
2018-04-13 23:08:00 +08:00
server_side_encryption =
storage_class =
2014-03-29 06:34:13 +08:00
--------------------
y) Yes this is OK
e) Edit this remote
d) Delete this remote
2018-04-13 23:08:00 +08:00
y/e/d>
2014-03-29 06:34:13 +08:00
```
2017-06-06 23:40:00 +08:00
### --fast-list ###
This remote supports `--fast-list` which allows you to use fewer
transactions in exchange for more memory. See the [rclone
docs](/docs/#fast-list) for more details.
2018-04-13 20:32:17 +08:00
### --update and --use-server-modtime ###
As noted below, the modified time is stored on metadata on the object. It is
used by default for all operations that require checking the time a file was
last updated. It allows rclone to treat the remote more like a true filesystem,
but it is inefficient because it requires an extra API call to retrieve the
metadata.
For many operations, the time the object was last uploaded to the remote is
sufficient to determine if it is "dirty". By using `--update` along with
`--use-server-modtime` , you can avoid the extra API call and simply upload
files whose local modtime is newer than the time it was last uploaded.
2015-06-06 17:05:21 +08:00
### Modified time ###
2014-03-29 06:34:13 +08:00
The modified time is stored as metadata on the object as
`X-Amz-Meta-Mtime` as floating point since the epoch accurate to 1 ns.
2015-08-16 01:44:45 +08:00
### Multipart uploads ###
rclone supports multipart uploads with S3 which means that it can
2018-11-27 05:09:23 +08:00
upload files bigger than 5GB.
Note that files uploaded *both* with multipart upload *and* through
crypt remotes do not have MD5 sums.
Rclone switches from single part uploads to multipart uploads at the
point specified by `--s3-upload-cutoff` . This can be a maximum of 5GB
and a minimum of 0 (ie always upload mulipart files).
The chunk sizes used in the multipart upload are specified by
`--s3-chunk-size` and the number of chunks uploaded concurrently is
specified by `--s3-upload-concurrency` .
Multipart uploads will use `--transfers` * `--s3-upload-concurrency` *
`--s3-chunk-size` extra memory. Single part uploads to not use extra
memory.
Single part transfers can be faster than multipart transfers or slower
depending on your latency from S3 - the more latency, the more likely
single part transfers will be faster.
Increasing `--s3-upload-concurrency` will increase throughput (8 would
be a sensible value) and increasing `--s3-chunk-size` also increases
througput (16M would be sensible). Increasing either of these will
use more memory. The default values are high enough to gain most of
the possible performance without using too much memory.
2015-08-16 01:44:45 +08:00
2015-08-26 03:15:50 +08:00
### Buckets and Regions ###
With Amazon S3 you can list buckets (`rclone lsd`) using any region,
but you can only access the content of a bucket from the region it was
created in. If you attempt to access a bucket from the wrong region,
you will get an error, `incorrect region, the bucket is not in 'XXX'
region`.
2016-02-01 21:11:27 +08:00
### Authentication ###
2018-04-13 23:08:00 +08:00
2018-04-16 19:14:35 +08:00
There are a number of ways to supply `rclone` with a set of AWS
credentials, with and without using the environment.
The different authentication methods are tried in this order:
- Directly in the rclone configuration file (`env_auth = false` in the config file):
- `access_key_id` and `secret_access_key` are required.
- `session_token` can be optionally set when using AWS STS.
- Runtime configuration (`env_auth = true` in the config file):
- Export the following environment variables before running `rclone` :
2016-02-10 01:19:13 +08:00
- Access Key ID: `AWS_ACCESS_KEY_ID` or `AWS_ACCESS_KEY`
- Secret Access Key: `AWS_SECRET_ACCESS_KEY` or `AWS_SECRET_KEY`
2018-04-16 19:14:35 +08:00
- Session Token: `AWS_SESSION_TOKEN` (optional)
- Or, use a [named profile ](https://docs.aws.amazon.com/cli/latest/userguide/cli-multiple-profiles.html ):
- Profile files are standard files used by AWS CLI tools
- By default it will use the profile in your home directory (eg `~/.aws/credentials` on unix based systems) file and the "default" profile, to change set these environment variables:
- `AWS_SHARED_CREDENTIALS_FILE` to control which file.
- `AWS_PROFILE` to control which profile to use.
- Or, run `rclone` in an ECS task with an IAM role (AWS only).
- Or, run `rclone` on an EC2 instance with an IAM role (AWS only).
2016-02-01 21:11:27 +08:00
If none of these option actually end up providing `rclone` with AWS
credentials then S3 interaction will be non-authenticated (see below).
2017-06-02 19:06:06 +08:00
### S3 Permissions ###
2018-04-13 00:05:53 +08:00
When using the `sync` subcommand of `rclone` the following minimum
2017-06-02 19:06:06 +08:00
permissions are required to be available on the bucket being written to:
* `ListBucket`
* `DeleteObject`
2017-06-10 22:22:43 +08:00
* `GetObject`
2017-06-02 19:06:06 +08:00
* `PutObject`
* `PutObjectACL`
Example policy:
```
{
"Version": "2012-10-17",
"Statement": [
{
"Effect": "Allow",
"Principal": {
"AWS": "arn:aws:iam::USER_SID:user/USER_NAME"
},
"Action": [
"s3:ListBucket",
"s3:DeleteObject",
2017-06-10 22:22:43 +08:00
"s3:GetObject",
2017-06-02 19:06:06 +08:00
"s3:PutObject",
"s3:PutObjectAcl"
],
"Resource": [
"arn:aws:s3:::BUCKET_NAME/*",
"arn:aws:s3:::BUCKET_NAME"
]
}
]
}
```
Notes on above:
1. This is a policy that can be used when creating bucket. It assumes
that `USER_NAME` has been created.
2018-04-13 00:05:53 +08:00
2. The Resource entry must include both resource ARNs, as one implies
2017-06-02 19:06:06 +08:00
the bucket and the other implies the bucket's objects.
2018-04-13 00:05:53 +08:00
For reference, [here's an Ansible script ](https://gist.github.com/ebridges/ebfc9042dd7c756cd101cfa807b7ae2b )
2017-06-02 19:06:06 +08:00
that will generate one or more buckets that will work with `rclone sync` .
2018-03-17 18:51:45 +08:00
### Key Management System (KMS) ###
If you are using server side encryption with KMS then you will find
you can't transfer small objects. As a work-around you can use the
`--ignore-checksum` flag.
A proper fix is being worked on in [issue #1824 ](https://github.com/ncw/rclone/issues/1824 ).
2017-09-09 20:02:56 +08:00
### Glacier ###
2018-12-06 22:35:29 +08:00
You can upload objects using the glacier storage class or transition them to glacier using a [lifecycle policy ](http://docs.aws.amazon.com/AmazonS3/latest/user-guide/create-lifecycle.html ).
2017-09-09 20:02:56 +08:00
The bucket can still be synced or copied into normally, but if rclone
2018-12-06 22:35:29 +08:00
tries to access data from the glacier storage class you will see an error like below.
2017-09-09 20:02:56 +08:00
2017/09/11 19:07:43 Failed to sync: failed to open source object: Object in GLACIER, restore first: path/to/file
In this case you need to [restore ](http://docs.aws.amazon.com/AmazonS3/latest/user-guide/restore-archived-objects.html )
the object(s) in question before using rclone.
2018-10-02 03:48:54 +08:00
<!-- - autogenerated options start - DO NOT EDIT, instead edit fs.RegInfo in backend/s3/s3.go then run make backenddocs -->
### Standard Options
Here are the standard options specific to s3 (Amazon S3 Compliant Storage Providers (AWS, Ceph, Dreamhost, IBM COS, Minio)).
#### --s3-provider
Choose your S3 provider.
- Config: provider
- Env Var: RCLONE_S3_PROVIDER
- Type: string
- Default: ""
- Examples:
- "AWS"
- Amazon Web Services (AWS) S3
- "Ceph"
- Ceph Object Storage
- "DigitalOcean"
- Digital Ocean Spaces
- "Dreamhost"
- Dreamhost DreamObjects
- "IBMCOS"
- IBM COS S3
- "Minio"
- Minio Object Storage
- "Wasabi"
- Wasabi Object Storage
- "Other"
- Any other S3 compatible provider
#### --s3-env-auth
Get AWS credentials from runtime (environment variables or EC2/ECS meta data if no env vars).
Only applies if access_key_id and secret_access_key is blank.
- Config: env_auth
- Env Var: RCLONE_S3_ENV_AUTH
- Type: bool
- Default: false
- Examples:
- "false"
- Enter AWS credentials in the next step
- "true"
- Get AWS credentials from the environment (env vars or IAM)
#### --s3-access-key-id
AWS Access Key ID.
Leave blank for anonymous access or runtime credentials.
- Config: access_key_id
- Env Var: RCLONE_S3_ACCESS_KEY_ID
- Type: string
- Default: ""
#### --s3-secret-access-key
AWS Secret Access Key (password)
Leave blank for anonymous access or runtime credentials.
- Config: secret_access_key
- Env Var: RCLONE_S3_SECRET_ACCESS_KEY
- Type: string
- Default: ""
#### --s3-region
Region to connect to.
- Config: region
- Env Var: RCLONE_S3_REGION
- Type: string
- Default: ""
- Examples:
- "us-east-1"
- The default endpoint - a good choice if you are unsure.
- US Region, Northern Virginia or Pacific Northwest.
- Leave location constraint empty.
- "us-east-2"
- US East (Ohio) Region
- Needs location constraint us-east-2.
- "us-west-2"
- US West (Oregon) Region
- Needs location constraint us-west-2.
- "us-west-1"
- US West (Northern California) Region
- Needs location constraint us-west-1.
- "ca-central-1"
- Canada (Central) Region
- Needs location constraint ca-central-1.
- "eu-west-1"
- EU (Ireland) Region
- Needs location constraint EU or eu-west-1.
- "eu-west-2"
- EU (London) Region
- Needs location constraint eu-west-2.
- "eu-central-1"
- EU (Frankfurt) Region
- Needs location constraint eu-central-1.
- "ap-southeast-1"
- Asia Pacific (Singapore) Region
- Needs location constraint ap-southeast-1.
- "ap-southeast-2"
- Asia Pacific (Sydney) Region
- Needs location constraint ap-southeast-2.
- "ap-northeast-1"
- Asia Pacific (Tokyo) Region
- Needs location constraint ap-northeast-1.
- "ap-northeast-2"
- Asia Pacific (Seoul)
- Needs location constraint ap-northeast-2.
- "ap-south-1"
- Asia Pacific (Mumbai)
- Needs location constraint ap-south-1.
- "sa-east-1"
- South America (Sao Paulo) Region
- Needs location constraint sa-east-1.
#### --s3-region
Region to connect to.
Leave blank if you are using an S3 clone and you don't have a region.
- Config: region
- Env Var: RCLONE_S3_REGION
- Type: string
- Default: ""
- Examples:
- ""
- Use this if unsure. Will use v4 signatures and an empty region.
- "other-v2-signature"
- Use this only if v4 signatures don't work, eg pre Jewel/v10 CEPH.
#### --s3-endpoint
Endpoint for S3 API.
Leave blank if using AWS to use the default endpoint for the region.
- Config: endpoint
- Env Var: RCLONE_S3_ENDPOINT
- Type: string
- Default: ""
#### --s3-endpoint
Endpoint for IBM COS S3 API.
Specify if using an IBM COS On Premise.
- Config: endpoint
- Env Var: RCLONE_S3_ENDPOINT
- Type: string
- Default: ""
- Examples:
- "s3-api.us-geo.objectstorage.softlayer.net"
- US Cross Region Endpoint
- "s3-api.dal.us-geo.objectstorage.softlayer.net"
- US Cross Region Dallas Endpoint
- "s3-api.wdc-us-geo.objectstorage.softlayer.net"
- US Cross Region Washington DC Endpoint
- "s3-api.sjc-us-geo.objectstorage.softlayer.net"
- US Cross Region San Jose Endpoint
- "s3-api.us-geo.objectstorage.service.networklayer.com"
- US Cross Region Private Endpoint
- "s3-api.dal-us-geo.objectstorage.service.networklayer.com"
- US Cross Region Dallas Private Endpoint
- "s3-api.wdc-us-geo.objectstorage.service.networklayer.com"
- US Cross Region Washington DC Private Endpoint
- "s3-api.sjc-us-geo.objectstorage.service.networklayer.com"
- US Cross Region San Jose Private Endpoint
- "s3.us-east.objectstorage.softlayer.net"
- US Region East Endpoint
- "s3.us-east.objectstorage.service.networklayer.com"
- US Region East Private Endpoint
- "s3.us-south.objectstorage.softlayer.net"
- US Region South Endpoint
- "s3.us-south.objectstorage.service.networklayer.com"
- US Region South Private Endpoint
- "s3.eu-geo.objectstorage.softlayer.net"
- EU Cross Region Endpoint
- "s3.fra-eu-geo.objectstorage.softlayer.net"
- EU Cross Region Frankfurt Endpoint
- "s3.mil-eu-geo.objectstorage.softlayer.net"
- EU Cross Region Milan Endpoint
- "s3.ams-eu-geo.objectstorage.softlayer.net"
- EU Cross Region Amsterdam Endpoint
- "s3.eu-geo.objectstorage.service.networklayer.com"
- EU Cross Region Private Endpoint
- "s3.fra-eu-geo.objectstorage.service.networklayer.com"
- EU Cross Region Frankfurt Private Endpoint
- "s3.mil-eu-geo.objectstorage.service.networklayer.com"
- EU Cross Region Milan Private Endpoint
- "s3.ams-eu-geo.objectstorage.service.networklayer.com"
- EU Cross Region Amsterdam Private Endpoint
- "s3.eu-gb.objectstorage.softlayer.net"
- Great Britan Endpoint
- "s3.eu-gb.objectstorage.service.networklayer.com"
- Great Britan Private Endpoint
- "s3.ap-geo.objectstorage.softlayer.net"
- APAC Cross Regional Endpoint
- "s3.tok-ap-geo.objectstorage.softlayer.net"
- APAC Cross Regional Tokyo Endpoint
- "s3.hkg-ap-geo.objectstorage.softlayer.net"
- APAC Cross Regional HongKong Endpoint
- "s3.seo-ap-geo.objectstorage.softlayer.net"
- APAC Cross Regional Seoul Endpoint
- "s3.ap-geo.objectstorage.service.networklayer.com"
- APAC Cross Regional Private Endpoint
- "s3.tok-ap-geo.objectstorage.service.networklayer.com"
- APAC Cross Regional Tokyo Private Endpoint
- "s3.hkg-ap-geo.objectstorage.service.networklayer.com"
- APAC Cross Regional HongKong Private Endpoint
- "s3.seo-ap-geo.objectstorage.service.networklayer.com"
- APAC Cross Regional Seoul Private Endpoint
- "s3.mel01.objectstorage.softlayer.net"
- Melbourne Single Site Endpoint
- "s3.mel01.objectstorage.service.networklayer.com"
- Melbourne Single Site Private Endpoint
- "s3.tor01.objectstorage.softlayer.net"
- Toronto Single Site Endpoint
- "s3.tor01.objectstorage.service.networklayer.com"
- Toronto Single Site Private Endpoint
#### --s3-endpoint
Endpoint for S3 API.
Required when using an S3 clone.
- Config: endpoint
- Env Var: RCLONE_S3_ENDPOINT
- Type: string
- Default: ""
- Examples:
- "objects-us-west-1.dream.io"
- Dream Objects endpoint
- "nyc3.digitaloceanspaces.com"
- Digital Ocean Spaces New York 3
- "ams3.digitaloceanspaces.com"
- Digital Ocean Spaces Amsterdam 3
- "sgp1.digitaloceanspaces.com"
- Digital Ocean Spaces Singapore 1
- "s3.wasabisys.com"
2018-11-24 21:44:25 +08:00
- Wasabi US East endpoint
- "s3.us-west-1.wasabisys.com"
- Wasabi US West endpoint
2018-10-02 03:48:54 +08:00
#### --s3-location-constraint
Location constraint - must be set to match the Region.
Used when creating buckets only.
- Config: location_constraint
- Env Var: RCLONE_S3_LOCATION_CONSTRAINT
- Type: string
- Default: ""
- Examples:
- ""
- Empty for US Region, Northern Virginia or Pacific Northwest.
- "us-east-2"
- US East (Ohio) Region.
- "us-west-2"
- US West (Oregon) Region.
- "us-west-1"
- US West (Northern California) Region.
- "ca-central-1"
- Canada (Central) Region.
- "eu-west-1"
- EU (Ireland) Region.
- "eu-west-2"
- EU (London) Region.
- "EU"
- EU Region.
- "ap-southeast-1"
- Asia Pacific (Singapore) Region.
- "ap-southeast-2"
- Asia Pacific (Sydney) Region.
- "ap-northeast-1"
- Asia Pacific (Tokyo) Region.
- "ap-northeast-2"
- Asia Pacific (Seoul)
- "ap-south-1"
- Asia Pacific (Mumbai)
- "sa-east-1"
- South America (Sao Paulo) Region.
#### --s3-location-constraint
Location constraint - must match endpoint when using IBM Cloud Public.
For on-prem COS, do not make a selection from this list, hit enter
- Config: location_constraint
- Env Var: RCLONE_S3_LOCATION_CONSTRAINT
- Type: string
- Default: ""
- Examples:
- "us-standard"
- US Cross Region Standard
- "us-vault"
- US Cross Region Vault
- "us-cold"
- US Cross Region Cold
- "us-flex"
- US Cross Region Flex
- "us-east-standard"
- US East Region Standard
- "us-east-vault"
- US East Region Vault
- "us-east-cold"
- US East Region Cold
- "us-east-flex"
- US East Region Flex
- "us-south-standard"
- US Sout hRegion Standard
- "us-south-vault"
- US South Region Vault
- "us-south-cold"
- US South Region Cold
- "us-south-flex"
- US South Region Flex
- "eu-standard"
- EU Cross Region Standard
- "eu-vault"
- EU Cross Region Vault
- "eu-cold"
- EU Cross Region Cold
- "eu-flex"
- EU Cross Region Flex
- "eu-gb-standard"
- Great Britan Standard
- "eu-gb-vault"
- Great Britan Vault
- "eu-gb-cold"
- Great Britan Cold
- "eu-gb-flex"
- Great Britan Flex
- "ap-standard"
- APAC Standard
- "ap-vault"
- APAC Vault
- "ap-cold"
- APAC Cold
- "ap-flex"
- APAC Flex
- "mel01-standard"
- Melbourne Standard
- "mel01-vault"
- Melbourne Vault
- "mel01-cold"
- Melbourne Cold
- "mel01-flex"
- Melbourne Flex
- "tor01-standard"
- Toronto Standard
- "tor01-vault"
- Toronto Vault
- "tor01-cold"
- Toronto Cold
- "tor01-flex"
- Toronto Flex
#### --s3-location-constraint
Location constraint - must be set to match the Region.
Leave blank if not sure. Used when creating buckets only.
- Config: location_constraint
- Env Var: RCLONE_S3_LOCATION_CONSTRAINT
- Type: string
- Default: ""
#### --s3-acl
2018-11-24 21:44:25 +08:00
Canned ACL used when creating buckets and storing or copying objects.
2018-10-02 03:48:54 +08:00
For more info visit https://docs.aws.amazon.com/AmazonS3/latest/dev/acl-overview.html#canned-acl
2018-11-24 21:44:25 +08:00
Note that this ACL is applied when server side copying objects as S3
doesn't copy the ACL from the source but rather writes a fresh one.
2018-10-02 03:48:54 +08:00
- Config: acl
- Env Var: RCLONE_S3_ACL
- Type: string
- Default: ""
- Examples:
- "private"
- Owner gets FULL_CONTROL. No one else has access rights (default).
- "public-read"
- Owner gets FULL_CONTROL. The AllUsers group gets READ access.
- "public-read-write"
- Owner gets FULL_CONTROL. The AllUsers group gets READ and WRITE access.
- Granting this on a bucket is generally not recommended.
- "authenticated-read"
- Owner gets FULL_CONTROL. The AuthenticatedUsers group gets READ access.
- "bucket-owner-read"
- Object owner gets FULL_CONTROL. Bucket owner gets READ access.
- If you specify this canned ACL when creating a bucket, Amazon S3 ignores it.
- "bucket-owner-full-control"
- Both the object owner and the bucket owner get FULL_CONTROL over the object.
- If you specify this canned ACL when creating a bucket, Amazon S3 ignores it.
- "private"
- Owner gets FULL_CONTROL. No one else has access rights (default). This acl is available on IBM Cloud (Infra), IBM Cloud (Storage), On-Premise COS
- "public-read"
- Owner gets FULL_CONTROL. The AllUsers group gets READ access. This acl is available on IBM Cloud (Infra), IBM Cloud (Storage), On-Premise IBM COS
- "public-read-write"
- Owner gets FULL_CONTROL. The AllUsers group gets READ and WRITE access. This acl is available on IBM Cloud (Infra), On-Premise IBM COS
- "authenticated-read"
- Owner gets FULL_CONTROL. The AuthenticatedUsers group gets READ access. Not supported on Buckets. This acl is available on IBM Cloud (Infra) and On-Premise IBM COS
#### --s3-server-side-encryption
The server-side encryption algorithm used when storing this object in S3.
- Config: server_side_encryption
- Env Var: RCLONE_S3_SERVER_SIDE_ENCRYPTION
- Type: string
- Default: ""
- Examples:
- ""
- None
- "AES256"
- AES256
- "aws:kms"
- aws:kms
#### --s3-sse-kms-key-id
If using KMS ID you must provide the ARN of Key.
- Config: sse_kms_key_id
- Env Var: RCLONE_S3_SSE_KMS_KEY_ID
- Type: string
- Default: ""
- Examples:
- ""
- None
- "arn:aws:kms:us-east-1:*"
- arn:aws:kms:*
#### --s3-storage-class
The storage class to use when storing new objects in S3.
- Config: storage_class
- Env Var: RCLONE_S3_STORAGE_CLASS
- Type: string
- Default: ""
- Examples:
- ""
- Default
- "STANDARD"
- Standard storage class
- "REDUCED_REDUNDANCY"
- Reduced redundancy storage class
- "STANDARD_IA"
- Standard Infrequent Access storage class
- "ONEZONE_IA"
- One Zone Infrequent Access storage class
2018-12-06 22:35:29 +08:00
- "GLACIER"
- GLACIER storage class
2018-10-02 03:48:54 +08:00
### Advanced Options
Here are the advanced options specific to s3 (Amazon S3 Compliant Storage Providers (AWS, Ceph, Dreamhost, IBM COS, Minio)).
2018-11-27 05:09:23 +08:00
#### --s3-upload-cutoff
Cutoff for switching to chunked upload
Any files larger than this will be uploaded in chunks of chunk_size.
The minimum is 0 and the maximum is 5GB.
- Config: upload_cutoff
- Env Var: RCLONE_S3_UPLOAD_CUTOFF
- Type: SizeSuffix
- Default: 200M
2018-10-02 03:48:54 +08:00
#### --s3-chunk-size
Chunk size to use for uploading.
2018-11-27 05:09:23 +08:00
When uploading files larger than upload_cutoff they will be uploaded
as multipart uploads using this chunk size.
2018-10-02 03:48:54 +08:00
Note that "--s3-upload-concurrency" chunks of this size are buffered
in memory per transfer.
If you are transferring large files over high speed links and you have
enough memory, then increasing this will speed up the transfers.
- Config: chunk_size
- Env Var: RCLONE_S3_CHUNK_SIZE
- Type: SizeSuffix
- Default: 5M
#### --s3-disable-checksum
Don't store MD5 checksum with object metadata
- Config: disable_checksum
- Env Var: RCLONE_S3_DISABLE_CHECKSUM
- Type: bool
- Default: false
#### --s3-session-token
An AWS session token
- Config: session_token
- Env Var: RCLONE_S3_SESSION_TOKEN
- Type: string
- Default: ""
#### --s3-upload-concurrency
Concurrency for multipart uploads.
This is the number of chunks of the same file that are uploaded
concurrently.
If you are uploading small numbers of large file over high speed link
and these uploads do not fully utilize your bandwidth, then increasing
this may help to speed up the transfers.
- Config: upload_concurrency
- Env Var: RCLONE_S3_UPLOAD_CONCURRENCY
- Type: int
2018-12-03 01:51:14 +08:00
- Default: 4
2018-10-02 03:48:54 +08:00
#### --s3-force-path-style
If true use path style access if false use virtual hosted style.
If this is true (the default) then rclone will use path style access,
if false then rclone will use virtual path style. See [the AWS S3
docs](https://docs.aws.amazon.com/AmazonS3/latest/dev/UsingBucket.html#access-bucket-intro)
for more info.
Some providers (eg Aliyun OSS or Netease COS) require this set to false.
- Config: force_path_style
- Env Var: RCLONE_S3_FORCE_PATH_STYLE
- Type: bool
- Default: true
2018-10-09 20:03:37 +08:00
#### --s3-v2-auth
If true use v2 authentication.
If this is false (the default) then rclone will use v4 authentication.
If it is set then rclone will use v2 authentication.
Use this only if v4 signatures don't work, eg pre Jewel/v10 CEPH.
- Config: v2_auth
- Env Var: RCLONE_S3_V2_AUTH
- Type: bool
- Default: false
2018-10-02 03:48:54 +08:00
<!-- - autogenerated options stop -->
2018-06-13 18:02:56 +08:00
2015-09-29 16:58:03 +08:00
### Anonymous access to public buckets ###
If you want to use rclone to access a public bucket, configure with a
2018-04-13 23:08:00 +08:00
blank `access_key_id` and `secret_access_key` . Your config should end
up looking like this:
2015-09-29 16:58:03 +08:00
```
2018-04-13 23:08:00 +08:00
[anons3]
type = s3
provider = AWS
env_auth = false
access_key_id =
secret_access_key =
region = us-east-1
endpoint =
location_constraint =
acl = private
server_side_encryption =
storage_class =
2015-09-29 16:58:03 +08:00
```
Then use it as normal with the name of the public bucket, eg
rclone lsd anons3:1000genomes
You will be able to list and copy data but not upload it.
2015-08-16 01:44:45 +08:00
### Ceph ###
2018-03-14 04:47:29 +08:00
[Ceph ](https://ceph.com/ ) is an open source unified, distributed
storage system designed for excellent performance, reliability and
scalability. It has an S3 compatible object storage interface.
To use rclone with Ceph, configure as above but leave the region blank
and set the endpoint. You should end up with something like this in
your config:
2015-08-16 01:44:45 +08:00
```
2018-03-14 04:47:29 +08:00
[ceph]
type = s3
2018-04-13 23:08:00 +08:00
provider = Ceph
2018-03-14 04:47:29 +08:00
env_auth = false
access_key_id = XXX
secret_access_key = YYY
2018-04-13 00:05:53 +08:00
region =
2018-03-14 04:47:29 +08:00
endpoint = https://ceph.endpoint.example.com
2018-04-13 00:05:53 +08:00
location_constraint =
acl =
server_side_encryption =
storage_class =
2015-08-16 01:44:45 +08:00
```
Note also that Ceph sometimes puts `/` in the passwords it gives
users. If you read the secret access key using the command line tools
you will get a JSON blob with the `/` escaped as `\/` . Make sure you
only write `/` in the secret access key.
Eg the dump from Ceph looks something like this (irrelevant keys
removed).
```
{
"user_id": "xxx",
"display_name": "xxxx",
"keys": [
{
"user": "xxx",
"access_key": "xxxxxx",
"secret_key": "xxxxxx\/xxxx"
}
],
}
```
Because this is a json dump, it is encoding the `/` as `\/` , so if you
use the secret key as `xxxxxx/xxxx` it will work fine.
2016-07-11 19:12:28 +08:00
2018-03-14 04:47:29 +08:00
### Dreamhost ###
Dreamhost [DreamObjects ](https://www.dreamhost.com/cloud/storage/ ) is
an object storage system based on CEPH.
To use rclone with Dreamhost, configure as above but leave the region blank
and set the endpoint. You should end up with something like this in
your config:
```
[dreamobjects]
2018-04-13 23:08:00 +08:00
type = s3
provider = DreamHost
2018-03-14 04:47:29 +08:00
env_auth = false
access_key_id = your_access_key
secret_access_key = your_secret_key
region =
endpoint = objects-us-west-1.dream.io
location_constraint =
acl = private
server_side_encryption =
storage_class =
```
2017-10-20 00:31:25 +08:00
### DigitalOcean Spaces ###
[Spaces ](https://www.digitalocean.com/products/object-storage/ ) is an [S3-interoperable ](https://developers.digitalocean.com/documentation/spaces/ ) object storage service from cloud provider DigitalOcean.
To connect to DigitalOcean Spaces you will need an access key and secret key. These can be retrieved on the "[Applications & API](https://cloud.digitalocean.com/settings/api/tokens)" page of the DigitalOcean control panel. They will be needed when promted by `rclone config` for your `access_key_id` and `secret_access_key` .
2017-10-28 14:03:51 +08:00
When prompted for a `region` or `location_constraint` , press enter to use the default value. The region must be included in the `endpoint` setting (e.g. `nyc3.digitaloceanspaces.com` ). The defualt values can be used for other settings.
2017-10-20 00:31:25 +08:00
Going through the whole process of creating a new remote by running `rclone config` , each prompt should be answered as shown below:
```
2018-03-14 04:47:29 +08:00
Storage> s3
2017-10-20 00:31:25 +08:00
env_auth> 1
access_key_id> YOUR_ACCESS_KEY
secret_access_key> YOUR_SECRET_KEY
2018-04-13 00:05:53 +08:00
region>
2017-10-20 00:31:25 +08:00
endpoint> nyc3.digitaloceanspaces.com
2018-04-13 00:05:53 +08:00
location_constraint>
acl>
storage_class>
2017-10-20 00:31:25 +08:00
```
The resulting configuration file should look like:
```
[spaces]
type = s3
2018-04-13 23:08:00 +08:00
provider = DigitalOcean
2017-10-20 00:31:25 +08:00
env_auth = false
access_key_id = YOUR_ACCESS_KEY
secret_access_key = YOUR_SECRET_KEY
2018-04-13 00:05:53 +08:00
region =
2017-10-20 00:31:25 +08:00
endpoint = nyc3.digitaloceanspaces.com
2018-04-13 00:05:53 +08:00
location_constraint =
acl =
server_side_encryption =
storage_class =
2017-10-20 00:31:25 +08:00
```
Once configured, you can create a new Space and begin copying files. For example:
```
rclone mkdir spaces:my-new-space
rclone copy /path/to/files spaces:my-new-space
```
2018-03-15 22:11:32 +08:00
### IBM COS (S3) ###
2018-04-13 23:08:00 +08:00
2018-03-27 03:49:53 +08:00
Information stored with IBM Cloud Object Storage is encrypted and dispersed across multiple geographic locations, and accessed through an implementation of the S3 API. This service makes use of the distributed storage technologies provided by IBM’ s Cloud Object Storage System (formerly Cleversafe). For more information visit: (http://www.ibm.com/cloud/object-storage)
2018-03-15 22:11:32 +08:00
To configure access to IBM COS S3, follow the steps below:
1. Run rclone config and select n for a new remote.
```
2018/02/14 14:13:11 NOTICE: Config file "C:\\Users\\a\\.config\\rclone\\rclone.conf" not found - using defaults
No remotes found - make a new one
n) New remote
s) Set configuration password
q) Quit config
n/s/q> n
```
2. Enter the name for the configuration
```
2018-04-13 00:05:53 +08:00
name> < YOUR NAME >
2018-03-15 22:11:32 +08:00
```
3. Select "s3" storage.
```
2018-04-13 00:05:53 +08:00
Choose a number from below, or type in your own value
1 / Alias for a existing remote
\ "alias"
2 / Amazon Drive
2018-03-15 22:11:32 +08:00
\ "amazon cloud drive"
2018-04-13 00:05:53 +08:00
3 / Amazon S3 Complaint Storage Providers (Dreamhost, Ceph, Minio, IBM COS)
2018-03-15 22:11:32 +08:00
\ "s3"
2018-04-13 00:05:53 +08:00
4 / Backblaze B2
\ "b2"
[snip]
23 / http Connection
\ "http"
Storage> 3
2018-03-15 22:11:32 +08:00
```
2018-04-13 00:05:53 +08:00
4. Select IBM COS as the S3 Storage Provider.
2018-03-15 22:11:32 +08:00
```
2018-04-13 00:05:53 +08:00
Choose the S3 provider.
Choose a number from below, or type in your own value
1 / Choose this option to configure Storage to AWS S3
\ "AWS"
2 / Choose this option to configure Storage to Ceph Systems
\ "Ceph"
3 / Choose this option to configure Storage to Dreamhost
\ "Dreamhost"
4 / Choose this option to the configure Storage to IBM COS S3
\ "IBMCOS"
5 / Choose this option to the configure Storage to Minio
\ "Minio"
Provider>4
2018-03-15 22:11:32 +08:00
```
5. Enter the Access Key and Secret.
```
AWS Access Key ID - leave blank for anonymous access or runtime credentials.
access_key_id> < >
AWS Secret Access Key (password) - leave blank for anonymous access or runtime credentials.
secret_access_key> < >
```
2018-04-13 00:05:53 +08:00
6. Specify the endpoint for IBM COS. For Public IBM COS, choose from the option below. For On Premise IBM COS, enter an enpoint address.
2018-03-15 22:11:32 +08:00
```
2018-04-13 00:05:53 +08:00
Endpoint for IBM COS S3 API.
Specify if using an IBM COS On Premise.
2018-03-15 22:11:32 +08:00
Choose a number from below, or type in your own value
2018-04-13 00:05:53 +08:00
1 / US Cross Region Endpoint
\ "s3-api.us-geo.objectstorage.softlayer.net"
2 / US Cross Region Dallas Endpoint
\ "s3-api.dal.us-geo.objectstorage.softlayer.net"
3 / US Cross Region Washington DC Endpoint
\ "s3-api.wdc-us-geo.objectstorage.softlayer.net"
4 / US Cross Region San Jose Endpoint
\ "s3-api.sjc-us-geo.objectstorage.softlayer.net"
5 / US Cross Region Private Endpoint
\ "s3-api.us-geo.objectstorage.service.networklayer.com"
6 / US Cross Region Dallas Private Endpoint
\ "s3-api.dal-us-geo.objectstorage.service.networklayer.com"
7 / US Cross Region Washington DC Private Endpoint
\ "s3-api.wdc-us-geo.objectstorage.service.networklayer.com"
8 / US Cross Region San Jose Private Endpoint
\ "s3-api.sjc-us-geo.objectstorage.service.networklayer.com"
9 / US Region East Endpoint
\ "s3.us-east.objectstorage.softlayer.net"
10 / US Region East Private Endpoint
\ "s3.us-east.objectstorage.service.networklayer.com"
11 / US Region South Endpoint
[snip]
34 / Toronto Single Site Private Endpoint
\ "s3.tor01.objectstorage.service.networklayer.com"
endpoint>1
```
7. Specify a IBM COS Location Constraint. The location constraint must match endpoint when using IBM Cloud Public. For on-prem COS, do not make a selection from this list, hit enter
```
1 / US Cross Region Standard
\ "us-standard"
2 / US Cross Region Vault
\ "us-vault"
3 / US Cross Region Cold
\ "us-cold"
4 / US Cross Region Flex
\ "us-flex"
5 / US East Region Standard
\ "us-east-standard"
6 / US East Region Vault
\ "us-east-vault"
7 / US East Region Cold
\ "us-east-cold"
8 / US East Region Flex
\ "us-east-flex"
9 / US South Region Standard
\ "us-south-standard"
10 / US South Region Vault
\ "us-south-vault"
[snip]
32 / Toronto Flex
\ "tor01-flex"
location_constraint>1
2018-03-15 22:11:32 +08:00
```
2018-04-13 00:05:53 +08:00
9. Specify a canned ACL. IBM Cloud (Strorage) supports "public-read" and "private". IBM Cloud(Infra) supports all the canned ACLs. On-Premise COS supports all the canned ACLs.
2018-03-15 22:11:32 +08:00
```
2018-04-13 00:05:53 +08:00
Canned ACL used when creating buckets and/or storing objects in S3.
For more info visit https://docs.aws.amazon.com/AmazonS3/latest/dev/acl-overview.html#canned-acl
Choose a number from below, or type in your own value
1 / Owner gets FULL_CONTROL. No one else has access rights (default). This acl is available on IBM Cloud (Infra), IBM Cloud (Storage), On-Premise COS
\ "private"
2 / Owner gets FULL_CONTROL. The AllUsers group gets READ access. This acl is available on IBM Cloud (Infra), IBM Cloud (Storage), On-Premise IBM COS
\ "public-read"
3 / Owner gets FULL_CONTROL. The AllUsers group gets READ and WRITE access. This acl is available on IBM Cloud (Infra), On-Premise IBM COS
\ "public-read-write"
4 / Owner gets FULL_CONTROL. The AuthenticatedUsers group gets READ access. Not supported on Buckets. This acl is available on IBM Cloud (Infra) and On-Premise IBM COS
\ "authenticated-read"
acl> 1
2018-03-15 22:11:32 +08:00
```
2018-03-27 03:49:53 +08:00
12. Review the displayed configuration and accept to save the "remote" then quit. The config file should look like this
2018-03-15 22:11:32 +08:00
```
2018-04-13 00:05:53 +08:00
[xxx]
type = s3
Provider = IBMCOS
access_key_id = xxx
secret_access_key = yyy
2018-03-15 22:11:32 +08:00
endpoint = s3-api.us-geo.objectstorage.softlayer.net
location_constraint = us-standard
acl = private
```
13. Execute rclone commands
```
1) Create a bucket.
rclone mkdir IBM-COS-XREGION:newbucket
2) List available buckets.
rclone lsd IBM-COS-XREGION:
-1 2017-11-08 21:16:22 -1 test
-1 2018-02-14 20:16:39 -1 newbucket
3) List contents of a bucket.
rclone ls IBM-COS-XREGION:newbucket
18685952 test.exe
4) Copy a file from local to remote.
rclone copy /Users/file.txt IBM-COS-XREGION:newbucket
5) Copy a file from remote to local.
rclone copy IBM-COS-XREGION:newbucket/file.txt .
6) Delete a file on remote.
rclone delete IBM-COS-XREGION:newbucket/file.txt
```
2016-07-11 19:12:28 +08:00
### Minio ###
[Minio ](https://minio.io/ ) is an object storage server built for cloud application developers and devops.
It is very easy to install and provides an S3 compatible server which can be used by rclone.
2017-06-20 06:51:39 +08:00
To use it, install Minio following the instructions [here ](https://docs.minio.io/docs/minio-quickstart-guide ).
2016-07-11 19:12:28 +08:00
When it configures itself Minio will print something like this
```
2017-06-20 06:51:39 +08:00
Endpoint: http://192.168.1.106:9000 http://172.23.0.1:9000
AccessKey: USWUXHGYZQYFYFFIT3RE
SecretKey: MOJRH0mkL1IPauahWITSVvyDrQbEEIwljvmxdq03
Region: us-east-1
SQS ARNs: arn:minio:sqs:us-east-1:1:redis arn:minio:sqs:us-east-1:2:redis
Browser Access:
http://192.168.1.106:9000 http://172.23.0.1:9000
Command-line Access: https://docs.minio.io/docs/minio-client-quickstart-guide
$ mc config host add myminio http://192.168.1.106:9000 USWUXHGYZQYFYFFIT3RE MOJRH0mkL1IPauahWITSVvyDrQbEEIwljvmxdq03
Object API (Amazon S3 compatible):
Go: https://docs.minio.io/docs/golang-client-quickstart-guide
Java: https://docs.minio.io/docs/java-client-quickstart-guide
Python: https://docs.minio.io/docs/python-client-quickstart-guide
JavaScript: https://docs.minio.io/docs/javascript-client-quickstart-guide
.NET: https://docs.minio.io/docs/dotnet-client-quickstart-guide
Drive Capacity: 26 GiB Free, 165 GiB Total
2016-07-11 19:12:28 +08:00
```
These details need to go into `rclone config` like this. Note that it
is important to put the region in as stated above.
```
env_auth> 1
2017-06-20 06:51:39 +08:00
access_key_id> USWUXHGYZQYFYFFIT3RE
secret_access_key> MOJRH0mkL1IPauahWITSVvyDrQbEEIwljvmxdq03
2016-07-11 19:12:28 +08:00
region> us-east-1
2017-06-20 06:51:39 +08:00
endpoint> http://192.168.1.106:9000
2017-01-09 13:09:19 +08:00
location_constraint>
2016-07-11 19:12:28 +08:00
server_side_encryption>
```
Which makes the config file look like this
```
[minio]
2018-04-13 23:08:00 +08:00
type = s3
provider = Minio
2016-07-11 19:12:28 +08:00
env_auth = false
2017-06-20 06:51:39 +08:00
access_key_id = USWUXHGYZQYFYFFIT3RE
secret_access_key = MOJRH0mkL1IPauahWITSVvyDrQbEEIwljvmxdq03
2016-07-11 19:12:28 +08:00
region = us-east-1
2017-06-20 06:51:39 +08:00
endpoint = http://192.168.1.106:9000
2017-01-09 13:09:19 +08:00
location_constraint =
server_side_encryption =
2016-07-11 19:12:28 +08:00
```
So once set up, for example to copy files into a bucket
2017-06-20 06:51:39 +08:00
```
rclone copy /path/to/files minio:bucket
2017-09-12 05:49:59 +08:00
```
2017-08-30 22:55:51 +08:00
### Wasabi ###
2017-09-26 00:55:19 +08:00
[Wasabi ](https://wasabi.com ) is a cloud-based object storage service for a
2017-08-30 22:55:51 +08:00
broad range of applications and use cases. Wasabi is designed for
individuals and organizations that require a high-performance,
reliable, and secure data storage infrastructure at minimal cost.
Wasabi provides an S3 interface which can be configured for use with
rclone like this.
```
No remotes found - make a new one
n) New remote
s) Set configuration password
n/s> n
name> wasabi
Type of storage to configure.
Choose a number from below, or type in your own value
1 / Amazon Drive
\ "amazon cloud drive"
2 / Amazon S3 (also Dreamhost, Ceph, Minio)
\ "s3"
[snip]
Storage> s3
2017-11-23 05:21:36 +08:00
Get AWS credentials from runtime (environment variables or EC2/ECS meta data if no env vars). Only applies if access_key_id and secret_access_key is blank.
2017-08-30 22:55:51 +08:00
Choose a number from below, or type in your own value
1 / Enter AWS credentials in the next step
\ "false"
2 / Get AWS credentials from the environment (env vars or IAM)
\ "true"
env_auth> 1
AWS Access Key ID - leave blank for anonymous access or runtime credentials.
access_key_id> YOURACCESSKEY
AWS Secret Access Key (password) - leave blank for anonymous access or runtime credentials.
secret_access_key> YOURSECRETACCESSKEY
Region to connect to.
Choose a number from below, or type in your own value
/ The default endpoint - a good choice if you are unsure.
1 | US Region, Northern Virginia or Pacific Northwest.
| Leave location constraint empty.
\ "us-east-1"
[snip]
region> us-east-1
Endpoint for S3 API.
Leave blank if using AWS to use the default endpoint for the region.
Specify if using an S3 clone such as Ceph.
endpoint> s3.wasabisys.com
Location constraint - must be set to match the Region. Used when creating buckets only.
Choose a number from below, or type in your own value
1 / Empty for US Region, Northern Virginia or Pacific Northwest.
\ ""
[snip]
2018-04-13 00:05:53 +08:00
location_constraint>
2017-08-30 22:55:51 +08:00
Canned ACL used when creating buckets and/or storing objects in S3.
For more info visit https://docs.aws.amazon.com/AmazonS3/latest/dev/acl-overview.html#canned-acl
Choose a number from below, or type in your own value
1 / Owner gets FULL_CONTROL. No one else has access rights (default).
\ "private"
[snip]
2018-04-13 00:05:53 +08:00
acl>
2017-08-30 22:55:51 +08:00
The server-side encryption algorithm used when storing this object in S3.
Choose a number from below, or type in your own value
1 / None
\ ""
2 / AES256
\ "AES256"
2018-04-13 00:05:53 +08:00
server_side_encryption>
2017-08-30 22:55:51 +08:00
The storage class to use when storing objects in S3.
Choose a number from below, or type in your own value
1 / Default
\ ""
2 / Standard storage class
\ "STANDARD"
3 / Reduced redundancy storage class
\ "REDUCED_REDUNDANCY"
4 / Standard Infrequent Access storage class
\ "STANDARD_IA"
2018-04-13 00:05:53 +08:00
storage_class>
2017-08-30 22:55:51 +08:00
Remote config
--------------------
[wasabi]
env_auth = false
access_key_id = YOURACCESSKEY
secret_access_key = YOURSECRETACCESSKEY
region = us-east-1
endpoint = s3.wasabisys.com
2018-04-13 00:05:53 +08:00
location_constraint =
acl =
server_side_encryption =
storage_class =
2017-08-30 22:55:51 +08:00
--------------------
y) Yes this is OK
e) Edit this remote
d) Delete this remote
y/e/d> y
```
This will leave the config file looking like this.
```
[wasabi]
2018-04-13 23:08:00 +08:00
type = s3
provider = Wasabi
2017-08-30 22:55:51 +08:00
env_auth = false
access_key_id = YOURACCESSKEY
secret_access_key = YOURSECRETACCESSKEY
2018-04-13 23:08:00 +08:00
region =
2017-08-30 22:55:51 +08:00
endpoint = s3.wasabisys.com
2018-04-13 00:05:53 +08:00
location_constraint =
acl =
server_side_encryption =
storage_class =
2017-08-30 22:55:51 +08:00
```
2018-07-20 22:49:07 +08:00
### Aliyun OSS / Netease NOS ###
This describes how to set up Aliyun OSS - Netease NOS is the same
except for different endpoints.
Note this is a pretty standard S3 setup, except for the setting of
`force_path_style = false` in the advanced config.
```
# rclone config
e/n/d/r/c/s/q> n
name> oss
Type of storage to configure.
Enter a string value. Press Enter for the default ("").
Choose a number from below, or type in your own value
3 / Amazon S3 Compliant Storage Providers (AWS, Ceph, Dreamhost, IBM COS, Minio)
\ "s3"
Storage> s3
Choose your S3 provider.
Enter a string value. Press Enter for the default ("").
Choose a number from below, or type in your own value
8 / Any other S3 compatible provider
\ "Other"
provider> other
Get AWS credentials from runtime (environment variables or EC2/ECS meta data if no env vars).
Only applies if access_key_id and secret_access_key is blank.
Enter a boolean value (true or false). Press Enter for the default ("false").
Choose a number from below, or type in your own value
1 / Enter AWS credentials in the next step
\ "false"
2 / Get AWS credentials from the environment (env vars or IAM)
\ "true"
env_auth> 1
AWS Access Key ID.
Leave blank for anonymous access or runtime credentials.
Enter a string value. Press Enter for the default ("").
access_key_id> xxxxxxxxxxxx
AWS Secret Access Key (password)
Leave blank for anonymous access or runtime credentials.
Enter a string value. Press Enter for the default ("").
secret_access_key> xxxxxxxxxxxxxxxxx
Region to connect to.
Leave blank if you are using an S3 clone and you don't have a region.
Enter a string value. Press Enter for the default ("").
Choose a number from below, or type in your own value
1 / Use this if unsure. Will use v4 signatures and an empty region.
\ ""
2 / Use this only if v4 signatures don't work, eg pre Jewel/v10 CEPH.
\ "other-v2-signature"
region> 1
Endpoint for S3 API.
Required when using an S3 clone.
Enter a string value. Press Enter for the default ("").
Choose a number from below, or type in your own value
endpoint> oss-cn-shenzhen.aliyuncs.com
Location constraint - must be set to match the Region.
Leave blank if not sure. Used when creating buckets only.
Enter a string value. Press Enter for the default ("").
location_constraint>
Canned ACL used when creating buckets and/or storing objects in S3.
For more info visit https://docs.aws.amazon.com/AmazonS3/latest/dev/acl-overview.html#canned-acl
Enter a string value. Press Enter for the default ("").
Choose a number from below, or type in your own value
1 / Owner gets FULL_CONTROL. No one else has access rights (default).
\ "private"
acl> 1
Edit advanced config? (y/n)
y) Yes
n) No
y/n> y
Chunk size to use for uploading
Enter a size with suffix k,M,G,T. Press Enter for the default ("5M").
chunk_size>
Don't store MD5 checksum with object metadata
Enter a boolean value (true or false). Press Enter for the default ("false").
disable_checksum>
An AWS session token
Enter a string value. Press Enter for the default ("").
session_token>
Concurrency for multipart uploads.
Enter a signed integer. Press Enter for the default ("2").
upload_concurrency>
If true use path style access if false use virtual hosted style.
Some providers (eg Aliyun OSS or Netease COS) require this.
Enter a boolean value (true or false). Press Enter for the default ("true").
force_path_style> false
Remote config
--------------------
[oss]
type = s3
provider = Other
env_auth = false
access_key_id = xxxxxxxxx
secret_access_key = xxxxxxxxxxxxx
endpoint = oss-cn-shenzhen.aliyuncs.com
acl = private
force_path_style = false
--------------------
y) Yes this is OK
e) Edit this remote
d) Delete this remote
y/e/d> y
```