2013-01-09 02:53:35 +08:00
|
|
|
// S3 interface
|
2013-06-28 03:13:07 +08:00
|
|
|
package s3
|
2013-01-09 02:53:35 +08:00
|
|
|
|
2013-01-24 06:43:20 +08:00
|
|
|
// FIXME need to prevent anything but ListDir working for s3://
|
|
|
|
|
2013-01-09 02:53:35 +08:00
|
|
|
import (
|
|
|
|
"errors"
|
|
|
|
"fmt"
|
|
|
|
"io"
|
|
|
|
"mime"
|
|
|
|
"net/http"
|
|
|
|
"path"
|
|
|
|
"regexp"
|
|
|
|
"strconv"
|
|
|
|
"strings"
|
|
|
|
"time"
|
|
|
|
|
2014-03-16 00:06:11 +08:00
|
|
|
"github.com/ncw/goamz/aws"
|
|
|
|
"github.com/ncw/goamz/s3"
|
|
|
|
"github.com/ncw/rclone/fs"
|
|
|
|
"github.com/ncw/swift"
|
|
|
|
)
|
2013-06-28 03:13:07 +08:00
|
|
|
|
|
|
|
// Register with Fs
|
|
|
|
func init() {
|
2014-03-16 00:06:11 +08:00
|
|
|
fs.Register(&fs.FsInfo{
|
|
|
|
Name: "s3",
|
|
|
|
NewFs: NewFs,
|
|
|
|
// AWS endpoints: http://docs.amazonwebservices.com/general/latest/gr/rande.html#s3_region
|
|
|
|
Options: []fs.Option{{
|
|
|
|
Name: "access_key_id",
|
|
|
|
Help: "AWS Access Key ID.",
|
|
|
|
}, {
|
|
|
|
Name: "secret_access_key",
|
|
|
|
Help: "AWS Secret Access Key (password). ",
|
|
|
|
}, {
|
|
|
|
Name: "endpoint",
|
|
|
|
Help: "Endpoint for S3 API.",
|
|
|
|
Examples: []fs.OptionExample{{
|
|
|
|
Value: "https://s3.amazonaws.com/",
|
|
|
|
Help: "The default endpoint - a good choice if you are unsure.\nUS Region, Northern Virginia or Pacific Northwest.\nLeave location constraint empty.",
|
|
|
|
}, {
|
|
|
|
Value: "https://s3-external-1.amazonaws.com",
|
|
|
|
Help: "US Region, Northern Virginia only.\nLeave location constraint empty.",
|
|
|
|
}, {
|
|
|
|
Value: "https://s3-us-west-2.amazonaws.com",
|
|
|
|
Help: "US West (Oregon) Region\nNeeds location constraint us-west-2.",
|
|
|
|
}, {
|
|
|
|
Value: "https://s3-us-west-1.amazonaws.com",
|
|
|
|
Help: "US West (Northern California) Region\nNeeds location constraint us-west-1.",
|
|
|
|
}, {
|
|
|
|
Value: "https://s3-eu-west-1.amazonaws.com",
|
|
|
|
Help: "EU (Ireland) Region Region\nNeeds location constraint EU or eu-west-1.",
|
|
|
|
}, {
|
|
|
|
Value: "https://s3-ap-southeast-1.amazonaws.com",
|
|
|
|
Help: "Asia Pacific (Singapore) Region\nNeeds location constraint ap-southeast-1.",
|
|
|
|
}, {
|
|
|
|
Value: "https://s3-ap-southeast-2.amazonaws.com",
|
|
|
|
Help: "Asia Pacific (Sydney) Region\nNeeds location constraint .",
|
|
|
|
}, {
|
|
|
|
Value: "https://s3-ap-northeast-1.amazonaws.com",
|
|
|
|
Help: "Asia Pacific (Tokyo) Region\nNeeds location constraint ap-northeast-1.",
|
|
|
|
}, {
|
|
|
|
Value: "https://s3-sa-east-1.amazonaws.com",
|
|
|
|
Help: "South America (Sao Paulo) Region\nNeeds location constraint sa-east-1.",
|
|
|
|
}},
|
|
|
|
}, {
|
|
|
|
Name: "location_constraint",
|
|
|
|
Help: "Location constraint - must be set to match the Endpoint.",
|
|
|
|
Examples: []fs.OptionExample{{
|
|
|
|
Value: "",
|
|
|
|
Help: "Empty for US Region, Northern Virginia or Pacific Northwest.",
|
|
|
|
}, {
|
|
|
|
Value: "us-west-2",
|
|
|
|
Help: "US West (Oregon) Region.",
|
|
|
|
}, {
|
|
|
|
Value: "us-west-1",
|
|
|
|
Help: "US West (Northern California) Region.",
|
|
|
|
}, {
|
|
|
|
Value: "eu-west-1",
|
|
|
|
Help: "EU (Ireland) Region.",
|
|
|
|
}, {
|
|
|
|
Value: "EU",
|
|
|
|
Help: "EU Region.",
|
|
|
|
}, {
|
|
|
|
Value: "ap-southeast-1",
|
|
|
|
Help: "Asia Pacific (Singapore) Region.",
|
|
|
|
}, {
|
|
|
|
Value: "ap-southeast-2",
|
|
|
|
Help: "Asia Pacific (Sydney) Region.",
|
|
|
|
}, {
|
|
|
|
Value: "ap-northeast-1",
|
|
|
|
Help: "Asia Pacific (Tokyo) Region.",
|
|
|
|
}, {
|
|
|
|
Value: "sa-east-1",
|
|
|
|
Help: "South America (Sao Paulo) Region.",
|
|
|
|
}},
|
|
|
|
}},
|
|
|
|
})
|
2013-06-28 03:13:07 +08:00
|
|
|
}
|
|
|
|
|
2013-01-09 02:53:35 +08:00
|
|
|
// Constants
|
|
|
|
const (
|
|
|
|
metaMtime = "X-Amz-Meta-Mtime" // the meta key to store mtime in
|
|
|
|
)
|
|
|
|
|
|
|
|
// FsS3 represents a remote s3 server
|
|
|
|
type FsS3 struct {
|
|
|
|
c *s3.S3 // the connection to the s3 server
|
|
|
|
b *s3.Bucket // the connection to the bucket
|
|
|
|
bucket string // the bucket we are working on
|
|
|
|
perm s3.ACL // permissions for new buckets / objects
|
2014-05-06 01:25:32 +08:00
|
|
|
root string // root of the bucket - ignore all objects above this
|
2013-01-09 02:53:35 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
// FsObjectS3 describes a s3 object
|
|
|
|
type FsObjectS3 struct {
|
|
|
|
// Will definitely have everything but meta which may be nil
|
|
|
|
//
|
|
|
|
// List will read everything but meta - to fill that in need to call
|
|
|
|
// readMetaData
|
|
|
|
s3 *FsS3 // what this object is part of
|
|
|
|
remote string // The remote path
|
|
|
|
etag string // md5sum of the object
|
|
|
|
bytes int64 // size of the object
|
|
|
|
lastModified time.Time // Last modified
|
|
|
|
meta s3.Headers // The object metadata if known - may be nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// ------------------------------------------------------------
|
|
|
|
|
|
|
|
// String converts this FsS3 to a string
|
|
|
|
func (f *FsS3) String() string {
|
2014-05-06 01:25:32 +08:00
|
|
|
if f.root == "" {
|
|
|
|
return fmt.Sprintf("S3 bucket %s", f.bucket)
|
|
|
|
}
|
|
|
|
return fmt.Sprintf("S3 bucket %s path %s", f.bucket, f.root)
|
2013-01-09 02:53:35 +08:00
|
|
|
}
|
|
|
|
|
2014-03-16 00:06:11 +08:00
|
|
|
// Pattern to match a s3 path
|
|
|
|
var matcher = regexp.MustCompile(`^([^/]*)(.*)$`)
|
|
|
|
|
2013-01-09 02:53:35 +08:00
|
|
|
// parseParse parses a s3 'url'
|
|
|
|
func s3ParsePath(path string) (bucket, directory string, err error) {
|
2014-03-16 00:06:11 +08:00
|
|
|
parts := matcher.FindStringSubmatch(path)
|
|
|
|
if parts == nil {
|
|
|
|
err = fmt.Errorf("Couldn't parse bucket out of s3 path %q", path)
|
2013-01-09 02:53:35 +08:00
|
|
|
} else {
|
2014-03-16 00:06:11 +08:00
|
|
|
bucket, directory = parts[1], parts[2]
|
2013-01-09 02:53:35 +08:00
|
|
|
directory = strings.Trim(directory, "/")
|
|
|
|
}
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
// s3Connection makes a connection to s3
|
2014-03-16 00:06:11 +08:00
|
|
|
func s3Connection(name string) (*s3.S3, error) {
|
2013-01-09 02:53:35 +08:00
|
|
|
// Make the auth
|
2014-03-16 00:06:11 +08:00
|
|
|
accessKeyId := fs.ConfigFile.MustValue(name, "access_key_id")
|
|
|
|
if accessKeyId == "" {
|
|
|
|
return nil, errors.New("access_key_id not found")
|
2013-01-09 02:53:35 +08:00
|
|
|
}
|
2014-03-16 00:06:11 +08:00
|
|
|
secretAccessKey := fs.ConfigFile.MustValue(name, "secret_access_key")
|
|
|
|
if secretAccessKey == "" {
|
|
|
|
return nil, errors.New("secret_access_key not found")
|
2013-01-09 02:53:35 +08:00
|
|
|
}
|
2014-03-16 00:06:11 +08:00
|
|
|
auth := aws.Auth{AccessKey: accessKeyId, SecretKey: secretAccessKey}
|
2013-01-09 02:53:35 +08:00
|
|
|
|
|
|
|
// FIXME look through all the regions by name and use one of them if found
|
|
|
|
|
|
|
|
// Synthesize the region
|
2014-03-16 00:06:11 +08:00
|
|
|
s3Endpoint := fs.ConfigFile.MustValue(name, "endpoint")
|
|
|
|
if s3Endpoint == "" {
|
|
|
|
s3Endpoint = "https://s3.amazonaws.com/"
|
2013-01-09 02:53:35 +08:00
|
|
|
}
|
|
|
|
region := aws.Region{
|
|
|
|
Name: "s3",
|
2014-03-16 00:06:11 +08:00
|
|
|
S3Endpoint: s3Endpoint,
|
2013-01-09 02:53:35 +08:00
|
|
|
S3LocationConstraint: false,
|
|
|
|
}
|
2014-03-16 00:06:11 +08:00
|
|
|
s3LocationConstraint := fs.ConfigFile.MustValue(name, "location_constraint")
|
|
|
|
if s3LocationConstraint != "" {
|
|
|
|
region.Name = s3LocationConstraint
|
2013-01-09 02:53:35 +08:00
|
|
|
region.S3LocationConstraint = true
|
|
|
|
}
|
|
|
|
|
|
|
|
c := s3.New(auth, region)
|
|
|
|
return c, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// NewFsS3 contstructs an FsS3 from the path, bucket:path
|
2014-05-06 02:52:52 +08:00
|
|
|
func NewFs(name, root string) (fs.Fs, error) {
|
|
|
|
bucket, directory, err := s3ParsePath(root)
|
2013-01-09 02:53:35 +08:00
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
2014-03-16 00:06:11 +08:00
|
|
|
c, err := s3Connection(name)
|
2013-01-09 02:53:35 +08:00
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
f := &FsS3{
|
|
|
|
c: c,
|
|
|
|
bucket: bucket,
|
|
|
|
b: c.Bucket(bucket),
|
|
|
|
perm: s3.Private, // FIXME need user to specify
|
2014-05-06 01:25:32 +08:00
|
|
|
root: directory,
|
2013-01-09 02:53:35 +08:00
|
|
|
}
|
2014-05-06 02:52:52 +08:00
|
|
|
if f.root != "" {
|
|
|
|
f.root += "/"
|
|
|
|
// Check to see if the object exists
|
|
|
|
_, err = f.b.Head(directory, nil)
|
|
|
|
if err == nil {
|
|
|
|
remote := path.Base(directory)
|
|
|
|
f.root = path.Dir(directory)
|
|
|
|
if f.root == "." {
|
|
|
|
f.root = ""
|
|
|
|
} else {
|
|
|
|
f.root += "/"
|
|
|
|
}
|
|
|
|
obj := f.NewFsObject(remote)
|
|
|
|
// return a Fs Limited to this object
|
|
|
|
return fs.NewLimited(f, obj), nil
|
|
|
|
}
|
|
|
|
}
|
2013-01-09 02:53:35 +08:00
|
|
|
return f, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Return an FsObject from a path
|
|
|
|
//
|
|
|
|
// May return nil if an error occurred
|
2013-06-28 15:57:32 +08:00
|
|
|
func (f *FsS3) NewFsObjectWithInfo(remote string, info *s3.Key) fs.Object {
|
2013-06-28 03:13:07 +08:00
|
|
|
o := &FsObjectS3{
|
2013-01-09 02:53:35 +08:00
|
|
|
s3: f,
|
|
|
|
remote: remote,
|
|
|
|
}
|
|
|
|
if info != nil {
|
|
|
|
// Set info but not meta
|
|
|
|
var err error
|
2013-06-28 03:13:07 +08:00
|
|
|
o.lastModified, err = time.Parse(time.RFC3339, info.LastModified)
|
2013-01-09 02:53:35 +08:00
|
|
|
if err != nil {
|
2013-06-28 15:57:32 +08:00
|
|
|
fs.Log(o, "Failed to read last modified: %s", err)
|
2013-06-28 03:13:07 +08:00
|
|
|
o.lastModified = time.Now()
|
2013-01-09 02:53:35 +08:00
|
|
|
}
|
2013-06-28 03:13:07 +08:00
|
|
|
o.etag = info.ETag
|
|
|
|
o.bytes = info.Size
|
2013-01-09 02:53:35 +08:00
|
|
|
} else {
|
2013-06-28 03:13:07 +08:00
|
|
|
err := o.readMetaData() // reads info and meta, returning an error
|
2013-01-09 02:53:35 +08:00
|
|
|
if err != nil {
|
|
|
|
// logged already FsDebug("Failed to read info: %s", err)
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
}
|
2013-06-28 03:13:07 +08:00
|
|
|
return o
|
2013-01-09 02:53:35 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
// Return an FsObject from a path
|
|
|
|
//
|
|
|
|
// May return nil if an error occurred
|
2013-06-28 15:57:32 +08:00
|
|
|
func (f *FsS3) NewFsObject(remote string) fs.Object {
|
2013-01-09 02:53:35 +08:00
|
|
|
return f.NewFsObjectWithInfo(remote, nil)
|
|
|
|
}
|
|
|
|
|
2014-05-06 01:25:32 +08:00
|
|
|
// list the objects into the function supplied
|
|
|
|
//
|
|
|
|
// If directories is set it only sends directories
|
|
|
|
func (f *FsS3) list(directories bool, fn func(string, *s3.Key)) {
|
|
|
|
delimiter := ""
|
|
|
|
if directories {
|
|
|
|
delimiter = "/"
|
|
|
|
}
|
|
|
|
// FIXME need to implement ALL loop
|
|
|
|
objects, err := f.b.List(f.root, delimiter, "", 10000)
|
|
|
|
if err != nil {
|
|
|
|
fs.Stats.Error()
|
|
|
|
fs.Log(f, "Couldn't read bucket %q: %s", f.bucket, err)
|
|
|
|
} else {
|
|
|
|
rootLength := len(f.root)
|
|
|
|
if directories {
|
|
|
|
for _, remote := range objects.CommonPrefixes {
|
|
|
|
if !strings.HasPrefix(remote, f.root) {
|
|
|
|
fs.Log(f, "Odd name received %q", remote)
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
remote := remote[rootLength:]
|
2014-07-25 06:01:04 +08:00
|
|
|
if strings.HasSuffix(remote, "/") {
|
|
|
|
remote = remote[:len(remote)-1]
|
|
|
|
}
|
2014-05-06 01:25:32 +08:00
|
|
|
fn(remote, &s3.Key{Key: remote})
|
|
|
|
}
|
2013-01-09 02:53:35 +08:00
|
|
|
} else {
|
|
|
|
for i := range objects.Contents {
|
|
|
|
object := &objects.Contents[i]
|
2014-05-06 01:25:32 +08:00
|
|
|
if !strings.HasPrefix(object.Key, f.root) {
|
|
|
|
fs.Log(f, "Odd name received %q", object.Key)
|
|
|
|
continue
|
2013-01-09 02:53:35 +08:00
|
|
|
}
|
2014-05-06 01:25:32 +08:00
|
|
|
remote := object.Key[rootLength:]
|
|
|
|
fn(remote, object)
|
2013-01-09 02:53:35 +08:00
|
|
|
}
|
|
|
|
}
|
2014-05-06 01:25:32 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Walk the path returning a channel of FsObjects
|
|
|
|
func (f *FsS3) List() fs.ObjectsChan {
|
|
|
|
out := make(fs.ObjectsChan, fs.Config.Checkers)
|
|
|
|
if f.bucket == "" {
|
|
|
|
// Return no objects at top level list
|
2013-01-09 02:53:35 +08:00
|
|
|
close(out)
|
2014-05-06 01:25:32 +08:00
|
|
|
fs.Stats.Error()
|
|
|
|
fs.Log(f, "Can't list objects at root - choose a bucket using lsd")
|
|
|
|
} else {
|
|
|
|
go func() {
|
|
|
|
defer close(out)
|
|
|
|
f.list(false, func(remote string, object *s3.Key) {
|
|
|
|
if fs := f.NewFsObjectWithInfo(remote, object); fs != nil {
|
|
|
|
out <- fs
|
|
|
|
}
|
|
|
|
})
|
|
|
|
}()
|
|
|
|
}
|
2013-01-09 02:53:35 +08:00
|
|
|
return out
|
|
|
|
}
|
|
|
|
|
2013-01-24 06:43:20 +08:00
|
|
|
// Lists the buckets
|
2013-06-28 15:57:32 +08:00
|
|
|
func (f *FsS3) ListDir() fs.DirChan {
|
|
|
|
out := make(fs.DirChan, fs.Config.Checkers)
|
2014-05-06 01:25:32 +08:00
|
|
|
if f.bucket == "" {
|
|
|
|
// List the buckets
|
|
|
|
go func() {
|
|
|
|
defer close(out)
|
|
|
|
buckets, err := f.c.ListBuckets()
|
|
|
|
if err != nil {
|
|
|
|
fs.Stats.Error()
|
|
|
|
fs.Log(f, "Couldn't list buckets: %s", err)
|
|
|
|
} else {
|
|
|
|
for _, bucket := range buckets {
|
|
|
|
out <- &fs.Dir{
|
|
|
|
Name: bucket.Name,
|
|
|
|
When: bucket.CreationDate,
|
|
|
|
Bytes: -1,
|
|
|
|
Count: -1,
|
|
|
|
}
|
2013-01-24 06:43:20 +08:00
|
|
|
}
|
|
|
|
}
|
2014-05-06 01:25:32 +08:00
|
|
|
}()
|
|
|
|
} else {
|
2014-05-06 02:52:52 +08:00
|
|
|
// List the directories in the path in the bucket
|
2014-05-06 01:25:32 +08:00
|
|
|
go func() {
|
|
|
|
defer close(out)
|
|
|
|
f.list(true, func(remote string, object *s3.Key) {
|
|
|
|
out <- &fs.Dir{
|
|
|
|
Name: remote,
|
|
|
|
Bytes: object.Size,
|
|
|
|
Count: 0,
|
|
|
|
}
|
|
|
|
})
|
|
|
|
}()
|
|
|
|
}
|
2013-01-24 06:43:20 +08:00
|
|
|
return out
|
|
|
|
}
|
|
|
|
|
2013-01-09 02:53:35 +08:00
|
|
|
// Put the FsObject into the bucket
|
2013-06-28 15:57:32 +08:00
|
|
|
func (f *FsS3) Put(in io.Reader, remote string, modTime time.Time, size int64) (fs.Object, error) {
|
2013-01-09 02:53:35 +08:00
|
|
|
// Temporary FsObject under construction
|
2013-01-11 05:58:46 +08:00
|
|
|
fs := &FsObjectS3{s3: f, remote: remote}
|
2014-04-19 00:04:21 +08:00
|
|
|
return fs, fs.Update(in, modTime, size)
|
2013-01-09 02:53:35 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
// Mkdir creates the bucket if it doesn't exist
|
|
|
|
func (f *FsS3) Mkdir() error {
|
2013-01-09 06:31:16 +08:00
|
|
|
err := f.b.PutBucket(f.perm)
|
|
|
|
if err, ok := err.(*s3.Error); ok {
|
|
|
|
if err.Code == "BucketAlreadyOwnedByYou" {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return err
|
2013-01-09 02:53:35 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
// Rmdir deletes the bucket
|
|
|
|
//
|
|
|
|
// Returns an error if it isn't empty
|
|
|
|
func (f *FsS3) Rmdir() error {
|
|
|
|
return f.b.DelBucket()
|
|
|
|
}
|
|
|
|
|
2013-01-19 07:21:02 +08:00
|
|
|
// Return the precision
|
2013-06-28 03:13:07 +08:00
|
|
|
func (f *FsS3) Precision() time.Duration {
|
2013-01-19 07:21:02 +08:00
|
|
|
return time.Nanosecond
|
|
|
|
}
|
|
|
|
|
2013-01-09 02:53:35 +08:00
|
|
|
// ------------------------------------------------------------
|
|
|
|
|
2014-03-29 01:56:04 +08:00
|
|
|
// Return the parent Fs
|
|
|
|
func (o *FsObjectS3) Fs() fs.Fs {
|
|
|
|
return o.s3
|
|
|
|
}
|
|
|
|
|
|
|
|
// Return a string version
|
|
|
|
func (o *FsObjectS3) String() string {
|
|
|
|
if o == nil {
|
|
|
|
return "<nil>"
|
|
|
|
}
|
|
|
|
return o.remote
|
|
|
|
}
|
|
|
|
|
2013-01-09 02:53:35 +08:00
|
|
|
// Return the remote path
|
2013-06-28 03:13:07 +08:00
|
|
|
func (o *FsObjectS3) Remote() string {
|
|
|
|
return o.remote
|
2013-01-09 02:53:35 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
// Md5sum returns the Md5sum of an object returning a lowercase hex string
|
2013-06-28 03:13:07 +08:00
|
|
|
func (o *FsObjectS3) Md5sum() (string, error) {
|
|
|
|
return strings.Trim(strings.ToLower(o.etag), `"`), nil
|
2013-01-09 02:53:35 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
// Size returns the size of an object in bytes
|
2013-06-28 03:13:07 +08:00
|
|
|
func (o *FsObjectS3) Size() int64 {
|
|
|
|
return o.bytes
|
2013-01-09 02:53:35 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
// readMetaData gets the metadata if it hasn't already been fetched
|
|
|
|
//
|
2014-07-29 05:32:15 +08:00
|
|
|
// if we get a 404 error then we retry a few times for eventual
|
|
|
|
// consistency reasons
|
|
|
|
//
|
2013-01-09 02:53:35 +08:00
|
|
|
// it also sets the info
|
2013-06-28 03:13:07 +08:00
|
|
|
func (o *FsObjectS3) readMetaData() (err error) {
|
|
|
|
if o.meta != nil {
|
2013-01-09 02:53:35 +08:00
|
|
|
return nil
|
|
|
|
}
|
2014-07-29 05:32:15 +08:00
|
|
|
var headers s3.Headers
|
|
|
|
|
|
|
|
// Try reading the metadata a few times (with exponential
|
|
|
|
// backoff) to get around eventual consistency on 404 error
|
|
|
|
for tries := uint(0); tries < 10; tries++ {
|
|
|
|
headers, err = o.s3.b.Head(o.s3.root+o.remote, nil)
|
|
|
|
if s3Err, ok := err.(*s3.Error); ok {
|
|
|
|
if s3Err.StatusCode == http.StatusNotFound {
|
|
|
|
time.Sleep(5 * time.Millisecond << tries)
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
}
|
|
|
|
break
|
|
|
|
}
|
2013-01-09 02:53:35 +08:00
|
|
|
if err != nil {
|
2013-06-28 15:57:32 +08:00
|
|
|
fs.Debug(o, "Failed to read info: %s", err)
|
2013-01-09 02:53:35 +08:00
|
|
|
return err
|
|
|
|
}
|
2014-05-16 23:27:53 +08:00
|
|
|
var size int64
|
|
|
|
// Ignore missing Content-Length assuming it is 0
|
|
|
|
// Some versions of ceph do this due their apache proxies
|
|
|
|
if contentLength, ok := headers["Content-Length"]; ok {
|
|
|
|
size, err = strconv.ParseInt(contentLength, 10, 64)
|
|
|
|
if err != nil {
|
|
|
|
fs.Debug(o, "Failed to read size from: %q", headers)
|
|
|
|
return err
|
|
|
|
}
|
2013-01-09 02:53:35 +08:00
|
|
|
}
|
2013-06-28 03:13:07 +08:00
|
|
|
o.etag = headers["Etag"]
|
|
|
|
o.bytes = size
|
|
|
|
o.meta = headers
|
|
|
|
if o.lastModified, err = time.Parse(http.TimeFormat, headers["Last-Modified"]); err != nil {
|
2013-06-28 15:57:32 +08:00
|
|
|
fs.Log(o, "Failed to read last modified from HEAD: %s", err)
|
2013-06-28 03:13:07 +08:00
|
|
|
o.lastModified = time.Now()
|
2013-01-09 02:53:35 +08:00
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// ModTime returns the modification time of the object
|
|
|
|
//
|
|
|
|
// It attempts to read the objects mtime and if that isn't present the
|
|
|
|
// LastModified returned in the http headers
|
2013-06-28 03:13:07 +08:00
|
|
|
func (o *FsObjectS3) ModTime() time.Time {
|
|
|
|
err := o.readMetaData()
|
2013-01-09 02:53:35 +08:00
|
|
|
if err != nil {
|
2013-06-28 15:57:32 +08:00
|
|
|
fs.Log(o, "Failed to read metadata: %s", err)
|
2013-01-09 02:53:35 +08:00
|
|
|
return time.Now()
|
|
|
|
}
|
|
|
|
// read mtime out of metadata if available
|
2013-06-28 03:13:07 +08:00
|
|
|
d, ok := o.meta[metaMtime]
|
2013-01-09 02:53:35 +08:00
|
|
|
if !ok {
|
2013-06-28 15:57:32 +08:00
|
|
|
// fs.Debug(o, "No metadata")
|
2013-06-28 03:13:07 +08:00
|
|
|
return o.lastModified
|
2013-01-09 02:53:35 +08:00
|
|
|
}
|
|
|
|
modTime, err := swift.FloatStringToTime(d)
|
|
|
|
if err != nil {
|
2013-06-28 15:57:32 +08:00
|
|
|
fs.Log(o, "Failed to read mtime from object: %s", err)
|
2013-06-28 03:13:07 +08:00
|
|
|
return o.lastModified
|
2013-01-09 02:53:35 +08:00
|
|
|
}
|
|
|
|
return modTime
|
|
|
|
}
|
|
|
|
|
|
|
|
// Sets the modification time of the local fs object
|
2013-06-28 03:13:07 +08:00
|
|
|
func (o *FsObjectS3) SetModTime(modTime time.Time) {
|
|
|
|
err := o.readMetaData()
|
2013-01-09 06:31:16 +08:00
|
|
|
if err != nil {
|
2013-06-28 03:13:07 +08:00
|
|
|
fs.Stats.Error()
|
2013-06-28 15:57:32 +08:00
|
|
|
fs.Log(o, "Failed to read metadata: %s", err)
|
2013-01-09 06:31:16 +08:00
|
|
|
return
|
|
|
|
}
|
2013-06-28 03:13:07 +08:00
|
|
|
o.meta[metaMtime] = swift.TimeToFloatString(modTime)
|
2014-05-06 01:25:32 +08:00
|
|
|
_, err = o.s3.b.Update(o.s3.root+o.remote, o.s3.perm, o.meta)
|
2013-01-09 06:31:16 +08:00
|
|
|
if err != nil {
|
2013-06-28 03:13:07 +08:00
|
|
|
fs.Stats.Error()
|
2013-06-28 15:57:32 +08:00
|
|
|
fs.Log(o, "Failed to update remote mtime: %s", err)
|
2013-01-09 06:31:16 +08:00
|
|
|
}
|
2013-01-09 02:53:35 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
// Is this object storable
|
2013-06-28 03:13:07 +08:00
|
|
|
func (o *FsObjectS3) Storable() bool {
|
2013-01-09 02:53:35 +08:00
|
|
|
return true
|
|
|
|
}
|
|
|
|
|
|
|
|
// Open an object for read
|
2013-06-28 03:13:07 +08:00
|
|
|
func (o *FsObjectS3) Open() (in io.ReadCloser, err error) {
|
2014-05-06 01:25:32 +08:00
|
|
|
in, err = o.s3.b.GetReader(o.s3.root + o.remote)
|
2013-01-09 02:53:35 +08:00
|
|
|
return
|
|
|
|
}
|
|
|
|
|
2014-04-19 00:04:21 +08:00
|
|
|
// Update the Object from in with modTime and size
|
|
|
|
func (o *FsObjectS3) Update(in io.Reader, modTime time.Time, size int64) error {
|
|
|
|
// Set the mtime in the headers
|
|
|
|
headers := s3.Headers{
|
|
|
|
metaMtime: swift.TimeToFloatString(modTime),
|
|
|
|
}
|
|
|
|
|
|
|
|
// Guess the content type
|
|
|
|
contentType := mime.TypeByExtension(path.Ext(o.remote))
|
|
|
|
if contentType == "" {
|
|
|
|
contentType = "application/octet-stream"
|
|
|
|
}
|
|
|
|
|
2014-05-06 01:25:32 +08:00
|
|
|
_, err := o.s3.b.PutReaderHeaders(o.s3.root+o.remote, in, size, contentType, o.s3.perm, headers)
|
2014-07-19 19:37:11 +08:00
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
// Read the metadata from the newly created object
|
2014-07-20 18:23:05 +08:00
|
|
|
o.meta = nil // wipe old metadata
|
2014-07-19 19:37:11 +08:00
|
|
|
err = o.readMetaData()
|
2014-04-19 00:04:21 +08:00
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
2013-01-09 02:53:35 +08:00
|
|
|
// Remove an object
|
2013-06-28 03:13:07 +08:00
|
|
|
func (o *FsObjectS3) Remove() error {
|
2014-05-06 01:25:32 +08:00
|
|
|
return o.s3.b.Del(o.s3.root + o.remote)
|
2013-01-09 02:53:35 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
// Check the interfaces are satisfied
|
2013-06-28 03:13:07 +08:00
|
|
|
var _ fs.Fs = &FsS3{}
|
2013-06-28 15:57:32 +08:00
|
|
|
var _ fs.Object = &FsObjectS3{}
|