mirror of
https://github.com/rclone/rclone.git
synced 2024-11-26 02:09:55 +08:00
operations: use less memory when doing multithread uploads
For uploads which are coming from disk or going to disk or going to a backend which doesn't need to seek except for retries this doesn't buffer the input. This dramatically reduces rclone's memory usage. Fixes #7350
This commit is contained in:
parent
c0fb9ebfce
commit
1f9a79ef09
|
@ -11,7 +11,6 @@ import (
|
||||||
"github.com/rclone/rclone/fs/accounting"
|
"github.com/rclone/rclone/fs/accounting"
|
||||||
"github.com/rclone/rclone/lib/atexit"
|
"github.com/rclone/rclone/lib/atexit"
|
||||||
"github.com/rclone/rclone/lib/multipart"
|
"github.com/rclone/rclone/lib/multipart"
|
||||||
"github.com/rclone/rclone/lib/readers"
|
|
||||||
"golang.org/x/sync/errgroup"
|
"golang.org/x/sync/errgroup"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
@ -53,13 +52,13 @@ func doMultiThreadCopy(ctx context.Context, f fs.Fs, src fs.Object) bool {
|
||||||
|
|
||||||
// state for a multi-thread copy
|
// state for a multi-thread copy
|
||||||
type multiThreadCopyState struct {
|
type multiThreadCopyState struct {
|
||||||
ctx context.Context
|
ctx context.Context
|
||||||
partSize int64
|
partSize int64
|
||||||
size int64
|
size int64
|
||||||
src fs.Object
|
src fs.Object
|
||||||
acc *accounting.Account
|
acc *accounting.Account
|
||||||
numChunks int
|
numChunks int
|
||||||
noSeek bool // set if sure the receiving fs won't seek the input
|
noBuffering bool // set to read the input without buffering
|
||||||
}
|
}
|
||||||
|
|
||||||
// Copy a single chunk into place
|
// Copy a single chunk into place
|
||||||
|
@ -88,10 +87,11 @@ func (mc *multiThreadCopyState) copyChunk(ctx context.Context, chunk int, writer
|
||||||
defer fs.CheckClose(rc, &err)
|
defer fs.CheckClose(rc, &err)
|
||||||
|
|
||||||
var rs io.ReadSeeker
|
var rs io.ReadSeeker
|
||||||
if mc.noSeek {
|
if mc.noBuffering {
|
||||||
// Read directly if we are sure we aren't going to seek
|
// Read directly if we are sure we aren't going to seek
|
||||||
// and account with accounting
|
// and account with accounting
|
||||||
rs = readers.NoSeeker{Reader: mc.acc.WrapStream(rc)}
|
rc.SetAccounting(mc.acc.AccountRead)
|
||||||
|
rs = rc
|
||||||
} else {
|
} else {
|
||||||
// Read the chunk into buffered reader
|
// Read the chunk into buffered reader
|
||||||
rw := multipart.NewRW()
|
rw := multipart.NewRW()
|
||||||
|
@ -130,15 +130,25 @@ func calculateNumChunks(size int64, chunkSize int64) int {
|
||||||
func multiThreadCopy(ctx context.Context, f fs.Fs, remote string, src fs.Object, concurrency int, tr *accounting.Transfer, options ...fs.OpenOption) (newDst fs.Object, err error) {
|
func multiThreadCopy(ctx context.Context, f fs.Fs, remote string, src fs.Object, concurrency int, tr *accounting.Transfer, options ...fs.OpenOption) (newDst fs.Object, err error) {
|
||||||
openChunkWriter := f.Features().OpenChunkWriter
|
openChunkWriter := f.Features().OpenChunkWriter
|
||||||
ci := fs.GetConfig(ctx)
|
ci := fs.GetConfig(ctx)
|
||||||
noseek := false
|
noBuffering := false
|
||||||
if openChunkWriter == nil {
|
if openChunkWriter == nil {
|
||||||
openWriterAt := f.Features().OpenWriterAt
|
openWriterAt := f.Features().OpenWriterAt
|
||||||
if openWriterAt == nil {
|
if openWriterAt == nil {
|
||||||
return nil, errors.New("multi-thread copy: neither OpenChunkWriter nor OpenWriterAt supported")
|
return nil, errors.New("multi-thread copy: neither OpenChunkWriter nor OpenWriterAt supported")
|
||||||
}
|
}
|
||||||
openChunkWriter = openChunkWriterFromOpenWriterAt(openWriterAt, int64(ci.MultiThreadChunkSize), int64(ci.MultiThreadWriteBufferSize), f)
|
openChunkWriter = openChunkWriterFromOpenWriterAt(openWriterAt, int64(ci.MultiThreadChunkSize), int64(ci.MultiThreadWriteBufferSize), f)
|
||||||
// We don't seek the chunks with OpenWriterAt
|
// If we are using OpenWriterAt we don't seek the chunks so don't need to buffer
|
||||||
noseek = true
|
fs.Debugf(src, "multi-thread copy: disabling buffering because destination uses OpenWriterAt")
|
||||||
|
noBuffering = true
|
||||||
|
} else if src.Fs().Features().IsLocal {
|
||||||
|
// If the source fs is local we don't need to buffer
|
||||||
|
fs.Debugf(src, "multi-thread copy: disabling buffering because source is local disk")
|
||||||
|
noBuffering = true
|
||||||
|
} else if f.Features().ChunkWriterDoesntSeek {
|
||||||
|
// If the destination Fs promises not to seek its chunks
|
||||||
|
// (except for retries) then we don't need buffering.
|
||||||
|
fs.Debugf(src, "multi-thread copy: disabling buffering because destination has set ChunkWriterDoesntSeek")
|
||||||
|
noBuffering = true
|
||||||
}
|
}
|
||||||
|
|
||||||
if src.Size() < 0 {
|
if src.Size() < 0 {
|
||||||
|
@ -192,12 +202,12 @@ func multiThreadCopy(ctx context.Context, f fs.Fs, remote string, src fs.Object,
|
||||||
g.SetLimit(concurrency)
|
g.SetLimit(concurrency)
|
||||||
|
|
||||||
mc := &multiThreadCopyState{
|
mc := &multiThreadCopyState{
|
||||||
ctx: gCtx,
|
ctx: gCtx,
|
||||||
size: src.Size(),
|
size: src.Size(),
|
||||||
src: src,
|
src: src,
|
||||||
partSize: info.ChunkSize,
|
partSize: info.ChunkSize,
|
||||||
numChunks: numChunks,
|
numChunks: numChunks,
|
||||||
noSeek: noseek,
|
noBuffering: noBuffering,
|
||||||
}
|
}
|
||||||
|
|
||||||
// Make accounting
|
// Make accounting
|
||||||
|
|
Loading…
Reference in New Issue
Block a user