package mega import ( "bytes" "crypto/aes" "crypto/cipher" "crypto/rand" "crypto/sha512" "encoding/json" "errors" "fmt" "io" "io/ioutil" "log" "math/big" mrand "math/rand" "net/http" "os" "path/filepath" "strings" "sync" "time" "golang.org/x/crypto/pbkdf2" ) // Default settings const ( API_URL = "https://g.api.mega.co.nz" BASE_DOWNLOAD_URL = "https://mega.co.nz" RETRIES = 10 DOWNLOAD_WORKERS = 3 MAX_DOWNLOAD_WORKERS = 30 UPLOAD_WORKERS = 1 MAX_UPLOAD_WORKERS = 30 TIMEOUT = time.Second * 10 minSleepTime = 10 * time.Millisecond // for retries maxSleepTime = 5 * time.Second // for retries ) type config struct { baseurl string retries int dl_workers int ul_workers int timeout time.Duration } func newConfig() config { return config{ baseurl: API_URL, retries: RETRIES, dl_workers: DOWNLOAD_WORKERS, ul_workers: UPLOAD_WORKERS, timeout: TIMEOUT, } } // Set mega service base url func (c *config) SetAPIUrl(u string) { if strings.HasSuffix(u, "/") { u = strings.TrimRight(u, "/") } c.baseurl = u } // Set number of retries for api calls func (c *config) SetRetries(r int) { c.retries = r } // Set concurrent download workers func (c *config) SetDownloadWorkers(w int) error { if w <= MAX_DOWNLOAD_WORKERS { c.dl_workers = w return nil } return EWORKER_LIMIT_EXCEEDED } // Set connection timeout func (c *config) SetTimeOut(t time.Duration) { c.timeout = t } // Set concurrent upload workers func (c *config) SetUploadWorkers(w int) error { if w <= MAX_UPLOAD_WORKERS { c.ul_workers = w return nil } return EWORKER_LIMIT_EXCEEDED } type Mega struct { config // Version of the account accountVersion int // Salt for the account if accountVersion > 1 accountSalt []byte // Sequence number sn int64 // Server state sn ssn string // Session ID sid []byte // Master key k []byte // User handle uh []byte // Filesystem object FS *MegaFS // HTTP Client client *http.Client // Loggers logf func(format string, v ...interface{}) debugf func(format string, v ...interface{}) // serialize the API requests apiMu sync.Mutex // mutex to protext waitEvents waitEventsMu sync.Mutex // Outstanding channels to close to indicate events all received waitEvents []chan struct{} } // Filesystem node types const ( FILE = 0 FOLDER = 1 ROOT = 2 INBOX = 3 TRASH = 4 ) // Filesystem node type Node struct { fs *MegaFS name string hash string parent *Node children []*Node ntype int size int64 ts time.Time meta NodeMeta } func (n *Node) removeChild(c *Node) bool { index := -1 for i, v := range n.children { if v.hash == c.hash { index = i break } } if index >= 0 { n.children[index] = n.children[len(n.children)-1] n.children = n.children[:len(n.children)-1] return true } return false } func (n *Node) addChild(c *Node) { if n != nil { n.children = append(n.children, c) } } func (n *Node) getChildren() []*Node { return n.children } func (n *Node) GetType() int { n.fs.mutex.Lock() defer n.fs.mutex.Unlock() return n.ntype } func (n *Node) GetSize() int64 { n.fs.mutex.Lock() defer n.fs.mutex.Unlock() return n.size } func (n *Node) GetTimeStamp() time.Time { n.fs.mutex.Lock() defer n.fs.mutex.Unlock() return n.ts } func (n *Node) GetName() string { n.fs.mutex.Lock() defer n.fs.mutex.Unlock() return n.name } func (n *Node) GetHash() string { n.fs.mutex.Lock() defer n.fs.mutex.Unlock() return n.hash } type NodeMeta struct { key []byte compkey []byte iv []byte mac []byte } // Mega filesystem object type MegaFS struct { root *Node trash *Node inbox *Node sroots []*Node lookup map[string]*Node skmap map[string]string mutex sync.Mutex } // Get filesystem root node func (fs *MegaFS) GetRoot() *Node { fs.mutex.Lock() defer fs.mutex.Unlock() return fs.root } // Get filesystem trash node func (fs *MegaFS) GetTrash() *Node { fs.mutex.Lock() defer fs.mutex.Unlock() return fs.trash } // Get inbox node func (fs *MegaFS) GetInbox() *Node { fs.mutex.Lock() defer fs.mutex.Unlock() return fs.inbox } // Get a node pointer from its hash func (fs *MegaFS) HashLookup(h string) *Node { fs.mutex.Lock() defer fs.mutex.Unlock() return fs.hashLookup(h) } func (fs *MegaFS) hashLookup(h string) *Node { if node, ok := fs.lookup[h]; ok { return node } return nil } // Get the list of child nodes for a given node func (fs *MegaFS) GetChildren(n *Node) ([]*Node, error) { fs.mutex.Lock() defer fs.mutex.Unlock() var empty []*Node if n == nil { return empty, EARGS } node := fs.hashLookup(n.hash) if node == nil { return empty, ENOENT } return node.getChildren(), nil } // Retreive all the nodes in the given node tree path by name // This method returns array of nodes upto the matched subpath // (in same order as input names array) even if the target node is not located. func (fs *MegaFS) PathLookup(root *Node, ns []string) ([]*Node, error) { fs.mutex.Lock() defer fs.mutex.Unlock() if root == nil { return nil, EARGS } var err error var found bool = true nodepath := []*Node{} children := root.children for _, name := range ns { found = false for _, n := range children { if n.name == name { nodepath = append(nodepath, n) children = n.children found = true break } } if found == false { break } } if found == false { err = ENOENT } return nodepath, err } // Get top level directory nodes shared by other users func (fs *MegaFS) GetSharedRoots() []*Node { fs.mutex.Lock() defer fs.mutex.Unlock() return fs.sroots } func newMegaFS() *MegaFS { fs := &MegaFS{ lookup: make(map[string]*Node), skmap: make(map[string]string), } return fs } func New() *Mega { max := big.NewInt(0x100000000) bigx, _ := rand.Int(rand.Reader, max) cfg := newConfig() mgfs := newMegaFS() m := &Mega{ config: cfg, sn: bigx.Int64(), FS: mgfs, client: newHttpClient(cfg.timeout), } m.SetLogger(log.Printf) m.SetDebugger(nil) return m } // SetClient sets the HTTP client in use func (m *Mega) SetClient(client *http.Client) *Mega { m.client = client return m } // discardLogf discards the log messages func discardLogf(format string, v ...interface{}) { } // SetLogger sets the logger for important messages. By default this // is log.Printf. Use nil to discard the messages. func (m *Mega) SetLogger(logf func(format string, v ...interface{})) *Mega { if logf == nil { logf = discardLogf } m.logf = logf return m } // SetDebugger sets the logger for debug messages. By default these // messages are not output. func (m *Mega) SetDebugger(debugf func(format string, v ...interface{})) *Mega { if debugf == nil { debugf = discardLogf } m.debugf = debugf return m } // backOffSleep sleeps for the time pointed to then adjusts it by // doubling it up to a maximum of maxSleepTime. // // This produces a truncated exponential backoff sleep func backOffSleep(pt *time.Duration) { time.Sleep(*pt) *pt *= 2 if *pt > maxSleepTime { *pt = maxSleepTime } } // API request method func (m *Mega) api_request(r []byte) (buf []byte, err error) { var resp *http.Response // serialize the API requests m.apiMu.Lock() defer func() { m.sn++ m.apiMu.Unlock() }() url := fmt.Sprintf("%s/cs?id=%d", m.baseurl, m.sn) if m.sid != nil { url = fmt.Sprintf("%s&sid=%s", url, string(m.sid)) } sleepTime := minSleepTime // inital backoff time for i := 0; i < m.retries+1; i++ { if i != 0 { m.debugf("Retry API request %d/%d: %v", i, m.retries, err) backOffSleep(&sleepTime) } resp, err = m.client.Post(url, "application/json", bytes.NewBuffer(r)) if err != nil { continue } if resp.StatusCode != 200 { // err must be not-nil on a continue err = errors.New("Http Status: " + resp.Status) _ = resp.Body.Close() continue } buf, err = ioutil.ReadAll(resp.Body) if err != nil { _ = resp.Body.Close() continue } err = resp.Body.Close() if err != nil { continue } // at this point the body is read and closed if bytes.HasPrefix(buf, []byte("[")) == false && bytes.HasPrefix(buf, []byte("-")) == false { return nil, EBADRESP } if len(buf) < 6 { var emsg [1]ErrorMsg err = json.Unmarshal(buf, &emsg) if err != nil { err = json.Unmarshal(buf, &emsg[0]) } if err != nil { return buf, EBADRESP } err = parseError(emsg[0]) if err == EAGAIN { continue } return buf, err } if err == nil { return buf, nil } } return nil, err } // prelogin call func (m *Mega) prelogin(email string) error { var msg [1]PreloginMsg var res [1]PreloginResp email = strings.ToLower(email) // mega uses lowercased emails for login purposes - FIXME is this true for prelogin? msg[0].Cmd = "us0" msg[0].User = email req, err := json.Marshal(msg) if err != nil { return err } result, err := m.api_request(req) if err != nil { return err } err = json.Unmarshal(result, &res) if err != nil { return err } if res[0].Version == 0 { return errors.New("prelogin: no version returned") } else if res[0].Version > 2 { return fmt.Errorf("prelogin: version %d account not supported", res[0].Version) } else if res[0].Version == 2 { if len(res[0].Salt) == 0 { return errors.New("prelogin: no salt returned") } m.accountSalt = base64urldecode([]byte(res[0].Salt)) } m.accountVersion = res[0].Version return nil } // Authenticate and start a session func (m *Mega) login(email string, passwd string) error { var msg [1]LoginMsg var res [1]LoginResp var err error var result []byte email = strings.ToLower(email) // mega uses lowercased emails for login purposes passkey := password_key(passwd) uhandle := stringhash(email, passkey) m.uh = make([]byte, len(uhandle)) copy(m.uh, uhandle) msg[0].Cmd = "us" msg[0].User = email if m.accountVersion == 1 { msg[0].Handle = string(uhandle) } else { const derivedKeyLength = 2 * aes.BlockSize derivedKey := pbkdf2.Key([]byte(passwd), m.accountSalt, 100000, derivedKeyLength, sha512.New) authKey := derivedKey[aes.BlockSize:] passkey = derivedKey[:aes.BlockSize] sessionKey := make([]byte, aes.BlockSize) _, err = rand.Read(sessionKey) if err != nil { return err } msg[0].Handle = string(base64urlencode(authKey)) msg[0].SessionKey = string(base64urlencode(sessionKey)) } req, _ := json.Marshal(msg) result, err = m.api_request(req) if err != nil { return err } err = json.Unmarshal(result, &res) if err != nil { return err } m.k = base64urldecode([]byte(res[0].Key)) cipher, err := aes.NewCipher(passkey) cipher.Decrypt(m.k, m.k) m.sid, err = decryptSessionId([]byte(res[0].Privk), []byte(res[0].Csid), m.k) if err != nil { return err } return nil } // Authenticate and start a session func (m *Mega) Login(email string, passwd string) error { err := m.prelogin(email) if err != nil { return err } err = m.login(email, passwd) if err != nil { return err } waitEvent := m.WaitEventsStart() err = m.getFileSystem() if err != nil { return err } // Wait until the all the pending events have been received m.WaitEvents(waitEvent, 5*time.Second) return nil } // WaitEventsStart - call this before you do the action which might // generate events then use the returned channel as a parameter to // WaitEvents to wait for the event(s) to be received. func (m *Mega) WaitEventsStart() <-chan struct{} { ch := make(chan struct{}) m.waitEventsMu.Lock() m.waitEvents = append(m.waitEvents, ch) m.waitEventsMu.Unlock() return ch } // WaitEvents waits for all outstanding events to be received for a // maximum of duration. eventChan should be a channel as returned // from WaitEventStart. // // If the timeout elapsed then it returns true otherwise false. func (m *Mega) WaitEvents(eventChan <-chan struct{}, duration time.Duration) (timedout bool) { m.debugf("Waiting for events to be finished for %v", duration) timer := time.NewTimer(duration) select { case <-eventChan: m.debugf("Events received") timedout = false case <-timer.C: m.debugf("Timeout waiting for events") timedout = true } timer.Stop() return timedout } // waitEventsFire - fire the wait event func (m *Mega) waitEventsFire() { m.waitEventsMu.Lock() if len(m.waitEvents) > 0 { m.debugf("Signalling events received") for _, ch := range m.waitEvents { close(ch) } m.waitEvents = nil } m.waitEventsMu.Unlock() } // Get user information func (m *Mega) GetUser() (UserResp, error) { var msg [1]UserMsg var res [1]UserResp msg[0].Cmd = "ug" req, _ := json.Marshal(msg) result, err := m.api_request(req) if err != nil { return res[0], err } err = json.Unmarshal(result, &res) return res[0], err } // Get quota information func (m *Mega) GetQuota() (QuotaResp, error) { var msg [1]QuotaMsg var res [1]QuotaResp msg[0].Cmd = "uq" msg[0].Xfer = 1 msg[0].Strg = 1 req, _ := json.Marshal(msg) result, err := m.api_request(req) if err != nil { return res[0], err } err = json.Unmarshal(result, &res) return res[0], err } // Add a node into filesystem func (m *Mega) addFSNode(itm FSNode) (*Node, error) { var compkey, key []uint32 var attr FileAttr var node, parent *Node var err error master_aes, _ := aes.NewCipher(m.k) switch { case itm.T == FOLDER || itm.T == FILE: args := strings.Split(itm.Key, ":") switch { // File or folder owned by current user case args[0] == itm.User: buf := base64urldecode([]byte(args[1])) err = blockDecrypt(master_aes, buf, buf) if err != nil { return nil, err } compkey = bytes_to_a32(buf) // Shared folder case itm.SUser != "" && itm.SKey != "": sk := base64urldecode([]byte(itm.SKey)) err = blockDecrypt(master_aes, sk, sk) if err != nil { return nil, err } sk_aes, _ := aes.NewCipher(sk) m.FS.skmap[itm.Hash] = itm.SKey buf := base64urldecode([]byte(args[1])) err = blockDecrypt(sk_aes, buf, buf) if err != nil { return nil, err } compkey = bytes_to_a32(buf) // Shared file default: k := m.FS.skmap[args[0]] b := base64urldecode([]byte(k)) err = blockDecrypt(master_aes, b, b) if err != nil { return nil, err } block, _ := aes.NewCipher(b) buf := base64urldecode([]byte(args[1])) err = blockDecrypt(block, buf, buf) if err != nil { return nil, err } compkey = bytes_to_a32(buf) } switch { case itm.T == FILE: key = []uint32{compkey[0] ^ compkey[4], compkey[1] ^ compkey[5], compkey[2] ^ compkey[6], compkey[3] ^ compkey[7]} default: key = compkey } attr, err = decryptAttr(a32_to_bytes(key), []byte(itm.Attr)) // FIXME: if err != nil { attr.Name = "BAD ATTRIBUTE" } } n, ok := m.FS.lookup[itm.Hash] switch { case ok: node = n default: node = &Node{ fs: m.FS, ntype: itm.T, size: itm.Sz, ts: time.Unix(itm.Ts, 0), } m.FS.lookup[itm.Hash] = node } n, ok = m.FS.lookup[itm.Parent] switch { case ok: parent = n parent.removeChild(node) parent.addChild(node) default: parent = nil if itm.Parent != "" { parent = &Node{ fs: m.FS, children: []*Node{node}, ntype: FOLDER, } m.FS.lookup[itm.Parent] = parent } } switch { case itm.T == FILE: var meta NodeMeta meta.key = a32_to_bytes(key) meta.iv = a32_to_bytes([]uint32{compkey[4], compkey[5], 0, 0}) meta.mac = a32_to_bytes([]uint32{compkey[6], compkey[7]}) meta.compkey = a32_to_bytes(compkey) node.meta = meta case itm.T == FOLDER: var meta NodeMeta meta.key = a32_to_bytes(key) meta.compkey = a32_to_bytes(compkey) node.meta = meta case itm.T == ROOT: attr.Name = "Cloud Drive" m.FS.root = node case itm.T == INBOX: attr.Name = "InBox" m.FS.inbox = node case itm.T == TRASH: attr.Name = "Trash" m.FS.trash = node } // Shared directories if itm.SUser != "" && itm.SKey != "" { m.FS.sroots = append(m.FS.sroots, node) } node.name = attr.Name node.hash = itm.Hash node.parent = parent node.ntype = itm.T return node, nil } // Get all nodes from filesystem func (m *Mega) getFileSystem() error { m.FS.mutex.Lock() defer m.FS.mutex.Unlock() var msg [1]FilesMsg var res [1]FilesResp msg[0].Cmd = "f" msg[0].C = 1 req, _ := json.Marshal(msg) result, err := m.api_request(req) if err != nil { return err } err = json.Unmarshal(result, &res) if err != nil { return err } for _, sk := range res[0].Ok { m.FS.skmap[sk.Hash] = sk.Key } for _, itm := range res[0].F { _, err = m.addFSNode(itm) if err != nil { return err } } m.ssn = res[0].Sn go m.pollEvents() return nil } // Download contains the internal state of a download type Download struct { m *Mega src *Node resourceUrl string aes_block cipher.Block iv []byte mac_enc cipher.BlockMode mutex sync.Mutex // to protect the following chunks []chunkSize chunk_macs [][]byte } // an all nil IV for mac calculations var zero_iv = make([]byte, 16) // Create a new Download from the src Node // // Call Chunks to find out how many chunks there are, then for id = // 0..chunks-1 call DownloadChunk. Finally call Finish() to receive // the error status. func (m *Mega) NewDownload(src *Node) (*Download, error) { if src == nil { return nil, EARGS } var msg [1]DownloadMsg var res [1]DownloadResp m.FS.mutex.Lock() msg[0].Cmd = "g" msg[0].G = 1 msg[0].N = src.hash key := src.meta.key m.FS.mutex.Unlock() request, err := json.Marshal(msg) if err != nil { return nil, err } result, err := m.api_request(request) if err != nil { return nil, err } err = json.Unmarshal(result, &res) if err != nil { return nil, err } _, err = decryptAttr(key, []byte(res[0].Attr)) if err != nil { return nil, err } chunks := getChunkSizes(int64(res[0].Size)) aes_block, err := aes.NewCipher(key) if err != nil { return nil, err } mac_enc := cipher.NewCBCEncrypter(aes_block, zero_iv) m.FS.mutex.Lock() t := bytes_to_a32(src.meta.iv) m.FS.mutex.Unlock() iv := a32_to_bytes([]uint32{t[0], t[1], t[0], t[1]}) d := &Download{ m: m, src: src, resourceUrl: res[0].G, aes_block: aes_block, iv: iv, mac_enc: mac_enc, chunks: chunks, chunk_macs: make([][]byte, len(chunks)), } return d, nil } // Chunks returns The number of chunks in the download. func (d *Download) Chunks() int { return len(d.chunks) } // ChunkLocation returns the position in the file and the size of the chunk func (d *Download) ChunkLocation(id int) (position int64, size int, err error) { if id < 0 || id >= len(d.chunks) { return 0, 0, EARGS } d.mutex.Lock() defer d.mutex.Unlock() return d.chunks[id].position, d.chunks[id].size, nil } // DownloadChunk gets a chunk with the given number and update the // mac, returning the position in the file of the chunk func (d *Download) DownloadChunk(id int) (chunk []byte, err error) { if id < 0 || id >= len(d.chunks) { return nil, EARGS } chk_start, chk_size, err := d.ChunkLocation(id) if err != nil { return nil, err } var resp *http.Response chunk_url := fmt.Sprintf("%s/%d-%d", d.resourceUrl, chk_start, chk_start+int64(chk_size)-1) sleepTime := minSleepTime // inital backoff time for retry := 0; retry < d.m.retries+1; retry++ { resp, err = d.m.client.Get(chunk_url) if err == nil { if resp.StatusCode == 200 { break } err = errors.New("Http Status: " + resp.Status) _ = resp.Body.Close() } d.m.debugf("%s: Retry download chunk %d/%d: %v", d.src.name, retry, d.m.retries, err) backOffSleep(&sleepTime) } if err != nil { return nil, err } if resp == nil { return nil, errors.New("retries exceeded") } chunk, err = ioutil.ReadAll(resp.Body) if err != nil { _ = resp.Body.Close() return nil, err } err = resp.Body.Close() if err != nil { return nil, err } // body is read and closed here if len(chunk) != chk_size { return nil, errors.New("wrong size for downloaded chunk") } // Decrypt the block ctr_iv := bytes_to_a32(d.src.meta.iv) ctr_iv[2] = uint32(uint64(chk_start) / 0x1000000000) ctr_iv[3] = uint32(chk_start / 0x10) ctr_aes := cipher.NewCTR(d.aes_block, a32_to_bytes(ctr_iv)) ctr_aes.XORKeyStream(chunk, chunk) // Update the chunk_macs enc := cipher.NewCBCEncrypter(d.aes_block, d.iv) i := 0 block := make([]byte, 16) paddedChunk := paddnull(chunk, 16) for i = 0; i < len(paddedChunk); i += 16 { enc.CryptBlocks(block, paddedChunk[i:i+16]) } d.mutex.Lock() if len(d.chunk_macs) > 0 { d.chunk_macs[id] = make([]byte, 16) copy(d.chunk_macs[id], block) } d.mutex.Unlock() return chunk, nil } // Finish checks the accumulated MAC for each block. // // If all the chunks weren't downloaded then it will just return nil func (d *Download) Finish() (err error) { // Can't check a 0 sized file if len(d.chunk_macs) == 0 { return nil } mac_data := make([]byte, 16) for _, v := range d.chunk_macs { // If a chunk_macs hasn't been set then the whole file // wasn't downloaded and we can't check it if v == nil { return nil } d.mac_enc.CryptBlocks(mac_data, v) } tmac := bytes_to_a32(mac_data) if bytes.Equal(a32_to_bytes([]uint32{tmac[0] ^ tmac[1], tmac[2] ^ tmac[3]}), d.src.meta.mac) == false { return EMACMISMATCH } return nil } // Download file from filesystem reporting progress if not nil func (m *Mega) DownloadFile(src *Node, dstpath string, progress *chan int) error { defer func() { if progress != nil { close(*progress) } }() d, err := m.NewDownload(src) if err != nil { return err } _, err = os.Stat(dstpath) if os.IsExist(err) { err = os.Remove(dstpath) if err != nil { return err } } outfile, err := os.OpenFile(dstpath, os.O_RDWR|os.O_CREATE, 0600) if err != nil { return err } workch := make(chan int) errch := make(chan error, m.dl_workers) wg := sync.WaitGroup{} // Fire chunk download workers for w := 0; w < m.dl_workers; w++ { wg.Add(1) go func() { defer wg.Done() // Wait for work blocked on channel for id := range workch { chunk, err := d.DownloadChunk(id) if err != nil { errch <- err return } chk_start, _, err := d.ChunkLocation(id) if err != nil { errch <- err return } _, err = outfile.WriteAt(chunk, chk_start) if err != nil { errch <- err return } if progress != nil { *progress <- len(chunk) } } }() } // Place chunk download jobs to chan err = nil for id := 0; id < d.Chunks() && err == nil; { select { case workch <- id: id++ case err = <-errch: } } close(workch) wg.Wait() closeErr := outfile.Close() if err != nil { _ = os.Remove(dstpath) return err } if closeErr != nil { return closeErr } return d.Finish() } // Upload contains the internal state of a upload type Upload struct { m *Mega parenthash string name string uploadUrl string aes_block cipher.Block iv []byte kiv []byte mac_enc cipher.BlockMode kbytes []byte ukey []uint32 mutex sync.Mutex // to protect the following chunks []chunkSize chunk_macs [][]byte completion_handle []byte } // Create a new Upload of name into parent of fileSize // // Call Chunks to find out how many chunks there are, then for id = // 0..chunks-1 Call ChunkLocation then UploadChunk. Finally call // Finish() to receive the error status and the *Node. func (m *Mega) NewUpload(parent *Node, name string, fileSize int64) (*Upload, error) { if parent == nil { return nil, EARGS } var msg [1]UploadMsg var res [1]UploadResp parenthash := parent.GetHash() msg[0].Cmd = "u" msg[0].S = fileSize request, err := json.Marshal(msg) if err != nil { return nil, err } result, err := m.api_request(request) if err != nil { return nil, err } err = json.Unmarshal(result, &res) if err != nil { return nil, err } uploadUrl := res[0].P ukey := []uint32{0, 0, 0, 0, 0, 0} for i, _ := range ukey { ukey[i] = uint32(mrand.Int31()) } kbytes := a32_to_bytes(ukey[:4]) kiv := a32_to_bytes([]uint32{ukey[4], ukey[5], 0, 0}) aes_block, _ := aes.NewCipher(kbytes) mac_enc := cipher.NewCBCEncrypter(aes_block, zero_iv) iv := a32_to_bytes([]uint32{ukey[4], ukey[5], ukey[4], ukey[5]}) chunks := getChunkSizes(fileSize) // File size is zero // Do one empty request to get the completion handle if len(chunks) == 0 { chunks = append(chunks, chunkSize{position: 0, size: 0}) } u := &Upload{ m: m, parenthash: parenthash, name: name, uploadUrl: uploadUrl, aes_block: aes_block, iv: iv, kiv: kiv, mac_enc: mac_enc, kbytes: kbytes, ukey: ukey, chunks: chunks, chunk_macs: make([][]byte, len(chunks)), completion_handle: []byte{}, } return u, nil } // Chunks returns The number of chunks in the upload. func (u *Upload) Chunks() int { return len(u.chunks) } // ChunkLocation returns the position in the file and the size of the chunk func (u *Upload) ChunkLocation(id int) (position int64, size int, err error) { if id < 0 || id >= len(u.chunks) { return 0, 0, EARGS } return u.chunks[id].position, u.chunks[id].size, nil } // UploadChunk uploads the chunk of id func (u *Upload) UploadChunk(id int, chunk []byte) (err error) { chk_start, chk_size, err := u.ChunkLocation(id) if err != nil { return err } if len(chunk) != chk_size { return errors.New("upload chunk is wrong size") } ctr_iv := bytes_to_a32(u.kiv) ctr_iv[2] = uint32(uint64(chk_start) / 0x1000000000) ctr_iv[3] = uint32(chk_start / 0x10) ctr_aes := cipher.NewCTR(u.aes_block, a32_to_bytes(ctr_iv)) enc := cipher.NewCBCEncrypter(u.aes_block, u.iv) i := 0 block := make([]byte, 16) paddedchunk := paddnull(chunk, 16) for i = 0; i < len(paddedchunk); i += 16 { copy(block[0:16], paddedchunk[i:i+16]) enc.CryptBlocks(block, block) } var rsp *http.Response var req *http.Request ctr_aes.XORKeyStream(chunk, chunk) chk_url := fmt.Sprintf("%s/%d", u.uploadUrl, chk_start) chunk_resp := []byte{} sleepTime := minSleepTime // inital backoff time for retry := 0; retry < u.m.retries+1; retry++ { reader := bytes.NewBuffer(chunk) req, err = http.NewRequest("POST", chk_url, reader) if err != nil { return err } rsp, err = u.m.client.Do(req) if err == nil { if rsp.StatusCode == 200 { break } err = errors.New("Http Status: " + rsp.Status) _ = rsp.Body.Close() } u.m.debugf("%s: Retry upload chunk %d/%d: %v", u.name, retry, u.m.retries, err) backOffSleep(&sleepTime) } if err != nil { return err } if rsp == nil { return errors.New("retries exceeded") } chunk_resp, err = ioutil.ReadAll(rsp.Body) if err != nil { _ = rsp.Body.Close() return err } err = rsp.Body.Close() if err != nil { return err } if bytes.Equal(chunk_resp, nil) == false { u.mutex.Lock() u.completion_handle = chunk_resp u.mutex.Unlock() } // Update chunk MACs on success only u.mutex.Lock() if len(u.chunk_macs) > 0 { u.chunk_macs[id] = make([]byte, 16) copy(u.chunk_macs[id], block) } u.mutex.Unlock() return nil } // Finish completes the upload and returns the created node func (u *Upload) Finish() (node *Node, err error) { mac_data := make([]byte, 16) for _, v := range u.chunk_macs { u.mac_enc.CryptBlocks(mac_data, v) } t := bytes_to_a32(mac_data) meta_mac := []uint32{t[0] ^ t[1], t[2] ^ t[3]} attr := FileAttr{u.name} attr_data, err := encryptAttr(u.kbytes, attr) if err != nil { return nil, err } key := []uint32{u.ukey[0] ^ u.ukey[4], u.ukey[1] ^ u.ukey[5], u.ukey[2] ^ meta_mac[0], u.ukey[3] ^ meta_mac[1], u.ukey[4], u.ukey[5], meta_mac[0], meta_mac[1]} buf := a32_to_bytes(key) master_aes, err := aes.NewCipher(u.m.k) if err != nil { return nil, err } enc := cipher.NewCBCEncrypter(master_aes, zero_iv) enc.CryptBlocks(buf[:16], buf[:16]) enc = cipher.NewCBCEncrypter(master_aes, zero_iv) enc.CryptBlocks(buf[16:], buf[16:]) var cmsg [1]UploadCompleteMsg var cres [1]UploadCompleteResp cmsg[0].Cmd = "p" cmsg[0].T = u.parenthash cmsg[0].N[0].H = string(u.completion_handle) cmsg[0].N[0].T = FILE cmsg[0].N[0].A = string(attr_data) cmsg[0].N[0].K = string(base64urlencode(buf)) request, err := json.Marshal(cmsg) if err != nil { return nil, err } result, err := u.m.api_request(request) if err != nil { return nil, err } err = json.Unmarshal(result, &cres) if err != nil { return nil, err } u.m.FS.mutex.Lock() defer u.m.FS.mutex.Unlock() return u.m.addFSNode(cres[0].F[0]) } // Upload a file to the filesystem func (m *Mega) UploadFile(srcpath string, parent *Node, name string, progress *chan int) (*Node, error) { defer func() { if progress != nil { close(*progress) } }() var infile *os.File var fileSize int64 info, err := os.Stat(srcpath) if err == nil { fileSize = info.Size() } infile, err = os.OpenFile(srcpath, os.O_RDONLY, 0666) if err != nil { return nil, err } if name == "" { name = filepath.Base(srcpath) } u, err := m.NewUpload(parent, name, fileSize) if err != nil { return nil, err } workch := make(chan int) errch := make(chan error, m.ul_workers) wg := sync.WaitGroup{} // Fire chunk upload workers for w := 0; w < m.ul_workers; w++ { wg.Add(1) go func() { defer wg.Done() for id := range workch { chk_start, chk_size, err := u.ChunkLocation(id) if err != nil { errch <- err return } chunk := make([]byte, chk_size) n, err := infile.ReadAt(chunk, chk_start) if err != nil && err != io.EOF { errch <- err return } if n != len(chunk) { errch <- errors.New("chunk too short") return } err = u.UploadChunk(id, chunk) if err != nil { errch <- err return } if progress != nil { *progress <- chk_size } } }() } // Place chunk download jobs to chan err = nil for id := 0; id < u.Chunks() && err == nil; { select { case workch <- id: id++ case err = <-errch: } } close(workch) wg.Wait() if err != nil { return nil, err } return u.Finish() } // Move a file from one location to another func (m *Mega) Move(src *Node, parent *Node) error { m.FS.mutex.Lock() defer m.FS.mutex.Unlock() if src == nil || parent == nil { return EARGS } var msg [1]MoveFileMsg var err error msg[0].Cmd = "m" msg[0].N = src.hash msg[0].T = parent.hash msg[0].I, err = randString(10) if err != nil { return err } request, _ := json.Marshal(msg) _, err = m.api_request(request) if err != nil { return err } if src.parent != nil { src.parent.removeChild(src) } parent.addChild(src) src.parent = parent return nil } // Rename a file or folder func (m *Mega) Rename(src *Node, name string) error { m.FS.mutex.Lock() defer m.FS.mutex.Unlock() if src == nil { return EARGS } var msg [1]FileAttrMsg master_aes, _ := aes.NewCipher(m.k) attr := FileAttr{name} attr_data, _ := encryptAttr(src.meta.key, attr) key := make([]byte, len(src.meta.compkey)) err := blockEncrypt(master_aes, key, src.meta.compkey) if err != nil { return err } msg[0].Cmd = "a" msg[0].Attr = string(attr_data) msg[0].Key = string(base64urlencode(key)) msg[0].N = src.hash msg[0].I, err = randString(10) if err != nil { return err } req, _ := json.Marshal(msg) _, err = m.api_request(req) src.name = name return err } // Create a directory in the filesystem func (m *Mega) CreateDir(name string, parent *Node) (*Node, error) { m.FS.mutex.Lock() defer m.FS.mutex.Unlock() if parent == nil { return nil, EARGS } var msg [1]UploadCompleteMsg var res [1]UploadCompleteResp compkey := []uint32{0, 0, 0, 0, 0, 0} for i, _ := range compkey { compkey[i] = uint32(mrand.Int31()) } master_aes, _ := aes.NewCipher(m.k) attr := FileAttr{name} ukey := a32_to_bytes(compkey[:4]) attr_data, _ := encryptAttr(ukey, attr) key := make([]byte, len(ukey)) err := blockEncrypt(master_aes, key, ukey) if err != nil { return nil, err } msg[0].Cmd = "p" msg[0].T = parent.hash msg[0].N[0].H = "xxxxxxxx" msg[0].N[0].T = FOLDER msg[0].N[0].A = string(attr_data) msg[0].N[0].K = string(base64urlencode(key)) msg[0].I, err = randString(10) if err != nil { return nil, err } req, _ := json.Marshal(msg) result, err := m.api_request(req) if err != nil { return nil, err } err = json.Unmarshal(result, &res) if err != nil { return nil, err } node, err := m.addFSNode(res[0].F[0]) return node, err } // Delete a file or directory from filesystem func (m *Mega) Delete(node *Node, destroy bool) error { if node == nil { return EARGS } if destroy == false { return m.Move(node, m.FS.trash) } m.FS.mutex.Lock() defer m.FS.mutex.Unlock() var msg [1]FileDeleteMsg var err error msg[0].Cmd = "d" msg[0].N = node.hash msg[0].I, err = randString(10) if err != nil { return err } req, _ := json.Marshal(msg) _, err = m.api_request(req) parent := m.FS.lookup[node.hash] parent.removeChild(node) delete(m.FS.lookup, node.hash) return err } // process an add node event func (m *Mega) processAddNode(evRaw []byte) error { m.FS.mutex.Lock() defer m.FS.mutex.Unlock() var ev FSEvent err := json.Unmarshal(evRaw, &ev) if err != nil { return err } for _, itm := range ev.T.Files { _, err = m.addFSNode(itm) if err != nil { return err } } return nil } // process an update node event func (m *Mega) processUpdateNode(evRaw []byte) error { m.FS.mutex.Lock() defer m.FS.mutex.Unlock() var ev FSEvent err := json.Unmarshal(evRaw, &ev) if err != nil { return err } node := m.FS.hashLookup(ev.N) attr, err := decryptAttr(node.meta.key, []byte(ev.Attr)) if err == nil { node.name = attr.Name } else { node.name = "BAD ATTRIBUTE" } node.ts = time.Unix(ev.Ts, 0) return nil } // process a delete node event func (m *Mega) processDeleteNode(evRaw []byte) error { m.FS.mutex.Lock() defer m.FS.mutex.Unlock() var ev FSEvent err := json.Unmarshal(evRaw, &ev) if err != nil { return err } node := m.FS.hashLookup(ev.N) if node != nil && node.parent != nil { node.parent.removeChild(node) delete(m.FS.lookup, node.hash) } return nil } // Listen for server event notifications and play actions func (m *Mega) pollEvents() { var err error var resp *http.Response sleepTime := minSleepTime // inital backoff time for { if err != nil { m.debugf("pollEvents: error from server", err) backOffSleep(&sleepTime) } else { // reset sleep time to minimum on success sleepTime = minSleepTime } url := fmt.Sprintf("%s/sc?sn=%s&sid=%s", m.baseurl, m.ssn, string(m.sid)) resp, err = m.client.Post(url, "application/xml", nil) if err != nil { m.logf("pollEvents: Error fetching status: %s", err) continue } if resp.StatusCode != 200 { m.logf("pollEvents: Error from server: %s", resp.Status) _ = resp.Body.Close() continue } buf, err := ioutil.ReadAll(resp.Body) if err != nil { m.logf("pollEvents: Error reading body: %v", err) _ = resp.Body.Close() continue } err = resp.Body.Close() if err != nil { m.logf("pollEvents: Error closing body: %v", err) continue } // body is read and closed here // First attempt to parse an array var events Events err = json.Unmarshal(buf, &events) if err != nil { // Try parsing as a lone error message var emsg ErrorMsg err = json.Unmarshal(buf, &emsg) if err != nil { m.logf("pollEvents: Bad response received from server: %s", buf) } else { err = parseError(emsg) if err == EAGAIN { } else if err != nil { m.logf("pollEvents: Error received from server: %v", err) } } continue } // if wait URL is set, then fetch it and continue - we // don't expect anything else if we have a wait URL. if events.W != "" { m.waitEventsFire() if len(events.E) > 0 { m.logf("pollEvents: Unexpected event with w set: %s", buf) } resp, err = m.client.Get(events.W) if err == nil { _ = resp.Body.Close() } continue } m.ssn = events.Sn // For each event in the array, parse it for _, evRaw := range events.E { // First attempt to unmarshal as an error message var emsg ErrorMsg err = json.Unmarshal(evRaw, &emsg) if err == nil { m.logf("pollEvents: Error message received %s", evRaw) err = parseError(emsg) if err != nil { m.logf("pollEvents: Event from server was error: %v", err) } continue } // Now unmarshal as a generic event var gev GenericEvent err = json.Unmarshal(evRaw, &gev) if err != nil { m.logf("pollEvents: Couldn't parse event from server: %v: %s", err, evRaw) continue } m.debugf("pollEvents: Parsing event %q: %s", gev.Cmd, evRaw) // Work out what to do with the event var process func([]byte) error switch gev.Cmd { case "t": // node addition process = m.processAddNode case "u": // node update process = m.processUpdateNode case "d": // node deletion process = m.processDeleteNode case "s", "s2": // share addition/update/revocation case "c": // contact addition/update case "k": // crypto key request case "fa": // file attribute update case "ua": // user attribute update case "psts": // account updated case "ipc": // incoming pending contact request (to us) case "opc": // outgoing pending contact request (from us) case "upci": // incoming pending contact request update (accept/deny/ignore) case "upco": // outgoing pending contact request update (from them, accept/deny/ignore) case "ph": // public links handles case "se": // set email case "mcc": // chat creation / peer's invitation / peer's removal case "mcna": // granted / revoked access to a node case "uac": // user access control default: m.debugf("pollEvents: Unknown message %q received: %s", gev.Cmd, evRaw) } // process the event if we can if process != nil { err := process(evRaw) if err != nil { m.logf("pollEvents: Error processing event %q '%s': %v", gev.Cmd, evRaw, err) } } } } } func (m *Mega) getLink(n *Node) (string, error) { var msg [1]GetLinkMsg var res [1]string msg[0].Cmd = "l" msg[0].N = n.GetHash() req, _ := json.Marshal(msg) result, err := m.api_request(req) if err != nil { return "", err } err = json.Unmarshal(result, &res) if err != nil { return "", err } return res[0], nil } // Exports public link for node, with or without decryption key included func (m *Mega) Link(n *Node, includeKey bool) (string, error) { id, err := m.getLink(n) if err != nil { return "", err } if includeKey { m.FS.mutex.Lock() key := string(base64urlencode(n.meta.compkey)) m.FS.mutex.Unlock() return fmt.Sprintf("%v/#!%v!%v", BASE_DOWNLOAD_URL, id, key), nil } else { return fmt.Sprintf("%v/#!%v", BASE_DOWNLOAD_URL, id), nil } }