333 lines
8.0 KiB
Go
333 lines
8.0 KiB
Go
// S3Store is a storage backend used as a tusd.DataStore in tusd.NewHandler.
|
|
// It stores the uploads in a directory specified in two different files: The
|
|
// `[id].info` files are used to store the fileinfo in JSON format. The
|
|
// `[id].bin` files contain the raw binary data uploaded.
|
|
// No cleanup is performed so you may want to run a cronjob to ensure your disk
|
|
// is not filled up with old and finished uploads.
|
|
package s3store
|
|
|
|
import (
|
|
"bytes"
|
|
"encoding/json"
|
|
"errors"
|
|
"fmt"
|
|
"io"
|
|
"io/ioutil"
|
|
"os"
|
|
"strings"
|
|
|
|
"github.com/tus/tusd"
|
|
"github.com/tus/tusd/uid"
|
|
|
|
"github.com/aws/aws-sdk-go/aws"
|
|
"github.com/aws/aws-sdk-go/aws/awserr"
|
|
"github.com/aws/aws-sdk-go/aws/session"
|
|
"github.com/aws/aws-sdk-go/service/s3"
|
|
)
|
|
|
|
const MinPartSize = int64(5 * 1024 * 1024)
|
|
|
|
// See the tusd.DataStore interface for documentation about the different
|
|
// methods.
|
|
type S3Store struct {
|
|
Bucket string
|
|
Service *s3.S3
|
|
MaxPartSize int64
|
|
}
|
|
|
|
func New(bucket string, service *s3.S3) *S3Store {
|
|
return &S3Store{
|
|
Bucket: bucket,
|
|
Service: service,
|
|
MaxPartSize: 6 * 1024 * 1024,
|
|
}
|
|
}
|
|
|
|
func (store S3Store) NewUpload(info tusd.FileInfo) (id string, err error) {
|
|
uploadId := uid.Uid()
|
|
|
|
infoJson, err := json.Marshal(info)
|
|
if err != nil {
|
|
return id, err
|
|
}
|
|
|
|
// Create object on S3 containing information about the file
|
|
_, err = store.Service.PutObject(&s3.PutObjectInput{
|
|
Bucket: aws.String(store.Bucket),
|
|
Key: aws.String(uploadId + ".info"),
|
|
Body: bytes.NewReader(infoJson),
|
|
ContentLength: aws.Int64(int64(len(infoJson))),
|
|
})
|
|
if err != nil {
|
|
return id, err
|
|
}
|
|
|
|
// Create the actual multipart upload
|
|
res, err := store.Service.CreateMultipartUpload(&s3.CreateMultipartUploadInput{
|
|
Bucket: aws.String(store.Bucket),
|
|
Key: aws.String(uploadId),
|
|
})
|
|
if err != nil {
|
|
return id, err
|
|
}
|
|
|
|
id = uploadId + "+" + *res.UploadId
|
|
|
|
return
|
|
}
|
|
|
|
func (store S3Store) WriteChunk(id string, offset int64, src io.Reader) (int64, error) {
|
|
uploadId, multipartId := splitIds(id)
|
|
|
|
// Get the total size of the current upload
|
|
info, err := store.GetInfo(id)
|
|
if err != nil {
|
|
return 0, err
|
|
}
|
|
|
|
size := info.Size
|
|
bytesUploaded := int64(0)
|
|
|
|
fmt.Println("size:", size)
|
|
fmt.Println("offset:", offset)
|
|
|
|
// Get number of parts to generate next number
|
|
listPtr, err := store.Service.ListParts(&s3.ListPartsInput{
|
|
Bucket: aws.String(store.Bucket),
|
|
Key: aws.String(uploadId),
|
|
UploadId: aws.String(multipartId),
|
|
})
|
|
if err != nil {
|
|
return 0, err
|
|
}
|
|
|
|
list := *listPtr
|
|
numParts := len(list.Parts)
|
|
fmt.Println("numParts:", numParts)
|
|
nextPartNum := int64(numParts + 1)
|
|
|
|
fmt.Println("nextNum:", nextPartNum)
|
|
|
|
for {
|
|
// Create a temporary file to store the part in it
|
|
file, err := ioutil.TempFile("", "tusd-s3-tmp-")
|
|
if err != nil {
|
|
return bytesUploaded, err
|
|
}
|
|
defer os.Remove(file.Name())
|
|
defer file.Close()
|
|
|
|
fmt.Println("max:", store.MaxPartSize)
|
|
|
|
limitedReader := io.LimitReader(src, store.MaxPartSize)
|
|
n, err := io.Copy(file, limitedReader)
|
|
fmt.Println("err:", err)
|
|
if err != nil && err != io.EOF {
|
|
return bytesUploaded, err
|
|
}
|
|
|
|
fmt.Println("n:", n)
|
|
fmt.Println("offset:", offset)
|
|
|
|
if (size - offset) <= MinPartSize {
|
|
if (size - offset) != n {
|
|
return bytesUploaded, nil
|
|
}
|
|
} else if n < MinPartSize {
|
|
return bytesUploaded, nil
|
|
}
|
|
|
|
fmt.Println("upload…")
|
|
|
|
// Seek to the beginning of the file
|
|
file.Seek(0, 0)
|
|
|
|
_, err = store.Service.UploadPart(&s3.UploadPartInput{
|
|
Bucket: aws.String(store.Bucket),
|
|
Key: aws.String(uploadId),
|
|
UploadId: aws.String(multipartId),
|
|
PartNumber: aws.Int64(nextPartNum),
|
|
Body: file,
|
|
})
|
|
if err != nil {
|
|
return bytesUploaded, err
|
|
}
|
|
|
|
pos, _ := file.Seek(1, 0)
|
|
fmt.Println("pos:", pos)
|
|
|
|
offset += bytesUploaded
|
|
bytesUploaded += n
|
|
nextPartNum += 1
|
|
}
|
|
|
|
fmt.Println("bytes:", bytesUploaded)
|
|
|
|
return bytesUploaded, nil
|
|
}
|
|
|
|
func (store S3Store) GetInfo(id string) (info tusd.FileInfo, err error) {
|
|
uploadId, multipartId := splitIds(id)
|
|
|
|
// Get file info stored in seperate object
|
|
res, err := store.Service.GetObject(&s3.GetObjectInput{
|
|
Bucket: aws.String(store.Bucket),
|
|
Key: aws.String(uploadId + ".info"),
|
|
})
|
|
if err != nil {
|
|
if err, ok := err.(awserr.Error); ok && err.Code() == "NoSuchKey" {
|
|
return info, tusd.ErrNotFound
|
|
}
|
|
|
|
return info, err
|
|
}
|
|
|
|
if err := json.NewDecoder(res.Body).Decode(&info); err != nil {
|
|
return info, err
|
|
}
|
|
|
|
// Get uploaded parts and their offset
|
|
listPtr, err := store.Service.ListParts(&s3.ListPartsInput{
|
|
Bucket: aws.String(store.Bucket),
|
|
Key: aws.String(uploadId),
|
|
UploadId: aws.String(multipartId),
|
|
})
|
|
if err != nil {
|
|
// Check if the error is caused by the upload not being found. This happens
|
|
// when the multipart upload has already been completed or aborted. Since
|
|
// we already found the info object, we know that the upload has been
|
|
// completed and therefore can ensure the the offset is the size.
|
|
fmt.Println(err, err.(awserr.Error).Code())
|
|
if err, ok := err.(awserr.Error); ok && err.Code() == "NoSuchUpload" {
|
|
info.Offset = info.Size
|
|
return info, nil
|
|
} else {
|
|
return info, err
|
|
}
|
|
}
|
|
|
|
list := *listPtr
|
|
|
|
offset := int64(0)
|
|
|
|
for _, part := range list.Parts {
|
|
offset += *part.Size
|
|
}
|
|
|
|
info.Offset = offset
|
|
|
|
return
|
|
}
|
|
|
|
func (store S3Store) GetReader(id string) (io.Reader, error) {
|
|
uploadId, multipartId := splitIds(id)
|
|
|
|
// Get file info stored in seperate object
|
|
res, err := store.Service.GetObject(&s3.GetObjectInput{
|
|
Bucket: aws.String(store.Bucket),
|
|
Key: aws.String(uploadId),
|
|
})
|
|
if err == nil {
|
|
// No error occured, and we are able to stream the object
|
|
return res.Body, nil
|
|
}
|
|
|
|
if err, ok := err.(awserr.Error); !ok || err.Code() != "NoSuchKey" {
|
|
return nil, err
|
|
}
|
|
|
|
// Test whether the multipart upload exists to find out if the upload
|
|
// never existsted or just has not been finished yet
|
|
_, err := store.Service.ListParts(&s3.ListPartsInput{
|
|
Bucket: aws.String(store.Bucket),
|
|
Key: aws.String(uploadId),
|
|
UploadId: aws.String(multipartId),
|
|
MaxParts: aws.Int64(0),
|
|
})
|
|
if err == nil {
|
|
// The multipart upload still exists, which means we cannot download it yet
|
|
return nil, errors.New("cannot stream non-finished upload")
|
|
}
|
|
|
|
if err, ok := err.(awserr.Error); ok && err.Code() == "NoSuchUpload" {
|
|
// Neither the object nor the multipart upload exists, so we return a 404
|
|
return nil, tusd.ErrNotFound
|
|
}
|
|
|
|
return nil, err
|
|
}
|
|
|
|
func (store S3Store) Terminate(id string) error {
|
|
uploadId, multipartId := splitIds(id)
|
|
|
|
// Abort the multipart upload first
|
|
_, err := store.Service.AbortMultipartUpload(&s3.AbortMultipartUploadInput{
|
|
Bucket: aws.String(store.Bucket),
|
|
Key: aws.String(uploadId),
|
|
UploadId: aws.String(multipartId),
|
|
})
|
|
if err != nil {
|
|
if err, ok := err.(awserr.Error); ok && err.Code() == "NoSuchUpload" {
|
|
// Test whether the multipart upload exists to find out if the upload
|
|
// never existsted or just has not been finished yet (TODO)
|
|
return tusd.ErrNotFound
|
|
}
|
|
|
|
return err
|
|
}
|
|
|
|
// TODO delete info file
|
|
|
|
return nil
|
|
}
|
|
|
|
func (store S3Store) FinishUpload(id string) error {
|
|
uploadId, multipartId := splitIds(id)
|
|
|
|
println("Finish upload")
|
|
|
|
// Get uploaded parts
|
|
listPtr, err := store.Service.ListParts(&s3.ListPartsInput{
|
|
Bucket: aws.String(store.Bucket),
|
|
Key: aws.String(uploadId),
|
|
UploadId: aws.String(multipartId),
|
|
})
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// Transform the []*s3.Part slice to a []*s3.CompletedPart slice for the next
|
|
// request.
|
|
list := *listPtr
|
|
parts := make([]*s3.CompletedPart, len(list.Parts))
|
|
|
|
for index, part := range list.Parts {
|
|
parts[index] = &s3.CompletedPart{
|
|
ETag: part.ETag,
|
|
PartNumber: part.PartNumber,
|
|
}
|
|
}
|
|
|
|
_, err = store.Service.CompleteMultipartUpload(&s3.CompleteMultipartUploadInput{
|
|
Bucket: aws.String(store.Bucket),
|
|
Key: aws.String(uploadId),
|
|
UploadId: aws.String(multipartId),
|
|
MultipartUpload: &s3.CompletedMultipartUpload{
|
|
Parts: parts,
|
|
},
|
|
})
|
|
|
|
return err
|
|
}
|
|
|
|
func splitIds(id string) (uploadId, multipartId string) {
|
|
index := strings.Index(id, "+")
|
|
if index == -1 {
|
|
return
|
|
}
|
|
|
|
uploadId = id[:index]
|
|
multipartId = id[index+1:]
|
|
return
|
|
}
|