Chris Lu
6 years ago
8 changed files with 369 additions and 46 deletions
-
17weed/command/filer_replication.go
-
14weed/command/scaffold.go
-
29weed/replication/replicator.go
-
33weed/replication/sink/filersink/filer_sink.go
-
12weed/replication/sink/replication_sink.go
-
123weed/replication/sink/s3sink/s3_sink.go
-
165weed/replication/sink/s3sink/s3_write.go
-
22weed/replication/source/filer_source.go
@ -0,0 +1,123 @@ |
|||||
|
package S3Sink |
||||
|
|
||||
|
import ( |
||||
|
"fmt" |
||||
|
|
||||
|
"github.com/chrislusf/seaweedfs/weed/pb/filer_pb" |
||||
|
"github.com/chrislusf/seaweedfs/weed/replication/source" |
||||
|
"github.com/chrislusf/seaweedfs/weed/util" |
||||
|
"github.com/aws/aws-sdk-go/service/s3/s3iface" |
||||
|
"github.com/aws/aws-sdk-go/service/s3" |
||||
|
"github.com/aws/aws-sdk-go/aws/session" |
||||
|
"sync" |
||||
|
"github.com/chrislusf/seaweedfs/weed/filer2" |
||||
|
"github.com/chrislusf/seaweedfs/weed/replication/sink" |
||||
|
"github.com/aws/aws-sdk-go/aws" |
||||
|
"github.com/aws/aws-sdk-go/aws/credentials" |
||||
|
) |
||||
|
|
||||
|
type S3Sink struct { |
||||
|
err error |
||||
|
conn s3iface.S3API |
||||
|
region string |
||||
|
bucket string |
||||
|
dir string |
||||
|
filerSource *source.FilerSource |
||||
|
} |
||||
|
|
||||
|
func init() { |
||||
|
sink.Sinks = append(sink.Sinks, &S3Sink{}) |
||||
|
} |
||||
|
|
||||
|
func (s3sink *S3Sink) GetName() string { |
||||
|
return "filer" |
||||
|
} |
||||
|
|
||||
|
func (s3sink *S3Sink) GetSinkToDirectory() string { |
||||
|
return s3sink.dir |
||||
|
} |
||||
|
|
||||
|
func (s3sink *S3Sink) Initialize(configuration util.Configuration) error { |
||||
|
return s3sink.initialize( |
||||
|
configuration.GetString("aws_access_key_id"), |
||||
|
configuration.GetString("aws_secret_access_key"), |
||||
|
configuration.GetString("region"), |
||||
|
configuration.GetString("bucket"), |
||||
|
configuration.GetString("directory"), |
||||
|
) |
||||
|
} |
||||
|
|
||||
|
func (s3sink *S3Sink) SetSourceFiler(s *source.FilerSource) { |
||||
|
s3sink.filerSource = s |
||||
|
} |
||||
|
|
||||
|
func (s3sink *S3Sink) initialize(awsAccessKeyId, aswSecretAccessKey, region, bucket, dir string) (error) { |
||||
|
s3sink.region = region |
||||
|
s3sink.bucket = bucket |
||||
|
s3sink.dir = dir |
||||
|
|
||||
|
config := &aws.Config{ |
||||
|
Region: aws.String(s3sink.region), |
||||
|
} |
||||
|
if awsAccessKeyId != "" && aswSecretAccessKey != "" { |
||||
|
config.Credentials = credentials.NewStaticCredentials(awsAccessKeyId, aswSecretAccessKey, "") |
||||
|
} |
||||
|
|
||||
|
sess, err := session.NewSession(config) |
||||
|
if err != nil { |
||||
|
return fmt.Errorf("create aws session: %v", err) |
||||
|
} |
||||
|
s3sink.conn = s3.New(sess) |
||||
|
|
||||
|
return nil |
||||
|
} |
||||
|
|
||||
|
func (s3sink *S3Sink) DeleteEntry(key string, isDirectory, deleteIncludeChunks bool) error { |
||||
|
|
||||
|
if isDirectory { |
||||
|
key = key + "/" |
||||
|
} |
||||
|
|
||||
|
return s3sink.deleteObject(key) |
||||
|
|
||||
|
} |
||||
|
|
||||
|
func (s3sink *S3Sink) CreateEntry(key string, entry *filer_pb.Entry) error { |
||||
|
|
||||
|
uploadId, err := s3sink.createMultipartUpload(key, entry) |
||||
|
if err != nil { |
||||
|
return err |
||||
|
} |
||||
|
|
||||
|
totalSize := filer2.TotalSize(entry.Chunks) |
||||
|
chunkViews := filer2.ViewFromChunks(entry.Chunks, 0, int(totalSize)) |
||||
|
|
||||
|
var parts []*s3.CompletedPart |
||||
|
var wg sync.WaitGroup |
||||
|
for chunkIndex, chunk := range chunkViews { |
||||
|
partId := chunkIndex + 1 |
||||
|
wg.Add(1) |
||||
|
go func(chunk *filer2.ChunkView) { |
||||
|
defer wg.Done() |
||||
|
if part, uploadErr := s3sink.uploadPart(key, uploadId, partId, chunk); uploadErr != nil { |
||||
|
err = uploadErr |
||||
|
} else { |
||||
|
parts = append(parts, part) |
||||
|
} |
||||
|
}(chunk) |
||||
|
} |
||||
|
wg.Wait() |
||||
|
|
||||
|
if err != nil { |
||||
|
s3sink.abortMultipartUpload(key, uploadId) |
||||
|
return err |
||||
|
} |
||||
|
|
||||
|
return s3sink.completeMultipartUpload(key, uploadId, parts) |
||||
|
|
||||
|
} |
||||
|
|
||||
|
func (s3sink *S3Sink) UpdateEntry(key string, oldEntry, newEntry *filer_pb.Entry, deleteIncludeChunks bool) (foundExistingEntry bool, err error) { |
||||
|
// TODO improve efficiency
|
||||
|
return false, nil |
||||
|
} |
@ -0,0 +1,165 @@ |
|||||
|
package S3Sink |
||||
|
|
||||
|
import ( |
||||
|
"github.com/aws/aws-sdk-go/service/s3" |
||||
|
"github.com/aws/aws-sdk-go/aws" |
||||
|
"github.com/chrislusf/seaweedfs/weed/glog" |
||||
|
"github.com/aws/aws-sdk-go/aws/awserr" |
||||
|
"fmt" |
||||
|
"io" |
||||
|
"github.com/chrislusf/seaweedfs/weed/pb/filer_pb" |
||||
|
"github.com/chrislusf/seaweedfs/weed/filer2" |
||||
|
"github.com/chrislusf/seaweedfs/weed/util" |
||||
|
"bytes" |
||||
|
) |
||||
|
|
||||
|
func (s3sink *S3Sink) deleteObject(key string) error { |
||||
|
input := &s3.DeleteObjectInput{ |
||||
|
Bucket: aws.String(s3sink.bucket), |
||||
|
Key: aws.String(key), |
||||
|
} |
||||
|
|
||||
|
result, err := s3sink.conn.DeleteObject(input) |
||||
|
|
||||
|
if err == nil { |
||||
|
glog.V(0).Infof("[%s] delete %s: %v", s3sink.bucket, key, result) |
||||
|
} else { |
||||
|
glog.Errorf("[%s] delete %s: %v", s3sink.bucket, key, err) |
||||
|
} |
||||
|
|
||||
|
return err |
||||
|
|
||||
|
} |
||||
|
|
||||
|
func (s3sink *S3Sink) createMultipartUpload(key string, entry *filer_pb.Entry) (uploadId string, err error) { |
||||
|
input := &s3.CreateMultipartUploadInput{ |
||||
|
Bucket: aws.String(s3sink.bucket), |
||||
|
Key: aws.String(key), |
||||
|
ContentType: aws.String(entry.Attributes.Mime), |
||||
|
} |
||||
|
|
||||
|
result, err := s3sink.conn.CreateMultipartUpload(input) |
||||
|
|
||||
|
if err == nil { |
||||
|
glog.V(0).Infof("[%s] createMultipartUpload %s: %v", s3sink.bucket, key, result) |
||||
|
} else { |
||||
|
glog.Errorf("[%s] createMultipartUpload %s: %v", s3sink.bucket, key, err) |
||||
|
return "", err |
||||
|
} |
||||
|
|
||||
|
return *result.UploadId, err |
||||
|
} |
||||
|
|
||||
|
func (s3sink *S3Sink) abortMultipartUpload(key, uploadId string) error { |
||||
|
input := &s3.AbortMultipartUploadInput{ |
||||
|
Bucket: aws.String(s3sink.bucket), |
||||
|
Key: aws.String(key), |
||||
|
UploadId: aws.String(uploadId), |
||||
|
} |
||||
|
|
||||
|
result, err := s3sink.conn.AbortMultipartUpload(input) |
||||
|
if err != nil { |
||||
|
if aerr, ok := err.(awserr.Error); ok { |
||||
|
switch aerr.Code() { |
||||
|
case s3.ErrCodeNoSuchUpload: |
||||
|
glog.Errorf("[%s] abortMultipartUpload %s: %v %v", s3sink.bucket, key, s3.ErrCodeNoSuchUpload, aerr.Error()) |
||||
|
default: |
||||
|
glog.Errorf("[%s] abortMultipartUpload %s: %v", s3sink.bucket, key, aerr.Error()) |
||||
|
} |
||||
|
} else { |
||||
|
// Print the error, cast err to awserr.Error to get the Code and
|
||||
|
// Message from an error.
|
||||
|
glog.Errorf("[%s] abortMultipartUpload %s: %v", s3sink.bucket, key, aerr.Error()) |
||||
|
} |
||||
|
return err |
||||
|
} |
||||
|
|
||||
|
glog.V(0).Infof("[%s] abortMultipartUpload %s: %v", s3sink.bucket, key, result) |
||||
|
|
||||
|
return nil |
||||
|
} |
||||
|
|
||||
|
// To complete multipart upload
|
||||
|
func (s3sink *S3Sink) completeMultipartUpload(key, uploadId string, parts []*s3.CompletedPart) error { |
||||
|
input := &s3.CompleteMultipartUploadInput{ |
||||
|
Bucket: aws.String(s3sink.bucket), |
||||
|
Key: aws.String(key), |
||||
|
UploadId: aws.String(uploadId), |
||||
|
MultipartUpload: &s3.CompletedMultipartUpload{ |
||||
|
Parts: parts, |
||||
|
}, |
||||
|
} |
||||
|
|
||||
|
result, err := s3sink.conn.CompleteMultipartUpload(input) |
||||
|
if err == nil { |
||||
|
glog.V(0).Infof("[%s] completeMultipartUpload %s: %v", s3sink.bucket, key, result) |
||||
|
} else { |
||||
|
glog.Errorf("[%s] completeMultipartUpload %s: %v", s3sink.bucket, key, err) |
||||
|
} |
||||
|
|
||||
|
return err |
||||
|
} |
||||
|
|
||||
|
// To upload a part
|
||||
|
func (s3sink *S3Sink) uploadPart(key, uploadId string, partId int, chunk *filer2.ChunkView) (*s3.CompletedPart, error) { |
||||
|
var readSeeker io.ReadSeeker |
||||
|
|
||||
|
readSeeker, err := s3sink.buildReadSeeker(chunk) |
||||
|
if err != nil { |
||||
|
glog.Errorf("[%s] uploadPart %s %d read: %v", s3sink.bucket, key, partId, err) |
||||
|
return nil, fmt.Errorf("[%s] uploadPart %s %d read: %v", s3sink.bucket, key, partId, err) |
||||
|
} |
||||
|
|
||||
|
input := &s3.UploadPartInput{ |
||||
|
Body: readSeeker, |
||||
|
Bucket: aws.String(s3sink.bucket), |
||||
|
Key: aws.String(key), |
||||
|
PartNumber: aws.Int64(int64(partId)), |
||||
|
UploadId: aws.String(uploadId), |
||||
|
} |
||||
|
|
||||
|
result, err := s3sink.conn.UploadPart(input) |
||||
|
if err == nil { |
||||
|
glog.V(0).Infof("[%s] uploadPart %s %d upload: %v", s3sink.bucket, key, partId, result) |
||||
|
} else { |
||||
|
glog.Errorf("[%s] uploadPart %s %d upload: %v", s3sink.bucket, key, partId, err) |
||||
|
} |
||||
|
|
||||
|
part := &s3.CompletedPart{ |
||||
|
ETag: result.ETag, |
||||
|
PartNumber: aws.Int64(int64(partId)), |
||||
|
} |
||||
|
|
||||
|
return part, err |
||||
|
} |
||||
|
|
||||
|
// To upload a part by copying byte range from an existing object as data source
|
||||
|
func (s3sink *S3Sink) uploadPartCopy(key, uploadId string, partId int64, copySource string, sourceStart, sourceStop int) error { |
||||
|
input := &s3.UploadPartCopyInput{ |
||||
|
Bucket: aws.String(s3sink.bucket), |
||||
|
CopySource: aws.String(fmt.Sprintf("/%s/%s", s3sink.bucket, copySource)), |
||||
|
CopySourceRange: aws.String(fmt.Sprintf("bytes=%d-%d", sourceStart, sourceStop)), |
||||
|
Key: aws.String(key), |
||||
|
PartNumber: aws.Int64(partId), |
||||
|
UploadId: aws.String(uploadId), |
||||
|
} |
||||
|
|
||||
|
result, err := s3sink.conn.UploadPartCopy(input) |
||||
|
if err == nil { |
||||
|
glog.V(0).Infof("[%s] uploadPartCopy %s %d: %v", s3sink.bucket, key, partId, result) |
||||
|
} else { |
||||
|
glog.Errorf("[%s] uploadPartCopy %s %d: %v", s3sink.bucket, key, partId, err) |
||||
|
} |
||||
|
|
||||
|
return err |
||||
|
} |
||||
|
|
||||
|
func (s3sink *S3Sink) buildReadSeeker(chunk *filer2.ChunkView) (io.ReadSeeker, error) { |
||||
|
fileUrl, err := s3sink.filerSource.LookupFileId(chunk.FileId) |
||||
|
if err != nil { |
||||
|
return nil, err |
||||
|
} |
||||
|
buf := make([]byte, chunk.Size) |
||||
|
util.ReadUrl(fileUrl, chunk.Offset, int(chunk.Size), buf) |
||||
|
return bytes.NewReader(buf), nil |
||||
|
} |
Write
Preview
Loading…
Cancel
Save
Reference in new issue