go fmt
This commit is contained in:
@@ -107,12 +107,12 @@ func doFilerBackup(grpcDialOption grpc.DialOption, backupOption *FilerBackupOpti
|
|||||||
|
|
||||||
processEventFn := genProcessFunction(sourcePath, targetPath, dataSink, debug)
|
processEventFn := genProcessFunction(sourcePath, targetPath, dataSink, debug)
|
||||||
|
|
||||||
processEventFnWithOffset := pb.AddOffsetFunc(processEventFn, 3 * time.Second, func(counter int64, lastTsNs int64) error {
|
processEventFnWithOffset := pb.AddOffsetFunc(processEventFn, 3*time.Second, func(counter int64, lastTsNs int64) error {
|
||||||
glog.V(0).Infof("backup %s progressed to %v %0.2f/sec", sourceFiler, time.Unix(0, lastTsNs), float64(counter)/float64(3))
|
glog.V(0).Infof("backup %s progressed to %v %0.2f/sec", sourceFiler, time.Unix(0, lastTsNs), float64(counter)/float64(3))
|
||||||
return setOffset(grpcDialOption, sourceFiler, BackupKeyPrefix, int32(sinkId), lastTsNs)
|
return setOffset(grpcDialOption, sourceFiler, BackupKeyPrefix, int32(sinkId), lastTsNs)
|
||||||
})
|
})
|
||||||
|
|
||||||
return pb.FollowMetadata(sourceFiler, grpcDialOption, "backup_" + dataSink.GetName(),
|
return pb.FollowMetadata(sourceFiler, grpcDialOption, "backup_"+dataSink.GetName(),
|
||||||
sourcePath, startFrom.UnixNano(), 0, processEventFnWithOffset, false)
|
sourcePath, startFrom.UnixNano(), 0, processEventFnWithOffset, false)
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -189,7 +189,7 @@ func (metaBackup *FilerMetaBackupOptions) streamMetadataBackup() error {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
processEventFnWithOffset := pb.AddOffsetFunc(eachEntryFunc, 3 * time.Second, func(counter int64, lastTsNs int64) error {
|
processEventFnWithOffset := pb.AddOffsetFunc(eachEntryFunc, 3*time.Second, func(counter int64, lastTsNs int64) error {
|
||||||
lastTime := time.Unix(0, lastTsNs)
|
lastTime := time.Unix(0, lastTsNs)
|
||||||
glog.V(0).Infof("meta backup %s progressed to %v %0.2f/sec", *metaBackup.filerAddress, lastTime, float64(counter)/float64(3))
|
glog.V(0).Infof("meta backup %s progressed to %v %0.2f/sec", *metaBackup.filerAddress, lastTime, float64(counter)/float64(3))
|
||||||
return metaBackup.setOffset(lastTime)
|
return metaBackup.setOffset(lastTime)
|
||||||
|
|||||||
@@ -252,9 +252,9 @@ func updateLocalEntry(filerClient filer_pb.FilerClient, dir string, entry *filer
|
|||||||
entry.RemoteEntry = remoteEntry
|
entry.RemoteEntry = remoteEntry
|
||||||
return filerClient.WithFilerClient(func(client filer_pb.SeaweedFilerClient) error {
|
return filerClient.WithFilerClient(func(client filer_pb.SeaweedFilerClient) error {
|
||||||
_, err := client.UpdateEntry(context.Background(), &filer_pb.UpdateEntryRequest{
|
_, err := client.UpdateEntry(context.Background(), &filer_pb.UpdateEntryRequest{
|
||||||
Directory: dir,
|
Directory: dir,
|
||||||
Entry: entry,
|
Entry: entry,
|
||||||
})
|
})
|
||||||
return err
|
return err
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -165,12 +165,12 @@ func doSubscribeFilerMetaChanges(grpcDialOption grpc.DialOption, sourceFiler, so
|
|||||||
return persistEventFn(resp)
|
return persistEventFn(resp)
|
||||||
}
|
}
|
||||||
|
|
||||||
processEventFnWithOffset := pb.AddOffsetFunc(processEventFn, 3 * time.Second, func(counter int64, lastTsNs int64) error {
|
processEventFnWithOffset := pb.AddOffsetFunc(processEventFn, 3*time.Second, func(counter int64, lastTsNs int64) error {
|
||||||
glog.V(0).Infof("sync %s to %s progressed to %v %0.2f/sec", sourceFiler, targetFiler, time.Unix(0, lastTsNs), float64(counter)/float64(3))
|
glog.V(0).Infof("sync %s to %s progressed to %v %0.2f/sec", sourceFiler, targetFiler, time.Unix(0, lastTsNs), float64(counter)/float64(3))
|
||||||
return setOffset(grpcDialOption, targetFiler, SyncKeyPrefix, sourceFilerSignature, lastTsNs)
|
return setOffset(grpcDialOption, targetFiler, SyncKeyPrefix, sourceFilerSignature, lastTsNs)
|
||||||
})
|
})
|
||||||
|
|
||||||
return pb.FollowMetadata(sourceFiler, grpcDialOption, "syncTo_" + targetFiler,
|
return pb.FollowMetadata(sourceFiler, grpcDialOption, "syncTo_"+targetFiler,
|
||||||
sourcePath, sourceFilerOffsetTsNs, targetFilerSignature, processEventFnWithOffset, false)
|
sourcePath, sourceFilerOffsetTsNs, targetFilerSignature, processEventFnWithOffset, false)
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -27,5 +27,4 @@ import (
|
|||||||
_ "github.com/chrislusf/seaweedfs/weed/filer/redis"
|
_ "github.com/chrislusf/seaweedfs/weed/filer/redis"
|
||||||
_ "github.com/chrislusf/seaweedfs/weed/filer/redis2"
|
_ "github.com/chrislusf/seaweedfs/weed/filer/redis2"
|
||||||
_ "github.com/chrislusf/seaweedfs/weed/filer/sqlite"
|
_ "github.com/chrislusf/seaweedfs/weed/filer/sqlite"
|
||||||
|
)
|
||||||
)
|
|
||||||
|
|||||||
@@ -160,8 +160,6 @@ func RemoveRemoteStorageMapping(oldContent []byte, dir string) (newContent []byt
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
func ReadMountMappings(grpcDialOption grpc.DialOption, filerAddress string) (mappings *filer_pb.RemoteStorageMapping, readErr error) {
|
func ReadMountMappings(grpcDialOption grpc.DialOption, filerAddress string) (mappings *filer_pb.RemoteStorageMapping, readErr error) {
|
||||||
var oldContent []byte
|
var oldContent []byte
|
||||||
if readErr = pb.WithFilerClient(filerAddress, grpcDialOption, func(client filer_pb.SeaweedFilerClient) error {
|
if readErr = pb.WithFilerClient(filerAddress, grpcDialOption, func(client filer_pb.SeaweedFilerClient) error {
|
||||||
|
|||||||
@@ -31,4 +31,4 @@ func TestFilerRemoteStorage_FindRemoteStorageClient(t *testing.T) {
|
|||||||
|
|
||||||
_, _, found4 := rs.FindRemoteStorageClient("/a/b/cc")
|
_, _, found4 := rs.FindRemoteStorageClient("/a/b/cc")
|
||||||
assert.Equal(t, false, found4, "should not find storage client")
|
assert.Equal(t, false, found4, "should not find storage client")
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -9,7 +9,7 @@ func (entry *Entry) IsInRemoteOnly() bool {
|
|||||||
return len(entry.Chunks) == 0 && entry.Remote != nil && entry.Remote.Size > 0
|
return len(entry.Chunks) == 0 && entry.Remote != nil && entry.Remote.Size > 0
|
||||||
}
|
}
|
||||||
|
|
||||||
func (f *Filer) ReadRemote(entry *Entry, offset int64, size int64) (data[]byte, err error) {
|
func (f *Filer) ReadRemote(entry *Entry, offset int64, size int64) (data []byte, err error) {
|
||||||
client, _, found := f.RemoteStorage.GetRemoteStorageClient(entry.Remote.StorageName)
|
client, _, found := f.RemoteStorage.GetRemoteStorageClient(entry.Remote.StorageName)
|
||||||
if !found {
|
if !found {
|
||||||
return nil, fmt.Errorf("remote storage %v not found", entry.Remote.StorageName)
|
return nil, fmt.Errorf("remote storage %v not found", entry.Remote.StorageName)
|
||||||
|
|||||||
@@ -174,13 +174,13 @@ func toTagging(attributes map[string][]byte) *s3.Tagging {
|
|||||||
|
|
||||||
func (s *s3RemoteStorageClient) readFileRemoteEntry(loc *filer_pb.RemoteStorageLocation) (*filer_pb.RemoteEntry, error) {
|
func (s *s3RemoteStorageClient) readFileRemoteEntry(loc *filer_pb.RemoteStorageLocation) (*filer_pb.RemoteEntry, error) {
|
||||||
resp, err := s.conn.HeadObject(&s3.HeadObjectInput{
|
resp, err := s.conn.HeadObject(&s3.HeadObjectInput{
|
||||||
Bucket: aws.String(loc.Bucket),
|
Bucket: aws.String(loc.Bucket),
|
||||||
Key: aws.String(loc.Path[1:]),
|
Key: aws.String(loc.Path[1:]),
|
||||||
})
|
})
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
|
|
||||||
return &filer_pb.RemoteEntry{
|
return &filer_pb.RemoteEntry{
|
||||||
LastModifiedAt: resp.LastModified.Unix(),
|
LastModifiedAt: resp.LastModified.Unix(),
|
||||||
Size: *resp.ContentLength,
|
Size: *resp.ContentLength,
|
||||||
@@ -200,8 +200,8 @@ func (s *s3RemoteStorageClient) UpdateFileMetadata(loc *filer_pb.RemoteStorageLo
|
|||||||
})
|
})
|
||||||
} else {
|
} else {
|
||||||
_, err = s.conn.DeleteObjectTagging(&s3.DeleteObjectTaggingInput{
|
_, err = s.conn.DeleteObjectTagging(&s3.DeleteObjectTaggingInput{
|
||||||
Bucket: aws.String(loc.Bucket),
|
Bucket: aws.String(loc.Bucket),
|
||||||
Key: aws.String(loc.Path[1:]),
|
Key: aws.String(loc.Path[1:]),
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
return
|
return
|
||||||
|
|||||||
@@ -26,8 +26,8 @@ import (
|
|||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
SequencerType = "master.sequencer.type"
|
SequencerType = "master.sequencer.type"
|
||||||
SequencerEtcdUrls = "master.sequencer.sequencer_etcd_urls"
|
SequencerEtcdUrls = "master.sequencer.sequencer_etcd_urls"
|
||||||
SequencerSnowflakeId = "master.sequencer.sequencer_snowflake_id"
|
SequencerSnowflakeId = "master.sequencer.sequencer_snowflake_id"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|||||||
@@ -18,16 +18,16 @@ func (vs *VolumeServer) FetchAndWriteNeedle(ctx context.Context, req *volume_ser
|
|||||||
}
|
}
|
||||||
|
|
||||||
remoteConf := &filer_pb.RemoteConf{
|
remoteConf := &filer_pb.RemoteConf{
|
||||||
Type: req.RemoteType,
|
Type: req.RemoteType,
|
||||||
Name: req.RemoteName,
|
Name: req.RemoteName,
|
||||||
S3AccessKey: req.S3AccessKey,
|
S3AccessKey: req.S3AccessKey,
|
||||||
S3SecretKey: req.S3SecretKey,
|
S3SecretKey: req.S3SecretKey,
|
||||||
S3Region: req.S3Region,
|
S3Region: req.S3Region,
|
||||||
S3Endpoint: req.S3Endpoint,
|
S3Endpoint: req.S3Endpoint,
|
||||||
}
|
}
|
||||||
|
|
||||||
client, getClientErr := remote_storage.GetRemoteStorage(remoteConf)
|
client, getClientErr := remote_storage.GetRemoteStorage(remoteConf)
|
||||||
if getClientErr != nil {
|
if getClientErr != nil {
|
||||||
return nil, fmt.Errorf("get remote client: %v", getClientErr)
|
return nil, fmt.Errorf("get remote client: %v", getClientErr)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -187,7 +187,7 @@ func TestBalance(t *testing.T) {
|
|||||||
func TestVolumeSelection(t *testing.T) {
|
func TestVolumeSelection(t *testing.T) {
|
||||||
topologyInfo := parseOutput(topoData)
|
topologyInfo := parseOutput(topoData)
|
||||||
|
|
||||||
vids, err := collectVolumeIdsForTierChange(nil, topologyInfo, 1000, types.ToDiskType("hdd"), "", 20.0, 0);
|
vids, err := collectVolumeIdsForTierChange(nil, topologyInfo, 1000, types.ToDiskType("hdd"), "", 20.0, 0)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
t.Errorf("collectVolumeIdsForTierChange: %v", err)
|
t.Errorf("collectVolumeIdsForTierChange: %v", err)
|
||||||
}
|
}
|
||||||
|
|||||||
Reference in New Issue
Block a user