Merge branch 'master' into messaging
This commit is contained in:
@@ -6,6 +6,10 @@ import (
|
|||||||
"errors"
|
"errors"
|
||||||
"flag"
|
"flag"
|
||||||
"fmt"
|
"fmt"
|
||||||
|
"io"
|
||||||
|
"math"
|
||||||
|
"os"
|
||||||
|
|
||||||
"github.com/chrislusf/seaweedfs/weed/glog"
|
"github.com/chrislusf/seaweedfs/weed/glog"
|
||||||
"github.com/chrislusf/seaweedfs/weed/operation"
|
"github.com/chrislusf/seaweedfs/weed/operation"
|
||||||
"github.com/chrislusf/seaweedfs/weed/pb"
|
"github.com/chrislusf/seaweedfs/weed/pb"
|
||||||
@@ -16,9 +20,6 @@ import (
|
|||||||
"github.com/chrislusf/seaweedfs/weed/storage/types"
|
"github.com/chrislusf/seaweedfs/weed/storage/types"
|
||||||
"github.com/chrislusf/seaweedfs/weed/util"
|
"github.com/chrislusf/seaweedfs/weed/util"
|
||||||
"google.golang.org/grpc"
|
"google.golang.org/grpc"
|
||||||
"io"
|
|
||||||
"math"
|
|
||||||
"os"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
var (
|
var (
|
||||||
@@ -155,7 +156,7 @@ func getVolumeFiles(v uint32, addr pb.ServerAddress) (map[types.NeedleId]needleS
|
|||||||
|
|
||||||
var maxOffset int64
|
var maxOffset int64
|
||||||
files := map[types.NeedleId]needleState{}
|
files := map[types.NeedleId]needleState{}
|
||||||
err = idx.WalkIndexFile(idxFile, func(key types.NeedleId, offset types.Offset, size types.Size) error {
|
err = idx.WalkIndexFile(idxFile, 0, func(key types.NeedleId, offset types.Offset, size types.Size) error {
|
||||||
if offset.IsZero() || size.IsDeleted() {
|
if offset.IsZero() || size.IsDeleted() {
|
||||||
files[key] = needleState{
|
files[key] = needleState{
|
||||||
state: stateDeleted,
|
state: stateDeleted,
|
||||||
|
|||||||
@@ -3,11 +3,12 @@ package main
|
|||||||
import (
|
import (
|
||||||
"flag"
|
"flag"
|
||||||
"fmt"
|
"fmt"
|
||||||
"github.com/chrislusf/seaweedfs/weed/util"
|
|
||||||
"os"
|
"os"
|
||||||
"path"
|
"path"
|
||||||
"strconv"
|
"strconv"
|
||||||
|
|
||||||
|
"github.com/chrislusf/seaweedfs/weed/util"
|
||||||
|
|
||||||
"github.com/chrislusf/seaweedfs/weed/glog"
|
"github.com/chrislusf/seaweedfs/weed/glog"
|
||||||
"github.com/chrislusf/seaweedfs/weed/storage/idx"
|
"github.com/chrislusf/seaweedfs/weed/storage/idx"
|
||||||
"github.com/chrislusf/seaweedfs/weed/storage/types"
|
"github.com/chrislusf/seaweedfs/weed/storage/types"
|
||||||
@@ -36,7 +37,7 @@ func main() {
|
|||||||
}
|
}
|
||||||
defer indexFile.Close()
|
defer indexFile.Close()
|
||||||
|
|
||||||
idx.WalkIndexFile(indexFile, func(key types.NeedleId, offset types.Offset, size types.Size) error {
|
idx.WalkIndexFile(indexFile, 0, func(key types.NeedleId, offset types.Offset, size types.Size) error {
|
||||||
fmt.Printf("key:%v offset:%v size:%v(%v)\n", key, offset, size, util.BytesToHumanReadable(uint64(size)))
|
fmt.Printf("key:%v offset:%v size:%v(%v)\n", key, offset, size, util.BytesToHumanReadable(uint64(size)))
|
||||||
return nil
|
return nil
|
||||||
})
|
})
|
||||||
|
|||||||
@@ -32,6 +32,8 @@ type AbstractSqlStore struct {
|
|||||||
dbsLock sync.Mutex
|
dbsLock sync.Mutex
|
||||||
}
|
}
|
||||||
|
|
||||||
|
var _ filer.BucketAware = (*AbstractSqlStore)(nil)
|
||||||
|
|
||||||
func (store *AbstractSqlStore) CanDropWholeBucket() bool {
|
func (store *AbstractSqlStore) CanDropWholeBucket() bool {
|
||||||
return store.SupportBucketTable
|
return store.SupportBucketTable
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -121,23 +121,31 @@ func (store *LevelDB3Store) findDB(fullpath weed_util.FullPath, isForChildren bo
|
|||||||
}
|
}
|
||||||
|
|
||||||
store.dbsLock.RUnlock()
|
store.dbsLock.RUnlock()
|
||||||
// upgrade to write lock
|
|
||||||
|
db, err := store.createDB(bucket)
|
||||||
|
|
||||||
|
return db, bucket, shortPath, err
|
||||||
|
}
|
||||||
|
|
||||||
|
func (store *LevelDB3Store) createDB(bucket string) (*leveldb.DB, error) {
|
||||||
|
|
||||||
store.dbsLock.Lock()
|
store.dbsLock.Lock()
|
||||||
defer store.dbsLock.Unlock()
|
defer store.dbsLock.Unlock()
|
||||||
|
|
||||||
// double check after getting the write lock
|
// double check after getting the write lock
|
||||||
if db, found := store.dbs[bucket]; found {
|
if db, found := store.dbs[bucket]; found {
|
||||||
return db, bucket, shortPath, nil
|
return db, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// create db
|
// create db
|
||||||
db, err := store.loadDB(bucket)
|
db, err := store.loadDB(bucket)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, bucket, shortPath, err
|
return nil, err
|
||||||
}
|
}
|
||||||
|
|
||||||
store.dbs[bucket] = db
|
store.dbs[bucket] = db
|
||||||
|
|
||||||
return db, bucket, shortPath, nil
|
return db, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (store *LevelDB3Store) closeDB(bucket string) {
|
func (store *LevelDB3Store) closeDB(bucket string) {
|
||||||
|
|||||||
23
weed/filer/leveldb3/leveldb3_store_bucket.go
Normal file
23
weed/filer/leveldb3/leveldb3_store_bucket.go
Normal file
@@ -0,0 +1,23 @@
|
|||||||
|
package leveldb
|
||||||
|
|
||||||
|
import (
|
||||||
|
"github.com/chrislusf/seaweedfs/weed/filer"
|
||||||
|
"os"
|
||||||
|
)
|
||||||
|
|
||||||
|
var _ filer.BucketAware = (*LevelDB3Store)(nil)
|
||||||
|
|
||||||
|
func (store *LevelDB3Store) OnBucketCreation(bucket string) {
|
||||||
|
store.createDB(bucket)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (store *LevelDB3Store) OnBucketDeletion(bucket string) {
|
||||||
|
store.closeDB(bucket)
|
||||||
|
if bucket != "" { // just to make sure
|
||||||
|
os.RemoveAll(store.dir + "/" + bucket)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (store *LevelDB3Store) CanDropWholeBucket() bool {
|
||||||
|
return true
|
||||||
|
}
|
||||||
@@ -7,6 +7,7 @@ import (
|
|||||||
"github.com/chrislusf/seaweedfs/weed/pb/master_pb"
|
"github.com/chrislusf/seaweedfs/weed/pb/master_pb"
|
||||||
"github.com/chrislusf/seaweedfs/weed/util"
|
"github.com/chrislusf/seaweedfs/weed/util"
|
||||||
"io"
|
"io"
|
||||||
|
"strings"
|
||||||
"sync"
|
"sync"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
@@ -99,7 +100,11 @@ func (ma *MetaAggregator) loopSubscribeToOneFiler(f *Filer, self pb.ServerAddres
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
if err != nil {
|
if err != nil {
|
||||||
glog.V(0).Infof("subscribing remote %s meta change: %v", peer, err)
|
errLvl := glog.Level(0)
|
||||||
|
if strings.Contains(err.Error(), "duplicated local subscription detected") {
|
||||||
|
errLvl = glog.Level(1)
|
||||||
|
}
|
||||||
|
glog.V(errLvl).Infof("subscribing remote %s meta change: %v", peer, err)
|
||||||
}
|
}
|
||||||
if lastTsNs < nextLastTsNs {
|
if lastTsNs < nextLastTsNs {
|
||||||
lastTsNs = nextLastTsNs
|
lastTsNs = nextLastTsNs
|
||||||
|
|||||||
@@ -18,6 +18,8 @@ const (
|
|||||||
CONNECTION_URL_PATTERN = "%s:%s@tcp(%s:%d)/%s?charset=utf8"
|
CONNECTION_URL_PATTERN = "%s:%s@tcp(%s:%d)/%s?charset=utf8"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
var _ filer.BucketAware = (*MysqlStore2)(nil)
|
||||||
|
|
||||||
func init() {
|
func init() {
|
||||||
filer.Stores = append(filer.Stores, &MysqlStore2{})
|
filer.Stores = append(filer.Stores, &MysqlStore2{})
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -17,6 +17,8 @@ const (
|
|||||||
CONNECTION_URL_PATTERN = "host=%s port=%d sslmode=%s connect_timeout=30"
|
CONNECTION_URL_PATTERN = "host=%s port=%d sslmode=%s connect_timeout=30"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
var _ filer.BucketAware = (*PostgresStore2)(nil)
|
||||||
|
|
||||||
func init() {
|
func init() {
|
||||||
filer.Stores = append(filer.Stores, &PostgresStore2{})
|
filer.Stores = append(filer.Stores, &PostgresStore2{})
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -18,6 +18,12 @@ import (
|
|||||||
"github.com/chrislusf/seaweedfs/weed/wdclient"
|
"github.com/chrislusf/seaweedfs/weed/wdclient"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
var getLookupFileIdBackoffSchedule = []time.Duration{
|
||||||
|
150 * time.Millisecond,
|
||||||
|
600 * time.Millisecond,
|
||||||
|
1800 * time.Millisecond,
|
||||||
|
}
|
||||||
|
|
||||||
func HasData(entry *filer_pb.Entry) bool {
|
func HasData(entry *filer_pb.Entry) bool {
|
||||||
|
|
||||||
if len(entry.Content) > 0 {
|
if len(entry.Content) > 0 {
|
||||||
@@ -69,14 +75,22 @@ func StreamContent(masterClient wdclient.HasLookupFileIdFunction, writer io.Writ
|
|||||||
fileId2Url := make(map[string][]string)
|
fileId2Url := make(map[string][]string)
|
||||||
|
|
||||||
for _, chunkView := range chunkViews {
|
for _, chunkView := range chunkViews {
|
||||||
|
var urlStrings []string
|
||||||
urlStrings, err := masterClient.GetLookupFileIdFunction()(chunkView.FileId)
|
var err error
|
||||||
|
for _, backoff := range getLookupFileIdBackoffSchedule {
|
||||||
|
urlStrings, err = masterClient.GetLookupFileIdFunction()(chunkView.FileId)
|
||||||
|
if err == nil && len(urlStrings) > 0 {
|
||||||
|
time.Sleep(backoff)
|
||||||
|
break
|
||||||
|
}
|
||||||
|
}
|
||||||
if err != nil {
|
if err != nil {
|
||||||
glog.V(1).Infof("operation LookupFileId %s failed, err: %v", chunkView.FileId, err)
|
glog.V(1).Infof("operation LookupFileId %s failed, err: %v", chunkView.FileId, err)
|
||||||
return err
|
return err
|
||||||
} else if len(urlStrings) == 0 {
|
} else if len(urlStrings) == 0 {
|
||||||
glog.Errorf("operation LookupFileId %s failed, err: urls not found", chunkView.FileId)
|
errUrlNotFound := fmt.Errorf("operation LookupFileId %s failed, err: urls not found", chunkView.FileId)
|
||||||
return fmt.Errorf("operation LookupFileId %s failed, err: urls not found", chunkView.FileId)
|
glog.Error(errUrlNotFound)
|
||||||
|
return errUrlNotFound
|
||||||
}
|
}
|
||||||
fileId2Url[chunkView.FileId] = urlStrings
|
fileId2Url[chunkView.FileId] = urlStrings
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -320,6 +320,8 @@ func (store *YdbStore) Shutdown() {
|
|||||||
_ = store.DB.Close(context.Background())
|
_ = store.DB.Close(context.Background())
|
||||||
}
|
}
|
||||||
|
|
||||||
|
var _ filer.BucketAware = (*YdbStore)(nil)
|
||||||
|
|
||||||
func (store *YdbStore) CanDropWholeBucket() bool {
|
func (store *YdbStore) CanDropWholeBucket() bool {
|
||||||
return store.SupportBucketTable
|
return store.SupportBucketTable
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -219,8 +219,16 @@ func (iama *IamApiServer) PutUserPolicy(s3cfg *iam_pb.S3ApiConfiguration, values
|
|||||||
if userName != ident.Name {
|
if userName != ident.Name {
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
|
|
||||||
|
existedActions := make(map[string]bool, len(ident.Actions))
|
||||||
|
for _, action := range ident.Actions {
|
||||||
|
existedActions[action] = true
|
||||||
|
}
|
||||||
|
|
||||||
for _, action := range actions {
|
for _, action := range actions {
|
||||||
ident.Actions = append(ident.Actions, action)
|
if !existedActions[action] {
|
||||||
|
ident.Actions = append(ident.Actions, action)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
return resp, nil
|
return resp, nil
|
||||||
}
|
}
|
||||||
@@ -349,7 +357,8 @@ func (iama *IamApiServer) CreateAccessKey(s3cfg *iam_pb.S3ApiConfiguration, valu
|
|||||||
}
|
}
|
||||||
if !changed {
|
if !changed {
|
||||||
s3cfg.Identities = append(s3cfg.Identities,
|
s3cfg.Identities = append(s3cfg.Identities,
|
||||||
&iam_pb.Identity{Name: userName,
|
&iam_pb.Identity{
|
||||||
|
Name: userName,
|
||||||
Credentials: []*iam_pb.Credential{
|
Credentials: []*iam_pb.Credential{
|
||||||
{
|
{
|
||||||
AccessKey: accessKeyId,
|
AccessKey: accessKeyId,
|
||||||
|
|||||||
@@ -235,7 +235,7 @@ func (wfs *WFS) handleRenameResponse(ctx context.Context, resp *filer_pb.StreamR
|
|||||||
|
|
||||||
sourceInode, targetInode := wfs.inodeToPath.MovePath(oldPath, newPath)
|
sourceInode, targetInode := wfs.inodeToPath.MovePath(oldPath, newPath)
|
||||||
if sourceInode != 0 {
|
if sourceInode != 0 {
|
||||||
if fh, foundFh := wfs.fhmap.inode2fh[sourceInode]; foundFh && fh.entry != nil {
|
if fh, foundFh := wfs.fhmap.FindFileHandle(sourceInode); foundFh && fh.entry != nil {
|
||||||
fh.entry.Name = newName
|
fh.entry.Name = newName
|
||||||
}
|
}
|
||||||
// invalidate attr and data
|
// invalidate attr and data
|
||||||
|
|||||||
@@ -65,8 +65,8 @@ type MasterServer struct {
|
|||||||
|
|
||||||
boundedLeaderChan chan int
|
boundedLeaderChan chan int
|
||||||
|
|
||||||
onPeerUpdatDoneCn chan string
|
onPeerUpdateDoneCn chan string
|
||||||
onPeerUpdatDoneCnExist bool
|
onPeerUpdateDoneCnExist bool
|
||||||
|
|
||||||
// notifying clients
|
// notifying clients
|
||||||
clientChansLock sync.RWMutex
|
clientChansLock sync.RWMutex
|
||||||
@@ -118,7 +118,7 @@ func NewMasterServer(r *mux.Router, option *MasterOption, peers map[string]pb.Se
|
|||||||
Cluster: cluster.NewCluster(),
|
Cluster: cluster.NewCluster(),
|
||||||
}
|
}
|
||||||
ms.boundedLeaderChan = make(chan int, 16)
|
ms.boundedLeaderChan = make(chan int, 16)
|
||||||
ms.onPeerUpdatDoneCn = make(chan string)
|
ms.onPeerUpdateDoneCn = make(chan string)
|
||||||
|
|
||||||
ms.MasterClient.OnPeerUpdate = ms.OnPeerUpdate
|
ms.MasterClient.OnPeerUpdate = ms.OnPeerUpdate
|
||||||
|
|
||||||
@@ -366,14 +366,15 @@ func (ms *MasterServer) OnPeerUpdate(update *master_pb.ClusterNodeUpdate, startF
|
|||||||
hashicorpRaft.ServerAddress(peerAddress.ToGrpcAddress()), 0, 0)
|
hashicorpRaft.ServerAddress(peerAddress.ToGrpcAddress()), 0, 0)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if ms.onPeerUpdatDoneCnExist {
|
if ms.onPeerUpdateDoneCnExist {
|
||||||
ms.onPeerUpdatDoneCn <- peerName
|
ms.onPeerUpdateDoneCn <- peerName
|
||||||
}
|
}
|
||||||
} else if isLeader {
|
} else if isLeader {
|
||||||
go func(peerName string) {
|
go func(peerName string) {
|
||||||
|
raftServerRemovalTimeAfter := time.After(RaftServerRemovalTime)
|
||||||
for {
|
for {
|
||||||
select {
|
select {
|
||||||
case <-time.After(RaftServerRemovalTime):
|
case <-raftServerRemovalTimeAfter:
|
||||||
err := ms.MasterClient.WithClient(false, func(client master_pb.SeaweedClient) error {
|
err := ms.MasterClient.WithClient(false, func(client master_pb.SeaweedClient) error {
|
||||||
_, err := client.RaftRemoveServer(context.Background(), &master_pb.RaftRemoveServerRequest{
|
_, err := client.RaftRemoveServer(context.Background(), &master_pb.RaftRemoveServerRequest{
|
||||||
Id: peerName,
|
Id: peerName,
|
||||||
@@ -384,14 +385,16 @@ func (ms *MasterServer) OnPeerUpdate(update *master_pb.ClusterNodeUpdate, startF
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
glog.Warningf("failed to removing old raft server %s: %v", peerName, err)
|
glog.Warningf("failed to removing old raft server %s: %v", peerName, err)
|
||||||
}
|
}
|
||||||
|
glog.V(0).Infof("old raft server %s removed", peerName)
|
||||||
return
|
return
|
||||||
case peerDone := <-ms.onPeerUpdatDoneCn:
|
case peerDone := <-ms.onPeerUpdateDoneCn:
|
||||||
if peerName == peerDone {
|
if peerName == peerDone {
|
||||||
|
glog.V(0).Infof("raft server %s remove canceled", peerName)
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}(peerName)
|
}(peerName)
|
||||||
ms.onPeerUpdatDoneCnExist = true
|
ms.onPeerUpdateDoneCnExist = true
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -68,45 +68,39 @@ func (c *commandVolumeConfigureReplication) Do(args []string, commandEnv *Comman
|
|||||||
volumeFilter := getVolumeFilter(replicaPlacement, uint32(vid), *collectionPattern)
|
volumeFilter := getVolumeFilter(replicaPlacement, uint32(vid), *collectionPattern)
|
||||||
|
|
||||||
// find all data nodes with volumes that needs replication change
|
// find all data nodes with volumes that needs replication change
|
||||||
var allLocations []location
|
|
||||||
eachDataNode(topologyInfo, func(dc string, rack RackId, dn *master_pb.DataNodeInfo) {
|
eachDataNode(topologyInfo, func(dc string, rack RackId, dn *master_pb.DataNodeInfo) {
|
||||||
loc := newLocation(dc, string(rack), dn)
|
var targetVolumeIds []uint32
|
||||||
for _, diskInfo := range dn.DiskInfos {
|
for _, diskInfo := range dn.DiskInfos {
|
||||||
for _, v := range diskInfo.VolumeInfos {
|
for _, v := range diskInfo.VolumeInfos {
|
||||||
if volumeFilter(v) {
|
if volumeFilter(v) {
|
||||||
allLocations = append(allLocations, loc)
|
targetVolumeIds = append(targetVolumeIds, v.Id)
|
||||||
continue
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
})
|
if len(targetVolumeIds) == 0 {
|
||||||
|
return
|
||||||
if len(allLocations) == 0 {
|
}
|
||||||
return fmt.Errorf("no volume needs change")
|
err = operation.WithVolumeServerClient(false, pb.NewServerAddressFromDataNode(dn), commandEnv.option.GrpcDialOption, func(volumeServerClient volume_server_pb.VolumeServerClient) error {
|
||||||
}
|
for _, targetVolumeId := range targetVolumeIds {
|
||||||
|
resp, configureErr := volumeServerClient.VolumeConfigure(context.Background(), &volume_server_pb.VolumeConfigureRequest{
|
||||||
for _, dst := range allLocations {
|
VolumeId: targetVolumeId,
|
||||||
err := operation.WithVolumeServerClient(false, pb.NewServerAddressFromDataNode(dst.dataNode), commandEnv.option.GrpcDialOption, func(volumeServerClient volume_server_pb.VolumeServerClient) error {
|
Replication: replicaPlacement.String(),
|
||||||
resp, configureErr := volumeServerClient.VolumeConfigure(context.Background(), &volume_server_pb.VolumeConfigureRequest{
|
})
|
||||||
VolumeId: uint32(vid),
|
if configureErr != nil {
|
||||||
Replication: replicaPlacement.String(),
|
return configureErr
|
||||||
})
|
}
|
||||||
if configureErr != nil {
|
if resp.Error != "" {
|
||||||
return configureErr
|
return errors.New(resp.Error)
|
||||||
}
|
}
|
||||||
if resp.Error != "" {
|
|
||||||
return errors.New(resp.Error)
|
|
||||||
}
|
}
|
||||||
return nil
|
return nil
|
||||||
})
|
})
|
||||||
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return
|
||||||
}
|
}
|
||||||
|
})
|
||||||
|
|
||||||
}
|
return err
|
||||||
|
|
||||||
return nil
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func getVolumeFilter(replicaPlacement *super_block.ReplicaPlacement, volumeId uint32, collectionPattern string) func(message *master_pb.VolumeInformationMessage) bool {
|
func getVolumeFilter(replicaPlacement *super_block.ReplicaPlacement, volumeId uint32, collectionPattern string) func(message *master_pb.VolumeInformationMessage) bool {
|
||||||
|
|||||||
@@ -18,7 +18,9 @@ func init() {
|
|||||||
}
|
}
|
||||||
|
|
||||||
type commandVolumeServerEvacuate struct {
|
type commandVolumeServerEvacuate struct {
|
||||||
|
topologyInfo *master_pb.TopologyInfo
|
||||||
targetServer string
|
targetServer string
|
||||||
|
volumeRack string
|
||||||
}
|
}
|
||||||
|
|
||||||
func (c *commandVolumeServerEvacuate) Name() string {
|
func (c *commandVolumeServerEvacuate) Name() string {
|
||||||
@@ -47,7 +49,8 @@ func (c *commandVolumeServerEvacuate) Do(args []string, commandEnv *CommandEnv,
|
|||||||
|
|
||||||
vsEvacuateCommand := flag.NewFlagSet(c.Name(), flag.ContinueOnError)
|
vsEvacuateCommand := flag.NewFlagSet(c.Name(), flag.ContinueOnError)
|
||||||
volumeServer := vsEvacuateCommand.String("node", "", "<host>:<port> of the volume server")
|
volumeServer := vsEvacuateCommand.String("node", "", "<host>:<port> of the volume server")
|
||||||
c.targetServer = *vsEvacuateCommand.String("target", "", "<host>:<port> of target volume")
|
volumeRack := vsEvacuateCommand.String("rack", "", "source rack for the volume servers")
|
||||||
|
targetServer := vsEvacuateCommand.String("target", "", "<host>:<port> of target volume")
|
||||||
skipNonMoveable := vsEvacuateCommand.Bool("skipNonMoveable", false, "skip volumes that can not be moved")
|
skipNonMoveable := vsEvacuateCommand.Bool("skipNonMoveable", false, "skip volumes that can not be moved")
|
||||||
applyChange := vsEvacuateCommand.Bool("force", false, "actually apply the changes")
|
applyChange := vsEvacuateCommand.Bool("force", false, "actually apply the changes")
|
||||||
retryCount := vsEvacuateCommand.Int("retry", 0, "how many times to retry")
|
retryCount := vsEvacuateCommand.Int("retry", 0, "how many times to retry")
|
||||||
@@ -56,12 +59,18 @@ func (c *commandVolumeServerEvacuate) Do(args []string, commandEnv *CommandEnv,
|
|||||||
}
|
}
|
||||||
infoAboutSimulationMode(writer, *applyChange, "-force")
|
infoAboutSimulationMode(writer, *applyChange, "-force")
|
||||||
|
|
||||||
if err = commandEnv.confirmIsLocked(args); err != nil {
|
if err = commandEnv.confirmIsLocked(args); err != nil && *applyChange {
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
if *volumeServer == "" {
|
if *volumeServer == "" && *volumeRack == "" {
|
||||||
return fmt.Errorf("need to specify volume server by -node=<host>:<port>")
|
return fmt.Errorf("need to specify volume server by -node=<host>:<port> or source rack")
|
||||||
|
}
|
||||||
|
if *targetServer != "" {
|
||||||
|
c.targetServer = *targetServer
|
||||||
|
}
|
||||||
|
if *volumeRack != "" {
|
||||||
|
c.volumeRack = *volumeRack
|
||||||
}
|
}
|
||||||
for i := 0; i < *retryCount+1; i++ {
|
for i := 0; i < *retryCount+1; i++ {
|
||||||
if err = c.volumeServerEvacuate(commandEnv, *volumeServer, *skipNonMoveable, *applyChange, writer); err == nil {
|
if err = c.volumeServerEvacuate(commandEnv, *volumeServer, *skipNonMoveable, *applyChange, writer); err == nil {
|
||||||
@@ -80,44 +89,59 @@ func (c *commandVolumeServerEvacuate) volumeServerEvacuate(commandEnv *CommandEn
|
|||||||
|
|
||||||
// list all the volumes
|
// list all the volumes
|
||||||
// collect topology information
|
// collect topology information
|
||||||
topologyInfo, _, err := collectTopologyInfo(commandEnv, 0)
|
c.topologyInfo, _, err = collectTopologyInfo(commandEnv, 0)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
if err := c.evacuateNormalVolumes(commandEnv, topologyInfo, volumeServer, skipNonMoveable, applyChange, writer); err != nil {
|
if err := c.evacuateNormalVolumes(commandEnv, volumeServer, skipNonMoveable, applyChange, writer); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
if err := c.evacuateEcVolumes(commandEnv, topologyInfo, volumeServer, skipNonMoveable, applyChange, writer); err != nil {
|
if err := c.evacuateEcVolumes(commandEnv, volumeServer, skipNonMoveable, applyChange, writer); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (c *commandVolumeServerEvacuate) evacuateNormalVolumes(commandEnv *CommandEnv, topologyInfo *master_pb.TopologyInfo, volumeServer string, skipNonMoveable, applyChange bool, writer io.Writer) error {
|
func (c *commandVolumeServerEvacuate) evacuateNormalVolumes(commandEnv *CommandEnv, volumeServer string, skipNonMoveable, applyChange bool, writer io.Writer) error {
|
||||||
// find this volume server
|
// find this volume server
|
||||||
volumeServers := collectVolumeServersByDc(topologyInfo, "")
|
volumeServers := collectVolumeServersByDc(c.topologyInfo, "")
|
||||||
thisNode, otherNodes := nodesOtherThan(volumeServers, volumeServer)
|
thisNodes, otherNodes := c.nodesOtherThan(volumeServers, volumeServer)
|
||||||
if thisNode == nil {
|
if len(thisNodes) == 0 {
|
||||||
return fmt.Errorf("%s is not found in this cluster", volumeServer)
|
return fmt.Errorf("%s is not found in this cluster", volumeServer)
|
||||||
}
|
}
|
||||||
|
|
||||||
// move away normal volumes
|
// move away normal volumes
|
||||||
volumeReplicas, _ := collectVolumeReplicaLocations(topologyInfo)
|
for _, thisNode := range thisNodes {
|
||||||
for _, diskInfo := range thisNode.info.DiskInfos {
|
for _, diskInfo := range thisNode.info.DiskInfos {
|
||||||
for _, vol := range diskInfo.VolumeInfos {
|
if applyChange {
|
||||||
hasMoved, err := moveAwayOneNormalVolume(commandEnv, volumeReplicas, vol, thisNode, otherNodes, applyChange)
|
if topologyInfo, _, err := collectTopologyInfo(commandEnv, 0); err != nil {
|
||||||
if err != nil {
|
fmt.Fprintf(writer, "update topologyInfo %v", err)
|
||||||
return fmt.Errorf("move away volume %d from %s: %v", vol.Id, volumeServer, err)
|
|
||||||
}
|
|
||||||
if !hasMoved {
|
|
||||||
if skipNonMoveable {
|
|
||||||
replicaPlacement, _ := super_block.NewReplicaPlacementFromByte(byte(vol.ReplicaPlacement))
|
|
||||||
fmt.Fprintf(writer, "skipping non moveable volume %d replication:%s\n", vol.Id, replicaPlacement.String())
|
|
||||||
} else {
|
} else {
|
||||||
return fmt.Errorf("failed to move volume %d from %s", vol.Id, volumeServer)
|
_, otherNodesNew := c.nodesOtherThan(
|
||||||
|
collectVolumeServersByDc(topologyInfo, ""), volumeServer)
|
||||||
|
if len(otherNodesNew) > 0 {
|
||||||
|
otherNodes = otherNodesNew
|
||||||
|
c.topologyInfo = topologyInfo
|
||||||
|
fmt.Fprintf(writer, "topologyInfo updated %v\n", len(otherNodes))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
volumeReplicas, _ := collectVolumeReplicaLocations(c.topologyInfo)
|
||||||
|
for _, vol := range diskInfo.VolumeInfos {
|
||||||
|
hasMoved, err := moveAwayOneNormalVolume(commandEnv, volumeReplicas, vol, thisNode, otherNodes, applyChange)
|
||||||
|
if err != nil {
|
||||||
|
fmt.Fprintf(writer, "move away volume %d from %s: %v", vol.Id, volumeServer, err)
|
||||||
|
}
|
||||||
|
if !hasMoved {
|
||||||
|
if skipNonMoveable {
|
||||||
|
replicaPlacement, _ := super_block.NewReplicaPlacementFromByte(byte(vol.ReplicaPlacement))
|
||||||
|
fmt.Fprintf(writer, "skipping non moveable volume %d replication:%s\n", vol.Id, replicaPlacement.String())
|
||||||
|
} else {
|
||||||
|
return fmt.Errorf("failed to move volume %d from %s", vol.Id, volumeServer)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -125,26 +149,28 @@ func (c *commandVolumeServerEvacuate) evacuateNormalVolumes(commandEnv *CommandE
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (c *commandVolumeServerEvacuate) evacuateEcVolumes(commandEnv *CommandEnv, topologyInfo *master_pb.TopologyInfo, volumeServer string, skipNonMoveable, applyChange bool, writer io.Writer) error {
|
func (c *commandVolumeServerEvacuate) evacuateEcVolumes(commandEnv *CommandEnv, volumeServer string, skipNonMoveable, applyChange bool, writer io.Writer) error {
|
||||||
// find this ec volume server
|
// find this ec volume server
|
||||||
ecNodes, _ := collectEcVolumeServersByDc(topologyInfo, "")
|
ecNodes, _ := collectEcVolumeServersByDc(c.topologyInfo, "")
|
||||||
thisNode, otherNodes := ecNodesOtherThan(ecNodes, volumeServer)
|
thisNodes, otherNodes := c.ecNodesOtherThan(ecNodes, volumeServer)
|
||||||
if thisNode == nil {
|
if len(thisNodes) == 0 {
|
||||||
return fmt.Errorf("%s is not found in this cluster\n", volumeServer)
|
return fmt.Errorf("%s is not found in this cluster\n", volumeServer)
|
||||||
}
|
}
|
||||||
|
|
||||||
// move away ec volumes
|
// move away ec volumes
|
||||||
for _, diskInfo := range thisNode.info.DiskInfos {
|
for _, thisNode := range thisNodes {
|
||||||
for _, ecShardInfo := range diskInfo.EcShardInfos {
|
for _, diskInfo := range thisNode.info.DiskInfos {
|
||||||
hasMoved, err := c.moveAwayOneEcVolume(commandEnv, ecShardInfo, thisNode, otherNodes, applyChange)
|
for _, ecShardInfo := range diskInfo.EcShardInfos {
|
||||||
if err != nil {
|
hasMoved, err := c.moveAwayOneEcVolume(commandEnv, ecShardInfo, thisNode, otherNodes, applyChange)
|
||||||
return fmt.Errorf("move away volume %d from %s: %v", ecShardInfo.Id, volumeServer, err)
|
if err != nil {
|
||||||
}
|
fmt.Fprintf(writer, "move away volume %d from %s: %v", ecShardInfo.Id, volumeServer, err)
|
||||||
if !hasMoved {
|
}
|
||||||
if skipNonMoveable {
|
if !hasMoved {
|
||||||
fmt.Fprintf(writer, "failed to move away ec volume %d from %s\n", ecShardInfo.Id, volumeServer)
|
if skipNonMoveable {
|
||||||
} else {
|
fmt.Fprintf(writer, "failed to move away ec volume %d from %s\n", ecShardInfo.Id, volumeServer)
|
||||||
return fmt.Errorf("failed to move away ec volume %d from %s", ecShardInfo.Id, volumeServer)
|
} else {
|
||||||
|
return fmt.Errorf("failed to move away ec volume %d from %s", ecShardInfo.Id, volumeServer)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -160,9 +186,6 @@ func (c *commandVolumeServerEvacuate) moveAwayOneEcVolume(commandEnv *CommandEnv
|
|||||||
})
|
})
|
||||||
for i := 0; i < len(otherNodes); i++ {
|
for i := 0; i < len(otherNodes); i++ {
|
||||||
emptyNode := otherNodes[i]
|
emptyNode := otherNodes[i]
|
||||||
if c.targetServer != "" && c.targetServer != emptyNode.info.Id {
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
collectionPrefix := ""
|
collectionPrefix := ""
|
||||||
if ecShardInfo.Collection != "" {
|
if ecShardInfo.Collection != "" {
|
||||||
collectionPrefix = ecShardInfo.Collection + "_"
|
collectionPrefix = ecShardInfo.Collection + "_"
|
||||||
@@ -207,10 +230,16 @@ func moveAwayOneNormalVolume(commandEnv *CommandEnv, volumeReplicas map[uint32][
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
func nodesOtherThan(volumeServers []*Node, thisServer string) (thisNode *Node, otherNodes []*Node) {
|
func (c *commandVolumeServerEvacuate) nodesOtherThan(volumeServers []*Node, thisServer string) (thisNodes []*Node, otherNodes []*Node) {
|
||||||
for _, node := range volumeServers {
|
for _, node := range volumeServers {
|
||||||
if node.info.Id == thisServer {
|
if node.info.Id == thisServer || (c.volumeRack != "" && node.rack == c.volumeRack) {
|
||||||
thisNode = node
|
thisNodes = append(thisNodes, node)
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
if c.volumeRack != "" && c.volumeRack == node.rack {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
if c.targetServer != "" && c.targetServer != node.info.Id {
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
otherNodes = append(otherNodes, node)
|
otherNodes = append(otherNodes, node)
|
||||||
@@ -218,10 +247,16 @@ func nodesOtherThan(volumeServers []*Node, thisServer string) (thisNode *Node, o
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
func ecNodesOtherThan(volumeServers []*EcNode, thisServer string) (thisNode *EcNode, otherNodes []*EcNode) {
|
func (c *commandVolumeServerEvacuate) ecNodesOtherThan(volumeServers []*EcNode, thisServer string) (thisNodes []*EcNode, otherNodes []*EcNode) {
|
||||||
for _, node := range volumeServers {
|
for _, node := range volumeServers {
|
||||||
if node.info.Id == thisServer {
|
if node.info.Id == thisServer || (c.volumeRack != "" && string(node.rack) == c.volumeRack) {
|
||||||
thisNode = node
|
thisNodes = append(thisNodes, node)
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
if c.volumeRack != "" && c.volumeRack == string(node.rack) {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
if c.targetServer != "" && c.targetServer != node.info.Id {
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
otherNodes = append(otherNodes, node)
|
otherNodes = append(otherNodes, node)
|
||||||
|
|||||||
@@ -6,12 +6,11 @@ import (
|
|||||||
)
|
)
|
||||||
|
|
||||||
func TestVolumeServerEvacuate(t *testing.T) {
|
func TestVolumeServerEvacuate(t *testing.T) {
|
||||||
topologyInfo := parseOutput(topoData)
|
c := commandVolumeServerEvacuate{}
|
||||||
|
c.topologyInfo = parseOutput(topoData)
|
||||||
|
|
||||||
volumeServer := "192.168.1.4:8080"
|
volumeServer := "192.168.1.4:8080"
|
||||||
|
if err := c.evacuateNormalVolumes(nil, volumeServer, true, false, os.Stdout); err != nil {
|
||||||
c := commandVolumeServerEvacuate{}
|
|
||||||
if err := c.evacuateNormalVolumes(nil, topologyInfo, volumeServer, true, false, os.Stdout); err != nil {
|
|
||||||
t.Errorf("evacuate: %v", err)
|
t.Errorf("evacuate: %v", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -294,7 +294,7 @@ func readNeedleMap(baseFileName string) (*needle_map.MemDb, error) {
|
|||||||
defer indexFile.Close()
|
defer indexFile.Close()
|
||||||
|
|
||||||
cm := needle_map.NewMemDb()
|
cm := needle_map.NewMemDb()
|
||||||
err = idx.WalkIndexFile(indexFile, func(key types.NeedleId, offset types.Offset, size types.Size) error {
|
err = idx.WalkIndexFile(indexFile, 0, func(key types.NeedleId, offset types.Offset, size types.Size) error {
|
||||||
if !offset.IsZero() && size != types.TombstoneFileSize {
|
if !offset.IsZero() && size != types.TombstoneFileSize {
|
||||||
cm.Set(key, offset, size)
|
cm.Set(key, offset, size)
|
||||||
} else {
|
} else {
|
||||||
|
|||||||
@@ -9,8 +9,8 @@ import (
|
|||||||
|
|
||||||
// walks through the index file, calls fn function with each key, offset, size
|
// walks through the index file, calls fn function with each key, offset, size
|
||||||
// stops with the error returned by the fn function
|
// stops with the error returned by the fn function
|
||||||
func WalkIndexFile(r io.ReaderAt, fn func(key types.NeedleId, offset types.Offset, size types.Size) error) error {
|
func WalkIndexFile(r io.ReaderAt, startFrom uint64, fn func(key types.NeedleId, offset types.Offset, size types.Size) error) error {
|
||||||
var readerOffset int64
|
readerOffset := int64(startFrom * types.NeedleMapEntrySize)
|
||||||
bytes := make([]byte, types.NeedleMapEntrySize*RowsToRead)
|
bytes := make([]byte, types.NeedleMapEntrySize*RowsToRead)
|
||||||
count, e := r.ReadAt(bytes, readerOffset)
|
count, e := r.ReadAt(bytes, readerOffset)
|
||||||
if count == 0 && e == io.EOF {
|
if count == 0 && e == io.EOF {
|
||||||
|
|||||||
@@ -111,7 +111,7 @@ func (cm *MemDb) LoadFromIdx(idxName string) (ret error) {
|
|||||||
|
|
||||||
func (cm *MemDb) LoadFromReaderAt(readerAt io.ReaderAt) (ret error) {
|
func (cm *MemDb) LoadFromReaderAt(readerAt io.ReaderAt) (ret error) {
|
||||||
|
|
||||||
return idx.WalkIndexFile(readerAt, func(key NeedleId, offset Offset, size Size) error {
|
return idx.WalkIndexFile(readerAt, 0, func(key NeedleId, offset Offset, size Size) error {
|
||||||
if offset.IsZero() || size.IsDeleted() {
|
if offset.IsZero() || size.IsDeleted() {
|
||||||
return cm.Delete(key)
|
return cm.Delete(key)
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -9,6 +9,8 @@ import (
|
|||||||
"github.com/syndtr/goleveldb/leveldb/opt"
|
"github.com/syndtr/goleveldb/leveldb/opt"
|
||||||
|
|
||||||
"github.com/chrislusf/seaweedfs/weed/storage/idx"
|
"github.com/chrislusf/seaweedfs/weed/storage/idx"
|
||||||
|
"github.com/chrislusf/seaweedfs/weed/storage/types"
|
||||||
|
"github.com/chrislusf/seaweedfs/weed/util"
|
||||||
|
|
||||||
"github.com/syndtr/goleveldb/leveldb"
|
"github.com/syndtr/goleveldb/leveldb"
|
||||||
|
|
||||||
@@ -17,10 +19,16 @@ import (
|
|||||||
. "github.com/chrislusf/seaweedfs/weed/storage/types"
|
. "github.com/chrislusf/seaweedfs/weed/storage/types"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
//mark it every watermarkBatchSize operations
|
||||||
|
const watermarkBatchSize = 10000
|
||||||
|
|
||||||
|
var watermarkKey = []byte("idx_entry_watermark")
|
||||||
|
|
||||||
type LevelDbNeedleMap struct {
|
type LevelDbNeedleMap struct {
|
||||||
baseNeedleMapper
|
baseNeedleMapper
|
||||||
dbFileName string
|
dbFileName string
|
||||||
db *leveldb.DB
|
db *leveldb.DB
|
||||||
|
recordCount uint64
|
||||||
}
|
}
|
||||||
|
|
||||||
func NewLevelDbNeedleMap(dbFileName string, indexFile *os.File, opts *opt.Options) (m *LevelDbNeedleMap, err error) {
|
func NewLevelDbNeedleMap(dbFileName string, indexFile *os.File, opts *opt.Options) (m *LevelDbNeedleMap, err error) {
|
||||||
@@ -46,7 +54,14 @@ func NewLevelDbNeedleMap(dbFileName string, indexFile *os.File, opts *opt.Option
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
glog.V(1).Infof("Loading %s...", indexFile.Name())
|
glog.V(0).Infof("Loading %s... , watermark: %d", dbFileName, getWatermark(m.db))
|
||||||
|
m.recordCount = uint64(m.indexFileOffset / types.NeedleMapEntrySize)
|
||||||
|
watermark := (m.recordCount / watermarkBatchSize) * watermarkBatchSize
|
||||||
|
err = setWatermark(m.db, watermark)
|
||||||
|
if err != nil {
|
||||||
|
glog.Fatalf("set watermark for %s error: %s\n", dbFileName, err)
|
||||||
|
return
|
||||||
|
}
|
||||||
mm, indexLoadError := newNeedleMapMetricFromIndexFile(indexFile)
|
mm, indexLoadError := newNeedleMapMetricFromIndexFile(indexFile)
|
||||||
if indexLoadError != nil {
|
if indexLoadError != nil {
|
||||||
return nil, indexLoadError
|
return nil, indexLoadError
|
||||||
@@ -78,9 +93,20 @@ func generateLevelDbFile(dbFileName string, indexFile *os.File) error {
|
|||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
defer db.Close()
|
defer db.Close()
|
||||||
return idx.WalkIndexFile(indexFile, func(key NeedleId, offset Offset, size Size) error {
|
|
||||||
|
watermark := getWatermark(db)
|
||||||
|
if stat, err := indexFile.Stat(); err != nil {
|
||||||
|
glog.Fatalf("stat file %s: %v", indexFile.Name(), err)
|
||||||
|
return err
|
||||||
|
} else {
|
||||||
|
if watermark*types.NeedleMapEntrySize > uint64(stat.Size()) {
|
||||||
|
glog.Warningf("wrong watermark %d for filesize %d", watermark, stat.Size())
|
||||||
|
}
|
||||||
|
glog.V(0).Infof("generateLevelDbFile %s, watermark %d, num of entries:%d", dbFileName, watermark, (uint64(stat.Size())-watermark*types.NeedleMapEntrySize)/types.NeedleMapEntrySize)
|
||||||
|
}
|
||||||
|
return idx.WalkIndexFile(indexFile, watermark, func(key NeedleId, offset Offset, size Size) error {
|
||||||
if !offset.IsZero() && size.IsValid() {
|
if !offset.IsZero() && size.IsValid() {
|
||||||
levelDbWrite(db, key, offset, size)
|
levelDbWrite(db, key, offset, size, false, 0)
|
||||||
} else {
|
} else {
|
||||||
levelDbDelete(db, key)
|
levelDbDelete(db, key)
|
||||||
}
|
}
|
||||||
@@ -102,6 +128,7 @@ func (m *LevelDbNeedleMap) Get(key NeedleId) (element *needle_map.NeedleValue, o
|
|||||||
|
|
||||||
func (m *LevelDbNeedleMap) Put(key NeedleId, offset Offset, size Size) error {
|
func (m *LevelDbNeedleMap) Put(key NeedleId, offset Offset, size Size) error {
|
||||||
var oldSize Size
|
var oldSize Size
|
||||||
|
var watermark uint64
|
||||||
if oldNeedle, ok := m.Get(key); ok {
|
if oldNeedle, ok := m.Get(key); ok {
|
||||||
oldSize = oldNeedle.Size
|
oldSize = oldNeedle.Size
|
||||||
}
|
}
|
||||||
@@ -110,16 +137,54 @@ func (m *LevelDbNeedleMap) Put(key NeedleId, offset Offset, size Size) error {
|
|||||||
if err := m.appendToIndexFile(key, offset, size); err != nil {
|
if err := m.appendToIndexFile(key, offset, size); err != nil {
|
||||||
return fmt.Errorf("cannot write to indexfile %s: %v", m.indexFile.Name(), err)
|
return fmt.Errorf("cannot write to indexfile %s: %v", m.indexFile.Name(), err)
|
||||||
}
|
}
|
||||||
return levelDbWrite(m.db, key, offset, size)
|
m.recordCount++
|
||||||
|
if m.recordCount%watermarkBatchSize != 0 {
|
||||||
|
watermark = 0
|
||||||
|
} else {
|
||||||
|
watermark = (m.recordCount / watermarkBatchSize) * watermarkBatchSize
|
||||||
|
glog.V(1).Infof("put cnt:%d for %s,watermark: %d", m.recordCount, m.dbFileName, watermark)
|
||||||
|
}
|
||||||
|
return levelDbWrite(m.db, key, offset, size, watermark == 0, watermark)
|
||||||
}
|
}
|
||||||
|
|
||||||
func levelDbWrite(db *leveldb.DB, key NeedleId, offset Offset, size Size) error {
|
func getWatermark(db *leveldb.DB) uint64 {
|
||||||
|
data, err := db.Get(watermarkKey, nil)
|
||||||
|
if err != nil || len(data) != 8 {
|
||||||
|
glog.Warningf("get watermark from db error: %v, %d", err, len(data))
|
||||||
|
/*
|
||||||
|
if !strings.Contains(strings.ToLower(err.Error()), "not found") {
|
||||||
|
err = setWatermark(db, 0)
|
||||||
|
if err != nil {
|
||||||
|
glog.Errorf("failed to set watermark: %v", err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
*/
|
||||||
|
return 0
|
||||||
|
}
|
||||||
|
return util.BytesToUint64(data)
|
||||||
|
}
|
||||||
|
|
||||||
|
func setWatermark(db *leveldb.DB, watermark uint64) error {
|
||||||
|
glog.V(1).Infof("set watermark %d", watermark)
|
||||||
|
var wmBytes = make([]byte, 8)
|
||||||
|
util.Uint64toBytes(wmBytes, watermark)
|
||||||
|
if err := db.Put(watermarkKey, wmBytes, nil); err != nil {
|
||||||
|
return fmt.Errorf("failed to setWatermark: %v", err)
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func levelDbWrite(db *leveldb.DB, key NeedleId, offset Offset, size Size, updateWatermark bool, watermark uint64) error {
|
||||||
|
|
||||||
bytes := needle_map.ToBytes(key, offset, size)
|
bytes := needle_map.ToBytes(key, offset, size)
|
||||||
|
|
||||||
if err := db.Put(bytes[0:NeedleIdSize], bytes[NeedleIdSize:NeedleIdSize+OffsetSize+SizeSize], nil); err != nil {
|
if err := db.Put(bytes[0:NeedleIdSize], bytes[NeedleIdSize:NeedleIdSize+OffsetSize+SizeSize], nil); err != nil {
|
||||||
return fmt.Errorf("failed to write leveldb: %v", err)
|
return fmt.Errorf("failed to write leveldb: %v", err)
|
||||||
}
|
}
|
||||||
|
// set watermark
|
||||||
|
if updateWatermark {
|
||||||
|
return setWatermark(db, watermark)
|
||||||
|
}
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
func levelDbDelete(db *leveldb.DB, key NeedleId) error {
|
func levelDbDelete(db *leveldb.DB, key NeedleId) error {
|
||||||
@@ -129,6 +194,7 @@ func levelDbDelete(db *leveldb.DB, key NeedleId) error {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func (m *LevelDbNeedleMap) Delete(key NeedleId, offset Offset) error {
|
func (m *LevelDbNeedleMap) Delete(key NeedleId, offset Offset) error {
|
||||||
|
var watermark uint64
|
||||||
oldNeedle, found := m.Get(key)
|
oldNeedle, found := m.Get(key)
|
||||||
if !found || oldNeedle.Size.IsDeleted() {
|
if !found || oldNeedle.Size.IsDeleted() {
|
||||||
return nil
|
return nil
|
||||||
@@ -139,8 +205,13 @@ func (m *LevelDbNeedleMap) Delete(key NeedleId, offset Offset) error {
|
|||||||
if err := m.appendToIndexFile(key, offset, TombstoneFileSize); err != nil {
|
if err := m.appendToIndexFile(key, offset, TombstoneFileSize); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
m.recordCount++
|
||||||
return levelDbWrite(m.db, key, oldNeedle.Offset, -oldNeedle.Size)
|
if m.recordCount%watermarkBatchSize != 0 {
|
||||||
|
watermark = 0
|
||||||
|
} else {
|
||||||
|
watermark = (m.recordCount / watermarkBatchSize) * watermarkBatchSize
|
||||||
|
}
|
||||||
|
return levelDbWrite(m.db, key, oldNeedle.Offset, -oldNeedle.Size, watermark == 0, watermark)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (m *LevelDbNeedleMap) Close() {
|
func (m *LevelDbNeedleMap) Close() {
|
||||||
|
|||||||
@@ -33,7 +33,7 @@ func LoadCompactNeedleMap(file *os.File) (*NeedleMap, error) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func doLoading(file *os.File, nm *NeedleMap) (*NeedleMap, error) {
|
func doLoading(file *os.File, nm *NeedleMap) (*NeedleMap, error) {
|
||||||
e := idx.WalkIndexFile(file, func(key NeedleId, offset Offset, size Size) error {
|
e := idx.WalkIndexFile(file, 0, func(key NeedleId, offset Offset, size Size) error {
|
||||||
nm.MaybeSetMaxFileKey(key)
|
nm.MaybeSetMaxFileKey(key)
|
||||||
if !offset.IsZero() && size.IsValid() {
|
if !offset.IsZero() && size.IsValid() {
|
||||||
nm.FileCounter++
|
nm.FileCounter++
|
||||||
|
|||||||
@@ -32,7 +32,7 @@ func Retry(name string, job func() error) (err error) {
|
|||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
func RetryForever(name string, job func() error, onErrFn func(err error) bool) {
|
func RetryForever(name string, job func() error, onErrFn func(err error) (shouldContinue bool)) {
|
||||||
waitTime := time.Second
|
waitTime := time.Second
|
||||||
for {
|
for {
|
||||||
err := job()
|
err := job()
|
||||||
|
|||||||
@@ -2,6 +2,7 @@ package wdclient
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
|
"fmt"
|
||||||
"github.com/chrislusf/seaweedfs/weed/stats"
|
"github.com/chrislusf/seaweedfs/weed/stats"
|
||||||
"math/rand"
|
"math/rand"
|
||||||
"time"
|
"time"
|
||||||
@@ -46,7 +47,7 @@ func (mc *MasterClient) GetLookupFileIdFunction() LookupFileIdFunctionType {
|
|||||||
|
|
||||||
func (mc *MasterClient) LookupFileIdWithFallback(fileId string) (fullUrls []string, err error) {
|
func (mc *MasterClient) LookupFileIdWithFallback(fileId string) (fullUrls []string, err error) {
|
||||||
fullUrls, err = mc.vidMap.LookupFileId(fileId)
|
fullUrls, err = mc.vidMap.LookupFileId(fileId)
|
||||||
if err == nil {
|
if err == nil && len(fullUrls) > 0 {
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
err = pb.WithMasterClient(false, mc.currentMaster, mc.grpcDialOption, func(client master_pb.SeaweedClient) error {
|
err = pb.WithMasterClient(false, mc.currentMaster, mc.grpcDialOption, func(client master_pb.SeaweedClient) error {
|
||||||
@@ -54,7 +55,7 @@ func (mc *MasterClient) LookupFileIdWithFallback(fileId string) (fullUrls []stri
|
|||||||
VolumeOrFileIds: []string{fileId},
|
VolumeOrFileIds: []string{fileId},
|
||||||
})
|
})
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return fmt.Errorf("LookupVolume failed: %v", err)
|
||||||
}
|
}
|
||||||
for vid, vidLocation := range resp.VolumeIdLocations {
|
for vid, vidLocation := range resp.VolumeIdLocations {
|
||||||
for _, vidLoc := range vidLocation.Locations {
|
for _, vidLoc := range vidLocation.Locations {
|
||||||
@@ -67,7 +68,6 @@ func (mc *MasterClient) LookupFileIdWithFallback(fileId string) (fullUrls []stri
|
|||||||
fullUrls = append(fullUrls, "http://"+loc.Url+"/"+fileId)
|
fullUrls = append(fullUrls, "http://"+loc.Url+"/"+fileId)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
return nil
|
return nil
|
||||||
})
|
})
|
||||||
return
|
return
|
||||||
|
|||||||
Reference in New Issue
Block a user