Refine Bucket Size Metrics: Logical and Physical Size (#7943)
* refactor: implement logical size calculation with replication factor using dedicated helper * ui: update bucket list to show logical/physical size
This commit is contained in:
@@ -23,6 +23,7 @@ import (
|
|||||||
"github.com/seaweedfs/seaweedfs/weed/pb/mq_pb"
|
"github.com/seaweedfs/seaweedfs/weed/pb/mq_pb"
|
||||||
"github.com/seaweedfs/seaweedfs/weed/pb/schema_pb"
|
"github.com/seaweedfs/seaweedfs/weed/pb/schema_pb"
|
||||||
"github.com/seaweedfs/seaweedfs/weed/security"
|
"github.com/seaweedfs/seaweedfs/weed/security"
|
||||||
|
"github.com/seaweedfs/seaweedfs/weed/storage/super_block"
|
||||||
"github.com/seaweedfs/seaweedfs/weed/util"
|
"github.com/seaweedfs/seaweedfs/weed/util"
|
||||||
"github.com/seaweedfs/seaweedfs/weed/wdclient"
|
"github.com/seaweedfs/seaweedfs/weed/wdclient"
|
||||||
"google.golang.org/grpc"
|
"google.golang.org/grpc"
|
||||||
@@ -96,6 +97,11 @@ type AdminServer struct {
|
|||||||
|
|
||||||
// Worker gRPC server
|
// Worker gRPC server
|
||||||
workerGrpcServer *WorkerGrpcServer
|
workerGrpcServer *WorkerGrpcServer
|
||||||
|
|
||||||
|
// Collection statistics caching
|
||||||
|
collectionStatsCache map[string]collectionStats
|
||||||
|
lastCollectionStatsUpdate time.Time
|
||||||
|
collectionStatsCacheThreshold time.Duration
|
||||||
}
|
}
|
||||||
|
|
||||||
// Type definitions moved to types.go
|
// Type definitions moved to types.go
|
||||||
@@ -119,13 +125,14 @@ func NewAdminServer(masters string, templateFS http.FileSystem, dataDir string)
|
|||||||
go masterClient.KeepConnectedToMaster(ctx)
|
go masterClient.KeepConnectedToMaster(ctx)
|
||||||
|
|
||||||
server := &AdminServer{
|
server := &AdminServer{
|
||||||
masterClient: masterClient,
|
masterClient: masterClient,
|
||||||
templateFS: templateFS,
|
templateFS: templateFS,
|
||||||
dataDir: dataDir,
|
dataDir: dataDir,
|
||||||
grpcDialOption: grpcDialOption,
|
grpcDialOption: grpcDialOption,
|
||||||
cacheExpiration: 10 * time.Second,
|
cacheExpiration: 10 * time.Second,
|
||||||
filerCacheExpiration: 30 * time.Second, // Cache filers for 30 seconds
|
filerCacheExpiration: 30 * time.Second, // Cache filers for 30 seconds
|
||||||
configPersistence: NewConfigPersistence(dataDir),
|
configPersistence: NewConfigPersistence(dataDir),
|
||||||
|
collectionStatsCacheThreshold: 30 * time.Second,
|
||||||
}
|
}
|
||||||
|
|
||||||
// Initialize topic retention purger
|
// Initialize topic retention purger
|
||||||
@@ -236,57 +243,32 @@ func (s *AdminServer) GetCredentialManager() *credential.CredentialManager {
|
|||||||
|
|
||||||
// InvalidateCache method moved to cluster_topology.go
|
// InvalidateCache method moved to cluster_topology.go
|
||||||
|
|
||||||
|
// GetS3BucketsData retrieves all Object Store buckets and aggregates total storage metrics
|
||||||
|
func (s *AdminServer) GetS3BucketsData() (S3BucketsData, error) {
|
||||||
|
buckets, err := s.GetS3Buckets()
|
||||||
|
if err != nil {
|
||||||
|
return S3BucketsData{}, err
|
||||||
|
}
|
||||||
|
|
||||||
|
var totalSize int64
|
||||||
|
for _, bucket := range buckets {
|
||||||
|
totalSize += bucket.PhysicalSize
|
||||||
|
}
|
||||||
|
|
||||||
|
return S3BucketsData{
|
||||||
|
Buckets: buckets,
|
||||||
|
TotalBuckets: len(buckets),
|
||||||
|
TotalSize: totalSize,
|
||||||
|
LastUpdated: time.Now(),
|
||||||
|
}, nil
|
||||||
|
}
|
||||||
|
|
||||||
// GetS3Buckets retrieves all Object Store buckets from the filer and collects size/object data from collections
|
// GetS3Buckets retrieves all Object Store buckets from the filer and collects size/object data from collections
|
||||||
func (s *AdminServer) GetS3Buckets() ([]S3Bucket, error) {
|
func (s *AdminServer) GetS3Buckets() ([]S3Bucket, error) {
|
||||||
var buckets []S3Bucket
|
var buckets []S3Bucket
|
||||||
|
|
||||||
// Build a map of collection name to collection data
|
// Collect volume information by collection with caching
|
||||||
collectionMap := make(map[string]struct {
|
collectionMap, _ := s.getCollectionStats()
|
||||||
Size int64
|
|
||||||
FileCount int64
|
|
||||||
})
|
|
||||||
|
|
||||||
// Collect volume information by collection
|
|
||||||
err := s.WithMasterClient(func(client master_pb.SeaweedClient) error {
|
|
||||||
resp, err := client.VolumeList(context.Background(), &master_pb.VolumeListRequest{})
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
|
|
||||||
if resp.TopologyInfo != nil {
|
|
||||||
for _, dc := range resp.TopologyInfo.DataCenterInfos {
|
|
||||||
for _, rack := range dc.RackInfos {
|
|
||||||
for _, node := range rack.DataNodeInfos {
|
|
||||||
for _, diskInfo := range node.DiskInfos {
|
|
||||||
for _, volInfo := range diskInfo.VolumeInfos {
|
|
||||||
collection := volInfo.Collection
|
|
||||||
if collection == "" {
|
|
||||||
collection = "default"
|
|
||||||
}
|
|
||||||
|
|
||||||
if _, exists := collectionMap[collection]; !exists {
|
|
||||||
collectionMap[collection] = struct {
|
|
||||||
Size int64
|
|
||||||
FileCount int64
|
|
||||||
}{}
|
|
||||||
}
|
|
||||||
|
|
||||||
data := collectionMap[collection]
|
|
||||||
data.Size += int64(volInfo.Size)
|
|
||||||
data.FileCount += int64(volInfo.FileCount)
|
|
||||||
collectionMap[collection] = data
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
return nil
|
|
||||||
})
|
|
||||||
|
|
||||||
if err != nil {
|
|
||||||
return nil, fmt.Errorf("failed to get volume information: %w", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
// Get filer configuration (buckets path and filer group)
|
// Get filer configuration (buckets path and filer group)
|
||||||
filerConfig, err := s.getFilerConfig()
|
filerConfig, err := s.getFilerConfig()
|
||||||
@@ -324,10 +306,12 @@ func (s *AdminServer) GetS3Buckets() ([]S3Bucket, error) {
|
|||||||
collectionName := getCollectionName(filerConfig.FilerGroup, bucketName)
|
collectionName := getCollectionName(filerConfig.FilerGroup, bucketName)
|
||||||
|
|
||||||
// Get size and object count from collection data
|
// Get size and object count from collection data
|
||||||
var size int64
|
var physicalSize int64
|
||||||
|
var logicalSize int64
|
||||||
var objectCount int64
|
var objectCount int64
|
||||||
if collectionData, exists := collectionMap[collectionName]; exists {
|
if collectionData, exists := collectionMap[collectionName]; exists {
|
||||||
size = collectionData.Size
|
physicalSize = collectionData.PhysicalSize
|
||||||
|
logicalSize = collectionData.LogicalSize
|
||||||
objectCount = collectionData.FileCount
|
objectCount = collectionData.FileCount
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -363,7 +347,8 @@ func (s *AdminServer) GetS3Buckets() ([]S3Bucket, error) {
|
|||||||
bucket := S3Bucket{
|
bucket := S3Bucket{
|
||||||
Name: bucketName,
|
Name: bucketName,
|
||||||
CreatedAt: time.Unix(resp.Entry.Attributes.Crtime, 0),
|
CreatedAt: time.Unix(resp.Entry.Attributes.Crtime, 0),
|
||||||
Size: size,
|
LogicalSize: logicalSize,
|
||||||
|
PhysicalSize: physicalSize,
|
||||||
ObjectCount: objectCount,
|
ObjectCount: objectCount,
|
||||||
LastModified: time.Unix(resp.Entry.Attributes.Mtime, 0),
|
LastModified: time.Unix(resp.Entry.Attributes.Mtime, 0),
|
||||||
Quota: quota,
|
Quota: quota,
|
||||||
@@ -389,6 +374,7 @@ func (s *AdminServer) GetS3Buckets() ([]S3Bucket, error) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// GetBucketDetails retrieves detailed information about a specific bucket
|
// GetBucketDetails retrieves detailed information about a specific bucket
|
||||||
|
// Note: This no longer lists objects for performance reasons. Use GetS3Buckets for size/count data.
|
||||||
func (s *AdminServer) GetBucketDetails(bucketName string) (*BucketDetails, error) {
|
func (s *AdminServer) GetBucketDetails(bucketName string) (*BucketDetails, error) {
|
||||||
// Get filer configuration (buckets path)
|
// Get filer configuration (buckets path)
|
||||||
filerConfig, err := s.getFilerConfig()
|
filerConfig, err := s.getFilerConfig()
|
||||||
@@ -396,16 +382,25 @@ func (s *AdminServer) GetBucketDetails(bucketName string) (*BucketDetails, error
|
|||||||
glog.Warningf("Failed to get filer configuration, using defaults: %v", err)
|
glog.Warningf("Failed to get filer configuration, using defaults: %v", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
bucketPath := fmt.Sprintf("%s/%s", filerConfig.BucketsPath, bucketName)
|
|
||||||
|
|
||||||
details := &BucketDetails{
|
details := &BucketDetails{
|
||||||
Bucket: S3Bucket{
|
Bucket: S3Bucket{
|
||||||
Name: bucketName,
|
Name: bucketName,
|
||||||
},
|
},
|
||||||
Objects: []S3Object{},
|
|
||||||
UpdatedAt: time.Now(),
|
UpdatedAt: time.Now(),
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Get collection data for size and object count with caching
|
||||||
|
collectionName := getCollectionName(filerConfig.FilerGroup, bucketName)
|
||||||
|
stats, err := s.getCollectionStats()
|
||||||
|
if err != nil {
|
||||||
|
glog.Warningf("Failed to get collection data: %v", err)
|
||||||
|
// Continue without collection data - use zero values
|
||||||
|
} else if data, ok := stats[collectionName]; ok {
|
||||||
|
details.Bucket.LogicalSize = data.LogicalSize
|
||||||
|
details.Bucket.PhysicalSize = data.PhysicalSize
|
||||||
|
details.Bucket.ObjectCount = data.FileCount
|
||||||
|
}
|
||||||
|
|
||||||
err = s.WithFilerClient(func(client filer_pb.SeaweedFilerClient) error {
|
err = s.WithFilerClient(func(client filer_pb.SeaweedFilerClient) error {
|
||||||
// Get bucket info
|
// Get bucket info
|
||||||
bucketResp, err := client.LookupDirectoryEntry(context.Background(), &filer_pb.LookupDirectoryEntryRequest{
|
bucketResp, err := client.LookupDirectoryEntry(context.Background(), &filer_pb.LookupDirectoryEntryRequest{
|
||||||
@@ -456,8 +451,7 @@ func (s *AdminServer) GetBucketDetails(bucketName string) (*BucketDetails, error
|
|||||||
details.Bucket.ObjectLockDuration = objectLockDuration
|
details.Bucket.ObjectLockDuration = objectLockDuration
|
||||||
details.Bucket.Owner = owner
|
details.Bucket.Owner = owner
|
||||||
|
|
||||||
// List objects in bucket (recursively)
|
return nil
|
||||||
return s.listBucketObjects(client, bucketPath, bucketPath, "", details)
|
|
||||||
})
|
})
|
||||||
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@@ -467,106 +461,6 @@ func (s *AdminServer) GetBucketDetails(bucketName string) (*BucketDetails, error
|
|||||||
return details, nil
|
return details, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// listBucketObjects recursively lists all objects in a bucket
|
|
||||||
// bucketBasePath is the full path to the bucket (e.g., /buckets/mybucket)
|
|
||||||
func (s *AdminServer) listBucketObjects(client filer_pb.SeaweedFilerClient, bucketBasePath, directory, prefix string, details *BucketDetails) error {
|
|
||||||
stream, err := client.ListEntries(context.Background(), &filer_pb.ListEntriesRequest{
|
|
||||||
Directory: directory,
|
|
||||||
Prefix: prefix,
|
|
||||||
StartFromFileName: "",
|
|
||||||
InclusiveStartFrom: false,
|
|
||||||
Limit: 1000,
|
|
||||||
})
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
|
|
||||||
for {
|
|
||||||
resp, err := stream.Recv()
|
|
||||||
if err != nil {
|
|
||||||
if err.Error() == "EOF" {
|
|
||||||
break
|
|
||||||
}
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
|
|
||||||
entry := resp.Entry
|
|
||||||
if entry.IsDirectory {
|
|
||||||
// Check if this is a .versions directory (represents a versioned object)
|
|
||||||
if strings.HasSuffix(entry.Name, ".versions") {
|
|
||||||
// This directory represents an object, add it as an object without the .versions suffix
|
|
||||||
objectName := strings.TrimSuffix(entry.Name, ".versions")
|
|
||||||
objectKey := objectName
|
|
||||||
if directory != bucketBasePath {
|
|
||||||
relativePath := directory[len(bucketBasePath)+1:]
|
|
||||||
objectKey = fmt.Sprintf("%s/%s", relativePath, objectName)
|
|
||||||
}
|
|
||||||
|
|
||||||
// Extract latest version metadata from extended attributes
|
|
||||||
var size int64 = 0
|
|
||||||
var mtime int64 = entry.Attributes.Mtime
|
|
||||||
if entry.Extended != nil {
|
|
||||||
// Get size of latest version
|
|
||||||
if sizeBytes, ok := entry.Extended[s3_constants.ExtLatestVersionSizeKey]; ok && len(sizeBytes) == 8 {
|
|
||||||
size = int64(util.BytesToUint64(sizeBytes))
|
|
||||||
}
|
|
||||||
// Get mtime of latest version
|
|
||||||
if mtimeBytes, ok := entry.Extended[s3_constants.ExtLatestVersionMtimeKey]; ok && len(mtimeBytes) == 8 {
|
|
||||||
mtime = int64(util.BytesToUint64(mtimeBytes))
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
obj := S3Object{
|
|
||||||
Key: objectKey,
|
|
||||||
Size: size,
|
|
||||||
LastModified: time.Unix(mtime, 0),
|
|
||||||
ETag: "",
|
|
||||||
StorageClass: "STANDARD",
|
|
||||||
}
|
|
||||||
|
|
||||||
details.Objects = append(details.Objects, obj)
|
|
||||||
details.TotalCount++
|
|
||||||
details.TotalSize += size
|
|
||||||
// Don't recurse into .versions directories
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
|
|
||||||
// Recursively list subdirectories
|
|
||||||
subDir := fmt.Sprintf("%s/%s", directory, entry.Name)
|
|
||||||
err := s.listBucketObjects(client, bucketBasePath, subDir, "", details)
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
// Add file object
|
|
||||||
objectKey := entry.Name
|
|
||||||
if directory != bucketBasePath {
|
|
||||||
// Remove bucket prefix to get relative path
|
|
||||||
relativePath := directory[len(bucketBasePath)+1:]
|
|
||||||
objectKey = fmt.Sprintf("%s/%s", relativePath, entry.Name)
|
|
||||||
}
|
|
||||||
|
|
||||||
obj := S3Object{
|
|
||||||
Key: objectKey,
|
|
||||||
Size: int64(entry.Attributes.FileSize),
|
|
||||||
LastModified: time.Unix(entry.Attributes.Mtime, 0),
|
|
||||||
ETag: "", // Could be calculated from chunks if needed
|
|
||||||
StorageClass: "STANDARD",
|
|
||||||
}
|
|
||||||
|
|
||||||
details.Objects = append(details.Objects, obj)
|
|
||||||
details.TotalSize += obj.Size
|
|
||||||
details.TotalCount++
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// Update bucket totals
|
|
||||||
details.Bucket.Size = details.TotalSize
|
|
||||||
details.Bucket.ObjectCount = details.TotalCount
|
|
||||||
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// CreateS3Bucket creates a new S3 bucket
|
// CreateS3Bucket creates a new S3 bucket
|
||||||
func (s *AdminServer) CreateS3Bucket(bucketName string) error {
|
func (s *AdminServer) CreateS3Bucket(bucketName string) error {
|
||||||
return s.CreateS3BucketWithQuota(bucketName, 0, false)
|
return s.CreateS3BucketWithQuota(bucketName, 0, false)
|
||||||
@@ -2108,3 +2002,66 @@ func getBoolFromMap(m map[string]interface{}, key string) bool {
|
|||||||
}
|
}
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
|
|
||||||
|
type collectionStats struct {
|
||||||
|
PhysicalSize int64
|
||||||
|
LogicalSize int64
|
||||||
|
FileCount int64
|
||||||
|
}
|
||||||
|
|
||||||
|
func collectCollectionStats(topologyInfo *master_pb.TopologyInfo) map[string]collectionStats {
|
||||||
|
collectionMap := make(map[string]collectionStats)
|
||||||
|
for _, dc := range topologyInfo.DataCenterInfos {
|
||||||
|
for _, rack := range dc.RackInfos {
|
||||||
|
for _, node := range rack.DataNodeInfos {
|
||||||
|
for _, diskInfo := range node.DiskInfos {
|
||||||
|
for _, volInfo := range diskInfo.VolumeInfos {
|
||||||
|
collection := volInfo.Collection
|
||||||
|
if collection == "" {
|
||||||
|
collection = "default"
|
||||||
|
}
|
||||||
|
|
||||||
|
data := collectionMap[collection]
|
||||||
|
data.PhysicalSize += int64(volInfo.Size)
|
||||||
|
rp, _ := super_block.NewReplicaPlacementFromByte(byte(volInfo.ReplicaPlacement))
|
||||||
|
// NewReplicaPlacementFromByte never returns a nil rp. If there's an error,
|
||||||
|
// it returns a zero-valued ReplicaPlacement, for which GetCopyCount() is 1.
|
||||||
|
// This provides a safe fallback, so we can ignore the error.
|
||||||
|
replicaCount := int64(rp.GetCopyCount())
|
||||||
|
if volInfo.Size >= volInfo.DeletedByteCount {
|
||||||
|
data.LogicalSize += int64(volInfo.Size-volInfo.DeletedByteCount) / replicaCount
|
||||||
|
}
|
||||||
|
if volInfo.FileCount >= volInfo.DeleteCount {
|
||||||
|
data.FileCount += int64(volInfo.FileCount-volInfo.DeleteCount) / replicaCount
|
||||||
|
}
|
||||||
|
collectionMap[collection] = data
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return collectionMap
|
||||||
|
}
|
||||||
|
|
||||||
|
// getCollectionStats returns current collection statistics with caching
|
||||||
|
func (s *AdminServer) getCollectionStats() (map[string]collectionStats, error) {
|
||||||
|
now := time.Now()
|
||||||
|
if s.collectionStatsCache != nil && now.Sub(s.lastCollectionStatsUpdate) < s.collectionStatsCacheThreshold {
|
||||||
|
return s.collectionStatsCache, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
err := s.WithMasterClient(func(client master_pb.SeaweedClient) error {
|
||||||
|
resp, err := client.VolumeList(context.Background(), &master_pb.VolumeListRequest{})
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
if resp.TopologyInfo != nil {
|
||||||
|
s.collectionStatsCache = collectCollectionStats(resp.TopologyInfo)
|
||||||
|
s.lastCollectionStatsUpdate = now
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
})
|
||||||
|
|
||||||
|
return s.collectionStatsCache, err
|
||||||
|
}
|
||||||
|
|||||||
@@ -48,26 +48,13 @@ type CreateBucketRequest struct {
|
|||||||
func (s *AdminServer) ShowS3Buckets(c *gin.Context) {
|
func (s *AdminServer) ShowS3Buckets(c *gin.Context) {
|
||||||
username := c.GetString("username")
|
username := c.GetString("username")
|
||||||
|
|
||||||
buckets, err := s.GetS3Buckets()
|
data, err := s.GetS3BucketsData()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
c.JSON(http.StatusInternalServerError, gin.H{"error": "Failed to get Object Store buckets: " + err.Error()})
|
c.JSON(http.StatusInternalServerError, gin.H{"error": "Failed to get Object Store buckets: " + err.Error()})
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
// Calculate totals
|
data.Username = username
|
||||||
var totalSize int64
|
|
||||||
for _, bucket := range buckets {
|
|
||||||
totalSize += bucket.Size
|
|
||||||
}
|
|
||||||
|
|
||||||
data := S3BucketsData{
|
|
||||||
Username: username,
|
|
||||||
Buckets: buckets,
|
|
||||||
TotalBuckets: len(buckets),
|
|
||||||
TotalSize: totalSize,
|
|
||||||
LastUpdated: time.Now(),
|
|
||||||
}
|
|
||||||
|
|
||||||
c.JSON(http.StatusOK, data)
|
c.JSON(http.StatusOK, data)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -120,8 +120,10 @@ func (s *AdminServer) getTopologyViaGRPC(topology *ClusterTopology) error {
|
|||||||
|
|
||||||
// InvalidateCache forces a refresh of cached data
|
// InvalidateCache forces a refresh of cached data
|
||||||
func (s *AdminServer) InvalidateCache() {
|
func (s *AdminServer) InvalidateCache() {
|
||||||
s.lastCacheUpdate = time.Time{}
|
s.lastCacheUpdate = time.Now().Add(-s.cacheExpiration)
|
||||||
s.cachedTopology = nil
|
s.cachedTopology = nil
|
||||||
s.lastFilerUpdate = time.Time{}
|
s.lastFilerUpdate = time.Now().Add(-s.filerCacheExpiration)
|
||||||
s.cachedFilers = nil
|
s.cachedFilers = nil
|
||||||
|
s.lastCollectionStatsUpdate = time.Now().Add(-s.collectionStatsCacheThreshold)
|
||||||
|
s.collectionStatsCache = nil
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -73,7 +73,8 @@ type VolumeServerEcInfo struct {
|
|||||||
type S3Bucket struct {
|
type S3Bucket struct {
|
||||||
Name string `json:"name"`
|
Name string `json:"name"`
|
||||||
CreatedAt time.Time `json:"created_at"`
|
CreatedAt time.Time `json:"created_at"`
|
||||||
Size int64 `json:"size"`
|
LogicalSize int64 `json:"logical_size"` // Actual data size (used space)
|
||||||
|
PhysicalSize int64 `json:"physical_size"` // Total allocated volume space
|
||||||
ObjectCount int64 `json:"object_count"`
|
ObjectCount int64 `json:"object_count"`
|
||||||
LastModified time.Time `json:"last_modified"`
|
LastModified time.Time `json:"last_modified"`
|
||||||
Quota int64 `json:"quota"` // Quota in bytes, 0 means no quota
|
Quota int64 `json:"quota"` // Quota in bytes, 0 means no quota
|
||||||
@@ -94,11 +95,8 @@ type S3Object struct {
|
|||||||
}
|
}
|
||||||
|
|
||||||
type BucketDetails struct {
|
type BucketDetails struct {
|
||||||
Bucket S3Bucket `json:"bucket"`
|
Bucket S3Bucket `json:"bucket"`
|
||||||
Objects []S3Object `json:"objects"`
|
UpdatedAt time.Time `json:"updated_at"`
|
||||||
TotalSize int64 `json:"total_size"`
|
|
||||||
TotalCount int64 `json:"total_count"`
|
|
||||||
UpdatedAt time.Time `json:"updated_at"`
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// ObjectStoreUser is defined in admin_data.go
|
// ObjectStoreUser is defined in admin_data.go
|
||||||
|
|||||||
@@ -409,8 +409,8 @@ func (h *AdminHandlers) getS3BucketsData(c *gin.Context) dash.S3BucketsData {
|
|||||||
username = "admin"
|
username = "admin"
|
||||||
}
|
}
|
||||||
|
|
||||||
// Get Object Store buckets
|
// Get Object Store buckets data
|
||||||
buckets, err := h.adminServer.GetS3Buckets()
|
data, err := h.adminServer.GetS3BucketsData()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
// Return empty data on error
|
// Return empty data on error
|
||||||
return dash.S3BucketsData{
|
return dash.S3BucketsData{
|
||||||
@@ -422,19 +422,8 @@ func (h *AdminHandlers) getS3BucketsData(c *gin.Context) dash.S3BucketsData {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Calculate totals
|
data.Username = username
|
||||||
var totalSize int64
|
return data
|
||||||
for _, bucket := range buckets {
|
|
||||||
totalSize += bucket.Size
|
|
||||||
}
|
|
||||||
|
|
||||||
return dash.S3BucketsData{
|
|
||||||
Username: username,
|
|
||||||
Buckets: buckets,
|
|
||||||
TotalBuckets: len(buckets),
|
|
||||||
TotalSize: totalSize,
|
|
||||||
LastUpdated: time.Now(),
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// getAdminData retrieves admin data from the server (now uses consolidated method)
|
// getAdminData retrieves admin data from the server (now uses consolidated method)
|
||||||
|
|||||||
@@ -116,7 +116,8 @@ templ S3Buckets(data dash.S3BucketsData) {
|
|||||||
<th>Owner</th>
|
<th>Owner</th>
|
||||||
<th>Created</th>
|
<th>Created</th>
|
||||||
<th>Objects</th>
|
<th>Objects</th>
|
||||||
<th>Size</th>
|
<th>Logical Size</th>
|
||||||
|
<th>Physical Size</th>
|
||||||
<th>Quota</th>
|
<th>Quota</th>
|
||||||
<th>Versioning</th>
|
<th>Versioning</th>
|
||||||
<th>Object Lock</th>
|
<th>Object Lock</th>
|
||||||
@@ -127,7 +128,7 @@ templ S3Buckets(data dash.S3BucketsData) {
|
|||||||
for _, bucket := range data.Buckets {
|
for _, bucket := range data.Buckets {
|
||||||
<tr>
|
<tr>
|
||||||
<td>
|
<td>
|
||||||
<a href={templ.SafeURL(fmt.Sprintf("/files?path=/buckets/%s", bucket.Name))}
|
<a href={templ.SafeURL(fmt.Sprintf("/files?path=/buckets/%s", bucket.Name))}
|
||||||
class="text-decoration-none">
|
class="text-decoration-none">
|
||||||
<i class="fas fa-cube me-2"></i>
|
<i class="fas fa-cube me-2"></i>
|
||||||
{bucket.Name}
|
{bucket.Name}
|
||||||
@@ -144,16 +145,24 @@ templ S3Buckets(data dash.S3BucketsData) {
|
|||||||
</td>
|
</td>
|
||||||
<td>{bucket.CreatedAt.Format("2006-01-02 15:04")}</td>
|
<td>{bucket.CreatedAt.Format("2006-01-02 15:04")}</td>
|
||||||
<td>{fmt.Sprintf("%d", bucket.ObjectCount)}</td>
|
<td>{fmt.Sprintf("%d", bucket.ObjectCount)}</td>
|
||||||
<td>{formatBytes(bucket.Size)}</td>
|
<td>
|
||||||
|
<div>{formatBytes(bucket.LogicalSize)}</div>
|
||||||
|
if bucket.PhysicalSize > 0 && bucket.LogicalSize > 0 && bucket.PhysicalSize > bucket.LogicalSize {
|
||||||
|
<div class="small text-muted">
|
||||||
|
{fmt.Sprintf("%.1fx overhead", float64(bucket.PhysicalSize)/float64(bucket.LogicalSize))}
|
||||||
|
</div>
|
||||||
|
}
|
||||||
|
</td>
|
||||||
|
<td>{formatBytes(bucket.PhysicalSize)}</td>
|
||||||
<td>
|
<td>
|
||||||
if bucket.Quota > 0 {
|
if bucket.Quota > 0 {
|
||||||
<div>
|
<div>
|
||||||
<span class={fmt.Sprintf("badge bg-%s", getQuotaStatusColor(bucket.Size, bucket.Quota, bucket.QuotaEnabled))}>
|
<span class={fmt.Sprintf("badge bg-%s", getQuotaStatusColor(bucket.LogicalSize, bucket.Quota, bucket.QuotaEnabled))}>
|
||||||
{formatBytes(bucket.Quota)}
|
{formatBytes(bucket.Quota)}
|
||||||
</span>
|
</span>
|
||||||
if bucket.QuotaEnabled {
|
if bucket.QuotaEnabled {
|
||||||
<div class="small text-muted">
|
<div class="small text-muted">
|
||||||
{fmt.Sprintf("%.1f%% used", float64(bucket.Size)/float64(bucket.Quota)*100)}
|
{fmt.Sprintf("%.1f%% used", float64(bucket.LogicalSize)/float64(bucket.Quota)*100)}
|
||||||
</div>
|
</div>
|
||||||
} else {
|
} else {
|
||||||
<div class="small text-muted">Disabled</div>
|
<div class="small text-muted">Disabled</div>
|
||||||
@@ -192,7 +201,7 @@ templ S3Buckets(data dash.S3BucketsData) {
|
|||||||
</td>
|
</td>
|
||||||
<td>
|
<td>
|
||||||
<div class="btn-group btn-group-sm" role="group">
|
<div class="btn-group btn-group-sm" role="group">
|
||||||
<a href={templ.SafeURL(fmt.Sprintf("/files?path=/buckets/%s", bucket.Name))}
|
<a href={templ.SafeURL(fmt.Sprintf("/files?path=/buckets/%s", bucket.Name))}
|
||||||
class="btn btn-outline-success btn-sm"
|
class="btn btn-outline-success btn-sm"
|
||||||
title="Browse Files">
|
title="Browse Files">
|
||||||
<i class="fas fa-folder-open"></i>
|
<i class="fas fa-folder-open"></i>
|
||||||
@@ -230,7 +239,7 @@ templ S3Buckets(data dash.S3BucketsData) {
|
|||||||
}
|
}
|
||||||
if len(data.Buckets) == 0 {
|
if len(data.Buckets) == 0 {
|
||||||
<tr>
|
<tr>
|
||||||
<td colspan="9" class="text-center text-muted py-4">
|
<td colspan="10" class="text-center text-muted py-4">
|
||||||
<i class="fas fa-cube fa-3x mb-3 text-muted"></i>
|
<i class="fas fa-cube fa-3x mb-3 text-muted"></i>
|
||||||
<div>
|
<div>
|
||||||
<h5>No Object Store buckets found</h5>
|
<h5>No Object Store buckets found</h5>
|
||||||
@@ -880,9 +889,9 @@ templ S3Buckets(data dash.S3BucketsData) {
|
|||||||
'<div class="text-center py-4">' +
|
'<div class="text-center py-4">' +
|
||||||
'<div class="spinner-border text-primary" role="status">' +
|
'<div class="spinner-border text-primary" role="status">' +
|
||||||
'<span class="visually-hidden">Loading...</span>' +
|
'<span class="visually-hidden">Loading...</span>' +
|
||||||
'</div>' +
|
'<\\/div>' +
|
||||||
'<div class="mt-2">Loading bucket details...</div>' +
|
'<div class="mt-2">Loading bucket details...</div>' +
|
||||||
'</div>';
|
'<\\/div>';
|
||||||
|
|
||||||
detailsModalInstance.show();
|
detailsModalInstance.show();
|
||||||
|
|
||||||
@@ -895,7 +904,7 @@ templ S3Buckets(data dash.S3BucketsData) {
|
|||||||
'<div class="alert alert-danger">' +
|
'<div class="alert alert-danger">' +
|
||||||
'<i class="fas fa-exclamation-triangle me-2"></i>' +
|
'<i class="fas fa-exclamation-triangle me-2"></i>' +
|
||||||
'Error loading bucket details: ' + data.error +
|
'Error loading bucket details: ' + data.error +
|
||||||
'</div>';
|
'<\\/div>';
|
||||||
} else {
|
} else {
|
||||||
displayBucketDetails(data);
|
displayBucketDetails(data);
|
||||||
}
|
}
|
||||||
@@ -906,7 +915,7 @@ templ S3Buckets(data dash.S3BucketsData) {
|
|||||||
'<div class="alert alert-danger">' +
|
'<div class="alert alert-danger">' +
|
||||||
'<i class="fas fa-exclamation-triangle me-2"></i>' +
|
'<i class="fas fa-exclamation-triangle me-2"></i>' +
|
||||||
'Error loading bucket details: ' + error.message +
|
'Error loading bucket details: ' + error.message +
|
||||||
'</div>';
|
'<\\/div>';
|
||||||
});
|
});
|
||||||
});
|
});
|
||||||
});
|
});
|
||||||
@@ -938,146 +947,91 @@ templ S3Buckets(data dash.S3BucketsData) {
|
|||||||
});
|
});
|
||||||
}
|
}
|
||||||
|
|
||||||
function displayBucketDetails(data) {
|
function displayBucketDetails(data) {
|
||||||
const bucket = data.bucket;
|
const bucket = data.bucket;
|
||||||
const objects = data.objects || [];
|
|
||||||
|
|
||||||
// Helper function to escape HTML to prevent XSS
|
function escapeHtml(v) {
|
||||||
function escapeHtml(v) {
|
return String(v ?? '')
|
||||||
return String(v ?? '')
|
.replace(/&/g, '&')
|
||||||
.replace(/&/g, '&')
|
.replace(/</g, '<')
|
||||||
.replace(/</g, '<')
|
.replace(/>/g, '>')
|
||||||
.replace(/>/g, '>')
|
.replace(/"/g, '"')
|
||||||
.replace(/"/g, '"')
|
.replace(/'/g, ''');
|
||||||
.replace(/'/g, ''');
|
|
||||||
}
|
|
||||||
|
|
||||||
// Helper function to format bytes
|
|
||||||
function formatBytes(bytes) {
|
|
||||||
if (bytes === 0) return '0 Bytes';
|
|
||||||
const k = 1024;
|
|
||||||
const sizes = ['Bytes', 'KB', 'MB', 'GB', 'TB'];
|
|
||||||
const i = Math.floor(Math.log(bytes) / Math.log(k));
|
|
||||||
return parseFloat((bytes / Math.pow(k, i)).toFixed(2)) + ' ' + sizes[i];
|
|
||||||
}
|
|
||||||
|
|
||||||
// Helper function to format date
|
|
||||||
function formatDate(dateString) {
|
|
||||||
const date = new Date(dateString);
|
|
||||||
return date.toLocaleString();
|
|
||||||
}
|
|
||||||
|
|
||||||
// Generate objects table
|
|
||||||
let objectsTable = '';
|
|
||||||
if (objects.length > 0) {
|
|
||||||
objectsTable = '<div class="table-responsive">' +
|
|
||||||
'<table class="table table-sm table-striped">' +
|
|
||||||
'<thead>' +
|
|
||||||
'<tr>' +
|
|
||||||
'<th>Object Key</th>' +
|
|
||||||
'<th>Size</th>' +
|
|
||||||
'<th>Last Modified</th>' +
|
|
||||||
'<th>Storage Class</th>' +
|
|
||||||
'</tr>' +
|
|
||||||
'</thead>' +
|
|
||||||
'<tbody>' +
|
|
||||||
objects.map(obj =>
|
|
||||||
'<tr>' +
|
|
||||||
'<td><i class="fas fa-file me-1"></i>' + escapeHtml(obj.key) + '</td>' +
|
|
||||||
'<td>' + formatBytes(obj.size) + '</td>' +
|
|
||||||
'<td>' + formatDate(obj.last_modified) + '</td>' +
|
|
||||||
'<td><span class="badge bg-primary">' + escapeHtml(obj.storage_class) + '</span></td>' +
|
|
||||||
'</tr>'
|
|
||||||
).join('') +
|
|
||||||
'</tbody>' +
|
|
||||||
'</table>' +
|
|
||||||
'</div>';
|
|
||||||
} else {
|
|
||||||
objectsTable = '<div class="text-center py-4 text-muted">' +
|
|
||||||
'<i class="fas fa-file fa-3x mb-3"></i>' +
|
|
||||||
'<div>No objects found in this bucket</div>' +
|
|
||||||
'</div>';
|
|
||||||
}
|
|
||||||
|
|
||||||
const content = '<div class="row">' +
|
|
||||||
'<div class="col-md-6">' +
|
|
||||||
'<h6><i class="fas fa-info-circle me-2"></i>Bucket Information</h6>' +
|
|
||||||
'<table class="table table-sm">' +
|
|
||||||
'<tr>' +
|
|
||||||
'<td><strong>Name:</strong></td>' +
|
|
||||||
'<td>' + escapeHtml(bucket.name) + '</td>' +
|
|
||||||
'</tr>' +
|
|
||||||
'<tr>' +
|
|
||||||
'<td><strong>Owner:</strong></td>' +
|
|
||||||
'<td>' + (bucket.owner ? '<span class="badge bg-info"><i class="fas fa-user me-1"></i>' + escapeHtml(bucket.owner) + '</span>' : '<span class="text-muted">No owner (admin-only)</span>') + '</td>' +
|
|
||||||
'</tr>' +
|
|
||||||
'<tr>' +
|
|
||||||
'<td><strong>Created:</strong></td>' +
|
|
||||||
'<td>' + formatDate(bucket.created_at) + '</td>' +
|
|
||||||
'</tr>' +
|
|
||||||
'<tr>' +
|
|
||||||
'<td><strong>Last Modified:</strong></td>' +
|
|
||||||
'<td>' + formatDate(bucket.last_modified) + '</td>' +
|
|
||||||
'</tr>' +
|
|
||||||
'<tr>' +
|
|
||||||
'<td><strong>Total Size:</strong></td>' +
|
|
||||||
'<td>' + formatBytes(bucket.size) + '</td>' +
|
|
||||||
'</tr>' +
|
|
||||||
'<tr>' +
|
|
||||||
'<td><strong>Object Count:</strong></td>' +
|
|
||||||
'<td>' + bucket.object_count + '</td>' +
|
|
||||||
'</tr>' +
|
|
||||||
'</table>' +
|
|
||||||
'</div>' +
|
|
||||||
'<div class="col-md-6">' +
|
|
||||||
'<h6><i class="fas fa-cogs me-2"></i>Configuration</h6>' +
|
|
||||||
'<table class="table table-sm">' +
|
|
||||||
'<tr>' +
|
|
||||||
'<td><strong>Quota:</strong></td>' +
|
|
||||||
'<td>' +
|
|
||||||
(bucket.quota_enabled ?
|
|
||||||
'<span class="badge bg-success">' + formatBytes(bucket.quota) + '</span>' :
|
|
||||||
'<span class="badge bg-secondary">Disabled</span>'
|
|
||||||
) +
|
|
||||||
'</td>' +
|
|
||||||
'</tr>' +
|
|
||||||
'<tr>' +
|
|
||||||
'<td><strong>Versioning:</strong></td>' +
|
|
||||||
'<td>' +
|
|
||||||
(bucket.versioning_status === 'Enabled' ?
|
|
||||||
'<span class="badge bg-success"><i class="fas fa-check me-1"></i>Enabled</span>' :
|
|
||||||
bucket.versioning_status === 'Suspended' ?
|
|
||||||
'<span class="badge bg-warning"><i class="fas fa-pause me-1"></i>Suspended</span>' :
|
|
||||||
'<span class="text-muted">Not configured</span>'
|
|
||||||
) +
|
|
||||||
'</td>' +
|
|
||||||
'</tr>' +
|
|
||||||
'<tr>' +
|
|
||||||
'<td><strong>Object Lock:</strong></td>' +
|
|
||||||
'<td>' +
|
|
||||||
(bucket.object_lock_enabled ?
|
|
||||||
'<span class="badge bg-warning"><i class="fas fa-lock me-1"></i>Enabled</span>' +
|
|
||||||
(bucket.object_lock_mode && bucket.object_lock_duration > 0 ?
|
|
||||||
'<br><small class="text-muted">' + escapeHtml(bucket.object_lock_mode) + ' • ' + bucket.object_lock_duration + ' days</small>' :
|
|
||||||
'') :
|
|
||||||
'<span class="text-muted">Not configured</span>'
|
|
||||||
) +
|
|
||||||
'</td>' +
|
|
||||||
'</tr>' +
|
|
||||||
'</table>' +
|
|
||||||
'</div>' +
|
|
||||||
'</div>' +
|
|
||||||
'<hr>' +
|
|
||||||
'<div class="row">' +
|
|
||||||
'<div class="col-12">' +
|
|
||||||
'<h6><i class="fas fa-list me-2"></i>Objects (' + objects.length + ')</h6>' +
|
|
||||||
objectsTable +
|
|
||||||
'</div>' +
|
|
||||||
'</div>';
|
|
||||||
|
|
||||||
document.getElementById('bucketDetailsContent').innerHTML = content;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
function formatBytes(bytes) {
|
||||||
|
if (bytes === 0) return '0 Bytes';
|
||||||
|
const k = 1024;
|
||||||
|
const sizes = ['Bytes', 'KB', 'MB', 'GB', 'TB'];
|
||||||
|
const i = Math.floor(Math.log(bytes) / Math.log(k));
|
||||||
|
return parseFloat((bytes / Math.pow(k, i)).toFixed(2)) + ' ' + sizes[i];
|
||||||
|
}
|
||||||
|
|
||||||
|
function formatDate(dateString) {
|
||||||
|
const date = new Date(dateString);
|
||||||
|
return date.toLocaleString();
|
||||||
|
}
|
||||||
|
|
||||||
|
let ownerHtml = '<span class="text-muted">No owner (admin-only)</span>';
|
||||||
|
if (bucket.owner) {
|
||||||
|
ownerHtml = '<span class="badge bg-info"><i class="fas fa-user me-1"></i>' + escapeHtml(bucket.owner) + '</span>';
|
||||||
|
}
|
||||||
|
|
||||||
|
let usageHtml = '';
|
||||||
|
if (bucket.physical_size > 0 && bucket.logical_size > 0 && bucket.physical_size > bucket.logical_size) {
|
||||||
|
const overhead = (bucket.physical_size / bucket.logical_size).toFixed(1);
|
||||||
|
usageHtml = '<br><small class="text-muted">' + overhead + 'x overhead<\/small>';
|
||||||
|
}
|
||||||
|
|
||||||
|
let quotaHtml = '<span class="badge bg-secondary">Disabled</span>';
|
||||||
|
if (bucket.quota_enabled) {
|
||||||
|
quotaHtml = '<span class="badge bg-success">' + formatBytes(bucket.quota) + '</span>';
|
||||||
|
}
|
||||||
|
|
||||||
|
let versioningHtml = '<span class="text-muted">Not configured</span>';
|
||||||
|
if (bucket.versioning_status === 'Enabled') {
|
||||||
|
versioningHtml = '<span class="badge bg-success"><i class="fas fa-check me-1"></i>Enabled</span>';
|
||||||
|
} else if (bucket.versioning_status === 'Suspended') {
|
||||||
|
versioningHtml = '<span class="badge bg-warning"><i class="fas fa-pause me-1"></i>Suspended</span>';
|
||||||
|
}
|
||||||
|
|
||||||
|
let objectLockHtml = '<span class="text-muted">Not configured</span>';
|
||||||
|
if (bucket.object_lock_enabled) {
|
||||||
|
let details = '';
|
||||||
|
if (bucket.object_lock_mode && bucket.object_lock_duration > 0) {
|
||||||
|
details = '<br><small class="text-muted">' + escapeHtml(bucket.object_lock_mode) + ' • ' + bucket.object_lock_duration + ' days<\/small>';
|
||||||
|
}
|
||||||
|
objectLockHtml = '<span class="badge bg-warning"><i class="fas fa-lock me-1"></i>Enabled</span>' + details;
|
||||||
|
}
|
||||||
|
|
||||||
|
const rows = [
|
||||||
|
'<div class="row">',
|
||||||
|
'<div class="col-md-6">',
|
||||||
|
'<h6><i class="fas fa-info-circle me-2"></i>Bucket Information</h6>',
|
||||||
|
'<table class="table table-sm">',
|
||||||
|
'<tr><td><strong>Name:</strong></td><td>' + escapeHtml(bucket.name) + '<\/td><\/tr>',
|
||||||
|
'<tr><td><strong>Owner:</strong></td><td>' + ownerHtml + '<\/td><\/tr>',
|
||||||
|
'<tr><td><strong>Created:</strong></td><td>' + formatDate(bucket.created_at) + '<\/td><\/tr>',
|
||||||
|
'<tr><td><strong>Last Modified:</strong></td><td>' + formatDate(bucket.last_modified) + '<\/td><\/tr>',
|
||||||
|
'<tr><td><strong>Logical Size:</strong></td><td>' + formatBytes(bucket.logical_size) + '<\/td><\/tr>',
|
||||||
|
'<tr><td><strong>Physical Size:</strong></td><td>' + formatBytes(bucket.physical_size) + usageHtml + '<\/td><\/tr>',
|
||||||
|
'<tr><td><strong>Object Count:</strong></td><td>' + bucket.object_count + '<\/td><\/tr>',
|
||||||
|
'<\/table>',
|
||||||
|
'<\/div>',
|
||||||
|
'<div class="col-md-6">',
|
||||||
|
'<h6><i class="fas fa-cogs me-2"></i>Configuration</h6>',
|
||||||
|
'<table class="table table-sm">',
|
||||||
|
'<tr><td><strong>Quota:</strong></td><td>' + quotaHtml + '<\/td><\/tr>',
|
||||||
|
'<tr><td><strong>Versioning:</strong></td><td>' + versioningHtml + '<\/td><\/tr>',
|
||||||
|
'<tr><td><strong>Object Lock:</strong></td><td>' + objectLockHtml + '<\/td><\/tr>',
|
||||||
|
'<\/table>',
|
||||||
|
'<\/div>',
|
||||||
|
'<\/div>'
|
||||||
|
];
|
||||||
|
|
||||||
|
document.getElementById('bucketDetailsContent').innerHTML = rows.join('');
|
||||||
|
}
|
||||||
|
|
||||||
function exportBucketList() {
|
function exportBucketList() {
|
||||||
// RFC 4180 compliant CSV escaping: escape double quotes by doubling them
|
// RFC 4180 compliant CSV escaping: escape double quotes by doubling them
|
||||||
function escapeCsvField(value) {
|
function escapeCsvField(value) {
|
||||||
@@ -1097,23 +1051,26 @@ templ S3Buckets(data dash.S3BucketsData) {
|
|||||||
owner: cells[1].textContent.trim(),
|
owner: cells[1].textContent.trim(),
|
||||||
created: cells[2].textContent.trim(),
|
created: cells[2].textContent.trim(),
|
||||||
objects: cells[3].textContent.trim(),
|
objects: cells[3].textContent.trim(),
|
||||||
size: cells[4].textContent.trim(),
|
|
||||||
quota: cells[5].textContent.trim(),
|
logicalSize: cells[4].textContent.trim(),
|
||||||
versioning: cells[6].textContent.trim(),
|
physicalSize: cells[5].textContent.trim(),
|
||||||
objectLock: cells[7].textContent.trim()
|
quota: cells[6].textContent.trim(),
|
||||||
|
versioning: cells[7].textContent.trim(),
|
||||||
|
objectLock: cells[8].textContent.trim()
|
||||||
};
|
};
|
||||||
}
|
}
|
||||||
return null;
|
return null;
|
||||||
}).filter(bucket => bucket !== null);
|
}).filter(bucket => bucket !== null);
|
||||||
|
|
||||||
const csvContent = "data:text/csv;charset=utf-8," +
|
const csvContent = "data:text/csv;charset=utf-8," +
|
||||||
"Name,Owner,Created,Objects,Size,Quota,Versioning,Object Lock\n" +
|
"Name,Owner,Logical Size,Physical Size,Object Count,Created,Quota,Versioning,Object Lock\n" +
|
||||||
buckets.map(b => [
|
buckets.map(b => [
|
||||||
escapeCsvField(b.name),
|
escapeCsvField(b.name),
|
||||||
escapeCsvField(b.owner),
|
escapeCsvField(b.owner),
|
||||||
escapeCsvField(b.created),
|
escapeCsvField(b.logicalSize),
|
||||||
|
escapeCsvField(b.physicalSize),
|
||||||
escapeCsvField(b.objects),
|
escapeCsvField(b.objects),
|
||||||
escapeCsvField(b.size),
|
escapeCsvField(b.created),
|
||||||
escapeCsvField(b.quota),
|
escapeCsvField(b.quota),
|
||||||
escapeCsvField(b.versioning),
|
escapeCsvField(b.versioning),
|
||||||
escapeCsvField(b.objectLock)
|
escapeCsvField(b.objectLock)
|
||||||
@@ -1151,4 +1108,4 @@ func getQuotaInMB(quotaBytes int64) int64 {
|
|||||||
quotaBytes = -quotaBytes // Handle disabled quotas (negative values)
|
quotaBytes = -quotaBytes // Handle disabled quotas (negative values)
|
||||||
}
|
}
|
||||||
return quotaBytes / (1024 * 1024)
|
return quotaBytes / (1024 * 1024)
|
||||||
}
|
}
|
||||||
File diff suppressed because one or more lines are too long
Reference in New Issue
Block a user