* filer use context without cancellation * pass along context * fix: copy to bucket with default SSE-S3 encryption fails (#7562) When copying an object from an encrypted bucket to a temporary unencrypted bucket, then to another bucket with default SSE-S3 encryption, the operation fails with 'invalid SSE-S3 source key type' error. Root cause: When objects are copied from an SSE-S3 encrypted bucket to an unencrypted bucket, the 'X-Amz-Server-Side-Encryption: AES256' header is preserved but the actual encryption key (SeaweedFSSSES3Key) is stripped. This creates an 'orphaned' SSE-S3 header that causes IsSSES3EncryptedInternal() to return true, triggering decryption logic with a nil key. Fix: 1. Modified IsSSES3EncryptedInternal() to require BOTH the AES256 header AND the SeaweedFSSSES3Key to be present before returning true 2. Added isOrphanedSSES3Header() to detect orphaned SSE-S3 headers 3. Updated copy handler to strip orphaned headers during copy operations Fixes #7562 * fmt * refactor: simplify isOrphanedSSES3Header function logic Remove redundant existence check since the caller iterates through metadata map, making the check unnecessary. Improves readability while maintaining the same functionality.
This commit is contained in:
@@ -171,8 +171,14 @@ func (s3a *S3ApiServer) CopyObjectHandler(w http.ResponseWriter, r *http.Request
|
||||
// Skip encryption-specific headers that might conflict with destination encryption type
|
||||
skipHeader := false
|
||||
|
||||
// Skip orphaned SSE-S3 headers (header exists but key is missing)
|
||||
// This prevents confusion about the object's actual encryption state
|
||||
if isOrphanedSSES3Header(k, entry.Extended) {
|
||||
skipHeader = true
|
||||
}
|
||||
|
||||
// If we're doing cross-encryption, skip conflicting headers
|
||||
if len(entry.GetChunks()) > 0 {
|
||||
if !skipHeader && len(entry.GetChunks()) > 0 {
|
||||
// Detect source and destination encryption types
|
||||
srcHasSSEC := IsSSECEncrypted(entry.Extended)
|
||||
srcHasSSEKMS := IsSSEKMSEncrypted(entry.Extended)
|
||||
@@ -297,7 +303,7 @@ func (s3a *S3ApiServer) CopyObjectHandler(w http.ResponseWriter, r *http.Request
|
||||
// For non-versioned destination, use regular copy
|
||||
// Remove any versioning-related metadata from source that shouldn't carry over
|
||||
cleanupVersioningMetadata(dstEntry.Extended)
|
||||
|
||||
|
||||
dstPath := util.FullPath(fmt.Sprintf("%s/%s%s", s3a.option.BucketsPath, dstBucket, dstObject))
|
||||
dstDir, dstName := dstPath.DirAndName()
|
||||
|
||||
@@ -2350,6 +2356,27 @@ func shouldCreateVersionForCopy(versioningState string) bool {
|
||||
return versioningState == s3_constants.VersioningEnabled
|
||||
}
|
||||
|
||||
// isOrphanedSSES3Header checks if a header is an orphaned SSE-S3 encryption header.
|
||||
// An orphaned header is one where the encryption indicator exists but the actual key is missing.
|
||||
// This can happen when an object was previously encrypted but then copied without encryption,
|
||||
// leaving behind the header but removing the key. These orphaned headers should be stripped
|
||||
// during copy operations to prevent confusion about the object's actual encryption state.
|
||||
// Fixes GitHub issue #7562.
|
||||
func isOrphanedSSES3Header(headerKey string, metadata map[string][]byte) bool {
|
||||
if headerKey != s3_constants.AmzServerSideEncryption {
|
||||
return false
|
||||
}
|
||||
|
||||
// The header is AmzServerSideEncryption. Check if its value indicates SSE-S3.
|
||||
if string(metadata[headerKey]) == "AES256" {
|
||||
// It's an SSE-S3 header. It's orphaned if the actual encryption key is missing.
|
||||
_, hasKey := metadata[s3_constants.SeaweedFSSSES3Key]
|
||||
return !hasKey
|
||||
}
|
||||
|
||||
return false
|
||||
}
|
||||
|
||||
// shouldSkipEncryptionHeader determines if a header should be skipped when copying extended attributes
|
||||
// based on the source and destination encryption types. This consolidates the repetitive logic for
|
||||
// filtering encryption-related headers during copy operations.
|
||||
|
||||
Reference in New Issue
Block a user