* S3: Implement IAM defaults and STS signing key fallback logic * S3: Refactor startup order to init SSE-S3 key manager before IAM * S3: Derive STS signing key from KEK using HKDF for security isolation * S3: Document STS signing key fallback in security.toml * fix(s3api): refine anonymous access logic and secure-by-default behavior - Initialize anonymous identity by default in `NewIdentityAccessManagement` to prevent nil pointer exceptions. - Ensure `ReplaceS3ApiConfiguration` preserves the anonymous identity if not present in the new configuration. - Update `NewIdentityAccessManagement` signature to accept `filerClient`. - In legacy mode (no policy engine), anonymous defaults to Deny (no actions), preserving secure-by-default behavior. - Use specific `LookupAnonymous` method instead of generic map lookup. - Update tests to accommodate signature changes and verify improved anonymous handling. * feat(s3api): make IAM configuration optional - Start S3 API server without a configuration file if `EnableIam` option is set. - Default to `Allow` effect for policy engine when no configuration is provided (Zero-Config mode). - Handle empty configuration path gracefully in `loadIAMManagerFromConfig`. - Add integration test `iam_optional_test.go` to verify empty config behavior. * fix(iamapi): fix signature mismatch in NewIdentityAccessManagementWithStore * fix(iamapi): properly initialize FilerClient instead of passing nil * fix(iamapi): properly initialize filer client for IAM management - Instead of passing `nil`, construct a `wdclient.FilerClient` using the provided `Filers` addresses. - Ensure `NewIdentityAccessManagementWithStore` receives a valid `filerClient` to avoid potential nil pointer dereferences or limited functionality. * clean: remove dead code in s3api_server.go * refactor(s3api): improve IAM initialization, safety and anonymous access security * fix(s3api): ensure IAM config loads from filer after client init * fix(s3): resolve test failures in integration, CORS, and tagging tests - Fix CORS tests by providing explicit anonymous permissions config - Fix S3 integration tests by setting admin credentials in init - Align tagging test credentials in CI with IAM defaults - Added goroutine to retry IAM config load in iamapi server * fix(s3): allow anonymous access to health targets and S3 Tables when identities are present * fix(ci): use /healthz for Caddy health check in awscli tests * iam, s3api: expose DefaultAllow from IAM and Policy Engine This allows checking the global "Open by Default" configuration from other components like S3 Tables. * s3api/s3tables: support DefaultAllow in permission logic and handler Updated CheckPermissionWithContext to respect the DefaultAllow flag in PolicyContext. This enables "Open by Default" behavior for unauthenticated access in zero-config environments. Added a targeted unit test to verify the logic. * s3api/s3tables: propagate DefaultAllow through handlers Propagated the DefaultAllow flag to individual handlers for namespaces, buckets, tables, policies, and tagging. This ensures consistent "Open by Default" behavior across all S3 Tables API endpoints. * s3api: wire up DefaultAllow for S3 Tables API initialization Updated registerS3TablesRoutes to query the global IAM configuration and set the DefaultAllow flag on the S3 Tables API server. This completes the end-to-end propagation required for anonymous access in zero-config environments. Added a SetDefaultAllow method to S3TablesApiServer to facilitate this. * s3api: fix tests by adding DefaultAllow to mock IAM integrations The IAMIntegration interface was updated to include DefaultAllow(), breaking several mock implementations in tests. This commit fixes the build errors by adding the missing method to the mocks. * env * ensure ports * env * env * fix default allow * add one more test using non-anonymous user * debug * add more debug * less logs
706 lines
23 KiB
Go
706 lines
23 KiB
Go
package s3tables
|
|
|
|
import (
|
|
"context"
|
|
"fmt"
|
|
"net"
|
|
"net/http"
|
|
"os"
|
|
"path/filepath"
|
|
"strconv"
|
|
"testing"
|
|
"time"
|
|
|
|
cryptorand "crypto/rand"
|
|
"sync"
|
|
|
|
"github.com/stretchr/testify/assert"
|
|
"github.com/stretchr/testify/require"
|
|
|
|
"github.com/seaweedfs/seaweedfs/weed/command"
|
|
"github.com/seaweedfs/seaweedfs/weed/glog"
|
|
"github.com/seaweedfs/seaweedfs/weed/s3api/s3tables"
|
|
flag "github.com/seaweedfs/seaweedfs/weed/util/fla9"
|
|
)
|
|
|
|
var (
|
|
miniClusterMutex sync.Mutex
|
|
)
|
|
|
|
func TestS3TablesIntegration(t *testing.T) {
|
|
if testing.Short() {
|
|
t.Skip("Skipping integration test in short mode")
|
|
}
|
|
|
|
// Create and start test cluster
|
|
cluster, err := startMiniCluster(t)
|
|
require.NoError(t, err)
|
|
defer cluster.Stop()
|
|
|
|
// Create S3 Tables client
|
|
client := NewS3TablesClient(cluster.s3Endpoint, testRegion, testAccessKey, testSecretKey)
|
|
|
|
// Run test suite
|
|
t.Run("TableBucketLifecycle", func(t *testing.T) {
|
|
testTableBucketLifecycle(t, client)
|
|
})
|
|
|
|
t.Run("NamespaceLifecycle", func(t *testing.T) {
|
|
testNamespaceLifecycle(t, client)
|
|
})
|
|
|
|
t.Run("TableLifecycle", func(t *testing.T) {
|
|
testTableLifecycle(t, client)
|
|
})
|
|
|
|
t.Run("TableBucketPolicy", func(t *testing.T) {
|
|
testTableBucketPolicy(t, client)
|
|
})
|
|
|
|
t.Run("TablePolicy", func(t *testing.T) {
|
|
testTablePolicy(t, client)
|
|
})
|
|
|
|
t.Run("Tagging", func(t *testing.T) {
|
|
testTagging(t, client)
|
|
})
|
|
|
|
t.Run("TargetOperations", func(t *testing.T) {
|
|
testTargetOperations(t, client)
|
|
})
|
|
}
|
|
|
|
func testTableBucketLifecycle(t *testing.T, client *S3TablesClient) {
|
|
bucketName := "test-bucket-" + randomString(8)
|
|
|
|
// Create table bucket
|
|
createResp, err := client.CreateTableBucket(bucketName, nil)
|
|
require.NoError(t, err, "Failed to create table bucket")
|
|
assert.Contains(t, createResp.ARN, bucketName)
|
|
t.Logf("Created table bucket: %s", createResp.ARN)
|
|
|
|
// Get table bucket
|
|
getResp, err := client.GetTableBucket(createResp.ARN)
|
|
require.NoError(t, err, "Failed to get table bucket")
|
|
assert.Equal(t, bucketName, getResp.Name)
|
|
t.Logf("Got table bucket: %s", getResp.Name)
|
|
|
|
// List table buckets
|
|
listResp, err := client.ListTableBuckets("", "", 0)
|
|
require.NoError(t, err, "Failed to list table buckets")
|
|
found := false
|
|
for _, b := range listResp.TableBuckets {
|
|
if b.Name == bucketName {
|
|
found = true
|
|
break
|
|
}
|
|
}
|
|
assert.True(t, found, "Created bucket should appear in list")
|
|
t.Logf("Listed table buckets, found %d buckets", len(listResp.TableBuckets))
|
|
|
|
// Delete table bucket
|
|
err = client.DeleteTableBucket(createResp.ARN)
|
|
require.NoError(t, err, "Failed to delete table bucket")
|
|
t.Logf("Deleted table bucket: %s", bucketName)
|
|
|
|
// Verify bucket is deleted
|
|
_, err = client.GetTableBucket(createResp.ARN)
|
|
assert.Error(t, err, "Bucket should not exist after deletion")
|
|
}
|
|
|
|
func testNamespaceLifecycle(t *testing.T, client *S3TablesClient) {
|
|
bucketName := "test-ns-bucket-" + randomString(8)
|
|
namespaceName := "test_namespace"
|
|
|
|
// Create table bucket first
|
|
createBucketResp, err := client.CreateTableBucket(bucketName, nil)
|
|
require.NoError(t, err, "Failed to create table bucket")
|
|
defer client.DeleteTableBucket(createBucketResp.ARN)
|
|
|
|
bucketARN := createBucketResp.ARN
|
|
|
|
// Create namespace
|
|
createNsResp, err := client.CreateNamespace(bucketARN, []string{namespaceName})
|
|
require.NoError(t, err, "Failed to create namespace")
|
|
assert.Equal(t, []string{namespaceName}, createNsResp.Namespace)
|
|
t.Logf("Created namespace: %s", namespaceName)
|
|
|
|
// Get namespace
|
|
getNsResp, err := client.GetNamespace(bucketARN, []string{namespaceName})
|
|
require.NoError(t, err, "Failed to get namespace")
|
|
assert.Equal(t, []string{namespaceName}, getNsResp.Namespace)
|
|
t.Logf("Got namespace: %v", getNsResp.Namespace)
|
|
|
|
// List namespaces
|
|
listNsResp, err := client.ListNamespaces(bucketARN, "", "", 0)
|
|
require.NoError(t, err, "Failed to list namespaces")
|
|
found := false
|
|
for _, ns := range listNsResp.Namespaces {
|
|
if len(ns.Namespace) > 0 && ns.Namespace[0] == namespaceName {
|
|
found = true
|
|
break
|
|
}
|
|
}
|
|
assert.True(t, found, "Created namespace should appear in list")
|
|
t.Logf("Listed namespaces, found %d namespaces", len(listNsResp.Namespaces))
|
|
|
|
// Delete namespace
|
|
err = client.DeleteNamespace(bucketARN, []string{namespaceName})
|
|
require.NoError(t, err, "Failed to delete namespace")
|
|
t.Logf("Deleted namespace: %s", namespaceName)
|
|
|
|
// Verify namespace is deleted
|
|
_, err = client.GetNamespace(bucketARN, []string{namespaceName})
|
|
assert.Error(t, err, "Namespace should not exist after deletion")
|
|
}
|
|
|
|
func testTableLifecycle(t *testing.T, client *S3TablesClient) {
|
|
bucketName := "test-table-bucket-" + randomString(8)
|
|
namespaceName := "test_ns"
|
|
tableName := "test_table"
|
|
|
|
// Create table bucket
|
|
createBucketResp, err := client.CreateTableBucket(bucketName, nil)
|
|
require.NoError(t, err, "Failed to create table bucket")
|
|
defer client.DeleteTableBucket(createBucketResp.ARN)
|
|
|
|
bucketARN := createBucketResp.ARN
|
|
|
|
// Create namespace
|
|
_, err = client.CreateNamespace(bucketARN, []string{namespaceName})
|
|
require.NoError(t, err, "Failed to create namespace")
|
|
defer client.DeleteNamespace(bucketARN, []string{namespaceName})
|
|
|
|
// Create table with Iceberg schema
|
|
icebergMetadata := &s3tables.TableMetadata{
|
|
Iceberg: &s3tables.IcebergMetadata{
|
|
Schema: s3tables.IcebergSchema{
|
|
Fields: []s3tables.IcebergSchemaField{
|
|
{Name: "id", Type: "int", Required: true},
|
|
{Name: "name", Type: "string"},
|
|
{Name: "value", Type: "int"},
|
|
},
|
|
},
|
|
},
|
|
}
|
|
|
|
createTableResp, err := client.CreateTable(bucketARN, []string{namespaceName}, tableName, "ICEBERG", icebergMetadata, nil)
|
|
require.NoError(t, err, "Failed to create table")
|
|
assert.NotEmpty(t, createTableResp.TableARN)
|
|
assert.NotEmpty(t, createTableResp.VersionToken)
|
|
t.Logf("Created table: %s (version: %s)", createTableResp.TableARN, createTableResp.VersionToken)
|
|
|
|
// Get table
|
|
getTableResp, err := client.GetTable(bucketARN, []string{namespaceName}, tableName)
|
|
require.NoError(t, err, "Failed to get table")
|
|
assert.Equal(t, tableName, getTableResp.Name)
|
|
assert.Equal(t, "ICEBERG", getTableResp.Format)
|
|
t.Logf("Got table: %s (format: %s)", getTableResp.Name, getTableResp.Format)
|
|
|
|
// List tables
|
|
listTablesResp, err := client.ListTables(bucketARN, []string{namespaceName}, "", "", 0)
|
|
require.NoError(t, err, "Failed to list tables")
|
|
found := false
|
|
for _, tbl := range listTablesResp.Tables {
|
|
if tbl.Name == tableName {
|
|
found = true
|
|
break
|
|
}
|
|
}
|
|
assert.True(t, found, "Created table should appear in list")
|
|
t.Logf("Listed tables, found %d tables", len(listTablesResp.Tables))
|
|
|
|
// Delete table
|
|
err = client.DeleteTable(bucketARN, []string{namespaceName}, tableName)
|
|
require.NoError(t, err, "Failed to delete table")
|
|
t.Logf("Deleted table: %s", tableName)
|
|
|
|
// Verify table is deleted
|
|
_, err = client.GetTable(bucketARN, []string{namespaceName}, tableName)
|
|
assert.Error(t, err, "Table should not exist after deletion")
|
|
}
|
|
|
|
func testTableBucketPolicy(t *testing.T, client *S3TablesClient) {
|
|
bucketName := "test-policy-bucket-" + randomString(8)
|
|
|
|
// Create table bucket
|
|
createBucketResp, err := client.CreateTableBucket(bucketName, nil)
|
|
require.NoError(t, err, "Failed to create table bucket")
|
|
defer client.DeleteTableBucket(createBucketResp.ARN)
|
|
|
|
bucketARN := createBucketResp.ARN
|
|
|
|
// Put bucket policy
|
|
policy := `{"Version":"2012-10-17","Statement":[{"Effect":"Allow","Principal":"*","Action":"s3tables:*","Resource":"*"}]}`
|
|
err = client.PutTableBucketPolicy(bucketARN, policy)
|
|
require.NoError(t, err, "Failed to put table bucket policy")
|
|
t.Logf("Put table bucket policy")
|
|
|
|
// Get bucket policy
|
|
getPolicyResp, err := client.GetTableBucketPolicy(bucketARN)
|
|
require.NoError(t, err, "Failed to get table bucket policy")
|
|
assert.Equal(t, policy, getPolicyResp.ResourcePolicy)
|
|
t.Logf("Got table bucket policy")
|
|
|
|
// Delete bucket policy
|
|
err = client.DeleteTableBucketPolicy(bucketARN)
|
|
require.NoError(t, err, "Failed to delete table bucket policy")
|
|
t.Logf("Deleted table bucket policy")
|
|
|
|
// Verify policy is deleted
|
|
_, err = client.GetTableBucketPolicy(bucketARN)
|
|
assert.Error(t, err, "Policy should not exist after deletion")
|
|
}
|
|
|
|
func testTablePolicy(t *testing.T, client *S3TablesClient) {
|
|
bucketName := "test-table-policy-bucket-" + randomString(8)
|
|
namespaceName := "test_ns"
|
|
tableName := "test_table"
|
|
|
|
// Create table bucket
|
|
createBucketResp, err := client.CreateTableBucket(bucketName, nil)
|
|
require.NoError(t, err, "Failed to create table bucket")
|
|
defer client.DeleteTableBucket(createBucketResp.ARN)
|
|
|
|
bucketARN := createBucketResp.ARN
|
|
|
|
// Create namespace
|
|
_, err = client.CreateNamespace(bucketARN, []string{namespaceName})
|
|
require.NoError(t, err, "Failed to create namespace")
|
|
defer client.DeleteNamespace(bucketARN, []string{namespaceName})
|
|
|
|
// Create table
|
|
icebergMetadata := &s3tables.TableMetadata{
|
|
Iceberg: &s3tables.IcebergMetadata{
|
|
Schema: s3tables.IcebergSchema{
|
|
Fields: []s3tables.IcebergSchemaField{
|
|
{Name: "id", Type: "int", Required: true},
|
|
{Name: "name", Type: "string"},
|
|
},
|
|
},
|
|
},
|
|
}
|
|
|
|
createTableResp, err := client.CreateTable(bucketARN, []string{namespaceName}, tableName, "ICEBERG", icebergMetadata, nil)
|
|
require.NoError(t, err, "Failed to create table")
|
|
defer client.DeleteTable(bucketARN, []string{namespaceName}, tableName)
|
|
|
|
t.Logf("Created table: %s", createTableResp.TableARN)
|
|
|
|
// Verify no policy exists initially
|
|
_, err = client.GetTablePolicy(bucketARN, []string{namespaceName}, tableName)
|
|
assert.Error(t, err, "Policy should not exist initially")
|
|
t.Logf("Verified no policy exists initially")
|
|
|
|
// Put table policy
|
|
policy := `{"Version":"2012-10-17","Statement":[{"Effect":"Allow","Principal":"*","Action":"s3tables:*","Resource":"*"}]}`
|
|
err = client.PutTablePolicy(bucketARN, []string{namespaceName}, tableName, policy)
|
|
require.NoError(t, err, "Failed to put table policy")
|
|
t.Logf("Put table policy")
|
|
|
|
// Get table policy
|
|
getPolicyResp, err := client.GetTablePolicy(bucketARN, []string{namespaceName}, tableName)
|
|
require.NoError(t, err, "Failed to get table policy")
|
|
assert.Equal(t, policy, getPolicyResp.ResourcePolicy)
|
|
t.Logf("Got table policy")
|
|
|
|
// Delete table policy
|
|
err = client.DeleteTablePolicy(bucketARN, []string{namespaceName}, tableName)
|
|
require.NoError(t, err, "Failed to delete table policy")
|
|
t.Logf("Deleted table policy")
|
|
|
|
// Verify policy is deleted
|
|
_, err = client.GetTablePolicy(bucketARN, []string{namespaceName}, tableName)
|
|
assert.Error(t, err, "Policy should not exist after deletion")
|
|
t.Logf("Verified policy deletion")
|
|
}
|
|
|
|
func testTagging(t *testing.T, client *S3TablesClient) {
|
|
bucketName := "test-tag-bucket-" + randomString(8)
|
|
|
|
// Create table bucket with tags
|
|
initialTags := map[string]string{"Environment": "test"}
|
|
createBucketResp, err := client.CreateTableBucket(bucketName, initialTags)
|
|
require.NoError(t, err, "Failed to create table bucket")
|
|
defer client.DeleteTableBucket(createBucketResp.ARN)
|
|
|
|
bucketARN := createBucketResp.ARN
|
|
|
|
// List tags
|
|
listTagsResp, err := client.ListTagsForResource(bucketARN)
|
|
require.NoError(t, err, "Failed to list tags")
|
|
assert.Equal(t, "test", listTagsResp.Tags["Environment"])
|
|
t.Logf("Listed tags: %v", listTagsResp.Tags)
|
|
|
|
// Add more tags
|
|
newTags := map[string]string{"Department": "Engineering"}
|
|
err = client.TagResource(bucketARN, newTags)
|
|
require.NoError(t, err, "Failed to tag resource")
|
|
t.Logf("Added tags")
|
|
|
|
// Verify tags
|
|
listTagsResp, err = client.ListTagsForResource(bucketARN)
|
|
require.NoError(t, err, "Failed to list tags")
|
|
assert.Equal(t, "test", listTagsResp.Tags["Environment"])
|
|
assert.Equal(t, "Engineering", listTagsResp.Tags["Department"])
|
|
t.Logf("Verified tags: %v", listTagsResp.Tags)
|
|
|
|
// Remove a tag
|
|
err = client.UntagResource(bucketARN, []string{"Environment"})
|
|
require.NoError(t, err, "Failed to untag resource")
|
|
t.Logf("Removed tag")
|
|
|
|
// Verify tag is removed
|
|
listTagsResp, err = client.ListTagsForResource(bucketARN)
|
|
require.NoError(t, err, "Failed to list tags")
|
|
_, hasEnvironment := listTagsResp.Tags["Environment"]
|
|
assert.False(t, hasEnvironment, "Environment tag should be removed")
|
|
assert.Equal(t, "Engineering", listTagsResp.Tags["Department"])
|
|
t.Logf("Verified tag removal")
|
|
}
|
|
|
|
func testTargetOperations(t *testing.T, client *S3TablesClient) {
|
|
bucketName := "test-target-bucket-" + randomString(8)
|
|
|
|
var createResp s3tables.CreateTableBucketResponse
|
|
err := client.doTargetRequestAndDecode("CreateTableBucket", &s3tables.CreateTableBucketRequest{
|
|
Name: bucketName,
|
|
}, &createResp)
|
|
require.NoError(t, err, "Failed to create table bucket via target")
|
|
defer client.doTargetRequestAndDecode("DeleteTableBucket", &s3tables.DeleteTableBucketRequest{
|
|
TableBucketARN: createResp.ARN,
|
|
}, nil)
|
|
|
|
var listResp s3tables.ListTableBucketsResponse
|
|
err = client.doTargetRequestAndDecode("ListTableBuckets", &s3tables.ListTableBucketsRequest{}, &listResp)
|
|
require.NoError(t, err, "Failed to list table buckets via target")
|
|
found := false
|
|
for _, b := range listResp.TableBuckets {
|
|
if b.Name == bucketName {
|
|
found = true
|
|
break
|
|
}
|
|
}
|
|
assert.True(t, found, "Created bucket should appear in target list")
|
|
|
|
var getResp s3tables.GetTableBucketResponse
|
|
err = client.doTargetRequestAndDecode("GetTableBucket", &s3tables.GetTableBucketRequest{
|
|
TableBucketARN: createResp.ARN,
|
|
}, &getResp)
|
|
require.NoError(t, err, "Failed to get table bucket via target")
|
|
assert.Equal(t, bucketName, getResp.Name)
|
|
|
|
namespaceName := "target_ns"
|
|
var createNsResp s3tables.CreateNamespaceResponse
|
|
err = client.doTargetRequestAndDecode("CreateNamespace", &s3tables.CreateNamespaceRequest{
|
|
TableBucketARN: createResp.ARN,
|
|
Namespace: []string{namespaceName},
|
|
}, &createNsResp)
|
|
require.NoError(t, err, "Failed to create namespace via target")
|
|
defer client.doTargetRequestAndDecode("DeleteNamespace", &s3tables.DeleteNamespaceRequest{
|
|
TableBucketARN: createResp.ARN,
|
|
Namespace: []string{namespaceName},
|
|
}, nil)
|
|
|
|
var listNsResp s3tables.ListNamespacesResponse
|
|
err = client.doTargetRequestAndDecode("ListNamespaces", &s3tables.ListNamespacesRequest{
|
|
TableBucketARN: createResp.ARN,
|
|
}, &listNsResp)
|
|
require.NoError(t, err, "Failed to list namespaces via target")
|
|
|
|
tableName := "target_table"
|
|
var createTableResp s3tables.CreateTableResponse
|
|
err = client.doTargetRequestAndDecode("CreateTable", &s3tables.CreateTableRequest{
|
|
TableBucketARN: createResp.ARN,
|
|
Namespace: []string{namespaceName},
|
|
Name: tableName,
|
|
Format: "ICEBERG",
|
|
}, &createTableResp)
|
|
require.NoError(t, err, "Failed to create table via target")
|
|
defer client.doTargetRequestAndDecode("DeleteTable", &s3tables.DeleteTableRequest{
|
|
TableBucketARN: createResp.ARN,
|
|
Namespace: []string{namespaceName},
|
|
Name: tableName,
|
|
}, nil)
|
|
|
|
var listTablesResp s3tables.ListTablesResponse
|
|
err = client.doTargetRequestAndDecode("ListTables", &s3tables.ListTablesRequest{
|
|
TableBucketARN: createResp.ARN,
|
|
Namespace: []string{namespaceName},
|
|
}, &listTablesResp)
|
|
require.NoError(t, err, "Failed to list tables via target")
|
|
|
|
var getTableResp s3tables.GetTableResponse
|
|
err = client.doTargetRequestAndDecode("GetTable", &s3tables.GetTableRequest{
|
|
TableBucketARN: createResp.ARN,
|
|
Namespace: []string{namespaceName},
|
|
Name: tableName,
|
|
}, &getTableResp)
|
|
require.NoError(t, err, "Failed to get table via target")
|
|
assert.Equal(t, tableName, getTableResp.Name)
|
|
|
|
policy := `{"Version":"2012-10-17","Statement":[{"Effect":"Allow","Principal":"*","Action":"s3tables:*","Resource":"*"}]}`
|
|
err = client.doTargetRequestAndDecode("PutTableBucketPolicy", &s3tables.PutTableBucketPolicyRequest{
|
|
TableBucketARN: createResp.ARN,
|
|
ResourcePolicy: policy,
|
|
}, nil)
|
|
require.NoError(t, err, "Failed to put bucket policy via target")
|
|
|
|
var getPolicyResp s3tables.GetTableBucketPolicyResponse
|
|
err = client.doTargetRequestAndDecode("GetTableBucketPolicy", &s3tables.GetTableBucketPolicyRequest{
|
|
TableBucketARN: createResp.ARN,
|
|
}, &getPolicyResp)
|
|
require.NoError(t, err, "Failed to get bucket policy via target")
|
|
assert.Equal(t, policy, getPolicyResp.ResourcePolicy)
|
|
|
|
err = client.doTargetRequestAndDecode("DeleteTableBucketPolicy", &s3tables.DeleteTableBucketPolicyRequest{
|
|
TableBucketARN: createResp.ARN,
|
|
}, nil)
|
|
require.NoError(t, err, "Failed to delete bucket policy via target")
|
|
|
|
err = client.doTargetRequestAndDecode("TagResource", &s3tables.TagResourceRequest{
|
|
ResourceARN: createResp.ARN,
|
|
Tags: map[string]string{"Environment": "test"},
|
|
}, nil)
|
|
require.NoError(t, err, "Failed to tag resource via target")
|
|
|
|
var listTagsResp s3tables.ListTagsForResourceResponse
|
|
err = client.doTargetRequestAndDecode("ListTagsForResource", &s3tables.ListTagsForResourceRequest{
|
|
ResourceARN: createResp.ARN,
|
|
}, &listTagsResp)
|
|
require.NoError(t, err, "Failed to list tags via target")
|
|
assert.Equal(t, "test", listTagsResp.Tags["Environment"])
|
|
|
|
err = client.doTargetRequestAndDecode("UntagResource", &s3tables.UntagResourceRequest{
|
|
ResourceARN: createResp.ARN,
|
|
TagKeys: []string{"Environment"},
|
|
}, nil)
|
|
require.NoError(t, err, "Failed to untag resource via target")
|
|
}
|
|
|
|
// Helper functions
|
|
|
|
// findAvailablePorts finds n available ports by binding to port 0 multiple times
|
|
// It keeps the listeners open until all ports are found to ensure uniqueness
|
|
func findAvailablePorts(n int) ([]int, error) {
|
|
listeners := make([]*net.TCPListener, n)
|
|
ports := make([]int, n)
|
|
|
|
// Open all listeners to ensure we get unique ports
|
|
for i := 0; i < n; i++ {
|
|
listener, err := net.Listen("tcp", "127.0.0.1:0")
|
|
if err != nil {
|
|
// Close valid listeners before returning error
|
|
for j := 0; j < i; j++ {
|
|
listeners[j].Close()
|
|
}
|
|
return nil, err
|
|
}
|
|
listeners[i] = listener.(*net.TCPListener)
|
|
ports[i] = listeners[i].Addr().(*net.TCPAddr).Port
|
|
}
|
|
|
|
// Close all listeners
|
|
for _, l := range listeners {
|
|
l.Close()
|
|
}
|
|
|
|
return ports, nil
|
|
}
|
|
|
|
// startMiniCluster starts a weed mini instance directly without exec
|
|
func startMiniCluster(t *testing.T) (*TestCluster, error) {
|
|
// Find available ports
|
|
// We need 8 unique ports: Master(2), Volume(2), Filer(2), S3(2)
|
|
ports, err := findAvailablePorts(8)
|
|
if err != nil {
|
|
return nil, fmt.Errorf("failed to find available ports: %v", err)
|
|
}
|
|
|
|
masterPort := ports[0]
|
|
masterGrpcPort := ports[1]
|
|
volumePort := ports[2]
|
|
volumeGrpcPort := ports[3]
|
|
filerPort := ports[4]
|
|
filerGrpcPort := ports[5]
|
|
s3Port := ports[6]
|
|
s3GrpcPort := ports[7]
|
|
// Create temporary directory for test data
|
|
testDir := t.TempDir()
|
|
|
|
// Ensure no configuration file from previous runs
|
|
configFile := filepath.Join(testDir, "mini.options")
|
|
_ = os.Remove(configFile)
|
|
|
|
// Create context with timeout
|
|
ctx, cancel := context.WithCancel(context.Background())
|
|
|
|
s3Endpoint := fmt.Sprintf("http://127.0.0.1:%d", s3Port)
|
|
cluster := &TestCluster{
|
|
t: t,
|
|
dataDir: testDir,
|
|
ctx: ctx,
|
|
cancel: cancel,
|
|
masterPort: masterPort,
|
|
volumePort: volumePort,
|
|
filerPort: filerPort,
|
|
s3Port: s3Port,
|
|
s3Endpoint: s3Endpoint,
|
|
}
|
|
|
|
// Create empty security.toml to disable JWT authentication in tests
|
|
securityToml := filepath.Join(testDir, "security.toml")
|
|
err = os.WriteFile(securityToml, []byte("# Empty security config for testing\n"), 0644)
|
|
if err != nil {
|
|
cancel()
|
|
return nil, fmt.Errorf("failed to create security.toml: %v", err)
|
|
}
|
|
|
|
// Set environment variables for admin credentials safely for this test
|
|
if os.Getenv("AWS_ACCESS_KEY_ID") == "" {
|
|
t.Setenv("AWS_ACCESS_KEY_ID", "admin")
|
|
}
|
|
if os.Getenv("AWS_SECRET_ACCESS_KEY") == "" {
|
|
t.Setenv("AWS_SECRET_ACCESS_KEY", "admin")
|
|
}
|
|
|
|
// Start weed mini in a goroutine by calling the command directly
|
|
cluster.wg.Add(1)
|
|
go func() {
|
|
defer cluster.wg.Done()
|
|
|
|
// Protect global state mutation with a mutex
|
|
miniClusterMutex.Lock()
|
|
defer miniClusterMutex.Unlock()
|
|
|
|
// Save current directory and args
|
|
oldDir, _ := os.Getwd()
|
|
oldArgs := os.Args
|
|
defer func() {
|
|
os.Chdir(oldDir)
|
|
os.Args = oldArgs
|
|
}()
|
|
|
|
// Change to test directory so mini picks up security.toml
|
|
os.Chdir(testDir)
|
|
|
|
// Configure args for mini command
|
|
os.Args = []string{
|
|
"weed",
|
|
"-dir=" + testDir,
|
|
"-master.port=" + strconv.Itoa(masterPort),
|
|
"-master.port.grpc=" + strconv.Itoa(masterGrpcPort),
|
|
"-volume.port=" + strconv.Itoa(volumePort),
|
|
"-volume.port.grpc=" + strconv.Itoa(volumeGrpcPort),
|
|
"-filer.port=" + strconv.Itoa(filerPort),
|
|
"-filer.port.grpc=" + strconv.Itoa(filerGrpcPort),
|
|
"-s3.port=" + strconv.Itoa(s3Port),
|
|
"-s3.port.grpc=" + strconv.Itoa(s3GrpcPort),
|
|
"-webdav.port=0", // Disable WebDAV
|
|
"-admin.ui=false", // Disable admin UI
|
|
"-master.volumeSizeLimitMB=32", // Small volumes for testing
|
|
"-ip=127.0.0.1",
|
|
"-master.peers=none", // Faster startup
|
|
"-s3.iam.readOnly=false", // Enable IAM write operations for tests
|
|
}
|
|
|
|
// Suppress most logging during tests
|
|
glog.MaxSize = 1024 * 1024
|
|
|
|
// Find and run the mini command
|
|
for _, cmd := range command.Commands {
|
|
if cmd.Name() == "mini" && cmd.Run != nil {
|
|
cmd.Flag.Parse(os.Args[1:])
|
|
args := cmd.Flag.Args()
|
|
command.MiniClusterCtx = ctx
|
|
cmd.Run(cmd, args)
|
|
command.MiniClusterCtx = nil
|
|
return
|
|
}
|
|
}
|
|
}()
|
|
|
|
// Wait for S3 service to be ready
|
|
err = waitForS3Ready(cluster.s3Endpoint, 30*time.Second)
|
|
if err != nil {
|
|
cancel()
|
|
return nil, fmt.Errorf("S3 service failed to start: %v", err)
|
|
}
|
|
|
|
cluster.isRunning = true
|
|
|
|
t.Logf("Test cluster started successfully at %s", cluster.s3Endpoint)
|
|
return cluster, nil
|
|
}
|
|
|
|
// Stop stops the test cluster
|
|
func (c *TestCluster) Stop() {
|
|
if c.cancel != nil {
|
|
c.cancel()
|
|
}
|
|
// Give services time to shut down gracefully
|
|
if c.isRunning {
|
|
time.Sleep(500 * time.Millisecond)
|
|
}
|
|
// Wait for the mini goroutine to finish
|
|
done := make(chan struct{})
|
|
go func() {
|
|
c.wg.Wait()
|
|
close(done)
|
|
}()
|
|
timer := time.NewTimer(2 * time.Second)
|
|
defer timer.Stop()
|
|
select {
|
|
case <-done:
|
|
// Goroutine finished
|
|
case <-timer.C:
|
|
// Timeout - goroutine doesn't respond to context cancel
|
|
// This may indicate the mini cluster didn't shut down cleanly
|
|
c.t.Log("Warning: Test cluster shutdown timed out after 2 seconds")
|
|
}
|
|
|
|
// Reset the global cmdMini flags to prevent state leakage to other tests
|
|
for _, cmd := range command.Commands {
|
|
if cmd.Name() == "mini" {
|
|
// Reset flags to defaults
|
|
cmd.Flag.VisitAll(func(f *flag.Flag) {
|
|
// Reset to default value
|
|
f.Value.Set(f.DefValue)
|
|
})
|
|
break
|
|
}
|
|
}
|
|
}
|
|
|
|
// waitForS3Ready waits for the S3 service to be ready
|
|
func waitForS3Ready(endpoint string, timeout time.Duration) error {
|
|
client := &http.Client{Timeout: 1 * time.Second}
|
|
deadline := time.Now().Add(timeout)
|
|
|
|
for time.Now().Before(deadline) {
|
|
resp, err := client.Get(endpoint)
|
|
if err == nil {
|
|
resp.Body.Close()
|
|
// Wait a bit more to ensure service is fully ready
|
|
time.Sleep(500 * time.Millisecond)
|
|
return nil
|
|
}
|
|
time.Sleep(200 * time.Millisecond)
|
|
}
|
|
|
|
return fmt.Errorf("timeout waiting for S3 service at %s", endpoint)
|
|
}
|
|
|
|
// randomString generates a random string for unique naming
|
|
func randomString(length int) string {
|
|
const charset = "abcdefghijklmnopqrstuvwxyz0123456789"
|
|
b := make([]byte, length)
|
|
if _, err := cryptorand.Read(b); err != nil {
|
|
panic("failed to generate random string: " + err.Error())
|
|
}
|
|
for i := range b {
|
|
b[i] = charset[int(b[i])%len(charset)]
|
|
}
|
|
return string(b)
|
|
}
|