milvus/internal/storage/gcp_native_object_storage_test.go
wei liu 975c91df16
feat: Add comprehensive snapshot functionality for collections (#44361)
issue: #44358

Implement complete snapshot management system including creation,
deletion, listing, description, and restoration capabilities across all
system components.

Key features:
- Create snapshots for entire collections
- Drop snapshots by name with proper cleanup
- List snapshots with collection filtering
- Describe snapshot details and metadata

Components added/modified:
- Client SDK with full snapshot API support and options
- DataCoord snapshot service with metadata management
- Proxy layer with task-based snapshot operations
- Protocol buffer definitions for snapshot RPCs
- Comprehensive unit tests with mockey framework
- Integration tests for end-to-end validation

Technical implementation:
- Snapshot metadata storage in etcd with proper indexing
- File-based snapshot data persistence in object storage
- Garbage collection integration for snapshot cleanup
- Error handling and validation across all operations
- Thread-safe operations with proper locking mechanisms

<!-- This is an auto-generated comment: release notes by coderabbit.ai
-->
- Core invariant/assumption: snapshots are immutable point‑in‑time
captures identified by (collection, snapshot name/ID); etcd snapshot
metadata is authoritative for lifecycle (PENDING → COMMITTED → DELETING)
and per‑segment manifests live in object storage (Avro / StorageV2). GC
and restore logic must see snapshotRefIndex loaded
(snapshotMeta.IsRefIndexLoaded) before reclaiming or relying on
segment/index files.

- New capability added: full end‑to‑end snapshot subsystem — client SDK
APIs (Create/Drop/List/Describe/Restore + restore job queries),
DataCoord SnapshotWriter/Reader (Avro + StorageV2 manifests),
snapshotMeta in meta, SnapshotManager orchestration
(create/drop/describe/list/restore), copy‑segment restore
tasks/inspector/checker, proxy & RPC surface, GC integration, and
docs/tests — enabling point‑in‑time collection snapshots persisted to
object storage and restorations orchestrated across components.

- Logic removed/simplified and why: duplicated recursive
compaction/delta‑log traversal and ad‑hoc lookup code were consolidated
behind two focused APIs/owners (Handler.GetDeltaLogFromCompactTo for
delta traversal and SnapshotManager/SnapshotReader for snapshot I/O).
MixCoord/coordinator broker paths were converted to thin RPC proxies.
This eliminates multiple implementations of the same traversal/lookup,
reducing divergence and simplifying responsibility boundaries.

- Why this does NOT introduce data loss or regressions: snapshot
create/drop use explicit two‑phase semantics (PENDING → COMMIT/DELETING)
with SnapshotWriter writing manifests and metadata before commit; GC
uses snapshotRefIndex guards and
IsRefIndexLoaded/GetSnapshotBySegment/GetSnapshotByIndex checks to avoid
removing referenced files; restore flow pre‑allocates job IDs, validates
resources (partitions/indexes), performs rollback on failure
(rollbackRestoreSnapshot), and converts/updates segment/index metadata
only after successful copy tasks. Extensive unit and integration tests
exercise pending/deleting/GC/restore/error paths to ensure idempotence
and protection against premature deletion.
<!-- end of auto-generated comment: release notes by coderabbit.ai -->

---------

Signed-off-by: Wei Liu <wei.liu@zilliz.com>
2026-01-06 10:15:24 +08:00

524 lines
14 KiB
Go

// Licensed to the LF AI & Data foundation under one
// or more contributor license agreements. See the NOTICE file
// distributed with this work for additional information
// regarding copyright ownership. The ASF licenses this file
// to you under the Apache License, Version 2.0 (the
// "License"); you may not use this file except in compliance
// with the License. You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
package storage
import (
"bytes"
"context"
"fmt"
"io"
"testing"
"github.com/stretchr/testify/assert"
"github.com/stretchr/testify/require"
"github.com/milvus-io/milvus/pkg/v2/objectstorage"
)
func TestGcpNativeObjectStorage(t *testing.T) {
ctx := context.Background()
bucketName := "test-bucket"
config := objectstorage.Config{
Address: "storage.gcs.127.0.0.1.nip.io:4443",
BucketName: bucketName,
CreateBucket: true,
UseIAM: false,
CloudProvider: "gcpnative",
UseSSL: false,
GcpNativeWithoutAuth: true,
}
t.Run("test initialize", func(t *testing.T) {
var err error
config.BucketName = ""
_, err = newGcpNativeObjectStorageWithConfig(ctx, &config)
assert.Error(t, err)
config.BucketName = bucketName
_, err = newGcpNativeObjectStorageWithConfig(ctx, &config)
assert.Equal(t, err, nil)
})
t.Run("test load", func(t *testing.T) {
testCM, err := newGcpNativeObjectStorageWithConfig(ctx, &config)
assert.Equal(t, err, nil)
defer testCM.DeleteBucket(ctx, config.BucketName)
prepareTests := []struct {
key string
value []byte
}{
{"abc", []byte("123")},
{"abcd", []byte("1234")},
{"key_1", []byte("111")},
{"key_2", []byte("222")},
{"key_3", []byte("333")},
}
for _, test := range prepareTests {
err := testCM.PutObject(ctx, config.BucketName, test.key, bytes.NewReader(test.value),
int64(len(test.value)))
require.NoError(t, err)
}
loadTests := []struct {
isvalid bool
loadKey string
expectedValue []byte
description string
}{
{true, "abc", []byte("123"), "load valid key abc"},
{true, "abcd", []byte("1234"), "load valid key abcd"},
{true, "key_1", []byte("111"), "load valid key key_1"},
{true, "key_2", []byte("222"), "load valid key key_2"},
{true, "key_3", []byte("333"), "load valid key key_3"},
{false, "key_not_exist", []byte(""), "load invalid key key_not_exist"},
{false, "/", []byte(""), "load leading slash"},
}
for _, test := range loadTests {
t.Run(test.description, func(t *testing.T) {
if test.isvalid {
got, err := testCM.GetObject(ctx, config.BucketName, test.loadKey, 0, 1024)
assert.NoError(t, err)
contentData, err := io.ReadAll(got)
assert.NoError(t, err)
assert.Equal(t, len(contentData), len(test.expectedValue))
assert.Equal(t, test.expectedValue, contentData)
statSize, err := testCM.StatObject(ctx, config.BucketName, test.loadKey)
assert.NoError(t, err)
assert.Equal(t, statSize, int64(len(contentData)))
_, err = testCM.GetObject(ctx, config.BucketName, test.loadKey, 1, 1023)
assert.NoError(t, err)
} else {
got, err := testCM.GetObject(ctx, config.BucketName, test.loadKey, 0, 1024)
assert.Error(t, err)
assert.Empty(t, got)
}
})
}
loadWithPrefixTests := []struct {
isvalid bool
prefix string
expectedValue [][]byte
description string
}{
{true, "abc", [][]byte{[]byte("123"), []byte("1234")}, "load with valid prefix abc"},
{true, "key_", [][]byte{[]byte("111"), []byte("222"), []byte("333")}, "load with valid prefix key_"},
{true, "prefix", [][]byte{}, "load with valid but not exist prefix prefix"},
}
for _, test := range loadWithPrefixTests {
t.Run(test.description, func(t *testing.T) {
gotk, _, err := listAllObjectsWithPrefixAtBucket(ctx, testCM, config.BucketName,
test.prefix, false)
assert.NoError(t, err)
assert.Equal(t, len(test.expectedValue), len(gotk))
for _, key := range gotk {
err := testCM.RemoveObject(ctx, config.BucketName, key)
assert.NoError(t, err)
}
})
}
})
t.Run("test list", func(t *testing.T) {
testCM, err := newGcpNativeObjectStorageWithConfig(ctx, &config)
assert.Equal(t, err, nil)
defer testCM.DeleteBucket(ctx, config.BucketName)
prepareTests := []struct {
valid bool
key string
value []byte
}{
{false, "abc/", []byte("123")},
{true, "abc/d", []byte("1234")},
{true, "abc/e/d", []byte("12354")},
{true, "key_/1/1", []byte("111")},
{true, "key_/1/2", []byte("222")},
{true, "key_/2/3", []byte("333")},
}
for _, test := range prepareTests {
err := testCM.PutObject(ctx, config.BucketName, test.key, bytes.NewReader(test.value),
int64(len(test.value)))
require.Nil(t, err)
if !test.valid {
err := testCM.RemoveObject(ctx, config.BucketName, test.key)
require.Nil(t, err)
}
}
insertWithPrefixTests := []struct {
recursive bool
prefix string
expectedValue []string
}{
{true, "abc/", []string{"abc/d", "abc/e/d"}},
{true, "key_/", []string{"key_/1/1", "key_/1/2", "key_/2/3"}},
{false, "abc/", []string{"abc/d", "abc/e/"}},
{false, "key_/", []string{"key_/1/", "key_/2/"}},
}
for _, test := range insertWithPrefixTests {
t.Run(fmt.Sprintf("prefix: %s, recursive: %t", test.prefix, test.recursive), func(t *testing.T) {
gotk, _, err := listAllObjectsWithPrefixAtBucket(ctx, testCM, config.BucketName,
test.prefix, test.recursive)
assert.NoError(t, err)
assert.Equal(t, len(test.expectedValue), len(gotk))
for _, key := range gotk {
assert.Contains(t, test.expectedValue, key)
}
})
}
})
}
func TestGcpNativeReadFile(t *testing.T) {
ctx := context.Background()
bucketName := "test-bucket"
c := &objectstorage.Config{
Address: "storage.gcs.127.0.0.1.nip.io:4443",
BucketName: bucketName,
CreateBucket: true,
UseIAM: false,
CloudProvider: "gcpnative",
UseSSL: false,
GcpNativeWithoutAuth: true,
}
rcm, err := NewRemoteChunkManager(ctx, c)
t.Run("Read", func(t *testing.T) {
filePath := "test-Read"
data := []byte("Test data for Read.")
err = rcm.Write(ctx, filePath, data)
assert.NoError(t, err)
defer rcm.Remove(ctx, filePath)
reader, err := rcm.Reader(ctx, filePath)
assert.NoError(t, err)
buffer := make([]byte, 4)
n, err := reader.Read(buffer)
assert.NoError(t, err)
assert.Equal(t, 4, n)
assert.Equal(t, "Test", string(buffer))
buffer = make([]byte, 6)
n, err = reader.Read(buffer)
assert.NoError(t, err)
assert.Equal(t, 6, n)
assert.Equal(t, " data ", string(buffer))
buffer = make([]byte, 40)
n, err = reader.Read(buffer)
assert.Error(t, err)
assert.Equal(t, 9, n)
assert.Equal(t, "for Read.", string(buffer[:9]))
})
t.Run("ReadAt", func(t *testing.T) {
filePath := "test-ReadAt"
data := []byte("Test data for ReadAt.")
err = rcm.Write(ctx, filePath, data)
assert.NoError(t, err)
defer rcm.Remove(ctx, filePath)
reader, err := rcm.Reader(ctx, filePath)
assert.NoError(t, err)
buffer := make([]byte, 4)
n, err := reader.ReadAt(buffer, 5)
assert.NoError(t, err)
assert.Equal(t, 4, n)
assert.Equal(t, "data", string(buffer))
buffer = make([]byte, 4)
n, err = reader.Read(buffer)
assert.NoError(t, err)
assert.Equal(t, 4, n)
assert.Equal(t, "Test", string(buffer))
buffer = make([]byte, 4)
n, err = reader.ReadAt(buffer, 20)
assert.Error(t, err)
assert.Equal(t, 1, n)
assert.Equal(t, ".", string(buffer[:1]))
buffer = make([]byte, 4)
n, err = reader.ReadAt(buffer, 25)
assert.Error(t, err)
assert.Equal(t, 0, n)
})
t.Run("Seek start", func(t *testing.T) {
filePath := "test-SeekStart"
data := []byte("Test data for Seek start.")
err = rcm.Write(ctx, filePath, data)
assert.NoError(t, err)
defer rcm.Remove(ctx, filePath)
reader, err := rcm.Reader(ctx, filePath)
assert.NoError(t, err)
offset, err := reader.Seek(10, io.SeekStart)
assert.NoError(t, err)
assert.Equal(t, int64(10), offset)
buffer := make([]byte, 4)
n, err := reader.Read(buffer)
assert.NoError(t, err)
assert.Equal(t, 4, n)
assert.Equal(t, "for ", string(buffer))
offset, err = reader.Seek(40, io.SeekStart)
assert.NoError(t, err)
assert.Equal(t, int64(40), offset)
buffer = make([]byte, 4)
n, err = reader.Read(buffer)
assert.Error(t, err)
assert.Equal(t, 0, n)
})
t.Run("Seek current", func(t *testing.T) {
filePath := "test-SeekCurrent"
data := []byte("Test data for Seek current.")
err = rcm.Write(ctx, filePath, data)
assert.NoError(t, err)
defer rcm.Remove(ctx, filePath)
reader, err := rcm.Reader(ctx, filePath)
assert.NoError(t, err)
buffer := make([]byte, 4)
n, err := reader.Read(buffer)
assert.NoError(t, err)
assert.Equal(t, 4, n)
assert.Equal(t, "Test", string(buffer))
offset, err := reader.Seek(10, io.SeekCurrent)
assert.NoError(t, err)
assert.Equal(t, int64(14), offset)
buffer = make([]byte, 4)
n, err = reader.Read(buffer)
assert.NoError(t, err)
assert.Equal(t, 4, n)
assert.Equal(t, "Seek", string(buffer))
offset, err = reader.Seek(40, io.SeekCurrent)
assert.NoError(t, err)
assert.Equal(t, int64(58), offset)
buffer = make([]byte, 4)
n, err = reader.Read(buffer)
assert.Error(t, err)
assert.Equal(t, 0, n)
})
t.Run("Seek end", func(t *testing.T) {
filePath := "test-SeekEnd"
data := []byte("Test data for Seek end.")
err = rcm.Write(ctx, filePath, data)
assert.NoError(t, err)
defer rcm.Remove(ctx, filePath)
reader, err := rcm.Reader(ctx, filePath)
assert.NoError(t, err)
buffer := make([]byte, 4)
n, err := reader.Read(buffer)
assert.NoError(t, err)
assert.Equal(t, 4, n)
assert.Equal(t, "Test", string(buffer))
offset, err := reader.Seek(10, io.SeekEnd)
assert.NoError(t, err)
assert.Equal(t, int64(33), offset)
buffer = make([]byte, 4)
n, err = reader.Read(buffer)
assert.Error(t, err)
assert.Equal(t, 0, n)
offset, err = reader.Seek(10, 3) // Invalid whence
assert.Error(t, err)
assert.Equal(t, int64(0), offset)
})
t.Run("Close", func(t *testing.T) {
filePath := "test-Close"
data := []byte("Test data for Close.")
err = rcm.Write(ctx, filePath, data)
assert.NoError(t, err)
defer rcm.Remove(ctx, filePath)
reader, err := rcm.Reader(ctx, filePath)
assert.NoError(t, err)
err = reader.Close()
assert.NoError(t, err)
})
t.Run("test Copy", func(t *testing.T) {
testCopyRoot := "test_copy_gcp"
// Test successful copy
t.Run("copy file successfully", func(t *testing.T) {
srcKey := testCopyRoot + "/src/file1"
dstKey := testCopyRoot + "/dst/file1"
value := []byte("test data for gcp copy")
// Write source file
err := rcm.Write(ctx, srcKey, value)
require.NoError(t, err)
defer rcm.Remove(ctx, srcKey)
defer rcm.Remove(ctx, dstKey)
// Copy file
err = rcm.Copy(ctx, srcKey, dstKey)
assert.NoError(t, err)
// Verify destination file exists and has correct content
dstData, err := rcm.Read(ctx, dstKey)
assert.NoError(t, err)
assert.Equal(t, value, dstData)
// Verify source file still exists
srcData, err := rcm.Read(ctx, srcKey)
assert.NoError(t, err)
assert.Equal(t, value, srcData)
})
// Test copy with non-existent source
t.Run("copy non-existent source file", func(t *testing.T) {
srcKey := testCopyRoot + "/not_exist/file"
dstKey := testCopyRoot + "/dst/file"
err := rcm.Copy(ctx, srcKey, dstKey)
assert.Error(t, err)
})
// Test copy overwrite existing file
t.Run("copy and overwrite existing file", func(t *testing.T) {
srcKey := testCopyRoot + "/src3/file3"
dstKey := testCopyRoot + "/dst3/file3"
srcValue := []byte("new gcp content")
oldValue := []byte("old gcp content")
// Create destination with old content
err := rcm.Write(ctx, dstKey, oldValue)
require.NoError(t, err)
defer rcm.Remove(ctx, dstKey)
// Create source with new content
err = rcm.Write(ctx, srcKey, srcValue)
require.NoError(t, err)
defer rcm.Remove(ctx, srcKey)
// Copy (should overwrite)
err = rcm.Copy(ctx, srcKey, dstKey)
assert.NoError(t, err)
// Verify destination has new content
dstData, err := rcm.Read(ctx, dstKey)
assert.NoError(t, err)
assert.Equal(t, srcValue, dstData)
})
// Test copy large file
t.Run("copy large file", func(t *testing.T) {
srcKey := testCopyRoot + "/src4/large_file"
dstKey := testCopyRoot + "/dst4/large_file"
// Create 5MB file
largeData := make([]byte, 5*1024*1024)
for i := range largeData {
largeData[i] = byte(i % 256)
}
err := rcm.Write(ctx, srcKey, largeData)
require.NoError(t, err)
defer rcm.Remove(ctx, srcKey)
defer rcm.Remove(ctx, dstKey)
// Copy large file
err = rcm.Copy(ctx, srcKey, dstKey)
assert.NoError(t, err)
// Verify content
dstData, err := rcm.Read(ctx, dstKey)
assert.NoError(t, err)
assert.Equal(t, largeData, dstData)
})
// Test copy empty file
t.Run("copy empty file", func(t *testing.T) {
srcKey := testCopyRoot + "/src5/empty_file"
dstKey := testCopyRoot + "/dst5/empty_file"
emptyData := []byte{}
// Write empty file
err := rcm.Write(ctx, srcKey, emptyData)
require.NoError(t, err)
defer rcm.Remove(ctx, srcKey)
defer rcm.Remove(ctx, dstKey)
// Copy empty file
err = rcm.Copy(ctx, srcKey, dstKey)
assert.NoError(t, err)
// Verify destination exists and has size 0
size, err := rcm.Size(ctx, dstKey)
assert.NoError(t, err)
assert.Equal(t, int64(0), size)
})
// Test copy with nested path
t.Run("copy file with nested path", func(t *testing.T) {
srcKey := testCopyRoot + "/src6/file6"
dstKey := testCopyRoot + "/dst6/nested/deep/path/file6"
value := []byte("test data for nested path copy")
// Write source file
err := rcm.Write(ctx, srcKey, value)
require.NoError(t, err)
defer rcm.Remove(ctx, srcKey)
defer rcm.Remove(ctx, dstKey)
// Copy to nested path
err = rcm.Copy(ctx, srcKey, dstKey)
assert.NoError(t, err)
// Verify destination file exists and has correct content
dstData, err := rcm.Read(ctx, dstKey)
assert.NoError(t, err)
assert.Equal(t, value, dstData)
})
})
}