mirror of
https://gitee.com/milvus-io/milvus.git
synced 2025-12-06 09:08:43 +08:00
Cherry-pick from master pr: #45061 #45488 #45803 #46017 #44991 #45132 #45723 #45726 #45798 #45897 #45918 #44998 This feature integrates the Storage V2 (Loon) FFI interface as a unified storage layer for segment loading and index building in Milvus. It enables manifest-based data access, replacing the traditional binlog-based approach with a more efficient columnar storage format. Key changes: ### Segment Self-Managed Loading Architecture - Move segment loading orchestration from Go layer to C++ segcore - Add NewSegmentWithLoadInfo() API for passing load info during segment creation - Implement SetLoadInfo() and Load() methods in SegmentInterface - Support parallel loading of indexed and non-indexed fields - Enable both sealed and growing segments to self-manage loading ### Storage V2 FFI Integration - Integrate milvus-storage library's FFI interface for packed columnar data - Add manifest path support throughout the data path (SegmentInfo, LoadInfo) - Implement ManifestReader for generating manifests from binlogs - Support zero-copy data exchange using Arrow C Data Interface - Add ToCStorageConfig() for Go-to-C storage config conversion ### Manifest-Based Index Building - Extend FileManagerContext to carry loon_ffi_properties - Implement GetFieldDatasFromManifest() using Arrow C Stream interface - Support manifest-based reading in DiskFileManagerImpl and MemFileManagerImpl - Add fallback to traditional segment insert files when manifest unavailable ### Compaction Pipeline Updates - Include manifest path in all compaction task builders (clustering, L0, mix) - Update BulkPackWriterV2 to return manifest path - Propagate manifest metadata through compaction pipeline ### Configuration & Protocol - Add common.storageV2.useLoonFFI config option (default: false) - Add manifest_path field to SegmentLoadInfo and related proto messages - Add manifest field to compaction segment messages ### Bug Fixes - Fix mmap settings not applied during segment load (key typo fix) - Populate index info after segment loading to prevent redundant load tasks - Fix memory corruption by removing premature transaction handle destruction Related issues: #44956, #45060, #39173 ## Individual Cherry-Picked Commits 1. **e1c923b5cc** - fix: apply mmap settings correctly during segment load (#46017) 2. **63b912370b** - enhance: use milvus-storage internal C++ Reader API for Loon FFI (#45897) 3. **bfc192faa5** - enhance: Resolve issues integrating loon FFI (#45918) 4. **fb18564631** - enhance: support manifest-based index building with Loon FFI reader (#45726) 5. **b9ec2392b9** - enhance: integrate StorageV2 FFI interface for manifest-based segment loading (#45798) 6. **66db3c32e6** - enhance: integrate Storage V2 FFI interface for unified storage access (#45723) 7. **ae789273ac** - fix: populate index info after segment loading to prevent redundant load tasks (#45803) 8. **49688b0be2** - enhance: Move segment loading logic from Go layer to segcore for self-managed loading (#45488) 9. **5b2df88bac** - enhance: [StorageV2] Integrate FFI interface for packed reader (#45132) 10. **91ff5706ac** - enhance: [StorageV2] add manifest path support for FFI integration (#44991) 11. **2192bb4a85** - enhance: add NewSegmentWithLoadInfo API to support segment self-managed loading (#45061) 12. **4296b01da0** - enhance: update delta log serialization APIs to integrate storage V2 (#44998) ## Technical Details ### Architecture Changes - **Before**: Go layer orchestrated segment loading, making multiple CGO calls - **After**: Segments autonomously manage loading in C++ layer with single entry point ### Storage Access Pattern - **Before**: Read individual binlog files through Go storage layer - **After**: Read manifest file that references packed columnar data via FFI ### Benefits - Reduced cross-language call overhead - Better resource management at C++ level - Improved I/O performance through batched streaming reads - Cleaner separation of concerns between Go and C++ layers - Foundation for proactive schema evolution handling --------- Signed-off-by: Ted Xu <ted.xu@zilliz.com> Signed-off-by: Congqi Xia <congqi.xia@zilliz.com> Co-authored-by: Ted Xu <ted.xu@zilliz.com>
437 lines
13 KiB
Go
437 lines
13 KiB
Go
// Licensed to the LF AI & Data foundation under one
|
|
// or more contributor license agreements. See the NOTICE file
|
|
// distributed with this work for additional information
|
|
// regarding copyright ownership. The ASF licenses this file
|
|
// to you under the Apache License, Version 2.0 (the
|
|
// "License"); you may not use this file except in compliance
|
|
// with the License. You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
package storage
|
|
|
|
import (
|
|
"context"
|
|
"encoding/base64"
|
|
"fmt"
|
|
sio "io"
|
|
"path"
|
|
"sort"
|
|
|
|
"github.com/samber/lo"
|
|
|
|
"github.com/milvus-io/milvus-proto/go-api/v2/schemapb"
|
|
"github.com/milvus-io/milvus/internal/allocator"
|
|
"github.com/milvus-io/milvus/internal/storagecommon"
|
|
"github.com/milvus-io/milvus/internal/storagev2/packed"
|
|
"github.com/milvus-io/milvus/internal/util/hookutil"
|
|
"github.com/milvus-io/milvus/pkg/v2/log"
|
|
"github.com/milvus-io/milvus/pkg/v2/proto/datapb"
|
|
"github.com/milvus-io/milvus/pkg/v2/proto/indexcgopb"
|
|
"github.com/milvus-io/milvus/pkg/v2/proto/indexpb"
|
|
"github.com/milvus-io/milvus/pkg/v2/util/merr"
|
|
"github.com/milvus-io/milvus/pkg/v2/util/typeutil"
|
|
)
|
|
|
|
const (
|
|
StorageV1 int64 = 0
|
|
StorageV2 int64 = 2
|
|
)
|
|
|
|
type (
|
|
downloaderFn func(ctx context.Context, paths []string) ([][]byte, error)
|
|
uploaderFn func(ctx context.Context, kvs map[string][]byte) error
|
|
)
|
|
|
|
// rwOp is enum alias for rwOption op field.
|
|
type rwOp int32
|
|
|
|
const (
|
|
OpWrite rwOp = 0
|
|
OpRead rwOp = 1
|
|
)
|
|
|
|
type rwOptions struct {
|
|
version int64
|
|
op rwOp
|
|
bufferSize int64
|
|
downloader downloaderFn
|
|
uploader uploaderFn
|
|
multiPartUploadSize int64
|
|
columnGroups []storagecommon.ColumnGroup
|
|
collectionID int64
|
|
storageConfig *indexpb.StorageConfig
|
|
neededFields typeutil.Set[int64]
|
|
useLoonFFI bool
|
|
}
|
|
|
|
func (o *rwOptions) validate() error {
|
|
if o.collectionID == 0 {
|
|
log.Warn("storage config collection id is empty when init BinlogReader")
|
|
// return merr.WrapErrServiceInternal("storage config collection id is empty")
|
|
}
|
|
if o.op == OpWrite && o.uploader == nil {
|
|
return merr.WrapErrServiceInternal("uploader is nil for writer")
|
|
}
|
|
switch o.version {
|
|
case StorageV1:
|
|
if o.op == OpRead && o.downloader == nil {
|
|
return merr.WrapErrServiceInternal("downloader is nil for v1 reader")
|
|
}
|
|
case StorageV2:
|
|
if o.storageConfig == nil {
|
|
return merr.WrapErrServiceInternal("storage config is nil")
|
|
}
|
|
default:
|
|
return merr.WrapErrServiceInternal(fmt.Sprintf("unsupported storage version %d", o.version))
|
|
}
|
|
return nil
|
|
}
|
|
|
|
type RwOption func(*rwOptions)
|
|
|
|
func DefaultWriterOptions() *rwOptions {
|
|
return &rwOptions{
|
|
bufferSize: packed.DefaultWriteBufferSize,
|
|
multiPartUploadSize: packed.DefaultMultiPartUploadSize,
|
|
op: OpWrite,
|
|
}
|
|
}
|
|
|
|
func DefaultReaderOptions() *rwOptions {
|
|
return &rwOptions{
|
|
bufferSize: packed.DefaultReadBufferSize,
|
|
op: OpRead,
|
|
}
|
|
}
|
|
|
|
func WithCollectionID(collID int64) RwOption {
|
|
return func(options *rwOptions) {
|
|
options.collectionID = collID
|
|
}
|
|
}
|
|
|
|
func WithVersion(version int64) RwOption {
|
|
return func(options *rwOptions) {
|
|
options.version = version
|
|
}
|
|
}
|
|
|
|
func WithBufferSize(bufferSize int64) RwOption {
|
|
return func(options *rwOptions) {
|
|
options.bufferSize = bufferSize
|
|
}
|
|
}
|
|
|
|
func WithMultiPartUploadSize(multiPartUploadSize int64) RwOption {
|
|
return func(options *rwOptions) {
|
|
options.multiPartUploadSize = multiPartUploadSize
|
|
}
|
|
}
|
|
|
|
func WithDownloader(downloader func(ctx context.Context, paths []string) ([][]byte, error)) RwOption {
|
|
return func(options *rwOptions) {
|
|
options.downloader = downloader
|
|
}
|
|
}
|
|
|
|
func WithUploader(uploader func(ctx context.Context, kvs map[string][]byte) error) RwOption {
|
|
return func(options *rwOptions) {
|
|
options.uploader = uploader
|
|
}
|
|
}
|
|
|
|
func WithColumnGroups(columnGroups []storagecommon.ColumnGroup) RwOption {
|
|
return func(options *rwOptions) {
|
|
options.columnGroups = columnGroups
|
|
}
|
|
}
|
|
|
|
func WithStorageConfig(storageConfig *indexpb.StorageConfig) RwOption {
|
|
return func(options *rwOptions) {
|
|
options.storageConfig = storageConfig
|
|
}
|
|
}
|
|
|
|
func WithNeededFields(neededFields typeutil.Set[int64]) RwOption {
|
|
return func(options *rwOptions) {
|
|
options.neededFields = neededFields
|
|
}
|
|
}
|
|
|
|
func WithUseLoonFFI(useLoonFFI bool) RwOption {
|
|
return func(options *rwOptions) {
|
|
options.useLoonFFI = useLoonFFI
|
|
}
|
|
}
|
|
|
|
func makeBlobsReader(ctx context.Context, binlogs []*datapb.FieldBinlog, downloader downloaderFn) (ChunkedBlobsReader, error) {
|
|
if len(binlogs) == 0 {
|
|
return func() ([]*Blob, error) {
|
|
return nil, sio.EOF
|
|
}, nil
|
|
}
|
|
sort.Slice(binlogs, func(i, j int) bool {
|
|
return binlogs[i].FieldID < binlogs[j].FieldID
|
|
})
|
|
for _, binlog := range binlogs {
|
|
sort.Slice(binlog.Binlogs, func(i, j int) bool {
|
|
return binlog.Binlogs[i].LogID < binlog.Binlogs[j].LogID
|
|
})
|
|
}
|
|
nChunks := len(binlogs[0].Binlogs)
|
|
chunks := make([][]string, nChunks) // i is chunkid, j is fieldid
|
|
missingChunks := lo.Map(binlogs, func(binlog *datapb.FieldBinlog, _ int) int {
|
|
return nChunks - len(binlog.Binlogs)
|
|
})
|
|
for i := range nChunks {
|
|
chunks[i] = make([]string, 0, len(binlogs))
|
|
for j, binlog := range binlogs {
|
|
if i >= missingChunks[j] {
|
|
idx := i - missingChunks[j]
|
|
chunks[i] = append(chunks[i], binlog.Binlogs[idx].LogPath)
|
|
}
|
|
}
|
|
}
|
|
// verify if the chunks order is correct.
|
|
// the zig-zag order should have a (strict) increasing order on logids.
|
|
// lastLogID := int64(-1)
|
|
// for _, paths := range chunks {
|
|
// lastFieldID := int64(-1)
|
|
// for _, path := range paths {
|
|
// _, _, _, fieldID, logID, ok := metautil.ParseInsertLogPath(path)
|
|
// if !ok {
|
|
// return nil, merr.WrapErrIoFailedReason(fmt.Sprintf("malformed log path %s", path))
|
|
// }
|
|
// if fieldID < lastFieldID {
|
|
// return nil, merr.WrapErrIoFailedReason(fmt.Sprintf("unaligned log path %s, fieldID %d less than lastFieldID %d", path, fieldID, lastFieldID))
|
|
// }
|
|
// if logID < lastLogID {
|
|
// return nil, merr.WrapErrIoFailedReason(fmt.Sprintf("unaligned log path %s, logID %d less than lastLogID %d", path, logID, lastLogID))
|
|
// }
|
|
// lastLogID = logID
|
|
// lastFieldID = fieldID
|
|
// }
|
|
// }
|
|
|
|
chunkPos := 0
|
|
return func() ([]*Blob, error) {
|
|
if chunkPos >= nChunks {
|
|
return nil, sio.EOF
|
|
}
|
|
|
|
vals, err := downloader(ctx, chunks[chunkPos])
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
blobs := make([]*Blob, 0, len(vals))
|
|
for i := range vals {
|
|
blobs = append(blobs, &Blob{
|
|
Key: chunks[chunkPos][i],
|
|
Value: vals[i],
|
|
})
|
|
}
|
|
chunkPos++
|
|
return blobs, nil
|
|
}, nil
|
|
}
|
|
|
|
func NewBinlogRecordReader(ctx context.Context, binlogs []*datapb.FieldBinlog, schema *schemapb.CollectionSchema, option ...RwOption) (rr RecordReader, err error) {
|
|
rwOptions := DefaultReaderOptions()
|
|
for _, opt := range option {
|
|
opt(rwOptions)
|
|
}
|
|
if err := rwOptions.validate(); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
binlogReaderOpts := []BinlogReaderOption{}
|
|
var pluginContext *indexcgopb.StoragePluginContext
|
|
if hookutil.IsClusterEncyptionEnabled() {
|
|
if ez := hookutil.GetEzByCollProperties(schema.GetProperties(), rwOptions.collectionID); ez != nil {
|
|
binlogReaderOpts = append(binlogReaderOpts, WithReaderDecryptionContext(ez.EzID, ez.CollectionID))
|
|
|
|
unsafe := hookutil.GetCipher().GetUnsafeKey(ez.EzID, ez.CollectionID)
|
|
if len(unsafe) > 0 {
|
|
pluginContext = &indexcgopb.StoragePluginContext{
|
|
EncryptionZoneId: ez.EzID,
|
|
CollectionId: ez.CollectionID,
|
|
EncryptionKey: base64.StdEncoding.EncodeToString(unsafe),
|
|
}
|
|
}
|
|
}
|
|
}
|
|
switch rwOptions.version {
|
|
case StorageV1:
|
|
var blobsReader ChunkedBlobsReader
|
|
blobsReader, err = makeBlobsReader(ctx, binlogs, rwOptions.downloader)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
rr = newIterativeCompositeBinlogRecordReader(schema, rwOptions.neededFields, blobsReader, binlogReaderOpts...)
|
|
case StorageV2:
|
|
if len(binlogs) <= 0 {
|
|
return nil, sio.EOF
|
|
}
|
|
sort.Slice(binlogs, func(i, j int) bool {
|
|
return binlogs[i].GetFieldID() < binlogs[j].GetFieldID()
|
|
})
|
|
|
|
binlogLists := lo.Map(binlogs, func(fieldBinlog *datapb.FieldBinlog, _ int) []*datapb.Binlog {
|
|
return fieldBinlog.GetBinlogs()
|
|
})
|
|
bucketName := rwOptions.storageConfig.BucketName
|
|
paths := make([][]string, len(binlogLists[0]))
|
|
for _, binlogs := range binlogLists {
|
|
for j, binlog := range binlogs {
|
|
logPath := binlog.GetLogPath()
|
|
if rwOptions.storageConfig.StorageType != "local" {
|
|
logPath = path.Join(bucketName, logPath)
|
|
}
|
|
paths[j] = append(paths[j], logPath)
|
|
}
|
|
}
|
|
// FIXME: add needed fields support
|
|
rr = newIterativePackedRecordReader(paths, schema, rwOptions.bufferSize, rwOptions.storageConfig, pluginContext)
|
|
default:
|
|
return nil, merr.WrapErrServiceInternal(fmt.Sprintf("unsupported storage version %d", rwOptions.version))
|
|
}
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
return rr, nil
|
|
}
|
|
|
|
func NewManifestRecordReader(ctx context.Context, manifestPath string, schema *schemapb.CollectionSchema, option ...RwOption) (rr RecordReader, err error) {
|
|
rwOptions := DefaultReaderOptions()
|
|
for _, opt := range option {
|
|
opt(rwOptions)
|
|
}
|
|
if err := rwOptions.validate(); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
var pluginContext *indexcgopb.StoragePluginContext
|
|
if hookutil.IsClusterEncyptionEnabled() {
|
|
if ez := hookutil.GetEzByCollProperties(schema.GetProperties(), rwOptions.collectionID); ez != nil {
|
|
unsafe := hookutil.GetCipher().GetUnsafeKey(ez.EzID, ez.CollectionID)
|
|
if len(unsafe) > 0 {
|
|
pluginContext = &indexcgopb.StoragePluginContext{
|
|
EncryptionZoneId: ez.EzID,
|
|
CollectionId: ez.CollectionID,
|
|
EncryptionKey: string(unsafe),
|
|
}
|
|
}
|
|
}
|
|
}
|
|
return NewRecordReaderFromManifest(manifestPath, schema, rwOptions.bufferSize, rwOptions.storageConfig, pluginContext)
|
|
}
|
|
|
|
func NewBinlogRecordWriter(ctx context.Context, collectionID, partitionID, segmentID UniqueID,
|
|
schema *schemapb.CollectionSchema, allocator allocator.Interface, chunkSize uint64, maxRowNum int64,
|
|
option ...RwOption,
|
|
) (BinlogRecordWriter, error) {
|
|
rwOptions := DefaultWriterOptions()
|
|
option = append(option, WithCollectionID(collectionID))
|
|
for _, opt := range option {
|
|
opt(rwOptions)
|
|
}
|
|
|
|
if err := rwOptions.validate(); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
blobsWriter := func(blobs []*Blob) error {
|
|
kvs := make(map[string][]byte, len(blobs))
|
|
for _, blob := range blobs {
|
|
kvs[blob.Key] = blob.Value
|
|
}
|
|
return rwOptions.uploader(ctx, kvs)
|
|
}
|
|
|
|
opts := []StreamWriterOption{}
|
|
var pluginContext *indexcgopb.StoragePluginContext
|
|
if hookutil.IsClusterEncyptionEnabled() {
|
|
ez := hookutil.GetEzByCollProperties(schema.GetProperties(), collectionID)
|
|
if ez != nil {
|
|
encryptor, edek, err := hookutil.GetCipher().GetEncryptor(ez.EzID, ez.CollectionID)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
opts = append(opts, GetEncryptionOptions(ez.EzID, edek, encryptor)...)
|
|
|
|
unsafe := hookutil.GetCipher().GetUnsafeKey(ez.EzID, ez.CollectionID)
|
|
if len(unsafe) > 0 {
|
|
pluginContext = &indexcgopb.StoragePluginContext{
|
|
EncryptionZoneId: ez.EzID,
|
|
CollectionId: ez.CollectionID,
|
|
EncryptionKey: base64.StdEncoding.EncodeToString(unsafe),
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
switch rwOptions.version {
|
|
case StorageV1:
|
|
rootPath := rwOptions.storageConfig.GetRootPath()
|
|
return newCompositeBinlogRecordWriter(collectionID, partitionID, segmentID, schema,
|
|
blobsWriter, allocator, chunkSize, rootPath, maxRowNum, opts...,
|
|
)
|
|
case StorageV2:
|
|
if rwOptions.useLoonFFI {
|
|
return newPackedManifestRecordWriter(collectionID, partitionID, segmentID, schema,
|
|
blobsWriter, allocator, maxRowNum,
|
|
rwOptions.bufferSize, rwOptions.multiPartUploadSize, rwOptions.columnGroups,
|
|
rwOptions.storageConfig,
|
|
pluginContext)
|
|
} else {
|
|
return newPackedBinlogRecordWriter(collectionID, partitionID, segmentID, schema,
|
|
blobsWriter, allocator, maxRowNum,
|
|
rwOptions.bufferSize, rwOptions.multiPartUploadSize, rwOptions.columnGroups,
|
|
rwOptions.storageConfig,
|
|
pluginContext,
|
|
)
|
|
}
|
|
}
|
|
return nil, merr.WrapErrServiceInternal(fmt.Sprintf("unsupported storage version %d", rwOptions.version))
|
|
}
|
|
|
|
func NewDeltalogWriter(
|
|
ctx context.Context,
|
|
collectionID, partitionID, segmentID, logID UniqueID,
|
|
pkType schemapb.DataType,
|
|
path string,
|
|
option ...RwOption,
|
|
) (RecordWriter, error) {
|
|
rwOptions := DefaultWriterOptions()
|
|
for _, opt := range option {
|
|
opt(rwOptions)
|
|
}
|
|
if err := rwOptions.validate(); err != nil {
|
|
return nil, err
|
|
}
|
|
return NewLegacyDeltalogWriter(collectionID, partitionID, segmentID, logID, pkType, rwOptions.uploader, path)
|
|
}
|
|
|
|
func NewDeltalogReader(
|
|
pkField *schemapb.FieldSchema,
|
|
paths []string,
|
|
option ...RwOption,
|
|
) (RecordReader, error) {
|
|
rwOptions := DefaultReaderOptions()
|
|
for _, opt := range option {
|
|
opt(rwOptions)
|
|
}
|
|
if err := rwOptions.validate(); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
return NewLegacyDeltalogReader(pkField, rwOptions.downloader, paths)
|
|
}
|