mirror of
https://gitee.com/milvus-io/milvus.git
synced 2025-12-06 17:18:35 +08:00
Related to #44956 Add manifest_path field throughout the data path to support LOON Storage V2 manifest tracking. The manifest stores metadata for segment data files and enables the unified Storage V2 FFI interface. Changes include: - Add manifest_path field to SegmentInfo and SaveBinlogPathsRequest proto messages - Add UpdateManifest operator to datacoord meta operations - Update metacache, sync manager, and meta writer to propagate manifest paths - Include manifest_path in segment load info for query coordinator This is part of the Storage V2 FFI interface integration. Signed-off-by: Congqi Xia <congqi.xia@zilliz.com>
186 lines
6.5 KiB
Go
186 lines
6.5 KiB
Go
package syncmgr
|
|
|
|
import (
|
|
"context"
|
|
|
|
"github.com/cockroachdb/errors"
|
|
"github.com/samber/lo"
|
|
"go.uber.org/zap"
|
|
|
|
"github.com/milvus-io/milvus-proto/go-api/v2/commonpb"
|
|
"github.com/milvus-io/milvus/internal/flushcommon/broker"
|
|
"github.com/milvus-io/milvus/internal/flushcommon/metacache"
|
|
storage "github.com/milvus-io/milvus/internal/storage"
|
|
"github.com/milvus-io/milvus/pkg/v2/log"
|
|
"github.com/milvus-io/milvus/pkg/v2/proto/datapb"
|
|
"github.com/milvus-io/milvus/pkg/v2/util/commonpbutil"
|
|
"github.com/milvus-io/milvus/pkg/v2/util/merr"
|
|
"github.com/milvus-io/milvus/pkg/v2/util/retry"
|
|
)
|
|
|
|
// MetaWriter is the interface for SyncManager to write segment sync meta.
|
|
type MetaWriter interface {
|
|
UpdateSync(context.Context, *SyncTask) error
|
|
DropChannel(context.Context, string) error
|
|
}
|
|
|
|
type brokerMetaWriter struct {
|
|
broker broker.Broker
|
|
opts []retry.Option
|
|
serverID int64
|
|
}
|
|
|
|
func BrokerMetaWriter(broker broker.Broker, serverID int64, opts ...retry.Option) MetaWriter {
|
|
return &brokerMetaWriter{
|
|
broker: broker,
|
|
serverID: serverID,
|
|
opts: opts,
|
|
}
|
|
}
|
|
|
|
func (b *brokerMetaWriter) UpdateSync(ctx context.Context, pack *SyncTask) error {
|
|
checkPoints := []*datapb.CheckPoint{}
|
|
// only current segment checkpoint info
|
|
segment, ok := pack.metacache.GetSegmentByID(pack.segmentID)
|
|
if !ok {
|
|
return merr.WrapErrSegmentNotFound(pack.segmentID)
|
|
}
|
|
|
|
insertFieldBinlogs := append(segment.Binlogs(), storage.SortFieldBinlogs(pack.insertBinlogs)...)
|
|
statsFieldBinlogs := append(segment.Statslogs(), lo.MapToSlice(pack.statsBinlogs, func(_ int64, fieldBinlog *datapb.FieldBinlog) *datapb.FieldBinlog { return fieldBinlog })...)
|
|
|
|
deltaFieldBinlogs := segment.Deltalogs()
|
|
if pack.deltaBinlog != nil && len(pack.deltaBinlog.Binlogs) > 0 {
|
|
deltaFieldBinlogs = append(deltaFieldBinlogs, pack.deltaBinlog)
|
|
}
|
|
|
|
deltaBm25StatsBinlogs := segment.Bm25logs()
|
|
if len(pack.bm25Binlogs) > 0 {
|
|
deltaBm25StatsBinlogs = append(segment.Bm25logs(), lo.MapToSlice(pack.bm25Binlogs, func(_ int64, fieldBinlog *datapb.FieldBinlog) *datapb.FieldBinlog { return fieldBinlog })...)
|
|
}
|
|
|
|
checkPoints = append(checkPoints, &datapb.CheckPoint{
|
|
SegmentID: pack.segmentID,
|
|
NumOfRows: segment.FlushedRows() + pack.batchRows,
|
|
Position: pack.checkpoint,
|
|
})
|
|
|
|
// Get not reported L1's start positions
|
|
startPos := lo.Map(pack.metacache.GetSegmentsBy(
|
|
metacache.WithSegmentState(commonpb.SegmentState_Growing, commonpb.SegmentState_Sealed, commonpb.SegmentState_Flushing),
|
|
metacache.WithLevel(datapb.SegmentLevel_L1), metacache.WithStartPosNotRecorded()),
|
|
func(info *metacache.SegmentInfo, _ int) *datapb.SegmentStartPosition {
|
|
return &datapb.SegmentStartPosition{
|
|
SegmentID: info.SegmentID(),
|
|
StartPosition: info.StartPosition(),
|
|
}
|
|
})
|
|
|
|
// L0 brings its own start position
|
|
if segment.Level() == datapb.SegmentLevel_L0 {
|
|
startPos = append(startPos, &datapb.SegmentStartPosition{SegmentID: pack.segmentID, StartPosition: pack.StartPosition()})
|
|
}
|
|
|
|
getBinlogNum := func(fBinlog *datapb.FieldBinlog) int { return len(fBinlog.GetBinlogs()) }
|
|
log.Info("SaveBinlogPath",
|
|
zap.Int64("SegmentID", pack.segmentID),
|
|
zap.Int64("CollectionID", pack.collectionID),
|
|
zap.Int64("ParitionID", pack.partitionID),
|
|
zap.Any("startPos", startPos),
|
|
zap.Any("checkPoints", checkPoints),
|
|
zap.Int("binlogNum", lo.SumBy(insertFieldBinlogs, getBinlogNum)),
|
|
zap.Int("statslogNum", lo.SumBy(statsFieldBinlogs, getBinlogNum)),
|
|
zap.Int("deltalogNum", lo.SumBy(deltaFieldBinlogs, getBinlogNum)),
|
|
zap.Int("bm25logNum", lo.SumBy(deltaBm25StatsBinlogs, getBinlogNum)),
|
|
zap.String("manifestPath", pack.manifestPath),
|
|
zap.String("vChannelName", pack.channelName),
|
|
)
|
|
|
|
req := &datapb.SaveBinlogPathsRequest{
|
|
Base: commonpbutil.NewMsgBase(
|
|
commonpbutil.WithMsgType(0),
|
|
commonpbutil.WithMsgID(0),
|
|
commonpbutil.WithSourceID(b.serverID),
|
|
),
|
|
SegmentID: pack.segmentID,
|
|
CollectionID: pack.collectionID,
|
|
PartitionID: pack.partitionID,
|
|
Field2BinlogPaths: insertFieldBinlogs,
|
|
Field2StatslogPaths: statsFieldBinlogs,
|
|
Field2Bm25LogPaths: deltaBm25StatsBinlogs,
|
|
Deltalogs: deltaFieldBinlogs,
|
|
|
|
CheckPoints: checkPoints,
|
|
|
|
StartPositions: startPos,
|
|
Flushed: pack.pack.isFlush,
|
|
Dropped: pack.pack.isDrop,
|
|
Channel: pack.channelName,
|
|
SegLevel: pack.level,
|
|
StorageVersion: segment.GetStorageVersion(),
|
|
WithFullBinlogs: true,
|
|
ManifestPath: pack.manifestPath,
|
|
}
|
|
err := retry.Handle(ctx, func() (bool, error) {
|
|
err := b.broker.SaveBinlogPaths(ctx, req)
|
|
// Segment not found during stale segment flush. Segment might get compacted already.
|
|
// Stop retry and still proceed to the end, ignoring this error.
|
|
if !pack.pack.isFlush && errors.Is(err, merr.ErrSegmentNotFound) {
|
|
log.Warn("stale segment not found, could be compacted",
|
|
zap.Int64("segmentID", pack.segmentID))
|
|
log.Warn("failed to SaveBinlogPaths",
|
|
zap.Int64("segmentID", pack.segmentID),
|
|
zap.Error(err))
|
|
return false, nil
|
|
}
|
|
// meta error, datanode handles a virtual channel does not belong here
|
|
if errors.IsAny(err, merr.ErrSegmentNotFound, merr.ErrChannelNotFound) {
|
|
log.Warn("meta error found, skip sync and start to drop virtual channel", zap.String("channel", pack.channelName))
|
|
return false, nil
|
|
}
|
|
|
|
if err != nil {
|
|
return !merr.IsCanceledOrTimeout(err), err
|
|
}
|
|
|
|
return false, nil
|
|
}, b.opts...)
|
|
if err != nil {
|
|
log.Warn("failed to SaveBinlogPaths",
|
|
zap.Int64("segmentID", pack.segmentID),
|
|
zap.Error(err))
|
|
return err
|
|
}
|
|
|
|
pack.metacache.UpdateSegments(metacache.SetStartPosRecorded(true), metacache.WithSegmentIDs(lo.Map(startPos, func(pos *datapb.SegmentStartPosition, _ int) int64 { return pos.GetSegmentID() })...))
|
|
pack.metacache.UpdateSegments(metacache.MergeSegmentAction(
|
|
metacache.UpdateBinlogs(insertFieldBinlogs),
|
|
metacache.UpdateStatslogs(statsFieldBinlogs),
|
|
metacache.UpdateDeltalogs(deltaFieldBinlogs),
|
|
metacache.UpdateBm25logs(deltaBm25StatsBinlogs),
|
|
), metacache.WithSegmentIDs(pack.segmentID))
|
|
return nil
|
|
}
|
|
|
|
func (b *brokerMetaWriter) DropChannel(ctx context.Context, channelName string) error {
|
|
err := retry.Handle(ctx, func() (bool, error) {
|
|
status, err := b.broker.DropVirtualChannel(context.Background(), &datapb.DropVirtualChannelRequest{
|
|
Base: commonpbutil.NewMsgBase(
|
|
commonpbutil.WithSourceID(b.serverID),
|
|
),
|
|
ChannelName: channelName,
|
|
})
|
|
err = merr.CheckRPCCall(status, err)
|
|
if err != nil {
|
|
return !merr.IsCanceledOrTimeout(err), err
|
|
}
|
|
return false, nil
|
|
}, b.opts...)
|
|
if err != nil {
|
|
log.Warn("failed to DropChannel",
|
|
zap.String("channel", channelName),
|
|
zap.Error(err))
|
|
}
|
|
return err
|
|
}
|