mirror of
https://gitee.com/milvus-io/milvus.git
synced 2025-12-08 01:58:34 +08:00
Concurrent load segment stats log when initialize dataSyncService (#18125)
Signed-off-by: Congqi Xia <congqi.xia@zilliz.com>
This commit is contained in:
parent
dd310bbce9
commit
8f1ba6a0bb
@ -30,6 +30,7 @@ import (
|
|||||||
"github.com/milvus-io/milvus/internal/proto/datapb"
|
"github.com/milvus-io/milvus/internal/proto/datapb"
|
||||||
"github.com/milvus-io/milvus/internal/storage"
|
"github.com/milvus-io/milvus/internal/storage"
|
||||||
"github.com/milvus-io/milvus/internal/types"
|
"github.com/milvus-io/milvus/internal/types"
|
||||||
|
"github.com/milvus-io/milvus/internal/util/concurrency"
|
||||||
"github.com/milvus-io/milvus/internal/util/flowgraph"
|
"github.com/milvus-io/milvus/internal/util/flowgraph"
|
||||||
)
|
)
|
||||||
|
|
||||||
@ -52,6 +53,9 @@ type dataSyncService struct {
|
|||||||
flushManager flushManager // flush manager handles flush process
|
flushManager flushManager // flush manager handles flush process
|
||||||
chunkManager storage.ChunkManager
|
chunkManager storage.ChunkManager
|
||||||
compactor *compactionExecutor // reference to compaction executor
|
compactor *compactionExecutor // reference to compaction executor
|
||||||
|
|
||||||
|
// concurrent add segments, reduce time to load delta log from oss
|
||||||
|
ioPool *concurrency.Pool
|
||||||
}
|
}
|
||||||
|
|
||||||
func newDataSyncService(ctx context.Context,
|
func newDataSyncService(ctx context.Context,
|
||||||
@ -72,6 +76,14 @@ func newDataSyncService(ctx context.Context,
|
|||||||
return nil, errors.New("Nil input")
|
return nil, errors.New("Nil input")
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Initialize io cocurrency pool
|
||||||
|
log.Info("initialize io concurrency pool", zap.String("vchannel", vchan.GetChannelName()), zap.Int("ioConcurrency", Params.DataNodeCfg.IOConcurrency))
|
||||||
|
ioPool, err := concurrency.NewPool(Params.DataNodeCfg.IOConcurrency)
|
||||||
|
if err != nil {
|
||||||
|
log.Error("failed to create goroutine pool for dataSyncService", zap.Error(err))
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
ctx1, cancel := context.WithCancel(ctx)
|
ctx1, cancel := context.WithCancel(ctx)
|
||||||
|
|
||||||
service := &dataSyncService{
|
service := &dataSyncService{
|
||||||
@ -90,6 +102,7 @@ func newDataSyncService(ctx context.Context,
|
|||||||
flushingSegCache: flushingSegCache,
|
flushingSegCache: flushingSegCache,
|
||||||
chunkManager: chunkManager,
|
chunkManager: chunkManager,
|
||||||
compactor: compactor,
|
compactor: compactor,
|
||||||
|
ioPool: ioPool,
|
||||||
}
|
}
|
||||||
|
|
||||||
if err := service.initNodes(vchan); err != nil {
|
if err := service.initNodes(vchan); err != nil {
|
||||||
@ -189,6 +202,12 @@ func (dsService *dataSyncService) initNodes(vchanInfo *datapb.VchannelInfo) erro
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
flushedSegmentInfos, err := dsService.getSegmentInfos(vchanInfo.GetFlushedSegmentIds())
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
futures := make([]*concurrency.Future, 0, len(unflushedSegmentInfos)+len(flushedSegmentInfos))
|
||||||
for _, us := range unflushedSegmentInfos {
|
for _, us := range unflushedSegmentInfos {
|
||||||
if us.CollectionID != dsService.collectionID ||
|
if us.CollectionID != dsService.collectionID ||
|
||||||
us.GetInsertChannel() != vchanInfo.ChannelName {
|
us.GetInsertChannel() != vchanInfo.ChannelName {
|
||||||
@ -213,16 +232,18 @@ func (dsService *dataSyncService) initNodes(vchanInfo *datapb.VchannelInfo) erro
|
|||||||
pos: *us.GetDmlPosition(),
|
pos: *us.GetDmlPosition(),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if err := dsService.replica.addNormalSegment(us.GetID(), us.CollectionID, us.PartitionID, us.GetInsertChannel(),
|
// avoid closure capture iteration variable
|
||||||
us.GetNumOfRows(), us.Statslogs, cp, vchanInfo.GetSeekPosition().GetTimestamp()); err != nil {
|
segment := us
|
||||||
return err
|
future := dsService.ioPool.Submit(func() (interface{}, error) {
|
||||||
}
|
if err := dsService.replica.addNormalSegment(segment.GetID(), segment.GetCollectionID(), segment.GetPartitionID(), segment.GetInsertChannel(),
|
||||||
|
segment.GetNumOfRows(), segment.GetStatslogs(), cp, vchanInfo.GetSeekPosition().GetTimestamp()); err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
return nil, nil
|
||||||
|
})
|
||||||
|
futures = append(futures, future)
|
||||||
}
|
}
|
||||||
|
|
||||||
flushedSegmentInfos, err := dsService.getSegmentInfos(vchanInfo.GetFlushedSegmentIds())
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
for _, fs := range flushedSegmentInfos {
|
for _, fs := range flushedSegmentInfos {
|
||||||
if fs.CollectionID != dsService.collectionID ||
|
if fs.CollectionID != dsService.collectionID ||
|
||||||
fs.GetInsertChannel() != vchanInfo.ChannelName {
|
fs.GetInsertChannel() != vchanInfo.ChannelName {
|
||||||
@ -240,10 +261,21 @@ func (dsService *dataSyncService) initNodes(vchanInfo *datapb.VchannelInfo) erro
|
|||||||
zap.Int64("SegmentID", fs.GetID()),
|
zap.Int64("SegmentID", fs.GetID()),
|
||||||
zap.Int64("NumOfRows", fs.GetNumOfRows()),
|
zap.Int64("NumOfRows", fs.GetNumOfRows()),
|
||||||
)
|
)
|
||||||
if err := dsService.replica.addFlushedSegment(fs.GetID(), fs.CollectionID, fs.PartitionID, fs.GetInsertChannel(),
|
// avoid closure capture iteration variable
|
||||||
fs.GetNumOfRows(), fs.Statslogs, vchanInfo.GetSeekPosition().GetTimestamp()); err != nil {
|
segment := fs
|
||||||
return err
|
future := dsService.ioPool.Submit(func() (interface{}, error) {
|
||||||
}
|
if err := dsService.replica.addFlushedSegment(segment.GetID(), segment.GetCollectionID(), segment.GetPartitionID(), segment.GetInsertChannel(),
|
||||||
|
segment.GetNumOfRows(), segment.GetStatslogs(), vchanInfo.GetSeekPosition().GetTimestamp()); err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
return nil, nil
|
||||||
|
})
|
||||||
|
futures = append(futures, future)
|
||||||
|
}
|
||||||
|
|
||||||
|
err = concurrency.AwaitAll(futures...)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
c := &nodeConfig{
|
c := &nodeConfig{
|
||||||
|
|||||||
@ -1101,6 +1101,9 @@ type dataNodeConfig struct {
|
|||||||
// etcd
|
// etcd
|
||||||
ChannelWatchSubPath string
|
ChannelWatchSubPath string
|
||||||
|
|
||||||
|
// io concurrency to fetch stats logs
|
||||||
|
IOConcurrency int
|
||||||
|
|
||||||
CreatedTime time.Time
|
CreatedTime time.Time
|
||||||
UpdatedTime time.Time
|
UpdatedTime time.Time
|
||||||
}
|
}
|
||||||
@ -1114,6 +1117,7 @@ func (p *dataNodeConfig) init(base *BaseTable) {
|
|||||||
p.initInsertBinlogRootPath()
|
p.initInsertBinlogRootPath()
|
||||||
p.initStatsBinlogRootPath()
|
p.initStatsBinlogRootPath()
|
||||||
p.initDeleteBinlogRootPath()
|
p.initDeleteBinlogRootPath()
|
||||||
|
p.initIOConcurrency()
|
||||||
|
|
||||||
p.initChannelWatchPath()
|
p.initChannelWatchPath()
|
||||||
}
|
}
|
||||||
@ -1164,6 +1168,10 @@ func (p *dataNodeConfig) initChannelWatchPath() {
|
|||||||
p.ChannelWatchSubPath = "channelwatch"
|
p.ChannelWatchSubPath = "channelwatch"
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (p *dataNodeConfig) initIOConcurrency() {
|
||||||
|
p.IOConcurrency = p.Base.ParseIntWithDefault("dataNode.dataSync.ioConcurrency", 10)
|
||||||
|
}
|
||||||
|
|
||||||
func (p *dataNodeConfig) SetNodeID(id UniqueID) {
|
func (p *dataNodeConfig) SetNodeID(id UniqueID) {
|
||||||
p.NodeID.Store(id)
|
p.NodeID.Store(id)
|
||||||
}
|
}
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user