milvus/internal/indexnode/taskinfo_ops_test.go
cai.zhang 2c9bb4dfa3
feat: Support stats task to sort segment by PK (#35054)
issue: #33744 

This PR includes the following changes:
1. Added a new task type to the task scheduler in datacoord: stats task,
which sorts segments by primary key.
2. Implemented segment sorting in indexnode.
3. Added a new field `FieldStatsLog` to SegmentInfo to store token index
information.

---------

Signed-off-by: Cai Zhang <cai.zhang@zilliz.com>
2024-09-02 14:19:03 +08:00

111 lines
3.2 KiB
Go

// Licensed to the LF AI & Data foundation under one
// or more contributor license agreements. See the NOTICE file
// distributed with this work for additional information
// regarding copyright ownership. The ASF licenses this file
// to you under the Apache License, Version 2.0 (the
// "License"); you may not use this file except in compliance
// with the License. You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
package indexnode
import (
"context"
"testing"
"github.com/stretchr/testify/suite"
"github.com/milvus-io/milvus/internal/proto/datapb"
"github.com/milvus-io/milvus/internal/proto/indexpb"
)
type statsTaskInfoSuite struct {
suite.Suite
ctx context.Context
node *IndexNode
cluster string
taskID int64
}
func Test_statsTaskInfoSuite(t *testing.T) {
suite.Run(t, new(statsTaskInfoSuite))
}
func (s *statsTaskInfoSuite) SetupSuite() {
s.node = &IndexNode{
loopCtx: context.Background(),
statsTasks: make(map[taskKey]*statsTaskInfo),
}
s.cluster = "test"
s.taskID = 100
}
func (s *statsTaskInfoSuite) Test_Methods() {
s.Run("loadOrStoreStatsTask", func() {
_, cancel := context.WithCancel(s.node.loopCtx)
info := &statsTaskInfo{
cancel: cancel,
state: indexpb.JobState_JobStateInProgress,
}
reInfo := s.node.loadOrStoreStatsTask(s.cluster, s.taskID, info)
s.Nil(reInfo)
reInfo = s.node.loadOrStoreStatsTask(s.cluster, s.taskID, info)
s.Equal(indexpb.JobState_JobStateInProgress, reInfo.state)
})
s.Run("getStatsTaskState", func() {
s.Equal(indexpb.JobState_JobStateInProgress, s.node.getStatsTaskState(s.cluster, s.taskID))
s.Equal(indexpb.JobState_JobStateNone, s.node.getStatsTaskState(s.cluster, s.taskID+1))
})
s.Run("storeStatsTaskState", func() {
s.node.storeStatsTaskState(s.cluster, s.taskID, indexpb.JobState_JobStateFinished, "finished")
s.Equal(indexpb.JobState_JobStateFinished, s.node.getStatsTaskState(s.cluster, s.taskID))
})
s.Run("storeStatsResult", func() {
s.node.storeStatsResult(s.cluster, s.taskID, 1, 2, 3, "ch1", 65535,
[]*datapb.FieldBinlog{{FieldID: 100, Binlogs: []*datapb.Binlog{{LogID: 1}}}},
[]*datapb.FieldBinlog{{FieldID: 100, Binlogs: []*datapb.Binlog{{LogID: 2}}}},
map[int64]*datapb.TextIndexStats{
100: {
FieldID: 100,
Version: 1,
Files: []string{"file1"},
LogSize: 1024,
MemorySize: 1024,
},
},
)
})
s.Run("getStatsTaskInfo", func() {
taskInfo := s.node.getStatsTaskInfo(s.cluster, s.taskID)
s.Equal(indexpb.JobState_JobStateFinished, taskInfo.state)
s.Equal(int64(1), taskInfo.collID)
s.Equal(int64(2), taskInfo.partID)
s.Equal(int64(3), taskInfo.segID)
s.Equal("ch1", taskInfo.insertChannel)
s.Equal(int64(65535), taskInfo.numRows)
})
s.Run("deleteStatsTaskInfos", func() {
s.node.deleteStatsTaskInfos(s.ctx, []taskKey{{ClusterID: s.cluster, TaskID: s.taskID}})
s.Nil(s.node.getStatsTaskInfo(s.cluster, s.taskID))
})
}