mirror of
https://gitee.com/milvus-io/milvus.git
synced 2025-12-07 01:28:27 +08:00
1. Enable Milvus to read cipher configs 2. Enable cipher plugin in binlog reader and writer 3. Add a testCipher for unittests 4. Support pooling for datanode 5. Add encryption in storagev2 See also: #40321 Signed-off-by: yangxuan <xuan.yang@zilliz.com> --------- Signed-off-by: yangxuan <xuan.yang@zilliz.com>
238 lines
7.0 KiB
Go
238 lines
7.0 KiB
Go
// Licensed to the LF AI & Data foundation under one
|
|
// or more contributor license agreements. See the NOTICE file
|
|
// distributed with this work for additional information
|
|
// regarding copyright ownership. The ASF licenses this file
|
|
// to you under the Apache License, Version 2.0 (the
|
|
// "License"); you may not use this file except in compliance
|
|
// with the License. You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
package datacoord
|
|
|
|
import (
|
|
"context"
|
|
"time"
|
|
|
|
"go.uber.org/zap"
|
|
|
|
"github.com/milvus-io/milvus-proto/go-api/v2/schemapb"
|
|
"github.com/milvus-io/milvus/internal/datacoord/session"
|
|
globalTask "github.com/milvus-io/milvus/internal/datacoord/task"
|
|
"github.com/milvus-io/milvus/pkg/v2/log"
|
|
"github.com/milvus-io/milvus/pkg/v2/proto/indexpb"
|
|
"github.com/milvus-io/milvus/pkg/v2/proto/workerpb"
|
|
"github.com/milvus-io/milvus/pkg/v2/taskcommon"
|
|
)
|
|
|
|
type analyzeTask struct {
|
|
*indexpb.AnalyzeTask
|
|
|
|
times *taskcommon.Times
|
|
|
|
schema *schemapb.CollectionSchema
|
|
meta *meta
|
|
}
|
|
|
|
var _ globalTask.Task = (*analyzeTask)(nil)
|
|
|
|
func newAnalyzeTask(t *indexpb.AnalyzeTask, meta *meta) *analyzeTask {
|
|
task := &analyzeTask{
|
|
AnalyzeTask: t,
|
|
times: taskcommon.NewTimes(),
|
|
meta: meta,
|
|
}
|
|
coll := meta.GetCollection(t.CollectionID)
|
|
if coll != nil {
|
|
task.schema = coll.Schema
|
|
}
|
|
|
|
return task
|
|
}
|
|
|
|
func (at *analyzeTask) SetTaskTime(timeType taskcommon.TimeType, time time.Time) {
|
|
at.times.SetTaskTime(timeType, time)
|
|
}
|
|
|
|
func (at *analyzeTask) GetTaskTime(timeType taskcommon.TimeType) time.Time {
|
|
return timeType.GetTaskTime(at.times)
|
|
}
|
|
|
|
func (at *analyzeTask) GetTaskVersion() int64 {
|
|
return at.GetVersion()
|
|
}
|
|
|
|
func (at *analyzeTask) GetTaskType() taskcommon.Type {
|
|
return taskcommon.Analyze
|
|
}
|
|
|
|
func (at *analyzeTask) GetTaskState() taskcommon.State {
|
|
return at.State
|
|
}
|
|
|
|
func (at *analyzeTask) GetTaskSlot() int64 {
|
|
return Params.DataCoordCfg.AnalyzeTaskSlotUsage.GetAsInt64()
|
|
}
|
|
|
|
func (at *analyzeTask) SetState(state indexpb.JobState, failReason string) {
|
|
at.State = state
|
|
at.FailReason = failReason
|
|
}
|
|
|
|
func (at *analyzeTask) UpdateStateWithMeta(state indexpb.JobState, failReason string) error {
|
|
if err := at.meta.analyzeMeta.UpdateState(at.GetTaskID(), state, failReason); err != nil {
|
|
return err
|
|
}
|
|
at.SetState(state, failReason)
|
|
return nil
|
|
}
|
|
|
|
func (at *analyzeTask) UpdateVersion(nodeID int64) error {
|
|
if err := at.meta.analyzeMeta.UpdateVersion(at.GetTaskID(), nodeID); err != nil {
|
|
return err
|
|
}
|
|
at.Version++
|
|
at.NodeID = nodeID
|
|
return nil
|
|
}
|
|
|
|
func (at *analyzeTask) setJobInfo(result *workerpb.AnalyzeResult) error {
|
|
if err := at.meta.analyzeMeta.FinishTask(at.GetTaskID(), result); err != nil {
|
|
return err
|
|
}
|
|
at.SetState(indexpb.JobState(result.GetState()), result.GetFailReason())
|
|
return nil
|
|
}
|
|
|
|
func (at *analyzeTask) resetTask(reason string) {
|
|
at.UpdateStateWithMeta(indexpb.JobState_JobStateInit, reason)
|
|
}
|
|
|
|
func (at *analyzeTask) dropAndResetTaskOnWorker(cluster session.Cluster, reason string) {
|
|
if err := at.tryDropTaskOnWorker(cluster); err != nil {
|
|
return
|
|
}
|
|
at.resetTask(reason)
|
|
}
|
|
|
|
func (at *analyzeTask) CreateTaskOnWorker(nodeID int64, cluster session.Cluster) {
|
|
log := log.Ctx(context.TODO()).With(zap.Int64("taskID", at.GetTaskID()))
|
|
|
|
// Check if task still exists in meta
|
|
task := at.meta.analyzeMeta.GetTask(at.GetTaskID())
|
|
if task == nil {
|
|
log.Info("analyze task has not exist in meta table, remove task")
|
|
at.SetState(indexpb.JobState_JobStateNone, "analyze task has not exist in meta table")
|
|
return
|
|
}
|
|
|
|
// Update task version
|
|
if err := at.UpdateVersion(nodeID); err != nil {
|
|
log.Warn("failed to update task version", zap.Error(err))
|
|
return
|
|
}
|
|
req := &workerpb.AnalyzeRequest{
|
|
ClusterID: Params.CommonCfg.ClusterPrefix.GetValue(),
|
|
TaskID: at.GetTaskID(),
|
|
CollectionID: task.CollectionID,
|
|
PartitionID: task.PartitionID,
|
|
FieldID: task.FieldID,
|
|
FieldName: task.FieldName,
|
|
FieldType: task.FieldType,
|
|
Dim: task.Dim,
|
|
SegmentStats: make(map[int64]*indexpb.SegmentStats),
|
|
Version: task.Version + 1,
|
|
StorageConfig: createStorageConfig(),
|
|
}
|
|
WrapPluginContext(task.CollectionID, at.schema.GetProperties(), req)
|
|
|
|
var err error
|
|
defer func() {
|
|
if err != nil {
|
|
log.Warn("assign analyze task to worker failed, try drop task on worker", zap.Error(err))
|
|
at.tryDropTaskOnWorker(cluster)
|
|
}
|
|
}()
|
|
|
|
err = cluster.CreateAnalyze(nodeID, req)
|
|
if err != nil {
|
|
log.Warn("assign analyze task to worker failed", zap.Error(err))
|
|
return
|
|
}
|
|
|
|
log.Info("analyze task assigned successfully")
|
|
if err = at.UpdateStateWithMeta(indexpb.JobState_JobStateInProgress, ""); err != nil {
|
|
log.Warn("failed to update task state to inProgress", zap.Error(err))
|
|
}
|
|
log.Info("update task state to inProgress successfully")
|
|
}
|
|
|
|
func (at *analyzeTask) QueryTaskOnWorker(cluster session.Cluster) {
|
|
log := log.Ctx(context.TODO()).With(
|
|
zap.Int64("taskID", at.GetTaskID()),
|
|
zap.Int64("nodeID", at.NodeID),
|
|
)
|
|
|
|
resp, err := cluster.QueryAnalyze(at.NodeID, &workerpb.QueryJobsRequest{
|
|
ClusterID: Params.CommonCfg.ClusterPrefix.GetValue(),
|
|
TaskIDs: []int64{at.GetTaskID()},
|
|
})
|
|
if err != nil {
|
|
log.Warn("query analyze task result from worker failed", zap.Error(err))
|
|
at.dropAndResetTaskOnWorker(cluster, err.Error())
|
|
return
|
|
}
|
|
|
|
// Process query results
|
|
for _, result := range resp.GetResults() {
|
|
if result.GetTaskID() != at.GetTaskID() {
|
|
continue
|
|
}
|
|
|
|
state := result.GetState()
|
|
// Handle different task states
|
|
switch state {
|
|
case indexpb.JobState_JobStateFinished, indexpb.JobState_JobStateFailed:
|
|
log.Info("query analyze task result success",
|
|
zap.String("state", state.String()),
|
|
zap.String("failReason", result.GetFailReason()))
|
|
at.setJobInfo(result)
|
|
case indexpb.JobState_JobStateRetry, indexpb.JobState_JobStateNone:
|
|
log.Info("query analyze task result success",
|
|
zap.String("state", state.String()),
|
|
zap.String("failReason", result.GetFailReason()))
|
|
at.dropAndResetTaskOnWorker(cluster, result.GetFailReason())
|
|
}
|
|
// Otherwise (inProgress or unissued/init), keep current state
|
|
return
|
|
}
|
|
|
|
log.Warn("query analyze task info failed, worker does not have task info")
|
|
at.UpdateStateWithMeta(indexpb.JobState_JobStateInit, "analyze result is not in info response")
|
|
}
|
|
|
|
func (at *analyzeTask) tryDropTaskOnWorker(cluster session.Cluster) error {
|
|
log := log.Ctx(context.TODO()).With(
|
|
zap.Int64("taskID", at.GetTaskID()),
|
|
zap.Int64("nodeID", at.NodeID),
|
|
)
|
|
|
|
if err := cluster.DropAnalyze(at.NodeID, at.GetTaskID()); err != nil {
|
|
log.Warn("failed to drop analyze task on worker", zap.Error(err))
|
|
return err
|
|
}
|
|
|
|
log.Info("dropped analyze task on worker successfully")
|
|
return nil
|
|
}
|
|
|
|
func (at *analyzeTask) DropTaskOnWorker(cluster session.Cluster) {
|
|
at.tryDropTaskOnWorker(cluster)
|
|
}
|