mirror of
https://gitee.com/milvus-io/milvus.git
synced 2025-12-29 15:05:31 +08:00
* Remove redundant session startup Signed-off-by: sunby <bingyi.sun@zilliz.com> * Register datanode after start success Signed-off-by: sunby <bingyi.sun@zilliz.com> * fix meta snap shot Signed-off-by: yefu.chen <yefu.chen@zilliz.com> * fix datanode message stream channel Signed-off-by: yangxuan <xuan.yang@zilliz.com> * Fix bugs when drop empty collection Signed-off-by: sunby <bingyi.sun@zilliz.com> * Fix bug of getting pchan statistics from task scheduler Signed-off-by: dragondriver <jiquan.long@zilliz.com> * Fix i/dist/dataservice test code Signed-off-by: Congqi Xia <congqi.xia@zilliz.com> * Fix epoch lifetime not applied Signed-off-by: Congqi Xia <congqi.xia@zilliz.com> * fix datanode flowgraph dd node Signed-off-by: yangxuan <xuan.yang@zilliz.com> * Fix handle datanode timetick bug Signed-off-by: sunby <bingyi.sun@zilliz.com> * Remove repack function of dml stream Signed-off-by: dragondriver <jiquan.long@zilliz.com> * fix proxynode Signed-off-by: yefu.chen <yefu.chen@zilliz.com> * Apply extended seal policy Signed-off-by: Congqi Xia <congqi.xia@zilliz.com> * add check for time tick Signed-off-by: yefu.chen <yefu.chen@zilliz.com> * fix check Signed-off-by: yefu.chen <yefu.chen@zilliz.com> * Fix the repack function of dml stream Signed-off-by: dragondriver <jiquan.long@zilliz.com> * Fix the bug when send statistics of pchan Signed-off-by: dragondriver <jiquan.long@zilliz.com> * Fix the repack function when craete dml stream Signed-off-by: dragondriver <jiquan.long@zilliz.com> * fix bugs Signed-off-by: yefu.chen <yefu.chen@zilliz.com> * fix describe collection Signed-off-by: yefu.chen <yefu.chen@zilliz.com> * Fix bug when send timestamp statistics Signed-off-by: dragondriver <jiquan.long@zilliz.com> * fix data node Signed-off-by: yefu.chen <yefu.chen@zilliz.com> * Add length check before flush request Signed-off-by: Congqi Xia <congqi.xia@zilliz.com> * add log for data node Signed-off-by: yefu.chen <yefu.chen@zilliz.com> * Fix SaveBinlog bugs Signed-off-by: sunby <bingyi.sun@zilliz.com> * Add more log in datanode Signed-off-by: yangxuan <xuan.yang@zilliz.com> * Put SegmentState.Flushing as the last one in enum to fit the client Signed-off-by: sunby <bingyi.sun@zilliz.com> * Fix params in GetInsertBinlogPaths Signed-off-by: sunby <bingyi.sun@zilliz.com> * Rename policy Signed-off-by: sunby <bingyi.sun@zilliz.com> * Remove unused ddl functions and fields Signed-off-by: Congqi Xia <congqi.xia@zilliz.com> * Remove pchan when drop collection Signed-off-by: dragondriver <jiquan.long@zilliz.com> * Add balanced assignment policy Signed-off-by: sunby <bingyi.sun@zilliz.com> * fix master ut Signed-off-by: yefu.chen <yefu.chen@zilliz.com> * Add lock in session manager Signed-off-by: Congqi Xia <congqi.xia@zilliz.com> * add log for debug Signed-off-by: yefu.chen <yefu.chen@zilliz.com> * Fix some logic bug and typo Signed-off-by: Congqi Xia <congqi.xia@zilliz.com> * Fix recover bugs Signed-off-by: sunby <bingyi.sun@zilliz.com> * Get collection scheme of a specific timestamp Signed-off-by: yangxuan <xuan.yang@zilliz.com> * Change CheckPoint to SegmentInfo in VchannelInfo Signed-off-by: sunby <bingyi.sun@zilliz.com> * Recover Unflushed segment numOfRows Signed-off-by: yangxuan <xuan.yang@zilliz.com> * Fix dataservice unit tests Signed-off-by: sunby <bingyi.sun@zilliz.com> Co-authored-by: yefu.chen <yefu.chen@zilliz.com> Co-authored-by: yangxuan <xuan.yang@zilliz.com> Co-authored-by: dragondriver <jiquan.long@zilliz.com> Co-authored-by: Congqi Xia <congqi.xia@zilliz.com>
259 lines
7.1 KiB
Go
259 lines
7.1 KiB
Go
// Copyright (C) 2019-2020 Zilliz. All rights reserved.
|
|
//
|
|
// Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance
|
|
// with the License. You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software distributed under the License
|
|
// is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express
|
|
// or implied. See the License for the specific language governing permissions and limitations under the License.
|
|
|
|
package datanode
|
|
|
|
import (
|
|
"math"
|
|
"os"
|
|
"testing"
|
|
"time"
|
|
|
|
"github.com/stretchr/testify/assert"
|
|
"github.com/stretchr/testify/require"
|
|
"go.uber.org/zap"
|
|
|
|
"github.com/milvus-io/milvus/internal/log"
|
|
"github.com/milvus-io/milvus/internal/msgstream"
|
|
"github.com/milvus-io/milvus/internal/proto/commonpb"
|
|
"github.com/milvus-io/milvus/internal/proto/datapb"
|
|
"github.com/milvus-io/milvus/internal/proto/internalpb"
|
|
)
|
|
|
|
func TestMain(t *testing.M) {
|
|
Params.Init()
|
|
refreshChannelNames()
|
|
code := t.Run()
|
|
os.Exit(code)
|
|
}
|
|
|
|
func TestDataNode(t *testing.T) {
|
|
t.Skip()
|
|
node := newIDLEDataNodeMock()
|
|
node.Start()
|
|
|
|
t.Run("Test WatchDmChannels", func(t *testing.T) {
|
|
node1 := newIDLEDataNodeMock()
|
|
node1.Start()
|
|
vchannels := []*datapb.VchannelInfo{}
|
|
for _, ch := range []string{"datanode-01-test-WatchDmChannel",
|
|
"datanode-02-test-WatchDmChannels"} {
|
|
log.Debug("InsertChannels", zap.String("name", ch))
|
|
vchan := &datapb.VchannelInfo{
|
|
CollectionID: 1,
|
|
ChannelName: ch,
|
|
UnflushedSegments: []*datapb.SegmentInfo{},
|
|
}
|
|
vchannels = append(vchannels, vchan)
|
|
}
|
|
req := &datapb.WatchDmChannelsRequest{
|
|
Base: &commonpb.MsgBase{
|
|
MsgType: 0,
|
|
MsgID: 0,
|
|
Timestamp: 0,
|
|
SourceID: Params.NodeID,
|
|
},
|
|
Vchannels: vchannels,
|
|
}
|
|
|
|
_, err := node1.WatchDmChannels(node.ctx, req)
|
|
assert.NoError(t, err)
|
|
assert.NotNil(t, node1.vchan2FlushCh)
|
|
assert.NotNil(t, node1.vchan2SyncService)
|
|
sync, ok := node1.vchan2SyncService["datanode-01-test-WatchDmChannel"]
|
|
assert.True(t, ok)
|
|
assert.NotNil(t, sync)
|
|
assert.Equal(t, UniqueID(1), sync.collectionID)
|
|
assert.Equal(t, 2, len(node1.vchan2SyncService))
|
|
assert.Equal(t, len(node1.vchan2FlushCh), len(node1.vchan2SyncService))
|
|
|
|
_, err = node1.WatchDmChannels(node1.ctx, req)
|
|
assert.NoError(t, err)
|
|
assert.Equal(t, 2, len(node1.vchan2SyncService))
|
|
|
|
<-node1.ctx.Done()
|
|
node1.Stop()
|
|
})
|
|
|
|
t.Run("Test GetComponentStates", func(t *testing.T) {
|
|
stat, err := node.GetComponentStates(node.ctx)
|
|
assert.NoError(t, err)
|
|
assert.Equal(t, commonpb.ErrorCode_Success, stat.Status.ErrorCode)
|
|
})
|
|
|
|
t.Run("Test NewDataSyncService", func(t *testing.T) {
|
|
node2 := newIDLEDataNodeMock()
|
|
node2.Start()
|
|
dmChannelName := "fake-dm-channel-test-NewDataSyncService"
|
|
|
|
vchan := &datapb.VchannelInfo{
|
|
CollectionID: 1,
|
|
ChannelName: dmChannelName,
|
|
UnflushedSegments: []*datapb.SegmentInfo{},
|
|
}
|
|
|
|
require.Equal(t, 0, len(node2.vchan2FlushCh))
|
|
require.Equal(t, 0, len(node2.vchan2SyncService))
|
|
|
|
err := node2.NewDataSyncService(vchan)
|
|
assert.NoError(t, err)
|
|
assert.Equal(t, 1, len(node2.vchan2FlushCh))
|
|
assert.Equal(t, 1, len(node2.vchan2SyncService))
|
|
|
|
err = node2.NewDataSyncService(vchan)
|
|
assert.NoError(t, err)
|
|
assert.Equal(t, 1, len(node2.vchan2FlushCh))
|
|
assert.Equal(t, 1, len(node2.vchan2SyncService))
|
|
<-node2.ctx.Done()
|
|
node2.Stop()
|
|
})
|
|
|
|
t.Run("Test FlushSegments", func(t *testing.T) {
|
|
dmChannelName := "fake-dm-channel-test-HEALTHDataNodeMock"
|
|
|
|
node1 := newHEALTHDataNodeMock(dmChannelName)
|
|
|
|
sync, ok := node1.vchan2SyncService[dmChannelName]
|
|
assert.True(t, ok)
|
|
sync.replica.addSegment(0, 1, 1, dmChannelName)
|
|
// sync.replica.addSegment(1, 1, 1, dmChannelName) unable to deal with this.
|
|
|
|
req := &datapb.FlushSegmentsRequest{
|
|
Base: &commonpb.MsgBase{},
|
|
DbID: 0,
|
|
CollectionID: 1,
|
|
SegmentIDs: []int64{0},
|
|
}
|
|
|
|
status, err := node1.FlushSegments(node1.ctx, req)
|
|
assert.NoError(t, err)
|
|
assert.Equal(t, commonpb.ErrorCode_Success, status.ErrorCode)
|
|
|
|
timeTickMsgPack := msgstream.MsgPack{}
|
|
|
|
timeTickMsg := &msgstream.TimeTickMsg{
|
|
BaseMsg: msgstream.BaseMsg{
|
|
BeginTimestamp: Timestamp(0),
|
|
EndTimestamp: Timestamp(0),
|
|
HashValues: []uint32{0},
|
|
},
|
|
TimeTickMsg: internalpb.TimeTickMsg{
|
|
Base: &commonpb.MsgBase{
|
|
MsgType: commonpb.MsgType_TimeTick,
|
|
MsgID: UniqueID(0),
|
|
Timestamp: math.MaxUint64,
|
|
SourceID: 0,
|
|
},
|
|
},
|
|
}
|
|
timeTickMsgPack.Msgs = append(timeTickMsgPack.Msgs, timeTickMsg)
|
|
|
|
// pulsar produce
|
|
msFactory := msgstream.NewPmsFactory()
|
|
m := map[string]interface{}{
|
|
"pulsarAddress": Params.PulsarAddress,
|
|
"receiveBufSize": 1024,
|
|
"pulsarBufSize": 1024}
|
|
err = msFactory.SetParams(m)
|
|
assert.NoError(t, err)
|
|
insertStream, _ := msFactory.NewMsgStream(node1.ctx)
|
|
insertStream.AsProducer([]string{dmChannelName})
|
|
|
|
var insertMsgStream msgstream.MsgStream = insertStream
|
|
insertMsgStream.Start()
|
|
|
|
err = insertMsgStream.Broadcast(&timeTickMsgPack)
|
|
assert.NoError(t, err)
|
|
|
|
err = insertMsgStream.Broadcast(&timeTickMsgPack)
|
|
assert.NoError(t, err)
|
|
|
|
_, err = sync.replica.getSegmentByID(0)
|
|
assert.NoError(t, err)
|
|
|
|
defer func() {
|
|
<-node1.ctx.Done()
|
|
node1.Stop()
|
|
}()
|
|
})
|
|
|
|
t.Run("Test GetTimeTickChannel", func(t *testing.T) {
|
|
_, err := node.GetTimeTickChannel(node.ctx)
|
|
assert.NoError(t, err)
|
|
})
|
|
|
|
t.Run("Test GetStatisticsChannel", func(t *testing.T) {
|
|
_, err := node.GetStatisticsChannel(node.ctx)
|
|
assert.NoError(t, err)
|
|
})
|
|
|
|
t.Run("Test ReleaseDataSyncService", func(t *testing.T) {
|
|
dmChannelName := "fake-dm-channel-test-NewDataSyncService"
|
|
|
|
vchan := &datapb.VchannelInfo{
|
|
CollectionID: 1,
|
|
ChannelName: dmChannelName,
|
|
UnflushedSegments: []*datapb.SegmentInfo{},
|
|
}
|
|
|
|
err := node.NewDataSyncService(vchan)
|
|
assert.NoError(t, err)
|
|
assert.Equal(t, 1, len(node.vchan2FlushCh))
|
|
assert.Equal(t, 1, len(node.vchan2SyncService))
|
|
time.Sleep(time.Second)
|
|
|
|
node.ReleaseDataSyncService(dmChannelName)
|
|
assert.Equal(t, 0, len(node.vchan2FlushCh))
|
|
assert.Equal(t, 0, len(node.vchan2SyncService))
|
|
|
|
s, ok := node.vchan2SyncService[dmChannelName]
|
|
assert.False(t, ok)
|
|
assert.Nil(t, s)
|
|
|
|
})
|
|
|
|
t.Run("Test BackGroundGC", func(t *testing.T) {
|
|
collIDCh := make(chan UniqueID)
|
|
go node.BackGroundGC(collIDCh)
|
|
|
|
dmChannelName := "fake-dm-channel-test-BackGroundGC"
|
|
|
|
vchan := &datapb.VchannelInfo{
|
|
CollectionID: 1,
|
|
ChannelName: dmChannelName,
|
|
UnflushedSegments: []*datapb.SegmentInfo{},
|
|
}
|
|
require.Equal(t, 0, len(node.vchan2FlushCh))
|
|
require.Equal(t, 0, len(node.vchan2SyncService))
|
|
|
|
err := node.NewDataSyncService(vchan)
|
|
require.NoError(t, err)
|
|
time.Sleep(time.Second)
|
|
|
|
require.Equal(t, 1, len(node.vchan2FlushCh))
|
|
require.Equal(t, 1, len(node.vchan2SyncService))
|
|
|
|
collIDCh <- 1
|
|
assert.Eventually(t, func() bool {
|
|
return len(node.vchan2FlushCh) == 0
|
|
}, time.Second*4, time.Millisecond)
|
|
|
|
assert.Equal(t, 0, len(node.vchan2SyncService))
|
|
|
|
s, ok := node.vchan2SyncService[dmChannelName]
|
|
assert.False(t, ok)
|
|
assert.Nil(t, s)
|
|
})
|
|
|
|
<-node.ctx.Done()
|
|
node.Stop()
|
|
}
|