Add unittest and fix a bug in segment manager

Signed-off-by: sunby <bingyi.sun@zilliz.com>
This commit is contained in:
sunby 2020-11-21 17:11:05 +08:00 committed by yefu.chen
parent 859ff62bc9
commit 3afab6ecb9
7 changed files with 186 additions and 42 deletions

View File

@ -7,7 +7,6 @@
"remoteUser": "debugger",
"remoteEnv": {"CCACHE_COMPILERCHECK":"content", "CCACHE_MAXSIZE": "2G", "CCACHE_COMPRESS": "1", "CCACHE_COMPRESSLEVEL": "5"},
"extensions": [
"ms-vscode.cmake-tools",
"ms-vscode.cpptools",
"golang.go"
]

4
.env
View File

@ -3,6 +3,6 @@ ARCH=amd64
UBUNTU=18.04
DATE_VERSION=20201120-092740
LATEST_DATE_VERSION=latest
PULSAR_ADDRESS=pulsar://pulsar:6650
ETCD_ADDRESS=etcd:2379
PULSAR_ADDRESS=pulsar://localhost:6650
ETCD_ADDRESS=localhost:2379
MASTER_ADDRESS=localhost:53100

View File

@ -57,7 +57,7 @@ jobs:
- name: Start Service
shell: bash
run: |
cd ${GITHUB_WORKSPACE}/deployments/docker && docker-compose -p milvus-distributed up -d
cd ${GITHUB_WORKSPACE}/deployments/docker && docker-compose up -d
- name: Build and UnitTest
env:
CHECK_BUILDER: "1"

View File

@ -8,8 +8,6 @@ services:
- "2379:2379"
- "2380:2380"
- "4001:4001"
networks:
- milvus
pulsar:
image: apachepulsar/pulsar:latest
@ -17,11 +15,6 @@ services:
ports:
- "6650:6650"
- "18080:8080"
networks:
- milvus
networks:
milvus:
# pd0:
# image: pingcap/pd:latest

View File

@ -10,6 +10,7 @@ x-ccache: &ccache
services:
ubuntu:
image: ${REPO}:${ARCH}-ubuntu${UBUNTU}-${DATE_VERSION}
network_mode: "host"
build:
context: .
dockerfile: build/docker/env/cpu/ubuntu${UBUNTU}/Dockerfile
@ -28,8 +29,6 @@ services:
command: &ubuntu-command >
/bin/bash -c "
make check-proto-product && make verifiers && make unittest"
networks:
- milvus
gdbserver:
image: ${REPO}:${ARCH}-ubuntu${UBUNTU}-${DATE_VERSION}
@ -53,8 +52,3 @@ services:
ports:
- "7776:22"
- "7777:7777"
networks:
- milvus
networks:
milvus:

View File

@ -86,15 +86,14 @@ func (segMgr *SegmentManager) closeSegment(segMeta *etcdpb.SegmentMeta) error {
if segMeta.GetCloseTime() == 0 {
// close the segment and remove from collStatus
collStatus, ok := segMgr.collStatus[segMeta.GetCollectionID()]
if !ok {
return errors.Errorf("Can not find the status of collection %d", segMeta.GetCollectionID())
}
if ok {
openedSegments := collStatus.openedSegments
for i, openedSegID := range openedSegments {
if openedSegID == segMeta.SegmentID {
openedSegments[i] = openedSegments[len(openedSegments)-1]
collStatus.openedSegments = openedSegments[:len(openedSegments)-1]
return nil
break
}
}
}
ts, err := segMgr.globalTSOAllocator()
@ -108,7 +107,7 @@ func (segMgr *SegmentManager) closeSegment(segMeta *etcdpb.SegmentMeta) error {
if err != nil {
return err
}
return errors.Errorf("The segment %d is not opened in collection %d", segMeta.SegmentID, segMeta.GetCollectionID())
return nil
}
func (segMgr *SegmentManager) AssignSegmentID(segIDReq []*internalpb.SegIDRequest) ([]*internalpb.SegIDAssignment, error) {

View File

@ -1,11 +1,19 @@
package master
import (
"context"
"log"
"sync/atomic"
"testing"
"time"
"github.com/gogo/protobuf/proto"
"github.com/zilliztech/milvus-distributed/internal/proto/commonpb"
"github.com/zilliztech/milvus-distributed/internal/proto/masterpb"
"github.com/zilliztech/milvus-distributed/internal/proto/servicepb"
"github.com/zilliztech/milvus-distributed/internal/util/typeutil"
"google.golang.org/grpc"
"github.com/stretchr/testify/assert"
"github.com/zilliztech/milvus-distributed/internal/errors"
"github.com/zilliztech/milvus-distributed/internal/kv"
@ -23,6 +31,8 @@ var collName = "coll_segmgr_test"
var collID = int64(1001)
var partitionTag = "test"
var kvBase *kv.EtcdKV
var master *Master
var masterCancelFunc context.CancelFunc
func setup() {
Params.Init()
@ -32,15 +42,24 @@ func setup() {
if err != nil {
panic(err)
}
rootpath := "/etcd/test/root"
kvBase = kv.NewEtcdKV(cli, rootpath)
rootPath := "/test/root"
ctx, cancel := context.WithCancel(context.TODO())
defer cancel()
_, err = cli.Delete(ctx, rootPath, clientv3.WithPrefix())
if err != nil {
panic(err)
}
kvBase = kv.NewEtcdKV(cli, rootPath)
tmpMt, err := NewMetaTable(kvBase)
if err != nil {
panic(err)
}
mt = tmpMt
if mt.HasCollection(collID) {
mt.DeleteCollection(collID)
err := mt.DeleteCollection(collID)
if err != nil {
panic(err)
}
}
err = mt.AddCollection(&pb.CollectionMeta{
ID: collID,
@ -141,7 +160,7 @@ func TestSegmentManager_AssignSegmentID(t *testing.T) {
newReqs[0].Count = 1000000
_, err = segMgr.AssignSegmentID(newReqs)
assert.Error(t, errors.Errorf("request with count %d need about %d mem size which is larger than segment threshold",
1000000, Params.DefaultRecordSize()*1000000), err)
1000000, 1024*1000000), err)
}
func TestSegmentManager_SegmentStats(t *testing.T) {
@ -162,7 +181,7 @@ func TestSegmentManager_SegmentStats(t *testing.T) {
MsgType: internalpb.MsgType_kQueryNodeSegStats,
PeerID: 1,
SegStats: []*internalpb.SegmentStats{
{SegmentID: 100, MemorySize: 25000 * Params.DefaultRecordSize(), NumRows: 25000, RecentlyModified: true},
{SegmentID: 100, MemorySize: 2500000, NumRows: 25000, RecentlyModified: true},
},
}
baseMsg := msgstream.BaseMsg{
@ -183,19 +202,159 @@ func TestSegmentManager_SegmentStats(t *testing.T) {
err = segMgr.HandleQueryNodeMsgPack(&msgPack)
assert.Nil(t, err)
time.Sleep(1 * time.Second)
segMeta, _ := mt.GetSegmentByID(100)
assert.Equal(t, int64(100), segMeta.SegmentID)
assert.Equal(t, 25000*Params.DefaultRecordSize(), segMeta.MemSize)
assert.Equal(t, int64(2500000), segMeta.MemSize)
assert.Equal(t, int64(25000), segMeta.NumRows)
// close segment
stats.SegStats[0].NumRows = 520000
stats.SegStats[0].MemorySize = 520000 * Params.DefaultRecordSize()
stats.SegStats[0].NumRows = 600000
stats.SegStats[0].MemorySize = 600000000
err = segMgr.HandleQueryNodeMsgPack(&msgPack)
assert.Nil(t, err)
time.Sleep(1 * time.Second)
segMeta, _ = mt.GetSegmentByID(100)
assert.Equal(t, int64(100), segMeta.SegmentID)
assert.NotEqual(t, 0, segMeta.CloseTime)
assert.NotEqual(t, uint64(0), segMeta.CloseTime)
}
func startupMaster() {
Params.Init()
etcdAddress := Params.EtcdAddress()
rootPath := "/test/root"
ctx, cancel := context.WithCancel(context.TODO())
masterCancelFunc = cancel
cli, err := clientv3.New(clientv3.Config{Endpoints: []string{etcdAddress}})
if err != nil {
panic(err)
}
_, err = cli.Delete(ctx, rootPath, clientv3.WithPrefix())
if err != nil {
panic(err)
}
opt := &Option{
KVRootPath: "/test/root/kv",
MetaRootPath: "/test/root/meta",
EtcdAddr: []string{etcdAddress},
PulsarAddr: "pulsar://localhost:6650",
ProxyIDs: []typeutil.UniqueID{1, 2},
PulsarProxyChannels: []string{"proxy1", "proxy2"},
PulsarProxySubName: "proxyTopics",
SoftTTBInterval: 300,
WriteIDs: []typeutil.UniqueID{3, 4},
PulsarWriteChannels: []string{"write3", "write4"},
PulsarWriteSubName: "writeTopics",
PulsarDMChannels: []string{"dm0", "dm1"},
PulsarK2SChannels: []string{"k2s0", "k2s1"},
DefaultRecordSize: 1024,
MinimumAssignSize: 1048576,
SegmentThreshold: 536870912,
SegmentExpireDuration: 2000,
NumOfChannel: 5,
NumOfQueryNode: 3,
StatsChannels: "statistic",
}
master, err = CreateServer(ctx, opt)
if err != nil {
panic(err)
}
err = master.Run(10013)
if err != nil {
panic(err)
}
}
func shutdownMaster() {
masterCancelFunc()
master.Close()
}
func TestSegmentManager_RPC(t *testing.T) {
startupMaster()
defer shutdownMaster()
ctx, cancel := context.WithCancel(context.TODO())
defer cancel()
dialContext, err := grpc.DialContext(ctx, "127.0.0.1:10013", grpc.WithInsecure(), grpc.WithBlock())
assert.Nil(t, err)
defer dialContext.Close()
client := masterpb.NewMasterClient(dialContext)
schema := &schemapb.CollectionSchema{
Name: collName,
Description: "test coll",
AutoID: false,
Fields: []*schemapb.FieldSchema{},
}
schemaBytes, err := proto.Marshal(schema)
assert.Nil(t, err)
_, err = client.CreateCollection(ctx, &internalpb.CreateCollectionRequest{
MsgType: internalpb.MsgType_kCreateCollection,
ReqID: 1,
Timestamp: 100,
ProxyID: 1,
Schema: &commonpb.Blob{Value: schemaBytes},
})
assert.Nil(t, err)
_, err = client.CreatePartition(ctx, &internalpb.CreatePartitionRequest{
MsgType: internalpb.MsgType_kCreatePartition,
ReqID: 2,
Timestamp: 101,
ProxyID: 1,
PartitionName: &servicepb.PartitionName{
CollectionName: collName,
Tag: partitionTag,
},
})
assert.Nil(t, err)
resp, err := client.AssignSegmentID(ctx, &internalpb.AssignSegIDRequest{
PeerID: 1,
Role: internalpb.PeerRole_Proxy,
PerChannelReq: []*internalpb.SegIDRequest{
{Count: 10000, ChannelID: 0, CollName: collName, PartitionTag: partitionTag},
},
})
assert.Nil(t, err)
assert.Equal(t, commonpb.ErrorCode_SUCCESS, resp.Status.ErrorCode)
assignments := resp.GetPerChannelAssignment()
assert.Equal(t, 1, len(assignments))
assert.Equal(t, collName, assignments[0].CollName)
assert.Equal(t, partitionTag, assignments[0].PartitionTag)
assert.Equal(t, int32(0), assignments[0].ChannelID)
assert.Equal(t, uint32(10000), assignments[0].Count)
// test stats
segID := assignments[0].SegID
ms := msgstream.NewPulsarMsgStream(ctx, 1024)
ms.SetPulsarClient("pulsar://localhost:6650")
ms.CreatePulsarProducers([]string{"statistic"})
ms.Start()
defer ms.Close()
err = ms.Produce(&msgstream.MsgPack{
BeginTs: 102,
EndTs: 104,
Msgs: []msgstream.TsMsg{
&msgstream.QueryNodeSegStatsMsg{
QueryNodeSegStats: internalpb.QueryNodeSegStats{
MsgType: internalpb.MsgType_kQueryNodeSegStats,
PeerID: 1,
SegStats: []*internalpb.SegmentStats{
{SegmentID: segID, MemorySize: 600000000, NumRows: 1000000, RecentlyModified: true},
},
},
BaseMsg: msgstream.BaseMsg{
HashValues: []int32{0},
},
},
},
})
assert.Nil(t, err)
time.Sleep(500 * time.Millisecond)
segMeta, err := master.metaTable.GetSegmentByID(segID)
assert.Nil(t, err)
assert.NotEqual(t, uint64(0), segMeta.GetCloseTime())
assert.Equal(t, int64(600000000), segMeta.GetMemSize())
}