mirror of
https://gitee.com/milvus-io/milvus.git
synced 2025-12-07 09:38:39 +08:00
Open golangci-lint check
Signed-off-by: cai.zhang <cai.zhang@zilliz.com>
This commit is contained in:
parent
14d4a400dd
commit
b68322d469
@ -6,6 +6,7 @@ run:
|
|||||||
- docs
|
- docs
|
||||||
- scripts
|
- scripts
|
||||||
- internal/core
|
- internal/core
|
||||||
|
- internal/proto
|
||||||
|
|
||||||
linters-settings:
|
linters-settings:
|
||||||
golint:
|
golint:
|
||||||
|
|||||||
6
Makefile
6
Makefile
@ -36,9 +36,9 @@ fmt:
|
|||||||
lint:
|
lint:
|
||||||
@echo "Running $@ check"
|
@echo "Running $@ check"
|
||||||
@GO111MODULE=on ${GOPATH}/bin/golangci-lint cache clean
|
@GO111MODULE=on ${GOPATH}/bin/golangci-lint cache clean
|
||||||
@GO111MODULE=on ${GOPATH}/bin/golangci-lint run --timeout=1m --config ./.golangci.yml ./internal/... || true
|
@GO111MODULE=on ${GOPATH}/bin/golangci-lint run --timeout=1m --config ./.golangci.yml ./internal/...
|
||||||
@GO111MODULE=on ${GOPATH}/bin/golangci-lint run --timeout=1m --config ./.golangci.yml ./cmd/... || true
|
@GO111MODULE=on ${GOPATH}/bin/golangci-lint run --timeout=1m --config ./.golangci.yml ./cmd/...
|
||||||
@GO111MODULE=on ${GOPATH}/bin/golangci-lint run --timeout=1m --config ./.golangci.yml ./test/... || true
|
@GO111MODULE=on ${GOPATH}/bin/golangci-lint run --timeout=1m --config ./.golangci.yml ./test/...
|
||||||
|
|
||||||
ruleguard:
|
ruleguard:
|
||||||
@echo "Running $@ check"
|
@echo "Running $@ check"
|
||||||
|
|||||||
@ -25,4 +25,4 @@ services:
|
|||||||
working_dir: "/milvus-distributed"
|
working_dir: "/milvus-distributed"
|
||||||
command: &ubuntu-command >
|
command: &ubuntu-command >
|
||||||
/bin/bash -c "
|
/bin/bash -c "
|
||||||
make verifiers && make unittest"
|
make build-cpp && make verifiers && make unittest"
|
||||||
|
|||||||
@ -2,19 +2,19 @@ package kv
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
"github.com/zilliztech/milvus-distributed/internal/conf"
|
|
||||||
"path"
|
"path"
|
||||||
"strconv"
|
"strconv"
|
||||||
"testing"
|
"testing"
|
||||||
|
|
||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
|
"github.com/zilliztech/milvus-distributed/internal/conf"
|
||||||
"go.etcd.io/etcd/clientv3"
|
"go.etcd.io/etcd/clientv3"
|
||||||
)
|
)
|
||||||
|
|
||||||
func TestEtcdKV_Load(t *testing.T) {
|
func TestEtcdKV_Load(t *testing.T) {
|
||||||
conf.LoadConfig("config.yaml")
|
conf.LoadConfig("config.yaml")
|
||||||
etcd_port := strconv.Itoa(int(conf.Config.Etcd.Port))
|
etcdPort := strconv.Itoa(int(conf.Config.Etcd.Port))
|
||||||
cli, err := clientv3.New(clientv3.Config{Endpoints: []string{"127.0.0.1:" + etcd_port}})
|
cli, err := clientv3.New(clientv3.Config{Endpoints: []string{"127.0.0.1:" + etcdPort}})
|
||||||
assert.Nil(t, err)
|
assert.Nil(t, err)
|
||||||
rootpath := "/etcd/test/root"
|
rootpath := "/etcd/test/root"
|
||||||
kv := NewEtcdKV(cli, rootpath)
|
kv := NewEtcdKV(cli, rootpath)
|
||||||
@ -69,8 +69,8 @@ func TestEtcdKV_Load(t *testing.T) {
|
|||||||
|
|
||||||
func TestEtcdKV_MultiSave(t *testing.T) {
|
func TestEtcdKV_MultiSave(t *testing.T) {
|
||||||
conf.LoadConfig("config.yaml")
|
conf.LoadConfig("config.yaml")
|
||||||
etcd_port := strconv.Itoa(int(conf.Config.Etcd.Port))
|
etcdPort := strconv.Itoa(int(conf.Config.Etcd.Port))
|
||||||
cli, err := clientv3.New(clientv3.Config{Endpoints: []string{"127.0.0.1:" + etcd_port}})
|
cli, err := clientv3.New(clientv3.Config{Endpoints: []string{"127.0.0.1:" + etcdPort}})
|
||||||
assert.Nil(t, err)
|
assert.Nil(t, err)
|
||||||
rootpath := "/etcd/test/root"
|
rootpath := "/etcd/test/root"
|
||||||
kv := NewEtcdKV(cli, rootpath)
|
kv := NewEtcdKV(cli, rootpath)
|
||||||
@ -98,8 +98,8 @@ func TestEtcdKV_MultiSave(t *testing.T) {
|
|||||||
|
|
||||||
func TestEtcdKV_Remove(t *testing.T) {
|
func TestEtcdKV_Remove(t *testing.T) {
|
||||||
conf.LoadConfig("config.yaml")
|
conf.LoadConfig("config.yaml")
|
||||||
etcd_port := strconv.Itoa(int(conf.Config.Etcd.Port))
|
etcdPort := strconv.Itoa(int(conf.Config.Etcd.Port))
|
||||||
cli, err := clientv3.New(clientv3.Config{Endpoints: []string{"127.0.0.1:" + etcd_port}})
|
cli, err := clientv3.New(clientv3.Config{Endpoints: []string{"127.0.0.1:" + etcdPort}})
|
||||||
assert.Nil(t, err)
|
assert.Nil(t, err)
|
||||||
rootpath := "/etcd/test/root"
|
rootpath := "/etcd/test/root"
|
||||||
kv := NewEtcdKV(cli, rootpath)
|
kv := NewEtcdKV(cli, rootpath)
|
||||||
@ -167,8 +167,8 @@ func TestEtcdKV_Remove(t *testing.T) {
|
|||||||
|
|
||||||
func TestEtcdKV_MultiSaveAndRemove(t *testing.T) {
|
func TestEtcdKV_MultiSaveAndRemove(t *testing.T) {
|
||||||
conf.LoadConfig("config.yaml")
|
conf.LoadConfig("config.yaml")
|
||||||
etcd_port := strconv.Itoa(int(conf.Config.Etcd.Port))
|
etcdPort := strconv.Itoa(int(conf.Config.Etcd.Port))
|
||||||
cli, err := clientv3.New(clientv3.Config{Endpoints: []string{"127.0.0.1:" + etcd_port}})
|
cli, err := clientv3.New(clientv3.Config{Endpoints: []string{"127.0.0.1:" + etcdPort}})
|
||||||
assert.Nil(t, err)
|
assert.Nil(t, err)
|
||||||
rootpath := "/etcd/test/root"
|
rootpath := "/etcd/test/root"
|
||||||
kv := NewEtcdKV(cli, rootpath)
|
kv := NewEtcdKV(cli, rootpath)
|
||||||
|
|||||||
@ -74,7 +74,7 @@ func (t *createCollectionTask) Execute() error {
|
|||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
collectionID, err := allocGlobalId()
|
collectionID, err := allocGlobalID()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|||||||
@ -2,11 +2,11 @@ package master
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
"github.com/zilliztech/milvus-distributed/internal/master/tso"
|
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/zilliztech/milvus-distributed/internal/errors"
|
"github.com/zilliztech/milvus-distributed/internal/errors"
|
||||||
"github.com/zilliztech/milvus-distributed/internal/master/id"
|
"github.com/zilliztech/milvus-distributed/internal/master/id"
|
||||||
|
"github.com/zilliztech/milvus-distributed/internal/master/tso"
|
||||||
"github.com/zilliztech/milvus-distributed/internal/proto/commonpb"
|
"github.com/zilliztech/milvus-distributed/internal/proto/commonpb"
|
||||||
"github.com/zilliztech/milvus-distributed/internal/proto/internalpb"
|
"github.com/zilliztech/milvus-distributed/internal/proto/internalpb"
|
||||||
"github.com/zilliztech/milvus-distributed/internal/proto/servicepb"
|
"github.com/zilliztech/milvus-distributed/internal/proto/servicepb"
|
||||||
|
|||||||
@ -2,12 +2,12 @@ package master
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
"github.com/zilliztech/milvus-distributed/internal/conf"
|
|
||||||
"strconv"
|
"strconv"
|
||||||
"testing"
|
"testing"
|
||||||
|
|
||||||
"github.com/golang/protobuf/proto"
|
"github.com/golang/protobuf/proto"
|
||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
|
"github.com/zilliztech/milvus-distributed/internal/conf"
|
||||||
"github.com/zilliztech/milvus-distributed/internal/proto/commonpb"
|
"github.com/zilliztech/milvus-distributed/internal/proto/commonpb"
|
||||||
"github.com/zilliztech/milvus-distributed/internal/proto/internalpb"
|
"github.com/zilliztech/milvus-distributed/internal/proto/internalpb"
|
||||||
"github.com/zilliztech/milvus-distributed/internal/proto/masterpb"
|
"github.com/zilliztech/milvus-distributed/internal/proto/masterpb"
|
||||||
@ -21,15 +21,15 @@ func TestMaster_CreateCollection(t *testing.T) {
|
|||||||
ctx, cancel := context.WithCancel(context.TODO())
|
ctx, cancel := context.WithCancel(context.TODO())
|
||||||
defer cancel()
|
defer cancel()
|
||||||
|
|
||||||
etcd_port := strconv.Itoa(int(conf.Config.Etcd.Port))
|
etcdPort := strconv.Itoa(int(conf.Config.Etcd.Port))
|
||||||
etcd_addr := "127.0.0.1:" + etcd_port
|
etcdAddr := "127.0.0.1:" + etcdPort
|
||||||
|
|
||||||
etcdCli, err := clientv3.New(clientv3.Config{Endpoints: []string{etcd_addr}})
|
etcdCli, err := clientv3.New(clientv3.Config{Endpoints: []string{etcdAddr}})
|
||||||
assert.Nil(t, err)
|
assert.Nil(t, err)
|
||||||
_, err = etcdCli.Delete(ctx, "/test/root", clientv3.WithPrefix())
|
_, err = etcdCli.Delete(ctx, "/test/root", clientv3.WithPrefix())
|
||||||
assert.Nil(t, err)
|
assert.Nil(t, err)
|
||||||
|
|
||||||
svr, err := CreateServer(ctx, "/test/root/kv", "/test/root/meta", "/test/root/meta/tso", []string{etcd_addr})
|
svr, err := CreateServer(ctx, "/test/root/kv", "/test/root/meta", "/test/root/meta/tso", []string{etcdAddr})
|
||||||
assert.Nil(t, err)
|
assert.Nil(t, err)
|
||||||
err = svr.Run(10001)
|
err = svr.Run(10001)
|
||||||
assert.Nil(t, err)
|
assert.Nil(t, err)
|
||||||
@ -96,7 +96,7 @@ func TestMaster_CreateCollection(t *testing.T) {
|
|||||||
},
|
},
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
schema_bytes, err := proto.Marshal(&sch)
|
schemaBytes, err := proto.Marshal(&sch)
|
||||||
assert.Nil(t, err)
|
assert.Nil(t, err)
|
||||||
|
|
||||||
req := internalpb.CreateCollectionRequest{
|
req := internalpb.CreateCollectionRequest{
|
||||||
@ -104,44 +104,44 @@ func TestMaster_CreateCollection(t *testing.T) {
|
|||||||
ReqId: 1,
|
ReqId: 1,
|
||||||
Timestamp: 11,
|
Timestamp: 11,
|
||||||
ProxyId: 1,
|
ProxyId: 1,
|
||||||
Schema: &commonpb.Blob{Value: schema_bytes},
|
Schema: &commonpb.Blob{Value: schemaBytes},
|
||||||
}
|
}
|
||||||
st, err := cli.CreateCollection(ctx, &req)
|
st, err := cli.CreateCollection(ctx, &req)
|
||||||
assert.Nil(t, err)
|
assert.Nil(t, err)
|
||||||
assert.Equal(t, st.ErrorCode, commonpb.ErrorCode_SUCCESS)
|
assert.Equal(t, st.ErrorCode, commonpb.ErrorCode_SUCCESS)
|
||||||
|
|
||||||
coll_meta, err := svr.mt.GetCollectionByName(sch.Name)
|
collMeta, err := svr.mt.GetCollectionByName(sch.Name)
|
||||||
assert.Nil(t, err)
|
assert.Nil(t, err)
|
||||||
t.Logf("collection id = %d", coll_meta.ID)
|
t.Logf("collection id = %d", collMeta.ID)
|
||||||
assert.Equal(t, coll_meta.CreateTime, uint64(11))
|
assert.Equal(t, collMeta.CreateTime, uint64(11))
|
||||||
assert.Equal(t, coll_meta.Schema.Name, "col1")
|
assert.Equal(t, collMeta.Schema.Name, "col1")
|
||||||
assert.Equal(t, coll_meta.Schema.AutoId, false)
|
assert.Equal(t, collMeta.Schema.AutoId, false)
|
||||||
assert.Equal(t, len(coll_meta.Schema.Fields), 2)
|
assert.Equal(t, len(collMeta.Schema.Fields), 2)
|
||||||
assert.Equal(t, coll_meta.Schema.Fields[0].Name, "col1_f1")
|
assert.Equal(t, collMeta.Schema.Fields[0].Name, "col1_f1")
|
||||||
assert.Equal(t, coll_meta.Schema.Fields[1].Name, "col1_f2")
|
assert.Equal(t, collMeta.Schema.Fields[1].Name, "col1_f2")
|
||||||
assert.Equal(t, coll_meta.Schema.Fields[0].DataType, schemapb.DataType_VECTOR_FLOAT)
|
assert.Equal(t, collMeta.Schema.Fields[0].DataType, schemapb.DataType_VECTOR_FLOAT)
|
||||||
assert.Equal(t, coll_meta.Schema.Fields[1].DataType, schemapb.DataType_VECTOR_BINARY)
|
assert.Equal(t, collMeta.Schema.Fields[1].DataType, schemapb.DataType_VECTOR_BINARY)
|
||||||
assert.Equal(t, len(coll_meta.Schema.Fields[0].TypeParams), 2)
|
assert.Equal(t, len(collMeta.Schema.Fields[0].TypeParams), 2)
|
||||||
assert.Equal(t, len(coll_meta.Schema.Fields[0].IndexParams), 2)
|
assert.Equal(t, len(collMeta.Schema.Fields[0].IndexParams), 2)
|
||||||
assert.Equal(t, len(coll_meta.Schema.Fields[1].TypeParams), 2)
|
assert.Equal(t, len(collMeta.Schema.Fields[1].TypeParams), 2)
|
||||||
assert.Equal(t, len(coll_meta.Schema.Fields[1].IndexParams), 2)
|
assert.Equal(t, len(collMeta.Schema.Fields[1].IndexParams), 2)
|
||||||
assert.Equal(t, coll_meta.Schema.Fields[0].TypeParams[0].Key, "col1_f1_tk1")
|
assert.Equal(t, collMeta.Schema.Fields[0].TypeParams[0].Key, "col1_f1_tk1")
|
||||||
assert.Equal(t, coll_meta.Schema.Fields[0].TypeParams[1].Key, "col1_f1_tk2")
|
assert.Equal(t, collMeta.Schema.Fields[0].TypeParams[1].Key, "col1_f1_tk2")
|
||||||
assert.Equal(t, coll_meta.Schema.Fields[0].TypeParams[0].Value, "col1_f1_tv1")
|
assert.Equal(t, collMeta.Schema.Fields[0].TypeParams[0].Value, "col1_f1_tv1")
|
||||||
assert.Equal(t, coll_meta.Schema.Fields[0].TypeParams[1].Value, "col1_f1_tv2")
|
assert.Equal(t, collMeta.Schema.Fields[0].TypeParams[1].Value, "col1_f1_tv2")
|
||||||
assert.Equal(t, coll_meta.Schema.Fields[0].IndexParams[0].Key, "col1_f1_ik1")
|
assert.Equal(t, collMeta.Schema.Fields[0].IndexParams[0].Key, "col1_f1_ik1")
|
||||||
assert.Equal(t, coll_meta.Schema.Fields[0].IndexParams[1].Key, "col1_f1_ik2")
|
assert.Equal(t, collMeta.Schema.Fields[0].IndexParams[1].Key, "col1_f1_ik2")
|
||||||
assert.Equal(t, coll_meta.Schema.Fields[0].IndexParams[0].Value, "col1_f1_iv1")
|
assert.Equal(t, collMeta.Schema.Fields[0].IndexParams[0].Value, "col1_f1_iv1")
|
||||||
assert.Equal(t, coll_meta.Schema.Fields[0].IndexParams[1].Value, "col1_f1_iv2")
|
assert.Equal(t, collMeta.Schema.Fields[0].IndexParams[1].Value, "col1_f1_iv2")
|
||||||
|
|
||||||
assert.Equal(t, coll_meta.Schema.Fields[1].TypeParams[0].Key, "col1_f2_tk1")
|
assert.Equal(t, collMeta.Schema.Fields[1].TypeParams[0].Key, "col1_f2_tk1")
|
||||||
assert.Equal(t, coll_meta.Schema.Fields[1].TypeParams[1].Key, "col1_f2_tk2")
|
assert.Equal(t, collMeta.Schema.Fields[1].TypeParams[1].Key, "col1_f2_tk2")
|
||||||
assert.Equal(t, coll_meta.Schema.Fields[1].TypeParams[0].Value, "col1_f2_tv1")
|
assert.Equal(t, collMeta.Schema.Fields[1].TypeParams[0].Value, "col1_f2_tv1")
|
||||||
assert.Equal(t, coll_meta.Schema.Fields[1].TypeParams[1].Value, "col1_f2_tv2")
|
assert.Equal(t, collMeta.Schema.Fields[1].TypeParams[1].Value, "col1_f2_tv2")
|
||||||
assert.Equal(t, coll_meta.Schema.Fields[1].IndexParams[0].Key, "col1_f2_ik1")
|
assert.Equal(t, collMeta.Schema.Fields[1].IndexParams[0].Key, "col1_f2_ik1")
|
||||||
assert.Equal(t, coll_meta.Schema.Fields[1].IndexParams[1].Key, "col1_f2_ik2")
|
assert.Equal(t, collMeta.Schema.Fields[1].IndexParams[1].Key, "col1_f2_ik2")
|
||||||
assert.Equal(t, coll_meta.Schema.Fields[1].IndexParams[0].Value, "col1_f2_iv1")
|
assert.Equal(t, collMeta.Schema.Fields[1].IndexParams[0].Value, "col1_f2_iv1")
|
||||||
assert.Equal(t, coll_meta.Schema.Fields[1].IndexParams[1].Value, "col1_f2_iv2")
|
assert.Equal(t, collMeta.Schema.Fields[1].IndexParams[1].Value, "col1_f2_iv2")
|
||||||
|
|
||||||
req.Timestamp = Timestamp(10)
|
req.Timestamp = Timestamp(10)
|
||||||
st, err = cli.CreateCollection(ctx, &req)
|
st, err = cli.CreateCollection(ctx, &req)
|
||||||
|
|||||||
@ -73,7 +73,7 @@ func Init() {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// CreateServer creates the UNINITIALIZED pd server with given configuration.
|
// CreateServer creates the UNINITIALIZED pd server with given configuration.
|
||||||
func CreateServer(ctx context.Context, kv_root_path string, meta_root_path, tso_root_path string, etcdAddr []string) (*Master, error) {
|
func CreateServer(ctx context.Context, kvRootPath string, metaRootPath, tsoRootPath string, etcdAddr []string) (*Master, error) {
|
||||||
rand.Seed(time.Now().UnixNano())
|
rand.Seed(time.Now().UnixNano())
|
||||||
Init()
|
Init()
|
||||||
|
|
||||||
@ -81,7 +81,7 @@ func CreateServer(ctx context.Context, kv_root_path string, meta_root_path, tso_
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
etcdkv := kv.NewEtcdKV(etcdClient, meta_root_path)
|
etcdkv := kv.NewEtcdKV(etcdClient, metaRootPath)
|
||||||
metakv, err := NewMetaTable(etcdkv)
|
metakv, err := NewMetaTable(etcdkv)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
@ -90,7 +90,7 @@ func CreateServer(ctx context.Context, kv_root_path string, meta_root_path, tso_
|
|||||||
m := &Master{
|
m := &Master{
|
||||||
ctx: ctx,
|
ctx: ctx,
|
||||||
startTimestamp: time.Now().Unix(),
|
startTimestamp: time.Now().Unix(),
|
||||||
kvBase: newKVBase(kv_root_path, etcdAddr),
|
kvBase: newKVBase(kvRootPath, etcdAddr),
|
||||||
scheduler: NewDDRequestScheduler(),
|
scheduler: NewDDRequestScheduler(),
|
||||||
mt: metakv,
|
mt: metakv,
|
||||||
ssChan: make(chan internalpb.SegmentStatistics, 10),
|
ssChan: make(chan internalpb.SegmentStatistics, 10),
|
||||||
|
|||||||
@ -2,11 +2,11 @@ package master
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
"github.com/zilliztech/milvus-distributed/internal/conf"
|
|
||||||
"strconv"
|
"strconv"
|
||||||
"testing"
|
"testing"
|
||||||
|
|
||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
|
"github.com/zilliztech/milvus-distributed/internal/conf"
|
||||||
"github.com/zilliztech/milvus-distributed/internal/kv"
|
"github.com/zilliztech/milvus-distributed/internal/kv"
|
||||||
pb "github.com/zilliztech/milvus-distributed/internal/proto/etcdpb"
|
pb "github.com/zilliztech/milvus-distributed/internal/proto/etcdpb"
|
||||||
"github.com/zilliztech/milvus-distributed/internal/proto/schemapb"
|
"github.com/zilliztech/milvus-distributed/internal/proto/schemapb"
|
||||||
@ -15,8 +15,8 @@ import (
|
|||||||
|
|
||||||
func TestMetaTable_Collection(t *testing.T) {
|
func TestMetaTable_Collection(t *testing.T) {
|
||||||
conf.LoadConfig("config.yaml")
|
conf.LoadConfig("config.yaml")
|
||||||
etcd_port := strconv.Itoa(int(conf.Config.Etcd.Port))
|
etcdPort := strconv.Itoa(int(conf.Config.Etcd.Port))
|
||||||
cli, err := clientv3.New(clientv3.Config{Endpoints: []string{"127.0.0.1:" + etcd_port}})
|
cli, err := clientv3.New(clientv3.Config{Endpoints: []string{"127.0.0.1:" + etcdPort}})
|
||||||
assert.Nil(t, err)
|
assert.Nil(t, err)
|
||||||
etcdKV := kv.NewEtcdKV(cli, "/etcd/test/root")
|
etcdKV := kv.NewEtcdKV(cli, "/etcd/test/root")
|
||||||
|
|
||||||
@ -138,8 +138,8 @@ func TestMetaTable_Collection(t *testing.T) {
|
|||||||
|
|
||||||
func TestMetaTable_DeletePartition(t *testing.T) {
|
func TestMetaTable_DeletePartition(t *testing.T) {
|
||||||
conf.LoadConfig("config.yaml")
|
conf.LoadConfig("config.yaml")
|
||||||
etcd_port := strconv.Itoa(int(conf.Config.Etcd.Port))
|
etcdPort := strconv.Itoa(int(conf.Config.Etcd.Port))
|
||||||
cli, err := clientv3.New(clientv3.Config{Endpoints: []string{"127.0.0.1:" + etcd_port}})
|
cli, err := clientv3.New(clientv3.Config{Endpoints: []string{"127.0.0.1:" + etcdPort}})
|
||||||
assert.Nil(t, err)
|
assert.Nil(t, err)
|
||||||
etcdKV := kv.NewEtcdKV(cli, "/etcd/test/root")
|
etcdKV := kv.NewEtcdKV(cli, "/etcd/test/root")
|
||||||
|
|
||||||
@ -221,8 +221,8 @@ func TestMetaTable_DeletePartition(t *testing.T) {
|
|||||||
|
|
||||||
func TestMetaTable_Segment(t *testing.T) {
|
func TestMetaTable_Segment(t *testing.T) {
|
||||||
conf.LoadConfig("config.yaml")
|
conf.LoadConfig("config.yaml")
|
||||||
etcd_port := strconv.Itoa(int(conf.Config.Etcd.Port))
|
etcdPort := strconv.Itoa(int(conf.Config.Etcd.Port))
|
||||||
cli, err := clientv3.New(clientv3.Config{Endpoints: []string{"127.0.0.1:" + etcd_port}})
|
cli, err := clientv3.New(clientv3.Config{Endpoints: []string{"127.0.0.1:" + etcdPort}})
|
||||||
assert.Nil(t, err)
|
assert.Nil(t, err)
|
||||||
etcdKV := kv.NewEtcdKV(cli, "/etcd/test/root")
|
etcdKV := kv.NewEtcdKV(cli, "/etcd/test/root")
|
||||||
|
|
||||||
|
|||||||
@ -21,7 +21,7 @@ func (rs *ddRequestScheduler) Enqueue(task task) error {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
//TODO, allocGlobalId
|
//TODO, allocGlobalID
|
||||||
func allocGlobalId() (UniqueID, error) {
|
func allocGlobalID() (UniqueID, error) {
|
||||||
return rand.Int63(), nil
|
return rand.Int63(), nil
|
||||||
}
|
}
|
||||||
|
|||||||
@ -2,11 +2,12 @@ package timesync
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
"github.com/stretchr/testify/assert"
|
|
||||||
ms "github.com/zilliztech/milvus-distributed/internal/msgstream"
|
|
||||||
"log"
|
"log"
|
||||||
"testing"
|
"testing"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
|
"github.com/stretchr/testify/assert"
|
||||||
|
ms "github.com/zilliztech/milvus-distributed/internal/msgstream"
|
||||||
)
|
)
|
||||||
|
|
||||||
type (
|
type (
|
||||||
@ -24,15 +25,8 @@ func (ttBarrier *TestTickBarrier) GetTimeTick() (Timestamp, error) {
|
|||||||
|
|
||||||
func (ttBarrier *TestTickBarrier) Start() error {
|
func (ttBarrier *TestTickBarrier) Start() error {
|
||||||
go func(ctx context.Context) {
|
go func(ctx context.Context) {
|
||||||
for {
|
<-ctx.Done()
|
||||||
select {
|
log.Printf("barrier context done, exit")
|
||||||
case <-ctx.Done():
|
|
||||||
{
|
|
||||||
log.Printf("barrier context done, exit")
|
|
||||||
return
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}(ttBarrier.ctx)
|
}(ttBarrier.ctx)
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
@ -90,7 +84,8 @@ func TestStream_PulsarMsgStream_TimeTick(t *testing.T) {
|
|||||||
producerChannels := []string{"proxyTtBarrier"}
|
producerChannels := []string{"proxyTtBarrier"}
|
||||||
consumerChannels := []string{"proxyTtBarrier"}
|
consumerChannels := []string{"proxyTtBarrier"}
|
||||||
consumerSubName := "proxyTtBarrier"
|
consumerSubName := "proxyTtBarrier"
|
||||||
ctx, _ := context.WithTimeout(context.Background(), 30*time.Second)
|
ctx, cancel := context.WithTimeout(context.Background(), 30*time.Second)
|
||||||
|
defer cancel()
|
||||||
proxyTtInputStream, proxyTtOutputStream := initTestPulsarStream(ctx, pulsarAddress, producerChannels, consumerChannels, consumerSubName)
|
proxyTtInputStream, proxyTtOutputStream := initTestPulsarStream(ctx, pulsarAddress, producerChannels, consumerChannels, consumerSubName)
|
||||||
|
|
||||||
producerChannels = []string{"writeNodeBarrier"}
|
producerChannels = []string{"writeNodeBarrier"}
|
||||||
@ -107,10 +102,10 @@ func TestStream_PulsarMsgStream_TimeTick(t *testing.T) {
|
|||||||
(*writeNodeOutputStream).Start()
|
(*writeNodeOutputStream).Start()
|
||||||
timeSyncProducer.Start()
|
timeSyncProducer.Start()
|
||||||
expected := []uint64{1, 2, 3, 4, 5, 6, 7, 8, 9, 10}
|
expected := []uint64{1, 2, 3, 4, 5, 6, 7, 8, 9, 10}
|
||||||
result_1 := receiveMsg(proxyTtOutputStream)
|
result1 := receiveMsg(proxyTtOutputStream)
|
||||||
assert.Equal(t, expected, result_1)
|
assert.Equal(t, expected, result1)
|
||||||
result_2 := receiveMsg(writeNodeOutputStream)
|
result2 := receiveMsg(writeNodeOutputStream)
|
||||||
assert.Equal(t, expected, result_2)
|
assert.Equal(t, expected, result2)
|
||||||
|
|
||||||
timeSyncProducer.Close()
|
timeSyncProducer.Close()
|
||||||
}
|
}
|
||||||
|
|||||||
@ -1,12 +1,12 @@
|
|||||||
package timesync
|
package timesync
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
"context"
|
||||||
|
"log"
|
||||||
|
|
||||||
"github.com/zilliztech/milvus-distributed/internal/errors"
|
"github.com/zilliztech/milvus-distributed/internal/errors"
|
||||||
ms "github.com/zilliztech/milvus-distributed/internal/msgstream"
|
ms "github.com/zilliztech/milvus-distributed/internal/msgstream"
|
||||||
internalPb "github.com/zilliztech/milvus-distributed/internal/proto/internalpb"
|
internalPb "github.com/zilliztech/milvus-distributed/internal/proto/internalpb"
|
||||||
"log"
|
|
||||||
|
|
||||||
"context"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
type timeSyncMsgProducer struct {
|
type timeSyncMsgProducer struct {
|
||||||
|
|||||||
@ -76,7 +76,7 @@ func (ttBarrier *softTimeTickBarrier) Start() error {
|
|||||||
log.Printf("[softTimeTickBarrier] peer(%d)=%d\n", ttmsg.PeerId, ttmsg.Timestamp)
|
log.Printf("[softTimeTickBarrier] peer(%d)=%d\n", ttmsg.PeerId, ttmsg.Timestamp)
|
||||||
|
|
||||||
if !ok {
|
if !ok {
|
||||||
log.Printf("[softTimeTickBarrier] Warning: peerId %d not exist\n", ttmsg.PeerId)
|
log.Printf("[softTimeTickBarrier] Warning: peerID %d not exist\n", ttmsg.PeerId)
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -108,7 +108,7 @@ func NewSoftTimeTickBarrier(ctx context.Context,
|
|||||||
minTtInterval Timestamp) *softTimeTickBarrier {
|
minTtInterval Timestamp) *softTimeTickBarrier {
|
||||||
|
|
||||||
if len(peerIds) <= 0 {
|
if len(peerIds) <= 0 {
|
||||||
log.Printf("[NewSoftTimeTickBarrier] Error: peerIds is emtpy!\n")
|
log.Printf("[NewSoftTimeTickBarrier] Error: peerIds is empty!\n")
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -193,7 +193,7 @@ func (ttBarrier *hardTimeTickBarrier) Start() error {
|
|||||||
|
|
||||||
oldT, ok := ttBarrier.peer2Tt[ttmsg.PeerId]
|
oldT, ok := ttBarrier.peer2Tt[ttmsg.PeerId]
|
||||||
if !ok {
|
if !ok {
|
||||||
log.Printf("[hardTimeTickBarrier] Warning: peerId %d not exist\n", ttmsg.PeerId)
|
log.Printf("[hardTimeTickBarrier] Warning: peerID %d not exist\n", ttmsg.PeerId)
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -233,7 +233,7 @@ func NewHardTimeTickBarrier(ctx context.Context,
|
|||||||
peerIds []UniqueID) *hardTimeTickBarrier {
|
peerIds []UniqueID) *hardTimeTickBarrier {
|
||||||
|
|
||||||
if len(peerIds) <= 0 {
|
if len(peerIds) <= 0 {
|
||||||
log.Printf("[NewSoftTimeTickBarrier] Error: peerIds is emtpy!")
|
log.Printf("[NewSoftTimeTickBarrier] Error: peerIds is empty!")
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -259,5 +259,4 @@ func (ttBarrier *hardTimeTickBarrier) Close() {
|
|||||||
ttBarrier.closeCh <- struct{}{}
|
ttBarrier.closeCh <- struct{}{}
|
||||||
}
|
}
|
||||||
ttBarrier.closed = true
|
ttBarrier.closed = true
|
||||||
return
|
|
||||||
}
|
}
|
||||||
|
|||||||
@ -9,18 +9,19 @@ import (
|
|||||||
|
|
||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
"github.com/stretchr/testify/require"
|
"github.com/stretchr/testify/require"
|
||||||
|
|
||||||
ms "github.com/zilliztech/milvus-distributed/internal/msgstream"
|
ms "github.com/zilliztech/milvus-distributed/internal/msgstream"
|
||||||
internalPb "github.com/zilliztech/milvus-distributed/internal/proto/internalpb"
|
internalPb "github.com/zilliztech/milvus-distributed/internal/proto/internalpb"
|
||||||
)
|
)
|
||||||
|
|
||||||
func getTtMsg(msgType internalPb.MsgType, peerId UniqueID, timeStamp uint64) *ms.TsMsg {
|
func getTtMsg(msgType internalPb.MsgType, peerID UniqueID, timeStamp uint64) *ms.TsMsg {
|
||||||
var tsMsg ms.TsMsg
|
var tsMsg ms.TsMsg
|
||||||
baseMsg := ms.BaseMsg{
|
baseMsg := ms.BaseMsg{
|
||||||
HashValues: []int32{int32(peerId)},
|
HashValues: []int32{int32(peerID)},
|
||||||
}
|
}
|
||||||
timeTickResult := internalPb.TimeTickMsg{
|
timeTickResult := internalPb.TimeTickMsg{
|
||||||
MsgType: internalPb.MsgType_kTimeTick,
|
MsgType: internalPb.MsgType_kTimeTick,
|
||||||
PeerId: peerId,
|
PeerId: peerID,
|
||||||
Timestamp: timeStamp,
|
Timestamp: timeStamp,
|
||||||
}
|
}
|
||||||
timeTickMsg := &ms.TimeTickMsg{
|
timeTickMsg := &ms.TimeTickMsg{
|
||||||
@ -393,6 +394,8 @@ func TestTt_HardTtBarrierGetTimeTick(t *testing.T) {
|
|||||||
|
|
||||||
// This will stuck
|
// This will stuck
|
||||||
ts, err = sttbarrierStuck.GetTimeTick()
|
ts, err = sttbarrierStuck.GetTimeTick()
|
||||||
|
assert.NotNil(t, err)
|
||||||
|
assert.Equal(t, Timestamp(0), ts)
|
||||||
|
|
||||||
// ---------------------context cancel------------------------
|
// ---------------------context cancel------------------------
|
||||||
channelsCancel := []string{"HardTtBarrierGetTimeTickCancel"}
|
channelsCancel := []string{"HardTtBarrierGetTimeTickCancel"}
|
||||||
@ -424,5 +427,6 @@ func TestTt_HardTtBarrierGetTimeTick(t *testing.T) {
|
|||||||
|
|
||||||
// This will stuck
|
// This will stuck
|
||||||
ts, err = sttbarrierCancel.GetTimeTick()
|
ts, err = sttbarrierCancel.GetTimeTick()
|
||||||
|
assert.NotNil(t, err)
|
||||||
|
assert.Equal(t, Timestamp(0), ts)
|
||||||
}
|
}
|
||||||
|
|||||||
@ -464,7 +464,7 @@ func TestStream_PulsarMsgStream_DefaultRepackFunc(t *testing.T) {
|
|||||||
Timestamp: uint64(1),
|
Timestamp: uint64(1),
|
||||||
}
|
}
|
||||||
timeTick := &TimeTickMsg{
|
timeTick := &TimeTickMsg{
|
||||||
BaseMsg: baseMsg,
|
BaseMsg: baseMsg,
|
||||||
TimeTickMsg: timeTickRequest,
|
TimeTickMsg: timeTickRequest,
|
||||||
}
|
}
|
||||||
var tsMsg TsMsg = timeTick
|
var tsMsg TsMsg = timeTick
|
||||||
|
|||||||
@ -74,11 +74,11 @@ func (queue *BaseTaskQueue) PopActiveTask(ts Timestamp) task {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (queue *BaseTaskQueue) getTaskByReqID(reqId UniqueID) task {
|
func (queue *BaseTaskQueue) getTaskByReqID(reqID UniqueID) task {
|
||||||
queue.utLock.Lock()
|
queue.utLock.Lock()
|
||||||
defer queue.utLock.Lock()
|
defer queue.utLock.Lock()
|
||||||
for e := queue.unissuedTasks.Front(); e != nil; e = e.Next() {
|
for e := queue.unissuedTasks.Front(); e != nil; e = e.Next() {
|
||||||
if e.Value.(task).ID() == reqId {
|
if e.Value.(task).ID() == reqID {
|
||||||
return e.Value.(task)
|
return e.Value.(task)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -86,7 +86,7 @@ func (queue *BaseTaskQueue) getTaskByReqID(reqId UniqueID) task {
|
|||||||
queue.atLock.Lock()
|
queue.atLock.Lock()
|
||||||
defer queue.atLock.Unlock()
|
defer queue.atLock.Unlock()
|
||||||
for ats := range queue.activeTasks {
|
for ats := range queue.activeTasks {
|
||||||
if queue.activeTasks[ats].ID() == reqId {
|
if queue.activeTasks[ats].ID() == reqID {
|
||||||
return queue.activeTasks[ats]
|
return queue.activeTasks[ats]
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -216,14 +216,14 @@ func (sched *TaskScheduler) scheduleDqTask() task {
|
|||||||
return sched.DqQueue.PopUnissuedTask()
|
return sched.DqQueue.PopUnissuedTask()
|
||||||
}
|
}
|
||||||
|
|
||||||
func (sched *TaskScheduler) getTaskByReqID(reqID UniqueID) task {
|
func (sched *TaskScheduler) getTaskByReqID(collMeta UniqueID) task {
|
||||||
if t := sched.DdQueue.getTaskByReqID(reqID); t != nil {
|
if t := sched.DdQueue.getTaskByReqID(collMeta); t != nil {
|
||||||
return t
|
return t
|
||||||
}
|
}
|
||||||
if t := sched.DmQueue.getTaskByReqID(reqID); t != nil {
|
if t := sched.DmQueue.getTaskByReqID(collMeta); t != nil {
|
||||||
return t
|
return t
|
||||||
}
|
}
|
||||||
if t := sched.DqQueue.getTaskByReqID(reqID); t != nil {
|
if t := sched.DqQueue.getTaskByReqID(collMeta); t != nil {
|
||||||
return t
|
return t
|
||||||
}
|
}
|
||||||
return nil
|
return nil
|
||||||
|
|||||||
@ -61,7 +61,7 @@ func (tt *timeTick) tick() error {
|
|||||||
tt.currentTick = ts
|
tt.currentTick = ts
|
||||||
}
|
}
|
||||||
|
|
||||||
if tt.areRequestsDelivered(tt.currentTick) == false {
|
if !tt.areRequestsDelivered(tt.currentTick) {
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
msgPack := msgstream.MsgPack{}
|
msgPack := msgstream.MsgPack{}
|
||||||
|
|||||||
@ -14,8 +14,8 @@ import (
|
|||||||
//----------------------------------------------------------------------------------------------------- collection
|
//----------------------------------------------------------------------------------------------------- collection
|
||||||
func TestColSegContainer_addCollection(t *testing.T) {
|
func TestColSegContainer_addCollection(t *testing.T) {
|
||||||
ctx := context.Background()
|
ctx := context.Background()
|
||||||
pulsarUrl := "pulsar://localhost:6650"
|
pulsarURL := "pulsar://localhost:6650"
|
||||||
node := NewQueryNode(ctx, 0, pulsarUrl)
|
node := NewQueryNode(ctx, 0, pulsarURL)
|
||||||
|
|
||||||
fieldVec := schemapb.FieldSchema{
|
fieldVec := schemapb.FieldSchema{
|
||||||
Name: "vec",
|
Name: "vec",
|
||||||
@ -66,8 +66,8 @@ func TestColSegContainer_addCollection(t *testing.T) {
|
|||||||
|
|
||||||
func TestColSegContainer_removeCollection(t *testing.T) {
|
func TestColSegContainer_removeCollection(t *testing.T) {
|
||||||
ctx := context.Background()
|
ctx := context.Background()
|
||||||
pulsarUrl := "pulsar://localhost:6650"
|
pulsarURL := "pulsar://localhost:6650"
|
||||||
node := NewQueryNode(ctx, 0, pulsarUrl)
|
node := NewQueryNode(ctx, 0, pulsarURL)
|
||||||
|
|
||||||
fieldVec := schemapb.FieldSchema{
|
fieldVec := schemapb.FieldSchema{
|
||||||
Name: "vec",
|
Name: "vec",
|
||||||
@ -122,8 +122,8 @@ func TestColSegContainer_removeCollection(t *testing.T) {
|
|||||||
|
|
||||||
func TestColSegContainer_getCollectionByID(t *testing.T) {
|
func TestColSegContainer_getCollectionByID(t *testing.T) {
|
||||||
ctx := context.Background()
|
ctx := context.Background()
|
||||||
pulsarUrl := "pulsar://localhost:6650"
|
pulsarURL := "pulsar://localhost:6650"
|
||||||
node := NewQueryNode(ctx, 0, pulsarUrl)
|
node := NewQueryNode(ctx, 0, pulsarURL)
|
||||||
|
|
||||||
fieldVec := schemapb.FieldSchema{
|
fieldVec := schemapb.FieldSchema{
|
||||||
Name: "vec",
|
Name: "vec",
|
||||||
@ -180,8 +180,8 @@ func TestColSegContainer_getCollectionByID(t *testing.T) {
|
|||||||
|
|
||||||
func TestColSegContainer_getCollectionByName(t *testing.T) {
|
func TestColSegContainer_getCollectionByName(t *testing.T) {
|
||||||
ctx := context.Background()
|
ctx := context.Background()
|
||||||
pulsarUrl := "pulsar://localhost:6650"
|
pulsarURL := "pulsar://localhost:6650"
|
||||||
node := NewQueryNode(ctx, 0, pulsarUrl)
|
node := NewQueryNode(ctx, 0, pulsarURL)
|
||||||
|
|
||||||
fieldVec := schemapb.FieldSchema{
|
fieldVec := schemapb.FieldSchema{
|
||||||
Name: "vec",
|
Name: "vec",
|
||||||
@ -239,8 +239,8 @@ func TestColSegContainer_getCollectionByName(t *testing.T) {
|
|||||||
//----------------------------------------------------------------------------------------------------- partition
|
//----------------------------------------------------------------------------------------------------- partition
|
||||||
func TestColSegContainer_addPartition(t *testing.T) {
|
func TestColSegContainer_addPartition(t *testing.T) {
|
||||||
ctx := context.Background()
|
ctx := context.Background()
|
||||||
pulsarUrl := "pulsar://localhost:6650"
|
pulsarURL := "pulsar://localhost:6650"
|
||||||
node := NewQueryNode(ctx, 0, pulsarUrl)
|
node := NewQueryNode(ctx, 0, pulsarURL)
|
||||||
|
|
||||||
fieldVec := schemapb.FieldSchema{
|
fieldVec := schemapb.FieldSchema{
|
||||||
Name: "vec",
|
Name: "vec",
|
||||||
@ -297,8 +297,8 @@ func TestColSegContainer_addPartition(t *testing.T) {
|
|||||||
|
|
||||||
func TestColSegContainer_removePartition(t *testing.T) {
|
func TestColSegContainer_removePartition(t *testing.T) {
|
||||||
ctx := context.Background()
|
ctx := context.Background()
|
||||||
pulsarUrl := "pulsar://localhost:6650"
|
pulsarURL := "pulsar://localhost:6650"
|
||||||
node := NewQueryNode(ctx, 0, pulsarUrl)
|
node := NewQueryNode(ctx, 0, pulsarURL)
|
||||||
|
|
||||||
fieldVec := schemapb.FieldSchema{
|
fieldVec := schemapb.FieldSchema{
|
||||||
Name: "vec",
|
Name: "vec",
|
||||||
@ -357,8 +357,8 @@ func TestColSegContainer_removePartition(t *testing.T) {
|
|||||||
|
|
||||||
func TestColSegContainer_getPartitionByTag(t *testing.T) {
|
func TestColSegContainer_getPartitionByTag(t *testing.T) {
|
||||||
ctx := context.Background()
|
ctx := context.Background()
|
||||||
pulsarUrl := "pulsar://localhost:6650"
|
pulsarURL := "pulsar://localhost:6650"
|
||||||
node := NewQueryNode(ctx, 0, pulsarUrl)
|
node := NewQueryNode(ctx, 0, pulsarURL)
|
||||||
|
|
||||||
fieldVec := schemapb.FieldSchema{
|
fieldVec := schemapb.FieldSchema{
|
||||||
Name: "vec",
|
Name: "vec",
|
||||||
@ -420,8 +420,8 @@ func TestColSegContainer_getPartitionByTag(t *testing.T) {
|
|||||||
//----------------------------------------------------------------------------------------------------- segment
|
//----------------------------------------------------------------------------------------------------- segment
|
||||||
func TestColSegContainer_addSegment(t *testing.T) {
|
func TestColSegContainer_addSegment(t *testing.T) {
|
||||||
ctx := context.Background()
|
ctx := context.Background()
|
||||||
pulsarUrl := "pulsar://localhost:6650"
|
pulsarURL := "pulsar://localhost:6650"
|
||||||
node := NewQueryNode(ctx, 0, pulsarUrl)
|
node := NewQueryNode(ctx, 0, pulsarURL)
|
||||||
|
|
||||||
fieldVec := schemapb.FieldSchema{
|
fieldVec := schemapb.FieldSchema{
|
||||||
Name: "vec",
|
Name: "vec",
|
||||||
@ -482,8 +482,8 @@ func TestColSegContainer_addSegment(t *testing.T) {
|
|||||||
|
|
||||||
func TestColSegContainer_removeSegment(t *testing.T) {
|
func TestColSegContainer_removeSegment(t *testing.T) {
|
||||||
ctx := context.Background()
|
ctx := context.Background()
|
||||||
pulsarUrl := "pulsar://localhost:6650"
|
pulsarURL := "pulsar://localhost:6650"
|
||||||
node := NewQueryNode(ctx, 0, pulsarUrl)
|
node := NewQueryNode(ctx, 0, pulsarURL)
|
||||||
|
|
||||||
fieldVec := schemapb.FieldSchema{
|
fieldVec := schemapb.FieldSchema{
|
||||||
Name: "vec",
|
Name: "vec",
|
||||||
@ -546,8 +546,8 @@ func TestColSegContainer_removeSegment(t *testing.T) {
|
|||||||
|
|
||||||
func TestColSegContainer_getSegmentByID(t *testing.T) {
|
func TestColSegContainer_getSegmentByID(t *testing.T) {
|
||||||
ctx := context.Background()
|
ctx := context.Background()
|
||||||
pulsarUrl := "pulsar://localhost:6650"
|
pulsarURL := "pulsar://localhost:6650"
|
||||||
node := NewQueryNode(ctx, 0, pulsarUrl)
|
node := NewQueryNode(ctx, 0, pulsarURL)
|
||||||
|
|
||||||
fieldVec := schemapb.FieldSchema{
|
fieldVec := schemapb.FieldSchema{
|
||||||
Name: "vec",
|
Name: "vec",
|
||||||
@ -611,8 +611,8 @@ func TestColSegContainer_getSegmentByID(t *testing.T) {
|
|||||||
|
|
||||||
func TestColSegContainer_hasSegment(t *testing.T) {
|
func TestColSegContainer_hasSegment(t *testing.T) {
|
||||||
ctx := context.Background()
|
ctx := context.Background()
|
||||||
pulsarUrl := "pulsar://localhost:6650"
|
pulsarURL := "pulsar://localhost:6650"
|
||||||
node := NewQueryNode(ctx, 0, pulsarUrl)
|
node := NewQueryNode(ctx, 0, pulsarURL)
|
||||||
|
|
||||||
fieldVec := schemapb.FieldSchema{
|
fieldVec := schemapb.FieldSchema{
|
||||||
Name: "vec",
|
Name: "vec",
|
||||||
|
|||||||
@ -13,8 +13,8 @@ import (
|
|||||||
|
|
||||||
func TestCollection_Partitions(t *testing.T) {
|
func TestCollection_Partitions(t *testing.T) {
|
||||||
ctx := context.Background()
|
ctx := context.Background()
|
||||||
pulsarUrl := "pulsar://localhost:6650"
|
pulsarURL := "pulsar://localhost:6650"
|
||||||
node := NewQueryNode(ctx, 0, pulsarUrl)
|
node := NewQueryNode(ctx, 0, pulsarURL)
|
||||||
|
|
||||||
fieldVec := schemapb.FieldSchema{
|
fieldVec := schemapb.FieldSchema{
|
||||||
Name: "vec",
|
Name: "vec",
|
||||||
|
|||||||
@ -3,13 +3,14 @@ package reader
|
|||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
"encoding/binary"
|
"encoding/binary"
|
||||||
|
"math"
|
||||||
|
"testing"
|
||||||
|
"time"
|
||||||
|
|
||||||
"github.com/golang/protobuf/proto"
|
"github.com/golang/protobuf/proto"
|
||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
"github.com/zilliztech/milvus-distributed/internal/proto/etcdpb"
|
"github.com/zilliztech/milvus-distributed/internal/proto/etcdpb"
|
||||||
"github.com/zilliztech/milvus-distributed/internal/proto/schemapb"
|
"github.com/zilliztech/milvus-distributed/internal/proto/schemapb"
|
||||||
"math"
|
|
||||||
"testing"
|
|
||||||
"time"
|
|
||||||
|
|
||||||
"github.com/zilliztech/milvus-distributed/internal/msgstream"
|
"github.com/zilliztech/milvus-distributed/internal/msgstream"
|
||||||
"github.com/zilliztech/milvus-distributed/internal/proto/commonpb"
|
"github.com/zilliztech/milvus-distributed/internal/proto/commonpb"
|
||||||
@ -33,8 +34,8 @@ func TestManipulationService_Start(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// init query node
|
// init query node
|
||||||
pulsarUrl := "pulsar://localhost:6650"
|
pulsarURL := "pulsar://localhost:6650"
|
||||||
node := NewQueryNode(ctx, 0, pulsarUrl)
|
node := NewQueryNode(ctx, 0, pulsarURL)
|
||||||
|
|
||||||
// init meta
|
// init meta
|
||||||
fieldVec := schemapb.FieldSchema{
|
fieldVec := schemapb.FieldSchema{
|
||||||
@ -157,7 +158,7 @@ func TestManipulationService_Start(t *testing.T) {
|
|||||||
producerChannels := []string{"insert"}
|
producerChannels := []string{"insert"}
|
||||||
|
|
||||||
insertStream := msgstream.NewPulsarMsgStream(ctx, receiveBufSize)
|
insertStream := msgstream.NewPulsarMsgStream(ctx, receiveBufSize)
|
||||||
insertStream.SetPulsarCient(pulsarUrl)
|
insertStream.SetPulsarCient(pulsarURL)
|
||||||
insertStream.CreatePulsarProducers(producerChannels)
|
insertStream.CreatePulsarProducers(producerChannels)
|
||||||
|
|
||||||
var insertMsgStream msgstream.MsgStream = insertStream
|
var insertMsgStream msgstream.MsgStream = insertStream
|
||||||
@ -171,10 +172,5 @@ func TestManipulationService_Start(t *testing.T) {
|
|||||||
|
|
||||||
node.Close()
|
node.Close()
|
||||||
|
|
||||||
for {
|
<-ctx.Done()
|
||||||
select {
|
|
||||||
case <-ctx.Done():
|
|
||||||
return
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|||||||
@ -13,8 +13,8 @@ import (
|
|||||||
|
|
||||||
func TestPartition_Segments(t *testing.T) {
|
func TestPartition_Segments(t *testing.T) {
|
||||||
ctx := context.Background()
|
ctx := context.Background()
|
||||||
pulsarUrl := "pulsar://localhost:6650"
|
pulsarURL := "pulsar://localhost:6650"
|
||||||
node := NewQueryNode(ctx, 0, pulsarUrl)
|
node := NewQueryNode(ctx, 0, pulsarURL)
|
||||||
|
|
||||||
fieldVec := schemapb.FieldSchema{
|
fieldVec := schemapb.FieldSchema{
|
||||||
Name: "vec",
|
Name: "vec",
|
||||||
|
|||||||
@ -539,8 +539,8 @@ func TestSegment_segmentDelete(t *testing.T) {
|
|||||||
//func TestSegment_segmentSearch(t *testing.T) {
|
//func TestSegment_segmentSearch(t *testing.T) {
|
||||||
// ctx := context.Background()
|
// ctx := context.Background()
|
||||||
// // 1. Construct node, collection, partition and segment
|
// // 1. Construct node, collection, partition and segment
|
||||||
// pulsarUrl := "pulsar://localhost:6650"
|
// pulsarURL := "pulsar://localhost:6650"
|
||||||
// node := NewQueryNode(ctx, 0, pulsarUrl)
|
// node := NewQueryNode(ctx, 0, pulsarURL)
|
||||||
// var collection = node.newCollection(0, "collection0", "")
|
// var collection = node.newCollection(0, "collection0", "")
|
||||||
// var partition = collection.newPartition("partition0")
|
// var partition = collection.newPartition("partition0")
|
||||||
// var segment = partition.newSegment(0)
|
// var segment = partition.newSegment(0)
|
||||||
|
|||||||
@ -1,8 +1,9 @@
|
|||||||
package flowgraph
|
package flowgraph
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"github.com/zilliztech/milvus-distributed/internal/msgstream"
|
|
||||||
"log"
|
"log"
|
||||||
|
|
||||||
|
"github.com/zilliztech/milvus-distributed/internal/msgstream"
|
||||||
)
|
)
|
||||||
|
|
||||||
type InputNode struct {
|
type InputNode struct {
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user