mirror of
https://gitee.com/milvus-io/milvus.git
synced 2025-12-28 22:45:26 +08:00
related: #45993 This commit extends nullable vector support to the proxy layer, querynode, and adds comprehensive validation, search reduce, and field data handling for nullable vectors with sparse storage. Proxy layer changes: - Update validate_util.go checkAligned() with getExpectedVectorRows() helper to validate nullable vector field alignment using valid data count - Update checkFloatVectorFieldData/checkSparseFloatVectorFieldData for nullable vector validation with proper row count expectations - Add FieldDataIdxComputer in typeutil/schema.go for logical-to-physical index translation during search reduce operations - Update search_reduce_util.go reduceSearchResultData to use idxComputers for correct field data indexing with nullable vectors - Update task.go, task_query.go, task_upsert.go for nullable vector handling - Update msg_pack.go with nullable vector field data processing QueryNode layer changes: - Update segments/result.go for nullable vector result handling - Update segments/search_reduce.go with nullable vector offset translation Storage and index changes: - Update data_codec.go and utils.go for nullable vector serialization - Update indexcgowrapper/dataset.go and index.go for nullable vector indexing Utility changes: - Add FieldDataIdxComputer struct with Compute() method for efficient logical-to-physical index mapping across multiple field data - Update EstimateEntitySize() and AppendFieldData() with fieldIdxs parameter - Update funcutil.go with nullable vector support functions <!-- This is an auto-generated comment: release notes by coderabbit.ai --> ## Summary by CodeRabbit * **New Features** * Full support for nullable vector fields (float, binary, float16, bfloat16, int8, sparse) across ingest, storage, indexing, search and retrieval; logical↔physical offset mapping preserves row semantics. * Client: compaction control and compaction-state APIs. * **Bug Fixes** * Improved validation for adding vector fields (nullable + dimension checks) and corrected search/query behavior for nullable vectors. * **Chores** * Persisted validity maps with indexes and on-disk formats. * **Tests** * Extensive new and updated end-to-end nullable-vector tests. <sub>✏️ Tip: You can customize this high-level summary in your review settings.</sub> <!-- end of auto-generated comment: release notes by coderabbit.ai --> --------- Signed-off-by: marcelo-cjl <marcelo.chen@zilliz.com>
559 lines
17 KiB
Go
559 lines
17 KiB
Go
// Licensed to the LF AI & Data foundation under one
|
|
// or more contributor license agreements. See the NOTICE file
|
|
// distributed with this work for additional information
|
|
// regarding copyright ownership. The ASF licenses this file
|
|
// to you under the Apache License, Version 2.0 (the
|
|
// "License"); you may not use this file except in compliance
|
|
// with the License. You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
package column
|
|
|
|
import (
|
|
"fmt"
|
|
|
|
"github.com/cockroachdb/errors"
|
|
"github.com/samber/lo"
|
|
|
|
"github.com/milvus-io/milvus-proto/go-api/v2/schemapb"
|
|
"github.com/milvus-io/milvus/client/v2/entity"
|
|
)
|
|
|
|
// Column interface field type for column-based data frame
|
|
type Column interface {
|
|
Name() string
|
|
Type() entity.FieldType
|
|
Len() int
|
|
Slice(int, int) Column
|
|
FieldData() *schemapb.FieldData
|
|
AppendValue(interface{}) error
|
|
Get(int) (interface{}, error)
|
|
GetAsInt64(int) (int64, error)
|
|
GetAsString(int) (string, error)
|
|
GetAsDouble(int) (float64, error)
|
|
GetAsBool(int) (bool, error)
|
|
// nullable related API
|
|
AppendNull() error
|
|
IsNull(int) (bool, error)
|
|
Nullable() bool
|
|
SetNullable(bool)
|
|
ValidateNullable() error
|
|
CompactNullableValues()
|
|
ValidCount() int
|
|
}
|
|
|
|
var errFieldDataTypeNotMatch = errors.New("FieldData type not matched")
|
|
|
|
// IDColumns converts schemapb.IDs to corresponding column
|
|
// currently Int64 / string may be in IDs
|
|
func IDColumns(schema *entity.Schema, ids *schemapb.IDs, begin, end int) (Column, error) {
|
|
var idColumn Column
|
|
pkField := schema.PKField()
|
|
if pkField == nil {
|
|
return nil, errors.New("PK Field not found")
|
|
}
|
|
switch pkField.DataType {
|
|
case entity.FieldTypeInt64:
|
|
data := ids.GetIntId().GetData()
|
|
if data == nil {
|
|
return NewColumnInt64(pkField.Name, nil), nil
|
|
}
|
|
if end >= 0 {
|
|
idColumn = NewColumnInt64(pkField.Name, data[begin:end])
|
|
} else {
|
|
idColumn = NewColumnInt64(pkField.Name, data[begin:])
|
|
}
|
|
case entity.FieldTypeVarChar, entity.FieldTypeString:
|
|
data := ids.GetStrId().GetData()
|
|
if data == nil {
|
|
return NewColumnVarChar(pkField.Name, nil), nil
|
|
}
|
|
if end >= 0 {
|
|
idColumn = NewColumnVarChar(pkField.Name, data[begin:end])
|
|
} else {
|
|
idColumn = NewColumnVarChar(pkField.Name, data[begin:])
|
|
}
|
|
default:
|
|
return nil, fmt.Errorf("unsupported id type %v", pkField.DataType)
|
|
}
|
|
return idColumn, nil
|
|
}
|
|
|
|
func parseScalarData[T any, COL Column, NCOL Column](
|
|
name string,
|
|
data []T,
|
|
start, end int,
|
|
validData []bool,
|
|
creator func(string, []T) COL,
|
|
nullableCreator func(string, []T, []bool, ...ColumnOption[T]) (NCOL, error),
|
|
) (Column, error) {
|
|
if end < 0 {
|
|
end = len(data)
|
|
}
|
|
data = data[start:end]
|
|
if len(validData) > 0 {
|
|
validData = validData[start:end]
|
|
ncol, err := nullableCreator(name, data, validData, WithSparseNullableMode[T](true))
|
|
return ncol, err
|
|
}
|
|
|
|
return creator(name, data), nil
|
|
}
|
|
|
|
func parseArrayData(fieldName string, elementType schemapb.DataType, fieldDataList []*schemapb.ScalarField, validData []bool, begin, end int) (Column, error) {
|
|
switch elementType {
|
|
case schemapb.DataType_Bool:
|
|
data := lo.Map(fieldDataList, func(fd *schemapb.ScalarField, _ int) []bool {
|
|
return fd.GetBoolData().GetData()
|
|
})
|
|
return parseScalarData(fieldName, data, begin, end, validData, NewColumnBoolArray, NewNullableColumnBoolArray)
|
|
|
|
case schemapb.DataType_Int8:
|
|
data := lo.Map(fieldDataList, func(fd *schemapb.ScalarField, _ int) []int8 {
|
|
return int32ToType[int8](fd.GetIntData().GetData())
|
|
})
|
|
return parseScalarData(fieldName, data, begin, end, validData, NewColumnInt8Array, NewNullableColumnInt8Array)
|
|
|
|
case schemapb.DataType_Int16:
|
|
data := lo.Map(fieldDataList, func(fd *schemapb.ScalarField, _ int) []int16 {
|
|
return int32ToType[int16](fd.GetIntData().GetData())
|
|
})
|
|
return parseScalarData(fieldName, data, begin, end, validData, NewColumnInt16Array, NewNullableColumnInt16Array)
|
|
|
|
case schemapb.DataType_Int32:
|
|
data := lo.Map(fieldDataList, func(fd *schemapb.ScalarField, _ int) []int32 {
|
|
return fd.GetIntData().GetData()
|
|
})
|
|
return parseScalarData(fieldName, data, begin, end, validData, NewColumnInt32Array, NewNullableColumnInt32Array)
|
|
|
|
case schemapb.DataType_Int64:
|
|
data := lo.Map(fieldDataList, func(fd *schemapb.ScalarField, _ int) []int64 {
|
|
return fd.GetLongData().GetData()
|
|
})
|
|
return parseScalarData(fieldName, data, begin, end, validData, NewColumnInt64Array, NewNullableColumnInt64Array)
|
|
|
|
case schemapb.DataType_Float:
|
|
data := lo.Map(fieldDataList, func(fd *schemapb.ScalarField, _ int) []float32 {
|
|
return fd.GetFloatData().GetData()
|
|
})
|
|
return parseScalarData(fieldName, data, begin, end, validData, NewColumnFloatArray, NewNullableColumnFloatArray)
|
|
|
|
case schemapb.DataType_Double:
|
|
data := lo.Map(fieldDataList, func(fd *schemapb.ScalarField, _ int) []float64 {
|
|
return fd.GetDoubleData().GetData()
|
|
})
|
|
return parseScalarData(fieldName, data, begin, end, validData, NewColumnDoubleArray, NewNullableColumnDoubleArray)
|
|
|
|
case schemapb.DataType_VarChar, schemapb.DataType_String:
|
|
data := lo.Map(fieldDataList, func(fd *schemapb.ScalarField, _ int) []string {
|
|
return fd.GetStringData().GetData()
|
|
})
|
|
return parseScalarData(fieldName, data, begin, end, validData, NewColumnVarCharArray, NewNullableColumnVarCharArray)
|
|
|
|
default:
|
|
return nil, fmt.Errorf("unsupported element type %s", elementType)
|
|
}
|
|
}
|
|
|
|
func parseStructArrayData(fieldName string, structArray *schemapb.StructArrayField, begin, end int) (Column, error) {
|
|
var fields []Column
|
|
for _, field := range structArray.GetFields() {
|
|
field, err := FieldDataColumn(field, begin, end)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
fields = append(fields, field)
|
|
}
|
|
return NewColumnStructArray(fieldName, fields), nil
|
|
}
|
|
|
|
func int32ToType[T ~int8 | int16](data []int32) []T {
|
|
return lo.Map(data, func(i32 int32, _ int) T {
|
|
return T(i32)
|
|
})
|
|
}
|
|
|
|
// FieldDataColumn converts schemapb.FieldData to Column, used int search result conversion logic
|
|
// begin, end specifies the start and end positions
|
|
func FieldDataColumn(fd *schemapb.FieldData, begin, end int) (Column, error) {
|
|
validData := fd.GetValidData()
|
|
|
|
switch fd.GetType() {
|
|
case schemapb.DataType_Bool:
|
|
return parseScalarData(fd.GetFieldName(), fd.GetScalars().GetBoolData().GetData(), begin, end, validData, NewColumnBool, NewNullableColumnBool)
|
|
|
|
case schemapb.DataType_Int8:
|
|
data := int32ToType[int8](fd.GetScalars().GetIntData().GetData())
|
|
return parseScalarData(fd.GetFieldName(), data, begin, end, validData, NewColumnInt8, NewNullableColumnInt8)
|
|
|
|
case schemapb.DataType_Int16:
|
|
data := int32ToType[int16](fd.GetScalars().GetIntData().GetData())
|
|
return parseScalarData(fd.GetFieldName(), data, begin, end, validData, NewColumnInt16, NewNullableColumnInt16)
|
|
|
|
case schemapb.DataType_Int32:
|
|
return parseScalarData(fd.GetFieldName(), fd.GetScalars().GetIntData().GetData(), begin, end, validData, NewColumnInt32, NewNullableColumnInt32)
|
|
|
|
case schemapb.DataType_Int64:
|
|
return parseScalarData(fd.GetFieldName(), fd.GetScalars().GetLongData().GetData(), begin, end, validData, NewColumnInt64, NewNullableColumnInt64)
|
|
|
|
case schemapb.DataType_Float:
|
|
return parseScalarData(fd.GetFieldName(), fd.GetScalars().GetFloatData().GetData(), begin, end, validData, NewColumnFloat, NewNullableColumnFloat)
|
|
|
|
case schemapb.DataType_Double:
|
|
return parseScalarData(fd.GetFieldName(), fd.GetScalars().GetDoubleData().GetData(), begin, end, validData, NewColumnDouble, NewNullableColumnDouble)
|
|
|
|
case schemapb.DataType_Timestamptz:
|
|
return parseScalarData(fd.GetFieldName(), fd.GetScalars().GetTimestamptzData().GetData(), begin, end, validData, NewColumnTimestamptz, NewNullableColumnTimestamptz)
|
|
|
|
case schemapb.DataType_String:
|
|
return parseScalarData(fd.GetFieldName(), fd.GetScalars().GetStringData().GetData(), begin, end, validData, NewColumnString, NewNullableColumnString)
|
|
|
|
case schemapb.DataType_VarChar:
|
|
return parseScalarData(fd.GetFieldName(), fd.GetScalars().GetStringData().GetData(), begin, end, validData, NewColumnVarChar, NewNullableColumnVarChar)
|
|
|
|
case schemapb.DataType_Array:
|
|
// handle struct array field
|
|
if fd.GetStructArrays() != nil {
|
|
return parseStructArrayData(fd.GetFieldName(), fd.GetStructArrays(), begin, end)
|
|
}
|
|
data := fd.GetScalars().GetArrayData()
|
|
return parseArrayData(fd.GetFieldName(), data.GetElementType(), data.GetData(), validData, begin, end)
|
|
|
|
case schemapb.DataType_JSON:
|
|
return parseScalarData(fd.GetFieldName(), fd.GetScalars().GetJsonData().GetData(), begin, end, validData, NewColumnJSONBytes, NewNullableColumnJSONBytes)
|
|
|
|
case schemapb.DataType_Geometry:
|
|
return parseScalarData(fd.GetFieldName(), fd.GetScalars().GetGeometryWktData().GetData(), begin, end, validData, NewColumnGeometryWKT, NewNullableColumnGeometryWKT)
|
|
|
|
case schemapb.DataType_FloatVector:
|
|
vectors := fd.GetVectors()
|
|
x, ok := vectors.GetData().(*schemapb.VectorField_FloatVector)
|
|
if !ok {
|
|
return nil, errFieldDataTypeNotMatch
|
|
}
|
|
data := x.FloatVector.GetData()
|
|
dim := int(vectors.GetDim())
|
|
|
|
if len(validData) > 0 {
|
|
if end < 0 {
|
|
end = len(validData)
|
|
}
|
|
vector := make([][]float32, 0, end-begin)
|
|
dataIdx := 0
|
|
for i := 0; i < begin; i++ {
|
|
if validData[i] {
|
|
dataIdx++
|
|
}
|
|
}
|
|
for i := begin; i < end; i++ {
|
|
if validData[i] {
|
|
v := make([]float32, dim)
|
|
copy(v, data[dataIdx*dim:(dataIdx+1)*dim])
|
|
vector = append(vector, v)
|
|
dataIdx++
|
|
} else {
|
|
vector = append(vector, nil)
|
|
}
|
|
}
|
|
col := NewColumnFloatVector(fd.GetFieldName(), dim, vector)
|
|
col.withValidData(validData[begin:end])
|
|
col.nullable = true
|
|
col.sparseMode = true
|
|
return col, nil
|
|
}
|
|
|
|
if end < 0 {
|
|
end = len(data) / dim
|
|
}
|
|
vector := make([][]float32, 0, end-begin)
|
|
for i := begin; i < end; i++ {
|
|
v := make([]float32, dim)
|
|
copy(v, data[i*dim:(i+1)*dim])
|
|
vector = append(vector, v)
|
|
}
|
|
return NewColumnFloatVector(fd.GetFieldName(), dim, vector), nil
|
|
|
|
case schemapb.DataType_BinaryVector:
|
|
vectors := fd.GetVectors()
|
|
x, ok := vectors.GetData().(*schemapb.VectorField_BinaryVector)
|
|
if !ok {
|
|
return nil, errFieldDataTypeNotMatch
|
|
}
|
|
data := x.BinaryVector
|
|
if data == nil {
|
|
return nil, errFieldDataTypeNotMatch
|
|
}
|
|
dim := int(vectors.GetDim())
|
|
blen := dim / 8
|
|
|
|
if len(validData) > 0 {
|
|
if end < 0 {
|
|
end = len(validData)
|
|
}
|
|
vector := make([][]byte, 0, end-begin)
|
|
dataIdx := 0
|
|
for i := 0; i < begin; i++ {
|
|
if validData[i] {
|
|
dataIdx++
|
|
}
|
|
}
|
|
for i := begin; i < end; i++ {
|
|
if validData[i] {
|
|
v := make([]byte, blen)
|
|
copy(v, data[dataIdx*blen:(dataIdx+1)*blen])
|
|
vector = append(vector, v)
|
|
dataIdx++
|
|
} else {
|
|
vector = append(vector, nil)
|
|
}
|
|
}
|
|
col := NewColumnBinaryVector(fd.GetFieldName(), dim, vector)
|
|
col.withValidData(validData[begin:end])
|
|
col.nullable = true
|
|
col.sparseMode = true
|
|
return col, nil
|
|
}
|
|
|
|
if end < 0 {
|
|
end = len(data) / blen
|
|
}
|
|
vector := make([][]byte, 0, end-begin)
|
|
for i := begin; i < end; i++ {
|
|
v := make([]byte, blen)
|
|
copy(v, data[i*blen:(i+1)*blen])
|
|
vector = append(vector, v)
|
|
}
|
|
return NewColumnBinaryVector(fd.GetFieldName(), dim, vector), nil
|
|
|
|
case schemapb.DataType_Float16Vector:
|
|
vectors := fd.GetVectors()
|
|
x, ok := vectors.GetData().(*schemapb.VectorField_Float16Vector)
|
|
if !ok {
|
|
return nil, errFieldDataTypeNotMatch
|
|
}
|
|
data := x.Float16Vector
|
|
dim := int(vectors.GetDim())
|
|
bytePerRow := dim * 2
|
|
|
|
if len(validData) > 0 {
|
|
if end < 0 {
|
|
end = len(validData)
|
|
}
|
|
vector := make([][]byte, 0, end-begin)
|
|
dataIdx := 0
|
|
for i := 0; i < begin; i++ {
|
|
if validData[i] {
|
|
dataIdx++
|
|
}
|
|
}
|
|
for i := begin; i < end; i++ {
|
|
if validData[i] {
|
|
v := make([]byte, bytePerRow)
|
|
copy(v, data[dataIdx*bytePerRow:(dataIdx+1)*bytePerRow])
|
|
vector = append(vector, v)
|
|
dataIdx++
|
|
} else {
|
|
vector = append(vector, nil)
|
|
}
|
|
}
|
|
col := NewColumnFloat16Vector(fd.GetFieldName(), dim, vector)
|
|
col.withValidData(validData[begin:end])
|
|
col.nullable = true
|
|
col.sparseMode = true
|
|
return col, nil
|
|
}
|
|
|
|
if end < 0 {
|
|
end = len(data) / bytePerRow
|
|
}
|
|
vector := make([][]byte, 0, end-begin)
|
|
for i := begin; i < end; i++ {
|
|
v := make([]byte, bytePerRow)
|
|
copy(v, data[i*bytePerRow:(i+1)*bytePerRow])
|
|
vector = append(vector, v)
|
|
}
|
|
return NewColumnFloat16Vector(fd.GetFieldName(), dim, vector), nil
|
|
|
|
case schemapb.DataType_BFloat16Vector:
|
|
vectors := fd.GetVectors()
|
|
x, ok := vectors.GetData().(*schemapb.VectorField_Bfloat16Vector)
|
|
if !ok {
|
|
return nil, errFieldDataTypeNotMatch
|
|
}
|
|
data := x.Bfloat16Vector
|
|
dim := int(vectors.GetDim())
|
|
bytePerRow := dim * 2
|
|
|
|
if len(validData) > 0 {
|
|
if end < 0 {
|
|
end = len(validData)
|
|
}
|
|
vector := make([][]byte, 0, end-begin)
|
|
dataIdx := 0
|
|
for i := 0; i < begin; i++ {
|
|
if validData[i] {
|
|
dataIdx++
|
|
}
|
|
}
|
|
for i := begin; i < end; i++ {
|
|
if validData[i] {
|
|
v := make([]byte, bytePerRow)
|
|
copy(v, data[dataIdx*bytePerRow:(dataIdx+1)*bytePerRow])
|
|
vector = append(vector, v)
|
|
dataIdx++
|
|
} else {
|
|
vector = append(vector, nil)
|
|
}
|
|
}
|
|
col := NewColumnBFloat16Vector(fd.GetFieldName(), dim, vector)
|
|
col.withValidData(validData[begin:end])
|
|
col.nullable = true
|
|
col.sparseMode = true
|
|
return col, nil
|
|
}
|
|
|
|
if end < 0 {
|
|
end = len(data) / bytePerRow
|
|
}
|
|
vector := make([][]byte, 0, end-begin)
|
|
for i := begin; i < end; i++ {
|
|
v := make([]byte, bytePerRow)
|
|
copy(v, data[i*bytePerRow:(i+1)*bytePerRow])
|
|
vector = append(vector, v)
|
|
}
|
|
return NewColumnBFloat16Vector(fd.GetFieldName(), dim, vector), nil
|
|
|
|
case schemapb.DataType_SparseFloatVector:
|
|
sparseVectors := fd.GetVectors().GetSparseFloatVector()
|
|
if sparseVectors == nil {
|
|
return nil, errFieldDataTypeNotMatch
|
|
}
|
|
data := sparseVectors.Contents
|
|
|
|
if len(validData) > 0 {
|
|
if end < 0 {
|
|
end = len(validData)
|
|
}
|
|
vectors := make([]entity.SparseEmbedding, 0, end-begin)
|
|
dataIdx := 0
|
|
for i := 0; i < begin; i++ {
|
|
if validData[i] {
|
|
dataIdx++
|
|
}
|
|
}
|
|
for i := begin; i < end; i++ {
|
|
if validData[i] {
|
|
vector, err := entity.DeserializeSliceSparseEmbedding(data[dataIdx])
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
vectors = append(vectors, vector)
|
|
dataIdx++
|
|
} else {
|
|
vectors = append(vectors, nil)
|
|
}
|
|
}
|
|
col := NewColumnSparseVectors(fd.GetFieldName(), vectors)
|
|
col.withValidData(validData[begin:end])
|
|
col.nullable = true
|
|
col.sparseMode = true
|
|
return col, nil
|
|
}
|
|
|
|
if end < 0 {
|
|
end = len(data)
|
|
}
|
|
data = data[begin:end]
|
|
vectors := make([]entity.SparseEmbedding, 0, len(data))
|
|
for _, bs := range data {
|
|
vector, err := entity.DeserializeSliceSparseEmbedding(bs)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
vectors = append(vectors, vector)
|
|
}
|
|
return NewColumnSparseVectors(fd.GetFieldName(), vectors), nil
|
|
|
|
case schemapb.DataType_Int8Vector:
|
|
vectors := fd.GetVectors()
|
|
x, ok := vectors.GetData().(*schemapb.VectorField_Int8Vector)
|
|
if !ok {
|
|
return nil, errFieldDataTypeNotMatch
|
|
}
|
|
data := x.Int8Vector
|
|
dim := int(vectors.GetDim())
|
|
|
|
if len(validData) > 0 {
|
|
if end < 0 {
|
|
end = len(validData)
|
|
}
|
|
vector := make([][]int8, 0, end-begin)
|
|
dataIdx := 0
|
|
for i := 0; i < begin; i++ {
|
|
if validData[i] {
|
|
dataIdx++
|
|
}
|
|
}
|
|
for i := begin; i < end; i++ {
|
|
if validData[i] {
|
|
v := make([]int8, dim)
|
|
for j := 0; j < dim; j++ {
|
|
v[j] = int8(data[dataIdx*dim+j])
|
|
}
|
|
vector = append(vector, v)
|
|
dataIdx++
|
|
} else {
|
|
vector = append(vector, nil)
|
|
}
|
|
}
|
|
col := NewColumnInt8Vector(fd.GetFieldName(), dim, vector)
|
|
col.withValidData(validData[begin:end])
|
|
col.nullable = true
|
|
col.sparseMode = true
|
|
return col, nil
|
|
}
|
|
|
|
if end < 0 {
|
|
end = len(data) / dim
|
|
}
|
|
vector := make([][]int8, 0, end-begin)
|
|
for i := begin; i < end; i++ {
|
|
v := make([]int8, dim)
|
|
for j := 0; j < dim; j++ {
|
|
v[j] = int8(data[i*dim+j])
|
|
}
|
|
vector = append(vector, v)
|
|
}
|
|
return NewColumnInt8Vector(fd.GetFieldName(), dim, vector), nil
|
|
|
|
default:
|
|
return nil, fmt.Errorf("unsupported data type %s", fd.GetType())
|
|
}
|
|
}
|
|
|
|
// getIntData get int32 slice from result field data
|
|
// also handles LongData bug (see also https://github.com/milvus-io/milvus/issues/23850)
|
|
func getIntData(fd *schemapb.FieldData) (*schemapb.ScalarField_IntData, bool) {
|
|
switch data := fd.GetScalars().GetData().(type) {
|
|
case *schemapb.ScalarField_IntData:
|
|
return data, true
|
|
case *schemapb.ScalarField_LongData:
|
|
// only alway empty LongData for backward compatibility
|
|
if len(data.LongData.GetData()) == 0 {
|
|
return &schemapb.ScalarField_IntData{
|
|
IntData: &schemapb.IntArray{},
|
|
}, true
|
|
}
|
|
return nil, false
|
|
default:
|
|
return nil, false
|
|
}
|
|
}
|