// Copyright (C) 2019-2020 Zilliz. All rights reserved. // // Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance // with the License. You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software distributed under the License // is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express // or implied. See the License for the specific language governing permissions and limitations under the License. #include #include #include #include #include #include #include "cache/CpuCacheMgr.h" #include "config/Config.h" #include "db/Constants.h" #include "db/DB.h" #include "db/DBFactory.h" #include "db/DBImpl.h" #include "db/IDGenerator.h" #include "db/meta/MetaConsts.h" #include "db/utils.h" #include "utils/CommonUtil.h" namespace { static const char* COLLECTION_NAME = "test_group"; static constexpr int64_t COLLECTION_DIM = 256; static constexpr int64_t VECTOR_COUNT = 5000; static constexpr int64_t INSERT_LOOP = 100; static constexpr int64_t SECONDS_EACH_HOUR = 3600; static constexpr int64_t DAY_SECONDS = 24 * 60 * 60; milvus::engine::meta::CollectionSchema BuildCollectionSchema() { milvus::engine::meta::CollectionSchema collection_info; collection_info.dimension_ = COLLECTION_DIM; collection_info.collection_id_ = COLLECTION_NAME; return collection_info; } void BuildVectors(uint64_t n, uint64_t batch_index, milvus::engine::VectorsData& vectors) { vectors.vector_count_ = n; vectors.float_data_.clear(); vectors.float_data_.resize(n * COLLECTION_DIM); float* data = vectors.float_data_.data(); for (uint64_t i = 0; i < n; i++) { for (int64_t j = 0; j < COLLECTION_DIM; j++) data[COLLECTION_DIM * i + j] = drand48(); data[COLLECTION_DIM * i] += i / 2000.; vectors.id_array_.push_back(n * batch_index + i); } } std::string CurrentTmDate(int64_t offset_day = 0) { time_t tt; time(&tt); tt = tt + 8 * SECONDS_EACH_HOUR; tt = tt + 24 * SECONDS_EACH_HOUR * offset_day; tm t; gmtime_r(&tt, &t); std::string str = std::to_string(t.tm_year + 1900) + "-" + std::to_string(t.tm_mon + 1) + "-" + std::to_string(t.tm_mday); return str; } void ConvertTimeRangeToDBDates(const std::string& start_value, const std::string& end_value, std::vector& dates) { dates.clear(); time_t tt_start, tt_end; tm tm_start, tm_end; if (!milvus::server::CommonUtil::TimeStrToTime(start_value, tt_start, tm_start)) { return; } if (!milvus::server::CommonUtil::TimeStrToTime(end_value, tt_end, tm_end)) { return; } int64_t days = (tt_end > tt_start) ? (tt_end - tt_start) / DAY_SECONDS : (tt_start - tt_end) / DAY_SECONDS; if (days == 0) { return; } for (int64_t i = 0; i < days; i++) { time_t tt_day = tt_start + DAY_SECONDS * i; tm tm_day; milvus::server::CommonUtil::ConvertTime(tt_day, tm_day); int64_t date = tm_day.tm_year * 10000 + tm_day.tm_mon * 100 + tm_day.tm_mday; // according to db logic dates.push_back(date); } } } // namespace TEST_F(DBTest, CONFIG_TEST) { { ASSERT_ANY_THROW(milvus::engine::ArchiveConf conf("wrong")); /* EXPECT_DEATH(engine::ArchiveConf conf("wrong"), ""); */ } { milvus::engine::ArchiveConf conf("delete"); ASSERT_EQ(conf.GetType(), "delete"); auto criterias = conf.GetCriterias(); ASSERT_EQ(criterias.size(), 0); } { milvus::engine::ArchiveConf conf("swap"); ASSERT_EQ(conf.GetType(), "swap"); auto criterias = conf.GetCriterias(); ASSERT_EQ(criterias.size(), 0); } { fiu_init(0); fiu_enable("ArchiveConf.ParseCritirias.OptionsParseCritiriasOutOfRange", 1, NULL, 0); ASSERT_ANY_THROW(milvus::engine::ArchiveConf conf("swap", "disk:")); fiu_disable("ArchiveConf.ParseCritirias.OptionsParseCritiriasOutOfRange"); } { fiu_enable("ArchiveConf.ParseCritirias.empty_tokens", 1, NULL, 0); milvus::engine::ArchiveConf conf("swap", ""); ASSERT_TRUE(conf.GetCriterias().empty()); fiu_disable("ArchiveConf.ParseCritirias.empty_tokens"); } { ASSERT_ANY_THROW(milvus::engine::ArchiveConf conf1("swap", "disk:")); ASSERT_ANY_THROW(milvus::engine::ArchiveConf conf2("swap", "disk:a")); milvus::engine::ArchiveConf conf("swap", "disk:1024"); auto criterias = conf.GetCriterias(); ASSERT_EQ(criterias.size(), 1); ASSERT_EQ(criterias["disk"], 1024); } { ASSERT_ANY_THROW(milvus::engine::ArchiveConf conf1("swap", "days:")); ASSERT_ANY_THROW(milvus::engine::ArchiveConf conf2("swap", "days:a")); milvus::engine::ArchiveConf conf("swap", "days:100"); auto criterias = conf.GetCriterias(); ASSERT_EQ(criterias.size(), 1); ASSERT_EQ(criterias["days"], 100); } { ASSERT_ANY_THROW(milvus::engine::ArchiveConf conf1("swap", "days:")); ASSERT_ANY_THROW(milvus::engine::ArchiveConf conf2("swap", "days:a")); milvus::engine::ArchiveConf conf("swap", "days:100;disk:200"); auto criterias = conf.GetCriterias(); ASSERT_EQ(criterias.size(), 2); ASSERT_EQ(criterias["days"], 100); ASSERT_EQ(criterias["disk"], 200); } } TEST_F(DBTest, DB_TEST) { milvus::engine::meta::CollectionSchema collection_info = BuildCollectionSchema(); auto stat = db_->CreateCollection(collection_info); milvus::engine::meta::CollectionSchema collection_info_get; collection_info_get.collection_id_ = COLLECTION_NAME; stat = db_->DescribeCollection(collection_info_get); ASSERT_TRUE(stat.ok()); ASSERT_EQ(collection_info_get.dimension_, COLLECTION_DIM); uint64_t qb = 5; milvus::engine::VectorsData qxb; BuildVectors(qb, 0, qxb); std::thread search([&]() { milvus::engine::ResultIds result_ids; milvus::engine::ResultDistances result_distances; int k = 10; std::this_thread::sleep_for(std::chrono::seconds(1)); INIT_TIMER; std::stringstream ss; uint64_t count = 0; uint64_t prev_count = 0; milvus::json json_params = {{"nprobe", 10}}; for (auto j = 0; j < 10; ++j) { ss.str(""); db_->Size(count); prev_count = count; if (count == 0) { continue; } START_TIMER; std::vector tags; stat = db_->Query(dummy_context_, COLLECTION_NAME, tags, k, json_params, qxb, result_ids, result_distances); ss << "Search " << j << " With Size " << count / milvus::engine::MB << " MB"; STOP_TIMER(ss.str()); ASSERT_TRUE(stat.ok()); ASSERT_EQ(result_ids.size(), qb * k); for (auto i = 0; i < qb; ++i) { ss.str(""); ss << "Result [" << i << "]:"; for (auto t = 0; t < k; t++) { ss << result_ids[i * k + t] << " "; } /* LOG(DEBUG) << ss.str(); */ } ASSERT_TRUE(count >= prev_count); std::this_thread::sleep_for(std::chrono::milliseconds(100)); } }); int loop = INSERT_LOOP; for (auto i = 0; i < loop; ++i) { if (i == 40) { db_->InsertVectors(COLLECTION_NAME, "", qxb); ASSERT_EQ(qxb.id_array_.size(), qb); } else { uint64_t nb = 50; milvus::engine::VectorsData xb; BuildVectors(nb, i, xb); db_->InsertVectors(COLLECTION_NAME, "", xb); ASSERT_EQ(xb.id_array_.size(), nb); } stat = db_->Flush(); ASSERT_TRUE(stat.ok()); std::this_thread::sleep_for(std::chrono::milliseconds(20)); } search.join(); uint64_t count; stat = db_->GetCollectionRowCount(COLLECTION_NAME, count); ASSERT_TRUE(stat.ok()); ASSERT_GT(count, 0); // test invalid build db { auto options = GetOptions(); options.meta_.backend_uri_ = "dummy"; ASSERT_ANY_THROW(BuildDB(options)); options.meta_.backend_uri_ = "mysql://root:123456@127.0.0.1:3306/test"; ASSERT_ANY_THROW(BuildDB(options)); options.meta_.backend_uri_ = "dummy://root:123456@127.0.0.1:3306/test"; ASSERT_ANY_THROW(BuildDB(options)); } } TEST_F(DBTest, SEARCH_TEST) { milvus::scheduler::OptimizerInst::GetInstance()->Init(); std::string config_path(CONFIG_PATH); config_path += CONFIG_FILE; milvus::server::Config& config = milvus::server::Config::GetInstance(); milvus::Status s = config.LoadConfigFile(config_path); milvus::engine::meta::CollectionSchema collection_info = BuildCollectionSchema(); auto stat = db_->CreateCollection(collection_info); milvus::engine::meta::CollectionSchema collection_info_get; collection_info_get.collection_id_ = COLLECTION_NAME; stat = db_->DescribeCollection(collection_info_get); ASSERT_TRUE(stat.ok()); ASSERT_EQ(collection_info_get.dimension_, COLLECTION_DIM); // prepare raw data size_t nb = VECTOR_COUNT; size_t nq = 10; size_t k = 5; milvus::engine::VectorsData xb, xq; xb.vector_count_ = nb; xb.float_data_.resize(nb * COLLECTION_DIM); xq.vector_count_ = nq; xq.float_data_.resize(nq * COLLECTION_DIM); xb.id_array_.resize(nb); std::random_device rd; std::mt19937 gen(rd()); std::uniform_real_distribution<> dis_xt(-1.0, 1.0); for (size_t i = 0; i < nb * COLLECTION_DIM; i++) { xb.float_data_[i] = dis_xt(gen); if (i < nb) { xb.id_array_[i] = i; } } for (size_t i = 0; i < nq * COLLECTION_DIM; i++) { xq.float_data_[i] = dis_xt(gen); } // result data // std::vector nns_gt(k*nq); std::vector nns(k * nq); // nns = nearst neg search // std::vector dis_gt(k*nq); std::vector dis(k * nq); // insert data stat = db_->InsertVectors(COLLECTION_NAME, "", xb); ASSERT_TRUE(stat.ok()); milvus::json json_params = {{"nprobe", 10}}; milvus::engine::CollectionIndex index; index.engine_type_ = (int)milvus::engine::EngineType::FAISS_IDMAP; index.extra_params_ = {{"nlist", 16384}}; // db_->CreateIndex(dummy_context_, COLLECTION_NAME, index); // wait until build index finish // // { // std::vector tags; // milvus::engine::ResultIds result_ids; // milvus::engine::ResultDistances result_distances; // stat = db_->Query(dummy_context_, COLLECTION_NAME, tags, k, json_params, xq, result_ids, result_distances); // ASSERT_TRUE(stat.ok()); // } // // index.engine_type_ = (int)milvus::engine::EngineType::FAISS_IVFFLAT; // index.extra_params_ = {{"nlist", 16384}}; // db_->CreateIndex(dummy_context_, COLLECTION_NAME, index); // wait until build index finish // // { // std::vector tags; // milvus::engine::ResultIds result_ids; // milvus::engine::ResultDistances result_distances; // stat = db_->Query(dummy_context_, COLLECTION_NAME, tags, k, json_params, xq, result_ids, result_distances); // ASSERT_TRUE(stat.ok()); // } index.engine_type_ = (int)milvus::engine::EngineType::FAISS_IVFSQ8; index.extra_params_ = {{"nlist", 16384}}; db_->CreateIndex(dummy_context_, COLLECTION_NAME, index); // wait until build index finish { std::vector tags; milvus::engine::ResultIds result_ids; milvus::engine::ResultDistances result_distances; stat = db_->Query(dummy_context_, COLLECTION_NAME, tags, k, json_params, xq, result_ids, result_distances); ASSERT_TRUE(stat.ok()); } #ifdef MILVUS_GPU_VERSION index.engine_type_ = (int)milvus::engine::EngineType::FAISS_IVFSQ8H; db_->CreateIndex(dummy_context_, COLLECTION_NAME, index); // wait until build index finish { std::vector tags; milvus::engine::ResultIds result_ids; milvus::engine::ResultDistances result_distances; stat = db_->Query(dummy_context_, COLLECTION_NAME, tags, k, json_params, xq, result_ids, result_distances); ASSERT_TRUE(stat.ok()); } #endif { // search by specify index file std::vector file_ids; // sometimes this case run fast to merge file and build index, old file will be deleted immediately, // so the QueryByFileID cannot get files to search // input 100 files ids to avoid random failure of this case for (int i = 0; i < 100; i++) { file_ids.push_back(std::to_string(i)); } milvus::engine::ResultIds result_ids; milvus::engine::ResultDistances result_distances; stat = db_->QueryByFileID(dummy_context_, file_ids, k, json_params, xq, result_ids, result_distances); ASSERT_TRUE(stat.ok()); } // TODO(zhiru): PQ build takes forever #if 0 { std::vector tags; milvus::engine::ResultIds result_ids; milvus::engine::ResultDistances result_distances; stat = db_->Query(dummy_context_, COLLECTION_NAME, tags, k, json_params, xq, result_ids, result_distances); ASSERT_TRUE(stat.ok()); stat = db_->Query(dummy_context_, COLLECTION_NAME, tags, k, json_params, xq, result_ids, result_distances); ASSERT_TRUE(stat.ok()); FIU_ENABLE_FIU("SqliteMetaImpl.FilesToSearch.throw_exception"); stat = db_->Query(dummy_context_, COLLECTION_NAME, tags, k, json_params, xq, result_ids, result_distances); ASSERT_FALSE(stat.ok()); fiu_disable("SqliteMetaImpl.FilesToSearch.throw_exception"); } #endif #ifdef MILVUS_GPU_VERSION // test FAISS_IVFSQ8H optimizer index.engine_type_ = (int)milvus::engine::EngineType::FAISS_IVFSQ8H; db_->CreateIndex(dummy_context_, COLLECTION_NAME, index); // wait until build index finish std::vector partition_tag; milvus::engine::ResultIds result_ids; milvus::engine::ResultDistances result_dists; { result_ids.clear(); result_dists.clear(); stat = db_->Query(dummy_context_, COLLECTION_NAME, partition_tag, k, json_params, xq, result_ids, result_dists); ASSERT_TRUE(stat.ok()); } { // search by specify index file std::vector file_ids; // sometimes this case run fast to merge file and build index, old file will be deleted immediately, // so the QueryByFileID cannot get files to search // input 100 files ids to avoid random failure of this case for (int i = 0; i < 100; i++) { file_ids.push_back(std::to_string(i)); } result_ids.clear(); result_dists.clear(); stat = db_->QueryByFileID(dummy_context_, file_ids, k, json_params, xq, result_ids, result_dists); ASSERT_TRUE(stat.ok()); } #endif } TEST_F(DBTest, PRELOAD_TEST) { fiu_init(0); milvus::engine::meta::CollectionSchema collection_info = BuildCollectionSchema(); auto stat = db_->CreateCollection(collection_info); milvus::engine::meta::CollectionSchema collection_info_get; collection_info_get.collection_id_ = COLLECTION_NAME; stat = db_->DescribeCollection(collection_info_get); ASSERT_TRUE(stat.ok()); ASSERT_EQ(collection_info_get.dimension_, COLLECTION_DIM); int loop = 5; for (auto i = 0; i < loop; ++i) { uint64_t nb = VECTOR_COUNT; milvus::engine::VectorsData xb; BuildVectors(nb, i, xb); db_->InsertVectors(COLLECTION_NAME, "", xb); ASSERT_EQ(xb.id_array_.size(), nb); } milvus::engine::CollectionIndex index; index.engine_type_ = (int)milvus::engine::EngineType::FAISS_IDMAP; db_->CreateIndex(dummy_context_, COLLECTION_NAME, index); // wait until build index finish int64_t prev_cache_usage = milvus::cache::CpuCacheMgr::GetInstance()->CacheUsage(); stat = db_->PreloadCollection(dummy_context_, COLLECTION_NAME); ASSERT_TRUE(stat.ok()); int64_t cur_cache_usage = milvus::cache::CpuCacheMgr::GetInstance()->CacheUsage(); ASSERT_TRUE(prev_cache_usage < cur_cache_usage); FIU_ENABLE_FIU("SqliteMetaImpl.FilesToSearch.throw_exception"); stat = db_->PreloadCollection(dummy_context_, COLLECTION_NAME); ASSERT_FALSE(stat.ok()); fiu_disable("SqliteMetaImpl.FilesToSearch.throw_exception"); // create a partition stat = db_->CreatePartition(COLLECTION_NAME, "part0", "0"); ASSERT_TRUE(stat.ok()); stat = db_->PreloadCollection(dummy_context_, COLLECTION_NAME); ASSERT_TRUE(stat.ok()); FIU_ENABLE_FIU("DBImpl.PreloadCollection.null_engine"); stat = db_->PreloadCollection(dummy_context_, COLLECTION_NAME); ASSERT_FALSE(stat.ok()); fiu_disable("DBImpl.PreloadCollection.null_engine"); FIU_ENABLE_FIU("DBImpl.PreloadCollection.exceed_cache"); stat = db_->PreloadCollection(dummy_context_, COLLECTION_NAME); ASSERT_FALSE(stat.ok()); fiu_disable("DBImpl.PreloadCollection.exceed_cache"); FIU_ENABLE_FIU("DBImpl.PreloadCollection.engine_throw_exception"); stat = db_->PreloadCollection(dummy_context_, COLLECTION_NAME); ASSERT_FALSE(stat.ok()); fiu_disable("DBImpl.PreloadCollection.engine_throw_exception"); } TEST_F(DBTest, SHUTDOWN_TEST) { db_->Stop(); milvus::engine::meta::CollectionSchema collection_info = BuildCollectionSchema(); auto stat = db_->CreateCollection(collection_info); ASSERT_FALSE(stat.ok()); stat = db_->DescribeCollection(collection_info); ASSERT_FALSE(stat.ok()); stat = db_->UpdateCollectionFlag(COLLECTION_NAME, 0); ASSERT_FALSE(stat.ok()); stat = db_->CreatePartition(COLLECTION_NAME, "part0", "0"); ASSERT_FALSE(stat.ok()); stat = db_->DropPartition("part0"); ASSERT_FALSE(stat.ok()); stat = db_->DropPartitionByTag(COLLECTION_NAME, "0"); ASSERT_FALSE(stat.ok()); std::vector partition_schema_array; stat = db_->ShowPartitions(COLLECTION_NAME, partition_schema_array); ASSERT_FALSE(stat.ok()); std::vector collection_infos; stat = db_->AllCollections(collection_infos); ASSERT_EQ(stat.code(), milvus::DB_ERROR); bool has_collection = false; stat = db_->HasCollection(collection_info.collection_id_, has_collection); ASSERT_FALSE(stat.ok()); milvus::engine::VectorsData xb; stat = db_->InsertVectors(collection_info.collection_id_, "", xb); ASSERT_FALSE(stat.ok()); stat = db_->Flush(); ASSERT_FALSE(stat.ok()); stat = db_->DeleteVector(collection_info.collection_id_, 0); ASSERT_FALSE(stat.ok()); milvus::engine::IDNumbers ids_to_delete{0}; stat = db_->DeleteVectors(collection_info.collection_id_, ids_to_delete); ASSERT_FALSE(stat.ok()); stat = db_->Compact(dummy_context_, collection_info.collection_id_); ASSERT_FALSE(stat.ok()); std::vector vectors; std::vector id_array = {0}; stat = db_->GetVectorsByID(collection_info, id_array, vectors); ASSERT_FALSE(stat.ok()); stat = db_->PreloadCollection(dummy_context_, collection_info.collection_id_); ASSERT_FALSE(stat.ok()); uint64_t row_count = 0; stat = db_->GetCollectionRowCount(collection_info.collection_id_, row_count); ASSERT_FALSE(stat.ok()); milvus::engine::CollectionIndex index; stat = db_->CreateIndex(dummy_context_, collection_info.collection_id_, index); ASSERT_FALSE(stat.ok()); stat = db_->DescribeIndex(collection_info.collection_id_, index); ASSERT_FALSE(stat.ok()); stat = db_->DropIndex(COLLECTION_NAME); ASSERT_FALSE(stat.ok()); std::vector tags; milvus::engine::ResultIds result_ids; milvus::engine::ResultDistances result_distances; milvus::json json_params = {{"nprobe", 1}}; stat = db_->Query(dummy_context_, collection_info.collection_id_, tags, 1, json_params, xb, result_ids, result_distances); ASSERT_FALSE(stat.ok()); std::vector file_ids; stat = db_->QueryByFileID(dummy_context_, file_ids, 1, json_params, xb, result_ids, result_distances); ASSERT_FALSE(stat.ok()); stat = db_->Query(dummy_context_, collection_info.collection_id_, tags, 1, json_params, milvus::engine::VectorsData(), result_ids, result_distances); ASSERT_FALSE(stat.ok()); stat = db_->DropCollection(collection_info.collection_id_); ASSERT_FALSE(stat.ok()); } TEST_F(DBTest, BACK_TIMER_THREAD_1) { fiu_init(0); milvus::engine::meta::CollectionSchema collection_info = BuildCollectionSchema(); milvus::Status stat; // test background timer thread { FIU_ENABLE_FIU("DBImpl.StartMetricTask.InvalidTotalCache"); FIU_ENABLE_FIU("SqliteMetaImpl.FilesToMerge.throw_exception"); stat = db_->CreateCollection(collection_info); ASSERT_TRUE(stat.ok()); // insert some vector to create some collection files int loop = 10; for (auto i = 0; i < loop; ++i) { int64_t nb = VECTOR_COUNT; milvus::engine::VectorsData xb; BuildVectors(nb, i, xb); db_->InsertVectors(COLLECTION_NAME, "", xb); ASSERT_EQ(xb.id_array_.size(), nb); } db_->Stop(); fiu_disable("DBImpl.StartMetricTask.InvalidTotalCache"); fiu_disable("SqliteMetaImpl.FilesToMerge.throw_exception"); } FIU_ENABLE_FIU("DBImpl.StartMetricTask.InvalidTotalCache"); db_->Start(); db_->Stop(); fiu_disable("DBImpl.StartMetricTask.InvalidTotalCache"); FIU_ENABLE_FIU("options_metric_enable"); db_->Start(); db_->Stop(); fiu_disable("options_metric_enable"); } TEST_F(DBTest, BACK_TIMER_THREAD_2) { fiu_init(0); milvus::Status stat; milvus::engine::meta::CollectionSchema collection_info = BuildCollectionSchema(); stat = db_->CreateCollection(collection_info); ASSERT_TRUE(stat.ok()); // insert some vector to create some collection files int loop = 10; for (auto i = 0; i < loop; ++i) { int64_t nb = VECTOR_COUNT; milvus::engine::VectorsData xb; BuildVectors(nb, i, xb); db_->InsertVectors(COLLECTION_NAME, "", xb); ASSERT_EQ(xb.id_array_.size(), nb); } FIU_ENABLE_FIU("SqliteMetaImpl.CreateCollectionFile.throw_exception"); db_->Stop(); fiu_disable("SqliteMetaImpl.CreateCollectionFile.throw_exception"); } TEST_F(DBTest, BACK_TIMER_THREAD_3) { fiu_init(0); milvus::Status stat; milvus::engine::meta::CollectionSchema collection_info = BuildCollectionSchema(); stat = db_->CreateCollection(collection_info); ASSERT_TRUE(stat.ok()); // insert some vector to create some collection files int loop = 10; for (auto i = 0; i < loop; ++i) { int64_t nb = VECTOR_COUNT; milvus::engine::VectorsData xb; BuildVectors(nb, i, xb); db_->InsertVectors(COLLECTION_NAME, "", xb); ASSERT_EQ(xb.id_array_.size(), nb); } FIU_ENABLE_FIU("DBImpl.MergeFiles.Serialize_ThrowException"); db_->Start(); db_->Stop(); fiu_disable("DBImpl.MergeFiles.Serialize_ThrowException"); } TEST_F(DBTest, BACK_TIMER_THREAD_4) { fiu_init(0); milvus::Status stat; milvus::engine::meta::CollectionSchema collection_info = BuildCollectionSchema(); stat = db_->CreateCollection(collection_info); ASSERT_TRUE(stat.ok()); // insert some vector to create some collection files int loop = 10; for (auto i = 0; i < loop; ++i) { int64_t nb = VECTOR_COUNT; milvus::engine::VectorsData xb; BuildVectors(nb, i, xb); db_->InsertVectors(COLLECTION_NAME, "", xb); ASSERT_EQ(xb.id_array_.size(), nb); } FIU_ENABLE_FIU("DBImpl.MergeFiles.Serialize_ErrorStatus"); db_->Start(); db_->Stop(); fiu_disable("DBImpl.MergeFiles.Serialize_ErrorStatus"); } TEST_F(DBTest, INDEX_TEST) { milvus::engine::meta::CollectionSchema collection_info = BuildCollectionSchema(); auto stat = db_->CreateCollection(collection_info); uint64_t nb = VECTOR_COUNT; milvus::engine::VectorsData xb; BuildVectors(nb, 0, xb); db_->InsertVectors(COLLECTION_NAME, "", xb); ASSERT_EQ(xb.id_array_.size(), nb); milvus::engine::CollectionIndex index; index.engine_type_ = (int)milvus::engine::EngineType::FAISS_IVFSQ8; index.metric_type_ = (int)milvus::engine::MetricType::IP; stat = db_->CreateIndex(dummy_context_, collection_info.collection_id_, index); ASSERT_TRUE(stat.ok()); index.engine_type_ = (int)milvus::engine::EngineType::FAISS_IVFFLAT; stat = db_->CreateIndex(dummy_context_, collection_info.collection_id_, index); ASSERT_TRUE(stat.ok()); fiu_init(0); FIU_ENABLE_FIU("SqliteMetaImpl.DescribeCollectionIndex.throw_exception"); stat = db_->CreateIndex(dummy_context_, collection_info.collection_id_, index); ASSERT_FALSE(stat.ok()); fiu_disable("SqliteMetaImpl.DescribeCollectionIndex.throw_exception"); index.engine_type_ = (int)milvus::engine::EngineType::FAISS_PQ; FIU_ENABLE_FIU("DBImpl.UpdateCollectionIndexRecursively.fail_update_collection_index"); stat = db_->CreateIndex(dummy_context_, collection_info.collection_id_, index); ASSERT_FALSE(stat.ok()); fiu_disable("DBImpl.UpdateCollectionIndexRecursively.fail_update_collection_index"); #ifdef MILVUS_GPU_VERSION index.engine_type_ = (int)milvus::engine::EngineType::FAISS_IVFSQ8H; stat = db_->CreateIndex(dummy_context_, collection_info.collection_id_, index); ASSERT_TRUE(stat.ok()); #endif milvus::engine::CollectionIndex index_out; stat = db_->DescribeIndex(collection_info.collection_id_, index_out); ASSERT_TRUE(stat.ok()); ASSERT_EQ(index.engine_type_, index_out.engine_type_); ASSERT_EQ(index.extra_params_, index_out.extra_params_); ASSERT_EQ(collection_info.metric_type_, index_out.metric_type_); stat = db_->DropIndex(collection_info.collection_id_); ASSERT_TRUE(stat.ok()); } TEST_F(DBTest, PARTITION_TEST) { milvus::engine::meta::CollectionSchema collection_info = BuildCollectionSchema(); auto stat = db_->CreateCollection(collection_info); ASSERT_TRUE(stat.ok()); // create partition and insert data const int64_t PARTITION_COUNT = 5; const int64_t INSERT_BATCH = 2000; std::string collection_name = COLLECTION_NAME; for (int64_t i = 0; i < PARTITION_COUNT; i++) { std::string partition_tag = std::to_string(i); std::string partition_name = collection_name + "_" + partition_tag; stat = db_->CreatePartition(collection_name, partition_name, partition_tag); ASSERT_TRUE(stat.ok()); // not allow nested partition stat = db_->CreatePartition(partition_name, "dumy", "dummy"); ASSERT_FALSE(stat.ok()); // not allow duplicated partition stat = db_->CreatePartition(collection_name, partition_name, partition_tag); ASSERT_FALSE(stat.ok()); milvus::engine::VectorsData xb; BuildVectors(INSERT_BATCH, i, xb); milvus::engine::IDNumbers vector_ids; vector_ids.resize(INSERT_BATCH); for (int64_t k = 0; k < INSERT_BATCH; k++) { vector_ids[k] = i * INSERT_BATCH + k; } db_->InsertVectors(collection_name, partition_tag, xb); ASSERT_EQ(vector_ids.size(), INSERT_BATCH); // insert data into not existed partition stat = db_->InsertVectors(COLLECTION_NAME, "notexist", xb); ASSERT_FALSE(stat.ok()); } // duplicated partition is not allowed stat = db_->CreatePartition(collection_name, "", "0"); ASSERT_FALSE(stat.ok()); std::vector partition_schema_array; stat = db_->ShowPartitions(collection_name, partition_schema_array); ASSERT_TRUE(stat.ok()); ASSERT_EQ(partition_schema_array.size(), PARTITION_COUNT); for (int64_t i = 0; i < PARTITION_COUNT; i++) { ASSERT_EQ(partition_schema_array[i].collection_id_, collection_name + "_" + std::to_string(i)); } // check collection existence std::string special_part = "special"; stat = db_->CreatePartition(collection_name, special_part, special_part); ASSERT_TRUE(stat.ok()); bool has_collection = false; stat = db_->HasNativeCollection(special_part, has_collection); ASSERT_FALSE(has_collection); stat = db_->HasCollection(special_part, has_collection); ASSERT_TRUE(has_collection); { // build index milvus::engine::CollectionIndex index; index.engine_type_ = (int)milvus::engine::EngineType::FAISS_IVFFLAT; index.metric_type_ = (int)milvus::engine::MetricType::L2; stat = db_->CreateIndex(dummy_context_, collection_info.collection_id_, index); ASSERT_TRUE(stat.ok()); fiu_init(0); FIU_ENABLE_FIU("DBImpl.WaitCollectionIndexRecursively.fail_build_collection_Index_for_partition"); stat = db_->CreateIndex(dummy_context_, collection_info.collection_id_, index); ASSERT_FALSE(stat.ok()); fiu_disable("DBImpl.WaitCollectionIndexRecursively.fail_build_collection_Index_for_partition"); FIU_ENABLE_FIU("DBImpl.WaitCollectionIndexRecursively.not_empty_err_msg"); stat = db_->CreateIndex(dummy_context_, collection_info.collection_id_, index); ASSERT_FALSE(stat.ok()); fiu_disable("DBImpl.WaitCollectionIndexRecursively.not_empty_err_msg"); uint64_t row_count = 0; stat = db_->GetCollectionRowCount(COLLECTION_NAME, row_count); ASSERT_TRUE(stat.ok()); ASSERT_EQ(row_count, INSERT_BATCH * PARTITION_COUNT); FIU_ENABLE_FIU("SqliteMetaImpl.Count.throw_exception"); stat = db_->GetCollectionRowCount(COLLECTION_NAME, row_count); ASSERT_FALSE(stat.ok()); fiu_disable("SqliteMetaImpl.Count.throw_exception"); FIU_ENABLE_FIU("DBImpl.GetCollectionRowCountRecursively.fail_get_collection_rowcount_for_partition"); stat = db_->GetCollectionRowCount(COLLECTION_NAME, row_count); ASSERT_FALSE(stat.ok()); fiu_disable("DBImpl.GetCollectionRowCountRecursively.fail_get_collection_rowcount_for_partition"); } { // search const int64_t nq = 5; const int64_t topk = 10; const int64_t nprobe = 10; milvus::engine::VectorsData xq; BuildVectors(nq, 0, xq); // specify partition tags std::vector tags = {"0", std::to_string(PARTITION_COUNT - 1)}; milvus::engine::ResultIds result_ids; milvus::engine::ResultDistances result_distances; milvus::json json_params = {{"nprobe", nprobe}}; stat = db_->Query(dummy_context_, COLLECTION_NAME, tags, topk, json_params, xq, result_ids, result_distances); ASSERT_TRUE(stat.ok()); ASSERT_EQ(result_ids.size() / topk, nq); // search in whole collection tags.clear(); result_ids.clear(); result_distances.clear(); stat = db_->Query(dummy_context_, COLLECTION_NAME, tags, topk, json_params, xq, result_ids, result_distances); ASSERT_TRUE(stat.ok()); ASSERT_EQ(result_ids.size() / topk, nq); // search in all partitions(tag regex match) tags.push_back("\\d"); result_ids.clear(); result_distances.clear(); stat = db_->Query(dummy_context_, COLLECTION_NAME, tags, topk, json_params, xq, result_ids, result_distances); ASSERT_TRUE(stat.ok()); ASSERT_EQ(result_ids.size() / topk, nq); } stat = db_->DropPartition(collection_name + "_0"); ASSERT_TRUE(stat.ok()); stat = db_->DropPartitionByTag(collection_name, "1"); ASSERT_TRUE(stat.ok()); FIU_ENABLE_FIU("DBImpl.DropCollectionIndexRecursively.fail_drop_collection_Index_for_partition"); stat = db_->DropIndex(collection_info.collection_id_); ASSERT_FALSE(stat.ok()); fiu_disable("DBImpl.DropCollectionIndexRecursively.fail_drop_collection_Index_for_partition"); FIU_ENABLE_FIU("DBImpl.DropCollectionIndexRecursively.fail_drop_collection_Index_for_partition"); stat = db_->DropIndex(collection_info.collection_id_); ASSERT_FALSE(stat.ok()); fiu_disable("DBImpl.DropCollectionIndexRecursively.fail_drop_collection_Index_for_partition"); stat = db_->DropIndex(collection_name); ASSERT_TRUE(stat.ok()); stat = db_->DropCollection(collection_name); ASSERT_TRUE(stat.ok()); } TEST_F(DBTest2, ARHIVE_DISK_CHECK) { milvus::engine::meta::CollectionSchema collection_info = BuildCollectionSchema(); auto stat = db_->CreateCollection(collection_info); std::vector collection_schema_array; stat = db_->AllCollections(collection_schema_array); ASSERT_TRUE(stat.ok()); bool bfound = false; for (auto& schema : collection_schema_array) { if (schema.collection_id_ == COLLECTION_NAME) { bfound = true; break; } } ASSERT_TRUE(bfound); milvus::engine::meta::CollectionSchema collection_info_get; collection_info_get.collection_id_ = COLLECTION_NAME; stat = db_->DescribeCollection(collection_info_get); ASSERT_TRUE(stat.ok()); ASSERT_EQ(collection_info_get.dimension_, COLLECTION_DIM); uint64_t size; db_->Size(size); int loop = INSERT_LOOP; for (auto i = 0; i < loop; ++i) { uint64_t nb = 10; milvus::engine::VectorsData xb; BuildVectors(nb, i, xb); db_->InsertVectors(COLLECTION_NAME, "", xb); } db_->Flush(); db_->Size(size); LOG(DEBUG) << "size=" << size; ASSERT_LE(size, 1 * milvus::engine::GB); } TEST_F(DBTest2, DELETE_TEST) { milvus::engine::meta::CollectionSchema collection_info = BuildCollectionSchema(); auto stat = db_->CreateCollection(collection_info); milvus::engine::meta::CollectionSchema collection_info_get; collection_info_get.collection_id_ = COLLECTION_NAME; stat = db_->DescribeCollection(collection_info_get); ASSERT_TRUE(stat.ok()); bool has_collection = false; db_->HasCollection(COLLECTION_NAME, has_collection); ASSERT_TRUE(has_collection); uint64_t size; db_->Size(size); uint64_t nb = VECTOR_COUNT; milvus::engine::VectorsData xb; BuildVectors(nb, 0, xb); milvus::engine::IDNumbers vector_ids; stat = db_->InsertVectors(COLLECTION_NAME, "", xb); milvus::engine::CollectionIndex index; stat = db_->CreateIndex(dummy_context_, COLLECTION_NAME, index); // create partition, drop collection will drop partition recursively stat = db_->CreatePartition(COLLECTION_NAME, "part0", "0"); ASSERT_TRUE(stat.ok()); // fail drop collection fiu_init(0); stat = db_->DropCollection(COLLECTION_NAME); ASSERT_TRUE(stat.ok()); db_->HasCollection(COLLECTION_NAME, has_collection); ASSERT_FALSE(has_collection); } TEST_F(DBTest2, SHOW_COLLECTION_INFO_TEST) { std::string collection_name = COLLECTION_NAME; milvus::engine::meta::CollectionSchema collection_schema = BuildCollectionSchema(); auto stat = db_->CreateCollection(collection_schema); uint64_t nb = VECTOR_COUNT; milvus::engine::VectorsData xb; BuildVectors(nb, 0, xb); milvus::engine::IDNumbers vector_ids; stat = db_->InsertVectors(collection_name, "", xb); // create partition and insert data const int64_t PARTITION_COUNT = 2; const int64_t INSERT_BATCH = 2000; for (int64_t i = 0; i < PARTITION_COUNT; i++) { std::string partition_tag = std::to_string(i); std::string partition_name = collection_name + "_" + partition_tag; stat = db_->CreatePartition(collection_name, partition_name, partition_tag); ASSERT_TRUE(stat.ok()); milvus::engine::VectorsData xb; BuildVectors(INSERT_BATCH, i, xb); db_->InsertVectors(collection_name, partition_tag, xb); } stat = db_->Flush(); ASSERT_TRUE(stat.ok()); { std::string collection_info; stat = db_->GetCollectionInfo(collection_name, collection_info); ASSERT_TRUE(stat.ok()); int64_t row_count = 0; nlohmann::json json_info = nlohmann::json::parse(collection_info); // for (auto& part : collection_info.partitions_stat_) { // row_count = 0; // for (auto& stat : part.segments_stat_) { // row_count += stat.row_count_; // ASSERT_EQ(stat.index_name_, "IDMAP"); // ASSERT_GT(stat.data_size_, 0); // } // if (part.tag_ == milvus::engine::DEFAULT_PARTITON_TAG) { // ASSERT_EQ(row_count, VECTOR_COUNT); // } else { // ASSERT_EQ(row_count, INSERT_BATCH); // } // } } } TEST_F(DBTestWAL, DB_INSERT_TEST) { milvus::engine::meta::CollectionSchema collection_info = BuildCollectionSchema(); auto stat = db_->CreateCollection(collection_info); ASSERT_TRUE(stat.ok()); uint64_t qb = 100; milvus::engine::VectorsData qxb; BuildVectors(qb, 0, qxb); std::string partition_name = "part_name"; std::string partition_tag = "part_tag"; stat = db_->CreatePartition(collection_info.collection_id_, partition_name, partition_tag); ASSERT_TRUE(stat.ok()); stat = db_->InsertVectors(collection_info.collection_id_, partition_tag, qxb); ASSERT_TRUE(stat.ok()); stat = db_->InsertVectors(collection_info.collection_id_, "", qxb); ASSERT_TRUE(stat.ok()); stat = db_->InsertVectors(collection_info.collection_id_, "not exist", qxb); ASSERT_FALSE(stat.ok()); db_->Flush(collection_info.collection_id_); stat = db_->DropCollection(collection_info.collection_id_); ASSERT_TRUE(stat.ok()); } TEST_F(DBTestWAL, DB_STOP_TEST) { milvus::engine::meta::CollectionSchema collection_info = BuildCollectionSchema(); auto stat = db_->CreateCollection(collection_info); ASSERT_TRUE(stat.ok()); uint64_t qb = 100; for (int i = 0; i < 5; i++) { milvus::engine::VectorsData qxb; BuildVectors(qb, i, qxb); stat = db_->InsertVectors(collection_info.collection_id_, "", qxb); ASSERT_TRUE(stat.ok()); } db_->Stop(); db_->Start(); const int64_t topk = 10; const int64_t nprobe = 10; milvus::json json_params = {{"nprobe", nprobe}}; milvus::engine::ResultIds result_ids; milvus::engine::ResultDistances result_distances; milvus::engine::VectorsData qxb; BuildVectors(qb, 0, qxb); stat = db_->Query(dummy_context_, collection_info.collection_id_, {}, topk, json_params, qxb, result_ids, result_distances); ASSERT_TRUE(stat.ok()); ASSERT_EQ(result_ids.size() / topk, qb); stat = db_->DropCollection(collection_info.collection_id_); ASSERT_TRUE(stat.ok()); } TEST_F(DBTestWALRecovery, RECOVERY_WITH_NO_ERROR) { milvus::engine::meta::CollectionSchema collection_info = BuildCollectionSchema(); auto stat = db_->CreateCollection(collection_info); ASSERT_TRUE(stat.ok()); uint64_t qb = 100; for (int i = 0; i < 5; i++) { milvus::engine::VectorsData qxb; BuildVectors(qb, i, qxb); stat = db_->InsertVectors(collection_info.collection_id_, "", qxb); ASSERT_TRUE(stat.ok()); } const int64_t topk = 10; const int64_t nprobe = 10; milvus::json json_params = {{"nprobe", nprobe}}; milvus::engine::ResultIds result_ids; milvus::engine::ResultDistances result_distances; milvus::engine::VectorsData qxb; BuildVectors(qb, 0, qxb); fiu_init(0); fiu_enable("DBImpl.ExexWalRecord.return", 1, nullptr, 0); db_ = nullptr; // don't use FreeDB(), this case needs keep the meta fiu_disable("DBImpl.ExexWalRecord.return"); auto options = GetOptions(); BuildDB(options); result_ids.clear(); result_distances.clear(); stat = db_->Query(dummy_context_, collection_info.collection_id_, {}, topk, json_params, qxb, result_ids, result_distances); ASSERT_TRUE(stat.ok()); ASSERT_EQ(result_ids.size(), 0); db_->Flush(); result_ids.clear(); result_distances.clear(); stat = db_->Query(dummy_context_, collection_info.collection_id_, {}, topk, json_params, qxb, result_ids, result_distances); ASSERT_TRUE(stat.ok()); ASSERT_EQ(result_ids.size() / topk, qb); } TEST_F(DBTestWALRecovery_Error, RECOVERY_WITH_INVALID_LOG_FILE) { milvus::engine::meta::CollectionSchema collection_info = BuildCollectionSchema(); auto stat = db_->CreateCollection(collection_info); ASSERT_TRUE(stat.ok()); uint64_t qb = 100; milvus::engine::VectorsData qxb; BuildVectors(qb, 0, qxb); stat = db_->InsertVectors(collection_info.collection_id_, "", qxb); ASSERT_TRUE(stat.ok()); fiu_init(0); fiu_enable("DBImpl.ExexWalRecord.return", 1, nullptr, 0); FreeDB(); fiu_disable("DBImpl.ExexWalRecord.return"); auto options = GetOptions(); // delete wal log file so that recovery will failed when start db next time. boost::filesystem::remove(options.mxlog_path_ + "0.wal"); ASSERT_ANY_THROW(BuildDB(options)); } TEST_F(DBTest2, FLUSH_NON_EXISTING_COLLECTION) { auto status = db_->Flush("non_existing"); ASSERT_FALSE(status.ok()); } TEST_F(DBTest2, GET_VECTOR_NON_EXISTING_COLLECTION) { std::vector vectors; std::vector id_array = {0}; milvus::engine::meta::CollectionSchema collection_info; collection_info.collection_id_ = "non_existing"; auto status = db_->GetVectorsByID(collection_info, id_array, vectors); ASSERT_FALSE(status.ok()); } TEST_F(DBTest2, GET_VECTOR_BY_ID_TEST) { milvus::engine::meta::CollectionSchema collection_info = BuildCollectionSchema(); auto stat = db_->CreateCollection(collection_info); ASSERT_TRUE(stat.ok()); uint64_t qb = 1000; milvus::engine::VectorsData qxb; BuildVectors(qb, 0, qxb); std::string partition_name = "part_name"; std::string partition_tag = "part_tag"; stat = db_->CreatePartition(collection_info.collection_id_, partition_name, partition_tag); ASSERT_TRUE(stat.ok()); std::vector vectors; std::vector empty_array; stat = db_->GetVectorsByID(collection_info, empty_array, vectors); ASSERT_FALSE(stat.ok()); stat = db_->InsertVectors(collection_info.collection_id_, partition_tag, qxb); ASSERT_TRUE(stat.ok()); db_->Flush(collection_info.collection_id_); stat = db_->GetVectorsByID(collection_info, qxb.id_array_, vectors); ASSERT_TRUE(stat.ok()); ASSERT_EQ(vectors.size(), qxb.id_array_.size()); ASSERT_EQ(vectors[0].float_data_.size(), COLLECTION_DIM); for (int64_t i = 0; i < COLLECTION_DIM; i++) { ASSERT_FLOAT_EQ(vectors[0].float_data_[i], qxb.float_data_[i]); } std::vector invalid_array = {-1, -1}; stat = db_->GetVectorsByID(collection_info, empty_array, vectors); ASSERT_TRUE(stat.ok()); for (auto& vector : vectors) { ASSERT_EQ(vector.vector_count_, 0); ASSERT_TRUE(vector.float_data_.empty()); ASSERT_TRUE(vector.binary_data_.empty()); } } TEST_F(DBTest2, GET_VECTOR_BY_ID_INVALID_TEST) { fiu_init(0); milvus::engine::meta::CollectionSchema collection_info = BuildCollectionSchema(); auto stat = db_->CreateCollection(collection_info); ASSERT_TRUE(stat.ok()); uint64_t qb = 1000; milvus::engine::VectorsData qxb; BuildVectors(qb, 0, qxb); std::string partition_name = "part_name"; std::string partition_tag = "part_tag"; stat = db_->CreatePartition(collection_info.collection_id_, partition_name, partition_tag); ASSERT_TRUE(stat.ok()); std::vector vectors; std::vector empty_array; stat = db_->GetVectorsByID(collection_info, empty_array, vectors); ASSERT_FALSE(stat.ok()); stat = db_->InsertVectors(collection_info.collection_id_, partition_tag, qxb); ASSERT_TRUE(stat.ok()); db_->Flush(collection_info.collection_id_); fiu_enable("bloom_filter_nullptr", 1, NULL, 0); stat = db_->GetVectorsByID(collection_info, qxb.id_array_, vectors); ASSERT_FALSE(stat.ok()); fiu_disable("bloom_filter_nullptr"); } TEST_F(DBTest2, GET_VECTOR_IDS_TEST) { milvus::engine::meta::CollectionSchema collection_schema = BuildCollectionSchema(); auto stat = db_->CreateCollection(collection_schema); ASSERT_TRUE(stat.ok()); uint64_t BATCH_COUNT = 1000; milvus::engine::VectorsData vector_1; BuildVectors(BATCH_COUNT, 0, vector_1); stat = db_->InsertVectors(COLLECTION_NAME, "", vector_1); ASSERT_TRUE(stat.ok()); std::string partition_tag = "part_tag"; stat = db_->CreatePartition(COLLECTION_NAME, "", partition_tag); ASSERT_TRUE(stat.ok()); milvus::engine::VectorsData vector_2; BuildVectors(BATCH_COUNT, 1, vector_2); stat = db_->InsertVectors(COLLECTION_NAME, partition_tag, vector_2); ASSERT_TRUE(stat.ok()); db_->Flush(); std::string collection_info; stat = db_->GetCollectionInfo(COLLECTION_NAME, collection_info); ASSERT_TRUE(stat.ok()); ASSERT_FALSE(collection_info.empty()); auto json = nlohmann::json::parse(collection_info); std::string default_segment = json["partitions"].at(0)["segments"].at(0)["name"]; std::string partition_segment = json["partitions"].at(1)["segments"].at(0)["name"]; milvus::engine::IDNumbers vector_ids; stat = db_->GetVectorIDs(COLLECTION_NAME, default_segment, vector_ids); ASSERT_TRUE(stat.ok()); ASSERT_EQ(vector_ids.size(), BATCH_COUNT); stat = db_->GetVectorIDs(COLLECTION_NAME, partition_segment, vector_ids); ASSERT_TRUE(stat.ok()); ASSERT_EQ(vector_ids.size(), BATCH_COUNT); milvus::engine::IDNumbers ids_to_delete{0, 100, 999, 1000, 1500, 1888, 1999}; stat = db_->DeleteVectors(COLLECTION_NAME, ids_to_delete); ASSERT_TRUE(stat.ok()); db_->Flush(); stat = db_->GetVectorIDs(COLLECTION_NAME, default_segment, vector_ids); ASSERT_TRUE(stat.ok()); ASSERT_EQ(vector_ids.size(), BATCH_COUNT - 3); stat = db_->GetVectorIDs(COLLECTION_NAME, partition_segment, vector_ids); ASSERT_TRUE(stat.ok()); // ASSERT_EQ(vector_ids.size(), BATCH_COUNT - 4); } TEST_F(DBTest2, INSERT_DUPLICATE_ID) { auto options = GetOptions(); options.wal_enable_ = false; BuildDB(options); milvus::engine::meta::CollectionSchema collection_schema = BuildCollectionSchema(); auto stat = db_->CreateCollection(collection_schema); ASSERT_TRUE(stat.ok()); uint64_t size = 20; milvus::engine::VectorsData vector; BuildVectors(size, 0, vector); vector.id_array_.clear(); for (int i = 0; i < size; ++i) { vector.id_array_.emplace_back(0); } stat = db_->InsertVectors(COLLECTION_NAME, "", vector); ASSERT_TRUE(stat.ok()); stat = db_->Flush(COLLECTION_NAME); ASSERT_TRUE(stat.ok()); } /* TEST_F(DBTest2, SEARCH_WITH_DIFFERENT_INDEX) { milvus::engine::meta::CollectionSchema collection_info = BuildCollectionSchema(); // collection_info.index_file_size_ = 1 * milvus::engine::M; auto stat = db_->CreateCollection(collection_info); int loop = 10; uint64_t nb = 100000; for (auto i = 0; i < loop; ++i) { milvus::engine::VectorsData xb; BuildVectors(nb, i, xb); db_->InsertVectors(COLLECTION_NAME, "", xb); stat = db_->Flush(); ASSERT_TRUE(stat.ok()); } std::random_device rd; std::mt19937 gen(rd()); std::uniform_int_distribution dis(0, nb * loop - 1); int64_t num_query = 10; std::vector ids_to_search; for (int64_t i = 0; i < num_query; ++i) { int64_t index = dis(gen); ids_to_search.emplace_back(index); } milvus::engine::CollectionIndex index; // index.metric_type_ = (int)milvus::engine::MetricType::IP; index.engine_type_ = (int)milvus::engine::EngineType::FAISS_IVFFLAT; stat = db_->CreateIndex(dummy_context_, collection_info.collection_id_, index); ASSERT_TRUE(stat.ok()); stat = db_->PreloadCollection(dummy_context_, collection_info.collection_id_); ASSERT_TRUE(stat.ok()); int topk = 10, nprobe = 10; milvus::json json_params = {{"nprobe", nprobe}}; for (auto id : ids_to_search) { // std::cout << "xxxxxxxxxxxxxxxxxxxx " << i << std::endl; std::vector tags; milvus::engine::ResultIds result_ids; milvus::engine::ResultDistances result_distances; stat = db_->QueryByID(dummy_context_, collection_info.collection_id_, tags, topk, json_params, id, result_ids, result_distances); ASSERT_TRUE(stat.ok()); ASSERT_EQ(result_ids[0], id); ASSERT_LT(result_distances[0], 1e-4); } db_->DropIndex(collection_info.collection_id_); index.engine_type_ = (int)milvus::engine::EngineType::FAISS_IVFSQ8; stat = db_->CreateIndex(dummy_context_, collection_info.collection_id_, index); ASSERT_TRUE(stat.ok()); stat = db_->PreloadCollection(dummy_context_, collection_info.collection_id_); ASSERT_TRUE(stat.ok()); for (auto id : ids_to_search) { // std::cout << "xxxxxxxxxxxxxxxxxxxx " << i << std::endl; std::vector tags; milvus::engine::ResultIds result_ids; milvus::engine::ResultDistances result_distances; stat = db_->QueryByID(dummy_context_, collection_info.collection_id_, tags, topk, json_params, id, result_ids, result_distances); ASSERT_TRUE(stat.ok()); ASSERT_EQ(result_ids[0], id); ASSERT_LT(result_distances[0], 1e-4); } } */