Loading cpp/src/db/meta/MetaConsts.h +0 −5 Original line number Diff line number Diff line Loading @@ -21,11 +21,6 @@ namespace milvus { namespace engine { namespace meta { const size_t K = 1024UL; const size_t M = K * K; const size_t G = K * M; const size_t T = K * G; const size_t S_PS = 1UL; const size_t MS_PS = 1000 * S_PS; const size_t US_PS = 1000 * MS_PS; Loading cpp/src/scheduler/JobMgr.cpp +4 −6 Original line number Diff line number Diff line Loading @@ -66,16 +66,14 @@ JobMgr::worker_function() { } auto tasks = build_task(job); auto disk_list = res_mgr_->GetDiskResources(); if (!disk_list.empty()) { if (auto disk = disk_list[0].lock()) { // disk resources NEVER be empty. if (auto disk = res_mgr_->GetDiskResources()[0].lock()) { for (auto& task : tasks) { disk->task_table().Put(task); } } } } } std::vector<TaskPtr> JobMgr::build_task(const JobPtr& job) { Loading cpp/src/scheduler/ResourceMgr.cpp +15 −9 Original line number Diff line number Diff line Loading @@ -79,9 +79,7 @@ ResourceMgr::Add(ResourcePtr&& resource) { gpu_resources_.emplace_back(ResourceWPtr(resource)); break; } default: { break; } default: { break; } } resources_.emplace_back(resource); Loading @@ -104,6 +102,10 @@ ResourceMgr::Connect(const std::string& name1, const std::string& name2, Connect void ResourceMgr::Clear() { std::lock_guard<std::mutex> lck(resources_mutex_); if (running_) { ENGINE_LOG_ERROR << "ResourceMgr is running, cannot clear."; return; } disk_resources_.clear(); cpu_resources_.clear(); gpu_resources_.clear(); Loading Loading @@ -196,13 +198,19 @@ bool ResourceMgr::check_resource_valid() { { // TODO: check one disk-resource, one cpu-resource, zero or more gpu-resource; if (GetDiskResources().size() != 1) return false; if (GetCpuResources().size() != 1) return false; if (GetDiskResources().size() != 1) { return false; } if (GetCpuResources().size() != 1) { return false; } } { // TODO: one compute-resource at least; if (GetNumOfComputeResource() < 1) return false; if (GetNumOfComputeResource() < 1) { return false; } } { Loading Loading @@ -233,9 +241,7 @@ void ResourceMgr::event_process() { while (running_) { std::unique_lock<std::mutex> lock(event_mutex_); event_cv_.wait(lock, [this] { return !queue_.empty(); }); event_cv_.wait(lock, [this] { return !queue_.empty(); }); auto event = queue_.front(); queue_.pop(); Loading cpp/src/scheduler/task/SearchTask.cpp +5 −8 Original line number Diff line number Diff line Loading @@ -22,6 +22,7 @@ #include "utils/Log.h" #include "utils/TimeRecorder.h" #include <algorithm> #include <string> #include <thread> #include <utility> Loading @@ -34,6 +35,7 @@ static constexpr size_t PARALLEL_REDUCE_BATCH = 1000; std::mutex XSearchTask::merge_mutex_; // TODO(wxyu): remove unused code // bool // NeedParallelReduce(uint64_t nq, uint64_t topk) { // server::ServerConfig &config = server::ServerConfig::GetInstance(); Loading Loading @@ -229,13 +231,8 @@ XSearchTask::Execute() { } Status XSearchTask::TopkResult(const std::vector<long> &input_ids, const std::vector<float> &input_distance, uint64_t input_k, uint64_t nq, uint64_t topk, bool ascending, scheduler::ResultSet &result) { XSearchTask::TopkResult(const std::vector<int64_t>& input_ids, const std::vector<float>& input_distance, uint64_t input_k, uint64_t nq, uint64_t topk, bool ascending, scheduler::ResultSet& result) { scheduler::ResultSet result_buf; if (result.empty()) { Loading cpp/src/scheduler/task/SearchTask.h +2 −7 Original line number Diff line number Diff line Loading @@ -39,13 +39,8 @@ class XSearchTask : public Task { public: static Status TopkResult(const std::vector<long> &input_ids, const std::vector<float> &input_distance, uint64_t input_k, uint64_t nq, uint64_t topk, bool ascending, scheduler::ResultSet &result); TopkResult(const std::vector<int64_t>& input_ids, const std::vector<float>& input_distance, uint64_t input_k, uint64_t nq, uint64_t topk, bool ascending, scheduler::ResultSet& result); public: TableFileSchemaPtr file_; Loading Loading
cpp/src/db/meta/MetaConsts.h +0 −5 Original line number Diff line number Diff line Loading @@ -21,11 +21,6 @@ namespace milvus { namespace engine { namespace meta { const size_t K = 1024UL; const size_t M = K * K; const size_t G = K * M; const size_t T = K * G; const size_t S_PS = 1UL; const size_t MS_PS = 1000 * S_PS; const size_t US_PS = 1000 * MS_PS; Loading
cpp/src/scheduler/JobMgr.cpp +4 −6 Original line number Diff line number Diff line Loading @@ -66,16 +66,14 @@ JobMgr::worker_function() { } auto tasks = build_task(job); auto disk_list = res_mgr_->GetDiskResources(); if (!disk_list.empty()) { if (auto disk = disk_list[0].lock()) { // disk resources NEVER be empty. if (auto disk = res_mgr_->GetDiskResources()[0].lock()) { for (auto& task : tasks) { disk->task_table().Put(task); } } } } } std::vector<TaskPtr> JobMgr::build_task(const JobPtr& job) { Loading
cpp/src/scheduler/ResourceMgr.cpp +15 −9 Original line number Diff line number Diff line Loading @@ -79,9 +79,7 @@ ResourceMgr::Add(ResourcePtr&& resource) { gpu_resources_.emplace_back(ResourceWPtr(resource)); break; } default: { break; } default: { break; } } resources_.emplace_back(resource); Loading @@ -104,6 +102,10 @@ ResourceMgr::Connect(const std::string& name1, const std::string& name2, Connect void ResourceMgr::Clear() { std::lock_guard<std::mutex> lck(resources_mutex_); if (running_) { ENGINE_LOG_ERROR << "ResourceMgr is running, cannot clear."; return; } disk_resources_.clear(); cpu_resources_.clear(); gpu_resources_.clear(); Loading Loading @@ -196,13 +198,19 @@ bool ResourceMgr::check_resource_valid() { { // TODO: check one disk-resource, one cpu-resource, zero or more gpu-resource; if (GetDiskResources().size() != 1) return false; if (GetCpuResources().size() != 1) return false; if (GetDiskResources().size() != 1) { return false; } if (GetCpuResources().size() != 1) { return false; } } { // TODO: one compute-resource at least; if (GetNumOfComputeResource() < 1) return false; if (GetNumOfComputeResource() < 1) { return false; } } { Loading Loading @@ -233,9 +241,7 @@ void ResourceMgr::event_process() { while (running_) { std::unique_lock<std::mutex> lock(event_mutex_); event_cv_.wait(lock, [this] { return !queue_.empty(); }); event_cv_.wait(lock, [this] { return !queue_.empty(); }); auto event = queue_.front(); queue_.pop(); Loading
cpp/src/scheduler/task/SearchTask.cpp +5 −8 Original line number Diff line number Diff line Loading @@ -22,6 +22,7 @@ #include "utils/Log.h" #include "utils/TimeRecorder.h" #include <algorithm> #include <string> #include <thread> #include <utility> Loading @@ -34,6 +35,7 @@ static constexpr size_t PARALLEL_REDUCE_BATCH = 1000; std::mutex XSearchTask::merge_mutex_; // TODO(wxyu): remove unused code // bool // NeedParallelReduce(uint64_t nq, uint64_t topk) { // server::ServerConfig &config = server::ServerConfig::GetInstance(); Loading Loading @@ -229,13 +231,8 @@ XSearchTask::Execute() { } Status XSearchTask::TopkResult(const std::vector<long> &input_ids, const std::vector<float> &input_distance, uint64_t input_k, uint64_t nq, uint64_t topk, bool ascending, scheduler::ResultSet &result) { XSearchTask::TopkResult(const std::vector<int64_t>& input_ids, const std::vector<float>& input_distance, uint64_t input_k, uint64_t nq, uint64_t topk, bool ascending, scheduler::ResultSet& result) { scheduler::ResultSet result_buf; if (result.empty()) { Loading
cpp/src/scheduler/task/SearchTask.h +2 −7 Original line number Diff line number Diff line Loading @@ -39,13 +39,8 @@ class XSearchTask : public Task { public: static Status TopkResult(const std::vector<long> &input_ids, const std::vector<float> &input_distance, uint64_t input_k, uint64_t nq, uint64_t topk, bool ascending, scheduler::ResultSet &result); TopkResult(const std::vector<int64_t>& input_ids, const std::vector<float>& input_distance, uint64_t input_k, uint64_t nq, uint64_t topk, bool ascending, scheduler::ResultSet& result); public: TableFileSchemaPtr file_; Loading