Files
ANSCORE/engines/OpenVINOEngine/include/pipelines/async_pipeline.h

124 lines
4.9 KiB
C++

/*
// Copyright (C) 2020-2024 Intel Corporation
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
*/
#pragma once
#include <stdint.h>
#include <condition_variable>
#include <exception>
#include <memory>
#include <mutex>
#include <unordered_map>
#include <openvino/openvino.hpp>
#include <models/results.h>
#include <utils/performance_metrics.hpp>
#include "pipelines/requests_pool.h"
#include "models/model_base.h"
class ModelBase;
struct InputData;
struct MetaData;
struct ModelConfig;
/// This is base class for asynchronous pipeline
/// Derived classes should add functions for data submission and output processing
class AsyncPipeline {
public:
/// Loads model and performs required initialization
/// @param modelInstance pointer to model object. Object it points to should not be destroyed manually after passing
/// pointer to this function.
/// @param config - fine tuning configuration for model
/// @param core - reference to ov::Core instance to use.
/// If it is omitted, new instance of ov::Core will be created inside.
AsyncPipeline(std::unique_ptr<ModelBase>&& modelInstance, const ModelConfig& config, ov::Core& core);
virtual ~AsyncPipeline();
/// Waits until either output data becomes available or pipeline allows to submit more input data.
/// @param shouldKeepOrder if true, function will treat results as ready only if next sequential result (frame) is
/// ready (so results can be extracted in the same order as they were submitted). Otherwise, function will return if
/// any result is ready.
void waitForData(bool shouldKeepOrder = true);
/// @returns true if there's available infer requests in the pool
/// and next frame can be submitted for processing, false otherwise.
bool isReadyToProcess() {
return requestsPool->isIdleRequestAvailable();
}
/// Waits for all currently submitted requests to be completed.
///
void waitForTotalCompletion() {
if (requestsPool)
requestsPool->waitForTotalCompletion();
}
void updateModelImageSize(const cv::Size& inputImgSize) {
model->updateImageSize(inputImgSize);
}
/// Submits data to the model for inference
/// @param inputData - input data to be submitted
/// @param metaData - shared pointer to metadata container.
/// Might be null. This pointer will be passed through pipeline and put to the final result structure.
/// @returns -1 if image cannot be scheduled for processing (there's no free InferRequest available).
/// Otherwise returns unique sequential frame ID for this particular request. Same frame ID will be written in the
/// result structure.
virtual int64_t submitData(const InputData& inputData, const std::shared_ptr<MetaData>& metaData);
/// Gets available data from the queue
/// @param shouldKeepOrder if true, function will treat results as ready only if next sequential result (frame) is
/// ready (so results can be extracted in the same order as they were submitted). Otherwise, function will return if
/// any result is ready.
virtual std::unique_ptr<ResultBase> getResult(bool shouldKeepOrder = true);
PerformanceMetrics getInferenceMetircs() {
return inferenceMetrics;
}
PerformanceMetrics getPreprocessMetrics() {
return preprocessMetrics;
}
PerformanceMetrics getPostprocessMetrics() {
return postprocessMetrics;
}
protected:
/// Returns processed result, if available
/// @param shouldKeepOrder if true, function will return processed data sequentially,
/// keeping original frames order (as they were submitted). Otherwise, function will return processed data in random
/// order.
/// @returns InferenceResult with processed information or empty InferenceResult (with negative frameID) if there's
/// no any results yet.
virtual InferenceResult getInferenceResult(bool shouldKeepOrder);
std::unique_ptr<RequestsPool> requestsPool;
std::unordered_map<int64_t, InferenceResult> completedInferenceResults;
ov::CompiledModel compiledModel;
std::mutex mtx;
std::condition_variable condVar;
int64_t inputFrameId = 0;
int64_t outputFrameId = 0;
std::exception_ptr callbackException = nullptr;
std::unique_ptr<ModelBase> model;
PerformanceMetrics inferenceMetrics;
PerformanceMetrics preprocessMetrics;
PerformanceMetrics postprocessMetrics;
};