Fix mutex lock issues
This commit is contained in:
@@ -58,6 +58,7 @@ namespace ANSCENTER {
|
||||
bool ANSRTYOLO::LoadModel(const std::string& modelZipFilePath,
|
||||
const std::string& modelZipPassword) {
|
||||
std::lock_guard<std::recursive_mutex> lock(_mutex);
|
||||
ModelLoadingGuard mlg(_modelLoading);
|
||||
try {
|
||||
_isFixedBatch = false;
|
||||
bool result = ANSODBase::LoadModel(modelZipFilePath, modelZipPassword);
|
||||
@@ -148,6 +149,7 @@ namespace ANSCENTER {
|
||||
std::string modelName, std::string className,
|
||||
const std::string& modelFolder, std::string& labelMap) {
|
||||
std::lock_guard<std::recursive_mutex> lock(_mutex);
|
||||
ModelLoadingGuard mlg(_modelLoading);
|
||||
try {
|
||||
_isFixedBatch = false;
|
||||
bool result = ANSODBase::LoadModelFromFolder(licenseKey, modelConfig,
|
||||
@@ -250,6 +252,7 @@ namespace ANSCENTER {
|
||||
const std::string& modelZipPassword,
|
||||
std::string& labelMap) {
|
||||
std::lock_guard<std::recursive_mutex> lock(_mutex);
|
||||
ModelLoadingGuard mlg(_modelLoading);
|
||||
try {
|
||||
const bool engineAlreadyLoaded = _modelLoadValid && _isInitialized && m_trtEngine != nullptr;
|
||||
_modelLoadValid = false;
|
||||
@@ -351,23 +354,7 @@ namespace ANSCENTER {
|
||||
|
||||
std::vector<Object> ANSRTYOLO::RunInference(const cv::Mat& inputImgBGR,
|
||||
const std::string& camera_id) {
|
||||
{
|
||||
std::lock_guard<std::recursive_mutex> lock(_mutex);
|
||||
if (!_modelLoadValid) {
|
||||
_logger.LogError("ANSRTYOLO::RunInference", "Cannot load TensorRT model", __FILE__, __LINE__);
|
||||
return {};
|
||||
}
|
||||
if (!_licenseValid) {
|
||||
_logger.LogError("ANSRTYOLO::RunInference", "Invalid license", __FILE__, __LINE__);
|
||||
return {};
|
||||
}
|
||||
if (!_isInitialized) {
|
||||
_logger.LogError("ANSRTYOLO::RunInference", "Model not initialized", __FILE__, __LINE__);
|
||||
return {};
|
||||
}
|
||||
if (inputImgBGR.empty() || inputImgBGR.cols < 10 || inputImgBGR.rows < 10)
|
||||
return {};
|
||||
}
|
||||
if (!PreInferenceCheck("ANSRTYOLO::RunInference")) return {};
|
||||
try { return DetectObjects(inputImgBGR, camera_id); }
|
||||
catch (const std::exception& e) {
|
||||
_logger.LogFatal("ANSRTYOLO::RunInference", e.what(), __FILE__, __LINE__);
|
||||
@@ -377,25 +364,7 @@ namespace ANSCENTER {
|
||||
|
||||
std::vector<std::vector<Object>> ANSRTYOLO::RunInferencesBatch(
|
||||
const std::vector<cv::Mat>& inputs, const std::string& camera_id) {
|
||||
{
|
||||
std::lock_guard<std::recursive_mutex> lock(_mutex);
|
||||
if (!_modelLoadValid) {
|
||||
_logger.LogFatal("ANSRTYOLO::RunInferencesBatch", "Cannot load the TensorRT model", __FILE__, __LINE__);
|
||||
return {};
|
||||
}
|
||||
if (!_licenseValid) {
|
||||
_logger.LogFatal("ANSRTYOLO::RunInferencesBatch", "Runtime license is not valid or expired", __FILE__, __LINE__);
|
||||
return {};
|
||||
}
|
||||
if (!_isInitialized) {
|
||||
_logger.LogFatal("ANSRTYOLO::RunInferencesBatch", "Initialisation is not valid", __FILE__, __LINE__);
|
||||
return {};
|
||||
}
|
||||
if (inputs.empty()) {
|
||||
_logger.LogFatal("ANSRTYOLO::RunInferencesBatch", "Input images vector is empty", __FILE__, __LINE__);
|
||||
return {};
|
||||
}
|
||||
}
|
||||
if (!PreInferenceCheck("ANSRTYOLO::RunInferencesBatch")) return {};
|
||||
try {
|
||||
if (_isFixedBatch) return ANSODBase::RunInferencesBatch(inputs, camera_id);
|
||||
else return DetectObjectsBatch(inputs, camera_id);
|
||||
|
||||
Reference in New Issue
Block a user