30 lines
3.2 KiB
Markdown
30 lines
3.2 KiB
Markdown
# Sync Benchmark C++ Sample
|
|
|
|
This sample demonstrates how to estimate performance of a model using Synchronous Inference Request API. It makes sense to use synchronous inference only in latency oriented scenarios. Models with static input shapes are supported. Unlike [demos](https://github.com/openvinotoolkit/open_model_zoo/tree/master/demos) this sample doesn't have other configurable command line arguments. Feel free to modify sample's source code to try out different options.
|
|
|
|
For more detailed information on how this sample works, check the dedicated [article](https://docs.openvino.ai/2025/get-started/learn-openvino/openvino-samples/sync-benchmark.html)
|
|
|
|
## Requirements
|
|
|
|
| Options | Values |
|
|
| -------------------------------| -------------------------------------------------------------------------------------------------------------------------|
|
|
| Validated Models | [yolo-v3-tf](https://github.com/openvinotoolkit/open_model_zoo/tree/master/models/public/yolo-v3-tf), |
|
|
| | [face-detection-0200](https://github.com/openvinotoolkit/open_model_zoo/tree/master/models/intel/face-detection-0200) |
|
|
| Model Format | OpenVINO™ toolkit Intermediate Representation |
|
|
| | (\*.xml + \*.bin), ONNX (\*.onnx) |
|
|
| Supported devices | [All](https://docs.openvino.ai/2025/documentation/compatibility-and-support/supported-devices.html) |
|
|
| Other language realization | [Python](https://docs.openvino.ai/2025/get-started/learn-openvino/openvino-samples/sync-benchmark.html) |
|
|
|
|
The following C++ API is used in the application:
|
|
|
|
| Feature | API | Description |
|
|
| -------------------------| ---------------------------------------------|----------------------------------------------|
|
|
| OpenVINO Runtime Version | ``ov::get_openvino_version`` | Get Openvino API version. |
|
|
| Basic Infer Flow | ``ov::Core``, ``ov::Core::compile_model``, | Common API to do inference: compile a model, |
|
|
| | ``ov::CompiledModel::create_infer_request``, | create an infer request, |
|
|
| | ``ov::InferRequest::get_tensor`` | configure input tensors. |
|
|
| Synchronous Infer | ``ov::InferRequest::infer``, | Do synchronous inference. |
|
|
| Model Operations | ``ov::CompiledModel::inputs`` | Get inputs of a model. |
|
|
| Tensor Operations | ``ov::Tensor::get_shape``, | Get a tensor shape and its data. |
|
|
| | ``ov::Tensor::data`` | |
|