13 base_detection::base_detection(
15 const std::string &pathToModel,
16 int maxBatch,
bool isBatchDynamic,
bool isAsync,
17 bool doRawOutputMessages
19 : topoName( topoName ), pathToModel( pathToModel )
20 , maxBatch( maxBatch ), isBatchDynamic( isBatchDynamic ), isAsync( isAsync )
21 , enablingChecked( false ), _enabled( false ), doRawOutputMessages( doRawOutputMessages )
41 _request->Wait( IInferRequest::WaitMode::RESULT_READY );
63 std::map<std::string, std::string>
config = { };
66 bool isPossibleDynBatch =
67 deviceName.find(
"CPU" ) != std::string::npos ||
68 deviceName.find(
"GPU" ) != std::string::npos;
69 if( isPossibleDynBatch )
70 config[PluginConfigParams::KEY_DYN_BATCH_ENABLED] = PluginConfigParams::YES;
InferenceEngine::InferRequest::Ptr _request
InferenceEngine::ExecutableNetwork net
LOG(INFO)<< "Log message to default logger"
void load_into(InferenceEngine::Core &ie, const std::string &deviceName)
virtual void submit_request()
virtual InferenceEngine::CNNNetwork read_network()=0