diff options
author | Inki Dae <inki.dae@samsung.com> | 2021-12-22 16:28:46 +0900 |
---|---|---|
committer | Inki Dae <inki.dae@samsung.com> | 2021-12-22 16:28:46 +0900 |
commit | 919fd9404cdbc934628a4b97c6fca21b6443fb3e (patch) | |
tree | a1d2ce5e80b38e5beea21ec5a855ad5118c555c2 | |
parent | 540c264c7ce5340bbe14207e4eeb5d38468d94c6 (diff) | |
download | inference-engine-tflite-919fd9404cdbc934628a4b97c6fca21b6443fb3e.tar.gz inference-engine-tflite-919fd9404cdbc934628a4b97c6fca21b6443fb3e.tar.bz2 inference-engine-tflite-919fd9404cdbc934628a4b97c6fca21b6443fb3e.zip |
fix seg. fault issue with GPU delegatesubmit/tizen/20220105.081745submit/tizen/20220105.080154accepted/tizen/unified/20220110.140032
Fixed seg. fault issue with GPU delegate of Tensorflow lite 2.7.
With Tensorflow lite 2.7 GPU delegate creation needs a default option.
So this patch gets a default option for creating GPU delegate by calling
TfLiteGpuDelegateOptionsV2Default().
[Versin] 0.0.2-0
[Issue type] bug fix
Change-Id: I4a99377e393aacb415de62ac9004c3095623d923
Signed-off-by: Yongjoo Ahn <yongjoo1.ahn@samsung.com>
Signed-off-by: Inki Dae <inki.dae@samsung.com>
-rw-r--r-- | packaging/inference-engine-tflite.spec | 4 | ||||
-rw-r--r-- | src/inference_engine_tflite.cpp | 3 |
2 files changed, 4 insertions, 3 deletions
diff --git a/packaging/inference-engine-tflite.spec b/packaging/inference-engine-tflite.spec index 587924a..6bea41e 100644 --- a/packaging/inference-engine-tflite.spec +++ b/packaging/inference-engine-tflite.spec @@ -1,7 +1,7 @@ Name: inference-engine-tflite Summary: Tensorflow-Lite based implementation of inference-engine-interface -Version: 0.0.1 -Release: 14 +Version: 0.0.2 +Release: 0 Group: Multimedia/Libraries License: Apache-2.0 Source0: %{name}-%{version}.tar.gz diff --git a/src/inference_engine_tflite.cpp b/src/inference_engine_tflite.cpp index 33e7c04..2043ff5 100644 --- a/src/inference_engine_tflite.cpp +++ b/src/inference_engine_tflite.cpp @@ -116,7 +116,8 @@ namespace TFLiteImpl LOGI("Inferece targets are: [%d]", mTargetTypes); if (mTargetTypes == INFERENCE_TARGET_GPU) { - TfLiteDelegate *delegate = TfLiteGpuDelegateV2Create(nullptr); + TfLiteGpuDelegateOptionsV2 options = TfLiteGpuDelegateOptionsV2Default(); + TfLiteDelegate *delegate = TfLiteGpuDelegateV2Create(&options); if (!delegate){ LOGE("Failed to GPU delegate"); return INFERENCE_ENGINE_ERROR_INVALID_OPERATION; |