- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Hello everyone.
I successfully transformed faster_rcnn_resnet101_coco model to IR using code bellow
<path_to_file>/mo_tf.py --input_model=<path_to_file>/faster_rcnn_resnet101_coco_2018_01_28/frozen_inference_graph.pb --tensorflow_use_custom_operations_config <path_to_file>/faster_rcnn_support.json --tensorflow_object_detection_api_pipeline_config <path_to_file>/faster_rcnn_resnet101_coco_2018_01_28/pipeline.config --reverse_input_channels --data_type FP16 --output=detection_boxes,detection_scores,num_detections
When I try to run the object_detection_demo_ssd_async.py using NCSV2
python3 <path_to_file>/object_detection_demo_ssd_async/object_detection_demo_ssd_async.py -i cam -m <path_to_file>/frozen_inference_graph.xml --labels <path_to_file>/frozen_inference_graph.mapping -d MYRIAD
I get an error:
[ INFO ] Reading IR...
Traceback (most recent call last):
File "/opt/intel/openvino_2019.1.144/deployment_tools/inference_engine/samples/python_samples/object_detection_demo_ssd_async/object_detection_demo_ssd_async.py", line 185, in <module>
sys.exit(main() or 0)
File "/opt/intel/openvino_2019.1.144/deployment_tools/inference_engine/samples/python_samples/object_detection_demo_ssd_async/object_detection_demo_ssd_async.py", line 75, in main
assert len(net.inputs.keys()) == 1, "Demo supports only single input topologies"
AssertionError: Demo supports only single input topologies
It possible run faster_rcnn_resnet101_coco on the python sample/demo?
Link Copied
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Same issue here - anyone have a solution?
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Hi,
I have successfully run the demo for ssd_mobilenet_v2_coco model on CPU however unable to run inference on the faster_rcnn models (both resnet101 and inception). The same error as you have posted above is returned when running inference for both model variants.
Have you had any success in resolving this?
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Dear kovrigin, vladimir and others - the non-SSD should work with faster_rcnn, http://docs.openvinotoolkit.org/latest/_inference_engine_samples_object_detection_demo_README.html
Unfortunately there is a bug which will be fixed in 2019R2 - which should be released "Any Day Now".
Thanks,
Shubha
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Shubha R. (Intel) wrote:Dear kovrigin, vladimir and others - the non-SSD should work with faster_rcnn, http://docs.openvinotoolkit.org/latest/_inference_engine_samples_object_detection_demo_README.html
Unfortunately there is a bug which will be fixed in 2019R2 - which should be released "Any Day Now".
Thanks,
Shubha
Hi Shubha,
same problem here. Both OpenVINO and OpenCV python libraries cannot use faster rcnn models converted with MO for Inference.
I have tested it also with latest OpenVINO version R2, but the problem is the same about inputs length>1.
I have successfully converted the model, after many issues, by using TF 1.13.1 and OV 2019.R1.1 latest, because with R2 version conversion give me the same error here https://software.intel.com/en-us/forums/computer-vision/topic/813854#comment-1942022 ).
Do you have a workaround for this solution?
I found this solution https://github.com/opencv/cvat/pull/545#issuecomment-508151724, but I didn't understand if it works and what this snippet do. Is valid also for IE library?
Thank you
Davide
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Dear Maffei, Davide,
I apologize for all of the confusion. But as of OpenVino 2019R2, Object Detection Sample SSD will work for faster rcnn. Other customers have tried it and succeeded. Please read this IDZ post for a detailed explanation.
Also 2019R2.01 is actually available now. Please try it !
I hope it helps.
Shubha

- Subscribe to RSS Feed
- Mark Topic as New
- Mark Topic as Read
- Float this Topic for Current User
- Bookmark
- Subscribe
- Printer Friendly Page