diff --git a/docs/source/index.rst b/docs/source/index.rst
index 2d51b886a..a99402d98 100644
--- a/docs/source/index.rst
+++ b/docs/source/index.rst
@@ -92,7 +92,6 @@ Now, pick a tutorial or code sample and start utilizing Gen2 capabilities
samples/16_device_queue_event.rst
samples/17_video_mobilenet.rst
samples/18_rgb_encoding_mobilenet.rst
- samples/21_mobilenet_decoding_on_device.rst
samples/22_1_tiny_yolo_v3_decoding_on_device.rst
samples/22_2_tiny_yolo_v4_decoding_on_device.rst
samples/23_autoexposure_roi.rst
diff --git a/docs/source/samples/09_mono_mobilenet.rst b/docs/source/samples/09_mono_mobilenet.rst
index 9cb99f9c5..f08954948 100644
--- a/docs/source/samples/09_mono_mobilenet.rst
+++ b/docs/source/samples/09_mono_mobilenet.rst
@@ -10,7 +10,7 @@ Demo
.. raw:: html
-
+
Setup
diff --git a/docs/source/samples/10_mono_depth_mobilenetssd.rst b/docs/source/samples/10_mono_depth_mobilenetssd.rst
index da36998b5..9c0f749bd 100644
--- a/docs/source/samples/10_mono_depth_mobilenetssd.rst
+++ b/docs/source/samples/10_mono_depth_mobilenetssd.rst
@@ -1,5 +1,5 @@
-10 - Mono & MobilenetSSD & Encoding
-===================================
+10 - Mono & MobilenetSSD & Depth
+================================
This example shows how to run MobileNetv2SSD on the left grayscale camera in parallel with running
the disparity depth results, displaying both the depth map and the right grayscale stream, with the
@@ -11,7 +11,7 @@ Demo
.. raw:: html
-
+
Setup
diff --git a/docs/source/samples/11_rgb_encoding_mono_mobilenet.rst b/docs/source/samples/11_rgb_encoding_mono_mobilenet.rst
index 925c15386..3f1862bb1 100644
--- a/docs/source/samples/11_rgb_encoding_mono_mobilenet.rst
+++ b/docs/source/samples/11_rgb_encoding_mono_mobilenet.rst
@@ -18,7 +18,7 @@ Demo
.. raw:: html
-
+
Setup
diff --git a/docs/source/samples/12_rgb_encoding_mono_mobilenet_depth.rst b/docs/source/samples/12_rgb_encoding_mono_mobilenet_depth.rst
index 9bbb030c1..7053567da 100644
--- a/docs/source/samples/12_rgb_encoding_mono_mobilenet_depth.rst
+++ b/docs/source/samples/12_rgb_encoding_mono_mobilenet_depth.rst
@@ -20,7 +20,7 @@ Demo
.. raw:: html
-
+
Setup
diff --git a/docs/source/samples/15_rgb_mobilenet_4k.rst b/docs/source/samples/15_rgb_mobilenet_4k.rst
index 7639a1a7a..a1e652a4b 100644
--- a/docs/source/samples/15_rgb_mobilenet_4k.rst
+++ b/docs/source/samples/15_rgb_mobilenet_4k.rst
@@ -11,7 +11,7 @@ Demo
.. raw:: html
-
+
diff --git a/docs/source/samples/17_video_mobilenet.rst b/docs/source/samples/17_video_mobilenet.rst
index 4b51a2e4f..e84bda816 100644
--- a/docs/source/samples/17_video_mobilenet.rst
+++ b/docs/source/samples/17_video_mobilenet.rst
@@ -12,7 +12,7 @@ Demo
.. raw:: html
-
+
@@ -23,7 +23,7 @@ Setup
This example also requires MobilenetSDD blob (:code:`mobilenet.blob` file) and prerecorded video
(:code:`construction_vest.mp4` file) to work - you can download them
-here: `mobilenet.blob `__
+here: `mobilenet.blob `__
and `construction_vest.mp4 `__
Source code
diff --git a/docs/source/samples/18_rgb_encoding_mobilenet.rst b/docs/source/samples/18_rgb_encoding_mobilenet.rst
index 86e6e25d3..e44e5caa3 100644
--- a/docs/source/samples/18_rgb_encoding_mobilenet.rst
+++ b/docs/source/samples/18_rgb_encoding_mobilenet.rst
@@ -18,7 +18,7 @@ Demo
.. raw:: html
-
+
Setup
diff --git a/docs/source/samples/21_mobilenet_decoding_on_device.rst b/docs/source/samples/21_mobilenet_decoding_on_device.rst
index 291dcb0f5..fe1925e09 100644
--- a/docs/source/samples/21_mobilenet_decoding_on_device.rst
+++ b/docs/source/samples/21_mobilenet_decoding_on_device.rst
@@ -1,32 +1 @@
-21 - RGB & MobilenetSSD decoding on device
-==========================================
-
-This example shows how to run MobileNetv2SSD on the RGB input frame, and how to display both the RGB
-preview and the metadata results from the MobileNetv2SSD on the preview. It's similar to example '08_rgb_mobilenet' except
-decoding is done on Myriad instead on the host.
-
-
-setConfidenceThreshold - confidence threshold above which objects are detected
-
-Demo
-####
-
-
-Setup
-#####
-
-.. include:: /includes/install_from_pypi.rst
-
-This example also requires MobilenetSDD blob (:code:`mobilenet.blob` file) to work - you can download it from
-`here `__
-
-Source code
-###########
-
-Also `available on GitHub `__
-
-.. literalinclude:: ../../../examples/21_mobilenet_device_side_decoding.py
- :language: python
- :linenos:
-
-.. include:: /includes/footer-short.rst
+This example was moved to :ref:`08 - RGB & MobilenetSSD`
\ No newline at end of file
diff --git a/docs/source/samples/22_1_tiny_yolo_v3_decoding_on_device.rst b/docs/source/samples/22_1_tiny_yolo_v3_decoding_on_device.rst
index b8609f35e..835a002d5 100644
--- a/docs/source/samples/22_1_tiny_yolo_v3_decoding_on_device.rst
+++ b/docs/source/samples/22_1_tiny_yolo_v3_decoding_on_device.rst
@@ -15,6 +15,12 @@ setConfidenceThreshold - confidence threshold above which objects are detected
Demo
####
+.. raw:: html
+
+
+
+
+
Setup
#####
diff --git a/docs/source/samples/22_2_tiny_yolo_v4_decoding_on_device.rst b/docs/source/samples/22_2_tiny_yolo_v4_decoding_on_device.rst
index b39859ae2..9770a8603 100644
--- a/docs/source/samples/22_2_tiny_yolo_v4_decoding_on_device.rst
+++ b/docs/source/samples/22_2_tiny_yolo_v4_decoding_on_device.rst
@@ -15,6 +15,12 @@ setConfidenceThreshold - confidence threshold above which objects are detected
Demo
####
+.. raw:: html
+
+
+
+
+
Setup
#####
diff --git a/docs/source/samples/24_opencv_support.rst b/docs/source/samples/24_opencv_support.rst
index 4316e0208..f0088618c 100644
--- a/docs/source/samples/24_opencv_support.rst
+++ b/docs/source/samples/24_opencv_support.rst
@@ -5,6 +5,15 @@ This example shows API which exposes both numpy and OpenCV compatible image type
It uses ColorCamera node to retrieve both BGR interleaved 'preview' and NV12 encoded 'video' frames.
Both are displayed using functions `getFrame` and `getCvFrame`.
+Demo
+####
+
+.. raw:: html
+
+
+
+
+
Setup
#####
diff --git a/docs/source/samples/25_system_information.rst b/docs/source/samples/25_system_information.rst
index fa89e19f2..c4dff03e8 100644
--- a/docs/source/samples/25_system_information.rst
+++ b/docs/source/samples/25_system_information.rst
@@ -3,6 +3,43 @@
This example shows how to get system information (memory usage, cpu usage and temperature) from the board.
+Demo
+####
+
+Example script output
+
+.. code-block::
+
+ Drr used / total - 0.13 / 414.80 MiB
+ Cmx used / total - 2.24 / 2.50 MiB
+ LeonCss heap used / total - 4.17 / 46.41 MiB
+ LeonMss heap used / total - 2.87 / 27.58 MiB
+ Chip temperature - average: 38.59, css: 39.81, mss: 37.71, upa0: 38.65, upa1: 38.18
+ Cpu usage - Leon OS: 7.08%, Leon RT: 1.48 %
+ ----------------------------------------
+ Drr used / total - 0.13 / 414.80 MiB
+ Cmx used / total - 2.24 / 2.50 MiB
+ LeonCss heap used / total - 4.17 / 46.41 MiB
+ LeonMss heap used / total - 2.87 / 27.58 MiB
+ Chip temperature - average: 38.59, css: 39.58, mss: 37.94, upa0: 38.18, upa1: 38.65
+ Cpu usage - Leon OS: 1.55%, Leon RT: 0.30 %
+ ----------------------------------------
+ Drr used / total - 0.13 / 414.80 MiB
+ Cmx used / total - 2.24 / 2.50 MiB
+ LeonCss heap used / total - 4.17 / 46.41 MiB
+ LeonMss heap used / total - 2.87 / 27.58 MiB
+ Chip temperature - average: 38.94, css: 40.04, mss: 38.18, upa0: 39.35, upa1: 38.18
+ Cpu usage - Leon OS: 0.56%, Leon RT: 0.06 %
+ ----------------------------------------
+ Drr used / total - 0.13 / 414.80 MiB
+ Cmx used / total - 2.24 / 2.50 MiB
+ LeonCss heap used / total - 4.17 / 46.41 MiB
+ LeonMss heap used / total - 2.87 / 27.58 MiB
+ Chip temperature - average: 39.46, css: 40.28, mss: 38.88, upa0: 39.81, upa1: 38.88
+ Cpu usage - Leon OS: 0.51%, Leon RT: 0.06 %
+ ----------------------------------------
+
+
Setup
#####
diff --git a/docs/source/samples/26_1_spatial_mobilenet.rst b/docs/source/samples/26_1_spatial_mobilenet.rst
index 26b49fb86..6bfa75037 100644
--- a/docs/source/samples/26_1_spatial_mobilenet.rst
+++ b/docs/source/samples/26_1_spatial_mobilenet.rst
@@ -11,6 +11,12 @@ setConfidenceThreshold - confidence threshold above which objects are detected
Demo
####
+.. raw:: html
+
+
+
+
+
Setup
#####
diff --git a/docs/source/samples/26_2_spatial_mobilenet_mono.rst b/docs/source/samples/26_2_spatial_mobilenet_mono.rst
index 89a9eb08d..4084be99c 100644
--- a/docs/source/samples/26_2_spatial_mobilenet_mono.rst
+++ b/docs/source/samples/26_2_spatial_mobilenet_mono.rst
@@ -1,4 +1,4 @@
-26.2 - MONO & MobilenetSSD with spatial data
+26.2 - Mono & MobilenetSSD with spatial data
============================================
This example shows how to run MobileNetv2SSD on the rectified right input frame, and
@@ -11,6 +11,12 @@ setConfidenceThreshold - confidence threshold above which objects are detected
Demo
####
+.. raw:: html
+
+
+
+
+
Setup
#####
diff --git a/docs/source/samples/26_3_spatial_tiny_yolo.rst b/docs/source/samples/26_3_spatial_tiny_yolo.rst
index bfdaf7a45..125e131d7 100644
--- a/docs/source/samples/26_3_spatial_tiny_yolo.rst
+++ b/docs/source/samples/26_3_spatial_tiny_yolo.rst
@@ -17,6 +17,12 @@ setConfidenceThreshold - confidence threshold above which objects are detected
Demo
####
+.. raw:: html
+
+
+
+
+
Setup
#####
diff --git a/docs/source/samples/28_camera_video_example.rst b/docs/source/samples/28_camera_video_example.rst
index 5288f0dbf..b3f6340e2 100644
--- a/docs/source/samples/28_camera_video_example.rst
+++ b/docs/source/samples/28_camera_video_example.rst
@@ -4,6 +4,15 @@
This example shows how to use high resolution video at low latency. Compared to :ref:`01 - RGB Preview`, this demo outputs NV12 frames whereas
preview frames are BGR and are not suited for larger resoulution (eg. 2000x1000). Preview is more suitable for either NN or visualization purposes.
+Demo
+####
+
+.. raw:: html
+
+
+
+
+
Setup
#####
diff --git a/examples/12_rgb_encoding_mono_mobilenet_depth.py b/examples/12_rgb_encoding_mono_mobilenet_depth.py
index b1cff44f1..57f713a57 100755
--- a/examples/12_rgb_encoding_mono_mobilenet_depth.py
+++ b/examples/12_rgb_encoding_mono_mobilenet_depth.py
@@ -34,7 +34,7 @@
camRight.setResolution(dai.MonoCameraProperties.SensorResolution.THE_400_P)
depth = pipeline.createStereoDepth()
-depth.setConfidenceThreshold(200)
+depth.setConfidenceThreshold(255)
# Note: the rectified streams are horizontally mirrored by default
depth.setOutputRectified(True)
depth.setRectifyEdgeFillColor(0) # Black, to better see the cutout
@@ -80,11 +80,11 @@
# Start pipeline
device.startPipeline()
- queue_size = 8
- qRight = device.getOutputQueue("right", queue_size)
- qDepth = device.getOutputQueue("depth", queue_size)
- qManip = device.getOutputQueue("manip", queue_size)
- qDet = device.getOutputQueue("nn", queue_size)
+ queueSize = 8
+ qRight = device.getOutputQueue("right", queueSize)
+ qDepth = device.getOutputQueue("depth", queueSize)
+ qManip = device.getOutputQueue("manip", queueSize)
+ qDet = device.getOutputQueue("nn", queueSize)
qRgbEnc = device.getOutputQueue('h265', maxSize=30, blocking=True)
frame = None
diff --git a/examples/README.md b/examples/README.md
index ab5d2164c..aaed656c2 100644
--- a/examples/README.md
+++ b/examples/README.md
@@ -57,28 +57,65 @@ TEST_TIMEOUT=0 ctest -R "01_rgb_preview" --verbose

### 08_rgb_mobilenet.py
-
+
### 09_mono_mobilenet.py
-
+
### 10_mono_depth_mobilenetssd.py
**NOTE** For clearness, a `disparity` output is used instead of `depth`, as disparity is better to visualize
-
+
### 11_rgb_encoding_mono_mobilenet.py
-
+
### 12_rgb_encoding_mono_mobilenet_depth.py
**NOTE** For clearness, a `disparity` output is used instead of `depth`, as disparity is better to visualize
-
+
### 13_encoding_max_limit.py

### 14_color_camera_control.py

+
+### 15_rgb_mobilenet_4k.py
+
+
+### 16_device_queue_event.py
+
+
+### 17_video_mobilenet.py
+
+
+### 18_rgb_encoding_mobilenet.py
+
+
+### 22_1_tiny_yolo_v3_device_side_decoding.py
+
+
+### 22_2_tiny_yolo_v4_device_side_decoding.py
+
+
+### 23_autoexposure_roi.py
+
+
+### 24_opencv_support.py
+
+
+### 26_1_spatial_mobilenet.py
+
+
+### 26_2_spatial_mobilenet_mono.py
+
+
+### 26_3_spatial_tiny_yolo.py
+
+
+### 28_camera_video_example.py
+
+