Skip to content
This repository has been archived by the owner on Aug 26, 2024. It is now read-only.

index.rst page edits #337

Open
wants to merge 2 commits into
base: master
Choose a base branch
from
Open
Show file tree
Hide file tree
Changes from 1 commit
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Prev Previous commit
Additional docs updates
  • Loading branch information
stuartsmoore committed Aug 12, 2022
commit cd8c612e03f3854e5b4932efdd92ff36861c5fcd
6 changes: 3 additions & 3 deletions source/pages/depth.rst
Original file line number Diff line number Diff line change
@@ -1,7 +1,7 @@
Depth perception
================

DepthAI platform supports different ways of perceiving depth:
The DepthAI platform supports different ways of perceiving depth:

#. :ref:`Passive stereo depth perception` - Used by non-Pro version of OAK-D cameras
#. :ref:`Active stereo depth perception` - Used by `Pro version <https://docs.luxonis.com/projects/hardware/en/latest/pages/articles/oak-s2.html#pro-version>`__ of OAK-D cameras
Expand All @@ -16,14 +16,14 @@ stereo camera pair - left and right monocular cameras - and the VPU (brains of t
**disparity matching** to estimate the depth of objects and scenes.

The **disparity** is the distance (in pixels) between the same point in the left and right image of the stereo pair camera.
A great demo of disparity is below - the disparity is visualized with a red line between points/features - which in this
A great demo of disparity is seen below. The disparity is visualized with a red line between points/features, which in this
case are facial landmarks (eyes, nose, mouth).

.. image:: https://user-images.githubusercontent.com/59799831/132098832-70a2d0b9-1a30-4994-8dad-dc880a803fb3.gif
:target: https://github.com/luxonis/depthai-experiments/tree/master/gen2-triangulation
:alt: Triangulation Demo

OAK-D camera does that for every pixel in the mono frame - it goes through pixels on the first mono frame,
The OAK-D camera does that for every pixel in the mono frame; it goes through pixels on the first mono frame,
finds the same point/feature on the second mono frame, and assigns a disparity value (in pixels) with some confidence for every pixel.
This all happens inside the `StereoDepth node <https://docs.luxonis.com/projects/api/en/latest/components/nodes/stereo_depth/>`__.
The depth map is calculated from the disparity map (on the camera) using `this formula <https://docs.luxonis.com/projects/api/en/latest/components/nodes/stereo_depth/#calculate-depth-using-disparity-map>`__.
Expand Down
14 changes: 7 additions & 7 deletions source/pages/spatial-ai.rst
Original file line number Diff line number Diff line change
Expand Up @@ -3,8 +3,8 @@
Spatial AI
==========

**Spatial AI** allows robots or computers to perceive the world as a human can - what objects or features are - and where
they are in the physical world. DepthAI platform **leverages Spatial AI by fusing** :ref:`AI capabilities <AI / ML / NN>` with
**Spatial AI** allows robots or computers to perceive the world like a human can, such as what objects or features are and where
they are in the physical world. The DepthAI platform **leverages Spatial AI by fusing** :ref:`AI capabilities <AI / ML / NN>` with
:ref:`depth perception <Depth perception>` on the OAK camera itself.

There are a few different approaches to achieve AI + depth fusion:
Expand Down Expand Up @@ -56,8 +56,8 @@ a depth map to provide spatial coordinates (XYZ) of objects it sees: person, pot
------------------------

An example would be a hand landmark detector on DepthAI. With a regular camera, this network returns
the 2D (XY) coordinates of all 21 hand landmarks (contours of all joints in fingers). Using this same network
with DepthAI, these 21 hand landmarks are now 3D points in physical space instead of 2D points in pixel space
the 2D (XY) coordinates of all 21 hand landmarks (contours of all joints of the fingers). Using this same network
with DepthAI, these 21 hand landmarks are now 3D points in physical space instead of 2D points in pixel space.

.. image:: https://user-images.githubusercontent.com/18037362/156813885-146602d9-e5f9-4afa-9542-7904c7e236b3.gif
:target: https://www.youtube.com/watch?v=xXXsT6afW6E
Expand All @@ -72,7 +72,7 @@ detection demos.
#################

One of the classic problems in autonomous robotic navigation/actuation are **unknown objects**.
Known objects are specified before the installation to be encountered - such as tools, other machines,
Known objects are specified before the installation to be encountered, such as tools, other machines,
workers, equipment, and facilities.

We cannot anticipate unknown objects - including those
Expand Down Expand Up @@ -104,8 +104,8 @@ only to include the person’s depth.
##########################

In this mode, the neural inference (landmark detection) is run on the left **and** right cameras to
produce stereo inference results. Unlike monocular neural inference fused with stereo depth - there is no max disparity
search limit - so the minimum distance is purely limited by the greater of (a) horizontal field of view (HFOV) of the
produce stereo inference results. Unlike monocular neural inference fused with stereo depth there is no max disparity
search limit, so the minimum distance is purely limited by the greater of (a) horizontal field of view (HFOV) of the
stereo cameras themselves and (b) the hyperfocal distance of the cameras (minimal distance for objects to be in focus).

After we have 2D positions of landmarks from both left/right cameras, we can calculate the disparity of the results,
Expand Down
Loading