Skip to content

Latest commit

 

History

History
67 lines (40 loc) · 1.26 KB

README.md

File metadata and controls

67 lines (40 loc) · 1.26 KB

PoseHID

hand pose control PC or some other thing (I mean anything)

key point

  1. mediapipe(0.9.3.0)
  2. pyautogui(0.9.53)

Architecture


    hand landmark detect and gesture regression based on `mediapipe`

      |   python   |   tflite  |  onnx(tf2onnx)  |  TensorRT  |

-----------------------------------------------------------------------------

              operation based on hand pose & landmark

PC | PC with GPU | Jetson | RPI4 | some other device without python or linux

-----------------------------------------------------------------------------

    mouse and keyboard simulation based on HID(Human Interface Device)

          map Pose to Operation, based on a json config file

Technology Selection

PC(Windows)

python + mediapipe

  1. Get hand landmark and gesture by mediapipe
  2. Control mouse and keyboard using pyautogui

Get Start

Config

Follow the instruction of readme/中文文档

Run

# need conda, need a conda env called 'PoseHID'
conda activate PoseHID
pip install -r requirements.txt
python main.py

Test

# need conda, need a conda env called 'PoseHID'
conda activate PoseHID
pip install -r requirements.txt
pytest

Limitations

WIP