a hand gesture recognition using mediapipe hand pose estimation
test.py
a sample program for inferencingkeyclassifier.ipynb
jupyter notebook for preprocessing,building neural network, training model and inference testinglabel.csv
csv file containing labels for the training datakeypoint.csv
contains the logged landmarkskeypoint_classifier.py
module for inferencing
test.py
sample program allows you to key in landmarks from the MediaPipe hand estimation into a keypoint.csv
file.
-
- When you run the program, there's a prompt that tells you to press "n" which sets
mode
to 1 which is the keying mode - You get to choose between (0-9) as the classifying labels for the data points collected. i.e say you make a peace sign and in the
label.csv
your peace sign is first on the list, you press (0) when keying the points(due to indexing) - the keypoints are collected into
keypoint.csv
file
- When you run the program, there's a prompt that tells you to press "n" which sets
- The collected landmarks points in
keypoint.csv
are preprocessed inkeyclassifier.ipynb
using the sci-kit learn library. - Using keras, a training module is built. An input layer of 42 nodes, with 2 hidden layers.
- The output layer has a number nodes equal to
num_classes =4
which is equivalent to the number of training classes.
- After a model has been built, the model and weights are saved into
models/savedkeypoints.h5
- The model is converted into a tflite.
- Inference testing is then performed.