Skip to content

A proof of concept to manage Microsoft Teams using hands gesture, powered by Mediapipe

Notifications You must be signed in to change notification settings

FedericoMontana/computer-vision-teams

Repository files navigation

This is a simple proof of concept to manage Microsoft Teams using hands gesture (touchless)

I have create a couple of handy classes that abstract much of the complexity of the computer vision logic. Thre are several jupyter notebooks to run the PoC:

Notebooks starting with 01 creates the "data" to train a model. You can execute either of the following parts:

  • data_cam creates the data by using your own cam, and you decide what gestures you want to include
  • data_imgs creates the data by using images. I have used a public repository of hands images as an example

Notebooks starting with 02 train the model, using the data created in step 01. You can execute either of the following parts:

Notebooks 03 grabs the model created, and uses it with your webcam to command Teams.

Note that I have created just a few combinations to manage Teams, you can easily extend the Teams class to add more commands, and associate them with whatever gesture you want..

About

A proof of concept to manage Microsoft Teams using hands gesture, powered by Mediapipe

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published