Shipped this project!
I build a ML model to detect your hand and see which gestures you are doing including “Pinch, Cursor, Scroll Up, and Scroll Down”. This project has the goal to help individuals who are disabled or injured still be able to interact with their computers without having to be restricted to keyboard and mouse. It was a huge challenge for me to create a fast backend and frontend so that people would actually be able to interact with a demo of my model at a reasonable framerate (~10fps). I did this through moving the mediapipe logic to run locally on the JS and the Python would just send light packages with position and gesture data rather than full heavy images. In the end I just migrated all the python code to js so that it could all run locally without having to communicate between frontend and backend (m. I am proud of this project because it gave me hands on experience with ML models and big libraries like opencv, mediapipe, and working through a fast backend -> frontend with flask, and socketio and was overall very fun to work on!