This repository is a collection of real-time computer vision mini-projects built with Python, OpenCV, and MediaPipe.
The goal is simple: detect hands, faces, and body pose from a webcam, then use that data for practical demos like finger counting, volume control, and sign-language action recognition.
- Hand tracking and hand landmark detection
- Face detection and face mesh tracking
- Pose estimation
- Finger counting demo
- Hand gesture volume control demo
- Sign-language/action recognition training and inference files
- Python 3.11
- OpenCV
- MediaPipe
- NumPy
- TensorFlow
- scikit-learn
HandTrackingModule.py,FaceDetectionModule.py,FaceMeshModule.py,PoseModule.py: reusable detection modules*Project.pyand*Basics.pyfiles: runnable demo scriptsAiTrainerProject.pyandSign-Language.py: model training and sign-language workflowMP_DATA/: saved training sequence dataimages/andvideos/: assets used by demos
- Clone the repo
- Install dependencies
- Run any demo script
pip install -r requirements.txt
python HandTrackingProject.pyOther examples:
python FaceDetectionBasics.py
python FaceMeshBasics.py
python PoseProject.py
python FingerCountingProject.py
python VolumeHandControl.pyOr use UV
uv sync
uv run HandTrackingProject.pySame for Other examples
- Most scripts use your webcam.
- If one script does not run on your machine immediately, check webcam permissions and Python package versions first.
Active learning/project repository with multiple independent experiments.