This Python application utilizes real-time facial tracking to control music volume based on facial gestures. Specifically:
Volume increases as the user’s mouth opens wider.
Music stops when the eyes are closed.
The interface displays a live webcam feed with facial landmarks, volume percentage, and a vertical volume bar.
Uses
MediaPipefor face mesh detection,PyQt5for GUI, andOpenCVfor image processing.Compatible with macOS (using
osascriptfor volume control). Windows may require additional setup.
This project is intended for educational and experimental purposes. In real-world applications, it can be extended to:
Monitor eye state to detect drowsiness.
Track mouth opening frequency — which can be useful in areas such as:
Automatic warning systems
Statistics on students sleeping or talking during class
Monitoring focus in training or workplace environments
This application shows potential for developing contactless interactive systems, especially in education, healthcare, and occupational safety.

Tiếng Việt