Files
EyePy/README.md
T
2024-11-04 09:32:53 +08:00

84 lines
1.9 KiB
Markdown

# EyePy
![made-with-python](https://img.shields.io/badge/Made%20with-Python-1f425f.svg)
![Open Source Love](https://badges.frapsoft.com/os/v1/open-source.svg?v=103)
![License: MIT](https://img.shields.io/badge/License-MIT-yellow.svg)
![GitHub stars](https://img.shields.io/github/stars/ck-zhang/EyePy.svg?style=social)
This is a Python library that provides **webcam-based eye tracking**.
Extract facial features, train eye tracking model and predict gaze with super easy to use interface.
## Usage Showcase
![Demo](https://github.com/user-attachments/assets/7e95c1f6-c56a-4760-a316-13d656fa24d2)
## Installation and Interactive Demo
Clone this project:
```shell
git clone https://github.com/ck-zhang/EyePy
```
### Using Pip
```shell
pip install -r requirements.txt
python demo.py
```
### Using uv
```shell
pip install uv
uv sync
uv run demo.py
```
### Options
- `--filter {kalman,kde}`: Filter method; `kalman` for Kalman Filter, or `kde` for a cool contour like in the demo
- `--background BACKGROUND`: Path to background image file
- `--confidence CONFIDENCE`: Set confidence interval for KDE contour; value must be between 0 and 1
- `--camera CAMERA`: Specify camera index
## Usage
### Initialization
```python
from EyePy import GazeEstimator
gaze_estimator = GazeEstimator()
```
### Feature Extraction
```python
import cv2
image = cv2.imread('image.jpg')
features, blink_detected = gaze_estimator.extract_features(image)
if blink_detected:
print("Blink detected!")
else:
print(features)
```
### Training the Model
```python
X = [...] # Features
y = [...] # Gaze coordinates
gaze_estimator.train(X, y)
```
### Predicting Gaze Location
```python
predicted_gaze = gaze_estimator.predict([features])
print(predicted_gaze)
```
## Future Work
### TODO
- [ ] Integrate with OBS
- [ ] Integrate with opentrack
Any suggestions for features and improvements are welcome.
If you enjoyed using EyePy, consider giving it a star.