This document provides detailed instructions on how to record your own dataset using the slr_collecting_data.ipynb Jupyter notebook. This is essential for creating and customizing a dataset for training the sign language recognition model in the Singa project.
-
Running the Notebook:
- Open and run the Jupyter notebook file
slr_collecting_data.ipynb.
- Open and run the Jupyter notebook file
-
Adjusting Parameters:
- Videos per Label: You can change the number of videos per label/action by adjusting the
videos_per_labelvariable. The default value is60. - Frames per Video: You can change the number of frames per video by adjusting the
frames_per_videovariable. The default value is60.
- Videos per Label: You can change the number of videos per label/action by adjusting the
-
Using Mediapipe TaskVision:
- For more information on how to use Mediapipe TaskVision, refer to the following resources:
-
Video Capture:
- The script will capture the specified number of videos (60 by default) per label action.
- Each video will consist of
60 frames by default. - There will be a pause of approximately
1.5 secondsbetween each video capture to allow the actor to adjust for the next video.
-
Debugging:
- For debugging purposes, you can set
debug_landmark=Trueto visualize the keypoints detected by Mediapipe TaskVision. This helps the actor verify if their sign actions are being detected properly.
- For debugging purposes, you can set
-
Exiting the Recording Session:
- To quit the recording session, press
qon the keyboard.
- To quit the recording session, press
By following these instructions, you can efficiently record and customize your dataset for training the sign language recognition model. Ensure that the parameters are set according to your requirements and refer to the Mediapipe documentation for additional customization options.