-
Notifications
You must be signed in to change notification settings - Fork 2
Home
Welcome to the JeFaPaTo wiki!
This page contains an overview of the tutorials on how to use JeFaPaTo.
We focus on these tutorials to be short when using the GUI version of our tool.
After that, you should be able to use the tool independently and adapt the settings or workflow to your needs.
For now, we explain how to extract the EAR score and, afterward, the actual blinks from the time series.
All steps include annotated images with red
boxes, arrows, or numbers to guide you through the UI and UX.
If you have any problem with something being unclear, please open an issue so we can improve the wiki tutorials together for other users 😄
The current tutorials are:
This section will discuss some of the details of the inner workings of JeFaPaTo
.
Especially information about expected input file and generated output formats, computer vision models, and the correctness of the calculation will be addressed.
Also, we give more insights how and which facial features are computed and how to read the visual summary of the statistics.
For the Facial Feature and EAR score
extraction only a video file is needed.
As the actual extraction is independent of the frames per second
, you can choose which ever you fell works rights in your experimental setting.
Regarding the video files, we currently support the same ones as OpenCV (the internal encodings of the video containers is handeld by ffmpeg)
- .mp4
- .flv
- .ts
- .mts
- .avi
- .mov
- .wmv
We tested JeFaPaTo's
capabilites regarding different video resolution and all tests with 1280x720
, 1920x1080
, and 3840x2160
succeeded.
In our custom loading threads in the background we limit the usage RAM to not crash a computer.
Therefore, loading and processing speed are rely on CPU, GPU, and hard disk space of the computer.
If a video was recorded in a different orientation, you can use the Rotation
setting inside the extraction GUI to adjust that.
This method only expects .csv
file as input with a header naming the according columns.
Each row describes the frame id
and the columns shall the be the actual features.
If you use JeFaPaTo
for the EAR score
extraction, we automatially detect the EAR2D_*
or EAR3D_*
correct columns.
If you use your own or modified .csv
file, you have to manually select the custom in the acording menu area.
The output file for the facial feature extraction is always a .csv
file with ;
as chosen separator.
The file will be created automatcially next to video file with the current timestamp.
The file only uses ascii symbols and contains a header row.
Each column in the header descirbes the extracted facial feature and gives a state about the extraction validity.
Each row represents a single frame with including the according frame id
inside the video
In the image below you can see such a extraction result:
-
frame: the actual
frame id
inside the video -
EAR2D6_l: the according
EAR2D6
score for the left eye at the according frame -
EAR2D6_r: the according
EAR2D6
score for the right eye at the according frame -
EAR2D6_valid: is
True
orFalse
, if the computation of the EAR score was valid eg. all landmarks were in the image, values are withing a logical range, etc. -
BS_Valid: is
True
orFalse
if the extraction of the facial blendshapes was valid eg. the face was fully visible.