Q     | In the Readme file, I have found the description of the data. |
Each video have four annotations: [video name]_drowsiness.txt : 0 means Stillness and 1 means Drowsy [video name]_head.txt : 0 means Stillness, 1 means Nodding and 2 means Looking aside. [video name]_mouth.txt : 0 means Stillness and 1 means Yawning and 2 means Talking & Laughing. [video name]_eye.txt : 0 means Stillness and 1 means Sleepy-eyes. Do you consider Yawning, Sleepy-eyes and Nodding as Drowsy behavior? If not, do we need to classify all them (Drowsy, Nodding, Yawn, Sleepy-eyes) completely ? |
A     | Thanks for your question. Following is our answer: |
1. We consider Yawning, Sleepy-eyes and Nodding as Drowsy behavior while Looking aside and Talking & Laughing as Non-drowsy behavior. |
|
2. The required results of challenge session is only the classification of drowsy and non-drowsy. You don't need to submit the detail behaviors, such as Yawning, Sleepy-eyes, Nodding, Looking aside or Talking & Laughing. |