Skip to content

Open Face 1 integration

Philippe Gauthier edited this page Apr 8, 2021 · 3 revisions

AUParserFileReader

To reproduce the facial expressions and head movements extracted via OpenFace 1, it is possible use the AUParserFileReader module. It takes as input a csv file containing the timestamp, the values of all AUs, gaze, head pose and head rotation. The output is a video that reproduce frame by frame the values contained in the csv file (example csv: https://github.com/isir/greta/blob/master/bin/Examples/AUParser/AUParser_example.csv).

Modules Connections

The module has to be connected to the SimpleAUPerformer to send the AUs to the agent and MPEG4Animatable to send the BAP values for the Head rotation. One more connection that has to be create is from the ParserCaptureControllerAU to the AUParserFileReader. (see figure below)

drawing

How it works

  1. Pushing the button Open it is possible to select the directory where are stored the csv files.

drawing

  1. Then pushing the button Send, the list of all variables in the csv files will appear in the window below (Headers).

drawing

  1. Select the variables you want to reproduce, chekcing the related checkbox (or just push the button Select All).

  2. Push the button Set. After that you are ready to record.

  3. To start the recording go to the ParserCaptureControllerAU interface and push the button Record all files

drawing

Some info:

  • the frame rate for the video is computed taking the difference between the timestamp value of the first two rows
  • keep the "Real Time" checkbox selected if you want the duration of the video be as long as the real time duration of the file

Getting started with Greta

Greta Architecture

Quick start

Advanced

Functionalities

Core functionality

Auxiliary functionalities

Preview functionality

Nothing to show here

Previous functionality (possibly it still works, but not supported anymore)

Clone this wiki locally