Article Preview
TopBackground
The field of human movements and gesture analysis has, for a long time now, attracted the interest of many researchers, choreographers and dancers. Thus, since the end of the last century, a significant corpus of work has been conducted relating movement perception with music (Fraisse, 1982).
Among the research community on this subject, there are works that stand out as important references on how video analysis technologies have provided interesting ways of movement-music interaction. Early works of composers Todd Winkler (Winkler, 1995) and Richard Povall (Povall, 1998), or the choreographer Robert Weschler work with Palindrome1. Also, Mark Coniglio continued development of his Isadora2 programming environment, plus the groundbreaking work Troika Ranch3 has done in interactive dance.
Other example of research in this field is the seminal work of Camurri, with several studies published, including an approach for the recognition of acted emotional states based on the analysis of body movement and gesture expressivity (Castellano, Villalba, & Camurri, 2007) and one of the most remarkable and recognized works, the EyesWeb software (Camurri et al., 2000).