The long-term objective of SIMILAR is to merge European research in the human-machine interface field
into a single European taskforce.
The immediate objectives are:
1) to integrate the traditionally separated research communities in visual interaction, interactive speech,
interactive haptics, learning, and human-computer interaction (HCI) across Europe;
2) to integrate national and regional research communities in the above fields in a larger community;
3) to promote standards, solutions and dissemination of knowledge in multimodal interaction and
4) to eventually lead to the establishment of a single ‘European virtual research centre in multimodal interaction
with strong industrial support.
Université Catholique de Louvain-la-Neuve, Belgium
Aristotle Univ. of Thessaloniki, Greece
Ecole Polytechnique Fédérale de Lausanne, Switzerland
Institut Eurécom, France
Faculté Polytechnique de Mons, Belgium CReATIVe-I3S Laboratory, University of Nice-Sophia Antipolis and CNRS
National Institute for R. and D. in Informatics-ICI Bucharest, Romania
Imperial College London
Institut National de Recherche en Informatique et en Automatique, France
Consiglio Nazionale delle Ricerche-ISTI, Italy
University of Linkoping, Sweden
Universidad de Valladolid-LPI-UVA, Spain
University of Southern Denmark-NISLAB, Denmark
University College Dublin, Ireland
University of Las Palmas Gran Canaria,Spain
University of Geneva-CVML, Switzerland
Universitat Politècnica de Catalunya, Spain Zentrum für Graphische Datenverarbeitung e.V., Germany
eNTERFACE workshops aim at establishing a tradition of collaborative,
localized research and development work by gathering, in a single
place, a team of leading professionals in multimodal interfaces
together with students (both graduate and undergraduate), to work on a
prespecified list of challenges, for 4 complete weeks.
is held on a yearly basis, in July-August, and organized around several
research Projects dealing with multimodal human-machine interfaces
design. It is thus radically different from traditional, scientific
workshops, in which only specialists meet for a few days to discuss
state-of-the art problems, but do not really work together.
The senior reseachers are mainly members of the SIMILAR Network of
Excellence, together with other university professors and industrial
and governmental researchers presently working in widely dispersed
locations. A small number (usually 7-9) of undergraduates are
also selected through an open Call for Participation, set on
outstanding academic promise.
About thirty graduate students familiar with the field are selected in accordance with their demonstrated performance.
Members of the ITI/AVRLab team were selected in both the 2005 and 2006
editions of the eNTERFACE summer workshops to contribute on the
following projects :
A Multimodal (Gesture+Speech) Interface for 3D Model Search and Retrieval Integrated in a Virtual Assembly Application :
development of a multimodal interface for content-based search of 3D
objects based on sketches. This user interface will integrate
graphical, gesture and speech modalities to aid the user in sketching
the outline of the 3D object he/she wants to search from a large
database. Finally, the system will be integrated in a virtual assembly
application, where the user will be able to assemble a machine from its
spare parts using only speech and specific gestures.
(July, August 2005, Mons, Belgium)
Multimodal tools and interfaces for the intercommunication between visually impaired and “deaf and mute” people :
building of a multimodal interface combining visual, aural and haptic
interaction with gesture-speech recognition, speech synthesis and sign
language recognition and synthesis, in order to enable the
communication of people exhibiting different kinds of disabilities.
(July-August 2006, Dubrovnik, Croatia)
Multimodal user interface for the communication of the disabled:
Upgrade of the treasure hunting game application that is jointly played by the blind and deaf-and-mute user. The
integration of the multimodal interfaces into a game application serves both as an entertainment and a pleasant education tool to
the users. The proposed application integrates haptics, audio, visual output and computer vision, sign language analysis and
synthesis, speech recognition and synthesis, in order to provide an interactive environment where the blind and deaf-and-mute
users can collaborate to play the treasure hunting game.
Game Demonstration Video Files