Filters and Search 🔍
  • What makes a smiling face look happy? Visual saliency, distinctiveness, and affect

    GatedPeer-Reviewed29/11/2016

    Abstract: We investigated the relative contribution of (a) perceptual (eyes and mouth visual saliency), (b) conceptual or categorical (eye expression distinctiveness), and (c) affective (rated valence and arousal) factors, and (d) specific morphological facial features (Action Units; AUs), to the recognition of facial happiness. The face stimuli conveyed truly happy expressions with a smiling mouth […]

  • HuSIS: A Dedicated Space for Studying Human Interactions

    GatedPeer-Reviewed21/11/2016

    Abstract: The Human-Surrogate Interaction Space (HuSIS) consists of a dedicated physical space, structures, and components designed specifically for carrying out controlled studies related to human-surrogate interactions. This article discusses the primary factors considered in the HuSIS design and the benefits of the common data-collection and analysis framework for HuSIS research.

  • Automatic expression recognition and expertise prediction in Bharatnatyam

    GatedPeer-Reviewed03/11/2016

    Abstract: Bharatnatyam is an ancient Indian Classical Dance form consisting of complex postures and expressions. One of the main challenges in this dance form is to perform expression recognition and use the resulting data to predict the expertise of a test dancer. In this paper, expression recognition is carried out for the 6 basic expressions […]

  • Alexithymia, but Not Autism Spectrum Disorder, May Be Related to the Production of Emotional Facial Expressions

    GatedPeer-Reviewed01/11/2016

    Abstract: Background: A prominent diagnostic criterion of autism spectrum disorder (ASD) relates to the abnormal or diminished use of facial expressions. Yet little is known about the mechanisms that contribute to this feature of ASD. Methods: We showed children with and without ASD emotionally charged video clips in order to parse out individual differences in spontaneous production […]

  • Deep Multimodal Fusion for Persuasiveness Prediction

    Open AccessPeer-Reviewed01/11/2016University of Central Florida + 2

    Abstract: Persuasiveness is a high-level personality trait that quantifies the influence a speaker has on the beliefs, attitudes, intentions , motivations, and behavior of the audience. With social multimedia becoming an important channel in propagating ideas and opinions, analyzing persuasiveness is very important. In this work, we use the publicly available Persuasive Opinion Multimedia (POM) […]

  • Asynchronous Video Interviews vs. Face-To-Face Interviews for Communication Skill Measurement: A Systematic Study

    GatedPeer-Reviewed31/10/2016

    Abstract: Communication skill is an important social variable in employment interviews. As recent trends point to, increasingly asynchronous or interface-based video interviews are becoming popular. Also getting increasing interest is automatic hiring analysis, of which automatic communication skill prediction is one such task. In this context, a research gap that exists and which our paper […]

  • Automatic assessment of communication skill in interface-based employment interviews using audio-visual cues

    GatedPeer-Reviewed26/10/2016

    Abstract: Being an effective communicator plays a major role in employment interviews. In this paper, we provide a computational framework to automatically predict the communication skill of a person in an interface-based interview setting. The advantage of interface-based interview setting compared to that of a face-to-face setting is, the participants get assessed without any human […]

  • Objective, computerized video-based rating of blepharospasm severity

    GatedPeer-Reviewed21/10/2016Salk Institute for Biological Studies; Institute for Neural Computation

    Objective: To compare clinical rating scales of blepharospasm severity with involuntary eye closures measured automatically from patient videos with contemporary facial expression software. Methods: We evaluated video recordings of a standardized clinical examination from 50 patients with blepharospasm in the Dystonia Coalition’s Natural History and Biorepository study. Eye closures were measured on a frame-by-frame basis […]

  • Eye Tracking Architecture: A Pilot Study of Buildings in Boston

    GatedPeer-Reviewed23/09/2016

    Abstract: In a collaboration between architecture, interior design, and cognitive science, we conducted an eye tracking study at the Institute for Human Centered Design, a non-profit in Boston. Our thirtythree volunteer viewers, ages 18 to 80 and from various occupations, looked at 60 images on a computer screen for 15 seconds each. Half of the […]

  • Advanced Driver Monitoring for Assistance System (ADMAS) based on emotions

    GatedPeer-Reviewed17/09/2016Tecnológico de Monterrey

    Abstract: This work presents advances in research of emotions recognition by using facial expressions to be used in active security system focused in driver monitoring systems to provide efficient assistance through Advanced Driver Assistance Systems to drivers when poor driving performance is detected; researchers have called to this approach Advanced Driver Monitoring for Assistance Systems […]

Share Your Research

850+ universities worldwide with an iMotions human behavior lab 

73 of the top 100 highest ranked universities 

710+ published research papers using iMotions 

The authors of these publications have used iMotions as a software tool within their research.

“Software should be cited on the same basis as any other research product such as a paper or a book; that is, authors should cite the appropriate set of software products just as they cite the appropriate set of papers” (Katz et al., 2020).

Publications

Read publications made possible with iMotions

Blog

Get inspired and learn more from our expert content writers

Newsletter

A monthly close up of latest product and research news