Facial Action Coding System (FACS) – A Visual Guidebook

In this article, we have put together a visual guidebook to better showcase the power of the facial action coding system (FACS). All facial action units are presented with animations to give a first-hand understanding and valuable reference point in the future. So if you are working with facial expression analysis (FEA) we encourage you to bookmark this page – it could just save you a lot of work.

The Facial Action Coding System

The Facial Action Coding System (FACS) refers to a set of facial muscle movements that correspond to a displayed emotion. Originally created by Carl-Herman Hjortsjö with 23 facial motion units in 1970, it was subsequently developed further by Paul Ekman, and Wallace Friesen. The FACS as we know it today was first published in 1978, but was substantially updated in 2002.

Using FACS, we are able to determine the displayed emotion of a participant. This analysis of facial expressions is one of very few techniques available for assessing emotions in real-time (fEMG is another option). Other measures, such as interviews and psychometric tests, must be completed after a stimulus has been presented. This delay ultimately adds another barrier to measuring how a participant truly feels in direct response to a stimulus.

Researchers have for a long time been limited to manually coding video recordings of participants according to the action units described by the FACS. This process is now possible to complete with automatic facial expression analysis. This saves vast amounts of time and money, as scoring no longer requires analysis of each frame by a trained researcher – the software simply does the work for you.

Below we have listed the major action units that are used to determine emotions. Roll your mouse over the image to start the movement!

The 3 main benefits of facial coding

Facial coding offers several key benefits in the field of emotion analysis and human behavior research:

  1. Non-Intrusive Emotion Measurement: Facial coding allows for non-intrusive and natural emotion measurement, as it captures emotional expressions in real time without requiring participants to report their feelings verbally or in writing. This provides more authentic and unbiased insights into emotional responses.
  2. High Temporal Precision: Facial coding provides high temporal precision, enabling researchers to analyze micro-expressions and subtle changes in facial expressions, which may occur in fractions of a second. This level of detail is valuable for understanding the dynamics of emotional responses and their triggers.
  3. Objective and Quantifiable Data: Facial coding generates objective and quantifiable data, making it suitable for both research and commercial applications. By converting facial expressions into numerical data, researchers can conduct rigorous statistical analyses and track changes in emotions over time, contributing to a deeper understanding of human behavior and consumer preferences.

Main Action Units

Action UnitDescriptionFacial MuscleExample
 1Inner Brow RaiserFrontalis, pars medialisAU1 FACS
 2Outer Brow Raiser (unilateral, right side)Frontalis, pars lateralisAU2 right only FACS
 4Brow LowererDepressor Glabellae, Depressor Supercilli, Currugatorau4 brow lowerer
 5Upper Lid RaiserLevator palpebrae superiorisAU5 FACS
 6Cheek RaiserOrbicularis oculi, pars orbitalisAU6 cheek raiser
 7Lid TightenerOrbicularis oculi, pars palpebralisAU7 lid tightener
 9 (also shows slight AU4 and AU10)Nose Wrinkler Levator labii superioris alaquae nasiAU9 with 4+10
 10 (also shows slight AU25)Upper Lip Raiser Levator Labii Superioris, Caput infraorbitalis AU10 with 25
 11Nasolabial DeepenerZygomatic Minor AU11 - nasolabial deepener
 12Lip Corner PullerZygomatic MajorAU12
 13Cheek PufferLevator anguli oris (Caninus)AU13 cheek puffer
 14DimplerBuccinatorAU14 dimpler
 15Lip Corner DepressorDepressor anguli oris (Triangularis)AU15 FACS
 16 (with AU25)Lower Lip DepressorDepressor labii inferiorisAU16 with 25
 17Chin RaiserMentalisAU17 FACS guide
 18 (with slight AU22 and AU25)Lip PuckererIncisivii labii superioris and Incisivii labii inferiorisAU18 with 22A and 25A
 20Lip stretcherRisorius AU20 lip stretcher
 22 (with AU25)Lip FunnelerOrbicularis orisAU22 with 25 FACS
 23Lip TightenerOrbicularis orisAU23 lip tightener
 24Lip PressorOrbicularis orisAU24 image FACS guide
 25Lips partDepressor Labii, Relaxation of Mentalis (AU17), Orbicularis OrisAU25 lips part
 26 (with AU25)Jaw DropMasetter; Temporal and Internal Pterygoid relaxedAU26 with 25 FACS affectiva
 27Mouth StretchPterygoids, DigastricAU27 mouth stretcher
 28 (with AU26)Lip SuckOrbicularis orisAU28 with 26 FACS affectiva
 41Lid droopRelaxation of Levator Palpebrae SuperiorisAU41 lid droop
42SlitOrbicularis oculiAU42 slit
 43Eyes ClosedRelaxation of Levator Palpebrae SuperiorisAU43 eyes closed
 44SquintOrbicularis oculi, pars palpebralis AU44 squint
 45BlinkRelaxation of Levator Palpebrae and Contraction of Orbicularis Oculi, Pars Palpebralis. AU45 blink
 46WinkLevator palpebrae superioris; Orbicularis oculi, pars palpebralisAU46 wink

Free 42-page Facial Expression Analysis Guide

For Beginners and Intermediates

  • Get a thorough understanding of all aspects
  • Valuable facial expression analysis insights
  • Learn how to take your research to the next level

Head Movement Action Units

Action UnitDescriptionExample
 51Head Turn LeftAU51 head turn left
 52Head Turn RightAU52 head turn right
 53Head UpAU53 head up
 54Head DownAU54 head down
 55Head Tilt LeftAU55 head tilt left
 56Head Tilt RightAU56 head tilt right
 57Head ForwardAU57 head forward
 58Head Back AU58 head back

Interested in Human Behavior and Psychology?

Sign up to our newsletter to get the latest articles and research send to you

Eye Movement Action Units

Action UnitDescriptionExample
 61Eyes Turn LeftAU61-eyes-turn-left
 62Eyes Turn RightAU62 eyes right
 63Eyes UpAU63 eyes up
 64Eyes DownAU64 eyes down

Emotions and Action Units

The Action Units described above show the different movements of facial muscles. Certain combined movements of these facial muscles pertain to a displayed emotion. Emotion recognition is completed in iMotions using Affectiva, which uses the collection of certain action units to provide information about which emotion is being displayed. For example, happiness is calculated from the combination of action unit 6 (cheek raiser) and 12 (lip corner puller). A complete list of these combinations and the emotion that they relate to is shown below. The gifs on the right are shown in the same order that the action units listed.

EmotionAction UnitsDescriptionExamples
Happiness / Joy6 + 12Cheek Raiser, Lip Corner PullerAU6 cheek raiser

 

AU12 lip corner puller

Sadness1 + 4 + 15Inner Brow Raiser, Brow Lowerer, Lip Corner DepressorAU1 inner brow raiser

 

au4 brow lowerer

AU15 lip corner depressor

Surprise1 + 2 + 5 + 26Inner Brow Raiser, Outer Brow Raiser, Upper Lid Raiser, Jaw DropAU1 inner brow raiser

 

au2 outer brow raiser

au5 upper lid raiser

AU26 jaw drop

Fear1 + 2 + 4 + 5 + 7 + 20 + 26Inner Brow Raiser, Outer Brow Raiser, Brow Lowerer, Upper Lid Raiser, Lid Tightener, Lip Stretcher, Jaw DropAU1 inner brow raiser

 

au2 outer brow raiser

au4 brow lowerer

au5 upper lid raiser

AU7 lid tightener

AU20 lip stretcher

AU26 jaw drop

Anger4 + 5 + 7 + 23Brow Lowerer, Upper Lid Raiser, Lid Tightener, Lip Tightenerau4 brow lowerer

 

au5 upper lid raiser

AU7 lid tightener

AU23 lip tightener

Disgust9 + 15 + 16Nose Wrinkler, Lip Corner Depressor, Lower Lip DepressorAU9 nose wrinkler

 

AU15 lip corner depressor

AU16 lower lip depressor

Contempt12 + 14 (on one side of the face)Lip Corner Puller, DimplerAU12 lip corner puller

 

AU14 dimpler

Putting it all together

When measuring facial expressions within iMotions, the stimuli are paired automatically to the FACS analysis, allowing you to pinpoint the exact moment that the stimulus triggered a certain emotion. The FACS is also graded on a scale of intensity, which gives a measure of how strongly the emotion is displayed. These measurements can also be synchronized with recordings of galvanic skin response, which provides a measure of arousal. With this information combined, it’s possible to start drawing conclusions about how strongly an individual felt, and what those emotions consisted of, in response to a set stimulus.

The screenshot below shows how the facial expression data is displayed while a participant watches an advertisement.

participant in an imotions experiment where his facial expressions are registered and analyzed
Facial Expression data displayed while participant watches advertisement in iMotions

If we zoom in, we can see the intensity of the displayed emotion. There are five emotions displayed in the image below, however iMotions provides a measure of the seven central emotions (shown in the table above), alongside, and in conjunction with measurements of action units.

facial expression analysis in imotions software

Best Facial Action Coding System Software

To choose the best facial action coding software, start by defining your research objectives and budget. Assess essential features like real-time capabilities, interoperability, and ease of use. Research the software’s accuracy, user community, and vendor reputation, and consider trying demo or trial versions when available. Ultimately, select software that aligns with your specific research needs, provides necessary support and training, and fits within your budget constraints.

Integrated into iMotions is the Affectiva Emotion AI, which is a highly specialized emotion recognition technology. The Emotion AI platform is designed for understanding and analyzing human emotions from facial expressions. Here are some key features and details:

  • Emotion Recognition: Affectiva’s Emotion AI uses deep learning and computer vision to detect and classify facial expressions in real-time. It can identify a range of emotions such as happiness, surprise, anger, and more.
  • Emotion Metrics: The software provides various metrics related to emotions, including intensity, valence (positive or negative emotion), and engagement levels. This data can be useful in market research, advertising, and user experience analysis.
  • Real-World Applications: Affectiva’s Emotion AI has been used in a wide range of industries, including market research, automotive (for driver monitoring and safety), healthcare (for patient feedback and engagement monitoring), and human-computer interaction (for improving user experiences in digital applications).
  • Cloud-Based Solution: Affectiva offers cloud-based solutions that can be integrated into various applications and platforms, making it accessible for developers and researchers.

Facial Expression Analysis

Decode emotional expressions as they happen.

Find Out More

Frequently Asked Question

How accurate is the facial action coding system?

The accuracy of the Facial Action Coding System (FACS) is highly dependent on the training and expertise of the individuals using it. When administered by experienced FACS coders under controlled conditions, FACS can achieve a high level of accuracy in coding facial expressions. However, accuracy can vary based on factors like the complexity of emotions and the quality of data collection conditions.

Is training required to use the Facial Action Coding System (FACS)?

Both yes and no, if you want to become an expert, then training is highly recommended and typically encouraged to use the Facial Action Coding System (FACS) effectively at a high level. If you use software such as Affectiva/iMotions where the software does the coding of the visual stimuli, you typically do not need to be an expert in FACS in order to use it.

Can Facial Coding be used for emotion recognition in real-time applications?

Yes, facial coding can be used for real-time emotion recognition in applications such as human-computer interaction, market research, and affective computing. Advances in computer vision and facial recognition technology have enabled the development of real-time facial coding systems that can analyze facial expressions as they occur. These systems use machine learning algorithms to detect and classify emotions based on the movements and patterns of facial muscles in real-time video streams. Real-time emotion recognition through facial coding has applications in areas like user experience design, virtual reality, and customer feedback analysis.


I hope this explanation of action units and FACS has been helpful, and informative. If you’d like to learn even more about facial expressions, then we also have a free pocket guide that you can download for free below!

Keep up with the latest in human behavior research and technological developments

Free 42-page Facial Expression Analysis Guide

For Beginners and Intermediates

  • Get a thorough understanding of all aspects
  • Valuable facial expression analysis insights
  • Learn how to take your research to the next level

iMotions Software Products

Full-scale human behavior
research solutions

A software research suite that integrates an ecosystem of virtually any biosensor technology to provide the most robust data and insights available.

iMotions Lab

Trusted by researchers, iMotions Lab lets you collect and analyze data from biosensors and create studies with any stimuli

iMotions Online

Webcam Eye Tracking and Facial Expression Analysis for deploying studies quickly and easily around the world, supported by all major browsers

About the author


See what is next in human behavior research

Follow our newsletter to get the latest insights and events send to your inbox.