Nothing Special   »   [go: up one dir, main page]

Next Article in Journal
Compressive Sensing Hyperspectral Imaging by Spectral Multiplexing with Liquid Crystal
Next Article in Special Issue
Enhancement and Segmentation Workflow for the Developing Zebrafish Vasculature
Previous Article in Journal
Recent Trends in Compressive Raman Spectroscopy Using DMD-Based Binary Detection
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

What’s in a Smile? Initial Analyses of Dynamic Changes in Facial Shape and Appearance †

1
School of Dentistry, Cardiff University, Heath Park, Cardiff CF14 4XY, UK
2
School of Computer Science and Informatics, Cardiff University, Cardiff CF24 3AA, UK
3
Research Unit of Oral Health Sciences, Faculty of Medicine, University of Oulu, FI-90014 Oulu, Finland
4
Medical Research Center Oulu (MRC Oulu), Oulu University Hospital, FI-90014 Oulu, Finland
*
Author to whom correspondence should be addressed.
This article is an extended version of our paper published in Farnell, D.J.J.; Galloway, J.; Zhurov, A.; Richmond, S.; Pirttiniemi, P.; Lähdesmäki, R. What’s in a Smile? Initial Results of Multilevel Principal Components Analysis of Facial Shape and Image Texture. In Medical Image Understanding and Analysis; Springer: Cham, Switzerland, 2018; Volume 894, pp. 177–188.
J. Imaging 2019, 5(1), 2; https://doi.org/10.3390/jimaging5010002
Submission received: 15 November 2018 / Revised: 13 December 2018 / Accepted: 18 December 2018 / Published: 21 December 2018
(This article belongs to the Special Issue Medical Image Understanding and Analysis 2018)
Figure 1
<p>Flowchart illustrating the multilevel model of facial shape for dataset 1.</p> ">
Figure 2
<p>Illustration of the 21 landmark points for dataset 1 ((1) Glabella (g); (2) Nasion (n); (3) Endocanthion left (enl); (4) Endocanthion right (enr); (5) Exocanthion left (exl); (6) Exocanthion right (exr); (7) Palpebrale superius left (psl); (8) Palpebrale superius right (psr); (9) Palpebrale inferius left (pil); (10) Palpebrale in-ferius right (pir); (11) Pronasale (prn); (12) Subnasale (sn); (13) Alare left (all); (14) Alare right (alr); (15) Labiale superius (ls); (16) Crista philtri left (cphl); (17) Crista philtri right (cphr); (18) Labiale inferius (li); (19) Cheilion left (chl); (20) Cheilion right (chr); (21) Pogonion (pg)).</p> ">
Figure 3
<p>Schematic illustration of a time series of smile amplitudes from Equation (1) for 3D shape data in dataset 2. Including rest phases, seven phases can be identified manually: rest pre-smile, onset acceleration, onset deceleration, apex, offset acceleration, offset deceleration, and rest post-smile.</p> ">
Figure 4
<p>Eigenvalues for dataset 1 from single-level PCA and from mPCA level 1 (biological sex), level 2 (between-subject variation), and level 3 (within-subject variation: facial expression). (<b>a</b>) Shape data; (<b>b</b>) Image texture data (All shapes have been scaled so that the average point-to-centroid distance equals 1.).</p> ">
Figure 5
<p>Modes of variation for shape for dataset 1 for the first three modes from single-level PCA in the upper set of images: (<b>a</b>) = mode 1; (<b>b</b>) = mode 2; (<b>c</b>) = mode 3. The first modes from levels 1 to 3 mPCA in the bottom set of images: (<b>d</b>) = mode 1, level 1 (biological sex); (<b>e</b>) = mode 1, level 2 (between subjects); (<b>f</b>) = mode 1, level 3 (facial expression).</p> ">
Figure 6
<p>Modes of variation for image texture for dataset 1 for the first three modes ((<b>a</b>) = mode 1; (<b>b</b>) = mode 2; (<b>c</b>) = mode 3) from single-level PCA in the left-hand set of images, and the first modes from levels 1 to 3 ((<b>a</b>) = level 1; (<b>b</b>) = level 2; (<b>c</b>) = level 3) from mPCA in the right-hand set of images. Note that for each set of three images: left image = mean − SD; middle image = mean; right image = mean + SD.</p> ">
Figure 7
<p>Standardized component scores with respect to shape for dataset 1: (<b>a</b>) Components 1 and 2 for single-level PCA; (<b>b</b>) Components 1 and 3 for single-level PCA; (<b>c</b>) Component 1 for level 1 (biological sex) for mPCA; (<b>d</b>) Components 1 and 2 for level 3 (facial expression) for mPCA.</p> ">
Figure 8
<p>Standardized component scores with respect to image texture for dataset 1: (<b>a</b>) Components 1 and 2 for single-level PCA; (<b>b</b>) Components 1 and 3 for single-level PCA; (<b>c</b>) Component 1 for level 1 (biological sex) for mPCA; (<b>d</b>) Components 1 and 2 for level 3 (facial expression) for mPCA.</p> ">
Figure 8 Cont.
<p>Standardized component scores with respect to image texture for dataset 1: (<b>a</b>) Components 1 and 2 for single-level PCA; (<b>b</b>) Components 1 and 3 for single-level PCA; (<b>c</b>) Component 1 for level 1 (biological sex) for mPCA; (<b>d</b>) Components 1 and 2 for level 3 (facial expression) for mPCA.</p> ">
Figure 9
<p>Eigenvalues for dataset 2 (shape data only) from single-level PCA and from mPCA level 1 (between-subject variation), level 2 (variation between smile phases), and level 3 (variation within smile phases).</p> ">
Figure 10
<p>Modes of variation for dataset 2 from mPCA: (<b>a</b>) level 1 (between-subject variation), mode 1, coronal plane; (<b>b</b>); level 1 (between-subject variation), mode 1, transverse plane; (<b>c</b>) level 2 (variation between smile phases), mode 1, coronal plane; (<b>d</b>) level 2 (variation between smile phases), mode 1, transverse plane.</p> ">
Figure 11
<p>Centroids over smile phases for standardized component scores with respect to shape for dataset 2: (<b>a</b>) Components 1 and 2 for single-level PCA; (<b>b</b>) Components 1 and 2 at level 2 (variation between smile phases) for mPCA.</p> ">
Review Reports Versions Notes

Abstract

:
Single-level principal component analysis (PCA) and multi-level PCA (mPCA) methods are applied here to a set of (2D frontal) facial images from a group of 80 Finnish subjects (34 male; 46 female) with two different facial expressions (smiling and neutral) per subject. Inspection of eigenvalues gives insight into the importance of different factors affecting shapes, including: biological sex, facial expression (neutral versus smiling), and all other variations. Biological sex and facial expression are shown to be reflected in those components at appropriate levels of the mPCA model. Dynamic 3D shape data for all phases of a smile made up a second dataset sampled from 60 adult British subjects (31 male; 29 female). Modes of variation reflected the act of smiling at the correct level of the mPCA model. Seven phases of the dynamic smiles are identified: rest pre-smile, onset 1 (acceleration), onset 2 (deceleration), apex, offset 1 (acceleration), offset 2 (deceleration), and rest post-smile. A clear cycle is observed in standardized scores at an appropriate level for mPCA and in single-level PCA. mPCA can be used to study static shapes and images, as well as dynamic changes in shape. It gave us much insight into the question “what’s in a smile?”.

1. Introduction

Human faces are central to our identity and they are important in expressing emotion. The act of smiling is important in this context and the exploration of facial changes and dynamics during the act of smiling [1,2] is an ongoing topic of investigation in fields of research in orthodontics and prosthodontics, both of which aim to improve the function and appearance of dentition. Aesthetics (e.g., of smiles [3]) are therefore an important aspect of these fields. Much research into the “science of a smile” also focuses on the effects of aging and biological sex on the shape and appearance [4,5] and also the dynamics [6,7,8] of smiling. Recent investigations have been greatly enhanced by the use of three-dimensional (3D) imaging techniques [9,10,11,12,13] that allow both static and dynamic imaging of the face. Clinically, this research has led to improved understanding of orthognathic surgery [9], malocclusion [10], associations between facial morphology and cardiometabolic risk [11], lip shape during speech [12], facial asymmetry [13], and sleep apnea [14] (to name but a few examples). Clearly also, facial simulation is of much interest for human–computer interfaces (see, e.g., [15,16,17]). The role of genetic factors on facial shape has also been the subject of much recent attention [18,19,20,21] and many factors (sex, age, and genetic factors) across a set of subjects can affect the shape and dynamics of facial and/or mouth shape.
In this article, we wish to explore the question “what’s in a smile?” by using multilevel principal components analysis (mPCA) to adjust for covariates such as natural mouth shape and/or sex. Indeed, the mPCA approach has previously been shown [22,23,24,25,26] to provide a simple and straightforward method of modeling shape. The mPCA approach is potentially also of much use in active shape models (ASMs) [27,28,29,30,31] and active appearance models (AAMs) [32,33,34,35,36,37] (see also [38]). We remark that one such previous application of mPCA to ASMs related to the segmentation of the human spine [22]. The authors stated that their results showed that “such a modelization offers more flexibility and allows deformations that classical statistical models can simply not generate”. They also noted that “the idea is to decompose the data into a within-individual and a between-individual component”. Hence, we remark again that mPCA provides one method of adjusting for external and/or confounding factors or covariates that can strongly affect shapes (or images). Other recent applications of mPCA [23,24,25] allowed us to determine the relative importance of biological sex and ethnicity on facial shape by examination of eigenvalues. Modes of variation made sense because changes in shape were seen to correspond to biological sex and ethnicity at the correct levels of the model and no ‘mixing’ of these effects was observed. Finally, principal component ‘scores’ also showed strong clustering, which were again at the correct levels of the mPCA model.
Another method that allows us to investigate the effects of covariates on facial shape to be modeled is called bootstrapped response-based imputation modeling (BRIM) [20,21]. The effects of covariates such as sex and genomic ancestry on facial shape were summarized in [20] as response-based imputed predictor (RIP) variables and the independent effects of particular alleles on facial features were uncovered. Indeed, the importance of modeling the effects of covariates in images is also becoming increasingly recognized, e.g., such as in variational auto-encoders (see, e.g., [39,40,41]) in which the effects of covariates are modeled as latent variables sandwiched between encoding (convolution) and decoding (deconvolution) layers. However, the topic of variational auto-encoders lies beyond the scope of this article. Linear discriminant functions have also been used previously (see, e.g., [42,43]) to explore groupings in the subject population for image data.
The work presented here is also an expansion of [26] that extended the mPCA approach from shape data also to include image data, where a set of (frontal) facial images from a group of 80 Finnish subjects (34 male; 46 female) each for two different facial expressions (smiling and neutral) were considered. A three-level model illustrated by Figure 1 was constructed that contains biological sex, facial expression, and ‘between-subject variation’ at different levels of the model and we use this model again here for this dataset. However, we also compare results of mPCA to those results of single-level PCA, which was not carried out in [26]. Furthermore, the dynamic aspects of a smile are considered here in a new (time series) dataset of 3D mouth shape captured during all phases of a smile, which was also not considered in [26]. We present here firstly the subject characteristics and details of image capture and preprocessing. We then consider the mathematical detail of the mPCA method. We then present our results for both datasets. Finally, we offer a discussion of these results in the concluding section.

2. Materials and Methods

2.1. Image Capture, Preprocessing, and Subject Characteristics

Dataset 1 consisted of photographic images of the 80 adult Finnish subjects (34 female; 46 male) who were captured with two expressions (neutral and smiling). Patients were from the Northern Finland Birth Cohort NFBC-66 (http://www.oulu.fi/nfbc/) and all patients were 46 years old when the images were collected. The Ethical Committee of the Northern Ostrobothnia Hospital District, Oulu University Hospital, Oulu, Finland has approved this study. Twenty-one reliable facial landmarks, illustrated by Figure 2, were positioned manually for each image and these points are used here in the analysis of shape. Preprocessing of the shapes included centering, alignment, and adjustment of overall scale only. Preprocessing of image texture [26] also included the definition of a region of interest (ROI) around edges of the face (7339 pixels) and the overall illumination of the grayscale images was standardized.
Dataset 2 consisted of 3D video shape data during all phases of a smile, where 13 points placed (and tracked) along the outer boundary of mouth. Subjects were 60 adult staff and students at Cardiff University (31 male and 29 female). The number of frames in the video was between approximately 100 and 250 frames during all phases of a smile for each subject. In these initial calculations, preprocessing consisted of centering the 3D shapes only. The (normalized) smile amplitude was defined by using the following equation.
Smile   Amplitude = Distance   between   the   left   and   right   chelion   at   time   t 2 × Distance   between   the   left   and   right   chelion   at   time   t = 0
Phases of the smile (rest pre-smile, onset acceleration, onset deceleration, apex, offset acceleration, offset deceleration, and rest post-smile) were identified manually by inspection of time series of smile amplitudes for each subject individually. This is illustrated schematically in Figure 3. Acceleration and deceleration phases are divided by points of inflexion in the amplitude. Inspection also of the (smoothed) first and second derivatives of the smile amplitude with respect to time allowed accurate estimation of the boundaries between these phases.

2.2. Multilevel Principal Components Analysis (mPCA)

ASMs [26,27,28,29,30] are statistical models of shape only and AAMs [32,33,34,35,36,37] are statistical models of both shape and appearance. The term ‘image texture’ is taken to refer to the pattern of intensities or colors across an image (or image patch) as in AAMs and we adopt this usage here. For ASMs, features may be segmented from an image by firstly forming the shape model over some ‘training’ set of shapes. Single-level principal component analysis (PCA) may be used to define the distributions of points or intensities (and/or color) at given pixel positions, respectively. For single-level PCA of shape, the mean shape vector (averaged over all N subjects) is given by z ¯ and a covariance matrix, C, is then found by evaluating
C k 1 , k 2 = 1 N 1 i = 1 N ( z i k 1 z ¯ i k 1 ) ( z i k 2 z ¯ i k 2 ) ,
where k 1 and k 2 indicate elements of this covariance matrix. We find the eigenvalues λ l and (orthonormal) eigenvectors u l of this matrix. All eigenvalues are non-negative, real numbers because covariance matrices are symmetric and positive semi-definite. For PCA, one ranks all of the eigenvalues λ l into descending order and we retain the first l 1 components in the model. Any new shape is then modeled by
z m o d e l = z ¯ + l = 1 l 1 a l u l ,
The coefficients, a l , for a fit of the model to a new shape vector, z , are found readily by using a scalar product with respect to the set of orthonormal eigenvectors u l , i.e.,
a l = u l ( z z ¯ ) .
A similar process is carried out for image texture or appearance [32,33,34,35,36,37]. For ASMs, features of interest in a new image can then be segmented by obtaining a trial shape from an image (e.g., defined along strong edges), which is then projected into the shape model in order to find an improved estimate that is consistent with the model via some constraint on the coefficients, a l . This process is iterated until convergence and so the final segmentation never ‘strays too far’ from a plausible solution with respect to the underlying shape model. In this article, we are concerned only with the modeling of shape and appearance and we do not carry out the ‘active’ image searches. However, the methods presented here could be extended to such image searches via ASMs or AAMs, although this is not the primary focus of this article.
Multilevel PCA (mPCA) allows us to isolate the effects of various influences on shape or image texture at different levels of the model. For the case of the act of “smiling”, this allows us to adjust for each subjects’ individual shape or appearance (and biological sex also for dataset 1) in order can get a clearer picture of these general changes due to a primary factor (here, facial expression due to smiling). This is illustrated schematically in Figure 1 for dataset 1. Multiple levels are used in mPCA to model the data and covariance matrices are formed at each level. For the model for dataset 1 illustrated by Figure 1, the covariance matrix at level 3 is formed with respect to the two expressions (neutral smiling) for each subject and then these covariance matrices are averaged over all 80 subjects. By contrast, the covariance matrix at level 2 is formed with respect to shapes or image texture averaged over the two expressions for each of the 80 subjects. Covariance matrices are formed for males and females separately and then they are averaged over two sexes to find the final covariance matrix at level 2 of this model. Finally, the covariance matrix at level 2 is formed with respect to shapes or image texture sex only, i.e., averaged over all subjects and expressions for the two sex groups separately. The number of shapes or images equals 2 at this level of the model and so the rank of this matrix is 1. Clearly, any restriction on the rank of the covariance matrix is a limitation of the mPCA model, although other multilevel methods (such as multilevel Bayesian approaches) ought not to be as strongly constrained as mPCA. An exploration of these topics will form the contents of future research. A three-level mPCA model was also used in this initial exploration for dataset 2: level 1, “between subject” due to natural face shape not attributed to smiling; level 2, “between smile phases” variation due to differences between seven different phases of a smile; level 3, “within smile phases” variation due to residual differences within the different phases of a smile. Hence, we explicitly model subjective variation across participants at specific levels of the mPCA model for both datasets.
mPCA then uses PCA with respect to these covariance matrices at each of the three levels separately. The l-th eigenvalue at level 1 is denoted by λ l 1 with associated eigenvector u l 1 , whereas the l-th eigenvalue at level 2 is denoted by λ l 2 with associated eigenvector is denoted by u l 2 , and so on. We rank all of the eigenvalues into descending order at each level of the model separately, and then we retain the first l 1 , l 2 , and l 3 eigenvectors of largest magnitude for all three levels, respectively. Any new shape is modeled by
z m o d e l = z ¯ + l = 1 l 1 a l 1 u l 1 + l = 1 l 2 a l 2 u l 2 + l = 1 l 3 a l 3 u l 3 ,
where z ¯ is now the ‘grand mean’. The coefficients { a l 1 } , { a l 2 } , and { a l 3 } (also referred to here as ‘component scores’) are determined for any new shape, z , by using a global optimization procedure in MATLAB R2017 with respect to the cost function
Δ = k = 1 ( z k z k m o d e l ) 2 = k = 1 ( z k z ¯ k l = 1 l 1 a l 1 u l k 1 l = 1 l 2 a l 2 u l k 2 l = 1 l 3 a l 3 u l k 3 ) 2 .
note again that z k is the k-th element of the shape vector z , and u l k 1 indicates the k-th element of the l-th eigenvector at level 1, etc. Another method of obtaining a solution is to iterate the following equations directly,
a l α a l α κ Δ a l α ,
(where α = 1, 2, 3, here and for all values l appropriately) for our three-level models from some ‘starting point’ (often taken to be the average shape, i.e., all coefficients are zero) until convergence. An appropriate choice of κ is found to be κ = 0.01. It is straightforward to see that
Δ a l α = 2 k u l k α ( z k z ¯ k m = 1 l 1 a m 1 u m k 1 m = 1 l 2 a m 2 u m k 2 m = 1 l 3 a m 3 u m k 3 ) .
Note that this approach finds identical solutions to that provided by MATLAB. Finally, standardized coefficients may be found by dividing the { a l } coefficients by the square root of the corresponding eigenvalue λ l for single-level PCA and by dividing the { a l 1 } coefficients by the square root of the corresponding eigenvalue λ l 1 at level 1 for mPCA (and similarly for the other levels). These standardized coefficients or ‘scores’ are useful in exploring clustering that may occur by biological sex, facial expression, and so on.

3. Results

Eigenvalues for both shape and also image texture via mPCA are shown in Figure 4 for dataset 1. The results for mPCA demonstrated a single non-zero eigenvalue for the level 1 (biological sex). A single large eigenvalue for the level 3 (facial expression) for mPCA occurs for shape and also for image texture, although many non-zero (albeit of much smaller magnitude) eigenvalues occur for image texture. Level 2 (between-subject variation) via mPCA tends to have the largest number of non-zero eigenvalues for both shape and image texture. The first two eigenvalues are (relatively) large at level 2, mPCA for image texture. mPCA results suggest that biological sex seems be the least important for this group of subjects for both shape and texture, although caution needs to be exercised as the rank of the matrix is 1 at this level for both shape and texture. Results for the eigenvalues from single-level PCA are of comparable magnitude to those results of mPCA, as one would expect, and they follow a very similar pattern. Inspection of these results for the eigenvalues tell us broadly that facial expression and natural facial shape (not dependent on sex or expression) are strong influences on facial shapes in the dataset. Biological sex was found to be a weaker effect comparatively, especially for shape, although again caution needs to be exercised in interpreting eigenvalues at this level for mPCA.
Modes of variation of shape for dataset 1 are presented in Figure 5. The first mode at level 3 (facial expression) for mPCA and mode 1 in single-level PCA both capture changes in facial expression (i.e., neutral to smiling and vice versa). Changes in mouth shape in Figure 5 can be seen that relate clearly to the act of smiling in both graphs. For example, obvious effects such as widening of the mouth, corners of the mouth raised slightly, exposure of teeth can be clearly seen. However, subtle effects such as narrowing of the eyes [44] and a slight widening at bottom of nose during smile are also seen clearly, especially for mPCA. Eyes become further apart (relatively) and the mouth becomes wider for the first mode at level 1 (biological sex) for mPCA in Figure 4. All shapes have been scaled so that the average point-to-centroid distance is equal to 1 and so this result makes sense because men have generally thinner faces than women on average [45,46]. This first mode via mPCA at level 3 probably corresponds to mode 3 or mode 2 (or a combination of both) in single-level PCA. However, modes 2 and 3 from single-level PCA are harder to interpret, and one can never preclude the possibility of mixing of different influences (e.g., sex, expression, etc.) in modes in single-level PCA. By contrast, mPCA should focus more clearly on individual influences because they are modeled at different levels of the mPCA model. The first mode at level 2 (between-subject variation) for mPCA in Figure 5 (middle row) corresponds to changes in the relative thinness/width of the face (presumably) that can occur irrespective of sex.
Modes of variation for image texture for dataset 1 are presented in Figure 6. The first modes at each level are relatively straightforward to understand for mPCA. We see that mode 1 for level 1 (biological sex) mPCA does indeed correspond to changes in appearance due to biological sex (e.g., females tend to have more prominent eyes and cheeks [45,46]), as required. Mode 1 for level 2 (between-subject) mPCA corresponds to residual changes due to left/right position (possibly) and also illumination, although this mode is slightly harder to interpret. Mode 1 for level 3 (facial expression) mPCA corresponds to changes due to the act of smiling (i.e., mean − SD = not smiling, mean = half smile, and mean + SD = full smile). We see clear evidence of a smile that exposes the teeth in this mode. Furthermore, subtle effects are seen for mPCA at this level such as increased prominence of the cheeks, increased nose width, and narrowing of the eyes [44]. The first three modes for single-level PCA are also relatively straightforward to interpret, although arguably less so than for the first mode at each level from mPCA. For example, mode 1 possibly corresponds to residual changes in illumination and/or also to slight changes to the nose and prominence of the cheeks, which might be associated with biological sex [45,46]. Modes 2 and 3 correspond clearly to changes due to the act of smiling.
Results for the standardized component ‘scores’ for mPCA for shape are shown in Figure 7. Results for component 1 for single-level PCA demonstrate differences due to facial expression clearly because the centroids are strongly separated between smiling and neutral expressions. By contrast, component 2 for single-level PCA does not seem to reflect changes due to either facial expression or biological sex very strongly. Finally, component 3 for single-level PCA reflects differences due to biological sex (albeit mildly), as there is some separation in the centroids between males and females. The centroids in Figure 7 at level 1 (biological sex) for mPCA are strongly separated by biological sex, although not by facial expression. Hence, strong clustering by biological sex (alone) is observed for shape at level 1 (biological sex) for mPCA. The centroids in Figure 7 at level 3 (facial expression) for mPCA are strongly separated by facial expression (neutral, smiling), although not by biological sex. Strong clustering by facial expression (alone) is therefore observed at level 3 (facial expression) for mPCA, also as required. Strong clustering by facial expression or biological sex is not seen at level 2 (between-subject variation) mPCA (not shown here), i.e., all centroids by biological sex and facial expression are congruent on the origin.
Results for the standardized component ‘scores’ for mPCA for image texture are shown in Figure 8. Component 1 for single-level PCA reflects differences due to biological sex and components 2 and 3 reflect changes due to facial expression. Again, level 1 for mPCA clearly reflects differences by biological sex and level 3 reflects differences by expression (neutral or smiling). Again, strong clustering by facial expression or biological sex is not seen at level 2 (between-subject variation) for mPCA also for image texture (not shown here), i.e., all centroids by biological sex and facial expression are congruent on the origin.
Eigenvalues via single-level PCA and mPCA for dataset 2 (‘dynamic’ 3D shape data) are shown in Figure 9. A single large eigenvalue for the level 2 (facial expression/smiling) occurs for mPCA, although the following eigenvalues are relatively larger than for the analysis of shapes for dataset 1. All phases of a smile are captured here in the 3D video shape data in dataset 2 and this result is to be expected. Variation at level 1 (between-subject variation) for mPCA tends to have larger eigenvalues compared to those for levels 2 and 3 (facial expression/smiling). Level 3 eigenvalues via mPCA are found to be minimal, thus indicating that residual variations within each smile phase are small. Finally, results of single-level PCA are of similar magnitude and follow a similar pattern to those eigenvalues from mPCA.
Modes of variation of shape for dataset 2 are presented in Figure 10. Results for the first mode at level 1 (between-subject variation) via mPCA in the coronal plane correspond to changes between upturned and downturned lip shape, which is consistent with changes due to subjects’ natural lip shape. Results for the first mode at level 1, mPCA in the transverse plane appear to indicate changes in the prominence of the upper and lower lips. By contrast, results for the first mode of variation at level 2 (i.e., between smile phases level) for mPCA correspond to increased mouth size and a strong drawing back (and slight upturn) of the corners of the mouth, which is consistent with the act of smiling. Mode 1 from single-level PCA is broadly similar to mode 1 at level 1 (between-subject variation) via mPCA, whereas mode 2 from single-level PCA is broadly similar to mode 1 at level 2 (between smile phases) via mPCA. Results for the modes of variation via single-level PCA for dataset 2 are therefore not presented here.
Standardized component scores from both single-level PCA and mPCA at level 2 (variation between smile phases) with respect to shape for dataset 2 are shown in Figure 11. Very little difference between centroids divided by smile phase is seen at levels 1 or 3 for mPCA (not shown here). The centroids of component scores at level 2 via mPCA are clearly separated in Figure 11 with respect to the seven phases of a smile (i.e., rest pre-smile, onset 1 (acceleration), onset 2 (deceleration), apex, offset 1 (acceleration), offset 2 (deceleration), and rest post-smile). Indeed, we see clear evidence of a cycle in these centroids in Figure 11 over all of these smile phases for both single-level PCA and at level 2 for mPCA. These results are strong evidence that seven phases of a smile do indeed exist.

4. Discussion

We have shown in this article that mPCA provides a viable method of accounting for groupings in our population subject set and/or for adjusting for potential confounding covariates. For example, natural face or lip shape was represented at one level of the mPCA model and shapes changes due to the act of smiling at another level (or levels) of the model. By capturing these different sources of variation we represented at different levels of the model, we are able to isolate those changes in expression due to smiling that are consistent over the entire populate much more effectively than single-level PCA. All results were found to agree with results of single-level PCA, although mPCA results were (arguably) easier to interpret than those results of single-level PCA.
For the first dataset considered here that contained two ‘expressions’ per subject (neutral or smiling), both obvious effects (widening of the mouth, corners of mouth raised slightly, exposure of teeth, and increased prominence of cheeks), and subtle effects (narrowing of the eyes and a slight widening at bottom of nose during smile) were detected in major modes of variation for the facial expression level of the mPCA model. Inspection of eigenvalues suggested that facial expression and ‘between-subject’ effects were strong influences on shape and image texture, although biological sex was a weaker effect especially for shape. Indeed, another study [24] has noted that sexual dimorphism was weakest for a Finnish population in comparison to other ethnicities (i.e., English, Welsh, and Croatian). Furthermore, the first major mode for shape showed clearly that males have longer/thinner faces on average than women [45,46] at an appropriate level of the mPCA model. Changes in image texture also clearly corresponded to biological sex, again at an appropriate level of the mPCA model. Model fits gave standardized scores for each principal component/mode of variation that show strong clustering for both shape and texture by biological sex and facial expression also at appropriate levels of the model. mPCA correctly decomposes sources of variation due to biological sex and facial expression (etc.). These results are an excellent initial test of the usefulness of mPCA in terms of modeling either shape or image texture.
For the second dataset that contained 3D time-series shape data, results of major modes of variation via mPCA were seen to correspond to the act of smiling. Inspection of eigenvalues again showed that both ‘natural lip shape’ and facial expression are strong sources of shape variation, as one would expect. Previous studies of 3D facial shape changes have posited that there are three phases to a smile [8,15,16,17], namely, onset, apex, and offset. However, if one includes rests pre and post smiling, standardized component scores from both mPCA (at the appropriate level of the model) and single-level PCA demonstrated clear evidence of a cycle containing seven phases of a smile: rest pre-smile, onset 1 (acceleration), onset 2 (deceleration), apex, offset 1 (acceleration), offset 2 (deceleration), and rest post-smile. This is strong evidence that seven phases of a smile do indeed exist and it is another excellent test of the mPCA method, now also for dynamic 3D shape data.
Future research will focus on modeling the effects of ethnicity, gender, age, genetic information, or diseases (e.g., effects perhaps previously hidden in the ‘final 5% of variation’) on facial shape or appearance. The present study has not considered the effects of “outliers” in the shape or image data. Clearly, the effects of outliers (either as isolated points, subjects or indeed even entire groups of subjects) might strongly affect mean averages used to estimate centroids of groups and also covariance matrices at the various levels of the model. The simplest method of addressing this problem is to use robust centroid and covariance matrix estimation [47,48,49] and then to carry out PCA as normal at each level. Note that robust covariance matrix estimation is included in MATLAB (2017a) and so this may be implemented easily, although a sample size of at least twice the length of the feature vector z is required. Furthermore, the mPCA method uses averages of covariance matrices (e.g., over all subjects in the population or over specific subgroups) and robust averaging of these matrices might also be beneficial. Clearly also, we can use other forms of robust PCA [50,51,52] and M-estimators [53,54] might also to deal with the problem of outliers. Finally, future research will attempt to extend existing single-level probabilistic methods of modeling shape and/or appearance (e.g., mixtures models [30,55] and extensions of Bayesian methods used in ASMs or AAMs [56,57]) to multilevel formulations and to active learning [58]. The use of schematics such as Figure 1 will hopefully prove just as useful in visualizing these models as they have for mPCA.

5. Conclusions

We have demonstrated in this article that mPCA can be used to study dynamic changes of facial shape and appearance. Our analyses yielded great insight into the question “what’s in a smile?”. PCA-based methods have the advantage that one can obtain an idea of the importance of modes of variation by the inspection of eigenvalues. This also appears to be the case for mPCA, although more caution must be exercised as the rank of covariance matrices might be constrained at some levels of the model. Although eigenvectors are orthogonal to each other within a given level, it should be noted that they do not need to be orthogonal between levels for mPCA. Potentially, this could lead to difficulty in carrying out model fits when obtaining component ‘scores’, especially for small numbers of mark-up points, although this did not seem to be a problem here. However, it might be that certain covariates affect facial or mouth shape in ways that are, in fact, inherently non-orthogonal. mPCA method provides a way of addressing this issue, whereas we would strongly expect single-level PCA to mix effects between different covariates in the principal components in such cases.

Author Contributions

Conceptualization: S.R., D.M., P.L.R., R.L., and P.P.; Methodology: D.J.J.F. and J.G.; Software: D.J.J.F. and A.I.Z.; Investigation: R.L., P.P., K.A.-M., and D.J.J.F.; Writing—Original Draft Preparation: D.J.J.F.; Writing—Review & Editing: All Authors. Supervision: D.M. and P.L.R.; Funding Acquisition: R.L.

Funding

NFBC1966 received financial support from University of Oulu, grant no. 24000692; Oulu University Hospital, grant no. 24301140; ERDF European Regional Development Fund, grant no. 539/2010 A31592.

Acknowledgments

We thank all cohort members and researchers who participated in the “46 years” study. We also wish acknowledge the work of the NFBC project center.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Sarver, D.M.; Ackerman, M.B. Dynamic smile visualization and quantification: Part 1. Evolution of the concept and dynamic records for smile capture. Am. J. Orthod. Dentofac. Orthop. 2003, 124, 4–12. [Google Scholar] [CrossRef]
  2. Sarver, D.M.; Ackerman, M.B. Dynamic smile visualization and quantification: Part 2. Smile analysis and treatment strategies. Am. J. Orthod. Dentofac. Orthop. 2003, 124, 116–127. [Google Scholar] [CrossRef]
  3. Dong, J.K.; Jin, T.H.; Cho, H.W.; Oh, S.C. The esthetics of the smile: A review of some recent studies. Int. J. Prosthodont. 1999, 12, 9–19. [Google Scholar]
  4. Otta, E. Sex differences over age groups in self-posed smiling in photographs. Psychol. Rep. 1998, 83, 907–913. [Google Scholar] [CrossRef]
  5. Drummond, S.; Capelli, J., Jr. Incisor display during speech and smile: Age and gender correlations. Angle Orthod. 2015, 86, 631–637. [Google Scholar] [CrossRef] [PubMed]
  6. Chetan, P.; Tandon, P.; Singh, G.K.; Nagar, A.; Prasad, V.; Chugh, V.K. Dynamics of a smile in different age groups. Angle Orthod. 2012, 83, 90–96. [Google Scholar] [CrossRef] [PubMed]
  7. Dibeklioğlu, H.; Gevers, T.; Salah, A.A.; Valenti, R. A smile can reveal your age: Enabling facial dynamics in age estimation. In Proceedings of the 20th ACM International Conference on Multimedia, Nara, Japan, 29 October–2 November 2012; pp. 209–218. [Google Scholar] [CrossRef]
  8. Dibeklioğlu, H.; Alnajar, F.; Salah, A.A.; Gevers, T. Combining facial dynamics with appearance for age estimation. IEEE Trans. Image Process. 2015, 24, 1928–1943. [Google Scholar] [CrossRef] [PubMed]
  9. Kau, C.H.; Cronin, A.; Durning, P.; Zhurov, A.I.; Sandham, A.; Richmond, S. A new method for the 3D measurement of postoperative swelling following orthognathic surgery. Orthod. Craniofac. Res. 2006, 9, 31–37. [Google Scholar] [CrossRef] [PubMed]
  10. Krneta, B.; Primožič, J.; Zhurov, A.; Richmond, S.; Ovsenik, M. Three-dimensional evaluation of facial morphology in children aged 5–6 years with a Class III malocclusion. Eur. J. Orthod. 2012, 36, 133–139. [Google Scholar] [CrossRef]
  11. Djordjevic, J.; Lawlor, D.A.; Zhurov, A.I.; Toma, A.M.; Playle, R.; Richmond, S. A population-based cross-sectional study of the association between facial morphology and cardiometabolic risk factors in adolescence. BMJ Open 2013, 3, e002910. [Google Scholar] [CrossRef] [Green Version]
  12. Popat, H.; Zhurov, A.I.; Toma, A.M.; Richmond, S.; Marshall, D.; Rosin, P.L. Statistical modeling of lip movement in the clinical context. Orthod. Craniofac. Res. 2012, 15, 92–102. [Google Scholar] [CrossRef] [PubMed]
  13. Alqattan, M.; Djordjevic, J.; Zhurov, A.I.; Richmond, S. Comparison between landmark and surface-based three-dimensional analyses of facial asymmetry in adults. Eur. J. Orthod. 2013, 37, 1–2. [Google Scholar] [CrossRef] [PubMed]
  14. Al Ali, A.; Richmond, S.; Popat, H.; Playle, R.; Pickles, T.; Zhurov, A.I.; Marshall, D.; Rosin, P.L.; Henderson, J.; Bonuck, K. The influence of snoring, mouth breathing and apnoea on facial morphology in late childhood: A three-dimensional study. BMJ Open 2015, 5, e009027. [Google Scholar] [CrossRef] [PubMed]
  15. Vandeventer, J. 4D (3D Dynamic) Statistical Models of Conversational Expressions and the Synthesis of Highly-Realistic 4D Facial Expression Sequences. Ph.D. Thesis, Cardiff University, Cardiff, UK, 2015. [Google Scholar]
  16. Vandeventer, J.; Graser, L.; Rychlowska, M.; Rosin, P.L.; Marshall, D. Towards 4D coupled models of conversational facial expression interactions. In Proceedings of the British Machine Vision Conference; BMVA Press: Durham, UK, 2015; p. 142-1. [Google Scholar] [CrossRef]
  17. Al-Meyah, K.; Marshall, D.; Rosin, P. 4D Analysis of Facial Ageing Using Dynamic Features. Procedia Comput. Sci. 2017, 112, 790–799. [Google Scholar] [CrossRef]
  18. Paternoster, L.; Zhurov, A.I.; Toma, A.M.; Kemp, J.P.; St. Pourcain, B.; Timpson, N.J.; McMahon, G.; McArdle, W.; Ring, S.M.; Davey Smith, G.; et al. Genome-wide Association Study of Three–Dimensional Facial Morphology Identifies a Variant in PAX3 Associated with Nasion Position. Am. J. Hum. Genet. 2012, 90, 478–485. [Google Scholar] [CrossRef] [PubMed]
  19. Fatemifar, G.; Hoggart, C.J.; Paternoster, L.; Kemp, J.P.; Prokopenko, I.; Horikoshi, M.; Wright, V.J.; Tobias, J.H.; Richmond, S.; Zhurov, A.I.; et al. Genome-wide association study of primary tooth eruption identifies pleiotropic loci associated with height and craniofacial distances. Hum. Mol. Gen. 2013, 22, 3807–3817. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  20. Claes, P.; Hill, H.; Shriver, M.D. Toward DNA-based facial composites: Preliminary results and validation. Forensic. Sci. Int. Genet. 2014, 13, 208–216. [Google Scholar] [CrossRef]
  21. Djordjevic, J.; Zhurov, A.I.; Richmond, S.; Visigen Consortium. Genetic and Environmental Contributions to Facial Morphological Variation: A 3D Population-Based Twin Study. PLoS ONE 2016, 11, e0162250. [Google Scholar] [CrossRef]
  22. Lecron, F.; Boisvert, J.; Benjelloun, M.; Labelle, H.; Mahmoudi, S. Multilevel statistical shape models: A new framework for modeling hierarchical structures. In Proceedings of the 2012 9th IEEE International Symposium on Biomedical Imaging (ISBI), Barcelona, Spain, 2–5 May 2012; pp. 1284–1287. [Google Scholar] [CrossRef]
  23. Farnell, D.J.J.; Popat, H.; Richmond, S. Multilevel principal component analysis (mPCA) in shape analysis: A feasibility study in medical and dental imaging. Comput. Methods Programs Biomed. 2016, 129, 149–159. [Google Scholar] [CrossRef] [Green Version]
  24. Farnell, D.J.J.; Galloway, J.; Zhurov, A.; Richmond, S.; Perttiniemi, P.; Katic, V. Initial results of multilevel principal components analysis of facial shape. In Annual Conference on Medical Image Understanding and Analysis; Springer: Cham, Switzerland, 2017; pp. 674–685. [Google Scholar]
  25. Farnell, D.J.J.; Galloway, J.; Zhurov, A.; Richmond, S.; Perttiniemi, P.; Katic, V. An Initial Exploration of Ethnicity, Sex, and Subject Variation on Facial Shape. in preparation.
  26. Farnell, D.J.J.; Galloway, J.; Zhurov, A.; Richmond, S.; Pirttiniemi, P.; Lähdesmäki, R. What’s in a Smile? Initial Results of Multilevel Principal Components Analysis of Facial Shape and Image Texture. In Annual Conference on Medical Image Understanding and Analysis; Springer: Cham, Switzerland, 2018; Volume 894, pp. 177–188. [Google Scholar]
  27. Cootes, T.F.; Hill, A.; Taylor, C.J.; Haslam, J. Use of Active Shape Models for Locating Structure in Medical Images. Image Vis. Comput. 1994, 12, 355–365. [Google Scholar] [CrossRef]
  28. Cootes, T.F.; Taylor, C.J.; Cooper, D.H.; Graham, J. Active Shape Models—Their Training and Application. Comput. Vis. Image Underst. 1995, 61, 38–59. [Google Scholar] [CrossRef]
  29. Hill, A.; Cootes, T.F.; Taylor, C.J. Active shape models and the shape approximation problem. Image Vis. Comput. 1996, 12, 601–607. [Google Scholar] [CrossRef]
  30. Cootes, T.F.; Taylor, C.J. A mixture model for representing shape variation. Image Vis. Comput. 1999, 17, 567–573. [Google Scholar] [CrossRef] [Green Version]
  31. Allen, P.D.; Graham, J.; Farnell, D.J.J.; Harrison, E.J.; Jacobs, R.; Nicopolou-Karayianni, K.; Lindh, C.; van der Stelt, P.F.; Horner, K.; Devlin, H. Detecting reduced bone mineral density from dental radiographs using statistical shape models. IEEE Trans. Inf. Technol. Biomed. 2007, 11, 601–610. [Google Scholar] [CrossRef] [PubMed]
  32. Edwards, G.J.; Lanitis, A.; Taylor, C.J.; Cootes, T. Statistical Models of Face Images: Improving Specificity. In Proceedings of the British Machine Vision Conference, Edinburgh, UK, 9–12 September 1996. [Google Scholar] [CrossRef]
  33. Taylor, C.J.; Cootes, T.F.; Lanitis, A.; Edwards, G.; Smyth, P.; Kotcheff, A.C. Model-based interpretation of complex and variable images. Philos. Trans. R. Soc. Lond. Biol. 1997, 352, 1267–1274. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  34. Edwards, G.J.; Cootes, T.F.; Taylor, C.J. Face recognition using active appearance models. In Computer Vision; Burkhardt, H., Neumann, B., Eds.; Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 1998; Volume 1407, pp. 581–595. [Google Scholar]
  35. Cootes, T.F.; Edwards, G.J.; Taylor, C.J. Active appearance models. IEEE Trans. Pattern Anal. Mach. Intell. 2001, 23, 681–685. [Google Scholar] [CrossRef] [Green Version]
  36. Cootes, T.F.; Taylor, C.J. Anatomical statistical models and their role in feature extraction. Br. J. Radiol. 2004, 77, S133–S139. [Google Scholar] [CrossRef]
  37. Matthews, I.; Baker, S. Active Appearance Models Revisited. Int. J. Comput. Vis. 2004, 60, 135–164. [Google Scholar] [CrossRef] [Green Version]
  38. Candemir, S.; Borovikov, E.; Santosh, K.C.; Antani, S.K.; Thoma, G.R. RSILC: Rotation- and Scale-Invariant, Line-based Color-aware descriptor. Image Vision Comput. 2005, 42, 1–12. [Google Scholar] [CrossRef]
  39. Doersch, C. Tutorial on variational autoencoders. arXiv, 2016; arXiv:1606.05908. [Google Scholar]
  40. Pu, Y.; Gan, Z.; Henao, R.; Yuan, X.; Li, C.; Stevens, A.; Carin, L. Variational autoencoder for deep learning of images, labels and captions. In Advances in Neural Information Processing Systems; Curran Associates, Inc.: Red Hook, NY, USA, 2016; pp. 2352–2360. [Google Scholar]
  41. Wetzel, S.J. Unsupervised learning of phase transitions: From principal component analysis to variational autoencoders. Phys. Rev. E. 2017, 96, 022140. [Google Scholar] [CrossRef] [PubMed]
  42. Etemad, K.; Chellappa, R. Discriminant analysis for recognition of human face images. J. Opt. Soc. Am. A 1997, 14, 1724–1733. [Google Scholar] [CrossRef] [Green Version]
  43. Kim, T.-K.; Kittler, J. Locally Linear Discriminant Analysis for Multimodally Distributed Classes for Face Recognition with a Single Model Image. IEEE Trans. Pattern Anal. Mach. Intell. 2005, 27, 318–327. [Google Scholar] [CrossRef] [PubMed]
  44. Dibeklioglu, H.; Valenti, R.; Salah, A.A.; Gevers, T. Eyes do not lie: Spontaneous versus posed smiles. In Proceedings of the 18th ACM International Conference on Multimedia, Firenze, Italy, 25–29 October 2010; pp. 703–706. [Google Scholar] [CrossRef]
  45. Velemínská, J.; Bigoni, L.; Krajíček, V.; Borský, J.; Šmahelová, D.; Cagáňová, V.; Peterka, M. Surface facial modeling and allometry in relation to sexual dimorphism. HOMO 2012, 63, 81–93. [Google Scholar] [CrossRef] [PubMed]
  46. Toma, A.M.; Zhurov, A.; Playle, R.; Richmond, S. A three-dimensional look for facial differences between males and females in a British-Caucasian sample aged 15 ½ years old. Orthol. Craniofac. Res. 2008, 11, 180–185. [Google Scholar] [CrossRef] [PubMed]
  47. Rousseeuw, P.J.; Driessen, K.V. A fast algorithm for the minimum covariance determinant estimator. Technometrics 1999, 41, 212–223. [Google Scholar] [CrossRef]
  48. Maronna, R.; Zamar, R.H. Robust estimates of location and dispersion for high dimensional datasets. Technometrics 2002, 50, 307–317. [Google Scholar] [CrossRef]
  49. Olive, D.J. A resistant estimator of multivariate location and dispersion. Comput. Stat. Data Anal. 2004, 46, 93–102. [Google Scholar] [CrossRef] [Green Version]
  50. Candès, E.J.; Li, X.; Ma, Y.; Wright, J. Robust principal component analysis? JACM 2011, 58, 11. [Google Scholar] [CrossRef]
  51. Wright, J.; Ganesh, A.; Rao, S.; Peng, Y.; Ma, Y. Robust principal component analysis: Exact recovery of corrupted low-rank matrices via convex optimization. In Advances in Neural Information Processing Systems; Curran Associates, Inc.: Red Hook, NY, USA, 2009; pp. 2080–2088. [Google Scholar]
  52. Hubert, M.; Engelen, S. Robust PCA and classification in biosciences. Bioinformatics 2004, 20, 1728–1736. [Google Scholar] [CrossRef] [Green Version]
  53. Andersen, R. Modern Methods for Robust Regression; Quantitative Applications in the Social Sciences. 152; Sage Publications: Los Angeles, CA, USA, 2008; ISBN 978-1-4129-4072-6. [Google Scholar]
  54. Godambe, V.P. Estimating Functions; Oxford Statistical Science Series. 7; Clarendon Press: New York, NY, USA, 1991; ISBN 978-0-19-852228-7. [Google Scholar]
  55. Ulukaya, S.; Erdem, C.E. Gaussian mixture model based estimation of the neutral face shape for emotion recognition. Digit. Signal Process. 2014, 32, 11–23. [Google Scholar] [CrossRef]
  56. Patenaude, B.; Smith, S.M.; Kennedy, D.N.; Jenkinson, M. A Bayesian model of shape and appearance for subcortical brain segmentation. Neuroimage 2011, 56, 907–922. [Google Scholar] [CrossRef] [Green Version]
  57. Li, F.; Fergus, R.; Perona, P. Learning generative visual models from few training examples: An incremental bayesian approach tested on 101 object categories. Comput. Vis. Image Underst. 2007, 106, 59–70. [Google Scholar] [CrossRef]
  58. Bouguelia, M.-R.; Nowaczyk, S.; Santosh, K.C.; Verikas, A. Agreeing to disagree: Active learning with noisy labels without crowdsourcing. Int. J. Mach. Learn. Cybern. 2018, 9, 1307–1319. [Google Scholar] [CrossRef]
Figure 1. Flowchart illustrating the multilevel model of facial shape for dataset 1.
Figure 1. Flowchart illustrating the multilevel model of facial shape for dataset 1.
Jimaging 05 00002 g001
Figure 2. Illustration of the 21 landmark points for dataset 1 ((1) Glabella (g); (2) Nasion (n); (3) Endocanthion left (enl); (4) Endocanthion right (enr); (5) Exocanthion left (exl); (6) Exocanthion right (exr); (7) Palpebrale superius left (psl); (8) Palpebrale superius right (psr); (9) Palpebrale inferius left (pil); (10) Palpebrale in-ferius right (pir); (11) Pronasale (prn); (12) Subnasale (sn); (13) Alare left (all); (14) Alare right (alr); (15) Labiale superius (ls); (16) Crista philtri left (cphl); (17) Crista philtri right (cphr); (18) Labiale inferius (li); (19) Cheilion left (chl); (20) Cheilion right (chr); (21) Pogonion (pg)).
Figure 2. Illustration of the 21 landmark points for dataset 1 ((1) Glabella (g); (2) Nasion (n); (3) Endocanthion left (enl); (4) Endocanthion right (enr); (5) Exocanthion left (exl); (6) Exocanthion right (exr); (7) Palpebrale superius left (psl); (8) Palpebrale superius right (psr); (9) Palpebrale inferius left (pil); (10) Palpebrale in-ferius right (pir); (11) Pronasale (prn); (12) Subnasale (sn); (13) Alare left (all); (14) Alare right (alr); (15) Labiale superius (ls); (16) Crista philtri left (cphl); (17) Crista philtri right (cphr); (18) Labiale inferius (li); (19) Cheilion left (chl); (20) Cheilion right (chr); (21) Pogonion (pg)).
Jimaging 05 00002 g002
Figure 3. Schematic illustration of a time series of smile amplitudes from Equation (1) for 3D shape data in dataset 2. Including rest phases, seven phases can be identified manually: rest pre-smile, onset acceleration, onset deceleration, apex, offset acceleration, offset deceleration, and rest post-smile.
Figure 3. Schematic illustration of a time series of smile amplitudes from Equation (1) for 3D shape data in dataset 2. Including rest phases, seven phases can be identified manually: rest pre-smile, onset acceleration, onset deceleration, apex, offset acceleration, offset deceleration, and rest post-smile.
Jimaging 05 00002 g003
Figure 4. Eigenvalues for dataset 1 from single-level PCA and from mPCA level 1 (biological sex), level 2 (between-subject variation), and level 3 (within-subject variation: facial expression). (a) Shape data; (b) Image texture data (All shapes have been scaled so that the average point-to-centroid distance equals 1.).
Figure 4. Eigenvalues for dataset 1 from single-level PCA and from mPCA level 1 (biological sex), level 2 (between-subject variation), and level 3 (within-subject variation: facial expression). (a) Shape data; (b) Image texture data (All shapes have been scaled so that the average point-to-centroid distance equals 1.).
Jimaging 05 00002 g004
Figure 5. Modes of variation for shape for dataset 1 for the first three modes from single-level PCA in the upper set of images: (a) = mode 1; (b) = mode 2; (c) = mode 3. The first modes from levels 1 to 3 mPCA in the bottom set of images: (d) = mode 1, level 1 (biological sex); (e) = mode 1, level 2 (between subjects); (f) = mode 1, level 3 (facial expression).
Figure 5. Modes of variation for shape for dataset 1 for the first three modes from single-level PCA in the upper set of images: (a) = mode 1; (b) = mode 2; (c) = mode 3. The first modes from levels 1 to 3 mPCA in the bottom set of images: (d) = mode 1, level 1 (biological sex); (e) = mode 1, level 2 (between subjects); (f) = mode 1, level 3 (facial expression).
Jimaging 05 00002 g005
Figure 6. Modes of variation for image texture for dataset 1 for the first three modes ((a) = mode 1; (b) = mode 2; (c) = mode 3) from single-level PCA in the left-hand set of images, and the first modes from levels 1 to 3 ((a) = level 1; (b) = level 2; (c) = level 3) from mPCA in the right-hand set of images. Note that for each set of three images: left image = mean − SD; middle image = mean; right image = mean + SD.
Figure 6. Modes of variation for image texture for dataset 1 for the first three modes ((a) = mode 1; (b) = mode 2; (c) = mode 3) from single-level PCA in the left-hand set of images, and the first modes from levels 1 to 3 ((a) = level 1; (b) = level 2; (c) = level 3) from mPCA in the right-hand set of images. Note that for each set of three images: left image = mean − SD; middle image = mean; right image = mean + SD.
Jimaging 05 00002 g006
Figure 7. Standardized component scores with respect to shape for dataset 1: (a) Components 1 and 2 for single-level PCA; (b) Components 1 and 3 for single-level PCA; (c) Component 1 for level 1 (biological sex) for mPCA; (d) Components 1 and 2 for level 3 (facial expression) for mPCA.
Figure 7. Standardized component scores with respect to shape for dataset 1: (a) Components 1 and 2 for single-level PCA; (b) Components 1 and 3 for single-level PCA; (c) Component 1 for level 1 (biological sex) for mPCA; (d) Components 1 and 2 for level 3 (facial expression) for mPCA.
Jimaging 05 00002 g007
Figure 8. Standardized component scores with respect to image texture for dataset 1: (a) Components 1 and 2 for single-level PCA; (b) Components 1 and 3 for single-level PCA; (c) Component 1 for level 1 (biological sex) for mPCA; (d) Components 1 and 2 for level 3 (facial expression) for mPCA.
Figure 8. Standardized component scores with respect to image texture for dataset 1: (a) Components 1 and 2 for single-level PCA; (b) Components 1 and 3 for single-level PCA; (c) Component 1 for level 1 (biological sex) for mPCA; (d) Components 1 and 2 for level 3 (facial expression) for mPCA.
Jimaging 05 00002 g008aJimaging 05 00002 g008b
Figure 9. Eigenvalues for dataset 2 (shape data only) from single-level PCA and from mPCA level 1 (between-subject variation), level 2 (variation between smile phases), and level 3 (variation within smile phases).
Figure 9. Eigenvalues for dataset 2 (shape data only) from single-level PCA and from mPCA level 1 (between-subject variation), level 2 (variation between smile phases), and level 3 (variation within smile phases).
Jimaging 05 00002 g009
Figure 10. Modes of variation for dataset 2 from mPCA: (a) level 1 (between-subject variation), mode 1, coronal plane; (b); level 1 (between-subject variation), mode 1, transverse plane; (c) level 2 (variation between smile phases), mode 1, coronal plane; (d) level 2 (variation between smile phases), mode 1, transverse plane.
Figure 10. Modes of variation for dataset 2 from mPCA: (a) level 1 (between-subject variation), mode 1, coronal plane; (b); level 1 (between-subject variation), mode 1, transverse plane; (c) level 2 (variation between smile phases), mode 1, coronal plane; (d) level 2 (variation between smile phases), mode 1, transverse plane.
Jimaging 05 00002 g010
Figure 11. Centroids over smile phases for standardized component scores with respect to shape for dataset 2: (a) Components 1 and 2 for single-level PCA; (b) Components 1 and 2 at level 2 (variation between smile phases) for mPCA.
Figure 11. Centroids over smile phases for standardized component scores with respect to shape for dataset 2: (a) Components 1 and 2 for single-level PCA; (b) Components 1 and 2 at level 2 (variation between smile phases) for mPCA.
Jimaging 05 00002 g011

Share and Cite

MDPI and ACS Style

Farnell, D.J.J.; Galloway, J.; Zhurov, A.I.; Richmond, S.; Marshall, D.; Rosin, P.L.; Al-Meyah, K.; Pirttiniemi, P.; Lähdesmäki, R. What’s in a Smile? Initial Analyses of Dynamic Changes in Facial Shape and Appearance. J. Imaging 2019, 5, 2. https://doi.org/10.3390/jimaging5010002

AMA Style

Farnell DJJ, Galloway J, Zhurov AI, Richmond S, Marshall D, Rosin PL, Al-Meyah K, Pirttiniemi P, Lähdesmäki R. What’s in a Smile? Initial Analyses of Dynamic Changes in Facial Shape and Appearance. Journal of Imaging. 2019; 5(1):2. https://doi.org/10.3390/jimaging5010002

Chicago/Turabian Style

Farnell, Damian J. J., Jennifer Galloway, Alexei I. Zhurov, Stephen Richmond, David Marshall, Paul L. Rosin, Khtam Al-Meyah, Pertti Pirttiniemi, and Raija Lähdesmäki. 2019. "What’s in a Smile? Initial Analyses of Dynamic Changes in Facial Shape and Appearance" Journal of Imaging 5, no. 1: 2. https://doi.org/10.3390/jimaging5010002

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop