Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Assessment of accuracy and recognition of three-dimensional computerized forensic craniofacial reconstruction

  • Geraldo Elias Miranda ,

    Roles Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Project administration, Resources, Validation, Writing – original draft

    geraldoelias@usp.br

    Affiliation Department of Social Dentistry, Laboratory of Forensic Anthropology and Odontology (OFLAB), School of Dentistry, University of São Paulo, São Paulo, Brazil

  • Caroline Wilkinson,

    Roles Supervision, Writing – review & editing

    Affiliation Face Lab, Liverpool John Moores University, Liverpool, United Kingdom

  • Mark Roughley,

    Roles Visualization, Writing – review & editing

    Affiliation Face Lab, Liverpool John Moores University, Liverpool, United Kingdom

  • Thiago Leite Beaini,

    Roles Formal analysis, Methodology, Writing – original draft

    Affiliation Department of Preventive and Social Dentistry, Uberlândia Federal University, Uberlândia, Brazil

  • Rodolfo Francisco Haltenhoff Melani

    Roles Conceptualization, Funding acquisition, Investigation, Methodology, Project administration, Supervision, Writing – review & editing

    Affiliation Department of Social Dentistry, Laboratory of Forensic Anthropology and Odontology (OFLAB), School of Dentistry, University of São Paulo, São Paulo, Brazil

Abstract

Facial reconstruction is a technique that aims to reproduce the individual facial characteristics based on interpretation of the skull, with the objective of recognition leading to identification. The aim of this paper was to evaluate the accuracy and recognition level of three-dimensional (3D) computerized forensic craniofacial reconstruction (CCFR) performed in a blind test on open-source software using computed tomography (CT) data from live subjects. Four CCFRs were produced by one of the researchers, who was provided with information concerning the age, sex, and ethnic group of each subject. The CCFRs were produced using Blender® with 3D models obtained from the CT data and templates from the MakeHuman® program. The evaluation of accuracy was carried out in CloudCompare, by geometric comparison of the CCFR to the subject 3D face model (obtained from the CT data). A recognition level was performed using the Picasa® recognition tool with a frontal standardized photography, images of the subject CT face model and the CCFR. Soft-tissue depth and nose, ears and mouth were based on published data, observing Brazilian facial parameters. The results were presented from all the points that form the CCFR model, with an average for each comparison between 63% and 74% with a distance -2.5 ≤ x ≤ 2.5 mm from the skin surface. The average distances were 1.66 to 0.33 mm and greater distances were observed around the eyes, cheeks, mental and zygomatic regions. Two of the four CCFRs were correctly matched by the Picasa® tool. Free software programs are capable of producing 3D CCFRs with plausible levels of accuracy and recognition and therefore indicate their value for use in forensic applications.

Introduction

Facial reconstruction or facial approximation is a technique that aims to reproduce the individual facial characteristics, prior to death, based on interpretation of the skull, with the objective of recognition leading to an identification [1]. The resulting reconstruction can be published in the media and therefore lead to recognition of the deceased by family or friend and consequently to identification by other means.

Nowadays, there are two techniques to perform a three-dimensional (3D) facial reconstruction: the manual or the computerized. 3D computerized craniofacial forensic reconstructions (CCFR) have some advantages provided by the use of a computer as the enhanced visualization tools allow for the display of bone and the skin together with many transparency adjustments [2]. It is possible to evaluate the reconstruction during the process, correcting mistakes [2, 3]. Other advantages include reduced risk of damaging the skull and the reassembly of skull fragments or the replacement of absent portions [4].

Nevertheless, some disadvantages of the CCFR are also observed. Firstly, some methods can be time consuming [4]. Secondly, many of the existing methodologies, may require a certain level of experience from the operator [4]. Some of these methods have undergone rigorous scientific evaluation of the accuracy and recognition levels [37]. There is a growing necessity to determine a clear and reproducible protocol to evaluate the quality of 3D CCFR in relation to the actual face of the subject [4, 8].

In order to overcome these limitations, this research highlights the importance of evaluating the results of facial reconstruction using new software. It is fundamental to clarify that the accuracy of the CCFR is directly related to the prediction of morphological facial traits, while recognition level is related to the capacity of automated recognition [5]. In this paper both of these variables were tested. The evaluation of a CCFR is important in the continuous enhancement of the methods involved, even though existing research presents little or no focus in the evaluation of the result quality achieved [9].

To our knowledge, there is no other research that evaluated the accuracy and recognition level of a CCFR with the software used in this study. With new methods, there is a need to achieve simple, reliable and automated ways to perform this task, returning more realistic images [8]. The 3D CCFR is in current need for verification of the reliability and reproducibility with available tools and others to be developed in the future [5, 6, 10], as there are no computerized tools globally accepted by the forensic community [11].

Therefore this work presents software not often used for forensic means, focusing more on the evaluation of the results than in the technique itself, as program handling can be learned. The use of such open source software might enable a wider cohort of researchers with the opportunity to produce CCFR, making CCFR more simple, efficient and accessible. The aim of this study is to evaluate the accuracy and recognition level of 3D CCFR using free open-source software programs.

Materials and methods

Four living volunteers donated their existing computer tomography (CT) exams and frontal standardized photographs: subject 1: Female, 22yo; subject 2: male, 24yo; subject 3: female, 49yo; subject 4: male, 21yo. They were classified as Brazilian Caucasian by the researcher and had no record of orthodontic treatment, craniofacial, plastic surgery or facial deformity. Written informed consent was obtained from all subjects and no subject was exposed to radiation for the purposes of this research. The Ethics committee of the University of São Paulo—Brazil, under protocol number 1.608.387, approved the procedures. The individual in this manuscript has given written informed consent (as outlined in PLOS consent form) to publish these case details.

The DICOM files (digital imaging and communications in medicine) from the CT data were first reconstructed in the open source software Horos® 3.0 version (www.horosproject.org), a viewer that can export stereolithographic (STL) files from selected surfaces, such as bone or skin tissues by selecting the respective density of the region of interest. The generated STL file was imported into Blender® 2.78 version (www.blender.org), a free and open-source 3D modeling and animation software where the CCFRs were performed.

The skull was positioned in reference to the Frankfurt plane and soft tissue markers were placed over its surface, using landmarks as described in a study of a Brazilian sample [1], that reports a table with male and female values for 10 mid-sagittal points and 11 bilateral points in a total of 32 references (Fig 1).

thumbnail
Fig 1. Soft tissue markers placement, eyes, mouth and nose guides.

https://doi.org/10.1371/journal.pone.0196770.g001

The shape and position of the eyes, nose and mouth also followed pre-established parameters:

  • Eyes: eyeballs of 24mm in diameters were placed inside the orbital cavity so that the eyeball and pupil were centrally located within the orbits, 4 mm from the roof and 4–5 mm from the lateral wall [12]. Protrusion was established as tangents from the margins of the orbital cavity crossing only the iris portion of the eye [13]. The endocanthus of the eye, was placed 2 mm lateral to the lacrimal bone crest at its middle and the exocanthus approximately 3–4 mm medially from the malar tubercle [12]. For the CCFR’s, the eyelids were modeled closed as the CT exams were produced in the same position.
  • Nose: the maximum width of the piriform cavity represents approximately 3/5 of the maximum width of the nose [14]. The tip of the nose was estimated to be at the crossing point of two lines: the nasal bone projection and a perpendicular line from the prosthion projection, in a 90° [15].
  • Mouth: The inter-canine distance is estimated as 75% of the total mouth width [16]. The height of the lips vermillion is 26% of the width [17]. The lower and upper lip intersection should be placed in the lower section of the upper central incisor [18].

In a forensic investigation, the anthropological examination will estimate the age, sex and ethnic group of the human remains. It is through this information that a 3D face model template was selected as appropriate on MakeHuman® 1.1.0 (makehuman.org). MakeHuman® is a 3D software with the goal to create humanoid models. The chosen template generated by the software was exported as an STL files and imported into Blender® and placed over the skull model to be reconstructed (Fig 2). The CCFRs were reconstructed as a blind study by one of the researchers, with the only known information as age, sex and ethnic group of the individual.

thumbnail
Fig 2. Template placement over the soft-tissue markers in lateral view.

https://doi.org/10.1371/journal.pone.0196770.g002

The addition of individual characterization was kept to a minimum due to the fact that incorrect skin color, hair and other modifiers can impair identification and lead to error [6, 19]. This research followed the recommendation of authors of existing CCFR research and the CCFR was presented without hair, the skin was left in a shade of gray and the eyes were closed.

The quantitative evaluation of the accuracy was carried out using CloudCompare® 2.6.3 software (www.cloudcompare.org), where two 3D point clouds, each corresponding to triangular faced mesh models, can be compared. To complete this task, the CT soft tissue model was placed as “reference” and the reconstruction as “compared” and the software calculated the distances between the two point clouds. An important step is to align the models, as described by Decker (2013) [20] using the nasion, and the deepest lateral points of the orbits as reference. Sequentially, one can use the “cut” tool to eliminate any lateral or lower parts that could be observed only in one of the models. Prior to attempting to align the CCFR 3D model with the respective CT 3D model, a test alignment with two identical models took place to verify the method of alignment. If the match was different to 100%, the method of alignment would be incorrect as the two 3D models were identical. (Fig 3). Therefore it was possible to calculate the discrepancy in millimeters between the meshes, which have minimum and maximum values, graphics, histograms and numerical tables as outputs and also can be observed through a color-map applied on the compared model’s surface.

The evaluation of the automated recognition level of the CCFR was performed using Picasa® 3.9 (https://picasa.google.com/), which is software for image storage. In the field “people” the standardized frontal photographs of each subject were inserted and named. Previously the frontal images captured from the CT soft-tissue model and from the CCFR were inserted into the data-base. Once the “people” field is calibrated by the imported images, the program can automatically scan existent ones. Consequently one of the four events occurs: 1- the image is not recognized as a face; 2 the image is recognized as a face but not a matching to any “people” (classified as unnamed); 3- the image is correlated to a known person and confirmation may be asked; 4- the image is incorrectly matched to the wrong person.

Results

The CloudCompare® showed the discrepancy (mm) between the surface of the CCFR and the surface of the CT face model. The positive numbers indicated that the surface of the reconstruction was overestimated in relation to the real face, while the negative numbers indicated that the surface was underestimated (Fig 4). The color map generated shows green when there is a distance of less than ± 2.5mm between the meshes, the colors that vary from yellow to red show when the distance is between 2.5mm to 10mm, and the blue to the dark blue colors show when the distance is -2.5mm to -10mm. The grey color areas are representative of the CT 3D model. In Fig 4D the CT 3D model and the CCFR 3D model are aligned on top of each other, overlapping. When the CT 3D model appears (the grey colour) it means that the surface of the produced CCFR 3D model was underestimated (Fig 4D). When analyzing each face region it can be noted that in all cases the greatest discrepancies occurred in the cheek and in the eyes, which were underestimated (blue), while the chin and zygomatic area were overestimated (red) (Fig 4E).

thumbnail
Fig 4. The columns show subjects 1, 2, 3, 4.

Line A: photograph; line B: CT surface; line C: reconstruction; line D: result of the comparison between CT and reconstruction with the respective color map. E: reconstruction with color map.

https://doi.org/10.1371/journal.pone.0196770.g004

The results were analyzed through descriptive statistics using the average, standard deviation and maximum and minimum deviation (mm) between the surface of the CCFR and the CT face model (Table 1). Table 2 shows the distribution (%) of the error deviation ± 2.5 mm, as performed by Lee et al [21].

thumbnail
Table 1. Average and standard deviation of the discrepancy between the facial surface of the reconstruction and corresponding subject.

https://doi.org/10.1371/journal.pone.0196770.t001

thumbnail
Table 2. Distribution (%) of the deviation error between the surfaces of the reconstruction and the subject within each defined range (2.5 mm).

https://doi.org/10.1371/journal.pone.0196770.t002

When evaluating the recognition level in Picasa® the program recognized 100% of imported images as a face. In addition, the program correctly assigned all photographs to the respective subjects, three of the four CTs (subjects 1, 2 and 3) and two of the four CCFRs (subjects 1 and 3) (Fig 5).

thumbnail
Fig 5. Analysis of the recognition level of photographs, CT and CCFR in Picasa®.

The CCFRs of subjects 2 and 4 plus the CT of subject 4 were classified as unnamed (circle).

https://doi.org/10.1371/journal.pone.0196770.g005

Discussion

The progress of computing and the improvement of imaging techniques in recent years has promoted the development of fast and flexible programs that can be used for CCFR [2, 4]. However, it is very important that researchers analyze the accuracy, reliability and reproducibility of these programs [2, 4] so that CCFR becomes accepted in the forensic field [4]. In addition, a constant reevaluation will promote improvement and increase reliability [4].

A critical element in the design of a CCFR is the evaluation of its accuracy. One-to-one comparison and face pools are useful methods for assessing CCFR accuracy, but other objective methods are required to assess quantitative accuracy, especially due to increased application of CCFR production methods [21]. The evaluation through the comparison of the geometric surface is an effective tool to evaluate the accuracy of the CCFR [21]. Two 3D facial models can be aligned so that differences in facial contour and reconstruction can be numerically computed, as well as providing a spatial map of the differences of each facial region reconstructed [2, 3, 21, 22]. Thus a quantitative evaluation of reconstruction error can be performed, observing three-dimensional surface differences [2, 3].

This was the method used in this research to evaluate the accuracy of the CCFR. For a more objective and quantitative evaluation, the CloudCompare® software was used, since the program quantitatively evaluates the morphological discrepancy of the surface between the reconstructed face and the subject. The purpose of these comparisons was not to quantify the similarity, but to measure the accuracy of the approximation via metric comparison [20]. Short et al. [23] evaluated the difference between landmarks, but the problem is that the program measures the distance between the closest points and that are not necessarily the landmarks. This was not a problem in this research since the distance between the CCFR landmarks and the real face was not evaluated [23], but the distance between the meshes in a holistic way.

The entire reconstruction was not analyzed, as the posterior region of the head, ears and below the mandible were removed, as was carried out by other authors [21, 23]. The results showed that the mean distance found in this study varied from -1.66 to 0.33mm (Table 1). Similar results were found in the literature with a mean error varying between 1.14mm [22]; -0.49 to -0.31mm [5]; -0.2 to 0.4mm [21] and 4.0mm [11].

When the total points of each reconstruction were analyzed, 63.20% to 73.67% presented a distance within the range of -2.5 ≤ x ≤ 2.5 mm between the CCFR (compared) and the real face (reference) (Table 2). Similar results were described in previous studies with percentages varying from 54% to 76% [5], 52% to 60% [7], 79% to 87% [21] e 56 to 90% [23]; when an error of ± 2.5mm was applied using a sample of three, two, three and ten subjects, respectively. The soft tissue thickness of an individual will never be completely accurate since the tissue thicknesses markers employed are averages [23]. In addition, minor errors may occur when converting CT files for viewing and exporting to programs [21, 24]. Another factor that influences the success of the reconstruction is the lack of information that BMI has on soft tissue thickness [20]. The level of precision required for a facial reconstruction to result in recognition is not clearly known [11]. Laboratory studies suggest that it is not possible to produce an exact portrait, but it should be possible to estimate the facial morphology with enough accuracy to allow recognition [10]. Therefore, the result of the accuracy found in this research is similar to other studies and can indicate value for use in forensic investigations.

Accuracy assessment also assists in the analysis of specific regions of the face. Prediction of location, size and morphology of facial features—eyes, nose, mouth and ears—is critical to the accuracy of CCFRs [21]. Some measures of the anatomical regions of the reconstruction are underestimated (they are smaller compared to the target model) and other areas are overestimated (larger than the target) [23]. When evaluating each region of the face, it can be observed that in all cases the greatest discrepancies occurred in the cheek and in the eyes that were underestimated (blue), while the chin and the zygomatic area were overestimated (red). This demonstrates that these regions of the CCFRs in this research need to be improved. Some variation in the cheek was already expected due to the gravitational detachment caused in the patient lying supine position in the CT [20, 22, 25]. In addition, the soft tissue thickness markers utilized in this research comes from cadavers, so thanatological effects may have influenced the accuracy of the reconstructions since of the four subjects, three had a negative average (Table 1). Any measure taken after death is slightly reduced and the position of features may be subject to high levels of post-mortem change [25]. This can be verified by replicating this work and using different tables of soft tissue thickness to evaluate which one presents better accuracy.

When comparing these results with the literature it was observed that the studies differ when specific areas of the face are analyzed, due to different methodologies and software used. These differences may also be the result of positional effects when performing CT scans making unmatched facial features, for example, facial expression, mouth and open eyes.

One study observed that, on average, the area of the chin was the one with the greatest similarity and the area of the eyes smaller [9]. Guyomarc’h et al (2014) found a greater error in the region of the ear (7mm), moderate in the mouth (4.5mm) and smaller in the nasal regions (3.1mm) and eyes (2.9mm) [11]. The reconstructions had larger cheeks, more prominent upper lip and a pattern of error in the ear, and the ear and tip of the nose were the areas with greater error (> 5 mm) [7].

Decker [20] carried out four reconstructions of the same individual, two being computerized using the FaceIT and ReFace programs and two manual clay craniofacial reconstructions and found out that there was great variation between the reconstruction methods [20]. All techniques performed well on nasal width, but there was variability in the angle / length of the nose, underestimating the width of the mouth, glabella and chin [20]. Another study showed that the nose and mouth were the most difficult areas to reconstruct [24]. The nose and mouth areas were overestimated with the biggest difference in the nose (7mm) [23]. In CCFR of subject 2 there was a positive difference on the right side and negative on the left side of the nose, which may have contributed to its non-recognition by the program.

The accuracy of facial reconstruction may directly affect the success of recognition [23] since the ultimate goal of reconstruction is not accuracy, but the recognition and subsequent success of identification [2, 3, 22]. Thus, recognition performance seems to be the most appropriate way of evaluating its effectiveness [19]. The automated recognition test consists of comparing the result of the CCFR with a candidate base including the actual person [2, 3]. The most valuable validation of reconstruction techniques is still based on human recognition, because it has a more practical relevance [24]. The majority of facial reconstruction studies evaluate the recognition level subjectively (examiners).

However, the future of facial recognition especially when there is a large base of images undoubtedly depends on facial recognition software that is far less cumbersome and more efficient than having human reviewers [26]. Thus for a more objective result the evaluation of recognition of the reconstructions was carried out by a program called Picasa®, a facial recognition software. The program correctly assigned all the photographs to the respective subjects, three of the four CTs (subjects 1, 2 and 3) and two of the four CCFRs (subjects 1 and 3) (Fig 5). Although all reconstructions have a close error rate (Table 2), several factors may have contributed to non-recognition by the program. For example, subjects 2 and 4 are with eyes open in CT and closed in CCFR, difference in nose, etc. A study showed that Picasa® was able to recognize 100% photographs and 27.5% of CT [26]. However, the program had not yet been used for CCFR recognition as performed in this work. Moreover, it was difficult to compare this similarity test with other studies since most of them use the subjective "pool face" methodology and the results vary widely. New researches evaluating the similarity of CCFRs using computational programs and with a larger sample are necessary.

Computational systems are dependent on facial templates, mean tissue depth markers, and specific population data [2, 27]. In this study, soft tissue thickness markers [1], lip prediction [17] and nose tip prediction methods [15] were used from studies with Brazilians in order to apply regional measures for better reconstruction efficiency [21, 24, 27].

The first step in reconstruction is the anthropological examination of the skull to estimate age, sex, and ancestry. It is through the anthropological examination that the 3D face model template will be chosen to be used in the CCFR. The template specification will strongly influence the final result, if an inappropriate template is chosen, bias may occur [2, 3, 8, 24]. Using only a generic template the potential to produce a biased model is high [2, 3]. When the difference between the 3D face model template and the target skull is large, the required deformation will be more pronounced, which can result in an implausible, unrealistic or cartoon-like facial reconstructions [24].

The performance of the reconstruction depends on the quality of the database used. More precisely, successful CCFRs can depend on the available CT scan data in a database that represents any skull of a given population [25]. Images in a database from CT provide good definitions of structures. However, this imaging technique is invasive and for ethical and legal reasons it is difficult to build a large database of healthy people [25]. Templates of 3D face models from the MakeHuman® database were used in this project as it was possible to obtain an appropriate template for the anthropological profile of the skull to be reconstructed. So the problem of a limited template base [4] has been eliminated.

One of the disadvantages of using CT is the sensitivity to amalgam and metal dental restorations that produce artifacts in the images produced during the CT exam [3]. This research focuses on CCFRs of living individuals however, with CCFRs of deceased individuals, photogrammetry can be used to obtain the 3D model of the skull from photographs. However photogrammetry does not show internal regions of the skull that are important for other forensic analysis. One study obtained a 3D skull model through photogrammetry and subsequently performed the reconstructions in Blender® [28]. The authors highlighted some advantages such as the ease of visualization, storage and sharing of the reconstruction [28]. The technique presented here also allows remote access and so it is possible to work with experts who are not in the same physical environment, avoiding the transport of the remains.

Turner et al. [29] developed a computer system called ReFace, a program that uses a dense placement of landmarks rather than the placement of sparse landmarks. The problem is that ReFace is an automated program developed by the Federal Bureau of Investigation (FBI) and is not available to the scientific community, so it cannot be used by other experts or be tested by other researchers.

All software programs used in this research study are free and open source, which can be downloaded and used by anyone. The use of free software eliminates the initial cost and contributes to the dissemination of the technique. Most CCFR methods make use of non-rigid generic deformation, which is mathematically well defined and easy to use but if not used correctly they can deform the face presenting it in an unrealistic way [3]. Another disadvantage of free software is that they can be disabled by the developer leaving the user unsupported or without updates.

Conclusion

The results from this study demonstrate that free open-source software programs are capable of producing 3D computerized forensic craniofacial reconstructions with plausible level of accuracy and recognition and therefore indicate value for use in forensic investigations. Subsequent studies are required using different tissue thickness and with a larger sample or using alternative software programs to evaluate the method further.

Acknowledgments

The authors would like to show their gratitude to the subjects who provided their facial photographs and CT exams. This research was carried out as part of a PhD study at the University of São Paulo by dental student Geraldo Elias Miranda with cooperation Face Lab at Liverpool John Moores University.

References

  1. 1. Tedeschi-Oliveira SV, Melani RF, de Almeida NH, de Paiva LA. Facial soft tissue thickness of Brazilian adults. Forensic Sci Int. 2009;193(1–3):127.e1–7.
  2. 2. Wilkinson C, Rynn C. Craniofacial Identification. 1st ed. New York: Cambridge University Press; 2012.
  3. 3. Claes P, Vandermeulen D, De Greef S, Willems G, Clement JG, Suetens P. Computerized craniofacial reconstruction: Conceptual framework and review. Forensic Sci Int. 2010;201(1–3):138–145. pmid:20359837
  4. 4. Wilkinson C. Computerized forensic facial reconstruction: A review of current systems. Forensic Sci Med Pathol. 2005;1(3):173–177. pmid:25870042
  5. 5. Lee WJ, Wilkinson CM, Hwang HS. An accuracy assessment of forensic computerized facial reconstruction employing cone-beam computed tomography from live subjects. J Forensic Sci. 2012;57(2):318–327. pmid:22073932
  6. 6. Parks CL, Richard AH, Monson KL. Preliminary performance assessment of computer automated facial approximations using computed tomography scans of living individuals. Forensic Sci Int. 2013;233(1–3):133–139. pmid:24314512
  7. 7. Wilkinson C, Rynn C, Peters H, Taister M, Kau CH, Richmond S. A blind accuracy assessment of computer-modeled forensic facial reconstruction using computed tomography data from live subjects. Forensic Sci Med Pathol. 2006;2(3):179–187. pmid:25868696
  8. 8. Clement JG, Marks MK. Computer-graphic Facial reconstruction. 1st ed. Burlington: Academic Press; 2005.
  9. 9. Zhao J, Liu C, Wu Z, Duan F, Wang K, Jia T, et al. Craniofacial reconstruction evaluation by geodesic network. Comput Math Methods Med. 2014;2014:943647. pmid:25214890
  10. 10. Wilkinson C. Craniofacial analysis and identification. In: Valentine T, Davis JP, editors. Forensic facial identification: theory and practice of identification from eyewitnesses, composites and CCTV. Chichester: John Wiley & Sons; 2015. pp. 93–126.
  11. 11. Guyomarc’h P, Dutailly B, Charton J, Santos F, Desbarats P, Coqueugniot H. Anthropological facial approximation in three dimensions (AFA3D): computer-assisted estimation of the facial morphology using geometric morphometrics. J Forensic Sci. 2014;59(6):1502–1516. pmid:25088006
  12. 12. Whitnall S. The anatomy of the human orbit and accessory organs of vision. London: Oxford; 1932.
  13. 13. Wilkinson CM, Mautner SA. Measurement of eyeball protrusion and its application in facial reconstruction. J Forensic Sci. 2003;48(1):12–6. pmid:12570193
  14. 14. Rynn C, Wilkinson CM, Peters HL. Prediction of nasal morphology from the skull. Forensic Sci Med Pathol. 2010;6(1):20–34. pmid:19924578
  15. 15. Tedeschi-Oliveira SV, Beaini TL, Melani RF. Forensic facial reconstruction: Nasal projection in Brazilian adults. Forensic Sci Int. 2016;266:123–9. pmid:27258921
  16. 16. Stephan CN, Henneberg M. Predicting mouth width from inter-canine width-a 75% rule. J Forensic Sci. 2003;48(4):725–727. pmid:12877286
  17. 17. Dias PE, Miranda GE, Beaini TL, Melani RF. Practical application of anatomy of the oral cavity in forensic facial reconstruction. PLoS One. 2016;11(9):e0162732. pmid:27612284
  18. 18. Wilkinson C. Forensic facial reconstruction. 1st ed. New York:Cambridge University Press; 2004.
  19. 19. Richard AH, Parks CL, Monson KL. Assessment of presentation methods for ReFace computerized facial approximations. Forensic Sci Int. 2014;242:283–292. pmid:25128751
  20. 20. Decker S, Ford J, Davy-Jow S, Faraut P, Neville W, Hilbelink D. Who is this person? A comparison study of current three-dimensional facial approximation methods. Forensic Sci Int. 2013;229(1–3):161.e1–8.
  21. 21. Lee WJ, Wilkinson CM, Hwang HS, Lee SM. Correlation between average tissue depth data and quantitative accuracy of forensic craniofacial reconstructions measured by geometric surface comparison method. J Forensic Sci. 2015;60(3):572–580. pmid:25739646
  22. 22. Claes P, Vandermeulen D, De Greef S, Willems G, Suetens P. Craniofacial reconstruction using a combined statistical model of face shape and soft tissue depths: methodology and validation. Forensic Sci Int. 2006;159 Suppl 1:S147–58.
  23. 23. Short LJ, Khambay B, Ayoub A, Erolin C, Rynn C, Wilkinson C. Validation of a computer modelled forensic facial reconstruction technique using CT data from live subjects: a pilot study. Forensic Sci Int. 2014;237:147.e1–e8.
  24. 24. Claes P, Vandermeulen D, De Greef S, Willems G, Clement JG, Suetens P. Bayesian estimation of optimal craniofacial reconstructions. Forensic Sci Int. 2010;201(1–3):146–152. pmid:20359836
  25. 25. Tilotta F, Richard F, Glaunès J, Berar M, Gey S, Verdeille S, et al. Construction and analysis of a head CT-scan database for craniofacial reconstruction. Forensic Sci Int. 2009;191(1–3):112.e1–12.
  26. 26. Mazura JC, Juluru K, Chen JJ, Morgan TA, John M, Siegel EL. Facial recognition software success rates for the identification of 3D surface reconstructed facial images: implications for patient privacy and security. J Digit Imaging. 2012;25(3):347–351. pmid:22065158
  27. 27. Suppersberger Hamre S, Ersland GA, Daux V, Parson W, Wilkinson C. Three individuals, three stories, three burials from medieval Trondheim, Norway. PLoS One. 2017;12(7):e0180277. pmid:28671986
  28. 28. Moraes CAC, Dias PEM, Melani RFH. Demonstration of protocol for computer-aided forensic facial reconstruction with free software and photogrammetry. J Res Dentistry. 2014;2(1):77–90.
  29. 29. Turner WD, Brown RE, Kelliher TP, Tu PH, Taister MA, Miller KW. A novel method of automated skull registration for forensic facial approximation. Forensic Sci Int. 2005;154(2–3):149–58. pmid:16140487