1. Introduction
Facial acquisition and 3D imaging are useful in many fields of medicine such as maxillo-facial surgery, the production of prostheses, forensic medicine, and orthodontics [1,2,3]. Precise acquisition of 3D facial scanning incorporated with dental design software may improve treatment planning predictability [1,4,5,6]. Facial anthropometry, which made use of calipers and protractors to measure indices, was traditionally employed for facial acquisition [7]. However facial anthropometry, despite its simplicity, is operator-dependent and time-consuming, potentially inducing discomfort to the patients [8].
2D digital photography may be useful during dental treatment planning, for example, to visualize some measurements used to communicate with the patients. However, it is not possible to simulate a 3D object such as the human face; therefore, 2D photography is not suitable for the detection of facial volumes, deformities, and asymmetries [4,8,9].
Modern technologies overcame the limitations of direct anthropometry and 2D photography. Recently, 3D optical facial scanners have been commercialized with the aim of the acquisition of 3D patient facial images [10]. A facial scanner may provide a reliable method of facial digitization in order to create a virtual patient for treatment planning. The different 3D facial scanning technologies can be classified as follows: photogrammetry, stereophotogrammetry [11,12,13,14,15,16], laser beam scanning [9,17,18], structured light scanning [19,20], and dual structured light with infra-red sensor. Photogrammetry and stereophotogrammetry are passive scanning systems consisting of taking two or more photos from different perspectives; the 3D point cloud is then obtained from the common points through a reverse engineering software program. On the contrary, laser beam scanning and structured light scanning are active scanning systems; more specifically, light patterns projected at the patient are detected by a camera to obtain the 3D point cloud [19,20,21,22,23].
Stereophotogrammetry has been one of the most employed face scanning acquisition systems [24], which is fast (rapid capture of the face’s shape and texture), non-invasive, and produces accurate 3D images [4,25,26]. However, most professional stereophotogrammetry scanners are complex and expensive, and often require a long learning curve to optimally perform the scanning protocols [27,28,29,30]. Recently, interest in the use of smartphones with 3D depth sensor cameras that use structured light or time of flight technology for face scanning is spreading. These devices have the advantage of being portable, inexpensive, and popular [31,32,33,34]. Other smartphone scanning advantages include reduced time consumption for scanning image processing, and technical learning [35]. Different smartphone applications have been created for face digitalization in order to obtain 3D facial models transferable to dental computer-aided design (CAD) software [31,33].
Facial scanning is more challenging in children for different reasons. First of all, they may be uncooperative patients. Facial scanning, to be optimal, requires the subject to remain still for the entire time of acquisition and this may not be easy with a child. Furthermore, in the case of syndromic patients, some current technologies may have greater difficulties in the specific acquisition of any facial deformities. All these factors could affect the accuracy of the facial scan and therefore it is necessary to create new technologies that may overcome these difficulties and make the acquisition time less but with greater diagnostic accuracy. In this regard, close-range photogrammetry integrated with machine learning has been proposed and could be promising to create 3D children’s facial models [36].
The accuracy of these systems is still under evaluation, therefore, the aim of this study is to systematically review and compare, in the light of current knowledge, the accuracy of smartphone scanners versus stereophotogrammetry technology for facial digitization.
2. Materials and Methods
2.1. Protocol, Registration and Search Strategy
This systematic review followed the Preferred Reporting Items for Systematic Reviews and Meta-analyses (PRISMA) guidelines [37] and it has also been registered on the PROSPERO database (registration number: CRD42021241229). The population, intervention, comparison, and outcomes (PICO) question has been formulated as follows: are extra-oral scannings for the reproduction of a virtual facial model (P) obtained by smartphone technology (I) comparable to those obtained by stereophotogrammetry (C) in terms of accuracy (O)? The search strategy involved a combination of Mesh terms and free text words pooled through boolean operators (‘AND’, ‘OR’) and it has been performed on the following databases: PUBMED, Scopus, Web of Science, Cochrane Library, LILACS, and OpenGrey (Table 1).
2.2. Inclusion, Exclusion Criteria and Outcomes
The studies included in this systematic review were: randomized and non-randomized controlled trials, cohort studies, case-control studies, cross-sectional studies, retrospective studies, and thesis, only in English language and for a publication period from 1 January 2010 to 30 August 2022. Case reports, opinion articles, and reviews were excluded. The population of interest involved human faces or objects reproducing the human face; in particular, 3D virtual facial models obtained from optical facial scanners based on stereophotogrammetry or smartphone technology, but the studies with the use of 2D model images only were excluded. The main outcome of this systematic review consisted of the accuracy of facial measurements obtained by stereophotogrammetry versus smartphone applications scannings. Accuracy is intended as the discrepancy between the virtual facial model obtained through the scanner and a reference model. The deviation was measured in terms of inter-landmarks linear distances or surface-to-surface deviations.
2.3. Study Selection and Data Extraction
Records identified through database searches underwent an initial screening (title and abstract evaluation) where potential relevant articles were selected. The eligible articles underwent a further review in the full-text versions for adherence to the inclusion criteria. The data extracted from the studies included in the study are: authors, year of publication, aim of the study, characteristics of the sample and of the scanning method, standard reference for validation, characteristics of the measurements carried out, and conclusions. Disagreements between the review authors over the selection of particular articles were resolved by discussion, with involvement of a third review author, where necessary.
2.4. Risk of Bias Assessment
Two reviewers independently assessed the risk of bias in included studies using the Quality Assessment Tool for Diagnostic Accuracy Studies-2 (QUADAS-2) [38], which comprises 4 domains: patient selection, index test, reference standard, and flow and timing. When 1 or more of the key domains are scored as high risk, the study in question is considered with a high risk of bias. When more than 2 key domains are judged as unclear, the study is regarded with an unclear risk of bias. Disagreements between the review authors over the risk of bias in particular articles were resolved by discussion, with involvement of a third review author where necessary.
3. Results
3.1. Study Selection
The database searching led to the identification of 908 articles: 244 from Pubmed, 215 from Scopus, 320 from Web of Science, 102 from Cochrane Library, 25 from LILACS, and 2 from OpenGrey (Table 1). 75 duplicates were removed through the title review and, after the abstract screening of the 833 remaining articles, 43 full-text studies were assessed for eligibility. 23 articles were considered as eligible for the review analysis, whereas the other 20 papers were excluded for: no clear assessment of accuracy, no stereophotogrammetry or smartphone scanner, no statistical analysis, no English full-text, and no facial measurements. The PRISMA flow diagram of the search and evaluation process is illustrated in Figure 1.
3.2. Risk of Bias and Applicability Concern
Table 2 shows the risk of bias and application concerns according to the QUADAS-2 guidelines. Among 23 articles [39,40,41,42,43,44,45,46,47,48,49,50,51,52,53,54,55,56,57,58,59,60,61], 14 papers showed an overall low risk of bias [40,41,43,45,46,47,48,49,50,51,53,56,59,60,61], 3 articles had unclear risk of bias [39,44,52,57] and 4 articles displayed a high risk of bias [42,54,55,58]. Concerning applicability concerns, all articles showed an overall low level of concern, however, the domain “index test” showed more frequently an unclear concern because some papers did not sufficiently describe the scanning procedures and the used devices [40,46,55]. Regarding the quality assessment, the domain “patient selection” showed the highest risk of bias because of the exclusion of maxillofacial abnormalities [42,45,54,55,58] (this may result in overestimation of diagnostic accuracy [38]), the small number of participants [40,41,46,51,52] and the unclear method of random sampling [40,51,57]. Weighted bar chart for the risk of bias and application concerns of the selected studies is shown in Figure 2.
3.3. Description of the Included Studies
The included studies’ characteristics are shown in Table 3. Among the 23 studies, 17 involved adult volunteers [39,42,43,44,45,48,49,50,53,54,55,56,57,58,59,60,61] (the range of the number of participants was 5–80). The other 6 studies were conducted on impression casts of the face [41,46,51], mannequin heads [40,52], and human cadaver heads [47]. However, 2 studies [48,49] included both volunteers and mannequin heads.
In terms of tested scanning methods, all included studies analyzed stereophotogrammetry, smartphone scanning, or both. More specifically, thirteen articles tested the accuracy of stereophotogrammetry [41,42,43,45,47,49,50,51,52,56,59,60,61], six articles a smartphone structured light scanner [39,40,43,44,57,58], three articles smartphone photogrammetry [46,54,56],two articles a portable dual-structured light scanner connected to a tablet [52,55], seven articles one or more laser scanners [40,41,42,46,47,48,58], three articles a structured light scanner [53,58,59], three articles computed tomography [40,47,61] and two articles 2D camera photogrammetry [42,45].
Furthermore, the most used reference method (which generated the reference model) was direct anthropometry [40,42,43,45,47,50,55,58,59], followed by stereophotogrammetry [39,44,48,49,51,52,53,57], structured light scanner [54,56] and computed tomography [41,46]. In one paper, a reference model generated by a laser scanner was used [60] and another study used a coordinate-measuring machine as the gold standard [61].
The facial landmarks considered in the articles ranged from 6 to 29. The linear distances and angles analyzed ranged, respectively, from 5 to 25 and from 3 to 17. In one study [53], two lego bricks attached to participants’ faces were employed as a reference object to measure the scanner accuracy. Most of the included articles measured the global surface-to-surface deviation between the reference and test images obtained from the scannings [40,41,46,48,49,53,54,56,57,60].
4. Discussion
The aim of this review was to investigate the accuracy of smartphone scanners to generate digital face models compared to stereophotogrammetry. Compared to a previous systematic review [62] on face scanning, our study specifically analyzed the effectiveness of modern systems integrated into smartphones, comparing them with stereophotogrammetry. The selection of the articles was carried out on a greater number of databases (including gray literature). Furthermore, this systematic review included a greater number of selected articles (11 vs. 23) and the literature search is updated to August 2022 (vs. May 2020). Finally, the previous review included four articles concerning the use of smartphones/tablets, while ours included eleven studies.
Stereophotogrammetry is a passive scanning system that consists of capturing face surfaces through multiple photoshoots all taken at the same time from different perspectives. The 3D digital face model consisting of a dense cloud of points is obtained through software that uses the information of the camera’s position (with set angles and distances) and camera-to-subject distance [29,63]. This technology is able to reproduce very realistic and colored face models, but the accuracy is highly dependent on the camera’s resolution and the light conditions [41,64]. This system requires the application of standardized flash units to eliminate interference from ambient light and the careful assessment of camera settings such as brightness level, aperture, and shutter speed [29]. All this requires expensive investments and specific skills to be acquired.
Thirteen included articles reported the results of stationary stereophotogrammetry scanners to reproduce 3D facial models [39,41,44,45,47,48,49,50,52,53,59,60,61]. Among these, ten studies concerned the accuracy of stationary stereophotogrammetry systems [39,44,45,47,48,52,53,59,60,61]. All the articles concluded that this system is really reliable for reproducing digital face models with an accuracy comparable or superior to other systems such as direct anthropometry, CMM, laser, and structured light scanners. However, not all stationary stereophotogrammetry systems showed the same accuracy and the reference method used for comparison differed through the included articles.
The other three included articles that compared the accuracy of stationary and portable stereophotogrammetry scanners [41,49,50]. Two studies reported that portable stereophotogrammetry scanners (Scanify and Vectra H1) showed high accuracy, but lower compared to stationary devices (Danae 100SP, 3dMDface, and Vectra M3); moreover, portable scanners suffered more the influence of involuntary facial movements [41,49]. However, in a similar study conducted by Kim et al. [50], portable Vectra M1 and stationary Vectra M3 yielded similar values. The contrasting results (despite the use of the same devices, in particular for the study of Gibelli et al. [49]), may be explained by the different number of participants, anthropometric landmarks, methodology, and statistical analysis adopted. Therefore, portable stereophotogrammetry scanners showed to be reliable for clinical use, however, further investigations must be conducted to better understand the limitations and advantages of handled stereophotogrammetry scanners.
Two other articles evaluated selectively, the accuracy of portable stereophotogrammetry scanners [42,51]. Dindaroglu et al. [45] reported that Vectra H1 showed higher accuracy and reliability compared to 2D photography and a laser scanner for the morphological evaluation of soft tissues. Liu et al. compared a very low-cost stereophotogrammetry portable scanner (Scanify) to Vectra H1 and the authors concluded that the first device needs future improvements for clinical application [51].
In the included studies we found that stationary stereophotogrammetry devices showed a mean accuracy that ranged from 0.087 to 0.860 mm, portable stereophotogrammetry scanners from 0.150 to 0.849 mm, and smartphones/tablets from 0.460 to 1.400 mm (2D photogrammetry reported the highest values) (Table 4). A digital face scanner can be considered:
highly reliable, if mean accuracy is <1.0 mm,
reliable, if mean accuracy is 1.0–1.5 mm,
moderately reliable, if mean accuracy is 1.5–2.0 mm,
unreliable, if mean accuracy is >2.0 mm [65].
Results of the scanners’ mean accuracy (expressed in millimeters) in the selected studies. NA: not available. * SD not reported; 1 Stationary stereophotogrammetry; 2 Portable stereophotogrammetry; a smartphone 3D scanner; b 2D photogrammetry; c tablet connected to a portable dual-structured light scanner; P: photographs.
Study | Stereophotogrammetry |
Smartphone |
Structured Light Scanner |
Laser Scanner |
2D Camera (Mean ± SD) |
---|---|---|---|---|---|
Akan et al. (2022) [39] | - | iPhoneX a: 0.753 ± 0.113 | - | - | - |
Amornvit et al. (2019) [40] | - | iPhoneX a: NA | - | EinScan Pro: NA |
- |
Aswehlee et al. |
Danae1: 0.087 ± 0.002 |
- | - | Vivid 910: 0.068 ± 0.001 | - |
Ayaz et al. (2020) [42] | Vectra H1 2: 0.280 * | - | - | Planmeca ProFace: 0.500 * | Nikon D800 2D camera: 0.780 * |
Chong et al. (2021) [43] | Vectra H1 2: NA | iPad/iPhone b: NA | - | - | - |
D’Ettorre et al. (2022) [44] | - | iPhoneXs b (Bellus3D Face App): NA |
- | - | - |
Dindaroglu et al. (2016) [45] | 3dMDface 1: NA | - | - | - | Canon EOS 40D 2D camera: NA |
Elbashti et al. (2019) [46] | - | iPhone6 b: |
- | Vivid 910: 0.068 ± 0.001 | - |
Fourie et al. (2011) [47] | Di3D 1: 0.860 ± 0.570 | - | - | Vivid 900: 0.890 ± 0.560 | - |
Gibelli et al. (2018) [48] | Vectra M3 1: 0.650 ± 0.120 | - | - | Sense: 0.420 ± 0.170 | - |
Gibelli et al. (2018) [49] | Vectra H1 2: 0.440 ± 0.360 |
- | - | - | - |
Kim et al. (2018) [50] | Vectra H1 2: NA |
- | - | - | - |
Liu et al. (2019) [51] | Vectra H1 2: 0.15 ± 0.015 |
- | - | - | - |
Liu et al. (2021) [66] | 3dMDface 1: 0.36 ± 0.20 | Face Camera Pro Bellus c: 0.61 ± 0.47 | - | - | - |
Modabber et al. (2016) [53] | FaceScan 3D 1: 0.523 ± 0.144 | - | Artec EVA: 0.228 ± 0.051 | - | - |
Nightingale et al. (2020) [54] | - | iPhone8S b: |
Artec Spider: NA | - | - |
Piedra-Cascon et al. (2020) [55] | - | Face Camera Pro Bellus c: 0.910 ± 0.320 | - | - | - |
Ross et al. (2018) [56] | - | iPhone7 b: |
Artec Spider: NA |
- | - |
Rudy et al. (2020) [57] | Vectra H1 2: NA | iPhoneX a: 0.460 ± 0.010 | - | - | - |
Wang et al. (2022) [58] | - | iPad Pro 2020 b: 1.17 ± 0.80 | ARC-7 Face Scanning System: 0.76 ± 0.61 | EinScan Pro 2X Plus: 0.69 ± 0.65 | - |
Ye et al. (2016) [59] | 3dMDface 1: 0.620 ± 0.390 | - | 3D CaMega: 0.580 ± 0.370 | - | - |
Zhao et al. (2017) [60] | 3dMDface 1: 0.580 ± 0.110 |
- | - | Faro LLP: NA | - |
Zhao et al. (2021) [61] | 3dMD face 1: NA | - | - | - | - |
In clinical practice, facial models with deviations < 1.5 mm can be considered acceptable [59,60,67]. Therefore, accurate digital face models are reproduced from both stereophotogrammetry and smartphone scanning technology; however, smartphones seem to be less accurate as reported in two studies, in particular in measuring depth [40,46]. Therefore, a more inaccurate reproduction of the anatomically complex surfaces of the face than the flat ones must be expected, and it is not known whether this could imply an alteration in terms of clinical applications.
At first, the smartphone scanning method was based on a 2D photogrammetry approach in which the 3D model was produced from the matching of several photographs from different perspectives through a smartphone app [43,46,54,56]. However, differently from smartphone cameras, stereophotogrammetry usually makes use of digital single-lens reflex cameras with higher ISO settings, better noise reduction software, and higher pixel densities [68]. Five included studies evaluated smartphone/tablet 2D photogrammetry using different iPhone/iPad models [43,46,54,56,58]. Elbashti et al. (2019) [46] reported a good mean accuracy (0.605 ± 0.124 mm) for digitizing an impression cast of a facial defect, using a 24P (photographs) photogrammetry protocol. However, a commercially available laser scanner showed higher accuracy (0.068 ± 0.001 mm) in reference to standard computed tomography imaging. In the study of Nightingale et al. (2020) [54] 20 participants’ faces were scanned by novice operators. The reported accuracies for 40P, 60P, and 80P photogrammetry protocols were 0.800 ± 0.200 mm, 0.900 ± 0.400 mm, and 0.800 ± 0.300, all values that indicate a reliable accuracy. Ross et al. [56] reported a mean discrepancy of 1.200 ± 0.300 mm, 1.200 ± 0.200 mm, and 1.400 ± 0.600 mm with 30P, 60P, and 90P photogrammetry protocols. These higher values, but still <1.5 mm, could be due to the fact that an anatomically complex structure in the participants (ear) was selectively scanned. In any case, these results demonstrated that the ear can be scanned quite accurately using iPhone photographs. Recently, Chong et al. (2021) [43] developed an application enabling patients to capture their 3D facial images. Compared to a portable stereophotogrammetry Vectra H1 device, the measurements obtained with the iPhone/iPad application were also significantly correlated to the direct anthropometric measurements. However, the authors observed that the subnasale area needs improvement with that method. Similarly, Wang et al. (2022) [58] detected a sufficient reliability with the use of an app in the iPad Pro 2020 device, however, this system was less accurate compared to a portable laser scanner and a stationary structured light device.
Recently, to improve scanning accuracy (specifically for anatomically more complex surfaces), infrared structured-light depth-sensing cameras have been incorporated into smartphone devices [69]. The working principle of 3D depth-sensing cameras is similar to professional laser scanners and consists of the time-of-flight technique: the sensor array detects the time interval for infrared light to travel to the object and return to the sensor [70,71,72]. However, professional laser scanning systems have high-tech sensors more sensitive to depths [46,73]. Four included articles in this review evaluated digital face model accuracy through iPhone X/iPhone Xs incorporated scanner [39,40,44,57]. Amornvit et al. (2019) [40] analyzed the face scans obtained from a mannequin head and reported that the iPhone X obtained the fastest scan (0.57 min, in contrast to 6.7 and 9.4 min of the professional laser scanners) but it showed less accuracy, in particular in recording depths > 2 mm for which this technology should be not recommended according to the authors. This may be due to the failure of passing the light into the depth during scanning. In contrast, in a study of 16 participants, Rudy et al. (2020) [57] compared the iPhone X scanner to a portable stereophotogrammetry scanner (Vectra H1) used as a reference method. They reported a reliable mean accuracy (0.460 ± 0.010 mm) with the iPhone X scanner. The different results of these two studies may be explained by the different reference methods (respectively, direct anthropometry and stereophotogrammetry), the different methodology, scanner involved, and scanned objects (respectively, one mannequin head and sixteen participants). More recently, Akan et al. (2021) [39], as reported by the study of Amornivit et al., found that the iPhone X device with a depth-sensing camera is quite accurate compared to stationary stereophotogrammetry, but complex structures visualization needs improvements. The most recent included publication on the accuracy of these devices [44] reported a good reliability of the iPhone Xs with two different applications compared to the 3dMD system. However, portable smartphones, although less expensive, need operator accuracy and patient compliance for the incremented time of acquisition.
Finally, external infrared structured-light depth-sensing cameras can be plugged into smartphones, tablets, or laptop computers [74,75,76,77]. Two included studies reported the use of this approach [52,55]. Piedra-Cascon et al. (2020) [55] evaluated the accuracy of a dual structured-light scanner connected to a tablet in reproducing 3D facial models of 10 participants, using direct anthropometry as a reference method. They reported a mean accuracy of 0.910 ± 0.320 mm which is considered acceptable for virtual treatment planning. Similarly, in the recent article of Liu et al. (2021) [52], the Face Camera Pro Bellus system connected to a tablet and stationary stereophotogrammetry both showed a good accuracy and precision for clinical purposes compared to direct anthropometry. However, the use of external structured-light scanners implies that the overall accuracy should be interpreted as a result that includes the performance of the compatible mobile device, therefore, the accuracy should be evaluated for each combination of external scanner and mobile device [62].
One of the main limitations of portable face-scanning systems is motion artifacts that are induced by involuntary facial movements and showed to be the main source of error in the results of these scanners [49,77,78,79]. Therefore, diagnostic accuracy studies on these devices must be conducted on human living subjects for a correct assessment. The use of devices that capture the information with a single scan tends to be naturally more effective for this reason and is especially suitable for children and patients with special needs (Table 5).
Currently, imaging methods and new technologies are moving forward very quickly. New methods are emerging proposing the integration of target-based close-range photogrammetry and facial landmark machine learning detection through smartphone-based approaches [36]. Other promising systems are based on the Simultaneous Localization and Mapping (SLAM) technique and moving camera. This new technology allows the detection of the object’s real-time position through the creation of 3D point clouds. The advantage of this approach is the ability to obtain a high-resolution real-time 3D point with a reduced scanning capture time [80].
This review included only the most recent literature on the topic, regarding publications subsequent to 2010, because of rapid technological changes, but is limited to English publications. We observed a great heterogeneity in the adopted methodology for diagnostic accuracy studies. For example, in most included articles direct anthropometry was the reference method, thereby limiting the measurements practically only to Euclidean distances. Some studies were conducted in vitro and therefore the accuracy of reported values may be overestimated. For research that aims to study the clinical applicability of devices, living persons must be used to include the possibility of motion artifacts, especially for portable devices. Some studies employed novice operators without scanning experience, while others did not specify the operators’ experience or used experienced operators. Furthermore, a diffuse risk of bias in the selected studies was the exclusion of maxillofacial abnormalities that may result in an overestimation of diagnostic accuracy.
5. Conclusions
All the analyzed devices showed sufficiently reliable accuracy for clinical application. Stationary stereophotogrammetry scanners, followed by portable ones, showed higher accuracy than smartphones, particularly in the reproduction of complex anatomies.
Different factors affected the accuracy of facial scanning. Stereophotogrammetry showed itself to be highly reliable, however, the quality of the 3D images is dependent on the camera’s resolution, pixel integrity, and the light conditions during photo acquisition. A direct light may induce a glare effect that affects the quality of the acquisition. Therefore, positioned flash units and camera settings must be standardized to perform a good image exposure. On the other hand, smartphones seemed to be less accurate in the acquisition of irregular face surfaces, but the use of depth-sensing cameras may improve 3D image quality. Motion artifacts, induced by involuntary movements, may significantly affect facial scanning accuracy. For this reason, the use of smartphones or other portable scanners may be less suitable for children and uncooperative patients compared to devices with a single scan acquisition. However, new technologies involving the integration of machine learning or SLAM technique and moving camera may be promising to overcome these limitations and perform higher quality 3D face scannings.
The studies included in the review showed that the use of smartphones and tablets is currently practicable in the clinic for 3D facial scanning. The big advantages are the low cost and portability. However, compared to other professional devices, they require greater attention from the clinician and greater compliance by the patient who must be able to remain motionless for the entire duration of the acquisition time of the photos from different perspectives (risk of motion artifacts). The use of devices with internal depth-sensing cameras or external infrared structured-light depth-sensing cameras plugged into smartphones/tablets showed higher accuracy compared to the classic 2D photogrammetry matching approach through the use of some applications. In general, it has been reported that all devices are quite reliable for clinical practice, however, portable and especially stationary stereophotogrammetry remains the gold standard for greater accuracy in detecting deep and irregular surfaces and for shorter acquisition time. Finally, the various limitations and biases of the included articles may have led to an overestimation of the real accuracy of these devices.
A.P. and G.I. selected and read the papers included in this systematic review and drafted the manuscript. A.L.G. and V.Q. read the papers included in this systematic review, analyzed and interpreted the data. V.R., C.C., S.S. and R.J.M. analyzed the data and gave scientific support. All authors have read and agreed to the published version of the manuscript.
Not applicable.
Not applicable.
Not applicable.
The authors declare no conflict of interest.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Figure 2. Weighted barchart for the risk of bias and application concerns of the selected studies according to the QUADAS-2 guidelines.
Searching strategy in the different databases.
Database | Boolean Operator | Results |
---|---|---|
Pubmed | (“virtual patient” OR “virtual face” OR “digital facial models” OR “3D virtual facial model”) AND (“scanner” [MeSH Terms] OR “digital face scan” OR “facial scan” OR “3D face scanning” OR “facial digitalization” OR “stereophotogrammetry” OR “smartphone scanner” OR “smartphone face scanning” OR “smartphone application”) AND (“accuracy” OR “trueness” OR “precision” OR “3D comparison”). | 244 |
Scopus | (“virtual patient” OR “virtual face” OR “digital facial models” OR “3D virtual facial model”) AND (“scanner” OR “digital face scan” OR “facial scan” OR “3D face scanning” OR “facial digitalization” OR “stereophotogrammetry” OR “smartphone scanner” OR “smartphone face scanning” OR “smartphone application”) AND (“accuracy” OR “trueness” OR “precision” OR “3D comparison”). | 215 |
Web of Science | TS = ((virtual face OR digital facial models OR 3D virtual facial model OR digital face OR face) AND (scanner OR digital face scan OR facial scan OR 3D face scanning OR facial digitalization OR face digitalization OR stereophotogrammetry OR smartphone scanner OR smartphone face scanning OR smartphone application) AND (accuracy OR trueness OR precision OR 3D comparison)). | 320 |
Cochrane Library | “virtual patient” OR “virtual face” OR “digital facial models” OR “3D virtual facial model” AND “scanner” OR “digital face scan” OR “facial scan” OR “3D face scanning” OR “facial digitalization” OR “stereophotogrammetry” OR “smartphone scanner” OR “smartphone face scanning” OR “smartphone application” AND “accuracy” OR “trueness” OR “precision” OR “3D comparison”. | 102 |
LILACS | (digital facial models) AND (scanner OR stereophotogrammetry OR smartphone face scanning) AND (accuracy). | 25 |
OpenGrey | (virtual face OR digital facial models OR 3D virtual facial model OR digital face OR face) AND (scanner OR digital face scan OR facial scan OR 3D face scanning OR facial digitalization OR face digitalization OR stereophotogrammetry OR smartphone scanner OR smartphone face scanning OR smartphone application) AND (accuracy OR trueness OR precision OR 3D comparison). | 2 |
Total | 908 |
Risk of bias and application concerns according to the QUADAS-2 guidelines. [Image omitted. Please see PDF.] High risk, [Image omitted. Please see PDF.] Unclear risk, [Image omitted. Please see PDF.] Low risk.
Study | Risk of Bias | Applicability Concerns | |||||||
---|---|---|---|---|---|---|---|---|---|
Patient selection | Index test | Reference Standard | Flow and Timing | Overall risk of bias | Patient selection | Index test | Reference Standard | Overall risk of bias | |
Akan et al. (2021) [ |
[Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] |
Amornvit et al. (2019) [ |
[Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] |
Aswehlee et al.(2018) [ |
[Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] |
Ayaz et al. (2020) [ |
[Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] |
Chong et al. (2021) [ |
[Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] |
D’Ettorre (2022) et al. [ |
[Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] |
Dindaroglu et al. (2016) [ |
[Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] |
Elbashti et al. (2019) [ |
[Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] |
Fourie et al. (2011) [ |
[Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] |
Gibelli et al. (2018) [ |
[Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] |
Gibelli et al. (2018) [ |
[Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] |
Kim et al. (2018) [ |
[Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] |
Liu et al. (2019) [ |
[Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] |
Liu et al. (2021) [ |
[Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] |
Modabber et al. (2016) [ |
[Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] |
Nightingale et al. (2020) [ |
[Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] |
Piedra-Cascon et al. (2020) [ |
[Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] |
Ross et al. (2018) [ |
[Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] |
Rudy et al. (2020) [ |
[Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] |
Wang et al. (2022) [ |
[Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] |
Ye et al. (2016) [ |
[Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] |
Zhao et al. (2017) [ |
[Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] |
Zhao et al. (2021) [ |
[Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] |
Characteristics of the included studies. IC: impression cast, IPX: iPhone X; MH: Mannequin head; NA: not applicable; SPG: stereophotogrammetry. Stationary Stereophotogrammetry a1/Portable Stereophotogrammetry a2; Smartphone structured light scanner b1/photogrammetry b2; Tablet connected to a portable dual-structured light scanner b3; Laser scanner c; Structured light scanner d; Computed tomography e; RMSE: root-mean-square error (surface-to-surface).
Study | Aim | Tested Method | Sample | Reference Method | N° Landmarks | Measurements (Metrics) | Conclusions |
---|---|---|---|---|---|---|---|
Akan et al. (2021) [ |
To analyze the accuracy of a depth sensor smartphone camera with conventional stationary stereophotogrammetry | iPhone X with depth-sensing camera b1 | 26 participants (16 F, 10 M) | 3dMD face a1 | 9 | 7 linear distances, 3 angles (mean deviations, RMSE) | Depth sensor smartphone camera may be employed for 3D facial scanning. However, complex structures visualization needs improvements. |
Amornvit et al. (2019) [ |
To compare the facial scannings acquired through 4 scanner systems with the measurements obtained with direct anthropometry. | EinScan Pro (EP) c, EinScan Pro 2X Plus (EP+) c, iPhone X (IPX) b1, Planmeca ProMax 3D Mid (PM) e | 1 MH | Direct anthropometry | NA | Δx, Δy, and Δz (mean linear deviations) | Mean Δx and Δy measurements ranged respectively 10–50 mm and 50–120 mm. The records in the z-axis were not possible to acquire. EP+ was the most accurate, EP the intermediate, whereas IPX and PM showed less accuracy. Furthermore, EP, IPX, and PM were less accurate in measuring depths of 2 mm. |
Aswehlee et al. |
To evaluate the reliability of different digitalization systems for capturing facial defects. | Vivid 910 c, |
1 IC | Toshiba TOSCANER-30000μCM e | 3D point |
Surface deviation (RMSE) | All systems were sufficiently reliable. Laser-beam, light-sectioning technology showed the best accuracy. |
Ayaz et al. (2020) [ |
To evaluate the reliability of 2D photography and 3D scanning systems for facial digitalization. | Nikon D800 2D camera, Planmeca ProFace c, Vectra H1 a2 | 50 participants (25 M, 25 F) | Direct anthropometry | 22 | 7 linear distances, 17 angles (linear and angular mean deviation) | Stereophotogrammetry showed the best accuracy for facial digitalization compared to 2D photography and laser scanner. |
Chong et al. (2021) [ |
To evaluate the accuracy of an iPad/iPhone application enabling patients to capture their 3D facial images compared to direct anthropometry. | iPad/iPhone camera through the application “MeinXuan” b2, Vectra H1 a2 | 20 participants | Direct anthropometry | 18 | 21 linear distances (RMSE, mean absolute difference and relative error measurement) | The measurements obtained with the iPhone/iPad application were significantly correlated to the direct anthropometric measurements. The iPhone/iPad and Vectra H1 mean RMSE were 0.08 and 0.67 mm respectively. The subnasale area needs improvement with the proposed method. |
D’Ettorre et al. (2022) [ |
To compare the accuracy of 3D face scanning from 3D stereophotogrammetry and two different applications in smartphones supporting the TrueDepth system. | iPhone Xs equipped with Bellus3D Face or Capture applications b1 | 40 participants (27 F, 13 M) | 3dMD face a1 | 13 | Δx, Δy, and Δz (mean linear deviations) | Stationary stereophotogrammetry is a reliable and fast system. The smartphone applications showed also a good accuracy, are portable and less expensive. However, they need operator accuracy and patient compliance for the incremented time of acquisition. |
Dindaroglu et al. (2016) [ |
To compare the stereophotogrammetry accuracy with the direct manual and digital photogrammetry methods. | Canon EOS 40D 2D camera, 3dMD face a1 | 80 participants (38 M, 42 F) | Direct anthropometry | 15 | 10 linear distances, 6 angles (mean linear and angular deviations) | 3D stereophotogrammetry performed accurate 3D facial images reliable in orthodontics. |
Elbashti et al. (2019) [ |
To evaluate the accuracy of a smartphone application as a low-cost approach for digitizing a facial defect for 3D modeling. | Vivid 910 c, |
1 IC | Toshiba TOSCANER-30000μCM e | 3D point |
Surface deviation (RMSE) | In reference to standard computed tomography imaging, data acquisition with a smartphone for 3D modeling is not as accurate as commercially available laser scanning. |
Fourie et al. (2011) [ |
To estimate the reliability of three different 3D scanning systems namely laser surface scanning (Minolta Vivid 900), CBCT, 3D SPG (Di3D system) and to compare them to physical linear measurements. | Di3D a1, Vivid |
7 cadaver heads | Direct anthropometry | 15 | 21 linear distances (mean linear deviations) | Measurements recorded by the three 3D systems appeared to be both sufficiently accurate and reliable. |
Gibelli et al. (2018) [ |
To compare the accuracy of SPG with a low-cost laser scanner. | Sense c | 50 participants (10 M, 40 F), 1 MH | Vectra M3 a1 | 17; 3D point clouds | 14 linear distances, 12 angles/volumes/surfaces (RMSE) | The low-cost laser scanner appeared sufficiently reliable for immovable objects but it is not suitable for 3D human face scanning. |
Gibelli et al. (2018) [ |
To validate VECTRA H1 portable SPG device to verify its applicability to 3D facial analysis. | Vectra H1 a2 | 50 participants (16 M, 34 F), 1 MH | Vectra M3 a1 | 12; 3D point clouds | 15 linear distances, 12 angles/volumes/surfaces (RMSE) | The portable Vectra H1 face scanning device proved reliable for assessing linear measurements, angles, and surface areas; conversely, the influence of involuntary facial movements on volumes and RMSE was higher compared to the stationary Vectra M3. |
Kim et al. (2018) [ |
To evaluate the accuracy and reliability of a 3D small-format, handheld camera. | Vectra H1 a2, Vectra M3 a1 | 5 participants (NA) | Direct anthropometry | 29 | 25 linear distances (mean linear deviations) | The 3D handheld camera showed high accuracy and reliability in comparison with traditional models. |
Liu et al. (2019) [ |
To evaluate reliability of a portable low-cost scanner (Scanify) for imaging facial casts compared to a previously validated portable digital stereophotogrammetry device (Vectra H1). | Scanify a2 | 2 IC (male) | Vectra H1 a2 | 13 | 11 linear distances (mean linear deviations) | Scanify showed to be a low-cost solution for facial digitalization but needs future improvements. |
Liu et al. (2021) [ |
To evaluate the reliability of Bellus3D Face Camera Pro compared to 3dMDface stereophotogrammetry for face scanning. | Face Camera Pro Bellus b3 (connected to a tablet), 3dMD face a1 | 1 MH | Direct anthropometry | 20 | 8 linear distances, 5 angles (absolute mean deviations) | Both systems showed good accuracy and precision for clinical purposes. |
Modabber et al. (2016) [ |
To evaluate the reliability of two devices for face digitalization. | Artec EVA d | 41 participants (16 M, 25 F) | FaceScan 3D a1 | 2 lego brick | Surface deviation (RMSE) | Artec EVA showed greater accuracy compared to FaceScan3D. |
Nightingale et al. (2020) [ |
To test the use of a low-cost scanner by inexperienced operators. | iPhone 8S b2 (40, 60 or 80 photographs) | 20 participants (11 M, 9 F) | Artec Spider d | 3D point clouds | Surface deviation (RMSE) | Smartphone photogrammetry showed to be useful for novice operators for its low cost and easy learning curve. |
Piedra-Cascon et al. (2020) [ |
To evaluate the reliability of Bellus3D Face Camera Pro for face digitalization. | Face Camera Pro Bellus b3 (connected to a tablet) | 10 participants (2 M, 8 F) | Direct anthropometry | 6 | 5 linear distances (precision and accuracy RMSE) | The new device showed to be clinically reliable for face scanning procedures. |
Ross et al. (2018) [ |
To estimate the performance of smartphones for external ear digitalization. | iPhone 7 b2 (30, 60 or |
16 participants (8 M, 8 F) | Artec Spider d | 6 | Surface deviation (RMSE) | The three protocols with 30-60-90 photographs showed a good and similar accuracy. The Intel RealSense showed the worst performance and resolution. |
Rudy et al. (2020) [ |
To compare the accuracy |
iPhone X b1 | 16 participants (NA) | Vectra H1 a2 | 10; 3D point cloud | Landmark-to-landmark surface distances; surface deviation (RMSE) | The iPhoneX accuracy stood around 0.5 mm when compared to Vectra H1. |
Wang et al. (2022) [ |
To compare the accuracy of two portable systems and a stationary scanner for facial digitalization. | iPad Pro 2020 b2 (Bellus 3D Dental Pro app), EinScan Pro 2X Plus (EP+) c, ARC-7 Face Scanning System d | 20 participants (5 M, 15 F) | Direct anthropometry | 12 | 14 linear distances (absolute error) | All the systems showed to be quite reliable for face digitalization. However, the iPad Pro 2020 system was the least accurate. |
Ye et al. (2016) [ |
To assess the accuracy, reliability and reproducibility of facial digitalization through stereophotogrammetry compared to a structured light scanner. | 3D CaMega d, |
10 participants (5 M, 5 F) | Direct anthropometry | 16 | 21 linear distances (mean linear deviations) | Both scanners showed to be quite accurate, reliable and reproducible to create 3D facial models. |
Zhao et al. (2017) [ |
To evaluate the accuracy of different scanning systems for face digitalization. | 3dMDfacea1, |
10 participants (NA) | Faro Edge LLP c | 3D point clouds | Surface deviation (RMSE) | All scanning systems showed to be sufficiently reliable for clinic purposes. |
Zhao et al. (2021) [ |
To evaluate the accuracy of stereophotogrammetry and a CBCT system to obtain facial models with deformities and partitions. | 3dMDfacea1, |
60 participants (28 M, 32 F) | Coordinate-measuring |
17 | 19 linear distances (mean linear deviations) | 3D stereophotogrammetry was more than CBCT in the acquisition of facial deformities. |
Comparison of stationary/portable stereophotogrammetry and smartphone scanning face systems.
Parameters | Stationary Stereophotogrammetry | Portable Stereophotogrammetry | Smartphones |
---|---|---|---|
Mean accuracy | 0.087–0.860 mm | 0.150–0.849 mm | 0.460–1.400 mm |
Costs | 8.000–26.000$ | 1.500–10.000$ | 350–1.200$ |
Capture time * | 1.5–600 ms | 3.5–250 ms | 2.5–350 ms |
Approach | Different cameras placed at different angulations capture various 2D images simultaneously to create a 3D face model. | A single camera capture one image at a time. Require several acquisitions from different angles to reconstruct the 3D facial model. | A single camera capture one image at a time. Require several acquisitions from different angles to reconstruct the 3D facial model. |
Advantages | Highly reliable system. Useful in children and uncooperative patients. | Reliable and portable system. | Comfortable, manageable, quite reliable, easy to use, inexpensive. |
Limitations | Expensive, requires space and ambient light control. | Expensive, ambient light control, motion artifacts, greater acquisition time, require patient collaboration. | Motion artifacts, less accurate on uneven and deep surfaces, greater acquisition time, require patient collaboration. |
* It refers to a single scan.
References
1. Launonen, A.M.; Vuollo, V.; Aarnivala, H.; Heikkinen, T.; Pirttiniemi, P.; Valkama, A.M.; Harila, V. Craniofacial asymmetry from one to three years of age: A prospective cohort study with 3d imaging. J. Clin. Med.; 2020; 9, 70. [DOI: https://dx.doi.org/10.3390/jcm9010070] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/31892114]
2. Lo Giudice, A.; Ortensi, L.; Farronato, M.; Lucchese, A.; Lo Castro, E.; Isola, G. The step further smile virtual planning: Milled versus prototyped mock-ups for the evaluation of the designed smile characteristics. BMC Oral Health; 2020; 20, 165. [DOI: https://dx.doi.org/10.1186/s12903-020-01145-z] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/32503567]
3. Giudice, A.L.; Spinuzza, P.; Rustico, L.; Messina, G.; Nucera, R. Short-term treatment effects produced by rapid maxillary expansion evaluated with computed tomography: A systematic review with meta-analysis. Korean J. Orthod.; 2020; 50, pp. 314-323. [DOI: https://dx.doi.org/10.4041/kjod.2020.50.5.314] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/32938824]
4. Elshewy, M. Assessment of 3D Facial Scan Integration in 3D Digital Workflow Using Radiographic Markers and Iterative Closest Point Algorithm. Ph.D. Thesis; Marquette University: Milwaukee, WI, USA, 2020.
5. Berssenbrügge, P.; Berlin, N.F.; Kebeck, G.; Runte, C.; Jung, S.; Kleinheinz, J.; Dirksen, D. 2D and 3D analysis methods of facial asymmetry in comparison. J. Cranio-Maxillofac. Surg.; 2014; 42, pp. e327-e334. [DOI: https://dx.doi.org/10.1016/j.jcms.2014.01.028]
6. Chu, Y.; Yang, J.; Ma, S.; Ai, D.; Li, W.; Song, H.; Li, L.; Chen, D.; Chen, L.; Wang, Y. Registration and fusion quantification of augmented reality based nasal endoscopic surgery. Med. Image Anal.; 2017; 42, pp. 241-256. [DOI: https://dx.doi.org/10.1016/j.media.2017.08.003]
7. Douglas, T.S. Image processing for craniofacial landmark identification and measurement: A review of photogrammetry and cephalometry. Comput. Med. Imaging Graph.; 2004; 28, pp. 401-409. [DOI: https://dx.doi.org/10.1016/j.compmedimag.2004.06.002]
8. Mai, H.-N.; Kim, J.; Choi, Y.-H.; Lee, D.-H. Accuracy of Portable Face-Scanning Devices for Obtaining Three-Dimensional Face Models: A Systematic Review and Meta-Analysis. Int. J. Environ. Res. Public Health; 2021; 18, 94. [DOI: https://dx.doi.org/10.3390/ijerph18010094]
9. Plooij, J.M.; Maal, T.J.; Haers, P.; Borstlap, W.A.; Kuijpers-Jagtman, A.M.; Bergé, S.J. Digital three-dimensional image fusion processes for planning and evaluating orthodontics and orthognathic surgery. A systematic review. Int. J. Oral Maxillofac. Surg.; 2011; 40, pp. 341-352. [DOI: https://dx.doi.org/10.1016/j.ijom.2010.10.013]
10. Joda, T.; Gallucci, G.O. The virtual patient in dental medicine. Clin. Oral Implant. Res.; 2015; 26, pp. 725-726. [DOI: https://dx.doi.org/10.1111/clr.12379]
11. Cascón, W.P.; de Gopegui, J.R.; Revilla-León, M. Facially generated and additively manufactured baseplate and occlusion rim for treatment planning a complete-arch rehabilitation: A dental technique. J. Prosthet. Dent.; 2019; 121, pp. 741-745. [DOI: https://dx.doi.org/10.1016/j.prosdent.2018.07.009]
12. Kau, C.H.; Richmond, S.; Zhurov, A.; Ovsenik, M.; Tawfik, W.; Borbely, P.; English, J.D. Use of 3-dimensional surface acquisition to study facial morphology in 5 populations. Am. J. Orthod. Dentofac. Orthop.; 2010; 137, pp. S56.e51-S56.e59. [DOI: https://dx.doi.org/10.1016/j.ajodo.2010.01.022]
13. Metzler, P.; Sun, Y.; Zemann, W.; Bartella, A.; Lehner, M.; Obwegeser, J.A.; Kruse-Gujer, A.L.; Lübbers, H.-T. Validity of the 3D VECTRA photogrammetric surface imaging system for cranio-maxillofacial anthropometric measurements. Oral Maxillofac. Surg.; 2014; 18, pp. 297-304. [DOI: https://dx.doi.org/10.1007/s10006-013-0404-7] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/23559195]
14. Nord, F.; Ferjencik, R.; Seifert, B.; Lanzer, M.; Gander, T.; Matthews, F.; Rücker, M.; Lübbers, H.-T. The 3dMD photogrammetric photo system in cranio-maxillofacial surgery: Validation of interexaminer variations and perceptions. J. Cranio-Maxillofac. Surg.; 2015; 43, pp. 1798-1803. [DOI: https://dx.doi.org/10.1016/j.jcms.2015.08.017] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/26421470]
15. Verhoeven, T.; Xi, T.; Schreurs, R.; Bergé, S.; Maal, T. Quantification of facial asymmetry: A comparative study of landmark-based and surface-based registrations. J. Cranio-Maxillofac. Surg.; 2016; 44, pp. 1131-1136. [DOI: https://dx.doi.org/10.1016/j.jcms.2016.07.017]
16. Weinberg, S.M.; Naidoo, S.; Govier, D.P.; Martin, R.A.; Kane, A.A.; Marazita, M.L. Anthropometric precision and accuracy of digital three-dimensional photogrammetry: Comparing the Genex and 3dMD imaging systems with one another and with direct anthropometry. J. Craniofacial Surg.; 2006; 17, pp. 477-483. [DOI: https://dx.doi.org/10.1097/00001665-200605000-00015]
17. Weinberg, S.M.; Kolar, J.C. Three-dimensional surface imaging: Limitations and considerations from the anthropometric perspective. J. Craniofacial Surg.; 2005; 16, pp. 847-851. [DOI: https://dx.doi.org/10.1097/01.SCS.0000164330.96675.0C]
18. Karatas, O.H.; Toy, E. Three-dimensional imaging techniques: A literature review. Eur. J. Dent.; 2014; 8, 132. [DOI: https://dx.doi.org/10.4103/1305-7456.126269]
19. Ma, L.; Xu, T.; Lin, J. Validation of a three-dimensional facial scanning system based on structured light techniques. Comput. Methods Programs Biomed.; 2009; 94, pp. 290-298. [DOI: https://dx.doi.org/10.1016/j.cmpb.2009.01.010]
20. Li, G.; Wei, J.; Wang, X.; Wu, G.; Ma, D.; Wang, B.; Liu, Y.; Feng, X. Three-dimensional facial anthropometry of unilateral cleft lip infants with a structured light scanning system. J. Plast. Reconstr. Aesthetic Surg.; 2013; 66, pp. 1109-1116. [DOI: https://dx.doi.org/10.1016/j.bjps.2013.04.007]
21. Vlaar, S.T.; van der Zel, J.M. Accuracy of dental digitizers. Int. Dent. J.; 2006; 56, pp. 301-309. [DOI: https://dx.doi.org/10.1111/j.1875-595X.2006.tb00105.x]
22. Beuer, F.; Schweiger, J.; Edelhoff, D. Digital dentistry: An overview of recent developments for CAD/CAM generated restorations. Br. Dent. J.; 2008; 204, pp. 505-511. [DOI: https://dx.doi.org/10.1038/sj.bdj.2008.350] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/18469768]
23. De Felice, M.E.; Nucci, L.; Fiori, A.; Flores-Mir, C.; Perillo, L.; Grassia, V. Accuracy of interproximal enamel reduction during clear aligner treatment. Prog. Orthod.; 2020; 21, 28. [DOI: https://dx.doi.org/10.1186/s40510-020-00329-1] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/32719906]
24. Gwilliam, J.R.; Cunningham, S.J.; Hutton, T. Reproducibility of soft tissue landmarks on three-dimensional facial scans. Eur. J. Orthod.; 2006; 28, pp. 408-415. [DOI: https://dx.doi.org/10.1093/ejo/cjl024] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/16901962]
25. Winder, R.; Darvann, T.A.; McKnight, W.; Magee, J.; Ramsay-Baggs, P. Technical validation of the Di3D stereophotogrammetry surface imaging system. Br. J. Oral Maxillofac. Surg.; 2008; 46, pp. 33-37. [DOI: https://dx.doi.org/10.1016/j.bjoms.2007.09.005]
26. Toma, A.M.; Zhurov, A.; Playle, R.; Richmond, S. A three-dimensional look for facial differences between males and females in a British-Caucasian sample aged 151/2 years old. Orthod. Craniofacial Res.; 2008; 11, pp. 180-185. [DOI: https://dx.doi.org/10.1111/j.1601-6343.2008.00428.x]
27. Sawyer, A.; See, M.; Nduka, C. Quantitative analysis of normal smile with 3D stereophotogrammetry–an aid to facial reanimation. J. Plast. Reconstr. Aesthetic Surg.; 2010; 63, pp. 65-72. [DOI: https://dx.doi.org/10.1016/j.bjps.2008.08.062]
28. Knoops, P.G.; Beaumont, C.A.; Borghi, A.; Rodriguez-Florez, N.; Breakey, R.W.; Rodgers, W.; Angullia, F.; Jeelani, N.O.; Schievano, S.; Dunaway, D.J. Comparison of three-dimensional scanner systems for craniomaxillofacial imaging. J. Plast. Reconstr. Aesthetic Surg.; 2017; 70, pp. 441-449. [DOI: https://dx.doi.org/10.1016/j.bjps.2016.12.015]
29. Tzou, C.-H.J.; Artner, N.M.; Pona, I.; Hold, A.; Placheta, E.; Kropatsch, W.G.; Frey, M. Comparison of three-dimensional surface-imaging systems. J. Plast. Reconstr. Aesthetic Surg.; 2014; 67, pp. 489-497. [DOI: https://dx.doi.org/10.1016/j.bjps.2014.01.003]
30. Dekel, E.; Nucci, L.; Weill, T.; Flores-Mir, C.; Becker, A.; Perillo, L.; Chaushu, S. Impaction of maxillary canines and its effect on the position of adjacent teeth and canine development: A cone-beam computed tomography study. Am. J. Orthod. Dentofac. Orthop.; 2021; 159, pp. e135-e147. [DOI: https://dx.doi.org/10.1016/j.ajodo.2020.10.015]
31. Granata, S.; Giberti, L.; Vigolo, P.; Stellini, E.; Di Fiore, A. Incorporating a facial scanner into the digital workflow: A dental technique. J. Prosthet. Dent.; 2020; 123, pp. 781-785. [DOI: https://dx.doi.org/10.1016/j.prosdent.2019.05.021]
32. Hong, S.-J.; Noh, K. Setting the sagittal condylar inclination on a virtual articulator by using a facial and intraoral scan of the protrusive interocclusal position: A dental technique. J. Prosthet. Dent.; 2020; 125, pp. 392-395. [DOI: https://dx.doi.org/10.1016/j.prosdent.2020.01.031] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/32220398]
33. Russo, L.L.; Salamini, A.; Troiano, G.; Guida, L. Digital dentures: A protocol based on intraoral scans. J. Prosthet. Dent.; 2020; 125, pp. 597-602. [DOI: https://dx.doi.org/10.1016/j.prosdent.2020.02.006] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/32331785]
34. Revilla-León, M.; Raney, L.; Piedra-Cascón, W.; Barrington, J.; Zandinejad, A.; Özcan, M. Digital workflow for an esthetic rehabilitation using a facial and intraoral scanner and an additive manufactured silicone index: A dental technique. J. Prosthet. Dent.; 2020; 123, pp. 564-570. [DOI: https://dx.doi.org/10.1016/j.prosdent.2019.03.014] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/31353112]
35. Liu, C.-H.; Lin, I.-C.; Lu, J.-J.; Cai, D. A Smartphone App for Improving Clinical Photography in Emergency Departments: Comparative Study. JMIR Mhealth Uhealth; 2019; 7, e14531. [DOI: https://dx.doi.org/10.2196/14531]
36. Barbero-García, I.; Pierdicca, R.; Paolanti, M.; Felicetti, A.; Lerma, J.L. Combining machine learning and close-range photogrammetry for infant’s head 3D measurement: A smartphone-based solution. Measurement; 2021; 182, 109686. [DOI: https://dx.doi.org/10.1016/j.measurement.2021.109686]
37. Liberati, A.; Altman, D.G.; Tetzlaff, J.; Mulrow, C.; Gøtzsche, P.C.; Ioannidis, J.P.; Clarke, M.; Devereaux, P.J.; Kleijnen, J.; Moher, D. The PRISMA statement for reporting systematic reviews and meta-analyses of studies that evaluate health care interventions: Explanation and elaboration. J. Clin. Epidemiol.; 2009; 62, pp. e1-e34. [DOI: https://dx.doi.org/10.1016/j.jclinepi.2009.06.006]
38. Whiting, P.F.; Rutjes, A.W.; Westwood, M.E.; Mallett, S.; Deeks, J.J.; Reitsma, J.B.; Leeflang, M.M.; Sterne, J.A.; Bossuyt, P.M. QUADAS-2: A revised tool for the quality assessment of diagnostic accuracy studies. Ann. Intern. Med.; 2011; 155, pp. 529-536. [DOI: https://dx.doi.org/10.7326/0003-4819-155-8-201110180-00009]
39. Akan, B.; Akan, E.; Sahan, A.O.; Kalak, M. Evaluation of 3D Face-Scan images obtained by stereophotogrammetry and smartphone camera. Int. Orthod.; 2021; 19, pp. 669-678. [DOI: https://dx.doi.org/10.1016/j.ortho.2021.08.007]
40. Amornvit, P.; Sanohkan, S. The Accuracy of Digital Face Scans Obtained from 3D Scanners: An In Vitro Study. Int. J. Environ. Res. Public Health; 2019; 16, 5061. [DOI: https://dx.doi.org/10.3390/ijerph16245061]
41. Aswehlee, A.M.; Elbashti, M.E.; Hattori, M.; Sumita, Y.I.; Taniguchi, H. Feasibility and Accuracy of Noncontact Three-Dimensional Digitizers for Geometric Facial Defects: An In Vitro Comparison. Int. J. Prosthodont.; 2018; 31, pp. 601-606. [DOI: https://dx.doi.org/10.11607/ijp.5855]
42. Ayaz, I.; Shaheen, E.; Aly, M.; Shujaat, S.; Gallo, G.; Coucke, W.; Politis, C.; Jacobs, R. Accuracy and reliability of 2-dimensional photography versus 3-dimensional soft tissue imaging. Imaging Sci. Dent.; 2020; 50, pp. 15-22. [DOI: https://dx.doi.org/10.5624/isd.2020.50.1.15] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/32206616]
43. Chong, Y.M.; Liu, X.Y.; Shi, M.; Huang, J.Z.; Yu, N.Z.; Long, X. Three-dimensional facial scanner in the hands of patients: Validation of a novel application on iPad/iPhone for three-dimensional imaging. Ann. Transl. Med.; 2021; 9, 1115. [DOI: https://dx.doi.org/10.21037/atm-21-1620] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/34430556]
44. D’Ettorre, G.; Farronato, M.; Candida, E.; Quinzi, V.; Grippaudo, C. A comparison between stereophotogrammetry and smartphone structured light technology for three-dimensional face scanning. Angle Orthod.; 2022; 92, pp. 358-363. [DOI: https://dx.doi.org/10.2319/040921-290.1]
45. Dindaroglu, F.; Kutlu, P.; Duran, G.S.; Gorgulu, S.; Aslan, E. Accuracy and reliability of 3D stereophotogrammetry: A comparison to direct anthropometry and 2D photogrammetry. Angle Orthod.; 2016; 86, pp. 487-494. [DOI: https://dx.doi.org/10.2319/041415-244.1] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/26267357]
46. Elbashti, M.E.; Sumita, Y.I.; Aswehlee, A.M.; Seelaus, R. Smartphone Application as a Low-Cost Alternative for Digitizing Facial Defects: Is It Accurate Enough for Clinical Application?. Int. J. Prosthodont.; 2019; 32, pp. 541-543. [DOI: https://dx.doi.org/10.11607/ijp.6347] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/31664272]
47. Fourie, Z.; Damstra, J.; Gerrits, P.O.; Ren, Y. Evaluation of anthropometric accuracy and reliability using different three-dimensional scanning systems. Forensic Sci. Int.; 2011; 207, pp. 127-134. [DOI: https://dx.doi.org/10.1016/j.forsciint.2010.09.018]
48. Gibelli, D.; Pucciarelli, V.; Caplova, Z.; Cappella, A.; Dolci, C.; Cattaneo, C.; Sforza, C. Validation of a low-cost laser scanner device for the assessment of three-dimensional facial anatomy in living subjects. J. Cranio-Maxillofac. Surg.; 2018; 46, pp. 1493-1499. [DOI: https://dx.doi.org/10.1016/j.jcms.2018.06.009]
49. Gibelli, D.; Pucciarelli, V.; Cappella, A.; Dolci, C.; Sforza, C. Are portable stereophotogrammetric devices reliable in facial imaging? A validation study of VECTRA H1 device. J. Oral Maxillofac. Surg.; 2018; 76, pp. 1772-1784. [DOI: https://dx.doi.org/10.1016/j.joms.2018.01.021]
50. Kim, A.J.; Gu, D.; Chandiramani, R.; Linjawi, I.; Deutsch, I.C.K.; Allareddy, V.; Masoud, M.I. Accuracy and reliability of digital craniofacial measurements using a small-format, handheld 3D camera. Orthod. Craniofac. Res.; 2018; 21, pp. 132-139. [DOI: https://dx.doi.org/10.1111/ocr.12228]
51. Liu, C.; Artopoulos, A. Validation of a low-cost portable 3-dimensional face scanner. Imaging Sci. Dent.; 2019; 49, pp. 35-43. [DOI: https://dx.doi.org/10.5624/isd.2019.49.1.35]
52. Liu, J.L.; Zhang, C.H.; Cai, R.L.; Yao, Y.; Zhao, Z.H.; Liao, W. Accuracy of 3-dimensional stereophotogrammetry: Comparison of the 3dMD and Bellus3D facial scanning systems with one another and with direct anthropometry. Am. J. Orthod. Dentofac. Orthop.; 2021; 160, pp. 862-871. [DOI: https://dx.doi.org/10.1016/j.ajodo.2021.04.020] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/34814981]
53. Modabber, A.; Peters, F.; Kniha, K.; Goloborodko, E.; Ghassemi, A.; Lethaus, B.; Holzle, F.; Mohlhenrich, S.C. Evaluation of the accuracy of a mobile and a stationary system for three-dimensional facial scanning. J. Cranio-Maxillofac. Surg.; 2016; 44, pp. 1719-1724. [DOI: https://dx.doi.org/10.1016/j.jcms.2016.08.008] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/27614543]
54. Nightingale, R.C.; Ross, M.T.; Allenby, M.C.; Woodruff, M.A.; Powell, S.K. A Method for Economical Smartphone-Based Clinical 3D Facial Scanning. J. Prosthodont. -Implant. Esthet. Reconstr. Dent.; 2020; 29, pp. 818-825. [DOI: https://dx.doi.org/10.1111/jopr.13274] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/33089546]
55. Piedra-Cascon, W.; Meyer, M.J.; Methani, M.M.; Revilla-Leon, M. Accuracy (trueness and precision) of a dual-structured light facial scanner and interexaminer reliability. J. Prosthet. Dent.; 2020; 124, pp. 567-574. [DOI: https://dx.doi.org/10.1016/j.prosdent.2019.10.010] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/31918895]
56. Ross, M.T.; Cruz, R.; Brooks-Richards, T.L.; Hafner, L.M.; Powell, S.K.; Woodruff, M.A. Comparison of three-dimensional surface scanning techniques for capturing the external ear. Virtual Phys. Prototyp.; 2018; 13, pp. 255-265. [DOI: https://dx.doi.org/10.1080/17452759.2018.1493803]
57. Rudy, H.L.; Wake, N.; Yee, J.; Garfein, E.S.; Tepper, O.M. Three-Dimensional Facial Scanning at the Fingertips of Patients and Surgeons: Accuracy and Precision Testing of iPhone X Three-Dimensional Scanner. Plast. Reconstr. Surg.; 2020; 146, pp. 1407-1417. [DOI: https://dx.doi.org/10.1097/PRS.0000000000007387]
58. Wang, C.; Shi, Y.F.; Xiong, Q.; Xie, P.J.; Wu, J.H.; Liu, W.C. Trueness of One Stationary and Two Mobile Systems for Three-Dimensional Facial Scanning. Int. J. Prosthodont.; 2022; 35, pp. 350-356. [DOI: https://dx.doi.org/10.11607/ijp.7746]
59. Ye, H.Q.; Lv, L.W.; Liu, Y.S.; Liu, Y.S.; Zhou, Y.S. Evaluation of the Accuracy, Reliability, and Reproducibility of Two Different 3D Face-Scanning Systems. Int. J. Prosthodont.; 2016; 29, pp. 213-218. [DOI: https://dx.doi.org/10.11607/ijp.4397]
60. Zhao, Y.J.; Xiong, Y.X.; Wang, Y. Three-Dimensional Accuracy of Facial Scan for Facial Deformities in Clinics: A New Evaluation Method for Facial Scanner Accuracy. PLoS ONE; 2017; 12, e0169402. [DOI: https://dx.doi.org/10.1371/journal.pone.0169402]
61. Zhao, Z.; Xie, L.; Cao, D.; Izadikhah, I.; Gao, P.; Zhao, Y.; Yan, B. Accuracy of three-dimensional photogrammetry and cone beam computed tomography based on linear measurements in patients with facial deformities. Dentomaxillofac. Radiol.; 2021; 50, 20200001. [DOI: https://dx.doi.org/10.1259/dmfr.20200001]
62. Mai, H.-N.; Lee, D.-H. Accuracy of Mobile Device–Compatible 3D Scanners for Facial Digitization: Systematic Review and Meta-Analysis. J. Med. Internet Res.; 2020; 22, e22228. [DOI: https://dx.doi.org/10.2196/22228] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/33095178]
63. Volonghi, P.; Baronio, G.; Signoroni, A. 3D scanning and geometry processing techniques for customised hand orthotics: An experimental assessment. Virtual Phys. Prototyp.; 2018; 13, pp. 105-116. [DOI: https://dx.doi.org/10.1080/17452759.2018.1426328]
64. Camison, L.; Bykowski, M.; Lee, W.W.; Carlson, J.C.; Roosenboom, J.; Goldstein, J.A.; Losee, J.E.; Weinberg, S.M. Validation of the Vectra H1 portable three-dimensional photogrammetry system for facial imaging. Int. J. Oral Maxillofac. Surg.; 2018; 47, pp. 403-410. [DOI: https://dx.doi.org/10.1016/j.ijom.2017.08.008]
65. Aung, S.; Ngim, R.; Lee, S. Evaluation of the laser scanner as a surface measuring tool and its accuracy compared with direct facial anthropometric measurements. Br. J. Plast. Surg.; 1995; 48, pp. 551-558. [DOI: https://dx.doi.org/10.1016/0007-1226(95)90043-8]
66. Liu, L.; Zhan, Q.; Zhou, J.; Kuang, Q.; Yan, X.; Zhang, X.; Shan, Y.; Li, X.; Lai, W.; Long, H. Effectiveness of an anterior mini-screw in achieving incisor intrusion and palatal root torque for anterior retraction with clear aligners. Angle Orthod.; 2021; 91, pp. 794-803. [DOI: https://dx.doi.org/10.2319/120420-982.1]
67. Secher, J.J.; Darvann, T.A.; Pinholt, E.M. Accuracy and reproducibility of the DAVID SLS-2 scanner in three-dimensional facial imaging. J. Cranio-Maxillofac. Surg.; 2017; 45, pp. 1662-1670. [DOI: https://dx.doi.org/10.1016/j.jcms.2017.07.006]
68. Lane, C.; Harrell, W., Jr. Completing the 3-dimensional picture. Am. J. Orthod. Dentofac. Orthop.; 2008; 133, pp. 612-620. [DOI: https://dx.doi.org/10.1016/j.ajodo.2007.03.023]
69. Yao, H.; Ge, C.; Xue, J.; Zheng, N. A high spatial resolution depth sensing method based on binocular structured light. Sensors; 2017; 17, 805. [DOI: https://dx.doi.org/10.3390/s17040805]
70. Sarbolandi, H.; Lefloch, D.; Kolb, A. Kinect range sensing: Structured-light versus Time-of-Flight Kinect. Comput. Vis. Image Underst.; 2015; 139, pp. 1-20. [DOI: https://dx.doi.org/10.1016/j.cviu.2015.05.006]
71. Jia, T.; Zhou, Z.; Gao, H. Depth measurement based on infrared coded structured light. J. Sens.; 2014; 2014, 852621. [DOI: https://dx.doi.org/10.1155/2014/852621]
72. Alfaro-Santafé, J.; Gómez-Bernal, A.; Lanuza-Cerzócimo, C.; Alfaro-Santafé, J.V.; Pérez-Morcillo, A.; Almenar-Arasanz, A.J. Three-axis measurements with a novel system for 3D plantar foot scanning: iPhone X. Footwear Sci.; 2020; 12, pp. 123-131. [DOI: https://dx.doi.org/10.1080/19424280.2020.1734867]
73. Kovacs, L.; Zimmermann, A.; Brockmann, G.; Baurecht, H.; Schwenzer-Zimmerer, K.; Papadopulos, N.A.; Papadopoulos, M.A.; Sader, R.; Biemer, E.; Zeilhofer, H.-F. Accuracy and precision of the three-dimensional assessment of the facial surface using a 3-D laser scanner. IEEE Trans. Med. Imaging; 2006; 25, pp. 742-754. [DOI: https://dx.doi.org/10.1109/TMI.2006.873624]
74. Koban, K.C.; Cotofana, S.; Frank, K.; Green, J.B.; Etzel, L.; Li, Z.; Giunta, R.E.; Schenck, T.L. Precision in 3-dimensional surface imaging of the face: A handheld scanner comparison performed in a cadaveric model. Aesthetic Surg. J.; 2019; 39, pp. NP36-NP44. [DOI: https://dx.doi.org/10.1093/asj/sjy242] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/30239592]
75. Maués, C.; Casagrande, M.; Almeida, R.; Almeida, M.; Carvalho, F. Three-dimensional surface models of the facial soft tissues acquired with a low-cost scanner. Int. J. Oral Maxillofac. Surg.; 2018; 47, pp. 1219-1225. [DOI: https://dx.doi.org/10.1016/j.ijom.2018.03.028]
76. Timen, C.; Speksnijder, C.M.; Van Der Heijden, F.; Beurskens, C.H.; Ingels, K.J.; Maal, T.J. Depth accuracy of the RealSense F200: Low-cost 4D facial imaging. Sci. Rep.; 2017; 7, 16263.
77. Koban, K.C.; Perko, P.; Etzel, L.; Li, Z.; Schenck, T.L.; Giunta, R.E. Validation of two handheld devices against a non-portable three-dimensional surface scanner and assessment of potential use for intraoperative facial imaging. J. Plast. Reconstr. Aesthetic Surg.; 2020; 73, pp. 141-148. [DOI: https://dx.doi.org/10.1016/j.bjps.2019.07.008]
78. Bakirman, T.; Gumusay, M.U.; Reis, H.C.; Selbesoglu, M.O.; Yosmaoglu, S.; Yaras, M.C.; Seker, D.Z.; Bayram, B. Comparison of low cost 3D structured light scanners for face modeling. Appl. Opt.; 2017; 56, pp. 985-992. [DOI: https://dx.doi.org/10.1364/AO.56.000985]
79. Perillo, L.; Padricelli, G.; Isola, G.; Femiano, F.; Chiodini, P.; Mataresei, G. Class II malocclusion division 1: A new classification method by cephalometric analysis. Eur. J. Paediatr. Dent.; 2012; 13, 192.
80. Kim, P.; Chen, J.; Cho, Y.K. SLAM-driven robotic mapping and registration of 3D point clouds. Autom. Constr.; 2018; 89, pp. 38-48. [DOI: https://dx.doi.org/10.1016/j.autcon.2018.01.009]
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
The aim of the study was to systematically review and compare the accuracy of smartphone scanners versus stereophotogrammetry technology for facial digitization in children. A systematic literature search strategy of articles published from 1 January 2010 to 30 August 2022 was adopted through a combination of Mesh terms and free text words pooled through boolean operators on the following databases: PubMed, Scopus, Web of Science, Cochrane Library, LILACS, and OpenGrey. Twenty-three articles met the inclusion criteria. Stationary stereophotogrammetry devices showed a mean accuracy that ranged from 0.087 to 0.860 mm, portable stereophotogrammetry scanners from 0.150 to 0.849 mm, and smartphones from 0.460 to 1.400 mm. Regarding the risk of bias assessment, fourteen papers showed an overall low risk, three articles had unclear risk and four articles had high risk. Although smartphones showed less performance on deep and irregular surfaces, all the analyzed devices were sufficiently accurate for clinical application. Internal depth-sensing cameras or external infrared structured-light depth-sensing cameras plugged into smartphones/tablets increased the accuracy. These devices are portable and inexpensive but require greater operator experience and patient compliance for the incremented time of acquisition. Stationary stereophotogrammetry is the gold standard for greater accuracy and shorter acquisition time, avoiding motion artifacts.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Details





1 Department of Life, Health & Environmental Sciences, Postgraduate School of Orthodontics, University of L’Aquila, 67100 L’Aquila, Italy
2 Department of Life, Health & Environmental Sciences, Postgraduate School of Orthodontics, University of L’Aquila, 67100 L’Aquila, Italy; Department of General Surgery and Surgical-Medical Specialties, School of Dentistry, University of Catania, 95124 Catania, Italy
3 Department of General Surgery and Surgical-Medical Specialties, School of Dentistry, University of Catania, 95124 Catania, Italy