Skip to main content

SYSTEMATIC REVIEW article

Front. Artif. Intell., 06 December 2022
Sec. Machine Learning and Artificial Intelligence

Applications of contemporary artificial intelligence technology in forensic odontology as primary forensic identifier: A scoping review

  • 1Center for Oral and Maxillofacial Diagnostics and Medicine Studies, Faculty of Dentistry, Universiti Teknologi MARA, Sungai Buloh Campus, Sungai Buloh, Malaysia
  • 2Center for Restorative Dentistry Studies, Universiti Teknologi MARA, Sungai Buloh Campus, Sungai Buloh, Malaysia
  • 3Department of Forensic Odontology, Faculty of Dental Medicine, Universitas Airlangga, Surabaya, Indonesia
  • 4Institute of Pathology, Laboratory and Forensic Medicine (I-PPerForM), Universiti Teknologi MARA, Sungai Buloh Campus, Sungai Buloh, Malaysia

Background: Forensic odontology may require a visual or clinical method during identification. Sometimes it may require forensic experts to refer to the existing technique to identify individuals, for example, by using the atlas to estimate the dental age. However, the existing technology can be a complicated procedure for a large-scale incident requiring a more significant number of forensic identifications, particularly during mass disasters. This has driven many experts to perform automation in their current practice to improve efficiency.

Objective: This article aims to evaluate current artificial intelligence applications and discuss their performance concerning the algorithm architecture used in forensic odontology.

Methods: This study summarizes the findings of 28 research papers published between 2010 and June 2022 using the Arksey and O'Malley framework, updated by the Joanna Briggs Institute Framework for Scoping Reviews methodology, highlighting the research trend of artificial intelligence technology in forensic odontology. In addition, a literature search was conducted on Web of Science (WoS), Scopus, Google Scholar, and PubMed, and the results were evaluated based on their content and significance.

Results: The potential application of artificial intelligence technology in forensic odontology can be categorized into four: (1) human bite marks, (2) sex determination, (3) age estimation, and (4) dental comparison. This powerful tool can solve humanity's problems by giving an adequate number of datasets, the appropriate implementation of algorithm architecture, and the proper assignment of hyperparameters that enable the model to perform the prediction at a very high level of performance.

Conclusion: The reviewed articles demonstrate that machine learning techniques are reliable for studies involving continuous features such as morphometric parameters. However, machine learning models do not strictly require large training datasets to produce promising results. In contrast, deep learning enables the processing of unstructured data, such as medical images, which require large volumes of data. Occasionally, transfer learning was used to overcome the limitation of data. In the meantime, this method's capacity to automatically learn task-specific feature representations has made it a significant success in forensic odontology.

Introduction

Primary identifiers are the most reliable method of confirming identification (Jeddy et al., 2017). Fingerprinting, forensic odontology, and DNA profiling are examples of these identifiers. These methods differ in complexity, but they all have the same level of certainty. Forensic odontology is the simplest and fastest of these methods (Jain et al., 2020). It is a subfield of dentistry that focuses primarily on identifying a person's identity by analyzing the distinctive anatomical structure of the oral cavity (Divakar, 2017; Johnson et al., 2018). The primary applications of this field of study are in medico-legal investigations during a mass disaster, identifying accidental remains through the examination of dental records, and determining an individual's identity based on human remains (Hachem et al., 2020). In this subfield of forensic science, human identification is possible through the deceased body, which usually includes teeth and jawbones. This field is also crucial for identifying human remains after disasters like tsunamis, earthquakes, landslides, bomb blasts, etc., when bodies are so severely damaged and broken up that they can't be identified (Krogman and Isçan, 1986; Hinchliffe, 2011).

The dead bodies are usually identified visually by a close family member or a familiar person who knew that person throughout their life. This is frequently accomplished by visually observing the features of the face, several body options, or personal belongings. However, this technique becomes unreliable if the body options are lost due to post-and perimortem changes, such as decomposition or incineration. In such cases, visual identification may be prone to error. For instance, in cases related to criminal or suspected criminal cases, forensic experts may be needed to conduct the identification process through specific methods to analyze, identify and classify the physical evidence. For instance, cases related to criminal or suspected criminal cases may involve lots of laboratory work. The accuracy of human expertise is unquestionable as they are well trained, which means they are less likely to make a mistake. However, when a significant number of forensic evaluations are needed, it may lengthen the investigation process, eventually causing a burden on the experts and leading to human error. In addition, human identification associated with digital or radiological images may be helpful when clinical dental records are unavailable. The possible images may be acquired from dental x-rays, such as panoramic dental images and digital photographs usually used for analyzing human bite marks. Furthermore, Maber et al. (2006) stated that the radiological observation of the tooth development of permanent teeth among children aged between 4 and 14.9 years gives the most accurate dental age estimation, except for the third molars.

Artificial intelligence (AI) is widely defined as a tool which encompasses any techniques that enable computers to mimic human behavior and excel over human decision-making to solve complex tasks independently or with minimal human intervention (Janiesch et al., 2021). Hence, it is always concerned with a range of central problems, including environmental systems (Krzywanski, 2022), intelligent transportation systems (ITS) (Phillips and Kenley, 2022) and the earth's systems (Sun et al., 2022), and refers to a variety of tools and methods such as artificial neural networks, genetic algorithms, fuzzy logic and expert systems. The emerging computer systems based on intelligent techniques that support complex activities enable the automation system, especially in the medical industry. However, intelligent systems that offer AI often rely on machine learning (ML), which this approach describes the capacity of systems to learn from problem-specific training data to automate the process of analytical model building and solve associated tasks. In contrast, deep learning (DL) is the ML concept based on artificial neural networks (ANN). DL models outperform shallow ML models and traditional data analysis approaches for many applications. A convolutional neural network (CNN) is a prime example of DL, which uses the image as an input to the architecture. This approach has been getting attention from forensic and AI practitioners and is widely used in forensic odontology, especially in identifying individuals and sex dimorphism through radiological examination. However, due to various types of ML architectures applied in the previous study, which varied according to several factors such as its applications, type and amount of dataset used, study setting, and various inclusion/exclusion that varied from another study set by the authors, the best AI technology that can be applied in forensic odontology remain unknown.

A recent comprehensive review on the application and performance of artificial intelligence technology in forensic odontology has been conducted by Khanagar et al. (2021), which involves articles published between January 2000 and June 2020. However, there has been a significant increase in the number of publications on the use of ML and DL methods in forensic odontology within the last 2 years as Google launched TensorFlow 2.0 in June 2019, which declared Keras as the official high-level API of TensorFlow for quick and easy model design and training. The new technology was user-friendly and a highly effective interface for solving machine learning issues which influenced scholars, ML and DL practitioners to iterate on their experiments faster. This seems to be one of the factors behind the increase in publications regarding the application of ML in forensic odontology. This scoping review was conducted to assess the current ML and DL architecture regardless of any computer vision or image processing techniques used in forensic odontology. Thus, the primary research question that guides this review is “What are the current AI technology and its application performance in the field of forensic odontology?”

Methods

A scoping review was conducted using the Arksey and O'Malley (2005) and updated by Joanna Briggs Institute (JBI) Framework for Scoping Reviews to clarify key concepts and identify gaps in the published literature. Scoping reviews map the available data from various sources to provide a broad overview of an ambiguous subject, in contrast to systematic reviews, which concentrate solely on a single question and review objective. In addition, because of the variety of recent ML and DL techniques used in forensic odontology among scholars, the authors decided to conduct a scoping review to identify research gaps of new knowledge and clarify the new concept of the proposed methods, which may also be valuable precursors to systematic reviews and can be used to confirm the relevance of inclusion criteria and potential questions, as stated by Munn et al. (2018).

The framework developed by Arksey and O'Malley has five components: defining the research question, identifying relevant studies, selecting studies, charting the data, and collating, summarizing, and reporting the results. This framework led to the development of the JBI protocol, which allows for systematic review and reporting while also making the process transparent (Peters et al., 2015). Furthermore, this review follows the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA-ScR) guidelines.

Search criteria

The review was structured around a PCC question, an acronym for population, concept, and context. The JBI recommends using this type of question for scoping reviews (Peters et al., 2020). Table 1 shows the search criteria based on the “PCC” mnemonic.

TABLE 1
www.frontiersin.org

Table 1. Description of the PCC elements.

The data for this study was gathered by searching for articles reported in the literature in renowned search engines, primarily PubMed, Google Scholar, Scopus, and Web of Science, published between March 2000 and June 2022. Based on that period, the databases were searched for the terms “artificial intelligence” OR “machine learning” OR “deep learning” OR “deep neural network” OR “convolutional neural network” AND forensic odontology. Table 2 summarizes the search terms used.

TABLE 2
www.frontiersin.org

Table 2. Summary of keyword terms.

Study identification and selection

The relevance and importance of the selected study were evaluated based on its content and publication type. Therefore, only full-text research articles were included in this review. Following identifying articles in the abovementioned databases, they were imported into the EndNote X9 software (Thompson Reuters, Philadelphia, PA, USA), where duplicates were removed. Next, based on the titles and abstracts of the articles, the eligibility criteria were used to perform a preliminary screening. As shown in Figure 1 for the PRISMA-ScR selection process flow diagram, the full text of articles was then accessed to determine which articles were eligible for inclusion in the review. In contrast, editorial notes, reviews, and conference abstracts were excluded from this review.

FIGURE 1
www.frontiersin.org

Figure 1. PRISMA-ScR flowchart of the study collection.

Eligibility criteria for the studies

Inclusion criteria

1. The article must concentrate on forensic odontology.

2. The AI technology employed in the study model should be explicitly stated.

3. There should be a clear statement of a predictive outcome.

4. The data sets utilized for training/validating or evaluating the AI model should be explicitly mentioned.

Exclusion criteria

1. Articles about subjects other than AI technology.

2. Articles that contain abstracts and no full-text articles.

3. Articles are written in languages other than English.

Data extraction

A data extraction form was used to extract the available study details, such as the author(s), year of publication, and AI technology used. In addition, study characteristics such as study factor, image type, feature extractor/preprocessing method, algorithm architecture, evaluation, and findings were extracted. Finally, a narrative synthesis of the results was conducted to address the objectives.

Results

Based on the initial search through selected databases using the keywords, we identified 605 articles, where 268 of them were excluded due to title, abstract, and duplicate removal screening. The remaining 337 articles were assessed for eligibility individually. Only 28 full-text articles fulfilled the inclusion criteria based on the final assessment. The comprehensive review process for the study collection is depicted in Figure 1. Table 3 summarizes the articles in the scoping review.

TABLE 3
www.frontiersin.org

Table 3. Summary of all reviewed articles.

General characteristics of the included studies for the scoping review

The dates of the publications ranged from 2010 to June 2022. Only one study was published in 2010. Another study was conducted after 6 years in 2016, three in 2017, one in 2019, nine in 2020, seven in 2021, and six in 2022. About 22 of the 28 articles were written in the last 2 years, which shows that interest in AI-based technology in forensic odontology applications started to rise in 2020.

As illustrated in Figure 2, forensic odontology can be classified into four significant thrusts: human bite marks, sex determination, age estimation, and dental comparison. With a total of 13 studies (47%), most studies focused on dental age estimation. In contrast, several publications involved dental practitioners (Johnson et al., 2018) and computer programming (Hinchliffe, 2011), and both approaches (Jain et al., 2020) in the image annotation stage or feature extraction. The second highest contribution was 38%, which is sex determination. Sex determination came in second (32%), with half of the publications employing a computer algorithm and half employing human experts to perform feature extraction, while one is not mentioned. Meanwhile, AI-based technology was rarely applied to human bite marks, where only two studies (7%) were included. Both studies employed human experts to annotate images during the feature extraction stage.

FIGURE 2
www.frontiersin.org

Figure 2. Forensic odontology classification.

AI-based method in forensic odontology

Frequent AI-based technologies employed in forensic odontology include deep neural networks, artificial neural networks, machine learning, and computational technology. As illustrated in Figure 3, deep neural networks are the most frequently used in age estimation (Khanagar et al., 2021), sex determination (Johnson et al., 2018), and dental comparison (Johnson et al., 2018), with five of the studies recently published in 2022, following six in 2021, one in 2020 and 2019, and two in 2017. Next, ANN is the second most frequent method, mainly employed in sex determination (Divakar, 2017) and human bite mark analysis (Jeddy et al., 2017). In contrast, ML and other computational technologies were not employed in age estimation and were least used in sex determination (Jain et al., 2020) and human bite mark analysis (Jeddy et al., 2017).

FIGURE 3
www.frontiersin.org

Figure 3. Publication of AI-based technology in forensic odontology.

Discussion

The conceptual distinction of the included studies for the scoping review

In this field, it is necessary to distinguish several relevant terms and concepts from each other. Three critical key terms that need to be well distinguished are machine learning algorithms, artificial neural networks, and deep neural networks. Despite the similarity between these terms, there are differences between them. The Venn diagram in Figure 4 depicts the hierarchical relationship between those terms.

FIGURE 4
www.frontiersin.org

Figure 4. Venn diagram illustrating machine learning concepts and classes.

Initially, AI research focused on hard-coded statements in formal languages, and a computer could automatically think about using rules for logical inference. Hence, according to Goodfellow et al., it is also known as the knowledge base approach (Goodfellow et al., 2016). However, Brynjolfsson and Mcafee (2017) stated that this approach has several limitations as it is difficult for humans to articulate all the implicit knowledge needed to carry out challenging tasks. Fortunately, ML is capable of overcoming these limitations. In terms of a class of tasks and performance measures, ML generally refers to the idea that a computer program performs better over time (Jordan and Mitchell, 2015). Hence, its main objective is to automate the development of analytic models to conduct cognitive tasks such as pattern recognition or object classification. Algorithms that can iteratively learn from training data specific to a given problem can be used to accomplish the goal. So, computers can find complex patterns and hidden information without being programmed (Aggarwal et al., 2022).

Machine learning shows good applicability when involving tasks related to high- dimensional data, such as classification, regression, and clustering. Moreover, it can improve reproducibility by learning from previous computations and extracting regularities from massive databases. Therefore, ML algorithms have been widely applied in many areas, such as image classification, speech and image recognition, or natural language processing. The ML algorithms are classified into three types (Saravanan and Sujatha, 2018): supervised, unsupervised, and reinforcement learning (RL). The overview of all ML types is presented in Table 4, and the comparison between these three types of ML is tabulated in Table 5.

TABLE 4
www.frontiersin.org

Table 4. An overview of the type of machine learning.

TABLE 5
www.frontiersin.org

Table 5. Comparison of all types of machine learning.

The field provides various classes of ML algorithms based on the learning task. Regression models, instance-based algorithms, decision trees, Bayesian methods, and ANNs are all different in their specifications and variants. The main distinction between the two approaches is that in supervised learning, the algorithm “learns” from the training dataset by making data predictions and adjusting for the correct answer iteratively. While supervised learning models are more accurate than unsupervised learning models, they require human intervention to label the data properly (Rudin, 2019). In contrast, unsupervised learning models work independently to discover the inherent structure of unlabeled data. It's essential to keep in mind that they still need human intervention to validate output variables.

Semi-supervised learning provides an advantageous balance between supervised and unsupervised learning (Reddy et al., 2018). During training, a smaller labeled data set is employed to guide classification and feature extraction from a larger unlabeled data set. Semi-supervised learning can solve the issue of insufficient labeled data for training a supervised learning algorithm.

Meanwhile, ANNs have gained wide attention due to their versatile structure, which enables them to be modified for a diverse range of situations among all three types of ML. Many functions, such as clustering, grouping, and regression, are made possible by ANN. For example, we can use ANN to group or sort unlabeled data based on similarities between the samples in the new dataset. In the case of classification, we can train the network on a labeled dataset to classify the objects in the dataset into several categories. The neural network's architecture and mechanisms are focused on the nature of the human brain. For example, humans use their brains to recognize patterns and distinguish various types of information, while NN can be trained to perform the same task on data. It consists of mathematical representations in which the biological neurons present in our brains inspire ANN. Each connection between neurons, like synapses in the brain, transmits signals whose strength can be amplified or attenuated by a weight constantly altered during the learning process (Janiesch et al., 2021).

The basic ANN architecture consists of input, hidden, and output layers. The input layer typically contains the input neurons that send the information/signal to the hidden layer. Then, in the hidden layer, the neurons will only process or fire the signals to the next neuron if those transmitted signals exceed specific threshold values determined by an activation function. This layer is useful for learning a non-linear mapping between input and output and can have zero or more hidden layers (Bishop and Nasrabadi, 2006; Goodfellow et al., 2016). Finally, the output layer generates final results such as image classification or binary input categorization. It should be noted that learning algorithms cannot learn the number of layers and neurons, the learning rate, or the activation function. Instead, they are the model's hyperparameters, which must be manually set or chosen by an optimization procedure.

Unlike simple ANNs, deep neural networks typically include sophisticated neurons of multiple hidden layers arranged in deeply nested network architectures. In this case, advanced operations such as convolutions may be used or have multiple activations in one neuron rather than a simple activation function. These features enable the deep neural network to process the raw input data and automatically learn the representation required for the corresponding learning task. This functionality, however, was not available in shallow ML, such as simple ANNs like shallow autoencoders and other ML algorithms like RF and decision trees. As some algorithms are innately interpretable by humans, shallow ML is identified as a white box. Most advanced ML algorithms, on the other hand, make decisions that can't be seen or understood. This makes them a “black box” (Janiesch et al., 2021).

According to LeCun (2015), deep neural networks outperformed shallow ML algorithms for most applications, including text, image, video, speech, and audio data. This is due to the ability of DL to deal with extensive and high-dimensional data. But in some situations, shallow ML can still do a better job (Zhang and Ling, 2018) than deep neural networks (Rudin, 2019) when there are few data points and low dimensionality. Hence, deciding which networks perform well is subjective and varies according to their applications. Nevertheless, various performance metrics can be used to evaluate the performance of ML algorithms. For example, metrics like log-loss, accuracy, confusion matrix, and AUC-ROC are some of the most common ways to measure how well classification works.

Role of forensic odontology in human identification

Tooth eruption structures can be a valuable source of information for determining the victim's chronological age (Uzuner et al., 2017). The development of dentition is more closely related to chronological age. Human dentition has four distinct developmental periods (Uzuner et al., 2017); the emergence of deciduous teeth in the second year of life, the eruption of two permanent incisors and the first permanent molar between the ages of 6 and 8 years, and the emergence of other remaining permanent dentitions except for the third molars between 10 and 12 years, and the eruption of the third molars around 18 years of age. However, they may remain impacted (Holobinko, 2012). The radiographic evaluation based on dental development and mineralization is considered one of the most reliable methods for determining an individual's age among children and adolescents (Panchbhai, 2011). Through the radiological observation method, chronological age was calculated using the period between the date of birth and the day of the panoramic X-ray study.

Meanwhile, the size and shape of the victim's jawbone can be used to estimate the victim's sex. Sex determination using skull bone analysis is up to 90% accurate (Guyomarc and Bruzek, 2011). In mass disaster cases, when the victim's body is severely damaged to the point where visual identification is no longer possible, the remains of the individual's teeth, jawbones, and skull have proven to be the most valuable source for identifying the individual. The conventional way of estimating sex is through radiographic estimation, whereby the radiographs of the jawbones are considered more practical due to the non-destructive and straightforward method that can be applied to dead and living cases (Patil et al., 2018, 2020).

Examining the diseased person's soft tissues, which primarily include palatal rugoscopy and cheiloscopy (Nagare et al., 2018), is another method of identifying a person. Palatal rugoscopy studies the patterns on the palatal rugae to identify a person. Trobo Hermosa was the first to propose palatal rugoscopy in 1932. Because of its internal position, stability, and perennity, or the fact that it lasts throughout life, it is used in forensics for human identification (Ramakrishnan et al., 2015). On the other hand, the study of lip prints is called cheiloscopy. Lip prints can be identified as early as the sixth week of pregnancy. These prints remain unchanged. Lip prints are thus distinct patterns on the lips that aid in identifying a person (Tsuchihashi, 1974; Ramakrishnan et al., 2015; Nagare et al., 2018). One benefit of this method is that it costs less to examine, which makes it easier to check on both living and dead people (Indira et al., 2012).

Meanwhile, studies involving hard tissue indicate that dentine translucency seems to be one of the most reliable methods of determining an individual's chronological age. The progressive sclerosing of the tubules at the root causes the development of root dentine transparency. This process begins at the root apex and then moves coronally. A previous study reported that dentin translucency increases with age. Therefore, this method is reliable for individuals over 20 years of age when all their permanent teeth have erupted. Also, in forensic odontology, bite mark analysis is the best way to identify a person because injuries caused by teeth and left on things like skin have a unique pattern.

AI-based method in forensic odontology

The application of the AI-based method in forensic odontology has proven to be a breakthrough in providing reliable information in decision-making in forensic sciences. Hence, we demonstrated in these papers that there are four primary areas which successfully employed AI technology at the moment: (Jeddy et al., 2017) human bite marks, (Jain et al., 2020) sex determination, (Divakar, 2017) age estimation, and (Johnson et al., 2018) dental comparison.

Human bite marks

Bite mark analysis is an important aspect and is the most prevalent type of dental evidence presented in criminal court. Matching bite marks to a suspect's dentition involves examining and measuring a person's teeth (Harvey, 1976). The principle of bite mark analysis is that “no two mouths are alike” (Gorea et al., 2014; Gopal and Anusha, 2018; Maji et al., 2018). The central doctrine of bite mark analysis is based on two assumptions: first, that human teeth are unique; second, sufficient detail of the uniqueness is rendered during the biting process to facilitate identification (Pretty and Turnbull, 2001; Lessig et al., 2006). Forensic odontologists can make appropriate decisions on personal identification and bite mark analysis due to the distinctiveness and uniqueness of human dentition. Bite marks can reveal individual tooth marks, a double-arched pattern, or multiple overlying bruises (Maji et al., 2018). In addition, bite marks can become deformed due to the skin's flexibility and elasticity. Bite marks can look different depending on how hard the bite was, where the body was, and how the upper and lower jaws were angled during the bite (Van der Velden et al., 2006; Osman et al., 2017).

Sörup (1926) published the first study on bite marks (Gill and Singh, 2015). Human bite marks are discovered when teeth are employed as weapons of rage, excitement, control, or murder (Pretty and Sweet, 2000). The imprints can also be found on the skin, stationery, musical instruments, cigarettes, and culinary items (Harvey, 1976). It can also be found in criminal cases, including homicides, quarrels, abductions, child abuse, and sexual assaults, as well as during sporting events, and is occasionally purposely caused to incriminate someone falsely (Van der Velden et al., 2006; Kashyap et al., 2015). Bite marks are a form of dental identification in and of themselves. It is now recognized that bite marks provide details comparable to fingerprints.

The usual term used in bite mark analysis is the victim, which indicates the recipient of the bite mark, and the perpetrator is the person who caused the bite marks (Chintala et al., 2018). Unlike bite marks on the body, which are usually caused on purpose, the offenders inadvertently leave bite marks on food at the crime scene. Hence, to identify the offender, dental casts of suspects are prepared and matched using dental material. The proper analysis of bite marks can prove the involvement of a specific person or persons in a meticulous crime (Kashyap et al., 2015). West et al. (1987) believed that bite marks on human skin could be experimentally created to a level that could be compared to bites delivered in aggressive or life-threatening situations. However, more research utilizing living subjects to explore a variety of experimental situations is required. Identifying, recovering, and analyzing bite marks from suspected biters is one of forensic dentistry's most unique, complex, and sometimes difficult challenges (Kashyap et al., 2015; Maji et al., 2018; Rizwal et al., 2021).

In contrast, to bite marks, analyzing the suspect's dentition includes measurement of individual teeth' size, shape, and position (Levine, 1972). Overlays are used in almost every comparison method (American Board of Forensic Odontology, 1986). Hand tracing from dental study casts (Sweet and Bowers, 1998), wax impressions (Luntz and Luntz, 1973), xerographic images (Dailey, 1991), the radio-opaque wax impression method (Naru and Dykes, 1996), and the computer-based method (Sweet et al., 1998; Kaur et al., 2013) are all methods for producing overlays from a suspect's dentition. Sweet and Bowers (1998) investigated the accuracy of these methods for producing bite mark overlays and concluded that computer-generated overlays produced the most accurate and reproducible exemplars. Intercanine distance (ICD) measurements are one part of the study since impressions of the front teeth are usually the most visible and most likely to be measured (Kashyap et al., 2015).

However, the use of ML methods in human bite mark analysis is still in the intermediate phase. Most previous methods, from manual to semi-automatic to fully automatic approaches, focus on computer vision systems that utilize image processing algorithms (Chen and Jain, 2005; Van der Velden et al., 2006; Flora et al., 2009). One of the reasons why this field is not getting a lot of attention from scholars and professionals is the growing doubts about how accurate bite marks can be used as evidence in court. Assumptions about the ability of bitemark comparisons to correctly identify the source of a disputed bite mark have progressed from widespread skepticism to pervasive credulity, with a growing return to skepticism (Saks et al., 2016). This growing skepticism stems from the realization that the field is built on a weak foundation of scientific proof, with a lack of valid evidence to support many of the assumptions and statements made by forensic odontologists during bite mark comparisons (Pretty and Sweet, 2001b; Bush and Bush, 2006; Franco et al., 2015) and that error rates by forensic dentists are possibly the highest of any forensic identification area of expertise still employed (Saks et al., 2016). Besides, the unsatisfactory nature of skin as a substrate for the registration of tooth impressions is one factor that raises doubts about the value and scientific validity of comparing and identifying bite marks (Council, 2009). The bite marks on the skin are easily changed over time and disrupted by skin elasticity, unevenness of the biting surface, swelling, and healing. Hence, these characteristics may strictly restrict the validity of forensic odontology (Janiesch et al., 2021).

However, Mahasantipiya et al. (2012) published a preliminary study on bite mark identification using ANN. The study aims to develop a ML model with high-performance accuracy and to overcome human bias during the analysis. The inclusion criteria include no missing lower and upper anterior teeth or fixed orthodontic appliances. Bite mark samples are then collected using the standard dental wax in five different biting positions. The bite marks of these samples were captured using the digital camera before the preprocessing algorithm. Selected features of the bite marks were chosen to undergo the learning process through the designed ML model. This study shows that the trained networks provided good matching accuracy. Although the accuracy of the proposed ANN was not so high, it shows that this approach has potential and should be investigated further to improve performance. Also, the authors suggested training the ML model on more features of the bite marks that could make it work better.

Although this pattern-matching evidence lacks the scientific foundation to justify continuing admission as trial evidence, most forensic odontologists believe that bite marks can demonstrate sufficient detail for positive identification (Saks et al., 2016). Molina et al. (2022) recently proposed a semi-automated analysis of human bite marks using two different software packages with the new intervention of computer software. DentalPrint generates biting edges from 3D dental cast images, while Biteprint is used to characterize the biting edges. The performance of the identification procedure was evaluated using the ROC curve. The authors reported that the highest area under the ROC curve (AUC) obtained for the Euclidean distance of lower tooth rotation was 0.73. Hence, their proposed method to measure lower tooth rotation may be helpful in identifying individuals responsible for bitemarks and may be relevant in forensic cases. In addition, this study established a new benchmark for future human bite mark analysis studies.

Sex determination

Sex determination is required when information about the deceased is unavailable. In the case of accidents, chemical and nuclear bomb explosions, natural disasters, crime investigations, and ethnic studies, determining a person's sex becomes the priority in the process of identification by a forensic investigator. Usually, forensic experts face a significant challenge when determining sex from skeletal remains, significantly when only body fragments are recovered. Using teeth and skull characteristics, forensic odontologists can help other experts determine the sex of the remains, as male and female teeth have different characteristics, such as morphology, crown size, and root length. In addition, the skull pattern and characteristics of the two sexes differed. Therefore, this will aid forensic odontologists in determining the sex of the remains.

There are several techniques for sex estimation, including odontological and anthropological methods. Both methods include various metric and non-metric variables and biochemical analyses (Capitaneanu et al., 2017). For example, dental methods for studying sexual dimorphism can be based on the morphology and measurements of teeth and other tissue structures, such as cheiloscopy (Karki, 2012; Kinra et al., 2014; Kaul et al., 2015), the palatal rugae (Bharath et al., 2011; Thabitha et al., 2015; Gadicherla et al., 2017), the mandible (Hu et al., 2006; Vinay et al., 2013), and sinuses (Kanthem et al., 2015; Akhlaghi et al., 2016). On the other hand, anthropological methods for figuring out a person's sex use the shape and size of bones like the skull, hip, sacrum, scapula, clavicle, sternum, humerus, and femur to confirm the individual's sex (Durić et al., 2005; Krishan et al., 2016).

Since automation trends in the medical field have been getting wide attention, computer science techniques such as ML, ANN, and DL are promising methods that can automate the conventional method and enhance reproducibility. Several studies have been published using ML techniques for sex determination. Akkoç et al. (2016) proposed a fully automated sex determination from individuals' maxillary tooth plaster model images. The image acquisition process is done before the segmentation and classification step. First, a standard image is obtained by fixing the camera angle on top of the mechanism and equipping it with cube-shaped light sources to absorb light from all directions. Based on the RGB color channel of the standard image, channel B provides significant features compared to others. Image segmentation includes converting color to a binary image, followed by morphological operations, which include binary dilation and erosion. Finally, the segmented plaster image was transformed into a gray-level image for feature extraction using a gray-level co-occurrence matrix (GLCM) method. Extracted features are then classified using the RF algorithm. Results show that the RF algorithm gives the highest classification accuracy compared to other methods such as SVM, ANN, and kNN. The authors made improvements by using the local discrete cosine transform (DCT) in extracting features and the RF algorithm for classifying images (Akkoç et al., 2017). Based on the 10-fold cross-validation, the average classification accuracy was 85.166%, and the area under the ROC curve was 91.75%

Meanwhile, a study using mandibular morphometric parameters that used digital panoramic radiographs has been proposed by Patil et al. (2020). Seven morphometric parameters were selected based on the previous studies (Raj and Ramesh, 2013; Kumar et al., 2016; More et al., 2017) for evaluation. Figure 5 shows the measurement of morphometric parameters done on digital panoramic radiographs. A feed-forward neural network with a backpropagation learning algorithm was proposed in this study. The NN model consists of an input layer, two hidden layers, and one output layer, where 70% of the dataset was assigned for training, 15% for validation, and 15% for testing. Based on the three analyses done on morphometric parameters, ANN analysis had a higher overall accuracy of 75% than Logistic Regression and Discriminant Analysis, both of which had an overall accuracy of 69.9%.

FIGURE 5
www.frontiersin.org

Figure 5. Morphometric parameter measurements done on digital panoramic radiographs by Raj and Ramesh (2013), Kumar et al. (2016), and More et al. (2017).

Similar research has been published by Ortiz et al. (2020), but different morphometric measurements are done on panoramic dental radiographs. Figure 6 shows the measurement of parameters on panoramic radiographs where each number marked on the image is defined as follows:

• AMD (D)

• AMD (E)

• C-Co (D)

• C-Co (E)

• C-C

• Go-Go

• C-Go (D)

• C-Go (E)

• FM – FM: FM - PSM (D)

• FM – PSM (E)

• FM – FM x PSM

• FM – BMD (D)

• FM – BMD (E

• X Me

FIGURE 6
www.frontiersin.org

Figure 6. Measurement of morphometric parameters marked on panoramic radiographs (Ortiz et al., 2020). R, Right; L, Left; AMD, right mandibular angle; C, Condyle; Co, Coronary process; Go, Gonys; FM, Mental Foramen; PSM, Medium Saginal Plane; Me, Mento; BMD, Mandible Base; X Me, Me Intersection Point; PSM, Me.

The authors performed the feature classification in this research using five different ML models: KNN, NN, stochastic gradient descent (SGD), Naïve Bayes, and logistic regression. Based on the predictive analysis, the KNN model exhibits the highest training accuracy of 93.7%, while NN shows the highest testing accuracy of 89.1%. Another promising ML model for sex classification was proposed by Esmaeilyfard et al. (2021) using the first molar teeth in Cone Beam Computed Tomography (CBCT) images. Feature extraction is performed before the classification. Nine parameters were measured in the centre of the corrected sagittal and coronal sections. These parameters are the roof, floor, and height of the pulp chamber, as well as marginal enamel thickness and dentin thickness at the height of contour (HOC), tooth width, and crown length in both buccolingual and mesiodistal aspects. This study experimented with three different classifiers: Naïve Bayesian, RF, and SVM. Based on the 10-fold cross-validation, Naïve Bayesian gives the best result, with an average accuracy of 92.31%.

Meanwhile, a recent publication employing DL methods for sex classification has been reported by Liang et al. (2021), where a CNN algorithm was proposed in which two pre-trained models are adopted: ResNet-34 (He et al., 2016) and Inception-ResNet (Szegedy et al., 2017). Based on the proposed method, the accuracy of Inception-ResNet is superior to the other pre-trained models, in which 59.62 and 50.57% were achieved in terms of mAP and rank-1 accuracy, respectively.

Another DL approach applied to the panoramic radiographs has been proposed by Milošević et al. (2021), in which several pre-trained models were tested: DenseNet201 (Huang et al., 2017), InceptionResNetV2 (Szegedy et al., 2017), ResNet50 (He et al., 2016), VGG16, VGG19 (Simonyan and Zisserman, 2015), and Xception (Chollet, 2017). Hyperparameter tuning was done to determine the best DL model for the study. VGG16 shows the most successful model based on tuned hyperparameters, as it can achieve a classification accuracy of up to 77%.

Early this year, in January 2022, Nithya and Sornam (2022) reported a detailed study with clear explanations of deep convolutional neural networks (DCNN) using dental x-ray images. The authors created their own CNN architecture, which comprises five layers of sequential networks, where the final layer is fully connected. In this paper, the hyperparameter values are listed. For example, batch size 50, Adam optimizer, and categorical cross-entropy loss are assigned to the algorithm. As a result, 95% accuracy was obtained. The authors stated that their proposed method was superior to the existing one, which utilized transfer learning through a pre-trained model of VGG16 (Ilić et al., 2019).

Although numerous tools for sexual dimorphism based on morphological dental traits in identifying sex are available, challenges still restrict their performance. Franco et al. (2022) presented a preliminary study on the applicability of an ML setup to distinguish males and females using dentomaxillofacial features from a panoramic radiograph dataset. Employing two different CNN architectures, one was the network built from scratch, and the other was transfer learning associated with DenseNet121, the authors reported that the classification accuracy of the transfer learning architecture was superior to the from-scratch model, which is 82% and 71%, respectively. The authors suggest that as this study aims to understand the discriminant power of dental morphology to distinguish between males and females, the current findings should not be applied in practice. However, the authors listed the hyperparameters, allowing other scholars to improve the ML architecture and prediction performance.

Age estimation

Extensive longitudinal studies in this field have influenced scholars and professionals to go beyond the conventional approach. A more sophisticated age estimation process can be done using computer science techniques with less human intervention. The automation process of age assessment proved to have exact reproducibility and be equally prominent, just like the conventional method. The most common method of dental age estimation based on tooth development is based on Demirjian et al.'s (1973) staging system, which uses digital panoramic dental imaging to estimate an individual's chronological age based on the mineralization of seven permanent left lower teeth. This method is also suitable for determining dental maturity states, whether the individual with a known age is advanced or delayed, rather than predicting an unknown age (Ismail et al., 2018).

De Tobel et al. (2017) proposed an automated technique for age estimation based on the mandibular third molar development using panoramic radiographs by employing pattern recognition and classification approaches to the target images. First, image contrast is normalized for all data and ROIs, which indicate the third molar, were cropped using the Photoshop software. Then, a pre-trained model of the AlexNet network was adopted, and the performance was evaluated in a 5-fold cross-validation scenario, using different validation metrics to obtain the accuracy, Rank-N recognition rate, mean absolute difference, and linear kappa coefficient. As a result, the proposed method can stage lower third molar development according to staging by human observers. But more training is needed with data because the pilot study was only done on 20 images.

Recently, the DCNN has also been introduced to perform automated tooth segmentation, in which the segmentation is done automatically. This type of ML has shown better performance compared to other mathematical approaches. Matsuda et al. (2020) improved as proposed in De Tobel et al. (2017). Instead of using the raw images as input data to the CNN architecture, the images were imported to Adobe Photoshop CC 2018 and segmented using the built-in tools before the classification process, using the DenseNet201 network for automatic stage allocation. The authors hypothesized that segmenting only the third molar could improve automated stage allocation performance based on the improvement. Another update on this research was proposed by Banar et al. (2020), which aims to develop a fully automated system to stage the third molar development. By providing the ground truth images, which are manually segmented as described in Merdietio Boedi et al. (2020), three main steps are proposed; third molar localization, segmentation, and classification. Image localization involves the prediction of the geometrical centre within the ground truth image cell using a YOLO-like (Redmon et al., 2016) CNN architecture, and the ImageNet (Russakovsky et al., 2015) pre-trained modal was employed to detect the rectangular ROI, which is the location of the third molar itself presented on the original input image. Then, another CNN was employed to segment the extracted ROI. A final CNN combines the third molar's ROI and segmentation to classify the third molar's developmental stage using two different CNN architectures: a simple CNN with ten layers and the more complex DenseNet201 (Huang et al., 2017), as proposed by Merdietio Boedi et al. (2020). Also, the authors said that future research should include the steps for estimating age instead of focusing on the proposed three-step procedure since this step was not included in the proposed framework.

Another study which utilized the third molar to perform the automatic developmental stage assessment has been proposed by Upalananda et al. (2021). In this study, third molar images in every developmental stage were segmented manually. The pre-trained GoogLeNet (Szegedy et al., 2015) architecture was employed. The stochastic gradient descent (SGD) function was used for training, with the hyperparameters set to default: learning rate of 0.001, training epochs of 10, and mini-batch size of 32. The authors reported that there was inconsistency in accuracy across developmental stages. For example, early developmental stages (Stages D and E) had higher accuracy, whereas later developmental stages had less accuracy (Stages F to H). This is because of the morphological variety of dentition at each stage of development, which grows in complexity as it approaches completion.

Another recent approach that utilizes the DCNN has been proposed by Lee et al. (2020) and Kahaki et al. (2020). However, concerning dental age estimation, the literature states that the variation of the third molar might affect the accuracy of age estimation in different populations (Tafrount et al., 2019). As such, regarding the automated approach, the classification accuracy may also be affected due to the morphology of the tooth and its surroundings. In addition, the unwanted ROI, such as periodontal ligament, bony structures, and mandibular nerve canal, influences the performance of automated stage allocation, as stated in Merdietio Boedi et al. (2020).

Variations in dental morphology may affect the performance of the automated system. For example, previous research reported that monoradicular teeth are more resistant to destruction. Besides, in terms of morphological appearance, they had excellent morphology associated with large pulpal areas present in the digital panoramic dental radiographs compared to incisors (Olze et al., 2012). Meanwhile, the superimposition of the normal dentition will superimpose the vertebrae in the centre of the panoramic image. Thus, they are prone to poor intensity, especially in the middle area of the image, due to the ghost image appearance of the spine and the bite-blocker effect of the x-ray procedure. Besides, molar teeth have a more complicated morphology, consisting of two or three roots, each containing one or two root canals. Moreover, as the root formation reaches approximately one-third of its development, the formation of inter-radicular bifurcation begins where this structure looks like a small clip has appeared in the lower-middle area of the tooth. Typically, these clips are not attached to the main object, which is the tooth structure itself. From a computer vision point of view, the image may have two connected parts: the main structure of the tooth and the clips. This makes it hard for the ML algorithm to pull out the features.

Despite the complicated morphological structure, the first molar is considered to be the most reliable tooth for estimating dental age (Shah and Venkatesh, 2016), Kim et al. (2021) developed a CNN model to determine an individual's age group by extracting the image patches consisting of all four first molar from the panoramic radiographs. First, the CNN architecture of ResNet152 (He et al., 2016) was employed, where the network weights were initialized using pre-trained weights from the ImageNet dataset. Then, according to the age and location of the tooth, the learned features of CNNs were visualized as a heatmap, which demonstrated that CNNs focus on various anatomical parameters, such as tooth pulp, alveolar bone level, or interdental space.

Meanwhile, Mohammad et al. (2021) proposed the deep neural networks associated with the pre-trained model called AlexNet to classify the first and second mandibular premolar teeth. Based on the original training dataset, significant ME was found in stage D of dental development over 1.0 years. Therefore, the authors proposed five new sub-stages to reduce the discrepancy: D1, D2, D3, D4, and D5. Then, the ratio of true classification to total observations is obtained using the same pre-trained model. AlexNet results in 92.5% of classification accuracy. An advanced DL approach has been proposed by the same authors, in which Chollet (2015) and TensorFlow Developers (2022) were used to classify dental developmental stages (Mohammad et al., 2022). In this research, a DL model was built from scratch. The robust model achieved an accuracy of 97.74, 96.63, and 78.13% on the training, validation, and testing sets, respectively. A customized CNN model indeed increased the performance of human identification (Mohammad et al., 2020). An automatic human identification system (DENT-net) (Fan et al., 2020) and a Learnable Connected Attention Network (LCANet) (Lai et al., 2020) are other examples of the recent customized model. These two models employed the same loss function, cosine loss, in which their approach achieves a more competitive result than the other losses function across their dataset.

A recent publication that utilized the largest panoramic dental X-ray image dataset in forensic odontology literature was trained on deep neural networks proposed by Milošević et al. (2022). This study aims to verify the deep neural network in solving age estimation problems. The authors successfully verified the literature for estimating the age of adult and senior subjects by using one of the most extensive datasets stated in the literature. The proposed CNN architecture consists of four parts. The first part is applying a pre-trained CNN for feature extraction where pre-trained models of DenseNet201, VGG16, InceptionResNetV2, ResNet50, VGG16, VGG19, and Xception were tested. The second part of the structure was a 1 × 1 convolutional layer used to adjust the number of channels in the final feature map. In contrast, the third part involved the optional attention mechanism, and the last part consisted of two fully connected layers. Finally, hyperparameter tuning was done to obtain the best performance of a model on a dataset. Based on the hyperparameter optimization, VGG16, with 40 channels in the final convolutional layer and 128 units in the second to last fully connected layers, with no attention mechanism and batch normalization, exhibits a high-performing model. In many articles, scholars rarely mention hyperparameter optimization to generate the optimal CNN model. However, as this step is essential in regulating ML behavior, this new publication can be used as a reference. Before this publication, Merdietio Boedi et al. (2020) reported that VGG16 could achieve the highest accuracy among those six CNN architectures, even with a small dataset, by employing two transfer learning methods: pretraining and fine-tuning. The ImageNet dataset, a large-scale image recognition dataset including over 14 million labeled images, has been utilized in both transfer learning methods.

Dental comparison

The establishment of an individual's identity is an essential aspect of forensic identification. In a large-scale disaster, forensic teams are challenged to perform effective and efficient identification by analyzing available human identifiers (Kurniawan et al., 2020). Human dentition is one of the recommended primary identifiers by Interpol. The scientific basis for dental identification is the comparison of antemortem and postmortem data based on the unique characteristics of human dentition. A previous study explained that the number and complexity of dental restorations increased with age (Andersen et al., 1995). In forensic dental identification, an individual who had a number of complicated dental treatments was easier to identify than someone who had little or no treatment (Pretty and Sweet, 2001a). Dental treatment patterns are regarded as a distinct and powerful feature that represents an individual's identity.

The application of AI in forensic dental identification can help to achieve a more efficient and effective identification process (Putra et al., 2022). The DL methods such as CNN and R-CNN have been developed and used for automatic tooth detection on dental radiographs to support individual identification (Miki et al., 2017; Chen et al., 2019; Mahdi et al., 2020). Choi et al. (2022) conducted a study of the automatic detection of teeth and dental treatment patterns on OPG using deep neural networks. The detection of natural teeth and dental treatment was done with a pre-trained object detection network, which was a CNN modified by EfficientDet-D3. The study reported the outstanding performance of CNN in automatic detection of natural teeth (99.1% precision), prostheses (80.6%), treated root canals (81.2%), and implants (96.8%).

A study by Heinrich et al. (2018) proposed an automatic comparison between antemortem and postmortem panoramic radiographs using computer vision. According to the findings of this study, the proposed technique could be a reliable method for comparing antemortem and postmortem OPG with an average accuracy of 85%. The systematic matching yielded a maximum of 259 corresponding points for successful identification between two different OPGs of the same person and a maximum of 12 points for other non-identical people. The challenge of the study by Heinrich et al. is associated with the identification of a person with only a few teeth or no special characteristics, such as dental fillings, dental implants, and prostheses. The inability to identify can be attributed to a lack of quality of OPG, as dental characteristics could not be extracted sufficiently from an overexposed radiograph. This study suggests that computer vision enables automated identification with short computation and reliable results.

Strength of the study

To our knowledge, this is the most recent comprehensive scoping review of AI technology in forensic odontology. Existing review articles have discussed AI technology published in January 2000 up till June 2020, which focused on the three potential applications of AI-based methods in forensic odontology, such as human bite marks, sex estimation and age estimation, except for the application of AI in dental comparison where this may result in omission of potential research articles that related to this field of study. In addition, this scoping review has screened publicly accessible resources worldwide based on the designated inclusion and exclusion criteria. Although this study was conducted as a scoping review, it followed a structured methodology that includes implementing AI technology in forensic odontology-related study factors, algorithm intervention, performance, and research. Therefore, this article may be the most recent study contributing to AI technology in forensic odontology.

Knowledge gaps

This article demonstrates shortcomings and a significant knowledge gap in prior research on AI technology in forensic odontology. Our review confirms the research community's interest in human identification assisted by ML architecture. However, the following constraints and challenges must be addressed for future development. First, studies that focused on transfer learning or using pre-trained models overestimate the estimation performance compared to its performance in real-world problems involving intra- and inter-observer agreement. In terms of the application of AI in age estimation, most studies applied staging techniques that considered morphological dental development, enabling human observers to compare the automated system and manual staging. However, comparison with human observers is unable when it comes to continuous data, which involves morphometric parameters such as volumes or the length of the anatomical structures.

In addition, studies involving preprocessing using several computer vision algorithms for segmentation exhibit multiple performance metrics that restrict the comparison of the reported study. For example, model accuracy was not the right metric to evaluate the performance of the segmentation algorithm. The appropriate metric to evaluate the method involving object segmentation was the intersection over union (IoU). Unfortunately, many studies have overlooked this metric that proposed object segmentation before classification. With the improperly reported details on the performance metric, the classification of the segmented ROI may remain doubtful. To address this issue, researchers are advised to report their research with diverse performance metrics that is significant to the proposed methodology.

This review shows a significant increase in publications over the previous 2 years, indicating a significant rise in knowledge, recognition, trends, and interest in using AI technology in forensic odontology for human identification. However, much of the existing literature is focused only on age and sex estimation, which uses dental radiographs to identify individuals. Hence, the potential application of AI technology in this field of study was restricted. Additionally, this study demonstrates that most of the architecture algorithm employed was based on the transfer learning approach, which uses the pre-trained model as the starting point for a model on a new task. As we know, transfer learning is an option for the small sample of the dataset.

Meanwhile, there is not much study developing a new ML model from scratch, which somehow may be an alternative to research which has limited Additionally, this study demonstrates that most of the architecture algorithm employed was based on the transfer learning approach, which uses the pre-trained model as the starting point for a model on a new task. As we know, transfer learning is an option for the small sample of the dataset. Meanwhile, few studies are developing a new ML model from scratch, which may be an alternative to research with a limited dataset.

Conclusion and future recommendations

Machine learning has proved its capability to predict as humans do. Therefore, ML's feasibility in the field of forensic odontology is undeniable. Furthermore, the reviewed articles show that ML techniques are reliable for studies that involve continuous features, such as morphometric parameters, that require fewer training datasets to be trained on the ML model with promising outcomes. Meanwhile, DL networks learn by observing complex patterns in the data they experience. Hence, large datasets may be required to be trained on the network as they need to learn as many features as possible to make a good prediction. In the meantime, this approach has been a significant success in the ML field due to its ability to learn task-specific feature representations automatically. Hence, it is one of the most frequently utilized types of ML in many applications nowadays.

Based on the presented studies, it is evident that AI technology has been successfully implemented in various aspects of forensic odontology. However, it is essential to highlight that all the studies published to date and reviewed in this present study were based on secondary data, which could not provide the actual performance of AI in real-world problems. Hence, in the future, the real dataset acquired from the criminal cases found at the scene or from the actual real-life incident, such as a mass disaster, was recommended to be tested on the proposed ML architecture, and a comparison should be made between these two approaches. In addition, instead of reporting the performance of the proposed method using simple per cent agreement calculation, Cohen's kappa coefficient is recommended as it is more robust that applies the statistic to measure inter-rater reliability. This can be done by introducing a cross-tabulation based on the result of the testing dataset, which includes two different raters: the computer prediction result and the other from a human observer.

Data availability statement

The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.

Author contributions

NM designed the study and drafted the manuscript, which was revised and completed by RA, AK, and MM. All authors have approved the final version of the manuscript for submission.

Funding

This work was supported by the GPK Research Grant, Universiti Teknologi MARA [600-RMC/GPK 5/3 (188/2020)].

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Aggarwal, K., Mijwil, M. M., Al-Mistarehi, A.-H., Alomari, S., Gök, M., Alaabdin, A. M. Z., et al. (2022). Has the future started? the current growth of artificial intelligence, machine learning, and deep learning. IRAQI J. Comp. Sci. Mathe. 3, 115–23. doi: 10.52866/ijcsm.2022.01.01.013

PubMed Abstract | CrossRef Full Text | Google Scholar

Akhlaghi, M., Bakhtavar, K., Moarefdoost, J., Kamali, A., and Rafeifar, S. (2016). Frontal sinus parameters in computed tomography and sex determination. Leg. Med. 19, 22–7. doi: 10.1016/j.legalmed.2016.01.008

PubMed Abstract | CrossRef Full Text | Google Scholar

Akkoç, B., Arslan, A., and Kök, H. (2016). Gray level co-occurrence and random forest algorithm-based gender determination with maxillary tooth plaster images. Comput. Biol. Med. 73, 102–7. doi: 10.1016/j.compbiomed.2016.04.003

PubMed Abstract | CrossRef Full Text | Google Scholar

Akkoç, B., Arslan, A., and Kök, H. (2017). Automatic gender determination from 3D digital maxillary tooth plaster models based on the random forest algorithm and discrete cosine transform. Comput. Method. Prog. Biomed. 143, 59–65. doi: 10.1016/j.cmpb.2017.03.001

PubMed Abstract | CrossRef Full Text | Google Scholar

American Board of Forensic Odontology (1986). Guidelines for bite mark analysis. J. Am. Dental Assoc. 112, 383–386. doi: 10.1016/S0002-817723021-4

CrossRef Full Text | Google Scholar

Andersen, L., Juhl, M., Solheim, T., and Borrman, H. (1995). Odontological identification of fire victims—potentialities and limitations. Int. J. Leg. Med. 107, 229–34. doi: 10.1007/BF01245479

PubMed Abstract | CrossRef Full Text | Google Scholar

Arksey, H., and O'Malley, L. (2005). Scoping studies: towards a methodological framework. Int. J. Soc. Res. Methodol. 8, 19–32. doi: 10.1080/1364557032000119616

CrossRef Full Text | Google Scholar

Banar, N., Bertels, J., Laurent, F., Boedi, R. M., De Tobel, J., Thevissen, P., et al. (2020). Towards fully automated third molar development staging in panoramic radiographs. Int. J. Leg. Med. 134, 1831–41. doi: 10.1007/s00414-020-02283-3

PubMed Abstract | CrossRef Full Text | Google Scholar

Bharath, S. T., Kumar, G. R., Dhanapal, R., and Saraswathi, T. (2011). Sex determination by discriminant function analysis of palatal rugae from a population of coastal Andhra. J. For. Dent. Sci. 3, 58. doi: 10.4103/0975-1475.92144

PubMed Abstract | CrossRef Full Text | Google Scholar

Bishop, C. M., and Nasrabadi, N. M. (2006). Pattern Recognition and Machine Learning. Berlin: Springer.

Google Scholar

Brynjolfsson, E., and Mcafee, A. (2017). Artificial intelligence, for real. Harvard Bus. Rev. 1, 1–31. Available online at: https://hbr.org/

Google Scholar

Bush, M. A., and Bush, P. J. (2006). Current Context of Bitemark Analysis and research. Bitemark Evidence. Boca Raton, FL: CRC Press.

Google Scholar

Capitaneanu, C., Willems, G., and Thevissen, P. A. (2017). systematic review of odontological sex estimation methods. J. For. Odonto. Stomatol. 35, 1. Available online at: http://www.iofos.eu/?page_id=255

PubMed Abstract | Google Scholar

Chen, H., and Jain, A. K. (2005). Dental biometrics: alignment and matching of dental radiographs. IEEE Trans. Patt. Anal. Mach. Int. 27, 1319–26. doi: 10.1109/TPAMI.2005.157

PubMed Abstract | CrossRef Full Text | Google Scholar

Chen, H., Zhang, K., Lyu, P., Li, H., Zhang, L., Wu, J., et al. (2019). A deep learning approach to automatic teeth detection and numbering based on object detection in dental periapical films. Sci. Rep. 9, 1–11. doi: 10.1038/s41598-019-40414-y

PubMed Abstract | CrossRef Full Text | Google Scholar

Chintala, L., Manjula, M., Goyal, S., Chaitanya, V., Hussain, M. K. A., Chaitanya, Y., et al. (2018). Human bite marks–a computer-based analysis using adobe photoshop. J. Ind. Acad. Oral Med. Radiol. 30, 58. doi: 10.4103/jiaomr.jiaomr_87_17

CrossRef Full Text | Google Scholar

Choi, H.-R., Siadari, T. S., Kim, J.-E., Huh, K.-H., Yi, W.-J., Lee, S.-S., et al. (2022). Automatic detection of teeth and dental treatment patterns on dental panoramic radiographs using deep neural networks. For. Sci. Res. 2022, 1–11. doi: 10.1080/20961790.2022.2034714

PubMed Abstract | CrossRef Full Text | Google Scholar

Chollet, F. (2015) Keras. Available online at: https://github.com/fchollet/keras.

Google Scholar

Chollet, F. (2017). “Deep learning with depthwise separable convolutions,” in Proceedings of the IEEE conference on computer vision and pattern recognition. Piscataway, NJ: IEEE.

Google Scholar

Council, N. R. (2009). Strengthening For. Science in the United States: A Path Forward. Washington, DC: National Academies Press.

Google Scholar

Dailey, J. A. (1991). Practical technique for the fabrication of transparent bite mark overlays. J. Fore. Sci. 36, 565–70. doi: 10.1520/JFS13059J

PubMed Abstract | CrossRef Full Text | Google Scholar

De Tobel, J., Radesh, P., Vandermeulen, D., and Thevissen, P. W. (2017). An automated technique to stage lower third molar development on panoramic radiographs for age estimation: a pilot study. J. For. Odonto. Stomatol. 35, 42. Available online at: http://www.iofos.eu/?page_id=255

PubMed Abstract | Google Scholar

Demirjian, A., Goldstein, H., and Tanner, J. M. (1973). A new system of dental age assessment. Hum. Biol. 1973, 211–27.

PubMed Abstract | Google Scholar

Divakar, K. (2017). Forensic odontology: the new dimension in dental analysis. Int. J. Biomed. Sci. 13, 1. Available online at: http://www.ijbs.org/

PubMed Abstract | Google Scholar

Durić, M., Rakočević, Z., and ÐDonić, D. (2005). The reliability of sex determination of skeletons from forensic context in the Balkans. For. Sci. Int. 147, 159–64. doi: 10.1016/j.forsciint.2004.09.111

PubMed Abstract | CrossRef Full Text | Google Scholar

Esmaeilyfard, R., Paknahad, M., and Dokohaki, S. (2021). Sex classification of first molar teeth in cone beam computed tomography images using data mining. For. Sci. Int. 318, 110633. doi: 10.1016/j.forsciint.2020.110633

PubMed Abstract | CrossRef Full Text | Google Scholar

Fan, F., Ke, W., Wu, W., Tian, X., Lyu, T., Liu, Y., et al. (2020). Automatic human identification from panoramic dental radiographs using the convolutional neural network. For. Sci. Int.. 314, 110416. doi: 10.1016/j.forsciint.2020.110416

PubMed Abstract | CrossRef Full Text | Google Scholar

Flora, G., Tuceryan, M., and Blitzer, H. (2009). “Forensic bite mark identification using image processing methods,” in Proceedings of the 2009 ACM Symposium on Applied Computing (Honolulu, HI: ACM).

Google Scholar

Franco, A., Porto, L., Heng, D., Murray, J., Lygate, A., Franco, R., et al. (2022). Radio-diagnostic application of convolutional neural networks for dental sexual dimorphism. Res. Squ. [Preprint]. doi: 10.21203/rs.3.rs-1821209/v1

PubMed Abstract | CrossRef Full Text | Google Scholar

Franco, A., Willems, G., Souza, P. H. C., Bekkering, G. E., and Thevissen, P. (2015). The uniqueness of the human dentition as forensic evidence: a systematic review on the technological methodology. Int. J. Leg. Med. 129, 1277–83. doi: 10.1007/s00414-014-1109-7

PubMed Abstract | CrossRef Full Text | Google Scholar

Gadicherla, P., Saini, D., and Bhaskar, M. (2017). Palatal rugae pattern: An aid for sex identification. J. For. Dent. Sci. 9, 48. doi: 10.4103/jfo.jfds_108_15

PubMed Abstract | CrossRef Full Text | Google Scholar

Gill, G., and Singh, R. (2015). Reality bites-demystifying crime. J. Foren. Res. 2015, 1. doi: 10.4172/2157-7145.1000S4-005

CrossRef Full Text | Google Scholar

Goodfellow, I., Bengio, Y., and Courville, A. (2016). Deep Learning. Cambridge, MA: MIT Press.

Google Scholar

Gopal, K. S., and Anusha, A. V. (2018). Evaluation of accuracy of human bite marks on skin and an inanimate object: a forensic-based cross-sectional study. Int. J. Forens. Odontol. 3, 2. doi: 10.4103/ijfo.ijfo_20_17

CrossRef Full Text | Google Scholar

Gorea, R., Jasuja, O., Abuderman, A. A., and Gorea, A. (2014). Bite marks on skin and clay: a comparative analysis. Egypt. J. Foren. Sci.. 4, 124–8. doi: 10.1016/j.ejfs.2014.09.002

CrossRef Full Text | Google Scholar

Guyomarc, P., and Bruzek, J. (2011). Accuracy and reliability in sex determination from skulls: a comparison of Fordisc® 3.0 and the discriminant function analysis. Forens. Sci. Int. 208, 180. doi: 10.1016/j.forsciint.2011.03.011

PubMed Abstract | CrossRef Full Text | Google Scholar

Hachem, M., Mohamed, A., Othayammadath, A., Gaikwad, J., and Hassanline, T. (2020). Emerging applications of dentistry in medico-legal practice-forensic odontology. Int. J. Emerg. Technol. 11, 66–70. Available online at: https://www.researchtrend.net/ijet/ijet.php

Google Scholar

Harvey, W. (1976). Bites and bite marks: dental identification and forensic odontology. Henry Kempt. Pub. 1, 90.

Google Scholar

Harvey, W. (1976). Dental Identification and For. Odontology. London: Kimpton.

Google Scholar

He, K., Zhang, X., Ren, S., and Sun, J. (2016). “Deep residual learning for image recognition, in Proceedings of the IEEE conference on computer vision and pattern recognition. Piscataway, NJ: IEEE.

Google Scholar

Heinrich, A., Güttler, F., Wendt, S., Schenkl, S., Hubig, M., Wagner, R, et al. (2018). Forensic odontology: automatic identification of persons comparing antemortem and postmortem panoramic radiographs using computer vision. RöFo-Fortschritte auf dem Gebiet der Röntgenstrahlen und der bildgebenden Verfahren. New York, NY: Georg Thieme Verlag KG.

Google Scholar

Hinchliffe, J. (2011). Forensic odontology, part 2. Major disasters. Br. Dental J. 210, 269–74. doi: 10.1038/sj.bdj.2011.199

PubMed Abstract | CrossRef Full Text | Google Scholar

Holobinko, A. (2012). Forensic human identification in the United States and Canada: a review of the law, admissible techniques, and the legal Implications of their application in forensic cases. Forens. Sci. Int. 222, 394doi: 10.1016/j.forsciint.2012.06.001

PubMed Abstract | CrossRef Full Text | Google Scholar

Hu, K. S., Koh, K. S., Han, S. H., Shin, K. J., and Kim, H. J. (2006). Sex determination using nonmetric characteristics of the mandible in Koreans. J. For. Sci. 51, 1376–82. doi: 10.1111/j.1556-4029.2006.00270.x

PubMed Abstract | CrossRef Full Text | Google Scholar

Huang, G., Liu, Z., Van Der Maaten, L., and Weinberger, K. Q. (2017). “Densely connected convolutional networks,” in Proceedings of the IEEE conference on computer vision and pattern recognition. Piscataway, NJ: IEEE.

Google Scholar

Ilić, I., Vodanović, M., and Subašić, M. (2019). “Gender estimation from panoramic dental X-ray images using deep convolutional networks,” in IEEE EUROCON 2019-18th International Conference on Smart Technologies. Piscataway, NJ: IEEE.

Google Scholar

Indira, A., Gupta, M., and David, M. P. (2012). Usefullness of palatal rugae patterns in establishing identity: preliminary results from Bengaluru city, India. J. For. Dent. Sci. 4, 2. doi: 10.4103/0975-1475.99149

PubMed Abstract | CrossRef Full Text | Google Scholar

Ismail, A. F., Othman, A., Mustafa, N. S., Kashmoola, M. A., Mustafa, B. E., Mohd Yusof, M. Y. P., et al. (2018). Accuracy of different dental age assessment methods to determine chronological age among malay children. J Phys Conf Ser. 1028, 012102. doi: 10.1088/1742-6596/1028/1/012102

CrossRef Full Text | Google Scholar

Jain, S., Singh, K., Gupta, M. R., Bagri, G., Vashistha, D. K., and Soangra, R. (2020). Role of forensic odontology in human identification: A review. Int. J. Appl. Dent. Sci. 6, 109–111.

Google Scholar

Janiesch, C., Zschech, P., and Heinrich, K. (2021). Machine learning and deep learning. Electr. Markets. 31, 685–95. doi: 10.1007/s12525-021-00475-2

CrossRef Full Text | Google Scholar

Jeddy, N., Ravi, S., and Radhika, T. (2017). Current trends in forensic odontology. J. Foren. Dent. Sci. 9, 115. doi: 10.4103/jfo.jfds_85_16

PubMed Abstract | CrossRef Full Text | Google Scholar

Johnson, A., Gandhi, B., and Joseph, S. E. A. (2018). Morphological study of tongue and its role in forensics odontology. J. Forens. Sci. Crim. Inves. 7, 1–5. doi: 10.19080/JFSCI.2018.07.555723

CrossRef Full Text | Google Scholar

Jordan, M. I., and Mitchell, T. M. (2015). Machine learning: trends, perspectives, and prospects. Science. 349, 255–60. doi: 10.1126/science.aaa8415

PubMed Abstract | CrossRef Full Text | Google Scholar

Kahaki, S. M., Nordin, M., Ahmad, N. S., Arzoky, M., and Ismail, W. (2020). Deep convolutional neural network designed for age assessment based on orthopantomography data. Neural Comp. App. 32, 9357–9368. doi: 10.1007/s00521-019-04449-6

CrossRef Full Text | Google Scholar

Kanthem, R. K., Guttikonda, V. R., Yeluri, S., and Kumari, G. (2015). Sex determination using maxillary sinus. J. For. Dent. Sci. 7, 163. doi: 10.4103/0975-1475.154595

PubMed Abstract | CrossRef Full Text | Google Scholar

Karki, R. (2012). Lip prints–an identification aid. Kathmandu Univ. Med. J. 10, 55–7. doi: 10.3126/kumj.v10i2.7345

PubMed Abstract | CrossRef Full Text | Google Scholar

Kashyap, B., Anand, S., Reddy, S., Sahukar, S. B., Supriya, N., Pasupuleti, S., et al. (2015). Comparison of the bite mark pattern and intercanine distance between humans and dogs. J. For. Dent Sci. 7, 175. doi: 10.4103/0975-1475.172419

PubMed Abstract | CrossRef Full Text | Google Scholar

Kaul, R., Padmashree, S. M., Shilpa, P. S., Sultana, N., and Bhat, S. (2015). Cheiloscopic patterns in Indian population and their efficacy in sex determination: a randomized cross-sectional study. J. For. Dent. Sci. 7, 101. doi: 10.4103/0975-1475.156192

PubMed Abstract | CrossRef Full Text | Google Scholar

Kaur, S., Krishan, K., Chatterjee, P. M., and Kanchan, T. (2013). Analysis and identification of bite marks in forensic casework. Oral Health Dent Manag. 12, 127–31. doi: 10.4172/2247-2452.1000500

PubMed Abstract | CrossRef Full Text | Google Scholar

Khanagar, S. B., Vishwanathaiah, S., Naik, S., Al-Kheraif, A. A., Divakar, D. D., Sarode, S. C., et al. (2021). Application and performance of artificial intelligence technology in forensic odontology–A systematic review. Legal Med. 48, 101826. doi: 10.1016/j.legalmed.2020.101826

PubMed Abstract | CrossRef Full Text | Google Scholar

Kim, S., Lee, Y.-H., Noh, Y.-K., Park, F. C., and Auh, Q. (2021). Age-group determination of living individuals using first molar images based on artificial intelligence. Sci. Rep. 11, 1–11. doi: 10.1038/s41598-020-80182-8

PubMed Abstract | CrossRef Full Text | Google Scholar

Kinra, M., Ramalingam, K., Sethuraman, S., Rehman, F., Lalawat, G., Pandey, A., et al. (2014). Cheiloscopy for sex determination: a study. J. Dent.. 4, 48–51. doi: 10.4103/2249-9725.127078

CrossRef Full Text | Google Scholar

Krishan, K., Chatterjee, P. M., Kanchan, T., Kaur, S., Baryah, N., Singh, R. A., et al. (2016). review of sex estimation techniques during examination of skeletal remains in forensic anthropology casework. For. Sci. Int. 261, 165. doi: 10.1016/j.forsciint.2016.02.007

PubMed Abstract | CrossRef Full Text | Google Scholar

Krogman, W. M., and Isçan, M. Y. (1986). The Human Skeleton in For. Medicine. Springfield, IL: Charles C. Thomas.

Google Scholar

Krzywanski, J. (2022). Advanced AI applications in energy and environmental engineering systems. MDPI 15, 5621. doi: 10.3390/en15155621

CrossRef Full Text | Google Scholar

Kumar, B., Ramesh, T., Reddy, R., Chennoju, S., Pavani, K., Praveen, K. A., et al. (2016). morphometric analysis of mandible for sex determination-A retrospective study. Int. J. Sci. Res. Methodol. 4, 115–22. Available online at: https://ijsrm.humanjournals.com/

PubMed Abstract | Google Scholar

Kurniawan, A., Yodokawa, K., Kosaka, M., Ito, K., Sasaki, K., Aoki, T., et al. (2020). Determining the effective number and surfaces of teeth for forensic dental identification through the 3D point cloud data analysis. Egyptian J. For. Sci. 10, 1–11. doi: 10.1186/s41935-020-0181-z

CrossRef Full Text | Google Scholar

Lai, Y., Fan, F., Wu, Q., Ke, W., Liao, P., Deng, Z., et al. (2020). Lcanet: learnable connected attention network for human identification using dental images. IEEE Trans. Med. Imag. 40, 905–15. doi: 10.1109/TMI.2020.3041452

PubMed Abstract | CrossRef Full Text | Google Scholar

LeCun, Y, Bengio, Y, and Hinton, G. (2015). Deep learning. Nature 521, 436–44. doi: 10.1038/nature14539

PubMed Abstract | CrossRef Full Text | Google Scholar

Lee, J.-H., Han, S.-S., Kim, Y. H., Lee, C., and Kim, I. (2020). Application of a fully deep convolutional neural network to the automation of tooth segmentation on panoramic radiographs. Oral Surg. Oral Med. Oral Pathol. Radiol. 129, 635–642. doi: 10.1016/j.oooo.2019.11.007

PubMed Abstract | CrossRef Full Text | Google Scholar

Lessig, R., Weber, M., and Wenzel, V. (2006). Bite Mark Analysis in Forensic Routine Case Work.

Google Scholar

Levine, L. J. (1972). For. Odontology Today: A“ new” For. Science: Federal Bureau of Investigation. Washington, DC: US Department of Justice.

Google Scholar

Liang, Y., Han, W., Qiu, L., Wu, C., Shao, Y., Wang, K., et al. (2021). Exploring Forensic Dental Identification with Deep Learning. Adv. Neural Inf. Proc. Syst. 34, 3244–58. Available online at: https://nips.cc/Conferences/2021

Google Scholar

Luntz, L., and Luntz, P. (1973). Antemortem Records. Handbook for Dental Identification. Philadelphia: JB Lippincott Company.

Google Scholar

Maber, M., Liversidge, H., and Hector, M. (2006). Accuracy of age estimation of radiographic methods using developing teeth. Foren. Sci. Int.. 159, S68–S73. doi: 10.1016/j.forsciint.2006.02.019

PubMed Abstract | CrossRef Full Text | Google Scholar

Mahasantipiya, P., Yeesarapat, U., Suriyadet, T., Sricharoen, J., Dumrongwanich, A., Thaiupathump, T., et al. (2012). Bite mark identification using neural networks: A preliminary study. World Congress on Engineering 2012 July 4-6, 2012. London: Citeseer.

Google Scholar

Mahdi, F. P., Motoki, K., and Kobashi, S. (2020). Optimization technique combined with deep learning method for teeth recognition in dental panoramic radiographs. Sci. Rep. 10, 1–12. doi: 10.1038/s41598-020-75887-9

PubMed Abstract | CrossRef Full Text | Google Scholar

Maji, A., Khaitan, T., Sinha, R., Sarkar, S., Verma, P., Shukla, A. K. A., et al. (2018). A novel computer-assisted method of bite mark analysis for gender determination. J. Environ. Pub. Health. 2018, 7130876. doi: 10.1155/2018/7130876

PubMed Abstract | CrossRef Full Text | Google Scholar

Matsuda, S., Miyamoto, T., Yoshimura, H., and Hasegawa, T. (2020). Personal identification with orthopantomography using simple convolutional neural networks: a preliminary study. Sci. Rep. 10, 1–7. doi: 10.1038/s41598-020-70474-4

PubMed Abstract | CrossRef Full Text | Google Scholar

Merdietio Boedi, R., Banar, N., De Tobel, J., Bertels, J., Vandermeulen, D., Thevissen, P. W., et al. (2020). Effect of lower third molar segmentations on automated tooth development staging using a convolutional neural network. J. For. Sciences. 65, 481–6. doi: 10.1111/1556-4029.14182

PubMed Abstract | CrossRef Full Text | Google Scholar

Miki, Y., Muramatsu, C., Hayashi, T., Zhou, X., Hara, T., Katsumata, A., et al. (2017). Tooth labeling in cone-beam CT using deep convolutional neural network for forensic identification. Comput. Aided Diag. 10314, 874–879. doi: 10.1117/12.2254332

CrossRef Full Text | Google Scholar

Milošević, D., Vodanović, M., Galić, I., and Subašić, M. (2021). “automated sex assessment of individual adult tooth X-Ray images,” in12th International Symposium on Image and Signal Processing and Analysis (ISPA). Piscataway, NJ: IEEE.

Google Scholar

Milošević, D., Vodanović, M., Galić, I., and Subašić, M. (2022). Automated estimation of chronological age from panoramic dental X-ray images using deep learning. Expert Syst. App. 189, 116038. doi: 10.1016/j.eswa.2021.116038

CrossRef Full Text | Google Scholar

Mohammad, N., Muad, A. M., Ahmad, R., and Yusof, M. Y. P. M. (2021). Reclassification of Demirjian's mandibular premolars staging for age estimation based on semi-automated segmentation of deep convolutional neural network. For. Imag. 24, 200440. doi: 10.1016/j.fri.2021.200440

CrossRef Full Text | Google Scholar

Mohammad, N., Muad, A. M., Ahmad, R., and Yusof, M. Y. P. M. (2022). Accuracy of advanced deep learning with tensorflow and keras for classifying teeth developmental stages in digital panoramic imaging. BMC Med. Imag. 22, 1–13. doi: 10.1186/s12880-022-00794-6

PubMed Abstract | CrossRef Full Text | Google Scholar

Mohammad, N., Yusof, M., Ahmad, R., and Muad, A. (2020). Region-based segmentation and classification of mandibular first molar tooth based on Demirjian's method. J. Phy. Conf. Series 1502, 12046. doi: 10.1088/1742-6596/1502/1/012046

CrossRef Full Text | Google Scholar

Molina, A., Ramos, B., and Torres, J. C. (2022). Dental parameter quantification with semi-automatized computational technology for the analysis of human bitemarks. Austr. J. For. Sci. 54, 247–57. doi: 10.1080/00450618.2020.1793385

CrossRef Full Text | Google Scholar

More, C. B., Vijayvargiya, R., and Saha, N. (2017). Morphometric analysis of mandibular ramus for sex determination on digital orthopantomogram. J. For. Dent. Sci. 9, 1. doi: 10.4103/jfo.jfds_25_15

PubMed Abstract | CrossRef Full Text | Google Scholar

Munn, Z., Peters, M. D., Stern, C., Tufanaru, C., McArthur, A., Aromataris, E., et al. (2018). Systematic review or scoping review? Guidance for authors when choosing between a systematic or scoping review approach. BMC Med. Res. Methodol. 18, 1–7. doi: 10.1186/s12874-018-0611-x

PubMed Abstract | CrossRef Full Text | Google Scholar

Nagare, S. P., Chaudhari, R. S., Birangane, R. S., and Parkarwar, P. C. (2018). Sex determination in forensic identification, a review. J. Forens. Dent. Sci. 10, 61. doi: 10.4103/jfo.jfds_55_17

PubMed Abstract | CrossRef Full Text | Google Scholar

Naru, A., and Dykes, E. (1996). The use of a digital imaging technique to aid bite mark analysis. Sci. Just. J. For. Sci. Soc. 36, 47–50. doi: 10.1016/S1355-0306(Heet al.,2016)72554-6

PubMed Abstract | CrossRef Full Text | Google Scholar

Nithya, L., and Sornam, M. (2022). Deep Convolutional Networks in Gender Classification Using Dental X-Ray Images. Artificial Intelligence and Evolutionary Computations in Engineering Systems. Berlin: Springer, 375–80.

Google Scholar

Olze, A., Hertel, J., Schulz, R., Wierer, T., and Schmeling, A. (2012). Radiographic evaluation of Gustafson's criteria for the purpose of forensic age diagnostics. Int. J. Leg. Med. 126, 615–21. doi: 10.1007/s00414-012-0701-y

PubMed Abstract | CrossRef Full Text | Google Scholar

Ortiz, A., Costa, C., Silva, R., Biazevic, M., and Michel-Crosato, E. (2020). Sex estimation: anatomical references on panoramic radiographs using Machine Learning. For. Imag. 20, 200356. doi: 10.1016/j.fri.2020.200356

CrossRef Full Text | Google Scholar

Osman, N. A., Omer, A. Z., and Abuaffan, A. H. (2017). Comparative study on two methods for bite mark analysis. ARC J. Forens. Sci. 2, 12–6. doi: 10.20431/2456-0049.0201004

PubMed Abstract | CrossRef Full Text | Google Scholar

Panchbhai, A. (2011). Dental radiographic indicators, a key to age estimation. Dentomaxillo. Radiolo. 40, 199–212. doi: 10.1259/dmfr/19478385

PubMed Abstract | CrossRef Full Text | Google Scholar

Patil, V., Pai, K. M., and Naik, N. (2018). Determination of pulp/tooth area ratio and age estimation: a pilot study. World J. Dent. 9, 424–8. doi: 10.5005/jp-journals-10015-1573

PubMed Abstract | CrossRef Full Text | Google Scholar

Patil, V., Vineetha, R., Vatsa, S., Shetty, D. K., Raju, A., Naik, N., et al. (2020). Artificial neural network for gender determination using mandibular morphometric parameters: a comparative retrospective study. Cogent Eng. 7, 1723783. doi: 10.1080/23311916.2020.1723783

CrossRef Full Text | Google Scholar

Peters, M., Godfrey, C., McInerney, P., Soares, C. B., Khalil, H., Parker, D., et al. (2015). Methodology for JBI Scoping Reviews. The Joanna Briggs Institute Reviewers Manual. Adelaide: Joanna Briggs Institute, 3–24.

Google Scholar

Peters, M., Ketter, W., Saar-Tsechansky, M., and Collins, J. A. (2013). Reinforcement learning approach to autonomous decision-making in smart electricity markets. Mach. Learn. 92, 5–39. doi: 10.1007/s10994-013-5340-0

CrossRef Full Text | Google Scholar

Peters, M. D., Marnie, C., Tricco, A. C., Pollock, D., Munn, Z., Alexander, L., et al. (2020). Updated methodological guidance for the conduct of scoping reviews. JBI Evid. Synth. 18, 2119–26. doi: 10.11124/JBIES-20-00167

PubMed Abstract | CrossRef Full Text | Google Scholar

Phillips, I., and Kenley, R. (2022). “Verification of intelligent transportation systems: challenges and possibilities,” in 2022 17th Annual System of Systems Engineering Conference (SOSE). Piscataway, NJ: IEEE.

Google Scholar

Pretty, I., and Sweet, D. A. (2001a). look at forensic dentistry–Part 1: the role of teeth in the determination of human identity. Br. Dent. J. 190, 359–66. doi: 10.1038/sj.bdj.4800972

PubMed Abstract | CrossRef Full Text | Google Scholar

Pretty, I. A., and Sweet, D. (2000). Anatomical location of bitemarks and associated findings in 101 cases from the United States. J. Foren. Sci. 45, 812–4. doi: 10.1520/JFS14775J

PubMed Abstract | CrossRef Full Text | Google Scholar

Pretty, I. A., and Sweet, D. (2001b). The scientific basis for human bitemark analyses–a critical review. Sci.: J. For. Sci. Soc. 41, 85–92. doi: 10.1016/S1355-0306(01)71859-X

PubMed Abstract | CrossRef Full Text | Google Scholar

Pretty, I. A., and Turnbull, M. D. (2001). Lack of dental uniqueness between two bite mark suspects. J. Foren. Sci. 46, 1487–91. doi: 10.1520/JFS15177J

PubMed Abstract | CrossRef Full Text | Google Scholar

Putra, R. H., Doi, C., Yoda, N., Astuti, E. R., and Sasaki, K. (2022). Current applications and development of artificial intelligence for digital dental radiography. Dentomaxillofacial Radiol. 51, 20210197. doi: 10.1259/dmfr.20210197

PubMed Abstract | CrossRef Full Text | Google Scholar

Raj, J. D., and Ramesh, S. (2013). Sexual dimorphism in mandibular ramus of South Indian population. Antrocom Online J. Anthropol. 9, 253–8. Available online at: http://www.antrocom.net/

Google Scholar

Ramakrishnan, K., Sharma, S., Sreeja, C., Pratima, D. B., Aesha, I., Vijayabanu, B., et al. (2015). Sex determination in forensic odontology: a review. J. Pharm. Bioall. Sci. 7, S398. doi: 10.4103/0975-7406.163469

PubMed Abstract | CrossRef Full Text | Google Scholar

Reddy, Y., Viswanath, P., and Reddy, B. E. (2018). Semi-supervised learning: a brief review. Int. J. Eng. Technol. 7, 81. doi: 10.14419/ijet.v7i1.8.9977

CrossRef Full Text | Google Scholar

Redmon, J., Divvala, S., Girshick, R., and Farhadi, A. (2016). “You only look once: Unified, real-time object detection,” in Proceedings of the IEEE conference on computer vision and pattern recognition. Piscataway, NJ: IEEE.

Google Scholar

Rizwal, A. A., Azahar, N., Reduwan, N. H., and Yusof, M. Y. P. M. (2021). Superimposed polygonal approximation analysis comparing 2D photography and 3D scanned images of bite marks on human skin. Egypt. J. Foren. Sci. 11, 18. doi: 10.1186/s41935-021-00233-7

CrossRef Full Text | Google Scholar

Rudin, C. (2019). Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nat. Machin. Int. 1, 206–15. doi: 10.1109/ICCONS.2018.8663155

PubMed Abstract | CrossRef Full Text | Google Scholar

Russakovsky, O., Deng, J., Su, H., Krause, J., Satheesh, S., Ma, S., et al. (2015). Imagenet large scale visual recognition challenge. Int. J. Comput. Vision. 115, 211–52. doi: 10.1007/s11263-015-0816-y

CrossRef Full Text | Google Scholar

Saks, M. J., Albright, T., Bohan, T. L., Bierer, B. E., Bowers, C. M., Bush, M. A., et al. (2016). Forensic bitemark identification: weak foundations, exaggerated claims. J. Law Biosci. 3, 538–75. doi: 10.1093/jlb/lsw045

PubMed Abstract | CrossRef Full Text | Google Scholar

Saravanan, R., and Sujatha, P. (2018). “A state of art techniques on machine learning algorithms: a perspective of supervised learning approaches in data classification,” in 2018 Second International Conference on Intelligent Computing and Control Systems (ICICCS). Piscataway, NJ: IEEE.

Google Scholar

Shah, P. H., and Venkatesh, R. (2016). Pulp/tooth ratio of mandibular first and second molars on panoramic radiographs: an aid for forensic age estimation. J. For. Dental Sci. 8, 112. doi: 10.4103/0975-1475.186374

PubMed Abstract | CrossRef Full Text | Google Scholar

Shanthamallu, U. S., and Spanias, A. (2021). Machine and deep learning algorithms and applications. Synth. Lect. Sig. Process. 12, 1–123. doi: 10.1007/978-3-031-03758-0

PubMed Abstract | CrossRef Full Text | Google Scholar

Silver, D., Hubert, T., Schrittwieser, J., Antonoglou, I., Lai, M., Guez, A., et al. (2018). A general reinforcement learning algorithm that masters chess, shogi, and go through self-play. Science. 362, 1140–4. doi: 10.1126/science.aar6404

PubMed Abstract | CrossRef Full Text | Google Scholar

Simonyan, K., and Zisserman, A. (2015). “Very deep convolutional networks for large-scale image recognition,” in 3rd International Conference on Learning Representations, ICLR 2015 - Conference Track Proceedings.

Google Scholar

Singh, A., Thakur, N., and Sharma, A. (2016). “A review of supervised machine learning algorithms,” in 2016 3rd International Conference on Computing for Sustainable Global Development (INDIACom). Piscataway, NJ: IEEE.

Google Scholar

Sörup, A. (1926). Odontoskopie ein beitrag zur zahnärztlichen kriminalistik. Dtsch Monatsschr Zahnheilkd. 21, 777–84.

Google Scholar

Sun, Z., Sandoval, L., Crystal-Ornelas, R., Mousavi, S. M., Wang, J., Lin, C., et al. (2022). A review of earth artificial intelligence. Comput. Geosci. 5, 105034. doi: 10.1109/SOSE55472.2022.9812702

PubMed Abstract | CrossRef Full Text | Google Scholar

Sweet, D., and Bowers, C. M. (1998). Accuracy of bite mark overlays: a comparison of five common methods to produce exemplars from a suspect's dentition. J. Forens. Sci. 43, 362–7. doi: 10.1520/JFS16146J

PubMed Abstract | CrossRef Full Text | Google Scholar

Sweet, D., Parhar, M., and Wood, R. E. (1998). Computer-based production of bite mark comparison overlays. J. For. Sci. 43, 1050–5. doi: 10.1520/JFS14356J

PubMed Abstract | CrossRef Full Text | Google Scholar

Szegedy, C., Ioffe, S., Vanhoucke, V., and Alemi, A. A. (2017). Inception-v4, inception-resnet and the impact of residual connections on learning. arXiv [Preprint]. arXiv:1602.07261. Available online at: https://arxiv.org/pdf/1602.07261.pdf

Google Scholar

Szegedy, C., Liu, W., Jia, Y., Sermanet, P., Reed, S., Anguelov, D., et al. (2015). “Going deeper with convolutions,” in Proceedings of the IEEE conference on computer vision and pattern recognition. Piscataway, NJ: IEEE.

Google Scholar

Tafrount, C., Galić, I., Franchi, A., Fanton, L., and Cameriere, R. (2019). Third molar maturity index for indicating the legal adult age in southeastern France. For. Sci. Int. 294, 218. doi: 10.1016/j.forsciint.2018.10.013

PubMed Abstract | CrossRef Full Text | Google Scholar

TensorFlow Developers. (2022). TensorFlow (v2.8.2). Zenodo. doi: 10.5281/zenodo.6574269

CrossRef Full Text | Google Scholar

Thabitha, R. S., Reddy, R. E., Manjula, M., Sreelakshmi, N., Rajesh, A., Kumar, V. L., et al. (2015). Evaluation of palatal rugae pattern in establishing identification and sex determination in Nalgonda children. J. For. Dent. Sci. 7, 232. doi: 10.4103/0975-1475.172447

PubMed Abstract | CrossRef Full Text | Google Scholar

Tsuchihashi, Y. (1974). Studies on personal identification by means of lip prints. Forens. Sci. 3, 233–48. doi: 10.1016/0300-9432(Molinaet al.,2022)90034-X

PubMed Abstract | CrossRef Full Text | Google Scholar

Upalananda, W., Wantanajittikul, K., Na Lampang, S., and Janhom, A. (2021). Semi-automated technique to assess the developmental stage of mandibular third molars for age estimation. Austr. J. For. Sci. 2021, 1–11. doi: 10.1080/00450618.2021.1882570

CrossRef Full Text | Google Scholar

Uzuner, F. D., Kaygisiz, E., and Darendeliler, N. (2017). Defining Dental Age for Chronological Age Determination. Post Mortem Examination and Autopsy-Current Issues From Death to Laboratory Analysis. London: IntechOpen.

Google Scholar

Van der Velden, A., Spiessens, M., and Willems, G. (2006). Bite mark analysis and comparison using image perception technology. J. Foren. Odontostomatol. 24, 14. Available online at: http://www.iofos.eu/?page_id=255

PubMed Abstract | Google Scholar

Vinay, G., Mangala Gowri, S., and Anbalagan, J. (2013). Sex determination of human mandible using metrical parameters. J. Clin. Diag. Res. 7, 2671. doi: 10.7860/JCDR/2013/7621.3728

PubMed Abstract | CrossRef Full Text | Google Scholar

West, M. H., Billings, J. D., and Frair, J. (1987). Ultraviolet photography: bite marks on human skin and suggested technique for the exposure and development of reflective ultraviolet photography. J. Foren. Sci. 32, 1204–13. doi: 10.1520/JFS11171J

CrossRef Full Text | Google Scholar

Zhang, Y., and Ling, C. A. (2018). Strategy to apply machine learning to small datasets in materials science. NPJ Comput. Mat. 4, 1–8. doi: 10.1038/s41524-018-0081-z

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: forensic odontology, human identification, dental age estimation (DAE), machine learning (ML), artificial neural network (ANN), deep learning

Citation: Mohammad N, Ahmad R, Kurniawan A and Mohd Yusof MYP (2022) Applications of contemporary artificial intelligence technology in forensic odontology as primary forensic identifier: A scoping review. Front. Artif. Intell. 5:1049584. doi: 10.3389/frai.2022.1049584

Received: 26 September 2022; Accepted: 21 November 2022;
Published: 06 December 2022.

Edited by:

Rashid Mehmood, King Abdulaziz University, Saudi Arabia

Reviewed by:

Muzafer Saracevic, University of Novi Pazar, Serbia
Alberto Ochoa Zezzatti, Universidad Autónoma de Ciudad Juárez, Mexico

Copyright © 2022 Mohammad, Ahmad, Kurniawan and Mohd Yusof. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Mohd Yusmiaidil Putera Mohd Yusof, yusmiaidil@uitm.edu.my

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.