Skip to main content

MINI REVIEW article

Front. Toxicol., 19 January 2024
Sec. Computational Toxicology and Informatics
This article is part of the Research Topic Next Generation Chemical Risk Assessment: Integration of Advances in Toxicology, Biology and Computation View all 5 articles

Computational models for predicting liver toxicity in the deep learning era

  • 1Department of Mathematics and Statistics, Texas Tech University, Lubbock, TX, United States
  • 2Division of Bioinformatics and Biostatistics, National Center for Toxicological Research, U.S. Food and Drug Administration, Jefferson, AR, United States

Drug-induced liver injury (DILI) is a severe adverse reaction caused by drugs and may result in acute liver failure and even death. Many efforts have centered on mitigating risks associated with potential DILI in humans. Among these, quantitative structure-activity relationship (QSAR) was proven to be a valuable tool for early-stage hepatotoxicity screening. Its advantages include no requirement for physical substances and rapid delivery of results. Deep learning (DL) made rapid advancements recently and has been used for developing QSAR models. This review discusses the use of DL in predicting DILI, focusing on the development of QSAR models employing extensive chemical structure datasets alongside their corresponding DILI outcomes. We undertake a comprehensive evaluation of various DL methods, comparing with those of traditional machine learning (ML) approaches, and explore the strengths and limitations of DL techniques regarding their interpretability, scalability, and generalization. Overall, our review underscores the potential of DL methodologies to enhance DILI prediction and provides insights into future avenues for developing predictive models to mitigate DILI risk in humans.

1 Introduction

Drug-induced liver injury (DILI) is a substantial safety concern, with a reported potential for more than 1,000 drugs or supplements to induce liver damage (Alempijevic et al., 2017; Zhu et al., 2018). DILI presents a significant challenge for healthcare professionals, pharmaceutical developers, and regulatory authorities (George et al., 2018), and frequently results in the discontinuation of drug candidates during their development (Weber and Gerbes, 2022). It also is a primary reason for the withdrawal of over 50 medications from the market (Devarbhavi, 2012; Wu et al., 2022) and ranks as a leading cause of acute liver failure in both the United States and Europe (Andrade et al., 2019). Despite notable advancements in drug safety, there is a continuing need for innovative approaches and methodologies to identify drugs candidates in development with potential hepatotoxicity in humans, and for reliable biomarkers to facilitate the early detection of DILI (Chen et al., 2014).

The need to enhance safety assessments in drug development has driven new approaches for predicting toxicity. Conventional methods often lack the precision and efficiency required to mitigate the risks associated with liver toxicity. However, machine learning, (ML), which includes Quantitative Structure-Activity Relationship (QSAR) modeling (Idakwo et al., 2019; Shin et al., 2023) as a pivotal component, harnesses extensive datasets, chemical structures, and biological assays to establish quantitative associations between molecular properties and toxicity outcomes (Wu et al., 2017). This approach has the potential to facilitate the detection of liver toxicity during the early stage drug development process, enabling screening of drug candidates and their analogs prior to chemical synthesis.

An advanced ML technique, deep learning (DL), signifies a transformative approach in the field of liver toxicity prediction, offering the potential for exceptionally accurate, data-driven insights. DL harnesses neural networks and extensive datasets, which encompass chemical data, biological assays, and omics information, to construct predictive models of outstanding performance (Xu et al., 2015; Goh et al., 2018). Its integration into liver toxicity prediction empowers researchers and pharmaceutical companies to identify potential risks associated with drug candidates at an early stage in the development process. Moreover, its capacity to analyze diverse and intricate data sources facilitates a better understanding of toxicity mechanisms. Consequently, DL not only advances patient safety by aiding in identifying harmful compounds, but also is cost-effective and contributes to the accelerated development of safer and more effective medications.

In this review, we focus on cutting-edge research using ML/DL applications to predict liver toxicity. We first examine the application of ML in liver toxicity prediction, with a particular emphasis on the development of QSAR models. Next, we provide a systematic evaluation of DL methods and their application for predicting liver toxicity, drawing comparisons with traditional ML approaches. Finally, we discuss the strengths and limitations of DL methods in interpretability, scalability, and generalization.

2 Machine learning for predicting liver toxicity

Machine learning (ML) algorithms have extensive applications in classification tasks, including the prediction of liver toxicity (Supplementary Table S1). In binary classification, compounds are typically categorized into two classes: a toxic class (commonly labeled class 1) and a non-toxic class (class 0). ML algorithms can learn from historical data and categorize new instances into one of these two classes by considering their observed characteristics, such as chemical structures. Among the various ML methods available, Naive Bayes Classifier (NBC), Support Vector Machines (SVM), and Random Forests are widely employed in this context.

2.1 Naive Bayes classifier

The Naive Bayes classifier (NBC) is a probabilistic ML algorithm widely used for both binary and multiclass classification tasks (Rish, 2001; Hastie et al., 2009). It is rooted in Bayes’ theorem, which quantifies the probability of an event based on prior knowledge of related events. The “Naive” component of its name comes from the assumption that input features are conditionally independent, simplifying calculations and enhancing computational efficiency. Thus, the NBC computes the conditional probability of a given instance belonging to a specific class by making the “naive” assumption of feature independence. Mathematically, it leverages Bayes’ theorem:

PCk|x=PCk|xi=1nPxi|CkPx,

where Ck is the class, x is the feature vectors.

The NBC makes the “naive” assumption that features are independent given the class. This strong assumption might not hold in all real-world scenarios. However, despite this simplification, it often performs surprisingly well and is computationally efficient. Critical steps to train and use NBC are listed below:

1. Calculate Class Priors: Estimate the prior probabilities PCk|x for each class based on the training data.

2. Calculate Feature Probabilities: Estimate the conditional probabilities Pxi|Ck for each feature and class pair based on the training data. This involves counting occurrences of features in each class.

3. Classification: Given a new instance with features x1,x2,,xn, calculate the posterior probability for each class using Bayes’ theorem. The class with the highest probability is the predicted class.

Variations of NBCs are based on types of data and assumptions. Some common variations are:

• Gaussian Naive Bayes: Assumes features follow a Gaussian (normal) distribution.

• Multinomial Naive Bayes: Suited for discrete features like text data, and often used for document classification.

• Bernoulli Naive Bayes: Designed for binary feature data (presence/absence), and often used for text classification tasks.

NBC is a straightforward yet effective classifier, particularly suitable for binary classification tasks, provided that the assumption of feature independence is reasonably met. While it may not be the optimal choice for all data types, it serves as a standardized baseline classifier and is extensively employed in the prediction DILI through QSAR modeling (Ai et al., 2019; Williams et al., 2019; Wu Y. et al., 2021). For instance, Zhang et al. (2016) employed NBC to construct a computational model for assessing DILI risk. Their model exhibited a 94.0% accuracy in 5-fold cross-validation during the training phase, with a concordance rate of 72.6% on an external test set. They identified key molecular characteristics associated with DILI risk.

Tang et al. (2020) developed QSAR models for mitochondrial toxicity using five machine learning methods, including NBC along with various chemical signatures. They adopted a threshold moving strategy to rectify data imbalance and implemented consensus models to enhance prediction performance, achieving up to 88.3% accuracy in external validation. Notably, the study highlighted the significance of substructures such as phenol, carboxylic acid, nitro compounds, and aryl chloride in classification. In another work, Rao et al. (2023), proposed an integrated artificial intelligence (AI)/ML model that employed physicochemical properties and in silico off-target interactions to predict the severity of DILI for small molecules. They utilized data from 603 compounds categorized by the U.S. Food and Drug Administration (FDA) as Most DILI, Less DILI, and No DILI, and combined the NBC with other ML approaches to enhance DILI prediction, surpassing the performance of QSAR models based solely on chemical properties.

2.2 Support vector machine classifier

The Support Vector Machine (SVM) is a potent and versatile machine learning algorithm used for a range of tasks, including regression, binary, and multiclass classification, as is the case in predicting DILI through QSAR modeling with chemical structures (Li et al., 2020a; Tang et al., 2020; Wu Z et al., 2021; Rao et al., 2023). Its primary goal is to identify a hyperplane that maximizes the margin between the nearest data points from the two classes. The fundamental concept is to optimize this margin between classes, resulting in improved generalization to new, unseen data. These closest data points are referred to as “support vectors.” The margin is defined as the distance between the hyperplane and these support vectors. Mathematically, the SVM tries to solve the following optimization problem:

12minw,bw2

Subject to

yiwxi+b1fori=1:n,

where:

• w is the weight vector perpendicular to the hyperplane.

• b is the bias term.

xi are the feature vectors.

yi are the class labels (1 or 0) for each data point.

• n is the number of data points.

The above optimization problem ensures that data points are correctly classified with a margin. Support vectors are the data points that lie on the margins or violate the margin constraint. In many cases, the data may not be linearly separable in the original feature space. To handle these cases, SVMs often use a kernel trick. A kernel function transforms the original feature space into a higher-dimensional space, where the data might become separable. Common kernel functions include:

• Linear Kernel: Kx,x=xx.

• Polynomial Kernel: Kx,x=xx+cd.

• Radial Basis Function (RBF) Kernel: Kx,x=eγxx2.

In some cases, the data might not be perfectly separable, or there could be outliers. In such situations, SVM employed a soft margin to allow for some misclassification by introducing a slack variable. The optimization problem becomes:

12minw,bw2+Cj=1nτj

Subject to:

yiwxi+b1τjfori,j=1:n,τj0,

where C is a hyperparameter that controls the trade-off between maximizing the margin and minimizing misclassification. Major steps to train and apply the SVM Classifier are listed below.

1. Data Preparation: Gather and preprocess the data, ensuring it is properly labeled, and features are appropriately represented.

2. Choose a Kernel: Decide on a kernel function based on the data characteristics, which can be critical to improving accuracy in DILI predictions (Wu Y. et al., 2021).

3. Train the SVM: Use an optimization algorithm to find the optimal hyperplane parameters (weights w and bias b) that minimize the objective function.

4. Classification: Given a new instance with features x, calculate the decision function fx=wx+b. If fx>0, classify as class 1; if fx<0, classify as class 0. SVMs can be computationally intensive for large datasets, and tuning the hyperparameters, such as the choice of kernel and the regularization parameter C, is essential for optimal performance.

SVM has been applied extensively in predicting DILI, particularly in scenarios with limited data, owing to its robust prediction accuracy and computational efficiency. Notable studies (Wang et al., 2019; Li et al., 2020b; Hemmerich et al., 2020; Mora et al., 2020; Wu Z. et al., 2021) have employed SVM for DILI prediction. In a comprehensive analysis conducted by Wu Y. et al. (2021), involving 14 sets of QSAR data and 16 ML algorithms, the radial basis function SVM (rbf-SVM) emerged as the top-performing method among all ML techniques, underscoring its efficacy in this domain.

However, certain limitations were associated with the SVM algorithm. SVM tends to be computationally expensive and may not be well-suited for very large datasets. When the dataset exhibits extra noise, such as overlapping target classes, SVM’s performance can be compromised. Furthermore, SVM may perform suboptimally when the number of features for each data point exceeds the number of training data samples. These considerations are critical when deciding on the suitability of SVM for specific DILI prediction tasks.

2.3 Random forest classifier

The Random Forest classifier is a robust ensemble ML algorithm frequently employed for both binary and multiclass classification tasks. It is an extension of the decision tree algorithm, having the primary objective of enhancing generalization and mitigating overfitting by forming an ensemble of multiple decision trees. In binary classification, the Random Forest classifier is designed to classify new instances into one of two classes based on their features. The process involves the construction of multiple decision trees during the training phase, and their collective predictions are amalgamated to reach the final classification decision. Key steps for training and applying the Random Forest classifier are outlined below.

1. Bootstrapped Sampling: For each tree in the forest, a random subset of the training data is selected with replacement. This process is known as bootstrapped sampling. It creates diversity among the trees, as each tree is trained on a slightly different data subset.

2. Random Feature Selection: At each split point in a decision tree, only a subset of the available features is considered for splitting. This introduces further randomness and prevents individual trees from relying on any one feature.

3. Tree Building: Each decision tree is constructed using the bootstrapped training data and random feature selection. The tree is grown until a stopping criterion is met, usually involving the maximum depth of the tree or the minimum number of samples required to split a node.

4. Voting for Classification: During prediction, each tree in the forest independently classifies the input data. The final classification decision is made by taking a majority vote among the individual tree predictions. In the case of binary classification, the class with the most votes wins.

Compared with other machine learning methods, random forest has several unique characteristics:

• Reduced Overfitting: The ensemble of trees helps to mitigate overfitting by averaging out the noise and biases present in individual trees.

• Improved Generalization: Random Forests are robust to outliers and noisy data due to the aggregation of multiple trees.

• Feature Importance: Random Forests can provide insights into feature importance by analyzing how much each feature contributes to the model’s performance.

• Non-linearity Handling: Random Forests can capture complex relationships in the data without requiring explicit feature extraction/selection.

Random Forest emerges as an invaluable machine learning technique for the classification of liver toxicity. Gadaleta et al. (2018) employed Random Forest classifiers and DRAGON molecular descriptors to create QSAR models designed to predict molecular initiating events leading to hepatic steatosis. They effectively used a Balanced Random Forest classifier, alongside the strategy of under-sampling, to construct robust QSAR models from unbalanced DILI datasets. Both techniques yielded comparable predictive results, achieving approximately 75% accuracy in toxicity prediction.

3 Deep learning for predicting liver toxicity

Deep learning (DL) represents a new class of machine learning methods characterized by the use of highly complex neural networks. Networks are structured in deeply nested architectures, often incorporating advanced operations like convolutions and multiple activation functions. These distinctive features empower DL with the unique capability to process raw input data and autonomously uncover hidden patterns for learning tasks. In the context of predicting liver toxicity, several DL methods are commonly employed for classification tasks. These methods deploy neural networks with diverse architectures and techniques to achieve precise and efficient classification (Supplementary Table S1). We provide a brief overview of various DL methods, including multi-layer perceptron (MLP), deep neural network (DNN), convolutional neural network (CNN), graph neural network (GNN), recurrent neural network (RNN), generative adversarial network (GAN), and transformer.

3.1 Multi-layer perceptron

The Multilayer Perceptron (MLP), also known as an Artificial Neural Network (ANN), is a fundamental neural network architecture used for a wide range of machine learning applications, including classification and regression. An MLP consists of multiple layers of artificial neurons, typically structured into an input layer, one or more hidden layers, and an output layer. Each neuron within a layer is connected to each neuron in the layers above and below it, creating a densely interconnected network. Connections between neurons, represented as weights (often denoted as W), are learned during the training process. The output of each neuron is determined by applying an activation function, such as the sigmoid, ReLU, or tanh function, to a weighted sum of its inputs. Mathematically, the output (O) of a neuron in a hidden or output layer is computed as follows:

y=fi=1nwi.xi+b,

where y is the output of the neuron. f is the activation function. wi represents the weight associated with the i-th input connection. xi is the i-th input to the neuron, and b is the bias term. The training process involves adjusting these weights and biases using techniques like backpropagation and gradient descent to minimize a loss function, allowing the MLP to learn complex relationships within the data.

MLPs are versatile and can approximate a wide range of functions, making them a popular choice for various ML applications. Cruz-Monteagudo et al. (2008) investigated computational approaches for predicting idiosyncratic hepatotoxicity using 3D chemical structures such as linear discriminant analysis (LDA) and ANNs. The RBF architecture was used in a neural network classification method that used the same descriptors as those in the LDA model. In the training series, this RBF neural network outperforms the LDA model, achieving an accuracy of 91.07%, sensitivity of 92.00%, and specificity of 90.32%. Examination of the Receiver Operating Characteristic (ROC) curve proved its continuously superior performance.

3.2 Deep neural networks

A Deep Neural Network (DNN) can be mathematically represented as a composition of functions (Schmidhuber, 2015). Given an input vector x, the output y of a DNN with L layers can be expressed as:

y=fLfL1f2f1x

Each layer l applies a linear transformation zl=Wlal1+bl followed by an activation function al=σzl, where Wl is the weight matrix and bl is the bias vector for that layer. The final output is obtained by applying an appropriate activation function at the last layer. Like MLP, DNN can be trained by adjusting the weights and biases to minimize a chosen loss function through techniques like backpropagation and optimization algorithms. DNNs are excellent for automatically extracting key features from large inputs, making them the perfect choice for transcriptomic data containing a wide variety of features.

DNNs (Hinton et al., 2006) have been effectively used to address the challenge of predicting various types of chemically induced liver injuries, including biliary hyperplasia, fibrosis, and necrosis, using DNA microarray data (Feng et al., 2019). Wang et al. (2019) used multi-task DNNs to evaluate gene and pathway-level feature selection strategies for these liver injuries. The DNN models exhibited high predictive accuracy and endpoint specificity, surpassing the performance of Random Forest and SVM models. In another study, Li et al. (2020b), developed a DNN model with eight layers using transcriptome profiles of human cell lines to predict DILI. The model leveraged a substantial binary DILI annotation dataset, achieving AUCs of 0.802 and 0.798 for the training and independent validation sets, respectively. These results outperformed traditional machine learning algorithms, including K-nearest neighbors, SVM, and Random Forest.

In a study conducted by Kang and Kang (2021), a DNN-based model was designed to predict DILI risk. This model used extended connectivity fingerprinting of diameter 4 (ECFP4) to represent molecular substructures. The data for this predictive model was meticulously collected from various sources, including publications like DILIrank and LiverTox. A model was developed through stratified 10-fold cross-validation, and the best DNN model showed an accuracy of 0.731, a sensitivity of 0.714, and a specificity of 0.750 when validated in the complete applicability domain. Jain et al. (2021) used a large-scale acute toxicity dataset encompassing over 80,000 compounds measured against 59 toxicity endpoints. They compared multiple single and multitask models using RF, DNN, CNN, and GNN approaches and found that multitask DL methods performed best.

3.3 Convolutional neural networks

Convolutional Neural Networks (CNNs) are mainly applied in image and speech recognition. These networks are well-suited for capturing spatial hierarchies and local patterns within images. CNNs typically incorporate convolutional layers for feature extraction, pooling layers for dimensionality reduction, and fully connected layers for classification. Architectures like AlexNet, VGG, ResNet, and InceptionNet have consistently demonstrated exceptional performance on various image classification tasks (Krizhevsky et al., 2012; Szegedy et al., 2015; Yamashita et al., 2018; Jain et al., 2021). Mathematically, CNN can be written as follows: Let X be the dataset having m number of images. The input feature size is denoted as n×n. Then the convolution layer is written as:

Aijl=σp=1fq=1fWp,qlAi+p1,j+q1l1+bl
Al=convAl1,Fl

Here, l=1,,L are the number of convolution layers. Next, the CNN has the pooling layer:

Pijl=maxApi,pj,lApi,pj+1l,,Api+p1,pj+p1l
Pl=maxpoolAl,P

Flatten the pooled feature maps to obtain a vector of size F. Now the fully connected layer is defined as:

Zl=WlAl1+bl
Zl=FCAl1,Hl

The output layer has a single neuron for binary classification or multiple neurons for multi-class classification:

Y=σZL

A suitable loss function, such as binary cross-entropy, is used for classification. The network is trained using gradient descent-based optimization to minimize the chosen loss function.

CNN was also used for DILI prediction (Nguyen-Vo et al., 2020; Chen X. et al., 2022). Nguyen-Vo et al. (2020) introduced a novel computational model for the prediction of DILI utilizing CNNs and molecular fingerprints based on 1,597 compounds. The model came up with an average accuracy of 0.89, a Matthews correlation coefficient of 0.80, and an AUC of 0.96.

3.4 Graph Neural Networks

Graph Neural Networks (GNNs) are a class of neural networks explicitly tailored for operating on graph data structures. They are particularly well-suited for tasks involving graphs, such as social network analysis, chemical structure analysis, and computational vision (Zhou et al., 2020). Node-level tasks are used in DILI prediction and chemical structure analysis, and involve predicting the properties or characteristics of individual chemical components, such as molecules, within a graph or network structure.

To learn node representations, GNNs combine information from nearby nodes, effectively capturing intricate relationships in graphs. For example, let G=V,E be the molecular graph, where V is the set of nodes (atoms) and E is the set of edges (bonds). The graph convolutional layer updates node representations based on their neighbors’ features. Let X be the initial node features (molecular fingerprint-embedded features) for all nodes in the graph. The output of the l th graph convolutional layer can be represented as Xl using the following equation:

Xl=σD^12A^D^12XlWl,

Here, A^=A+I is the adjacency matrix of the graph with added self-loops, D^ is the diagonal of matrix A^, Wl is the learnable weights for l th layers, and σ is the activation function. Pooling or aggregation layers were incorporated to combine node features across different neighborhoods. Similar with the CNN architecture, one or more fully-connected layers were used to learn higher-level representations from the aggregated features.

The final layer produces the network’s output. Depending on the task (e.g., regression, classification), the number of neurons and the activation function in the output layer can be adjusted. The forward pass through the GNN can be represented mathematically as written below. Let X0 be the molecular fingerprint-embedded features. The graph convolution is:

Xl=σD^12A^D^12XlWlAggregated Features=Pooling/Aggregation XlD1=Dense Aggregated Features,neurons,activation,,Dk=Dense Dk1,neurons,activationOutput=Dense Dk,output neurons,output activation

Here, k represents the number of dense layers in the network.

GNNs have demonstrated their efficacy in addressing node-level tasks related to DILI predictions (Hwang et al., 2020). Ma et al. (2020) used a MV-GNN based model as a backbone to propose a property augmentation approach to involving more data for four datasets with liver toxicity-relevant properties. The GNN-based approach significantly outperformed existing baselines on DILI datasets, achieving an impressive 81.4% accuracy using cross-validation with random splitting. Lim et al. (2023) introduced a novel technique known as supervised subgraph mining (SSM). SSM effectively identifies explicit subgraph features through iterative optimization of graph transitions. This approach surpasses conventional machine learning methods such as SVM, Random Forest, k-Nearest Neighbors, and deep learning neural networks in DILI classification using two datasets, DILIst and TDC-benchmark. By employing structure-based pattern matching, the proposed approach can also identify subgraph characteristics associated with specific medication groups.

3.5 Recurrent neural networks

Mathematically, a recurrent neural network (RNN) can be represented as follows: at each time step t, the RNN takes an input vector xt, and computes the hidden state ht and the output yt using the following equations:

ht=fWhh*ht1+Whx*xt+bh
yt=gWyh*ht+by

Here, ht represents the hidden state at time t, xt is the input at time t, yt is the output at time t, Whh, Whx, Wyh are weight matrices, bh and by are bias vectors, f and g are activation functions (typically sigmoid or hyperbolic tangent for f and softmax for g). The hidden state ht captures information from previous time steps, allowing RNNs to model temporal dependencies in sequential data.

Xu et al. (2015) employed undirected graph recursive neural networks (UGRNN) to develop DL models for predicting DILI for drugs and small moleculars. Their DL-combined model outperformed ANN and DNN models, achieving an accuracy of 86.9% and an AUC of 0.955 when predicting the DILI of 198 drugs in the external validation set. The model also successfully identified important molecular substructures relevant to DILI, demonstrating the power of DL in this context. In another study Ruiz Puentes et al. (2021), investigators proposed using PharmaNet, a machine learning method that employs RNNs, to search for novel pharmaceutical candidates. PharmaNet was applied to discover ligands for 102 cell receptors and achieved impressive performance with a 97.7% Receiver Operating Characteristic curve-Area Under the Curve (ROC-AUC).

3.6 Generative adversarial network

Generative Adversarial Network (GAN) is an advanced generative model composed of two neural networks: a generator and a discriminator. These networks are trained in opposition to each other. The generator’s objective is to create synthetic data that is virtually indistinguishable from genuine data, while the discriminator’s role is to differentiate between real and generated data. GANs operate through a minimax game where the generator and discriminator compete. As training progresses, the generator becomes increasingly skilled at generating realistic data, while the discriminator becomes better at distinguishing between real and fake data. This dynamic process drives the generator to produce high-quality synthetic data, establishing GANs as a foundational technology in a wide range of applications, such as picture production, style transfer, and data augmentation.

Chen Z. et al. (2022) developed Tox-GAN, which employed deep GANs to generate fresh animal study results without the need for extra tests. They demonstrated its effectiveness by creating transcriptome profiles with remarkable similarity (0.997 ± 0.002 in intensity and 0.740 ± 0.082 in fold change) to real-world data obtained from rat liver toxicogenomic studies. In a related study, Li et al. (2023) introduced the TransOrGAN framework, which aims to map gene expression patterns across multiple rodent organs, sexes, and ages. TransOrGAN generated synthetic transcriptomic profiles with an average cosine similarity of 0.984 compared to their corresponding real profiles. This proof-of-concept study involved 288 samples from nine different organs, showcasing the potential of TransOrGAN to generate realistic transcriptomic data for various research applications.

3.7 Transformers

The field of Natural Language Processing (NLP) has undergone a transformative shift with the introduction of transformer-based models (Kang et al., 2020). These models have enabled the automatic analysis and comprehension of text data in scientific literature. In the domain of DILI studies, NLP models have proven to be valuable tools for extracting insights from textual sources.

Zhan et al. (2022) developed NLP techniques specifically for biomedical texts, allowing the automated processing of 28,000 titles and abstracts retrieved from the PubMed database. By comparing five different text embedding techniques, they found that the model using term frequency-inverse document frequency and logistic regression performed best, with an accuracy of 0.957 on the validation set. Wu Z. et al. (2021) employed a NLP approach based on Bidirectional Encoder Representations from Transformers (BERT) to classify DILI and decipher the meanings of complex text in drug labeling documents. This AI-based model utilized BERT’s power to enhance understanding of text data, particularly in the context of drug safety assessments.

4 Comparison of machine learning and deep learning for DILI prediction

Machine learning and deep learning techniques have emerged as powerful tools for developing models to predict DILI (Table 1). Machine learning uses algorithms to discover patterns and make predictions based on labeled data, whereas deep learning, a subset of machine learning, uses artificial neural networks to replicate the sophisticated functioning of the human brain. Machine learning algorithms analyze a set of predefined features to identify patterns associated with liver injury in the context of DILI prediction, whereas deep learning models can automatically extract intricate features from raw data, providing a more nuanced understanding of complex relationships. The major distinction between the two is in the level of abstraction and data representation (Figure 1). Machine learning is based on feature engineering, in which the algorithm needs to select important features from high-dimensional dataset, whereas deep learning can develop hierarchical representations from raw data, possibly catching subtle nuances that typical machine learning algorithms may overlook. Both approaches provide important contributions to improving our ability to detect and alleviate DILI, giving essential insights for drug development and patient safety.

TABLE 1
www.frontiersin.org

TABLE 1. Comparative analysis of machine learning and deep learning for DILI prediction.

FIGURE 1
www.frontiersin.org

FIGURE 1. The general flowcharts for machine learning and deep learning techniques for developing models to predict DILI.

5 Conclusion

Deep learning approaches have indeed shown significant promise in predicting DILI, leveraging the advantages of large datasets and the ability to capture intricate patterns. In the context of QSAR modeling, DL methods have often been reported to outperform conventional machine learning methods. However, it is essential to recognize that DL’s superiority is not always guaranteed and can depend on the specific characteristics of the dataset and the problem. For instance, Liu et al. (2018) pointed out that global performance metrics, which typically show DNNs as superior to conventional machine learning, may not be appropriate for datasets with highly imbalanced sample distributions. They argued that for highly toxic chemicals, DNNs trained on all samples often perform worse than indicated by global performance metrics.

Imbalanced datasets can lead to misrepresentations of the actual performance, especially in cases where the minority class (highly toxic chemicals, in this example) is of particular interest. Similarly, Russo et al. (2018) compared DNN with conventional machine learning algorithms, including Naive Bayes, AdaBoost Decision Tree, Random Forest, and SVM, in the development of QSAR models for predicting endocrine disrupting endpoints using up to 7,500 compounds. Their results revealed that while DNNs may achieve higher accuracy on the training set, they did not consistently outperform classic machine learning methods in 5-fold cross-validation and predictions on external test sets. The performance of machine learning models can be influenced by various factors, including the nature of the data, the choice of molecular descriptors, and the specific problem being addressed.

Deep learning has specific characteristics for toxicity prediction. Scalability is a primary one, since DL models can handle vast amounts of data and understand nuanced correlations, enabling the discovery of small DILI risk variables that older approaches may overlook. Furthermore, by collecting latent characteristics across varied datasets, these models can accomplish impressive generalization, boosting the capacity to predict DILI across different chemicals and patient groups. However, interpretability is a key weakness of DL in this scenario. Because the models are intrinsically complex, deciphering the precise biological or chemical elements leading to DILI forecasts is difficult, limiting one’s capacity to grasp the underlying processes. Additionally, DL also requires a large amount of high-quality data for training, and like machine learning, is also prone to overfitting when the training data is noisy or when the model is too complex.

Researchers and practitioners in this field must carefully consider these advantages and challenges when choosing and implementing DL approaches for toxicity prediction. Balancing the needs for accuracy and interpretability is crucial in improving our understanding and prediction of DILI and other toxicities.

Author contributions

FM: Data curation, Formal Analysis, Methodology, Writing–original draft. MC: Conceptualization, Funding acquisition, Methodology, Supervision, Validation, Writing–review and editing.

Funding

The authors declare that no financial support was received for the research, authorship, and/or publication of this article.

Acknowledgments

The authors thank Joanne Berger, FDA Library, for manuscript editing assistance.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

The authors declared that they were an editorial board member of Frontiers, at the time of submission. This had no impact on the peer review process and the final decision.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Author disclaimer

This article reflects the views of the authors and does not necessarily reflect those of the U.S. Food and Drug Administration. Any mention of commercial products is for clarification only and is not intended as approval, endorsement, or recommendation.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/ftox.2023.1340860/full#supplementary-material

References

Ai, H., Wu, X., Zhang, L., Qi, M., Zhao, Y., Zhao, Q., et al. (2019). QSAR modelling study of the bioconcentration factor and toxicity of organic compounds to aquatic organisms using machine learning and ensemble methods. Ecotoxicol. Environ. Saf. 179, 71–78. doi:10.1016/j.ecoenv.2019.04.035

PubMed Abstract | CrossRef Full Text | Google Scholar

Alempijevic, T., Zec, S., and Milosavljevic, T. (2017). Drug-induced liver injury: do we know everything? World J. hepatology 9 (10), 491–502. doi:10.4254/wjh.v9.i10.491

PubMed Abstract | CrossRef Full Text | Google Scholar

Andrade, R. J., Chalasani, N., Björnsson, E. S., Suzuki, A., Kullak-Ublick, G. A., Watkins, P. B., et al. (2019). Drug-induced liver injury. Nat. Rev. Dis. Prim. 5 (1), 58. doi:10.1038/s41572-019-0105-0

PubMed Abstract | CrossRef Full Text | Google Scholar

Chen, M., Borlak, J., and Tong, W. (2014). Predicting idiosyncratic drug-induced liver injury–some recent advances. Expert Rev. Gastroenterology Hepatology 8 (7), 721–723. doi:10.1586/17474124.2014.922871

PubMed Abstract | CrossRef Full Text | Google Scholar

Chen, X., Roberts, R., Tong, W., and Liu, Z. (2022). Tox-GAN: an artificial intelligence approach alternative to animal studies—a case study with toxicogenomics. Toxicol. Sci. 186 (2), 242–259. doi:10.1093/toxsci/kfab157

PubMed Abstract | CrossRef Full Text | Google Scholar

Chen, Z., Jiang, Y., Zhang, X., Zheng, R., Qiu, R., Sun, Y., et al. (2022). ResNet18DNN: prediction approach of drug-induced liver injury by deep neural network with ResNet18. Briefings Bioinforma. 23 (1), bbab503. doi:10.1093/bib/bbab503

CrossRef Full Text | Google Scholar

Cruz-Monteagudo, M., Cordeiro, M. N. D., and Borges, F. (2008). Computational chemistry approach for the early detection of drug-induced idiosyncratic liver toxicity. J. Comput. Chem. 29 (4), 533–549. doi:10.1002/jcc.20812

PubMed Abstract | CrossRef Full Text | Google Scholar

Devarbhavi, H. (2012). An update on drug-induced liver injury. J. Clin. Exp. hepatology 2 (3), 247–259. doi:10.1016/j.jceh.2012.05.002

CrossRef Full Text | Google Scholar

Feng, C., Chen, H., Yuan, X., Sun, M., Chu, K., Liu, H., et al. (2019). Gene expression data based deep learning model for accurate prediction of drug-induced liver injury in advance. J. Chem. Inf. Model. 59 (7), 3240–3250. doi:10.1021/acs.jcim.9b00143

PubMed Abstract | CrossRef Full Text | Google Scholar

Gadaleta, D., Manganelli, S., Roncaglioni, A., Toma, C., Benfenati, E., and Mombelli, E. (2018). QSAR modeling of ToxCast assays relevant to the molecular initiating events of AOPs leading to hepatic steatosis. J. Chem. Inf. Model. 58 (8), 1501–1517. doi:10.1021/acs.jcim.8b00297

PubMed Abstract | CrossRef Full Text | Google Scholar

George, N., Chen, M., Yuen, N., Hunt, C. M., and Suzuki, A. (2018). Interplay of gender, age and drug properties on reporting frequency of drug-induced liver injury. Regul. Toxicol. Pharmacol. 94, 101–107. doi:10.1016/j.yrtph.2018.01.018

PubMed Abstract | CrossRef Full Text | Google Scholar

Goh, G. B., Siegel, C., Vishnu, A., Hodas, N., and Baker, N. (2018). “How much chemistry does a deep neural network need to know to make accurate predictions?,” in 2018 IEEE Winter Conference on Applications of Computer Vision (WACV) (IEEE).

CrossRef Full Text | Google Scholar

Hastie, T., Tibshirani, R., Friedman, J. H., and Friedman, J. H. (2009). The elements of statistical learning: data mining, inference, and prediction. Springer.

Google Scholar

Hemmerich, J., Asilar, E., and Ecker, G. F. (2020). COVER: conformational oversampling as data augmentation for molecules. J. cheminformatics 12 (1), 18. doi:10.1186/s13321-020-00420-z

CrossRef Full Text | Google Scholar

Hinton, G. E., Osindero, S., and Teh, Y.-W. (2006). A fast learning algorithm for deep belief nets. Neural Comput. 18 (7), 1527–1554. doi:10.1162/neco.2006.18.7.1527

PubMed Abstract | CrossRef Full Text | Google Scholar

Hwang, D., Jeon, M., and Kang, J. (2020). “A drug-induced liver injury prediction model using transcriptional response data with graph neural network,” in 2020 IEEE International Conference on Big Data and Smart Computing (BigComp) (IEEE).

CrossRef Full Text | Google Scholar

Idakwo, G., Luttrell IV, J., Chen, M., Hong, H., Gong, P., and Zhang, C. (2019). A review of feature reduction methods for QSAR-based toxicity prediction. Springer.

Google Scholar

Jain, S., Siramshetty, V. B., Alves, V. M., Muratov, E. N., Kleinstreuer, N., Tropsha, A., et al. (2021). Large-scale modeling of multispecies acute toxicity end points using consensus of multitask deep learning methods. J. Chem. Inf. Model. 61 (2), 653–663. doi:10.1021/acs.jcim.0c01164

PubMed Abstract | CrossRef Full Text | Google Scholar

Kang, M.-G., and Kang, N. S. (2021). Predictive model for drug-induced liver injury using deep neural networks based on substructure space. Molecules 26 (24), 7548. doi:10.3390/molecules26247548

PubMed Abstract | CrossRef Full Text | Google Scholar

Kang, Y., Cai, Z., Tan, C.-W., Huang, Q., and Liu, H. (2020). Natural language processing (NLP) in management research: a literature review. J. Manag. Anal. 7 (2), 139–172. doi:10.1080/23270012.2020.1756939

CrossRef Full Text | Google Scholar

Krizhevsky, A., Sutskever, I., and Hinton, G. E. (2012). Imagenet classification with deep convolutional neural networks, Adv. neural Inf. Process. Syst., 25. doi:10.1145/3065386

CrossRef Full Text | Google Scholar

Li, T., Roberts, R., Liu, Z., and Tong, W. (2023). TransOrGAN: an artificial intelligence mapping of rat transcriptomic profiles between organs, ages, and sexes. Chem. Res. Toxicol. 36, 916–925. doi:10.1021/acs.chemrestox.3c00037

PubMed Abstract | CrossRef Full Text | Google Scholar

Li, T., Tong, W., Roberts, R., Liu, Z., and Thakkar, S. (2020a). Deep learning on high-throughput transcriptomics to predict drug-induced liver injury. Front. Bioeng. Biotechnol. 8, 562677. doi:10.3389/fbioe.2020.562677

PubMed Abstract | CrossRef Full Text | Google Scholar

Li, T., Tong, W., Roberts, R., Liu, Z., and Thakkar, S. (2020b). DeepDILI: deep learning-powered drug-induced liver injury prediction using model-level representation. Chem. Res. Toxicol. 34 (2), 550–565. doi:10.1021/acs.chemrestox.0c00374

PubMed Abstract | CrossRef Full Text | Google Scholar

Lim, S., Kim, Y., Gu, J., Lee, S., Shin, W., and Kim, S. (2023). Supervised chemical graph mining improves drug-induced liver injury prediction. iScience 26 (1), 105677. doi:10.1016/j.isci.2022.105677

PubMed Abstract | CrossRef Full Text | Google Scholar

Liu, R., Madore, M., Glover, K. P., Feasel, M. G., and Wallqvist, A. (2018). Assessing deep and shallow learning methods for quantitative prediction of acute chemical toxicity. Toxicol. Sci. 164 (2), 512–526. doi:10.1093/toxsci/kfy111

PubMed Abstract | CrossRef Full Text | Google Scholar

Ma, H., An, W., Wang, Y., Sun, H., Huang, R., and Huang, J. (2020). Deep graph learning with property augmentation for predicting drug-induced liver injury. Chem. Res. Toxicol. 34 (2), 495–506. doi:10.1021/acs.chemrestox.0c00322

PubMed Abstract | CrossRef Full Text | Google Scholar

Mora, J. R., Marrero-Ponce, Y., García-Jacas, C. R., and Suarez Causado, A. (2020). Ensemble models based on QuBiLS-MAS features and shallow learning for the prediction of drug-induced liver toxicity: improving deep learning and traditional approaches. Chem. Res. Toxicol. 33 (7), 1855–1873. doi:10.1021/acs.chemrestox.0c00030

PubMed Abstract | CrossRef Full Text | Google Scholar

Nguyen-Vo, T.-H., Nguyen, L., Do, N., Le, P. H., Nguyen, T.-N., Nguyen, B. P., et al. (2020). Predicting drug-induced liver injury using convolutional neural network and molecular fingerprint-embedded features. ACS omega 5 (39), 25432–25439. doi:10.1021/acsomega.0c03866

PubMed Abstract | CrossRef Full Text | Google Scholar

Rao, M., Nassiri, V., Alhambra, C., Snoeys, J., Van Goethem, F., Irrechukwu, O., et al. (2023). AI/ML models to predict the severity of drug-induced liver injury for small molecules. Chem. Res. Toxicol. 36, 1129–1139. doi:10.1021/acs.chemrestox.3c00098

PubMed Abstract | CrossRef Full Text | Google Scholar

Rish, I. (2001). “An empirical study of the naive Bayes classifier,” in IJCAI 2001 workshop on empirical methods in artificial intelligence.

Google Scholar

Ruiz Puentes, P., Valderrama, N., González, C., Daza, L., Muñoz-Camargo, C., Cruz, J. C., et al. (2021). PharmaNet: pharmaceutical discovery with deep recurrent neural networks. Plos one 16 (4), e0241728. doi:10.1371/journal.pone.0241728

PubMed Abstract | CrossRef Full Text | Google Scholar

Russo, D. P., Zorn, K. M., Clark, A. M., Zhu, H., and Ekins, S. (2018). Comparing multiple machine learning algorithms and metrics for estrogen receptor binding prediction. Mol. Pharm. 15 (10), 4361–4370. doi:10.1021/acs.molpharmaceut.8b00546

PubMed Abstract | CrossRef Full Text | Google Scholar

Schmidhuber, J. (2015). Deep learning in neural networks: an overview. Neural Netw. 61, 85–117. doi:10.1016/j.neunet.2014.09.003

PubMed Abstract | CrossRef Full Text | Google Scholar

Shin, H. K., Huang, R., and Chen, M. (2023). In silico modeling-based new alternative methods to predict drug and herb-induced liver injury: a review. Food Chem. Toxicol. 179, 113948. doi:10.1016/j.fct.2023.113948

PubMed Abstract | CrossRef Full Text | Google Scholar

Szegedy, C., Liu, W., Jia, Y., Sermanet, P., Reed, S., Anguelov, D., et al. (2015). “Going deeper with convolutions,” in proceedings of the IEEE computer society conference on computer vision and pattern recognition.

CrossRef Full Text | Google Scholar

Tang, W., Chen, J., and Hong, H. (2020). Discriminant models on mitochondrial toxicity improved by consensus modeling and resolving imbalance in training. Chemosphere 253, 126768. doi:10.1016/j.chemosphere.2020.126768

PubMed Abstract | CrossRef Full Text | Google Scholar

Wang, H., Liu, R., Schyman, P., and Wallqvist, A. (2019). Deep neural network models for predicting chemically induced liver toxicity endpoints from transcriptomic responses. Front. Pharmacol. 10, 42. doi:10.3389/fphar.2019.00042

PubMed Abstract | CrossRef Full Text | Google Scholar

Weber, S., and Gerbes, A. L. (2022). Challenges and future of drug-induced liver injury research—laboratory tests. Int. J. Mol. Sci. 23 (11), 6049. doi:10.3390/ijms23116049

PubMed Abstract | CrossRef Full Text | Google Scholar

Williams, D. P., Lazic, S. E., Foster, A. J., Semenova, E., and Morgan, P. (2019). Predicting drug-induced liver injury with Bayesian machine learning. Chem. Res. Toxicol. 33 (1), 239–248. doi:10.1021/acs.chemrestox.9b00264

PubMed Abstract | CrossRef Full Text | Google Scholar

Wu, L., Liu, Z., Auerbach, S., Huang, R., Chen, M., McEuen, K., et al. (2017). Integrating drug’s mode of action into quantitative structure–activity relationships for improved prediction of drug-induced liver injury. J. Chem. Inf. Model. 57 (4), 1000–1006. doi:10.1021/acs.jcim.6b00719

PubMed Abstract | CrossRef Full Text | Google Scholar

Wu, Y., Liu, Z., Wu, L., Chen, M., and Tong, W. (2021). BERT-based Natural Language Processing of drug labeling documents: a case study for classifying drug-induced liver injury risk. Front. Artif. Intell. 4, 729834. doi:10.3389/frai.2021.729834

PubMed Abstract | CrossRef Full Text | Google Scholar

Wu, Y., Xiao, W., Tong, W., Borlak, J., and Chen, M. (2022). A systematic comparison of hepatobiliary adverse drug reactions in FDA and EMA drug labeling reveals discrepancies. Drug Discov. Today 27 (1), 337–346. doi:10.1016/j.drudis.2021.09.009

PubMed Abstract | CrossRef Full Text | Google Scholar

Wu, Z., Zhu, M., Kang, Y., Leung, E. L.-H., Lei, T., Shen, C., et al. (2021). Do we need different machine learning algorithms for QSAR modeling? A comprehensive assessment of 16 machine learning algorithms on 14 QSAR data sets. Briefings Bioinforma. 22 (4), bbaa321. doi:10.1093/bib/bbaa321

CrossRef Full Text | Google Scholar

Xu, Y., Dai, Z., Chen, F., Gao, S., Pei, J., and Lai, L. (2015). Deep learning for drug-induced liver injury. J. Chem. Inf. Model. 55 (10), 2085–2093. doi:10.1021/acs.jcim.5b00238

PubMed Abstract | CrossRef Full Text | Google Scholar

Yamashita, R., Nishio, M., Do, R. K. G., and Togashi, K. (2018). Convolutional neural networks: an overview and application in radiology. Insights into imaging 9, 611–629. doi:10.1007/s13244-018-0639-9

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhan, X., Wang, F., and Gevaert, O. (2022). Reliably filter drug-induced liver injury literature with Natural Language Processing and conformal prediction. IEEE J. Biomed. Health Inf. 26 (10), 5033–5041. doi:10.1109/JBHI.2022.3193365

CrossRef Full Text | Google Scholar

Zhang, H., Ding, L., Zou, Y., Hu, S.-Q., Huang, H.-G., Kong, W.-B., et al. (2016). Predicting drug-induced liver injury in human with Naïve Bayes classifier approach. J. computer-aided Mol. Des. 30, 889–898. doi:10.1007/s10822-016-9972-6

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhou, J., Cui, G., Hu, S., Zhang, Z., Yang, C., Liu, Z., et al. (2020). Graph neural networks: a review of methods and applications. AI open 1, 57–81. doi:10.1016/j.aiopen.2021.01.001

CrossRef Full Text | Google Scholar

Zhu, J., Seo, J.-E., Wang, S., Ashby, K., Ballard, R., Yu, D., et al. (2018). The development of a database for herbal and dietary supplement induced liver toxicity. Int. J. Mol. Sci. 19 (10), 2955. doi:10.3390/ijms19102955

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: drug-induced liver injury (DILI), machine learning, deep learning, drug safety, predictive model

Citation: Mostafa F and Chen M (2024) Computational models for predicting liver toxicity in the deep learning era. Front. Toxicol. 5:1340860. doi: 10.3389/ftox.2023.1340860

Received: 19 November 2023; Accepted: 22 December 2023;
Published: 19 January 2024.

Edited by:

Chao Ji, Indiana University, United States

Reviewed by:

Sijie Lin, Tongji University, China

Copyright © 2024 Mostafa and Chen. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Minjun Chen, bWluanVuLmNoZW5AZmRhLmhocy5nb3Y=

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.