Skip to main content

ORIGINAL RESEARCH article

Front. Psychiatry, 12 January 2024
Sec. Psychological Therapy and Psychosomatics

Integrating exploration and prediction in computational psychotherapy science: proof of concept

  • 1Department of Psychology, University of Haifa, Haifa, Israel
  • 2Department of Psychology, Ohio University, Athens, OH, United States
  • 3Integrative Pain Laboratory (iPainLab), School of Public Health, University of Haifa, Haifa, Israel

Introduction: Psychotherapy research has long preferred explanatory over predictive models. As a result, psychotherapy research is currently limited in the variability that can be accounted for in the process and outcome of treatment. The present study is a proof-of-concept approach to psychotherapy science that uses a datadriven approach to achieve robust predictions of the process and outcome of treatment.

Methods: A trial including 65 therapeutic dyads was designed to enable an adequate level of variability in therapist characteristics, overcoming the common problem of restricted range. A mixed-model, data-driven approach with cross-validation machine learning algorithms was used to predict treatment outcome and alliance (within- and between-clients; client- and therapist-rated alliance).

Results and discussion: Based on baseline predictors only, the models explained 52.8% of the variance for out-of-sample prediction in treatment outcome, and 24.1–52.8% in therapeutic alliance. The identified predictors were consistent with previous findings and point to directions for future investigation. Although limited by its sample size, this study serves as proof of the great potential of the presented approach to produce robust predictions regarding the process and outcome of treatment, offering a potential solution to problems such as p-hacking and lack of replicability. Findings should be replicated using larger samples and distinct populations and settings.

1 Introduction

Traditionally, statistical goals can be divided into two categories: explanation-oriented models, in which researchers are interested in testing an a priori hypothesized relationship between two or more variables, and prediction-oriented models, in which researchers are interested in finding an algorithm capable of recognizing which set of variables yields the best predictions about new observations (1, 2). For many years, psychotherapy research focused mainly on models aimed at understanding and explaining associations between variables (3). Many hundreds of studies testing predictors, moderators, and mediators have contributed greatly to our understanding of the process and outcome of psychotherapy (2). For instance, research has shown that the therapeutic alliance is strongly related to improved treatment outcomes, with a more pronounced effect observed in clients dealing with interpersonal problems (4).

Contemporary psychotherapy research is interested not only in driving understanding, but also in making accurate predictions about who may be a good therapist (to direct therapist selection and training efforts), and about treatment prognosis (to select the most efficient treatment for individual clients). Despite the statistical and mathematical similarities, the choice between explanatory models and predictive models can have a significant impact on researchers’ decisions regarding data collection, data preparation, and the statistical models they employ (1). For example, while explanatory models tend to emphasize goodness of fit based on a specific sample and the testing of statistical significance, predictive models assess how accurately the model can apply information about one sample to make correct predictions or decisions when applied to a new dataset (5). Researchers, therefore, must make a deliberate choice to identify the best models for achieving their aims (1, 6).

When following a hypothesis-driven approach, researchers rarely use tools to verify that the models they propose are capable of predicting the outcomes they are modeling (7). Instead, researchers focus largely on the statistical significance of various factors that may be related to the outcome (3). The statistical significance approach does not guarantee predictive accuracy when models are applied to new (i.e., out-of-sample) data. Furthermore, in the past years, there is an increasing concern that many findings in the field cannot be reliably reproduced in subsequent studies, raising questions about the credibility and robustness of psychological research findings (i.e., the replication crisis) (811). There is increasing consensus that this widespread replication failure is due largely to “p-hacking” and other questionable research practices that have been historically prevalent in the field (12). At the same time, the increased capacity to collect and analyze massive amounts of data, gathered through new technologies or the internet, has contributed to the adoption of a computational approach to analysis (13). While a hypothesis-driven approach is rooted in, and therefore constrained by, substantive theory, a computational approach prioritizes prediction, thus allowing model complexity to increase as long as it continues to enhance predictive performance (7). Inspired by the computational approach, researchers started to use predictability models by applying data-driven approaches and principles from the machine learning field to increase the predictive power of the models (7). Machine learning is a subset of artificial intelligence that focuses on developing algorithms and models capable of learning from data and making predictions or decisions without being explicitly programmed. It involves the use of statistical techniques to enable computers to recognize patterns, draw insights, and improve their performance based on experience (5). In this regard, the objectives of machine learning closely parallel those of psychotherapists. Both aim to accumulate knowledge from prior data, such as client histories, and apply that knowledge to new cases, even those that may be unique (14). Machine learning encompasses various approaches, including supervised learning (where models are trained on labeled data), unsupervised learning (for discovering patterns within data), and reinforcement learning (teaching machines to make sequential decisions). These algorithms have applications across a wide range of fields including healthcare, revolutionizing the way we analyze data and solve complex problems (15).

Some promising evidence for the potential utility of implementations of data-driven and machine learning approaches in psychotherapy research has emerged in recent years, answering questions of treatment personalization (1618) while providing critical solutions to p-hacking (19) and replicability issues (10). Moreover, data-driven approaches that use principles from the machine learning field may be better able to capture the richness and complexity of the therapeutic encounter than explanatory models. No single, independent predictor, or even three or four, can be as informative as a set of interrelated variables that jointly increase predictive power to explain variance in the process and outcome of treatment (16). Some of the explanatory approaches, which test each predictor as a separate hypothesis, can lead to erroneous conclusions because of multiple comparisons (inflated type I errors), model misspecification, and multicollinearity (20). Findings may also be affected by publication bias, as statistically significant predictors have a better chance of being reported in the literature (21). Machine learning approaches provide a robust solution to these problems by looking at data insights, developing a predictive model, and finally, applying self-validation procedures (22, 23). However, a significant drawback of some (but not all) machine learning approaches is their tendency to treat models mostly as a “black box,” resulting in a lack of interpretability of the findings (3, 7).

We sought to integrate the benefits of the interpretability and predictability of statistical models. As such, the present study was designed as a proof of concept of the ability of models to predict variability in treatment process and outcome, based on pre-treatment client and therapist characteristics. Our study focused on changes in psychological dysfunction from pre- to post-treatment as the treatment outcome. We chose the therapeutic alliance as an example of a process variable because it is one of the most researched constructs in psychotherapy research, and one of the most consistent predictors of treatment outcome (24). Two important aspects of the alliance are of interest when predicting clients’ and therapists’ reported alliance: (a) the trait-like tendency of the client and therapist to form a strong alliance across various phases of treatment (also known as the between-clients alliance), and (b) the state-like changes occurring during treatment in client and therapist reports of the alliance (the within-client alliance). The sample was chosen to avoid restriction of range, a substantial risk when testing therapist characteristics because in many trials only a narrow range of therapist characteristics is allowed. As the therapist effect may have been skewed in previous studies owing to restriction of range, to increase variance, we selected therapists prospectively, based on their level of social skills and performance on a simulated task. For the same reason, we also selected both trained and untrained therapists. Using prediction models in a sample that is less restricted in range, we expect the baseline characteristics of clients and therapists to predict a large portion of treatment outcome and within- and between-clients alliance.

For client characteristics that may affect the alliance and the outcome of treatment, we focused on those that have received extensive attention in the literature. Specifically, we focused on clients’ demographics (gender, age, and income), symptom severity, and functioning. We also focused on clients’ interpersonal skills and problems [for a review, (see 25, 26)]. Whereas previous studies showed little evidence that demographic variables significantly predict psychotherapy outcome or alliance, more complex results were found concerning symptom and interpersonal problem severity (25). Severity of symptoms and functional impairment predicted worse outcome and lower trait-like alliance [(e.g., 2730)]. At the same time, severity of symptoms also predicted greater improvement of both symptoms and state-like alliance [(e.g., 3133)]. The same pattern of results was found for interpersonal skills and problems [(e.g., 3436)].

With regard to therapists, in addition to including therapist training and clinical orientation as potential predictors of treatment process and outcome, we focused on therapist characteristics that are typically easily available in clinical practice, including demographics (age, gender, and income), and on those receiving theoretical and empirical support, such as previously being in treatment themselves (37), and Facilitative Interpersonal Skills [FIS; (38)]. Previous studies examining the association between such therapists’ characteristics and alliance or outcome have yielded mixed results. Specifically, whereas previous studies mostly found that training and experience were not related to outcome and alliance [(e.g., 3941)], other studies found them to be related to poorer outcome and lower alliance (4244). Manne et al. (45) reported that therapists’ experience negatively predicted mean alliance when alliance was rated by the client, but positively predicted alliance when alliance was rated by the therapist. These findings may suggest that a complex predictive model is needed to better explain the mixed results.

In the same vein, most studies examining the relationship between the therapist’s demographic variables and alliance or treatment outcome have found that these variables are not significantly related to therapy outcome or alliance [Outcome: (37, 46, 47); Alliance: (42, 45)]. At the same time, other studies suggested that therapists being older or a woman may be related to outcome and alliance, presumably because of their relation to superior interpersonal and social skills (48). For instance, results of a large naturalistic study suggested that gender may have an indirect role on alliance and outcome such that female therapists intervene more empathically, whereas male therapists tend to use more confrontational techniques (49). Similarly, Anderson et al. (38) found that older therapists produced superior outcomes. However, when therapists’ social skills and FIS were examined, age no longer predicted outcome. These results may point to the importance of including both demographic characteristics and interpersonal and social skills in the predictive model for better prediction of alliance and outcome. One way to measure therapists’ interpersonal and social skills is using the FIS. FIS includes common relational skills, such as the ability to respond empathetically to clients, express the appropriate amount and type of emotion when responding to clients, and efficiently respond to ruptures that may arise in the alliance with clients. FIS is measured using a performance task in which therapists record audio responses to a set of stimulus clips simulating various difficult interpersonal situations derived from real therapy sessions. Each therapist’s responses are then rated by trained observers for their coverage of the eight FIS items: verbal fluency, emotional expression, persuasiveness, warmth/acceptance/understanding, hope/positive expectations, empathy, alliance bond capacity, and alliance-rupture-repair responsiveness (38, 50). Therapist FIS has been found to predict treatment outcome across settings (38, 51, 52). Anderson et al. (28) found that therapist FIS predicted outcomes at a university counseling center. In a randomized clinical trial, Anderson et al. (51) found that helpers with significantly higher FIS showed better outcomes and developed stronger client-related alliances when counseling undergraduate students with heightened levels of clinical distress. In a naturalistic study, Anderson et al. (42) found that therapist FIS predicted the outcomes of clients treated by graduate students in clinical psychology over 1 year later.

In addition to FIS’s prediction of outcome in several studies, multiple metanalyses have revealed several other predictors of outcome [(e.g., 2)]. For example, a previous review by Beutler (53) identified multiple therapist characteristics, such as their training and experience, which give promise to predicting outcome. Nevertheless, at that point, these characteristics were impossible to consider jointly. It is reasonable to include these client and therapist characteristics in a model which allows them to combine with FIS. Furthermore, the notion of therapist responsiveness (54) recognizes that there are a very large number of possible variables within the “emergent context” (55) that might be accounted for to explain what works in any particular therapeutic situation.

Taken together, in the present study, we aim to benefit from the advantages of the two potentially complementary approaches, the interpretability and predictability approaches. We will do so by applying a data-driven approach and principles from the multilevel machine learning framework. Our approach allows: (a) identifying the models with the best fit to the data (given a set of potential predictors), considering the linear and quadratic shapes of the relationships as well as pairwise interactions, and taking into account their nested data structure; (b) providing an exact interpretation of all model effects; (c) cross-validating the proposed model; and (d) estimating the contribution of each model effect, as defined below.

2 Methods

2.1 Participants

The clients were 65 individuals who showed significant distress and met criteria for a DSM-IV psychological disorder, selected to participate in the study from 2,713 undergraduates. The participants were selected on the basis of significantly high scores (i.e., at least two standard deviations above the mean) across two administrations of a general symptom measure as well as a brief assessment interview. For more information about client selection procedure, see Anderson et al. (51). For ethical considerations, given the diversity of training and experience of the individuals playing the role of therapists in this trial, none of the participants who were recruited for the study were actively seeking clinical services at the time. Of the 65 individuals who completed the study as clients, 64 provided demographic data. Thirty-five clients (54.7%) identified as female, while 29 (45.3%) identified as male. Most of the sample was White (n = 60, 93.8%). Other clients self-identified their race/ethnicity as Asian/Pacific Islander (n = 1, 1.6%), Black (n = 1, 1.6%), and Hispanic (n = 2, 3.1%). The average client age was 19.2 (SD = 1.1). Diagnostic groupings were as follows: adjustment disorder (n = 12, 18.8%), dysthymia (n = 15, 23.4%), generalized anxiety (n = 14, 21.9%), major depression (n = 10, 15.6%), personality (n = 5, 7.8%), and miscellaneous (e.g., phobia, panic, eating; n = 8, 12.5%).

The therapists were 23 (eight male, 15 female) doctoral students, selected from a larger group of 56 applicants. Therapists were selected with the intention of creating independent groups on the basis of their interpersonal skills (i.e., low vs. high) and training status (i.e., trained vs. untrained). Low or high interpersonal skills was defined by scores on a self-report measure of social skills. The individuals that were selected as therapists for this study were taken from the highest and lowest quarter of performances on a social skills measure. In other words, those who deviated most from the (gender-specific) mean. For more information about skill selection procedure, see Anderson et al. (51). For training status, 11 therapists were in a clinical psychology doctoral program and had completed at least 2 years of training; these were considered the “trained” group. The other 12 therapists, constituting the “untrained” group, had no clinical or psychotherapeutic training but had completed at least 2 years in doctoral programs in various other disciplines. Therapists’ ages ranged from 23 to 53 years (mean = 30.61 years; SD = 9.32). Therapists self-identified as 83% Caucasian, 13% Asian, and 4% Hispanic. Each therapist worked with two clients.

2.2 Measures

2.2.1 Outcome questionnaire-45

The Outcome Questionnaire-45 (OQ-45) is a 45-item general symptom measure that was completed by clients (56). The items assess the following three primary dimensions: (a) subjective discomfort (e.g., anxiety and depression), (b) interpersonal relationships, and (c) social role performance. Each item is rated on a five-point Likert scale, ranging from 0 (never) to 4 (almost always). The sum of the items (after reverse-coding selected items) forms the total OQ-45 score, which was used in the current study. The measure has demonstrated good internal consistency in prior studies [αs ranging from 0.70 to 0.93; (57)], as well as in the current study (α = 0.96). In case of missing data, we followed the standard scoring rules, according to which only forms with fewer than four missing items were included in the following analysis.

2.2.2 Inventory of interpersonal problems

The Inventory of interpersonal problems (IIP-64) is a measure of interpersonal distress commonly used for measuring treatment changes in the interpersonal domain (58). The degree of distress associated with each item is rated on a five-point scale, ranging from 1 (not at all) to 5 (extremely). Test–retest reliability for the IIP for a 10-week period has been reported at 0.98 for the overall inventory, with internal consistency ranging from 0.82 to 0.93 (59). The present sample had good internal consistency as well (α = 0.93 at pretreatment and termination and α = 0.95 at 3-month follow-up).

2.2.3 Social skills inventory

The social skills inventory (SSI) is a 90-item self-report questionnaire that assessed self-reported social skills (60). Items were scored using five-point Likert scaling, from 1 = “not at all like me” to 5 = “exactly like me.” The SSI measures skills in expressivity, sensitivity, and control in verbal (social) and non-verbal (emotional) domains. The total of the items provided an overall indicator of social skills, which was used in this study. The scale has high internal consistency, and factor analytic studies have supported the multidimensional structure of the scale. Coefficient alphas range from 0.75 to 0.88. Test–retest correlations range from 0.81 to 0.96 for a 2-week interval, and alpha coefficients range from 0.62 to 0.87 (60). Convergent and discriminant validity for the SSI were supported in a series of studies conducted by Riggio (60). In the present study, the SSI was completed by both therapists (as a selection variable) and clients in the study (at pre-treatment). The SSI had good internal consistency in the present sample (α = 0.88).

2.2.4 Facilitative interpersonal skills

Facilitative interpersonal skills (FIS) is an observational rating of audio responses provided by therapists to difficult simulated clients (38, 51). There are eight items on the rating scale, all of which pertain to the therapist’s skill in fostering facilitative conditions. These eight skill domains are verbal fluency, emotional expression, persuasiveness, warmth/positive regard, hopefulness, empathy, alliance bond capacity, and alliance-rupture-repair responsiveness. Each of these domains was rated on a five-point Likert scale, ranging from 1 (skill deficit) to 5 (optimal presence of the skill). All ratings were initially anchored at 3 and were moved up or down the scale based on evidence of skills found in the audio responses. To increase reliability of these ratings, a coding manual was used that provided descriptions for each of the skills. The FIS ratings were made by four coders, which included one doctoral-level researcher (Caucasian male), two graduate students (Chinese female and Caucasian male), and one undergraduate student (Caucasian female). Instruction in the FIS rating method occurred weekly over a 2-month period. Once there appeared to be sufficient agreement, ratings for the study commenced. The prospective therapist responses were rated in random order and in sets of 10. Each coder made their ratings separately and independently. Calibration meetings were held after each set of ratings, where discussion focused on those ratings that were most discrepant (i.e., typically greater than 1 point discrepancy). Final ratings for analysis in the study were a mean of all eight items, which were averaged across the four coders. Interrater reliability was acceptable for total FIS (intraclass correlation coefficient = 0.86), and the internal consistency of the eight FIS items was high (α = 0.96).

2.2.5 Working alliance inventory (WAI-C and WAI-T)

The WAI is a 36-item measure of the quality of the therapeutic alliance (61). It contains subscales for measuring agreement on tasks, goals, and the existence of a therapeutic bond. Each subscale contains 12 items, which the participant rated on a seven-point Likert type scale, ranging from 1 (never) to 7 (always). Across all sessions, the WAI-C, administered to clients, had good internal consistency with alpha ranging from a low of 0.79 to a high of 0.90, and the WAI-T, administered to therapists, ranging from α = 0.80 to 0.81.

2.3 Procedure

Study procedures are described in a previous report (51). Selected therapists were randomly assigned to two clients each. After describing the study to the clients, written informed consent was obtained. Clients were given the opportunity to receive a referral for treatment elsewhere and were also notified that they may discontinue their involvement in the study at any time. Clients were told that they would meet with a “helper” for sessions, who would try to aid them with their problems. At the first session, clients received brief instructions to discuss their problems with the “helper.” Treatments lasted 7 weeks, one session per week. No clients chose to discontinue their involvement in the study at any time.

Treatment outcome was measured pre-treatment, at sessions 1, 3, 5, and 7, and post-termination, using the OQ-45 (56). Alliance was measured at sessions 1, 3, 5, and 7, using the WAI (61). Baseline predictors included: clients’ and therapists’ demographics, clients’ IIP-64 (58), clients’ SSI (60), therapists’ FIS (51, 52), and the OQ-45 total score and subscales. All methods were carried out in accordance with relevant guidelines and regulations. The study was approved by the Institutional Review Board at Ohio University and all ethical standards were followed.

2.4 Statistical analyses

2.4.1 Identifying the models with the best fit to the data

Model selection is a crucial step in data-driven modeling. In both linear and nonlinear modeling, selection criteria are generally used to identify a model that (a) fits the data well, (b) consists of model variables that can be easily interpreted, (c) involves a parsimonious representation, and (d) can be used for inference and model prediction. In the present study, we applied a search for the best subset of input parameters, considering all possible models that consist of up to nine terms: linear and quadratic variable effects, and pairwise interactions. To account for the nested structure of the data (therapist and client levels), we used multilevel models with the lmer function of the R package lme4 (62). This package is commonly employed by researchers for testing hypotheses in multilevel treatment data. However, by utilizing a data-driven search for the best model—a principle from the machine learning framework—we not only enhance the validity of model inferences but also bolster predictive power. Simultaneously, this process enhances the interpretability of the findings.

2.4.2 Training and cross-validation

We systematically searched for the best subset of input parameters, considering all possible models that consist of up to nine terms, due to the limited sample size: linear, quadratic effects, and pairwise interactions. To take into account the nested data structure, we applied multilevel models using the lmer function of the R package lme4 (62) inside the machine learning framework. The best model was selected based on the Akaike Information Criterion (AIC) (63) and cross-validation. Specifically, we employed leave-one-out cross-validation (LOOCV) to identify a model with the lowest AIC in the training phase while ensuring that the explained variance (R2) for the LOOCV did not decrease by more than 10%, which would indicate potential overfitting. Using LOOCV in model selection provides advantages such as robust evaluation by testing on multiple validation sets. Minimizing the AIC helps find a balanced model fit while considering complexity. LOOCV builds a model for each data point, creating as many models as data points. It leaves one data point out for validation in each iteration, training the model on the rest. The term “training” refers to the process of teaching the model to make predictions based on the dataset. During training, the model adjusts its parameters and internal settings using the training data to understand patterns and relationships. After training, the model can make accurate predictions or classifications when given new, unseen data. In LOOCV, model results are combined, typically by averaging or summing, to assess overall performance and generalization. It is recommended for smaller datasets but can be computationally expensive for large ones. LOOCV does not involve splitting data into traditional training and test sets; each model uses one data point as a test while others are for training, repeating for all data points (64). We opted LOOCV over the more conventional 10-fold cross-validation for several reasons. First, our dataset is relatively small, and with this limited data, LOOCV is less prone to overfitting as it employs nearly all available data for training in each iteration. Secondly, LOOCV typically yields less biased estimates of model performance, particularly in cases of limited data, which is essential for obtaining a precise evaluation of our model’s capabilities. Lastly, LOOCV maximizes data utilization by ensuring that every data point is used for both training and testing, which is particularly advantageous when working with a small dataset, allowing for a comprehensive understanding of the model’s behavior (65, 66). However, it should be noted that LOOVC runs the risk of overestimating prediction accuracy, or conversely, underestimating prediction error. We report model effect size based on the quasi-R2, as proposed by Nakagawa and Schielzeth (67).

2.4.3 Providing exact interpretation of all model effects

The proposed restriction for the potential model effects (described above) also helps to filter out complicated effects with a low level of interpretability. Linear effects are easy to interpret merely based on model coefficients. For quadratic effects and pairwise interactions, we provided plots of the effects to increase their interpretability (Figures 13).

Figure 1
www.frontiersin.org

Figure 1. Graphical presentation of the effects in models 1–2. Model 1: predicting change in treatment outcome based on baseline variables; Model 2: predicting between-client differences in client-reported alliance.

Figure 2
www.frontiersin.org

Figure 2. Graphical presentation of the effects in models 3–4. Model 3: predicting between-client differences in therapist-reported alliance; Model 4: predicting within-client changes during treatment in client-reported alliance.

Figure 3
www.frontiersin.org

Figure 3. Model 5: predicting within-client changes during treatment in therapist-reported alliance.

We used this approach to predict five client-level psychotherapy process and outcome variables. We used Model 1 to identify the best model for predicting change in treatment outcome. We operationalized change in treatment outcome as the client-specific slope of psychological dysfunction from pre- to post-treatment. Model 2 served to predict between-clients differences in the client-reported alliance, and Model 3 in the therapist-reported alliance. In Models 2 and 3, the aggregated level of alliance across time served as an estimation of between-clients alliance. Model 4 served to predict within-client changes in the client-reported alliance of the course of treatment. Model 5 served to predict within-client changes in the therapist-reported alliance. In Models 4 and 5, the client-level session-related random slopes (time effect), calculated based on the mixed models, served as an estimation of the within-client alliance.

In all models, pre-treatment client and therapist characteristics were used as predictors. We used the following predictors: client and therapist demographics (gender, age, and income), therapist characteristics (FIS, clinical orientation, and whether they had clinical training), client characteristics (Inventory of Interpersonal Problems, Social Skills Inventory, Outcome Questionnaire at baseline, and its three subscales, and whether they had treatment before). All continuous predictors were preprocessed by Z-score transformation, and all binary predictors were scored as 0.5 vs. −0.5. Four missing values in income were imputed using the k-nearest neighbor imputation approach (68). We tested normality assumption and heteroscedacity of the residuals in the final models. The assumptions were met. K-nearest neighbor imputation is performed by finding the k closest samples (Euclidean distance) in the training set. Imputation by bagging fits a bagged tree model for each predictor, as a function of all the others (69). KNN-based imputation demonstrated an adequate performance compared to other imputation approaches though using only a single imputed dataset may underestimates the standard error estimations. (70).

A data-driven approach was used to produce predictions in a rigorous test to increase the predictive power of the models, enhance the replicability of the findings, and enable the interpretation of each effect. We focused only on models that can be replicated outside the sub-sample used to build the models. Unlike many machine learning approaches, this one accounted for the nested structure of the data (clients nested within therapists) and provided interpretable results, rather than findings in a “black box” format.

3 Results

3.1 Model 1: predicting change in treatment outcome based on baseline variables

The model explained 52.8% of the variance in treatment outcome for out-of-sample prediction. Variables contributing to the explained variance were as follows (Table 1; Figure 1, Model 1): (a) therapists’ cognitive, dynamic, and humanistic orientations were related to less reduction in symptoms than were eclectic or undetected orientation; (b) intermediate level of client social skills predicted less symptom reduction than did low or high client social skills; (c) for clients with lower levels of baseline symptom severity, more severe psychological dysfunction was associated with less symptom reduction, whereas for clients with higher levels of baseline symptom severity, no significant association was found between psychological dysfunction and symptom reduction; and (d) for clients with more interpersonal dysfunction, more psychological dysfunction was associated with less symptom reduction, whereas for clients with lower levels of interpersonal dysfunction, no significant association was found between psychological dysfunction and symptom reduction.

Table 1
www.frontiersin.org

Table 1. A predictive model of treatment outcome.

3.2 Model 2: predicting between-clients differences in client-reported alliance

The model explained 32.2% of the variance for out-of-sample prediction of between-clients differences in client-reported alliance. Variables contributing to the explained variance were as follows (Table 2; Figure 1, Model 2): (a) interaction between age and psychological dysfunction such that for younger therapists, more psychological dysfunction was associated with a weaker alliance, whereas for older therapists, more psychological dysfunction was associated with a stronger alliance; (b) interaction between age and clinical training such that for younger therapists, those with clinical training had a stronger alliance than those without clinical training, whereas, for older therapists, those with clinical training had a weaker alliance than those without clinical training.

Table 2
www.frontiersin.org

Table 2. Model 2: predicting between-clients differences in client-reported alliance.

3.3 Model 3: predicting between-clients differences in therapist-reported alliance

The model explained 52.8% of the variance for out-of-sample prediction of between-clients differences in therapist-reported alliance. Variables contributing to the explained variance were as follows (Table 3; Figure 2, Model 3): (a) for poorer FIS, female therapists showed stronger alliance than did male therapists, whereas for better FIS, no significant differences were found between males and females; and (b) for males, higher therapist income was associated with weaker alliance, whereas for females, there was no significant association between income and alliance.

Table 3
www.frontiersin.org

Table 3. Model 3: predicting between-clients differences in therapist-reported alliance.

3.4 Model 4: predicting within-client changes during treatment in client-reported alliance

The model explained 43.8% of the variance for out-of-sample prediction of within-client changes in client-reported alliance. Variables contributing to the explained variance were as follows (Table 4; Figure 2, Model 4): (a) intermediate levels of interpersonal dysfunction were associated with less strengthening of the alliance throughout the course of treatment; (b) for better FIS, older clients showed more within-client strengthening of the alliance than did younger ones, whereas for poorer FIS, there was no association between client age and alliance; and (c) for younger clients, clients’ interpersonal dysfunction did not predict alliance, whereas, for older clients, higher interpersonal dysfunction was associated with less strengthening of within-client alliance.

Table 4
www.frontiersin.org

Table 4. Model 4: predicting within-client changes during treatment in client-reported alliance.

3.5 Model 5: predicting within-client changes during treatment In therapist-reported alliance

The model explained 24.1% of the variance for out-of-sample prediction of within-client changes in therapist-reported alliance. Variables contributing to the explained variance were as follows (Table 5; Figure 3): (a) intermediate levels of clients’ interpersonal skills predicted more strengthening in within-client alliance than did low or high levels of clients’ interpersonal skills; and (b) for male therapists, a history of previous psychological treatment predicted more strengthening in within-client alliance than no history of previous treatment, whereas, for females therapists, there was no significant association between history of previous psychological treatment and alliance.

Table 5
www.frontiersin.org

Table 5. Model 5: predicting within-client changes during treatment in therapist-reported alliance.

The R code and further information regarding the models employed in this study can be accessed at https://osf.io/uzjhy/. Additionally, model evaluation metrics and correlation metrics between the study variables are provided in the supplement.

4 Discussion

Results of the current study showed that a large portion of variance, both in alliance and outcome, can be explained based only on the pre-treatment characteristics of clients and therapists. Baseline characteristics (clients’ psychological dysfunction, interpersonal dysfunction, and social skills, and therapists’ general treatment orientation) were able to predict 52.8% of the variance in outcome and 24.1–52.8% of the variance in alliance for clients whose data were not used for building the models on which the predictions were based, to avoid overfitting. Thus, the present study demonstrates the potential utility of integrating explanatory and predictive models for psychotherapy science. Such integration may allow for deriving theoretical insights through selecting the variables that have the most relevance in terms of explanation (based on the theoretical framework) and including them in a model of prediction of treatment processes and outcomes. Whereas the most prominent explanatory studies in psychotherapy today can explain about 5–7% variance (71), the present study increased this range several fold.

Results of the current study showed that changes in symptoms from pre- to post-treatment were predicted mostly by clients’ baseline symptoms as well as psychological and interpersonal functioning. Previous studies using machine learning to predict treatment response found mixed results. For example, whereas Yin et al. (72) reported that greater baseline symptoms severity were among the most important predictors of treatment response, Ziobrowski et al. (73) did not find baseline symptom severity to be among the important predictors. Future studies should address this discrepancy in the literature. If replicated, the finding that baseline symptoms predict treatment response may suggest that in tailoring treatment to the individual client, considerable attention should be given to the clients’ symptomatic complaints (25, 26). The process of tailoring treatment to the individual client can be further enhanced using network analysis to recognize the key symptoms and their dynamics when identifying treatment targets (74, 75).

Concerning alliance prediction models, using both within- and between-clients prediction enabled us to recognize predictors of both state-like and trait-like components of the alliance. At the trait-like component, we found that therapist characteristics were better predictors of trait-like alliance than were client characteristics. Specifically, we found that therapist age moderated the effect of training on alliance, with training being associated with better alliance only for young therapists. In addition, we found that therapists’ gender moderated the association between therapists’ income and alliance, with higher income being associated with poorer alliance only in male therapists. Gender also moderated the association between FIS and alliance, with poorer FIS being associated with poorer alliance only in male therapists. Considering this proof-of-concept study, we will not individually elaborate on each outcome. However, in a broader sense, the observed interaction effects potentially indicate that cultural factors might influence how therapists’ characteristics impact the therapeutic alliance. For instance, income could affect males differently, implying that financial status might correlate with the therapeutic relationship in gender-specific ways. Moreover, interpersonal skills, often more culturally encouraged in women who tend to exhibit greater supportive communication, potentially leading to a reduction in the variance explained by FIS among women (76). In our sample, 66% of the females had high FIS scores, whereas only 37.5% of the men had high FIS scores. This is in accordance with a recent study indicating that female students demonstrate superior interpersonal skills compared to males in the context of medical consultations (77). Importantly, the centrality of therapists’ characteristics in predicting alliance is consistent with a recent meta-analysis by Del Re et al. (78), who found support for therapist effects on alliance, indicating that some therapists are better at forming strong alliances with their clients than others. Still, previous studies focused mainly on a given therapist’s characteristics at a certain point in time. For example, one study found that therapist age was positively associated with trait-like alliance but found no effect of gender or therapist years of experience on alliance (42). Using the proposed framework, we were able to identify a more nuanced understanding that has the potential to elucidate previously inconsistent findings in psychotherapy research.

The current findings, together with those reported in the literature (7982), highlight the importance of therapist characteristics in determining alliance and treatment outcome. Such characteristics, when supported by theory and clinical observations, can be used to identify candidates for clinical training who are expected to form strong alliances with their clients. Furthermore, characteristics that are amenable to change and will be found to be causally related to outcome should be the focus of empirically guided training programs for clinicians (83).

The findings suggest that whereas trait-like alliance was predicted mostly by therapist characteristics, state-like changes in alliance over the course of treatment were predicted by both client and therapist characteristics, and by the interaction between them. Here again, although previous studies reported no effect of client age on changes in alliance (28, 84), the findings of the current study suggest that client age may have a more complex effect on alliance. Specifically, older clients were found to benefit more from therapist FIS and were more negatively affected by their own interpersonal skills. Such nuanced understanding may facilitate progress toward personalization in both client treatment and therapist training.

Taken together, the explainable nature (a glass box vs. a black box) of the proposed framework of integrating explanation and prediction in psychotherapy may be instrumental in reaching a more nuanced understanding of the richness of clinical practice, where each predictor is not isolated from the others, but rather, interacts with them in complex ways in predicting the process and treatment outcome. Given the small sample size, the findings should be interpreted with caution, and serve mainly as a proof of concept demonstrating the potential utility of integrating explanation and prediction in computational psychotherapy science. Still, the results of these predictive models have the potential to assist in drawing a map of the factors contributing to the psychotherapy process and outcome and the complex interconnections between them. If replicated in future studies with large samples, the current findings offer instructive insights that expand the available literature on the predictors of psychotherapy process and outcome.

The most important limitation of the present study is the flip side of one of its main merits: its unique sample. Because of ethical considerations, maximizing variance in therapists results in minimizing variance in clients. To avoid restriction of range in therapist characteristics, we recruited a diverse set of treatment providers, who then provided counseling to a sample of individuals who had not actively sought treatment. For the same reasons, the effect of baseline characteristics was quite likely inflated. In the case of therapists who are more skilled in repairing alliance ruptures and who received appropriate alliance-focused training, it may be possible to rise above the deterministic view that the strengthening of alliance can be predicted based on the predispositions of the clients and bring about a real therapeutic change in the clients’ pre-treatment potential to form a strong alliance. In addition, since therapists represent both ends of highly skilled and unskilled communicators, this sample might have inflated effect sizes. Importantly, given the unique characteristics of the sample, it is questionable whether the results can be consistently generalized to psychotherapy settings. Another important limitation is that the small sample size restricted our ability to calculate therapist effects. Given the unique characteristics of the sample that were required, specifically, the richness of the variance and availability of baseline predictors (including therapist performance, as coded before the treatment, based on a standard evaluative task), we were limited by the available sample size. Future work on larger samples is critical. Thus, the current findings should be regarded as a proof of concept of our suggested framework of integrating explanation and prediction, rather than serving to inform clinical practice. In this work, we used a quasi-R2 estimate of Nakagawa and Shielzeth (67) because of the relatively simple data architecture and our interest in just the marginal explained variability. The approach suggested by Rights and Sterba (85) provides estimates with a better variance composition, taking into account various scenarios of the data and model structures. One should consider using these estimates.

Although the current proof of concept focused on client and therapist pretreatment characteristics, the proposed framework for integrating explanation and prediction in psychotherapy science can also be implemented on in-treatment data. Personalized treatment does not end with pretreatment clinical decision-making. Rather, ongoing tailoring is needed as well (86, 87). The literature on therapist responsiveness highlights the importance of ongoing tailoring of the treatment to the client. Effective therapists are responsive to client behaviors within the emerging context of the treatment (55). For example, observations of the many changing characteristics and behaviors of the client may prompt a therapist to use different interventions in the course of therapy (88). Likewise, the response of the client to a particular strategy may prompt the therapist’s next move, whether to stay the course or try something else—e.g., the therapist should not push an interpretation if the client responds to it defensively (55, 89). Therapeutic interactions exist in a constant loop of feedback and mutual influence (89). Using the proposed framework, ongoing data collection can be fed into the proposed models to support in-treatment decision-making processes (86). While the current proof of concept focused on pretreatment client and therapist characteristics, it is essential to note that the available data limited the construction of individualized models for each participant due to most variables having only one assessment per participant. Nevertheless, embracing idiographic approaches holds promise in potentially enhancing model predictability and optimizing treatment outcomes. Future studies adopting idiographic approaches are recommended to explore and leverage the benefits of constructing personalized models.

Explanatory models have contributed greatly to developments in the field of psychotherapy research, and today we know much more than we did 50 years ago about what drives therapeutic change and about the factors influencing it. This study demonstrates the great potential of the proposed approach to produce robust predictions of the process and outcome of treatment, offering a potential solution to the p-hacking and replicability problem. Such an approach is needed to answer questions in which future predictions are important, such as therapist selection, client prognosis to benefit from treatment, and so on. The present findings show how measures that have been used before in psychotherapy research can predict a large portion of the process (in this case, alliance) and outcome of psychotherapy, shedding new light on previously inconsistent findings.

Data availability statement

The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding author.

Ethics statement

The studies involving humans were approved by Institutional Review Board at Ohio University. The studies were conducted in accordance with the local legislation and institutional requirements. The participants provided their written informed consent to participate in this study.

Author contributions

HF: Conceptualization, Writing – original draft, Writing – review & editing. SS: Conceptualization, Writing – original draft, Writing – review & editing. SZ-M: Conceptualization, Writing – original draft, Writing – review & editing. PG: Formal analysis, Conceptualization, Writing – review & editing. TA: Conceptualization, Resources, Supervision, Writing – review & editing.

Funding

The author(s) declare financial support was received for the research, authorship, and/or publication of this article. The study was supported by the American Psychological Foundation (APF) Walter Katkovsky grant.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fpsyt.2023.1274764/full#supplementary-material

References

1. Shmueli, G. To explain or to predict? Stat Sci. (2010) 25:289–310. doi: 10.1214/10-STS330

Crossref Full Text | Google Scholar

2. Barkham, M, Lutz, W, and Castonguay, LG. Bergin and Garfield’s Handbook of Psychotherapy and Behavior Change. New York, NY: Wiley (2021).

Google Scholar

3. Yarkoni, T, and Westfall, J. Choosing prediction over explanation in psychology: lessons from machine learning. Perspect Psychol Sci. (2017) 12:1100–22. doi: 10.1177/1745691617693393

PubMed Abstract | Crossref Full Text | Google Scholar

4. Zilcha-Mano, S, and Fisher, H. Distinct roles of state-like and trait-like patient–therapist alliance in psychotherapy. Nat Rev Psychol. (2022) 1:194–210. doi: 10.1038/s44159-022-00029-z

Crossref Full Text | Google Scholar

5. Bi, Q, Goodman, KE, Kaminsky, J, and Lessler, J. What is machine learning? A primer for the epidemiologist. Am J Epidemiol. (2019) 188:2222–39. doi: 10.1093/aje/kwz189

PubMed Abstract | Crossref Full Text | Google Scholar

6. Wu, S, Harris, TJ, and McAuley, KB. The use of simplified or misspecified models: linear case. Can J Chem Eng. (2007) 85:386–98. doi: 10.1002/cjce.5450850401

Crossref Full Text | Google Scholar

7. Hofman, JM, Watts, DJ, Athey, S, Garip, F, Griffiths, TL, Kleinberg, J, et al. Integrating explanation and prediction in computational social science. Nature. (2021) 595:181–8. doi: 10.1038/s41586-021-03659-0

PubMed Abstract | Crossref Full Text | Google Scholar

8. Ebersole, CR, Atherton, OE, Belanger, AL, Skulborstad, HM, Allen, JM, Banks, JB, et al. Many labs 3: evaluating participant pool quality across the academic semester via replication. J Exp Soc Psychol. (2016) 67:68–82. doi: 10.1016/j.jesp.2015.10.012

Crossref Full Text | Google Scholar

9. Nosek, BA, and Lakens, D (2014). Registered reports: a method to increase the credibility of published results. Soc. Psychol 45, 137–141. doi: 10.1027/1864-9335/a000192

Crossref Full Text | Google Scholar

10. Open Science Collaboration. Estimating the reproducibility of psychological science. Science. (2015) 349:aac4716. doi: 10.1126/science.aac4716

Crossref Full Text | Google Scholar

11. Munafò, MR, Nosek, BA, Bishop, DV, Button, KS, Chambers, CD, Percie du Sert, N, et al. A manifesto for reproducible science. Nat Hum Behav. (2017) 1:1–9. doi: 10.1038/s41562-016-0021

Crossref Full Text | Google Scholar

12. Olsson-Collentine, A, van Aert, R, Bakker, M, and Wicherts, J. Meta-analyzing the multiverse: a peek under the hood of selective reporting. Psychol Methods. (2023). doi: 10.1037/met0000559 [Online ahead of print].

PubMed Abstract | Crossref Full Text | Google Scholar

13. Lazer, D, Pentland, A, Adamic, L, Aral, S, Barabási, AL, Brewer, D, et al. Computational social science. Science. (2009) 323:721–3. doi: 10.1126/science.1167742

PubMed Abstract | Crossref Full Text | Google Scholar

14. Aafjes-van Doorn, K, Kamsteeg, C, Bate, J, and Aafjes, M. A scoping review of machine learning in psychotherapy research. Psychother Res. (2021) 31:92–116. doi: 10.1080/10503307.2020.1808729

PubMed Abstract | Crossref Full Text | Google Scholar

15. Alanazi, A. Using machine learning for healthcare challenges and opportunities. Inform Med Unlock. (2022) 30:100924. doi: 10.1016/j.imu.2022.100924

Crossref Full Text | Google Scholar

16. Chekroud, AM, Bondar, J, Delgadillo, J, Doherty, G, Wasil, A, Fokkema, M, et al. The promise of machine learning in predicting treatment outcomes in psychiatry. World Psychiatry. (2021) 20:154–70. doi: 10.1002/wps.20882

PubMed Abstract | Crossref Full Text | Google Scholar

17. Penedo, JMG, Rubel, J, Meglio, M, Bornhauser, L, Krieger, T, Babl, A, et al. Using machine learning algorithms to predict the effects of change processes in psychotherapy: toward process-level treatment personalization. Psychotherapy. (2023) 60:536–47. doi: 10.1037/pst0000507

Crossref Full Text | Google Scholar

18. Rubel, JA, Zilcha-Mano, S, Giesemann, J, Prinz, J, and Lutz, W. Predicting personalized process-outcome associations in psychotherapy using machine learning approaches—a demonstration. Psychother Res. (2020) 30:300–9. doi: 10.1080/10503307.2019.1597994

PubMed Abstract | Crossref Full Text | Google Scholar

19. Simmons, JP, Nelson, LD, and Simonsohn, U. False-positive psychology: undisclosed flexibility in data collection and analysis allows presenting anything as significant. Psychol Sci. (2011) 22:1359–66. doi: 10.1177/0956797611417632

Crossref Full Text | Google Scholar

20. Békés, V, Aafjes-van Doorn, K, Zilcha-Mano, S, Prout, T, and Hoffman, L. Psychotherapists’ acceptance of telepsychotherapy during the COVID-19 pandemic: a machine learning approach. Clin Psychol Psychother. (2021) 28:1403–15. doi: 10.1002/cpp.2682

PubMed Abstract | Crossref Full Text | Google Scholar

21. Carter, EC, Schönbrodt, FD, Gervais, WM, and Hilgard, J. Correcting for bias in psychology: a comparison of meta-analytic methods. Adv Methods Pract Psychol Sci. (2019) 2:115–44. doi: 10.1177/2515245919847196

Crossref Full Text | Google Scholar

22. Zilcha-Mano, S, Roose, SP, Brown, PJ, and Rutherford, BR. A machine learning approach to identifying placebo responders in late-life depression trials. Am J Geriatr Psychiatry. (2018) 26:669–77. doi: 10.1016/j.jagp.2018.01.001

PubMed Abstract | Crossref Full Text | Google Scholar

23. Cohen, ZD, and DeRubeis, RJ. Treatment selection in depression. Annu Rev Clin Psychol. (2018) 14:209–36. doi: 10.1146/annurev-clinpsy-050817-084746

Crossref Full Text | Google Scholar

24. Flückiger, C, Del Re, AC, Wampold, BE, and Horvath, AO. The alliance in adult psychotherapy: a meta-analytic synthesis. Psychotherapy. (2018) 55:316–40. doi: 10.1037/pst0000172

PubMed Abstract | Crossref Full Text | Google Scholar

25. Bohart, AC, and Wade, AG. The client in psychotherapy In: MJ Lambert, editor. Bergin and Garfield’s Handbook of Psychotherapy and Behavior Change. Hoboken, NJ: Wiley (2013). 219–57.

Google Scholar

26. Clarkin, JF, and Levy, KN. Psychotherapy for patients with borderline personality disorder: focusing on the mechanisms of change. J Clin Psychol. (2006) 62:405–10. doi: 10.1002/jclp.20238

PubMed Abstract | Crossref Full Text | Google Scholar

27. Brewe, AM, Mazefsky, CA, and White, SW. Therapeutic Alliance formation for adolescents and young adults with autism: relation to treatment outcomes and client characteristics. J Autism Dev Disord. (2021) 51:1446–57. doi: 10.1007/s10803-020-04623-z

PubMed Abstract | Crossref Full Text | Google Scholar

28. Accurso, EC, Fitzsimmons-Craft, EE, Ciao, A, Cao, L, Crosby, RD, Smith, TL, et al. Therapeutic alliance in a randomized clinical trial for bulimia nervosa. J Consult Clin Psychol. (2015) 83:637–42. doi: 10.1037/ccp0000021

PubMed Abstract | Crossref Full Text | Google Scholar

29. Newman, MG, Crits-Christoph, P, Gibbons, MBC, and Erickson, TM. Participant factors in treating anxiety disorders In: LG Castonguay and LE Beutler, editors. Principles of Therapeutic Change That Work. New York, NY: Oxford University Press New York (2006). 121–54.

Google Scholar

30. Patel, KD, and Suhr, JA. The relationship of MMPI–2–RF scales to treatment engagement and alliance. J Pers Assess. (2020) 102:594–603. doi: 10.1080/00223891.2019.1635488

PubMed Abstract | Crossref Full Text | Google Scholar

31. Folmo, EJ, Stänicke, E, Johansen, MS, Pedersen, G, and Kvarstein, EH. Development of therapeutic alliance in mentalization-based treatment—goals, bonds, and tasks in a specialized treatment for borderline personality disorder. Psychother Res. (2021) 31:604–18. doi: 10.1080/10503307.2020.1831097

PubMed Abstract | Crossref Full Text | Google Scholar

32. Hansen, NB, and Lambert, MJ. An evaluation of the dose–response relationship in naturalistic treatment settings using survival analysis. Ment Health Serv Res. (2003) 5:1–12. doi: 10.1023/A:1021751307358

Crossref Full Text | Google Scholar

33. Hersoug, AG, Høglend, P, Havik, OE, von der Lippe, A, and Monsen, JT. Pretreatment patient characteristics related to the level and development of working alliance in long-term psychotherapy. Psychother Res. (2009) 19:172–80. doi: 10.1080/10503300802657374

PubMed Abstract | Crossref Full Text | Google Scholar

34. Constantino, M, and Smith-Hansen, L. Patient interpersonal factors and the therapeutic alliance in two treatments for bulimia nervosa. Psychother Res. (2008) 18:683–98. doi: 10.1080/10503300802183702

PubMed Abstract | Crossref Full Text | Google Scholar

35. Dinger, U, Zilcha-Mano, S, McCarthy, KS, Barrett, MS, and Barber, JP. Interpersonal problems as predictors of alliance, symptomatic improvement and premature termination in treatment of depression. J Affect Disord. (2013) 151:800–3. doi: 10.1016/j.jad.2013.07.003

PubMed Abstract | Crossref Full Text | Google Scholar

36. Nissen-Lie, HA, Havik, OE, Høglend, PA, Monsen, JT, and Rønnestad, MH. The contribution of the quality of therapists’ personal lives to the development of the working alliance. J Couns Psychol. (2013) 60:483–95. doi: 10.1037/a0033643

PubMed Abstract | Crossref Full Text | Google Scholar

37. Greenspan, M, and Kulish, NM. Factors in premature termination in long-term psychotherapy. Psychother Theory Res Pract Train. (1985) 22:75–82. doi: 10.1037/h0088529

Crossref Full Text | Google Scholar

38. Anderson, T, Ogles, BM, Patterson, CL, Lambert, MJ, and Vermeersch, DA. Therapist effects: facilitative interpersonal skills as a predictor of therapist success. J Clin Psychol. (2009) 65:755–68. doi: 10.1002/jclp.20583

PubMed Abstract | Crossref Full Text | Google Scholar

39. Budge, SL, Owen, JJ, Kopta, SM, Minami, T, Hanson, MR, and Hirsch, G. Differences among trainees in client outcomes associated with the phase model of change. Psychotherapy. (2013) 50:150–7. doi: 10.1037/a0029565

PubMed Abstract | Crossref Full Text | Google Scholar

40. Minami, T, Davies, DR, Tierney, SC, Bettmann, JE, McAward, SM, Averill, LA, et al. Preliminary evidence on the effectiveness of psychological treatments delivered at a university counseling center. J Couns Psychol. (2009) 56:309–20. doi: 10.1037/a0015398

Crossref Full Text | Google Scholar

41. Wampold, BE, and Brown, GSJ. Estimating variability in outcomes attributable to therapists: a naturalistic study of outcomes in managed care. J Consult Clin Psychol. (2005) 73:914–23. doi: 10.1037/0022-006X.73.5.914

PubMed Abstract | Crossref Full Text | Google Scholar

42. Hersoug, AG, Høglend, P, Havik, O, von der Lippe, A, and Monsen, J. Therapist characteristics influencing the quality of alliance in long-term psychotherapy. Clin Psychol Psychother. (2009) 16:100–10. doi: 10.1002/cpp.605

PubMed Abstract | Crossref Full Text | Google Scholar

43. Goldberg, SB, Rousmaniere, T, Miller, SD, Whipple, J, Nielsen, SL, Hoyt, WT, et al. Do psychotherapists improve with time and experience? A longitudinal analysis of outcomes in a clinical setting. J Couns Psychol. (2016) 63:1–11. doi: 10.1037/cou0000131

PubMed Abstract | Crossref Full Text | Google Scholar

44. Lorentzen, S, Bakali, JV, Hersoug, AG, Hagtvet, KA, Ruud, T, and Høglend, P. Psychodynamic group psychotherapy: impact of group length and therapist professional characteristics on development of therapeutic alliance. Clin Psychol Psychother. (2012) 19:420–33. doi: 10.1002/cpp.758

PubMed Abstract | Crossref Full Text | Google Scholar

45. Manne, SL, Kashy, DA, Rubin, S, Hernandez, E, and Bergman, C. Therapist and patient perceptions of alliance and progress in psychological therapy for women diagnosed with gynecological cancers. J Consult Clin Psychol. (2012) 80:800–10. doi: 10.1037/a0029158

PubMed Abstract | Crossref Full Text | Google Scholar

46. Huppert, JD, Bufka, LF, Barlow, DH, Gorman, JM, Shear, MK, and Woods, SW. Therapists, therapist variables, and cognitive-behavioral therapy outcome in a multicenter trial for panic disorder. J Consult Clin Psychol. (2001) 69:747–55. doi: 10.1037/0022-006X.69.5.747

PubMed Abstract | Crossref Full Text | Google Scholar

47. Pfeiffer, E, Ormhaug, SM, Tutus, D, Holt, T, Rosner, R, Wentzel Larsen, T, et al. Does the therapist matter? Therapist characteristics and their relation to outcome in trauma-focused cognitive behavioral therapy for children and adolescents. Eur J Psychotraumatol. (2020) 11:1776048. doi: 10.1080/20008198.2020.1776048

PubMed Abstract | Crossref Full Text | Google Scholar

48. Bowman, DO. Effects of therapist sex on the outcome of therapy. Psychother Theory Res Pract Train. (1993) 30:678–84. doi: 10.1037/0033-3204.30.4.678

Crossref Full Text | Google Scholar

49. Staczan, P, Schmuecker, R, Koehler, M, Berglar, J, Crameri, A, von Wyl, A, et al. Effects of sex and gender in ten types of psychotherapy. Psychother Res. (2017) 27:74–88. doi: 10.1080/10503307.2015.1072285

PubMed Abstract | Crossref Full Text | Google Scholar

50. Anderson, T, Finkelstein, JD, and Horvath, SA. The facilitative interpersonal skills method: difficult psychotherapy moments and appropriate therapist responsiveness. Couns Psychother Res. (2020) 20:463–9. doi: 10.1002/capr.12302

Crossref Full Text | Google Scholar

51. Anderson, T, Crowley, MEJ, Himawan, L, Holmberg, JK, and Uhlin, BD. Therapist facilitative interpersonal skills and training status: a randomized clinical trial on alliance and outcome. Psychother Res. (2016) 26:511–29. doi: 10.1080/10503307.2015.1049671

PubMed Abstract | Crossref Full Text | Google Scholar

52. Anderson, T, McClintock, AS, Himawan, L, Song, X, and Patterson, CL. A prospective study of therapist facilitative interpersonal skills as a predictor of treatment outcome. J Consult Clin Psychol. (2016) 84:57–66. doi: 10.1037/ccp0000060

PubMed Abstract | Crossref Full Text | Google Scholar

53. Beutler, LE. The psychotherapist as a neglected variable in psychotherapy: an illustration by reference to the role of therapist experience and training. Clin Psychol Sci Pract. (1997) 4:44–52. doi: 10.1111/j.1468-2850.1997.tb00098.x

Crossref Full Text | Google Scholar

54. Watson, JC, and Wiseman, HE. The Responsive Psychotherapist: Attuning to Clients in the Moment. American Psychological Association. (2021) doi: 10.1037/0000240-000

Crossref Full Text | Google Scholar

55. Stiles, WB. Responsiveness in psychotherapy research: Problems and ways forward. In. JC Watson and H Wiseman, editors. The responsive psychotherapist: Attuning to clients in the moment. American Psychological Association. (2021). 15–35. doi: 10.1037/0000240-002

Crossref Full Text | Google Scholar

56. Lambert, M, Morton, J, Hatfield, D, Harmon, C, Hamilton, S, Reid, R, et al. (2004). Administration and scoring manual for the outcome Questionnaire-45 OQ measures. Orem UT Am Prof Credentialing Services.

Google Scholar

57. Ogles, BM, Lambert, MJ, and Masters, KS. Assessing outcome in clinical practice. Boston: Allyn & Bacon. (1996).

Google Scholar

58. Horowitz, LM, Alden, LE, Wiggins, JS, and Pincus, AL. Manual for the Inventory of Interpersonal Problems. Oxford, UK: Harcourt Assessment, The Psychological Corporation (2000).

Google Scholar

59. Horowitz, LM, Rosenberg, SE, Baer, BA, Ureño, G, and Villaseñor, VS. Inventory of interpersonal problems: psychometric properties and clinical applications. J Consult Clin Psychol. (1988) 56:885–92. doi: 10.1037/0022-006X.56.6.885

PubMed Abstract | Crossref Full Text | Google Scholar

60. Riggio, RE. Assessment of basic social skills. J Pers Soc Psychol. (1986) 51:649–60. doi: 10.1037/0022-3514.51.3.649

Crossref Full Text | Google Scholar

61. Horvath, AO, and Greenberg, LS The Development of the Working Alliance Inventory: A Research Handbook. In. LS Greenberg and WM Pinsof, editors. New York: Guilford Press. (1986). 60, 529–556

Google Scholar

62. Bates, D, Mächler, M, Bolker, B, and Walker, S (2014). Fitting linear mixed-effects models using lme4. arXiv [Preprint]. doi: 10.48550/arXiv.1406.5823

Crossref Full Text | Google Scholar

63. Akaike, H. A new look at the statistical model identification. IEEE Trans. Autom. Control. (1974). 19:716–23.

Google Scholar

64. Vehtari, A, Gelman, A, and Gabry, J. Practical Bayesian model evaluation using leave-one-out cross-validation and WAIC. Stat Comput. (2017) 27:1413–32. doi: 10.1007/s11222-016-9696-4

Crossref Full Text | Google Scholar

65. Varma, S, and Simon, R. Bias in error estimation when using cross-validation for model selection. BMC Bioinformatics. (2006) 7:91. doi: 10.1186/1471-2105-7-91

PubMed Abstract | Crossref Full Text | Google Scholar

66. Bengio, Y, and Grandvalet, Y. No unbiased estimator of the variance of k-fold cross-validation. J Mach Learn Res. (2004) 16:1089–105.

Google Scholar

67. Nakagawa, S, and Schielzeth, H. A general and simple method for obtaining R2 from generalized linear mixed-effects models. Methods Ecol Evol. (2013) 4:133–42. doi: 10.1111/j.2041-210x.2012.00261.x

Crossref Full Text | Google Scholar

68. Beretta, L, and Santaniello, A. Nearest neighbor imputation algorithms: a critical evaluation. BMC Med Inform Decis Mak. (2016) 16:74. doi: 10.1186/s12911-016-0318-z

PubMed Abstract | Crossref Full Text | Google Scholar

69. Kuhn, M. Building predictive models in R using the caret package. J Stat Softw. (2008) 28:1–26. doi: 10.18637/jss.v028.i05

Crossref Full Text | Google Scholar

70. Platias, C, and Petasis, G (2020). “A comparison of machine learning methods for data imputation” in SETN 2020: 11th Hellenic Conference on Artificial Intelligence. 150–159.

Google Scholar

71. Lambert, MJ. Bergin and Garfield’s Handbook of Psychotherapy and Behavior Change. New York, NY: John Wiley & Sons (2013).

Google Scholar

72. Yin, Q, Stern, M, Kleiman, EM, and Rizvi, SL. Investigating predictors of treatment response in dialectical behavior therapy for borderline personality disorder using LASSO regression. Psychother Res. (2023) 33:455–67. doi: 10.1080/10503307.2022.2138790

PubMed Abstract | Crossref Full Text | Google Scholar

73. Ziobrowski, HN, Cui, R, Ross, EL, Liu, H, Puac-Polanco, V, Turner, B, et al. Development of a model to predict psychotherapy response for depression among veterans. Psychol Med. (2023) 53:3591–600. doi: 10.1017/S0033291722000228

PubMed Abstract | Crossref Full Text | Google Scholar

74. Fisher, AJ, Reeves, JW, Lawyer, G, Medaglia, JD, and Rubel, JA. Exploring the idiographic dynamics of mood and anxiety via network analysis. J Abnorm Psychol. (2017) 126:1044–56. doi: 10.1037/abn0000311

PubMed Abstract | Crossref Full Text | Google Scholar

75. Rodebaugh, TL, Tonge, NA, Piccirillo, ML, Fried, E, Horenstein, A, Morrison, AS, et al. Does centrality in a cross-sectional network suggest intervention targets for social anxiety disorder? J Consult Clin Psychol. (2018) 86:831–44. doi: 10.1037/ccp0000336

PubMed Abstract | Crossref Full Text | Google Scholar

76. Burleson, BR, and Kunkel, A. Revisiting the different cultures thesis: an assessment of sex differences and similarities in supportive communication In: Sex Differences and Similarities in Communication. In. DJ Canary and K Dindia, editors. New York: Routledge (2009). 133–54.

Google Scholar

77. Bellier, A, Chaffanjon, P, Morand, P, Palombi, O, Francois, P, and Labarère, J. Student characteristics associated with interpersonal skills in medical consultations. BMC Med Educ. (2022) 22:338. doi: 10.1186/s12909-022-03412-9

PubMed Abstract | Crossref Full Text | Google Scholar

78. Del Re, AC, Flückiger, C, Horvath, AO, and Wampold, BE. Examining therapist effects in the Alliance—outcome relationship: a multilevel meta-analysis. J Consult Clin Psychol. (2021) 89:371–8. doi: 10.1037/ccp0000637

PubMed Abstract | Crossref Full Text | Google Scholar

79. Baldwin, SA, and Imel, ZE. Therapist effects. In. MJ Lambert editor. Bergin Garfield’s Handb Psychother Behav Change. 6th edn. New York, NY: Wiley (2013) 258–97.

Google Scholar

80. Johns, RG, Barkham, M, Kellett, S, and Saxon, D. A systematic review of therapist effects: a critical narrative update and refinement to review. Clin Psychol Rev. (2019) 67:78–93. doi: 10.1016/j.cpr.2018.08.004

PubMed Abstract | Crossref Full Text | Google Scholar

81. Barkham, M, Lutz, W, Lambert, MJ, and Saxon, D. Therapist effects, effective therapists, and the law of variability In: LG Castonguay and CE Hill, editors. How and Why Are Some Therapists Better Than Others?: Understanding Therapist Effects. American Psychological Association (2017) 13–36. doi: 10.1037/0000034-002

Crossref Full Text | Google Scholar

82. Wampold, BE, Baldwin, SA, and Imel, ZE (2017). What characterizes effective therapists? In: LG Castonguay and CE Hill, editors. How and why are some therapists better than others?: Understanding therapist effects. American Psychological Association (2017) 37–53. doi: 10.1037/0000034-003

Crossref Full Text | Google Scholar

83. Anderson, T, Perlman, MR, McCarrick, SM, and McClintock, AS. Modeling therapist responses with structured practice enhances facilitative interpersonal skills. J Clin Psychol. (2020) 76:659–75. doi: 10.1002/jclp.22911

PubMed Abstract | Crossref Full Text | Google Scholar

84. Santoft, F, Salomonsson, S, Hesser, H, Lindsäter, E, Ljótsson, B, Lekander, M, et al. Mediators of change in cognitive behavior therapy for clinical burnout. Behav Ther. (2019) 50:475–88. doi: 10.1016/j.beth.2018.08.005

PubMed Abstract | Crossref Full Text | Google Scholar

85. Rights, JD, and Sterba, SK. Quantifying explained variance in multilevel models: an integrative framework for defining R-squared measures. Psychol Methods. (2019) 24:309–38. doi: 10.1037/met0000184

PubMed Abstract | Crossref Full Text | Google Scholar

86. Lutz, W, Rubel, JA, Schwartz, B, Schilling, V, and Deisenhofer, AK. Towards integrating personalized feedback research into clinical practice: development of the Trier treatment navigator (TTN). Behav Res Ther. (2019) 120:103438. doi: 10.1016/j.brat.2019.103438

PubMed Abstract | Crossref Full Text | Google Scholar

87. Zilcha-Mano, S, Constantino, MJ, and Eubanks, CF. Evidence-based tailoring of treatment to patients, providers, and processes: introduction to the special issue. J Consult Clin Psychol. (2022) 90:1–4. doi: 10.1037/ccp0000694

PubMed Abstract | Crossref Full Text | Google Scholar

88. Kramer, U, and Stiles, WB. The responsiveness problem in psychotherapy: a review of proposed solutions. Clin Psychol Sci Pract. (2015) 22:277–95. doi: 10.1111/cpsp.12107

Crossref Full Text | Google Scholar

89. Stiles, WB, Honos-Webb, L, and Surko, M. Responsiveness in psychotherapy. Clin Psychol Sci Pract. (1998) 5:439–58. doi: 10.1111/j.1468-2850.1998.tb00166.x

Crossref Full Text | Google Scholar

Keywords: psychotherapy process and outcome, predictive models, explanatory models, machine learning, alliance, data-driven approach

Citation: Fisher H, Stone SJ, Zilcha-Mano S, Goldstein P and Anderson T (2024) Integrating exploration and prediction in computational psychotherapy science: proof of concept. Front. Psychiatry. 14:1274764. doi: 10.3389/fpsyt.2023.1274764

Received: 08 August 2023; Accepted: 15 December 2023;
Published: 12 January 2024.

Edited by:

Veena Kumari, Brunel University London, United Kingdom

Reviewed by:

Ivo Rollmann, Heidelberg University Hospital, Germany
Warren Mansell, The University of Manchester, United Kingdom

Copyright © 2024 Fisher, Stone, Zilcha-Mano, Goldstein and Anderson. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Hadar Fisher, Hadar.fisher@gmail.com

These authors have contributed equally to this work

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.