- Department of Biostatistics and Data Science, School of Public Health, The University of Texas Health Science Center, Houston, TX, United States
As most great discoveries and advancements in science and technology invariably involve the cooperation of a group of researchers, effective collaboration is the key factor. Nevertheless, finding suitable scholars and researchers to work with is challenging and, mostly, time-consuming for many. A recommender who is capable of finding and recommending collaborators would prove helpful. In this work, we utilized a life science and biomedical research database, i.e., MEDLINE, to develop a collaboration recommendation system based on novel graph neural networks, i.e., GraphSAGE and Temporal Graph Network, which can capture intrinsic, complex, and changing dependencies among researchers, including temporal user–user interactions. The baseline methods based on LightGCN and gradient boosting trees were also developed in this work for comparison. Internal automatic evaluations and external evaluations through end-users' ratings were conducted, and the results revealed that our graph neural networks recommender exhibits consistently encouraging results.
Introduction
Academic collaboration can be termed as a research effort done by researchers either from different disciplines or from the same discipline across different groups, nationally or internationally (Katsouyanni, 2008). Collaborations between researchers often have a synergistic effect as follows: The combined expertise of a group of researchers can often yield results far surpassing the sum of the individual researchers' capabilities (Pavlov and Ichise, 2007). As important as it is, however, it is often challenging to find suitable collaborations, partly due to the sheer number of researchers out there and the fact that it is impractical for anyone to be fully aware of others' expertise within a limited time, if they have not known each other beforehand, especially for junior scholars. Adding to that, the constant change is researchers' expertise, areas of interest, and knowledge over time (more rapid for some researchers than others), all make it more challenging. Recommendation systems (RS), or recommenders, have proven to be an effective strategy to deal with such information overload.
Recommendation systems are information filtering systems that deploy data mining and analytics of users' behaviors, for predictions of users' interests in information, products, or services. Given their wide and successful use in commercial applications, RS is being explored and applied in the academic domain as well (Zhu et al., 2021).
In this work, we focused on the following two Graph Neural Networks (GNN) models that cater to our collaborator recommendation problems: GraphSAGE (Hamilton et al., 2017) and Temporal Graph Networks (TGN; Rossi et al., 2020). Both GraphSAGE and TGN are inductive GNNs that are able to generate embeddings for unseen nodes, in addition, TGN was specifically designed for graphs that are dynamic in nature. With their state-of-the-art performances on several transductive and inductive graph prediction tasks, we built on these foundations for recommending collaborators in the population health domain to study this under-explored area. We carefully designed and executed a series of experiments to develop our collaborators recommender. The evaluations of the experiments on the recommender were conducted both internally (using existing relationships within the data) and externally (by collecting end-users' ratings). Our main contributions are as follows:
• We employed novel inductive GNN networks (GraphSAGE and TGN), instead of the traditional machine learning techniques and transductive-only GNNs, to capture intrinsic, complex, and temporal dependencies among researchers for future link predictions.
• Unlike other GNNs commonly used for predicting static user–item interactions in the industry domain (e.g., restaurants, music, and movies to people), we focused on the temporal user–user recommendations in the academic domain.
• We crawled the following data suitable for real-world applications: Datasets from the MEDLINE database with a wide range of time periods and achieved consistently encouraging results for predictions.
• Finally, we developed a web-based application for our recommender, available at http://genestudy.org/recommends/#/collaborators, giving our research practical use. This also allowed us to collect feedback/ratings from end-users to conduct an external evaluation of the system.
The rest of the content in this manuscript is organized as follows: the “Related work” section summarizes the available literature on the topic of collaboration recommendation; the “Materials and Methods” section presents the data and methods we used and the evaluations we carried out for two scenarios of experiments; “Results” section summarizes our GNNs performances; and finally, “Discussion and conclusion” section reiterates our main contributions, extends an in-depth analysis of the results, and brings forward our future plans for the continuation of this work.
Related Work
Collaboration recommendation problems have been commonly addressed in the context of link predictions within networks. The link predictions aim to predict the likelihood of a future association (in our case, collaborations) between two nodes (in our case, researchers), based on the current states of links and nodes in the network (Yu et al., 2014). However, most of the link predictions have been solved using the network topological features for nodes in concern and the Markov process (random process index by time in the topological space), or a combination of both methods. Liben-Nowell and Kleinberg (2007) were the first that applied the topological features of node similarities measures toward linkage predictions, measures that they systematically compared included shortest path, common neighbors, preferential attachments, and six others. Afolabi et al. (2021) used the simple Adamic–Adar Index (Adamic and Adar, 2003) for the link predictions among computer science researchers. Pavlov and Ichise (2007) utilized multiple structural attributes, such as the shortest path from past collaborations, coupled with supervised machine learning algorithms, i.e., support vector machines (SVM) and decision trees, for predicting future research potentials. Yu et al. (2014) also utilized several structural attributes with logistic regression and SVM for collaboration recommendations in the medical domain. Lopes et al. (2010) employed topic modeling to build the profiles of researchers and the asymmetric variant of Jaccard' coefficients (Salton and McGill, 1983) for calculating the node proximity. Cho and Yu (2018) used node similarity scores with multi-networks as follows: Co-authorship, researcher–journal, and school networks for the collaboration predictions. Wang and Sukthankar (2013) introduced social features with “edge clustering” in addition to nodes' proximity measures (e.g., common neighbors, etc.) and employed machine learning models, i.e., Naïve Bayes, logistic regression, etc. for the collaboration predicting on DBLP dataset (Yang and Leskovec, 2012). Kuzmin et al. (2016) used the PageRank (Page et al., 1999), a random walk model, and multilayer networks that merged molecular interaction networks with authorship information for recommending biomedical collaboration potentials. Kong et al. (2016) adopted word2vec to identify the academic domains and the random walk with a restart to compute researchers' feature vectors for collaborator recommendations. Liu et al. (2018) combined neural networks based collaborative entity embedding with a hierarchical factorization model to produce context–aware collaborator results. However, there are still concerns about whether these techniques adequately and properly addressed the linkage issues. On the one hand, instead of relying on these hand-engineered topographical features or the Markov process, we are able to find an algorithm to capture the various information within the data elegantly and expressively for linkage predictions; on the other hand, more importantly, collaboration networks, in general, are not static; researchers and their associations evolve dynamically over time, and therefore ideal algorithms should capture change and iteratively be updated with new information, handle emerging researchers, and recommend potential collaborations accordingly.
With the advances in deep learning (DL), DL models have been popular methodologies for recommenders due to their ability to capture non-linear/complex relationships in data and easily incorporate temporal, contextual, and external information (Mu, 2018). Among these DL algorithms, GNN is undoubtedly the most attractive one due to its superior ability on graph/network structured data to combine connectivity and features of local neighbors, which fits naturally with recommenders' link predictions (Ying et al., 2018; Wu S. et al., 2020). The main idea of GNN is to iteratively aggregate feature information from local neighbors and integrate the aggregated information with the current node representation during the propagation process (Wu Z. et al., 2020; Zhou et al., 2020). There have been some efforts to use GNN for the link predictions (Zheng et al., 2018; Fan et al., 2019; Wang et al., 2019; Zhang et al., 2019; Chen et al., 2020; He et al., 2020; Sun et al., 2020; Zhang and Chen, 2020; Mandal and Maiti, 2021); nevertheless, they were focusing on encoding the static collaborative signal of bipartite user–item interactions instead of the evolving social user–user relationships. Among these, some of the following studies modified the graph operations to be more suitable for collaborative filtering: Chen et al. (2020) removed non-linearity and introduced residual network structures; LightGCN (He et al., 2020) was proposed for learning user–item interactions without feature transformation or non-linear activation; Zheng et al. (2018) introduced spectral convolution operation to discover and learn from the deep connections of user–item in the spectral domain; Zhang et al. (2019) developed a stacked GCN encoder–decoder architecture to solve the user–item matrix completion problem; Zhang and Chen (2020) extracted h-hop enclosing subgraphs to train a GNN for rating matrix completion. Some utilized fixed hierarchical bipartite graphs on the user–item interactions to produce embeddings: Wang et al. (2019) proposed Neural Graph Collaborative Filtering, a framework that explicitly encodes the user–item signal in the form of high-order connectivity using embedding propagation; Mandal and Maiti (2021) incorporated the degree of authenticity of reviews and customers in their trust-based social recommendation, a system which is composed of both customers networks and customer–item interaction networks. Sun et al. (2020) introduced a pairwise neighborhood aggregation layer to explicitly model the relational information between the neighbor nodes, and developed parallel GCNs to exploit the heterogeneous nature of the user–item bipartite graph for recommendation; Fan et al. (2019) developed GraphRec, which jointly captured the interactions and opinions in the user–item graph and heterogeneous social graph.
Materials and Methods
Experiment Scenario With Automatic Evaluations
We crawled MEDLINE, a public dataset of 33 million citations, from PubMed1, for the experiments of collaboration recommendations. Collaborations are defined as “two or more authors sharing a publication.” The number of publications (Figure 1) and collaborations (Figure 2) are surging each year.
We followed similar practices in the existing literature on linkage predictions (Pavlov and Ichise, 2007; Yu et al., 2014; Chuan et al., 2018) and sampled three time periods from the dataset to cover a wider time range. The three time periods are 2000–2002, 2010–2011, and 2019–2020. All follow the 70:15:15% train, validation, and test chronological split as discussed in Rossi et al. (2020) and Xu et al. (2020). For validation and test, inductive tasks (Hamilton et al., 2017) refer to the link predictions involving nodes that have never appeared during graph training. The GNN models were trained and ultimately evaluated on test data and an inductive subset2 of test data, while baseline methods were evaluated on the whole test data (because of its inability to perform in inductive tasks). The detailed statistics of the three sets of data are described in Table 1.
We experimented with two GNN methods (GraphSAGE and TGN) for producing node embeddings and attached a small neural network to feed on produced embedding for linkage predictions. We compared the performances with a transductive GNN method; LightGCN and a machine learning method; Gradient boosted classifier.
GraphSAGE
GraphSAGE (Hamilton et al., 2017) is a framework for inductive representation learning on large graphs. Specifically, GraphSAGE generates embeddings for each node (including unseen ones) by (1) uniformly sampling a fixed small number of neighbors and aggregating the neighbor embeddings using learnable aggregators, such as mean, graph convolution networks (GCN); (2) concatenating neighbor embeddings with the embedding of the node itself; (3) feeding the concatenated embeddings to fully connected networks, represented by the following equation:
where is the embeddings for node v at kth iteration; N(v): neighbors of node v; : the neighbor embeddings at kth iteration; σ: activation function; and Wk: weight matrix.
We defined the nodes as authors and links as collaborations, with the collaborations defined as “the two authors sharing an article” as stated earlier. We constructed the raw node features using Term Frequency–Inverse Document Frequency (TF–IDF) with the following two variations: (1) Mesh terms of the articles (Mesh or Medical Subject Headings3, is the National Library of Medicine controlled vocabulary thesaurus used for indexing articles for PubMed and (2) titles of the articles. The usage is illustrated in Figure 3. The training, validation, and test graph construction for GraphSAGE can be found in Figure 4. We used the Pytorch implementation of GraphSAGE by Deep Graph Library4. The detailed hyperparameters used can be found in Table 2.
Figure 3. Usage of GraphSAGE for collaborator recommendations, modified based on Hamilton et al. (2017). Nodes A, B, C, D, and E are authors; first constructed by using either (1) mesh terms or (2) titles of publications. Links are defined by “sharing of publications.” During training, the embeddings of each node are updated using GraphSAGE, as detailed in the text. Then the question of collaboration predictions becomes as follows: Given the embeddings of two nodes, how likely will there be a link between them?
Figure 4. An illustration of training, valid, and test graph construction for GraphSAGE, where collaborations are split on chronological order.
Temporal Graph Networks
Temporal graph networks (Rossi et al., 2020) is a generic framework for deep learning on dynamic graphs represented as sequences of timed events, which produces the embeddings of graph nodes Z(t) = (z1(t), …, zn(t)(t)). There are four important components that constitute the TGN, which are message function, memory, memory updater, and embeddings. Message functions compute two messages for the involving nodes (i, j) using the memories of i, j before the link, the link timestamp, and edge features, if any. Memory modules store the states of all nodes up till time t:si(t), acting as the compressed representations of the nodes' past interactions. Memory updater updates the memory with the new messages. Finally, the embedding module computes the temporal embeddings of the nodes by performing a graph aggregation over the spatial–temporal neighbors of the node in concern, similar to GraphSAGE.
Message function:
,
where t1, …, tb ≤ t.
mi(t) is the message for node i at time t; is the memory of node i before time t; ei, j(t) is the edge features of node i, j at time t; is the aggregated message for node i at time t after applying an aggregator function on the past messages for node i.
Memory updater:
where memfunc is a learnable memory update function, such as GRU.
Embedding:
where h is a learnable function such as graph attention (Rossi et al., 2020), Ni[0, t] is node i's neighbors up to time t.
The graph was defined similarly to the way used in GraphSAGE, with the timestamp of the links explicitly represented using the publication date. The usage is illustrated in Figure 5. The training, validation, and test graph construction for TGN can be found in Figure 6. We modified the architecture based on the original GitHub repo by twitter research5. The detailed hyperparameters used can be found in Table 3.
Figure 5. Usage of TGN for collaborator recommendations, modified based on Rossi et al. (2020). Nodes A, B, C, D, and E are authors; first constructed by using either (1) mesh terms or (2) titles of publications. Temporal links are defined by “sharing of publications” at time t. During training, each temporal collaboration is computed within a message between the involved nodes [e.g., at time t2, the collaboration between D and E is calculated in both messages, mD(t2) and mE(t2)]. Then the memory state of each author is updated using those temporal messages. The updated memories, together with embeddings constructed similar to GraphSAGE (concatenation of neighbor embeddings and its own embeddings) are then aggregated as the final embedding for each node. Then the question of collaboration predictions becomes as follows: Given the embeddings of two nodes at time t, how likely will there be a link between them?
Figure 6. An illustration of training, valid, and test graph construction for TGN, where the collaborations are split on chronological order.
Baseline: LightGCN
As a competitive transductive GNN baseline, LightGCN was chosen because of its efficiency in many static and transductive recommendation tasks (He et al., 2020; Ragesh et al., 2021). The most essential part of this model is a simplified graph convolution with neither feature transformations nor non-linear activations. The weighted sum of embeddings learned at different propagations layers was used for final node embeddings. Similar to GraphSAGE and TGN-based recommenders, the original features of each author were constructed using mesh terms first, then the titles of the articles. We used the implementation of LightGCN from Pytorch Geometric6 with hyperparameters as detailed in Table 4.
Baseline: Gradient Boosted Classifier (GBC)
Finally, GBC was chosen as another baseline because of its simplicity in implementation, scalability, and proven successes in many competitions (Chen and Guestrin, 2016), in consideration of bias–variance trade-off (Hastie et al., 2009). Simply, boosting is a type of ensemble method that builds weak/slow learners (usually trees) sequentially, to correct the previously wrongly predicted outcomes. Similar to GraphSAGE and TGN, the feature space of each author was constructed using mesh terms and titles of the articles separately for two scenarios. We used the implementation from XGboost7, hyperparameters are detailed in Table 5.
Evaluation Metrics
For all four methods, we used ROC–AUC (AUC) and average precision (AP) for evaluations, which were widely used in experiments (Yu et al., 2014; Rossi et al., 2020; Xu et al., 2020). Since GraphSAGE and TGN are able to handle new nodes, the two metrics were also used to evaluate the inductive subset of the test data.
• AUC: A receiver operating characteristic (ROC) curve is a graphical plot that illustrates the diagnostic ability of a binary classifier system as its discrimination threshold is varied; AUC is the area under an ROC curve, which provides an aggregated measure of performance across all possible classification thresholds (Fawcett, 2006).
• AP: It summarizes precision–recall curve as the weighted mean of precisions achieved at each threshold, with the increase in recall from the previous threshold used as the weight.
Experiment Scenario With External Evaluations
In addition to using existing relationships in crawled data, we further explored the external evaluations by collecting ratings and feedback directly from end-users. We used their PubMed publications available from their resumes from 2019 to 2021 and predicted the possibilities of future collaborations between the user and all authors in the existing database that the user had not interacted with yet.
We provided 1–3 stars for all users to rate our collaborator recommendations, with 3 stars being “the most satisfactory” and 1 star being “the least satisfactory.” See Figure 7 for an example of the evaluation collection page. We considered the recommendations with 2 stars and above as “relevant recommendations,” thus precision@k is calculated as follows: at the kth retrieved item, the proportion of the retrieved items are relevant. We look at both k = 1 and k = 5.
The codes for all current experiments can be found at the GitHub repo: https://github.com/ashraf-yaseen/VRA/tree/master/collaborator_rec.
Results
For the experiments with automatic evaluations, the AUC and AP for two GNN-based methods vs. baselines are shown in Tables 6, 7; all results were averaged over 5 runs. Node/author features for two tables were constructed using mesh terms (Table 6) and publication titles (Table 7) respectively, as detailed in the “Related work” section.
Table 6. AUC and AP for GraphSAGE, TGN, and baseline LightGCN, baseline GBC on sampled datasets from three time periods using mesh terms as node features, with best performances shown in bold.
Table 7. AUC and AP for GraphSAGE, TGN, and baseline LightGCN, baseline GBC on sampled datasets from three time periods using article titles as node features, with best performances shown in bold.
When constructing node/author features using mesh terms, we can see that the best transductive performances were all achieved by LightGCN, demonstrating its competitive ability in learning linkage representations when node features were confined. For inductive subsets, the best performances were split between GraphSAGE and TGN.
When constructing node/author features using publication titles, however, TGN was able to achieve consistently the best results, both transductive and inductive, across all time periods. Furthermore, we can see varying performance improvements over all three methods except for the LightGCN baseline as discussed in the following: A significant improvement on TGN (maximum 62.8% improvement on test AP), a moderate improvement on GraphSAGE (maximum 10.3% improvement on test AP, inductive subset) and on GBC baseline (maximum 15.2% improvement on test AP), and finally a small fluctuation on LightGCN (maximum 0.23% improvement on test AUC, and maximum 0.96% deterioration on test AP). Since the publication titles capture richer information than the mesh terms, this showed that when the raw node features were provided well, TGN was better fitted for the linkage predictions in addition to its inductive ability.
Given this improved performance when using publication titles as node features, we decided to adopt this representation to implement our GraphSAGE and TGN-based recommenders. The two methods were then deployed within a web-based application to allow for external evaluations. We asked several researchers in our institute to evaluate the recommendations of the two methods. At this time, we received ratings/feedback from seven users. Simple statistics of their publications, the average stars, P@1, and P@5 for both GraphSAGE and TGN are shown in Table 8.
Table 8. External evaluations collected from users, with “**” indicating the existence of 3-star rating.
Overall, we can see that GraphSAGE produced better top 1 hits than TGN (P@1 = 0.43 vs. 0.29), but less top-5 hits in general (P@5 = 0.29 vs. 0.37). However, TGN was able to make more strictly relevant recommendations (3-star rating) than GraphSAGE. The external evaluations are inherently more subjective and thus are not idealistic as the automatic evaluations. Considering this, we think that TGN especially has the potential to deliver satisfactory recommendations with proper adjustments, and it will be the foundational architecture of our collaborator recommender moving forward. Nevertheless, we do acknowledge that this is only a small sample evaluation due to low response rates (30%), and therefore larger sample evaluation should be carefully carried out for analysis in the future.
Discussion and Conclusion
In this work, a collaboration recommendation system using novel GNNs (GraphSAGE and TGN) have been developed. Moreover, baselines using both transductive GNN and GBC have also been developed for comparison. The ability of our method to capture intrinsic, complex, and changing dependencies among researchers, including temporal user–user interactions is crucial and important to the overall performance of the recommender. The internal evaluations using crawled collaborators networks revealed that our TGN, when well-supplied with node features, consistently exhibited better performance compared with the baselines, in addition to its strong inductive ability. The external evaluations also revealed that our models are of practical value with encouraging P@1 and P@5 on a small sample, which, we hope, could prove useful to the population health professionals on a larger scale with proper adjustments.
Furthermore, we believe that there is still room for improvement. First, in terms of raw node features, it would be interesting to investigate the possibility of presenting each author as a distribution of features instead of a fixed vector. Representing data with a distribution comes with many advantages; for example, it can allow better uncertainty encodings, and express asymmetries more naturally than dot product, cosine similarity, or Euclidean distance (Wu S. et al., 2020). Managing the uncertainty that is inherent in RS for predictive modeling is important for us to understand how we should interpret our recommendation results and act accordingly. In fact, many popular techniques in RS such as Upper Confidence Band (Auer et al., 2002) and Thomas sampling (Thompson, 1933) require uncertainty estimation to perform more efficient feature space exploration (Zeldes et al., 2017). Dos Santos et al. (2017) and Jiang et al. (2019) actually deployed Gaussian embeddings to capture users' uncertain preferences for improving user representations and recommendation performance. However, in general, the distribution-based representations have not been well-studied in the GNN-based recommendation models (Wu S. et al., 2020). Thus, we hope to work on modifying existing graph architectures in the future to add to this knowledge.
Second, in this work, GNN models were tuned with a random search on the validation dataset for the important hyperparameters centered on reported “best performing” values in the original articles (number of layers, embedding dimensions, learning rate, etc.), due to the large number of hyperparameters involved, as well as corresponding lengthy time consumptions. However, more efficient and thorough hyperparameter tuning should be implemented and better examined for the completeness of the study.
Finally, a detailed look into the external evaluations found that both GNN models favored the researchers from publications with long author lists. This was not surprising, since a long author list can create more links for the graph constructions. One possible next step is to analyze how placing penalty weights on the publications with long authors' lists might affect the performance of our recommender. Another improvement would be to collect users' text feedback in addition to ratings, though this might result in an even lower response rate. Users' text reviews could reveal additional features to consider for encoding, such as geographical locations or affiliations of researchers that could make our recommendations more personalized, thereby more fitting in the service scenario.
Data Availability Statement
The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.
Author Contributions
JZ was responsible for the background review, methodology development, experiment conduction, results summary, discussion, and wrote the article with guidance from AY. AY supervised the concept discussion, contributed to the manuscript writing, and approved the submitted version. Both authors contributed to the article and approved the submitted version.
Conflict of Interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher's Note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Acknowledgments
The authors would like to express thanks to all faculties that participated in the external evaluations and Hulin Wu for insightful discussions.
Footnotes
1. ^https://pubmed.ncbi.nlm.nih.gov/
2. ^Inductive subset: as long as one node (of the connecting two) of the link has never appeared in the training set. Also appeared as “inductive task” in the text earlier.
3. ^https://www.nlm.nih.gov/mesh/meshhome.html
5. ^https://github.com/twitter-research/tgn
References
Adamic, L. A., and Adar, E. (2003). Friends and neighbors on the web. Soc. Netw. 25, 211–230. doi: 10.1016/S0378-8733(03)00009-1
Afolabi, I. T., Ayo, A., and Odetunmibi, O. A. (2021). Academic collaboration recommendation for computer science researchers using social network analysis. Wireless Personal Commun. 121, 487–501. doi: 10.1007/s11277-021-08646-2
Auer, P., Cesa-Bianchi, N., and Fischer, P. (2002). Finite-time analysis of the multiarmed bandit problem. Machine Learn. 47, 235–256. doi: 10.1023/A:1013689704352
Chen, L., Wu, L., Hong, R., Zhang, K., and Wang, M. (2020). Revisiting graph based collaborative filtering: a linear residual graph convolutional network approach. Proc. AAAI Conf. Artif. Intelligence 34, 27–34. doi: 10.1609/aaai.v34i01.5330
Chen, T., and Guestrin, C. (2016). “Xgboost: a scalable tree boosting system,” in Proceedings of the 22nd ACM Sigkdd International Conference on Knowledge Discovery and Data Mining (San Francisco, CA), 785–794. doi: 10.1145/2939672.2939785
Cho, H., and Yu, Y. (2018). Link prediction for interdisciplinary collaboration via co-authorship network. Soc. Netw. Anal. Min. 8, 1–12. doi: 10.1007/s13278-018-0501-6
Chuan, P. M., Son, L. H., Ali, M., Khang, T. D., Huong, L. T., and Dey, N. (2018). Link prediction in co-authorship networks based on hybrid content similarity metric. Appl. Intelligence 48, 2470–2486. doi: 10.1007/s10489-017-1086-x
Dos Santos, L., Piwowarski, B., and Gallinari, P. (2017). “Gaussian embeddings for collaborative filtering,” in Proceedings of the 40th International ACM SIGIR Conference on Research and Development in Information Retrieval (Tokyo), 1065–1068. doi: 10.1145/3077136.3080722
Fan, W., Ma, Y., Li, Q., He, Y., Zhao, E., Tang, J., et al. (2019). “Graph neural networks for social recommendation,” in The World Wide Web Conference (San Francisco, CA), 417–426. doi: 10.1145/3308558.3313488
Fawcett, T. (2006). An introduction to ROC analysis. Patt. Recogn. Lett. 27, 861–874. doi: 10.1016/j.patrec.2005.10.010
Hamilton, W., Ying, Z., and Leskovec, J. (2017). Inductive representation learning on large graphs. Adv. Neural Inform. Proces. Syst. 30, 1024–1034. Available online at: https://papers.nips.cc/paper/2017/file/5dd9db5e033da9c6fb5ba83c7a7ebea9-Paper.pdf
Hastie, T., Tibshirani, R., and Friedman, J. (2009). The Elements of Statistical Learning: Data Mining, Inference, and Prediction. (Berlin: Springer), 219–223. doi: 10.1007/978-0-387-84858-7
He, X., Deng, K., Wang, X., Li, Y., Zhang, Y., and Wang, M. (2020). “Lightgcn: simplifying and powering graph convolution network for recommendation,” in Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval, 639–648. doi: 10.1145/3397271.3401063
Jiang, J., Yang, D., Xiao, Y., and Shen, C. (2019). “Convolutional Gaussian embeddings for personalized recommendation with uncertainty,” in Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, IJCAI-19. International Joint Conferences on Artificial Intelligence Organization (Macao. P.R.), 2642–2648. doi: 10.24963/ijcai.2019/367
Katsouyanni, K. (2008). Collaborative research: accomplishments & potential. Environ. Health 7, 1–7. doi: 10.1186/1476-069X-7-3
Kong, X., Jiang, H., Yang, Z., Xu, Z., Xia, F., and Tolba, A. (2016). Exploiting publication contents and collaboration networks for collaborator recommendation. PLoS ONE 11.2,e0e0148492. doi: 10.1371/journal.pone.0148492
Kuzmin, K., Lu, X., Mukherjee, P. S., Zhuang, J., Gaiteri, C., and Szymanski, B. K. (2016). Supporting novel biomedical research via multilayer collaboration networks. Appl. Netw. Sci. 1, 1–27. doi: 10.1007/s41109-016-0015-y
Liben-Nowell, D., and Kleinberg, J. (2007). The link-prediction problem for social networks. J. Assoc. Inform. Sci. Technol. 2007, 1019–1031. doi: 10.1002/asi.20591
Liu, Z., Xie, X., and Chen, L. (2018). “Context-aware academic collaborator recommendation,” in Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining (London), 1870–1879. doi: 10.1145/3219819.3220050
Lopes, G. R., Moro, M. M., Wives, L. K., and de Oliveira, J. P. M. (2010). “Collaboration recommendation on academic social networks,” in International Conference on Conceptual Modeling. (Berlin: Springer), 190–199. doi: 10.1007/978-3-642-16385-2_24
Mandal, S., and Maiti, A. (2021). “Graph neural networks for heterogeneous trust based social recommendation,” 2021 International Joint Conference on Neural Networks (IJCNN). (Piscataway, NJ: IEEE), 1–8. doi: 10.1109/IJCNN52387.2021.9533367
Mu, R. (2018). A survey of recommender systems based on deep learning. IEEE Access 6, 69009–69022. doi: 10.1109/ACCESS.2018.2880197
Page, L., Brin, S., Motwani, R., and Winograd, T. (1999). The PageRank Citation Ranking: Bringing Order to the Web. Tech. rep. Stanford, CA: Stanford InfoLab.
Pavlov, M., and Ichise, R. (2007). Finding experts by link prediction in co-authorship networks. FEWS 290, 42–55. doi: 10.5555/2889513.2889517
Ragesh, R., Sellamanickam, S., Lingam, V., Iyer, A., and Bairi, R. (2021). User embedding based neighborhood aggregation method for inductive recommendation. arXiv preprint arXiv:2102.07575. doi: 10.48550/arXiv.2102.07575
Rossi, E., Chamberlain, B., Frasca, F., Eynard, D., Monti, F., and Bronstein, M. (2020). Temporal graph networks for deep learning on dynamic graphs. arXiv preprint arXiv:2006.10637. doi: 10.48550/arXiv.2006.10637
Salton, G., and McGill, M. J. (1983). Introduction to Modern Information Retrieval. New York, NY: McGraw-Hill, Inc.
Sun, J., Zhang, Y., Guo, W., Guo, H., Tang, R., and He, X. (2020). “Neighbor interaction aware graph convolution networks for recommendation,” in Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval, 1289–1298. doi: 10.1145/3397271.3401123
Thompson, W. R. (1933). On the likelihood that one unknown probability exceeds another in view of the evidence of two samples. Biometrika 25, 285–294. doi: 10.1093/biomet/25.3-4.285
Wang, X., He, X., Wang, M., Feng, F., and Chua, T. F. (2019). “Neural graph collaborative filtering,” in Proceedings of the 42nd international ACM SIGIR conference on Research and development in Information Retrieval (Paris), 165–174. doi: 10.1145/3331184.3331267
Wang, X., and Sukthankar, G. (2013). “Link prediction in multi-relational collaboration networks,” in 2013 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining (ASONAM 2013). (Piscataway, NJ: IEEE), 1445–1447. doi: 10.1145/2492517.2492584
Wu, S., Sun, F., Zhang, W., and Cui, B. (2020). Graph Neural Networks in Recommender Systems: A Survey. ACM Computing Surveys (CSUR). doi: 10.1145/1122445.1122456
Wu, Z., Pan, S., Chen, F., Long, G., Zhang, C., and Yu, P. S. (2020). A comprehensive survey on graph neural networks. IEEE Trans. Neural Netw. Learn. Syst. 32, 4–24. doi: 10.1109/TNNLS.2020.2978386
Xu, D., Ruan, C., Korpeoglu, E., Kumar, S., and Achan, K. (2020). Inductive representation learning on temporal graphs. arXiv preprint arXiv:2002.07962. doi: 10.48550/arXiv.2002.07962
Yang, J., and Leskovec, J. (2012). Defining and evaluating network communities based on ground-truth. Knowl. Inform. Syst. 42, 181–213. doi: 10.1007/s10115-013-0693-z
Ying, R., He, R., Chen, K., Eksombatchai, P., Hamilto, W. S., and Leskovec, J. (2018). “Graph convolutional neural networks for web-scale recommender systems,” in Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining (London), 974–983. doi: 10.1145/3219819.3219890
Yu, Q., Long, C., Lv, Y., Shao, H., He, P., and Duan, Z. (2014). Predicting co-author relationship in medical co-authorship networks. PLoS ONE 9, e101214. doi: 10.1371/journal.pone.0101214
Zeldes, Y., Theodorakis, S., Solodnik, E., Rotman, A., Rotman, G., and Friedman, D. (2017). Deep density networks and uncertainty in recommender systems. arXiv preprint arXiv:1711.02487. doi: 10.48550/arXiv.1711.02487
Zhang, J., Shi, X., Zhao, S., and King, I. (2019). Star-gcn: Stacked and reconstructed graph convolutional networks for recommender systems. arXiv preprint arXiv:1905.13129. doi: 10.24963/ijcai.2019/592
Zhang, M., and Chen, Y. (2020). “Inductive matrix completion based on graph neural networks,” in International Conference on Learning Representations.
Zheng, L., Lu, C. T., Jiang, F., Zhang, J., and Yu, P. S. (2018). “Spectral collaborative filtering,” in Proceedings of the 12th ACM Conference on Recommender Systems (Vancouver, BC), 311–319. doi: 10.1145/3240323.3240343
Zhou, J., Cui, G., Hu, S., Zhang, Z., Yang, C., Liu, Z., et al. (2020). Graph neural networks: a review of methods and applications. AI Open 1, 57–81. doi: 10.1016/j.aiopen.2021.01.001
Keywords: graph neural networks (GNN), recommendation systems, collaborator recommendation, deep learning, artificial intelligence
Citation: Zhu J and Yaseen A (2022) A Recommender for Research Collaborators Using Graph Neural Networks. Front. Artif. Intell. 5:881704. doi: 10.3389/frai.2022.881704
Received: 22 February 2022; Accepted: 23 June 2022;
Published: 01 August 2022.
Edited by:
Chanyoung Park, Korea Advanced Institute of Science and Technology, South KoreaReviewed by:
Shuhao Shi, PLA Strategy Support Force Information Engineering University, ChinaSupriyo Mandal, Christian-Albrechts-Universität zu Kiel, Germany
Copyright © 2022 Zhu and Yaseen. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Ashraf Yaseen, Ashraf.yaseen@uth.tmc.edu