Skip to main content

ORIGINAL RESEARCH article

Front. Big Data, 22 December 2021
Sec. Big Data Networks
This article is part of the Research Topic Computational Behavioral Modeling for Big User Data View all 5 articles

Computational Modeling of Hierarchically Polarized Groups by Structured Matrix Factorization

Dachun Sun
&#x;Dachun Sun1*Chaoqi Yang&#x;Chaoqi Yang1Jinyang LiJinyang Li1Ruijie WangRuijie Wang1Shuochao YaoShuochao Yao2Huajie ShaoHuajie Shao3Dongxin LiuDongxin Liu1Shengzhong LiuShengzhong Liu1Tianshi WangTianshi Wang1Tarek F. AbdelzaherTarek F. Abdelzaher1
  • 1Computer Science Department, University of Illinois at Urbana-Champaign, Urbana, IL, United States
  • 2Computer Science Department, George Mason University, Fairfax, VA, United States
  • 3Computer Science Department, William and Mary, Williamsburg, VA, United States

The paper extends earlier work on modeling hierarchically polarized groups on social media. An algorithm is described that 1) detects points of agreement and disagreement between groups, and 2) divides them hierarchically to represent nested patterns of agreement and disagreement given a structural guide. For example, two opposing parties might disagree on core issues. Moreover, within a party, despite agreement on fundamentals, disagreement might occur on further details. We call such scenarios hierarchically polarized groups. An (enhanced) unsupervised Non-negative Matrix Factorization (NMF) algorithm is described for computational modeling of hierarchically polarized groups. It is enhanced with a language model, and with a proof of orthogonality of factorized components. We evaluate it on both synthetic and real-world datasets, demonstrating ability to hierarchically decompose overlapping beliefs. In the case where polarization is flat, we compare it to prior art and show that it outperforms state of the art approaches for polarization detection and stance separation. An ablation study further illustrates the value of individual components, including new enhancements.

1 Introduction

Extending a previous conference publication (Yang et al., 2020), this paper solves the problem of unsupervised computational modeling of hierarchically polarized groups. The model can accept, as input, a structural guide to the layout of groups and subgroups. The goal is to uncover the beliefs of each group and subgroup in that layout, divided into points of agreement and disagreement among the underlying communities and sub-communities, given their social media posts on polarizing topics. Most prior work clusters sources or beliefs into flat classes or stances (Küçük and Can, 2020). Instead, we focus on scenarios where the underlying social groups disagree on some issues but agree on others (i.e., their beliefs overlap). Moreover, we consider a (shallow) hierarchical structure, where communities can be further subdivided into subsets with their own agreement and disagreement points.

Our work is motivated, in part, by the increasing polarization on social media (Liu, 2012). Individuals tend to connect with like-minded sources (Bessi et al., 2016), ultimately producing echo-chambers (Bessi et al., 2016) and filter bubbles (Bakshy et al., 2015). Tools that could automatically extract social beliefs, and distinguish points of agreement and disagreement among them, might help generate future technologies (e.g., less biased search engines) that summarize information for consumption in a manner that gives individuals more control over (and better visibility into) the degree of bias in the information they consume.

The basic solution described in this paper is unsupervised. However, it does accept guidance on group/subgroup structure. Furthermore, the solution has an option for enhancement using prior knowledge of language models. By unsupervised, therefore, we mean that the (basic) approach does not need prior training, labeling, or remote supervision. This is in contrast, for example, to deep-learning solutions (Irsoy and Cardie, 2014; Liu et al., 2015; Wang et al., 2017) that usually require labeled data. The structural guidance, in this paper, is not meant to be obtained through training. Rather, it is meant as a mechanism for an analyst familiar with the situation to enter a template to match the inferred groups against. For example, the analyst might have the intuition that the community is divided into two conflicted factions, of which one is further divided into two subgroups with partial disagreement. They might be interested in understanding the current views of each faction/subgroup. The ability to exploit such analyst guidance (on the hierarchy of disagreement) is one of the distinguishing properties of our approach. In the absence of analyst intuitions, it is of course possible to skip structural guidance, as we show later in the paper. The basic algorithm can be configured so that it does not need language-specific prior knowledge (Liu, 2012; Hu Y. et al., 2013), distant-supervision (Srivatsa et al., 2012; Weninger et al., 2012), or prior embedding (Irsoy and Cardie, 2014; Liu et al., 2015), essentially making it language-agnostic. Instead, it relies on tokenization (the ability to separate individual words). Where applicable, however, we can utilize a BERTweet (Nguyen et al., 2020) variant that uses a pre-trained Tweet embedding to generate text features, if higher performance is desired. BERTweet is a specific language model for Tweets with the same structure as in the Bidirectional Encoder Representations from Transformers (BERT) (Devlin et al., 2019). While we test the solution only with English text, we conjecture that the its application can be easily extended to other languages (with the exception of those that do not have spaces between words, such as Chinese and Japanese, because we expect spaces as token separators). An advantage of unsupervised techniques is that they do not need to be retrained for new domains, jargon, or hash-tags.

The work is a significant generalization of approaches for polarization detection (Conover et al., 2011; Demartini et al., 2011; Bessi et al., 2016; Al Amin et al., 2017) that identify opposing positions in a debate but do not explicitly search for points of agreement. The unsupervised problem addressed in this paper is also different from unsupervised techniques for topic modeling (Litou and Kalogeraki, 2017; Ibrahim et al., 2018) and polarity detection (Al Amin et al., 2017; Cheng et al., 2017). Prior solutions to these problems aim to find orthogonal topic components (Cheng et al., 2017) or conflicting stances (Conover et al., 2011). In contrast, we aim to find components that adhere to a given (generic) overlap structure, presented as structural guidance from the user (e.g., from an analyst). Moreover, unlike solutions for hierarchical topic decomposition (Weninger et al., 2012; Zhang et al., 2018), we consider not only message content but also user attitudes towards it (e.g., who forwards it), thus allowing for better separation, because posts that share a specific stance are more likely to overlap in the target community (who end up spreading them).

This paper extends work originally presented at ASONAM 2020 (Yang et al., 2020). The extension augments the original paper in several aspects. First, we provide options for integrating a language model to improve outcomes. In this version, besides tokenization, we use language models that boost performance, compared to a purely lexical overlap-based approach. Second, we derive a new orthogonality property for the factorized components by our model. Finally, we conduct a simulation and new experiments that additionally verify our model in multiple scenarios involving both a flat group structure, and a hierarchical structure with complex sub-structure.

The work is evaluated using both synthetic data as well as real-life datasets, where it is compared to approaches that detect polarity by only considering who posted which claim (Al Amin et al., 2017), approaches that separate messages by content or sentiment analysis (Go et al., 2009), approaches that identify different communities in social hypergraphs (Zhou et al., 2007), and approaches that detect user stance by density-based feature clustering (Darwish et al., 2020). The results of this comparison show that our algorithm outperforms the state of the art. An ablation study further illustrates the impact of different design decisions on accomplishing this improvement.

The rest of the paper is organized as follows. Section 2 formulates the problem and summarizes the solution approach. Section 3 proposes our new belief structured matrix factorization model, and analyzes some model properties. Section 4 proves the property of orthogonality regarding one of the decomposed components. Section 5 and Section 6 present an experimental evaluation based on simulation and real data, respectively. We review the related literature on belief mining and matrix factorization in Section 7. The paper concludes with key observations and a statement on future directions in Section 8.

2 Problem Formulation

Consider an observed data set of posts collected from a social medium, such as Twitter, where each post is associated with a source and with semantic content, called a claim. Let S be the set of sources in our data set, and C be the set of claims made by those sources. While, in this paper, a claim is the content of a tweet (or retweet), the analytical treatment does not depend on this interpretation. Let matrix X, of dimension |S|×|C| be a matrix of binary entries denoting who claimed what. A claim can be made by multiple sources. In general, an entry of X could be a positive real number indicating a level of endorsement of a source to a claim. In this paper, we simplify by using binary entries; if source Si endorsed claim Cj, then xij = 1, otherwise xij = 0.

2.1 Problem Statement

We assume that the set of sources, S, is divided into a small number, K, of latent social groups, denoted by the subsets G1 G2, …, GK, that form a tree. In this tree, group Gi is a child (i.e., a subgroup) of Gk if GiGk. Children of the same parent are disjoint groups. Members of group Gk that do not belong to any of its children are denoted by the residual set Gk. Within each group, Gk, 1 ≤ kK, individuals have shared beliefs expressed by a set of claims. A shared belief of a group is a belief espoused by all members of the group. By definition, therefore, a child group inherits the shared beliefs of its parent. The child group may have additional shared beliefs within its group (not shared by the remaining members of the parent). Thus, we define the incremental belief set, Bk, of group Gk to be the beliefs held by group Gk, beyond what’s inherited from its parent. The overall belief set of group Gk is thus the union of incremental beliefs of its ancestors and itself. The problem addressed in this paper is to simultaneously 1) populate the membership of all latent groups, Gk, in the tree, and 2) populate their incremental belief sets, Bk, given a structural template, B (to be populated), that lays out the latent groups and latent belief sets.

Figure 1 illustrates an example, inspired by the first wave of the COVID-19 pandemic in 2020. In this figure, a hypothetical community is divided on whether to maintain social distancing (G1) or reopen everything and let natural selection take place (G2). Furthermore, while G1 agree on social distancing, they disagree on some implementation details, such as whether classes should be entirely online (G3) or hybrid (G4). In this example, an analyst might want to understand what groups/subgroups exist and what they disagree on. The analyst will enter a tree topology template, letting the algorithm populate who and which beliefs lie at which node of the tree.

FIGURE 1
www.frontiersin.org

FIGURE 1. A notional example of hierarchical overlapping beliefs.

2.2 Solution Approach

We use a non-negative matrix factorization algorithm to decompose the “who endorsed what” matrix, X{0,1}|S|×|C|, into 1) a matrix, UR+|S|×K, that maps sources to latent groups, 2) a matrix, B ∈ {0,1}K×K, that maps latent groups to latent incremental belief sets (called the belief structure matrix), entered as structural guidance from the analyst, and 3) a matrix, MR+|C|×K, that maps claims to latent incremental belief sets. Importantly, since groups and belief sets are latent, the belief structure matrix, B, in essence, specifies the latent structure of the solution space that the algorithm needs to populate with specific sources and claims, thereby guiding factorization.

3 Structured Matrix Factorization

The proposed structured matrix factorization algorithms allows the user (e.g., an analyst) to specify matrix, B, to represent the relation between the latent groups, Gk (that we wish to discover), and their incremental belief sets, Bk (that we wish to discover as well). An element, bij of the matrix, B, is 1 if group Gi adopts the belief set Bj. Otherwise, it is zero. In a typical (non-overlapping) clustering or matrix factorization framework, there is an one-to-one correspondence between groups and belief sets, reducing B to an identity matrix. Structured matrix factorization extends that structure to an arbitrary relation. Matrix B can be thought of as a template relating latent groups (to be discovered) and belief sets (to be identified). It is a way to describe the structure that one wants the factorization algorithm to populate with appropriate group members and claims. While it might seem confusing to presuppose that one knows the latent structure, B, before the groups and belief sets in question are populated, below we show why this problem formulation is very useful.

3.1 An Illustrative Example

Consider a conflict involving two opposing groups, say, a minority group G1 and a majority G2. Their incremental belief sets are denoted by B1 and B2, respectively. The two groups disagree on everything. Thus, sets B1 and B2 do not overlap. An unfriendly agent wants to weaken the majority and conjectures that the majority group might disagree on something internally. Thus, the unfriendly postulates that group G2 is predominantly made of subgroups G2a and G2b. While both subgroups agree on the shared beliefs, B2, each subgroup has its own incremental belief sets, B2a and B2b, respectively. The structure matrix in Figure 2 represents the belief structure postulated above.

FIGURE 2
www.frontiersin.org

FIGURE 2. The belief structure matrix, B.

For example, the second column indicates that the belief set B2 is shared by all members of group G2 (hence, there is a “1” in rows of subgroups G2a, G2b, and the residual G2). The third and fourth columns state that belief sets B2a and B2b are unique to subgroups G2a and G2b, respectively. Note how the beliefs espoused by different groups overlap. For example, from the third and fourth rows, we see that groups G2a and G2b overlap in the set of beliefs, B2.

An interesting question might be: which sources belong to which group/subgroup? What are the incremental belief sets B2a and B2b that divide group G2 (i.e., are shared only by the individual respective subgroups)? What are the shared beliefs B2 that unite it? What are the beliefs, B1, of group G1? These are the questions answered by our structured matrix factorization algorithm whose input is (only) matrix, X, and matrix, B (Figure 2).

3.2 Mathematical Formulation

To formulate the hierarchical overlapping belief estimation problem, we introduce the notion of claim endorsement. A source is said to endorse a claim if the source finds the claim agreeable with their belief. Endorsement, in this paper, represents a state of belief, not a physical act. A source might find a claim agreeable with their belief, even if the source did not explicitly post it. Let the probability that source Si endorses claim Cj be denoted by   Pr(SiCj). We further denote the proposition SiGp by Sip, and the proposition CjBq by Cjq. Thus, Pr(Sip) denotes the probability that source SiGp. Similarly, Pr(Cjq) denotes the probability that claim CjBq. Following the law of total probability:

PrSiCj=p,qPrSiCj|SipCjqPrSipCjq=p,qPrSiCj|SipCjqPrSipPrCjq.(1)

By definition of the belief structure matrix, B, we say that Pr(SiCj|SipCjq)=1 if bpq = 1 in the belief structure matrix. Otherwise, Pr(SiCj|SipCjq)=0. Let uip=Pr(Sip) and mjq=Pr(Cjq). Let ui and mj be the corresponding vectors, with elements ranging over values of p and q respectively. Thus, we get: Pr(SiCj)=uiBmj. Let the matrix XG be the matrix of probabilities,  Pr(SiCj), such that element xijG=Pr(SiCj). Thus:

XG=UBM(2)

where U is a matrix whose elements are uip and M is a matrix whose elements are mjq. Factorizing XG, given B, would directly yield U and M, whose elements are the probabilities we want: elements of matrix U yield the probabilities that a given source Si belongs to a given group Gp, whereas elements of matrix M yield the probabilities that a claim Cj belongs to a belief Bq. Each source is then assigned to a group and each claim to a belief set, based on the highest probability entry in their row of matrix U and M, respectively. In practice, B could be customized up to certain tree depth to meet different granularity of belief estimation.

3.3 Estimating XG

Unfortunately, we do not really have matrix XG to do the above factorization. Instead, we have the observed source-claim matrix X that is merely a sampling of what the sources endorse. (It is a sampling because a silent source might be in agreement with a claim even if they did not say so.) Using X directly is suboptimal because it is very sparse. It is desired to estimate that a source endorses a claim even if the source remains silent. We do so on two steps.

3.3.1 Message Interpolation (the M-Module)

First, while source Si might have not posted a specific claim, Cj, it may have posted similar ones. If a source Si posted, retweeted, or liked claim Cj in our data set (i.e., xij = 1 in matrix X), then we know that the source endorses that claim (i.e., xijM=1 in matrix XM). The question is, what to do when xij = 0? In other words, we need to estimate the likelihood that the source endorses a claim, when no explicit observations of such endorsement were made. We do so by considering the claim similarity matrix D, of dimension |C|×|C|. If source Si was observed to endorse claims Ck similar to Cj, then it will likely endorse Cj with a probability that depends on the degree of similarity between Cj and Ck. Thus, when xij = 0, we can estimate xijM by weighted sum interpolation:

xijM=k:xik0dkjxik(3)

To compute matrix D, in this work, we provide two configurations. For the language agnostic approach, we first compute a bag-of-words (BOW) vector wj for each claim j, and then normalize it using vector L2 norm, wj̄=wj/wj2. For the BERTweet approach, wj is the embedding for each claim. We select non-zero entries xij in each row i of X as medoids {wj̄xij0}. We assume that claims close to any of the medoids could also be endorsed by Si as well. Based on that, we use dkj=ϕ(wj̄wk̄) in Eq. 3. A Gaussian radial basis function (RBF) is used for ϕ(r)=e(ϵr)2. This is called message interpolation (the M-module). The output of this module is a matrix XM. If the resulting value of xijM is less than 0.2, we regard that it is far from all of the medoids and set it back to 0. In the experiments presented in our evaluation, ϵ is set to 0.5 for the synthetic dataset and 0.05 for each of the United States Election 2020, Eurovision 2016, and Global Warming datasets.

3.3.2 Social Graph Convolution (S-Module)

To further improve our estimation of matrix XG, we assume that sources generally hold similar beliefs to those in their immediate social neighborhood. Thus, we perform a smoothing of matrix XM by replacing each cell, xij, by a weighted average of itself and the entries pertaining to neighbors of its source, Si, in the social graph. Let matrix A, of dimension |S|×|S|, denote the social graph. Each entry, aij, denotes the influence of user Si on user Sj. A is thus the adjacency matrix of a social network graph. In this paper, we construct A by calculating the frequency of each source Si retweeting posts of source Sj. We call it the retweet graph. The update of entries of XM by smoothing over entries of neighboring sources is called social graph convolution (S-module). It results in an improved estimate, XMS.

More specifically, from the social dependency matrix A (user-user retweet frequency), we can compute the degree matrix F by summing each row of A. The random walk normalized adjacency is denoted as Arw̃=F1A. We define our propagation operator based on Arw̃ with self-loop re-normalization, Arw̄12Frw̃1(Arw̃+I). Thus, the new source-claim network is given by,

XMS=Arw̄XM,(4)

where each row of Arw̄ adds up to 1. The effect of the propagation operator is to convolve the information from 1-hop neighbors, while preserving half of the information from itself. Note that, we deem dependency beyond 1-hop too weak to capture, so we do not consider An, where n > 1. From a macroscopic perspective, this social graph convolution recovers some of the possible source-claim connections and also enforces the smoothness of matrix XMS. We can now take, XGXMS, and decompose it as presented in Section 3.2, Eq. 2, with L1 and L2 regularizations to enforce sparsity and prevent overfitting.

3.4 Loss and Optimization

Given a belief mixture matrix, B, we now factorize XMS to estimate matrices U and M that decide the belief regions associated with sources and claims, respectively. (e.g., the estimated belief for claim Cj is given by the index of maximum entry in the jth row of M).

Regularization. To avoid model overfitting, we include widely used L2 regularization. Also, we enforce the sparsity of U and M by introducing an L1 norm. The overall objective function becomes (defined by the Frobenius-norm),

J=XMSUBMF2+λ1UF2+λ1MF2+λ2U1+λ2M1.(5)

We rewrite J using matrix trace function tr(⋅)

J=trXMSXMS2trXMSUBM+trUBMMBU+λ1trUU+λ1trMM+λ2U1+λ2M1.(6)

We minimize J by gradient descent. Since only the non-negative region is of interest, derivatives of the L1 norm are differentiable in this setting. By referring to the gradient of traces of product with constant matrix A, ∇Xtr (AX) = A and ∇Xtr (XAX) = X (A+ A), the partial derivative of J w.r.t. U and M are calculated as,

U=2XMSMB+2UBMMB+2λ1U+λ21,M=2XMSUB+2MBUUB+2λ1M+λ21.

The gradient matrix ∇U is of dimension |S|×K, and ∇M is of dimension |C|×K. Estimation step begins by updating UUηU and MMηM, and η is the step size. Negative values might appear in the learning process, which are physically meaningless in this problem. Thus, we impose non-negativity constraints for U and M during the update. A ReLU-like strategy is utilized: when any entry of U or M becomes negative, it is set to be ξ. In the experiment, we set ξ = 10−8, λ1 = λ2 = 10−3. Note that the initial entry values of U and M are randomized uniformly from (0, 1).

www.frontiersin.org

Algorithm 1.: Belief Structured Matrix Factorization (BSMF).

3.4.1 Run-time Complexity

During the estimation, we generalize standard NMF multiplicative update rules (Lee and Seung, 2001) for our tri-factorization,

ηU=12UUBMMB,ηM=12MMBUUB.(7)

Note that, K (number of belief groups) is picked according to the dataset, and it typically satisfies Kmin(|S|,|C|). Algorithmically, updating U and M takes OK|SC| per iteration, similar to the typical NMF. The number of iterations before the empirical convergence is usually no more than 200 for random initialization in our experiments, and thus we claim that our model is scalable and efficient. In Figure 3, we present our measurements on time taken per iteration for our algorithm and FNMTF (Wang et al., 2011), which is known to be able to perform tri-factorization fast. We measured each iteration 20 times to make the results more accurate. Our algorithm, with the generalized multiplicative update rules, performs better. This is because our algorithm does not need to update the B matrix, and avoids using matrix inversion. The graph appears to be consistent with our time complexity analysis.

FIGURE 3
www.frontiersin.org

FIGURE 3. Running time under different configurations of dimensions of X and B. For each method, under the same |S|, |C| takes 10000, 20000, 30000, 40000, 50000 respectively. Red bar indicates standard deviation on time measurements.

4 Property of Orthogonality

The basic idea of belief structured matrix factorization (BSMF) is to decompose XMSUBM. What does that imply regarding the generated rows of decomposition matrices? Below, we show that the rows of matrix U are approximately orthogonal and that the rows of matrix M are also approximately orthogonal. This property suggests a sense of decomposition quality in that the latent factors produced constitute more efficient (non-redundant) encoding bases of the original information.

Theorem 1. After BSMF factorization, {Uk} are approximately orthogonal.

PROOF. We start the analysis by the recapitulation of loss function,

J=XMSUBMF2+λ1UF2+λ1MF2+λ2U1+λ2M1.=trXMSXMS2trXMSUBM+trUBMMBU+λ1trUU+λ1trMM+λ2U1+λ2M1.(8)

Let us first ignore the L1 and L2 term. Since tr(XMSXMS) is a constant, the problem of minimizing Eq. 8 is equivalent to,

maxU,M0trXMSUBM,(9)
andminU,M0trUBMMBU.(10)

The first objective, Eq. 9, is similar to K-means type clustering after expanding the trace function, which maximize within-cluster similarities,

maxi,juiBmjxij.(11)

The second objective is to enforce orthogonality approximately. Because UBMXMS, let us add L2-norm of U and M now, so that the scale of U, M are constrained Wang and Zhang (2012). Since B is positive and fixed, tr(UU)=UF2 and tr(BMMB)=MBF2MF2BF2 are approximately constants. So Eq. 10 becomes,

minU,M0trUUIBMMBI.(12)

By Cauchy’s inequality, we further have,

trUUIBMMBI=i=1,,K;j=1,,KUUIijBMMBIjiUUIFBMMBIF.(13)

Thus, the overall second objective is bounded by,

minU,M0UUIFBMMBIFminU0UUIFminM0BMMBIF.(14)

which enforces the orthogonality of rows in set {Uk}.

Corollary 1.1. After BSMF factorization, {Mk} are approximately orthogonal.

PROOF. Similar to the proof above, tr(MM)=MF2 and tr(BUUB)=UBF2UF2BF2 are approximately constants. Equation 10 becomes,

minU,M0trMMIBUUBI.(15)

Thus, the overall second objective is bounded by,

minM0MMIFminU0BUUBIF.(16)

which enforces the orthogonality of pre-belief bases {Mk}.In other words, the generated groups into which all sources and claims decomposed are maximally diverse and are not redundant in some sense.

5 Illustrative Simulation

To help visualize the belief structure developed in our approach and offer a concrete example of approximate orthogonality of groups (i.e., pre-belief bases), we created simulations to evaluate our models multiple times against two simpler variants. We also provide 3D visualizations of the factorized components to observe their spatial characteristics.

5.1 Dataset Construction

To offer a simplified and controllable setting in which we can visualize the novel factorization algorithm, we build a synthetic dataset, where two groups of users are created, a minority, G1 (100 users) of belief set B1, and a majority, G2 (300 users) of belief set B2. The majority includes two subgroups, G2a and G2b (100 users each) of incremental belief sets B2a and B2b, respectively. Essentially, the groups follow the hierarchical structure illustrated in Figure 2. For each group, we built disjoint claim corpora, denoted by c1, c2, c2a and c2b to express their respective belief sets. Users were simulated to post claims chosen from their group’s assigned corpus or from their parent’s corpus (we randomly generate 20 claims for each user). Thus, for example, users in group G2a could post claims generated from c2a or from the parent corpus c2, but users in group G1 only post claims from corpus c1. In sum, 400 users and 8,000 claims were created. To keep it simple, in this experiment, we do not impose social relations. Instead, we use the identity matrix for the adjacency A.

5.2 Method Comparison

The factorization algorithm uses the belief structure matrix in Figure 2. Two simpler variants are introduced: 1) the first variant substitutes B with an identity matrix, and takes a standard NMF formulation XG = UM; 2) the second variant substitutes B with an learnable matrix B̃, which takes a standard non-negative matrix tri-factorization (NMTF) form, XG=UB̃M. Obviously, NMTF offers more freedom. However, the need to learn parameters of matrix B̃ can cause overfitting. We use the same regularization settings for NMF, NMTF and our BSMF to make sure the comparison is fair. Empirically, after 150 ∼ 200 iterations, all three methods converge. The predicted belief set label for each claim is given by the index of the maximum value in this final representation from matrix, M.1

5.3 Results of 200 Rounds

We run each model for 200 times and compute the classification accuracy of users and claims for each algorithm in every run, then average them over all runs. We find that BSMF consistently outperforms NMF and NMTF. The average values of accuracy for BSMF, NMF and NMTF are 97.34%, 93.78%, 95.54%, respectively. As might be expected, specifying matrix, B, guides subsequent factorization to a better result compared to both NMF and NMTF.

5.4 Visualization of Results

We also visualize the computed matrix, M, for each algorithm. Colors are based on ground-truth labels. In Figure 4 and Figure 5, we project the estimated M into a 3-D space, where each data point represents a message. In each figure, all of the data points seem to lie in a regular tetrahedron (should be regular K-polyhedron for more general K-belief cases). It is interesting that for NMF, different colors are very closely collocated in the latent space (e.g., there is very little separation between the grey color and others). It is obviously difficult to draw a boundary for the crowded mass. NMTF is a little bit better: different colors are visually more separable. We also visualize the learned B̃R4×4, which turns out to be an SVD-like diagonal matrix, meaning that pure NMTF only learns the independent variances aligned with each belief.

FIGURE 4
www.frontiersin.org

FIGURE 4. Visualization of M for our BSMF (left) and NMF (right).

FIGURE 5
www.frontiersin.org

FIGURE 5. Visualization of M (left) and B̃ (right) for NMTF.

The projection result of BSMF are different: different color points are better separated and grouped by colors. We hypothesize that in the four-dimensional space, data points should be perfectly aligned with one of the belief bases/parts, and these four bases are conceivably orthogonal in that space. In a word, the results on synthetic data strongly suggest that our model disentangles the latent manifold leading to a better separation of messages by belief sets.

6 Experiments With Empirical Data

In the section, we evaluate Belief Structured Matrix Factorization (BSMF) using real-world Twitter datasets of different patterns of belief overlap, hierarchy, or disjointedness. Key statistics of these data sets are briefly summarized in Table 1. Our model is compared to (up to) six baselines and five model variants. Each experiment is run 20 times to acquire the mean and the standard deviation for measurements. We elaborate the experimental settings and results below.

TABLE 1
www.frontiersin.org

TABLE 1. Basic statistics for three twitter datasets.

6.1 Disjoint Belief Structure: United States Election 2020

We start with a dataset where we posit that agreements are too weak to the point where beliefs of key groups are assumed to be “disjoint” (i.e., share no overlap). Specifically, during the 2020 United States election, messages on social media tended to be very polarized, either supporting the democratic or the republican party, exclusively. The example demonstrates a simplest case of belief factorization for clarity.

6.1.1 Dataset

We use Apollo Social Sensing Toolkit2 to collect the United States Election 2020 dataset. The dataset contains tweets collected during the United States Election in 2020, where the support for candidates was split between former president Donald Trump and president Joe Biden. Basic statistics are reported in Table 1. Overall, the most retweeted 237 tweets from the dataset are manually annotated, which separate into 133 pro-Trump and 104 anti-Trump claims.

In this dataset, the sources are split into two groups G1 and G2, each with belief sets, B1 and B2, respectively. The belief structure matrix, therefore, is:

B=1001

where rows correspond to groups G1 and G2, respectively, and columns correspond to belief sets, B1 and B2, respectively.

6.1.2 Baselines

We select six baseline methods that encompass different perspectives on belief separation:

• Random: A trivial baseline that annotates posts randomly, giving equal probability to each label.

• DBSCAN (Darwish et al., 2020): A density-based clustering technique that extracts user-level features and performs DBSCAN for unsupervised user stance detection in Twitter. In this paper, we use DBSCAN and then map the user stance to claim stances with majority voting, as a baseline.

• Sentiment140 (Go et al., 2009): Content-aware solutions based on language or sentiment models. In the implementation, each of the claims is a query through Sentiment140 API, which responds with a polarity score. The API will respond with the same score upon multiple same requests.

• H-NCut (Zhou et al., 2007): The method views the bipartite structure of the source-claim network as a hypergraph, where claims are nodes and sources are hyperedges. The problem is thus seen as a hypergraph community detection problem, where community nodes represent posts. We implement H-NCut, a hypergraph normalized cut algorithm.

• Polarization (Al Amin et al., 2017): A baseline that uses an NMF-based solution for social network belief extraction to separate biased and neutral claims.

• NMTF: A baseline with a learnable mixture matrix. We compare our model with it to demonstrate that pure learning without a prior is not enough to unveil the true belief overlap structure in real-world applications.

• FNMTF (Wang et al., 2011): A baseline for data co-clustering with non-negative matrix tri-factorization. We compare to this model mainly to have a run-time complexity comparison and demonstrate the importance of structural guidance.

Different variants of BSMF are also evaluated to verify the effectiveness of message similarity interpolation (the M-module) and social graph convolution (the S-module). BSMFMS incorporates both modules. Models with only the M-module or the S-module are named BSMFM and BSMFS, respectively. BSMF denotes the model without either module. BSMFMSBERT and BSMFMBERT are two variants whose M-modules are configured to use BERTweet (Nguyen et al., 2020) embeddings, while BSMFMS and BSMFM are using lexical overlap enabled by tokenization.

6.1.3 Evaluation Metrics

We evaluate claim separation, since only claim labels are accessible. We use the Python scikit-learn package to help with the evaluation. Multiple metrics are employed. Since we only have two groups in the dataset, we use binary-metrics to calculate precision, recall and f1-score over the classification results, and we also use weighted-metrics to account for class imbalance by computing the average of metrics in which each class score is weighted by its presence in the true data sample. Standard precision, recall and f1-score are considered in both scenarios. Note that weighted averaging may produce an f1 that is not between precision and recall.

6.1.4 Result of United States Election 2020

The comparison results are shown in Table 2. It is not surprising that all baselines beat Random. Overall, matrix factorization methods work well for this problem. Among other baselines, Sentiment140 work poorly for this task, because 1) they use background language models that are pre-trained on another corpus; and 2) they do not user dependency information, which matters in real-world data. H-NCut and DBSCAN yield acceptable performance, but cannot compare with our BSMF algorithm with S-module, since they ignore the user dependencies. Considering weighted scores, NMTF outperforms the NMF-based algorithm, which is as expected. With the S-module, our BSMF algorithm ranks the top in terms of all metrics. Comparing to other variants, M-module in this dataset does not add benefit, mostly because several important keywords such as “president”, “Trump”, and “election” are shared by both sides. Therefore, variants using content similarity may experience confusion and not perform well. This is especially true of variants that use lexical similarity although variants that use BERTweet embeddings also suffer compared to BSMFS.

TABLE 2
www.frontiersin.org

TABLE 2. Binary- and weighted- metrics comparison (United States election 2020).

For illustrative purposes (to give a feel for the data), Table 3 shows the top 3 tweets from each belief set (B1 and B2) estimated by our model. Note that, due to an update of the Twitter API, the crawled text field is truncated to 140 characters. Our algorithm runs on the text within that range only. For human readability and interpretability, we manually fill in the rest of the tweet, showing the additional text in yellow (the same for Table 5 and Table 7). Note that, the labels shown in the first column, called Beliefs, are inserted manually after the fact (and not by our algorithm). The algorithm merely does the separation/clustering.

TABLE 3
www.frontiersin.org

TABLE 3. Top 3 tweets from separated beliefs (United States election 2020).

6.2 Star-like Belief Structure: Eurovision2016

Next, we consider a data set where we break sources into two subgroups but assume that they have overlapping beliefs. The example facilitates comparison with the prior state of the art on polarization that addresses a flat belief structure.

6.2.1 Dataset

We use the Eurovision2016 dataset, borrowed from (Al Amin et al., 2017). Eurovision2016 contains tweets about the Ukrainian singer, Jamala, who won the Eurovision (annual) song contest in 2016. Her success was a surprise to many as the expected winner had been from Russia according to pre-competition polls. The song was on a controversial political topic, telling a story about deportation of Crimean Tatars by Soviet Union forces in the 1940s. Tweets related to Jamala were collected within 5 days of the contest. Basic statistics are reported in Table 1. As pre-processed in (Al Amin et al., 2017), the most popular 1,000 claims were manually annotated. They were separated into 600 pro-Jamala, 239 anti-Jamala, and 161 neutral claims.

In the context of the dataset, the entire set of sources is regarded as a big group, G1, with belief set, B1, agreed among all users. The group is further divided into three disjoint groups, group G1a (with inherited belief set B1 and incremental belief set B1a), group G1b (with inherited belief set B1 and incremental belief set B1b), and the residual group G1 (with belief set B1). In this case, the belief structure is:

B=100110101

where rows correspond to groups G1, G1a, and G1b, respectively, whereas columns correspond to belief sets, B1, B1a and B1b respectively.

6.2.2 Baselines

We use the same baselines as in Section 6.1.2. Similarly, we also use all variants of BSMF algorithms.

6.2.3 Evaluation Metrics

In this dataset, we evaluate claim separation using metrics described in Section 6.1.3. Instead of binary-evaluation, we use macro-evaluation because there are three groups in this dataset, as opposed to only two (in the previous subsection). This metric calculates the mean of the metrics, giving equal weight to each class. It is used to highlight model performance of infrequent classes. Still, note that weighted averaging may produce an f1-score that is not between precision and recall.

6.2.4 Result of Eurovision2016

The comparison results are shown in Table 4. Similar to the results of United States Election 2020, all baselines beat Random, and matrix factorization methods work okay for this problem, but not as good. Sentiment140 still works poorly for the same reason as before. H-NCut and DBSCAN yield much weaker performance than before, likely because they fail to adequately consider the underlying overlapping belief structure. NMTF outperforms the NMF-based algorithm. The reason may be that its additional freedom allows it to capture the underlying structure better. With both the M-module and S-module, our BSMF algorithm ranks the top in all metrics. Both modules help in this experiment. We believe that it is due to the specific star-like belief structure and the existence of a residual group who beliefs can be better inferred by message interpolation. Further, we see that BERTweet variants perform better than lexical M-modules on several metrics in this case, because BERTweet is more language focused. As before, for illustration and to give a sense of the data, Table 5 shows the top 3 tweets from each belief set (B1, B1a, B1b) estimated by our model.

TABLE 4
www.frontiersin.org

TABLE 4. Macro- and weighted- metrics comparison (eurovision 2016).

TABLE 5
www.frontiersin.org

TABLE 5. Top 3 tweets from separated beliefs (eurovision 2016).

6.3 Joint Polarity and Topic Separation

Before we test our algorithm on a dataset with a hierarchical belief structure, we test it on something it is not strictly designed to do: namely, joint separation of topic and polarity. This problem may arise in instances where multiple interest groups simultaneously discuss different topics, expressing different opinions on them. We check whether our algorithm can simultaneously separate the different interest groups and their stances on their topics of interest. To do so, we artificially concatenated tweets from Eurovision2016 and the United States Election 2020 datasets. This operation creates a virtual hierarchy where each original topic is viewed as a different interest group. Inside each group, there are sub-structures as introduced before. Accordingly, we apply our algorithm on the following belief structure:

B=1000011000101000001000001

where rows correspond to groups G1, G1a, G1b, Gα, and Gβ respectively, whereas columns correspond to belief sets, B1, B1aB1b, Bα and Bβ respectively.

6.3.1 Evaluation Metrics

We first evaluate the accuracy of the dataset separation by collapsing the label to each 0 and 1 corresponding to two datasets. Then, for each dataset, we employ Macro-f1 score and Weighted-f1 score on Eurovision2016 dataset, Binary-f1 score and Weighted-f1 score on United States Election 2020 dataset.

6.3.2 Result of Separation

The comparison results are shown in Table 6. All BSMF variants, specifically those with M-module and S-module, performed well on separation of two datasets. In addition, comparing the f1-scores to the performance of the same variant in to Table 2 and Table 4, we are pleased to discover the f1-scores have not deteriorated, demonstrating the basic ability of our model to perform hierarchical belief estimation.

TABLE 6
www.frontiersin.org

TABLE 6. Separation accuracy and macro-/weighted- F1 scores comparison.

6.4 Hierarchical Belief Structure: Global Warming

We consider a real hierarchical scenario in this section, with a majority group, G1, and a minority group, G2, of beliefs B1 and B2 that do not overlap. Since more data is expected on G1 (by definition of majority), we opt to further divide it into subgroups G1a and G1b, who (besides believing in B1) hold the incremental belief sets B1a and B1b, respectively. The corresponding belief structure is reflected by the belief matrix:

B=1000110010100001

where rows represent groups G1, G1a, G1b, and G2 and columns represent the belief sets B1, B1a, B1b, and B2. The matrix is not an identity matrix because beliefs overlap (e.g, groups G1, G1a, and G1b share belief B1). It also features a hierarchical subdivision of G1, into G1, G1a, and G1b.

6.4.1 Datasets

We apply this belief structure to an unlabeled dataset, Global Warming, which is crawled in real time with the Apollo Social Sensing Toolkit. This dataset is about a twitter discussion of global warming in the wake of Australia wildfires that ravaged the continent, in September 2019, where at least 17.9 million acres of forest have burned in the fire. Our goal is to identify and separate posts according to the above abstract belief structure.

Table 7 shows the algorithm’s assignment of claims to belief groups (only the top 3 claims are shown for space limitations). The first column shows the abstract belief categories B1, B1a, B1b, and B2. While the algorithm allocates posts to categories based on the structure of matrix B, for readability, we manually inspect posts assigned to each category in the matrix, and give that category a human-readable name, also shown in the first column. For each belief category, the table also shows the top ranked statements. The table reveals that sources in our data set are polarized between a group, G1, that believes in global warming (offering statements that urge a serious response) and a group, G2, that does not (offering statements that oppose the thesis of man-made global warming). Within group, G1 (apart from G1), there are two subgroups, G1a and G1b. The former blames the fossil fuel industry, whereas the latter is concerned with rising sea levels. While we do not claim to have reached conclusions on global warming, the table shows how structured matrix factorization can fit data sets automatically to useful belief structures, thereby offering visibility into what individuals are concerned with, what actions they agree on, and what they disagree about.

TABLE 7
www.frontiersin.org

TABLE 7. Top 3 tweets from separated beliefs (global warming).

6.4.2 Quantitative Measurements

Next, we do a sanity check by measuring user grouping consistency. Specifically, we first identify the belief sets (by claim separation) and then assign belief labels to users by having a user inherit the assigned belief set label for each claim they made. The inherited labels are inconsistent if they belong to different groups according to matrix, B. For example, if the same user has been assigned belief labels B1 and B1a, then the labeling is coherent because both represent beliefs of G1 (remember that a group inherits the beliefs of its parent). If another user is labeled with both B1 and B2, then it is apparently wrong, since belief sets B1 and B2 belong to different groups.

The percentage of coherently labeled users was 96.08%. Note that, we do not conduct comparison in this dataset, since most baselines do not uncover hierarchical group/belief structures, whereas those that do generally break up the hierarchy differently (e.g., by hierarchical topic, not hierarchical stance) thus not offering an apples to apples comparison. In future work, we shall explore more comparison options.

7 Related Work

The problem of modeling social groups has been long researched. For example, the problem of belief mining has been a subject of study for decades (Liu, 2012). Solutions include such diverse approaches as detecting social polarization (Conover et al., 2011; Al Amin et al., 2017), opinion extraction (Srivatsa et al., 2012; Irsoy and Cardie, 2014; Liu et al., 2015), stance detection (Darwish et al., 2020) and sentiment analysis (Hu et al., 2013a,b), to name a few.

Pioneers, like Leman at el (Akoglu, 2014). and Bishan at el. (Yang and Cardie, 2012), had used Bayesian models and other basic classifiers to separate social beliefs. On the linguistic side, many efforts to extract user opinions based on domain-specific phrase chunks (Wu et al., 2018), and temporal expressions (Schulz et al., 2015). With the help of pre-trained embedding, like Glove (Liu et al., 2015) or word2vec (Wang et al., 2017), deep neural networks (e.g., variants of RNN (Irsoy and Cardie, 2014; Liu et al., 2015)) emerged as powerful tools (usually with attention modules (Wang et al., 2017)) for understanding the polarity or sentiment of user messages. In contrast to the above supervised language-specific solutions, we want to provide options and consider the challenge of developing an unsupervised approach that could also be language-agnostic.

In the domain of unsupervised algorithms, our problem is different from the related problems of unsupervised topic detection (Ibrahim et al., 2018; Litou and Kalogeraki, 2017), sentiment analysis (Hu et al., 2013a,b), truth discovery (Shao et al., 2018, 2020), and unsupervised community detection (Fortunato and Hric, 2016). Topic modeling assigns posts to polarities or topic mixtures (Han et al., 2007), independently of actions of users on this content. Hence, they often miss content nuances or context that helps better interpret the stance of the source. Community detection (Yang and Leskovec, 2013), on the other hand, groups nodes by their general interactions, maximizing intra-class links while minimizing inter-class links (Yang and Leskovec, 2013; Fortunato and Hric, 2016), or partitioning (hyper)graphs (Zhou et al., 2007). While different communities may adopt different beliefs, this formulation fails to distinguish regions of belief overlap from regions of disagreement.

The above suggests that belief mining must consider both sources (and forwarding patterns) and content. Prior solutions used a source-claim bipartite graph, and determined disjoint polarities by iterative factorization (Akoglu, 2014; Al Amin et al., 2017). Our work extends a conference publication that first introduced the hierarchical belief separation problem (Yang et al., 2020). This direction is novel in postulating a more generic and realistic view: social beliefs could overlap and can be hierarchically structured. In this context, we developed a new matrix factorization scheme that considers 1) the source-claim graph (Al Amin et al., 2017); 2) message word similarity (Weninger et al., 2012) and 3) user social dependency (Zhang et al., 2013) in a new class of non-negative matrix factorization techniques to solve the hierarchical overlapping belief estimation problem.

The work also contributes to non-negative matrix factorization. NMF was first introduced by Paatero and Tapper (Paatero and Tapper, 1994) as the concept of positive matrix factorization and was popularized by the work of Lee and Seung (Lee and Seung, 2001), who gave an interesting interpretation based on parts-based representation. Since then, NMF has been widely used in various applications, such as pattern recognition (Cichocki et al., 2009) and signal processing (Buciu, 2008).

Two main issues of NMF have been intensively discussed during the development of its theoretical properties: solution uniqueness (Donoho and Stodden, 2004; Klingenberg et al., 2009) and decomposition sparsity (Moussaoui et al., 2005; Laurberg et al., 2008). By only considering the standard formula X ˜ UM, it is usually not difficult to find a non-negative and non-singular matrix V, such that UV and V−1M could also be a valid solution. Uniqueness will be achieved if U and M are sufficiently sparse or if additional constraints are included (Wang and Zhang, 2012). Special constraints have been proposed in (Hoyer, 2004; Mohammadiha and Leijon, 2009) to improve the sparseness of the final representation.

Non-negative matrix tri-factorization (NMTF) is an extension of conventional NMF (i.e., X ˜ UBM (Yoo and Choi, 2010)). Unconstrained NMTF is theoretically identical to unconstrained NMF. However, when constrained, NMTF possesses more degrees of freedom (Wang and Zhang, 2012). NMF on a manifold emerges when the data lies in a nonlinear low-dimensional submanifold (Cai et al., 2008). Manifold Regularized Discriminative NMF (Ana et al., 2011; Guan et al., 2011) were proposed with special constraints to preserve local invariance, so as to reflect the multilateral characteristics.

In this work, instead of including constraints to impose structural properties, we adopt a novel belief structured matrix factorization by introducing the mixture matrix B. The structure of B can well reflect the latent belief structure and thus narrows the search space to a good enough region.

8 Conclusion

In this paper, we discuss computational modeling of polarized social groups using a class of NMF, where the structure of parts is already known (or assumed to follow some generic form). Specifically, we use a belief structure matrix B to describe the structure of the latent space and evaluate a novel Belief Structured Matrix Factorization algorithm (BSMF) that separates overlapping, hierarchically structured beliefs from large volumes of user-generated messages. The factorization could be briefly formulated as XMSUBM, where B is known. The soundness of the model is first tested on a synthetic dataset. A further evaluation is conducted on real Twitter events. The results show that our algorithm consistently outperforms baselines. The paper contributes to a research direction on automatically separating data sets according to arbitrary belief structures to enable more in-depth modeling and understanding of social groups, attitudes, and narratives on social media.

Data Availability Statement

The original contributions presented in the study are included in the article/Supplementary Material, further inquiries can be directed to the corresponding author.

Author Contributions

DS and CY contributed to the formulation of the problem and derivation of the algorithm and the proof. DS and CY performed the simulation experiment. JL, RW, SY, HS, DL, SL, and TW collected and organized datasets, and helped DS and CY performed experiments. DS and CY wrote the draft of the manuscript. All authors contributed to manuscript revision, read, and approved the submitted version.

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s Note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors, and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Acknowledgments

Research reported in this paper was sponsored in part by DARPA award W911NF-17-C-0099, DARPA award HR001121C0165, Basic Research Office award HQ00342110002, and the Army Research Laboratory under Cooperative Agreement W911NF-17-20196. The views and conclusion contained in this document are those of the authors and should not be interpreted as representing the official policies of the CCDC Army Research Laboratory, DARPA, or the United States government. The United States government is authorized to reproduce and distribute reprints for government purposes notwithstanding any copyright notation hereon.

Footnotes

1In practice, we permute the labels and pick the best matching as a result, since our approach does clustering not classification.

2http://apollo4.cs.illinois.edu/datasets.html

References

Akoglu, L. (2014). Quantifying Political Polarity Based on Bipartite Opinion Networks. Ann Arbor: ICWSM.

Google Scholar

Al Amin, M. T., Aggarwal, C., Yao, S., Abdelzaher, T., and Kaplan, L. (2017). “Unveiling Polarization in Social Networks: A Matrix Factorization Approach,” in 2017 - IEEE Conference on Computer Communications, Atlanta (INFOCOM). doi:10.1109/infocom.2017.8056959

CrossRef Full Text | Google Scholar

Ana, S., Yoob, J., and Choi, S. (2011). Manifold-respecting Discriminant Nonnegative Matrix Factorization [j]. Pattern Recognition Lett. 32 (6), 832–837. doi:10.1016/j.patrec.2011.01.012

CrossRef Full Text | Google Scholar

Bakshy, E., Messing, S., and Adamic, L. A. (2015). Exposure to Ideologically Diverse News and Opinion on Facebook. Science 348 (6239), 1130–1132. doi:10.1126/science.aaa1160

PubMed Abstract | CrossRef Full Text | Google Scholar

Bessi, A., Zollo, F., Del Vicario, M., Puliga, M., Scala, A., Caldarelli, G., et al. (2016). Users Polarization on Facebook and Youtube. PloS one 11 (8), e0159641. doi:10.1371/journal.pone.0159641

PubMed Abstract | CrossRef Full Text | Google Scholar

Buciu, I. (2008). “Non-negative Matrix Factorization, a New Tool for Feature Extraction: Theory and Applications,” in International Journal of Computers, Communications and Control.

Google Scholar

Cai, D., He, X., Wu, X., and Han, J. (2008). “Non-negative Matrix Factorization on Manifold,” in ICDM. doi:10.1109/icdm.2008.57

CrossRef Full Text | Google Scholar

Cheng, K., Li, J., Tang, J., and Liu, H. (2017). “Unsupervised Sentiment Analysis with Signed Social Networks,” in AAAI.

Google Scholar

Cichocki, A., Zdunek, R., Phan, A. H., and Amari, S.-i. (2009). Nonnegative Matrix and Tensor Factorizations: Applications to Exploratory Multi-Way Data Analysis and Blind Source Separation. John Wiley & Sons.

Google Scholar

Conover, M. D., Ratkiewicz, J., Francisco, M., Gonçalves, B., Menczer, F., and Flammini, A. (2011). “Political Polarization on Twitter,” in ICWSM.

Google Scholar

Darwish, K., Stefanov, P., Aupetit, M., and Nakov, P. (2020). “Unsupervised User Stance Detection on Twitter,” in Proceedings of the International AAAI Conference on Web and Social Media, 141–152.14

Google Scholar

Demartini, G., Siersdorfer, S., Chelaru, S., and Nejdl, W. (2011). “Analyzing Political Trends in the Blogosphere,” in Fifth International AAAI Conference on Weblogs and Social Media.

Google Scholar

Devlin, J., Chang, M.-W., Lee, K., and Toutanova, K. (2019). Bert: Pre-training of Deep Bidirectional Transformers for Language Understanding [Dataset].

Google Scholar

Donoho, D., and Stodden, V. (2004). “When Does Non-negative Matrix Factorization Give a Correct Decomposition into Parts,” in NIPS.

Google Scholar

Fortunato, S., and Hric, D. (2016). Community Detection in Networks: A User Guide. Phys. Rep. 659. doi:10.1016/j.physrep.2016.09.002

CrossRef Full Text | Google Scholar

Go, A., Bhayani, R., and Huang, L. (2009). Twitter Sentiment Classification Using Distant Supervision. Processing 150.

Google Scholar

Guan, N., Tao, D., Luo, Z., and Yuan, B. (2011). Manifold Regularized Discriminative Nonnegative Matrix Factorization with Fast Gradient Descent. IEEE Trans. Image Process. 20 (7), 2030–2048. doi:10.1109/tip.2011.2105496

PubMed Abstract | CrossRef Full Text | Google Scholar

Han, J., Cheng, H., Xin, D., and Yan, X. (2007). Frequent Pattern Mining: Current Status and Future Directions. TKDD 15 (1), 55–86. doi:10.1007/s10618-006-0059-1

CrossRef Full Text | Google Scholar

Hoyer, P. O. (2004). Non-negative Matrix Factorization with Sparseness Constraints. JMLR.

Google Scholar

Hu, X., Tang, J., Gao, H., and Liu, H. (2013a). “Unsupervised Sentiment Analysis with Emotional Signals,” in Proceedings of the 22nd international conference on World Wide Web - WWW. 13. doi:10.1145/2488388.2488442

CrossRef Full Text | Google Scholar

Hu, Y., Wang, F., and Kamb, S. (2013b). “Listening to the Crowd: Automated Analysis of Events via Aggregated Twitter Sentiment,” in IJCAI.

Google Scholar

Ibrahim, R., Elbagoury, A., Kamel, M. S., and Karray, F. (2018). “Tools and Approaches for Topic Detection from Twitter Streams: Survey,” in Knowledge and Information Systems. doi:10.1007/s10115-017-1081-x

CrossRef Full Text | Google Scholar

Irsoy, O., and Cardie, C. (2014). “Opinion Mining with Deep Recurrent Neural Networks,” in EMNLP. doi:10.3115/v1/d14-1080

CrossRef Full Text | Google Scholar

Klingenberg, B., Curry, J., and Dougherty, A. (2009). Non-negative Matrix Factorization: Ill-Posedness and a Geometric Algorithm. Pattern Recognition. 42(5):918–928. doi:10.1016/j.patcog.2008.08.026

CrossRef Full Text | Google Scholar

Küçük, D., and Can, F. (2020). Stance Detection. ACM Comput. Surv. 53, 1–37. doi:10.1145/3369026

CrossRef Full Text | Google Scholar

Laurberg, H., Christensen, M. G., Plumbley, M. D., Hansen, L. K., and Jensen, S. H. (2008). Theorems on Positive Data: On the Uniqueness of Nmf. Computational Intelligence and Neuroscience.

Google Scholar

Lee, D. D., and Seung, H. S. (2001). “Algorithms for Non-negative Matrix Factorization,” in NIPS.

Google Scholar

Litou, I., and Kalogeraki, V. (2017). “Pythia: A System for Online Topic Discovery of Social media Posts,” in ICDCS. doi:10.1109/icdcs.2017.289

CrossRef Full Text | Google Scholar

Liu, B. (2012). “Sentiment Analysis and Opinion Mining,” in Synthesis Lectures on Human Language Technologies.

CrossRef Full Text | Google Scholar

Liu, P., Joty, S., and Meng, H. (2015). “Fine-grained Opinion Mining with Recurrent Neural Networks and Word Embeddings,” in EMNLP. doi:10.18653/v1/d15-1168

CrossRef Full Text | Google Scholar

Mohammadiha, N., and Leijon, A. (2009). “Nonnegative Matrix Factorization Using Projected Gradient Algorithms with Sparseness Constraints,” in ISSPIT. doi:10.1109/isspit.2009.5407557

CrossRef Full Text | Google Scholar

Moussaoui, S., Brie, D., and Idier, J. (2005). “Non-negative Source Separation: Range of Admissible Solutions and Conditions for the Uniqueness of the Solution,” in ICASSP.

Google Scholar

Nguyen, D. Q., Vu, T., and Nguyen, A. T. (2020). “BERTweet: A Pre-trained Language Model for English Tweets,” in Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, 9–14. doi:10.18653/v1/2020.emnlp-demos.2

CrossRef Full Text | Google Scholar

Paatero, P., and Tapper, U. (1994). Positive Matrix Factorization: A Non-negative Factor Model with Optimal Utilization of Error Estimates of Data Values. Environmetrics 5 (2), 111–126. doi:10.1002/env.3170050203

CrossRef Full Text | Google Scholar

Schulz, A., Schmidt, B., and Strufe, T. (2015). “Small-scale Incident Detection Based on Microposts,” in Proceedings of the 26th ACM Conference on Hypertext & Social Media. doi:10.1145/2700171.2791038

CrossRef Full Text | Google Scholar

Shao, H., Sun, D., Yao, S., Su, L., Wang, Z., Liu, D., et al. (2021). Truth Discovery with Multi-Modal Data in Social Sensing. IEEE Trans. Comput. 70, 1325–1337. doi:10.1109/TC.2020.3008561

CrossRef Full Text | Google Scholar

Shao, H., Yao, S., Zhao, Y., Zhang, C., Han, J., Kaplan, L., et al. (2018). “A Constrained Maximum Likelihood Estimator for Unguided Social Sensing,” in IEEE INFOCOM 2018-IEEE Conference on Computer Communications (IEEE), 2429–2437. doi:10.1109/infocom.2018.8486306

CrossRef Full Text | Google Scholar

Srivatsa, M., Lee, S., and Abdelzaher, T. (2012). “Mining Diverse Opinions,” in MILCOM 2012-2012 IEEE Military Communications Conference. doi:10.1109/milcom.2012.6415602

CrossRef Full Text | Google Scholar

Wang, H., Nie, F., Huang, H., and Makedon, F. (2011). “Fast Nonnegative Matrix Tri-factorization for Large-Scale Data Co-clustering,” in Twenty-Second International Joint Conference on Artificial Intelligence.

Google Scholar

Wang, W., Pan, S. J., Dahlmeier, D., and Xiao, X. (2017). “Coupled Multi-Layer Attentions for Co-extraction of Aspect and Opinion Terms,” in AAAI.

Google Scholar

Wang, Y.-X., and Zhang, Y.-J. (2012). Nonnegative Matrix Factorization: A Comprehensive Review. San Francisco: TKDE.

Google Scholar

Weninger, T., Bisk, Y., and Han, J. (2012). “Document-topic Hierarchies from Document Graphs,” in CIKM. doi:10.1145/2396761.2396843

CrossRef Full Text | Google Scholar

Wu, C., Wu, F., Wu, S., Yuan, Z., and Huang, Y. (2018). A Hybrid Unsupervised Method for Aspect Term and Opinion Target Extraction. Knowledge-Based Syst. 148, 66–73. doi:10.1016/j.knosys.2018.01.019

CrossRef Full Text | Google Scholar

Yang, B., and Cardie, C. (2012). “Extracting Opinion Expressions with Semi-markov Conditional Random fields,” in EMNLP.

Google Scholar

Yang, C., Li, J., Wang, R., Yao, S., Shao, H., Liu, D., et al. (2020). “Hierarchical Overlapping Belief Estimation by Structured Matrix Factorization,” in 2020 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining (ASONAM), 81–88. doi:10.1109/ASONAM49781.2020.9381477

CrossRef Full Text | Google Scholar

Yang, J., and Leskovec, J. (2013). “Overlapping Community Detection at Scale: a Nonnegative Matrix Factorization Approach,” in WSDM.

Google Scholar

Yoo, J., and Choi, S. (2010). Orthogonal Nonnegative Matrix Tri-factorization for Co-clustering: Multiplicative Updates on Stiefel Manifolds. Elsevier.

Google Scholar

Zhang, C., Tao, F., Chen, X., Shen, J., Jiang, M., Sadler, B., et al. (2018). “Taxogen: Unsupervised Topic Taxonomy Construction by Adaptive Term Embedding and Clustering,” in SIGKDD.

Google Scholar

Zhang, H., Dinh, T. N., and Thai, M. T. (2013). “Maximizing the Spread of Positive Influence in Online Social Networks,” in ICDCS. doi:10.1109/icdcs.2013.37

CrossRef Full Text | Google Scholar

Zhou, D., Huang, J., and Schölkopf, B. (2007). “Learning with Hypergraphs: Clustering, Classification, and Embedding,” in NIPS.

Google Scholar

Keywords: polarization, belief estimation, hierarchical, matrix factorization, unsupervised

Citation: Sun D, Yang C, Li J, Wang R, Yao S, Shao H, Liu D, Liu S, Wang T and Abdelzaher TF (2021) Computational Modeling of Hierarchically Polarized Groups by Structured Matrix Factorization. Front. Big Data 4:729881. doi: 10.3389/fdata.2021.729881

Received: 24 June 2021; Accepted: 16 November 2021;
Published: 22 December 2021.

Edited by:

Neil Shah, Independent researcher, Santa Monica, CA, United States

Reviewed by:

Tong Zhao, University of Notre Dame, United States
Ekta Gujral, Walmart Labs, United States

Copyright © 2021 Sun, Yang, Li, Wang, Yao, Shao, Liu, Liu, Wang and Abdelzaher. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Dachun Sun, bWFpbHRvOmRzdW4xOEBpbGxpbm9pcy5lZHU=

These authors have contributed equally to this work

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.