The Opacity Problem in Social Contagion
Abstract
Fads, product adoption, mobs, rumors, memes, and emergent norms are diverse social contagions that have been modeled as network cascades. Empirical study of these cascades is vulnerable to what we describe as the “opacity problem”: the inability to observe the critical level of peer influence required to trigger an individual’s behavioral change. Even with maximal information, network cascades reveal intervals that bound critical levels of peer exposure, rather than critical values themselves. Existing practice uses interval maxima, which systematically overestimates the social influence required for behavioral change. Simulations reveal that the overestimation is likely common and large in magnitude. This is confirmed by an empirical study of hashtag cascades among 3.2 million Twitter users: one in five hashtag adoptions suffers critical value uncertainty due to the opacity problem. Different assumptions about these intervals lead to qualitatively different conclusions about the role of peer reinforcement in diffusion. We introduce a solution that combines identifying tightly bounded intervals with predicting uncertain critical values using nodelevel information.
Keywords: social contagion, social influence, peer effects, measurement
Like epidemic diseases [30, 45, 11], social contagions are ubiquitous, highly consequential, and widely studied [24, 39, 40, 13, 32, 42, 43, 16, 44]. Unlike nearly all diseases, social contagions often require multiple sources of infection, especially if adoption is costly, risky, motivated by affect, or entails positive network externalities [13].
We discover a challenge confronting observational studies of social contagion, which we term the “opacity problem”. Data generated by network cascades often fail to precisely indicate the critical value of social influence required for a node to change behavior. Instead, cascades reveal intervals in which critical values lie. Researchers typically infer the amount of social reinforcement required to trigger activation by recording node exposure (number of active neighbors) at activation time. This procedure is equivalent to taking the maximum of the critical value interval for each node. We note that this exposureatactivation (EAA) rule can be found in diverse studies across disciplines, including sociology [22, 43, 20], economics [19, 10, 6], medicine [16], and information science [1, 37, 38, 23, 41, 9, 48]. It is even present in “snapshot” studies [4, 18] which compare a network observed at two or more time intervals.
The implications of opacity and the EAA rule are potentially farreaching: the data used for empirical estimates of peer influence likely overstate the level of influence required to change behavior. This suggests that estimates of peer influence suffer bias, including the possibility of estimating peer effects when choices are in fact independent.
Simulated cascades on a variety of network topologies, using both deterministic and probabilistic activation rules, demonstrate that opacity and the EAA rule often produce overestimates of critical values. An empirical study of hashtag cascades among 3.2 million users on Twitter confirms this intuition: different responses to the opacity problem produce qualitatively different conclusions about whether peer reinforcement promotes diffusion.
Unlike Manski’s famous reflection problem [33], opacity cannot be solved by alternative model specification or more finegrained data (such as more frequent samples of node behavior). It is part of the data generated by cascades.
Fortunately, opacity can be addressed by: 1) identifying which node critical values are measured precisely based on the width of their activation intervals; and 2) using models trained on the precisely measured nodes to predict for nodes whose critical values are uncertain. In addition, we expect future research to develop novel ways to further reduce measurement error introduced by opacity and the EAA rule.
The importance of cascades and the difficulty of performing experiments on graphs means that cascade dynamics are often studied observationally. This is true even in settings where higherorder units are randomized, such as villages [6] or health discussion groups [14]. In cases where egonetworks are randomized [5, 2], the opacity problem we highlight can be avoided. In particular, this method solves the opacity problem by turning a large, complex network into many much simpler structures where dependencies between nodes are limited (see [21, 3] for a discussion of how to design experiments in networks while avoiding interference).
1 Intuition
1.1 Example
Suppose we wish to measure the level of social influence required for a focal node (“ego”) to adopt a hashtag on social media. Ego logs in without any neighbors using the hashtag (exposure 0) and then logs out for a period of time. While ego is logged out and “not looking”, three of ego’s network neighbors use the hashtag. Ego then logs in again, sees that three neighbors have adopted the hashtag, and adopts the hashtag (exposure 3). Given the available information, ego’s critical exposure could be 1, 2, or 3. Ego therefore is measured as having a critical value interval rather than a point.
Standard empirical practice in observational studies of network cascades is to record ego’s exposureatactivation (EAA) as the “critical value”, which in this example is 3. Yet we cannot rule out the possibility that ego would have activated at exposure 1 or 2, in which case the EAA rule overestimates the required level of social influence.
The opacity problem refers to uncertainty about the precise critical value within an interval. Put differently, opacity is the inability to observe a counterfactual: the number of activated neighbors at which ego would have activated had ego logged in sooner. The problem is not limited to cascades observed on social media, nor is the problem solved if ego were to continuously update the status of her neighbors. For example, suppose we are studying a cascade in real time on a small facetoface network such as a workplace or classroom. We observe that ego adopted only after three of ego’s neighbors simultaneously adopted. There is no way to know if ego might have adopted if only one or two neighbors had adopted and not all three. The critical value is still uncertain in an interval.
We may conjecture that if only ego had checked (“updated”) more frequently, the opacity problem could be avoided. However, simple graphs (Figure 1B) demonstrate that the opacity problem is often inevitable for at least one node. In an enumeration of all two, three, and fournode graphs with all viable critical value combinations (Figure 2), we find that 75% always suffer from opacity for at least one node.
1.2 Implications
The “worstcase” outcome of the opacity problem occurs when an “innovator” (critical value 0) activates after all of its neighbors have activated. In this case, the node has a large critical value interval, and the EAA rule has error equal to the node degree. This raises the alarming possibility that much prior research on diffusion greatly overstates the level of social reinforcement required for behavioral change.
Yet the seriousness of the opacity problem may be greatest when the error magnitude is minimal. Consider the case of an innovator who has exposure 1 at activation (Figure 1A). The EAA rule assigns this node a critical value of 1. This minimum possible overestimation leads to a qualitatively incorrect conclusion that social influence is responsible for behavioral change that is actually independent. The opacity problem therefore creates an additional problem for researchers seeking to determine the role of influence as opposed to homophily in empirical cascades [1].
The opacity problem also impacts graphlevel analyses. In large hashtag cascades on Twitter, one in five nodes was found to suffer from uncertainty. The effects of this uncertainty for cascade prediction and seed selection [15, 28] could be dire due to the sensitivity of cascades to even very small gaps in the distribution of critical values, as demonstrated by Granovetter [24]. Even a seemingly trivial overestimation of the critical value of a single node in a large population could make the difference between predicting early cascade failure and global adoption.
2 Findings
2.1 Terminology and notation
In a graph, each node has an activation status (also called adoption status) , where 1 indicates active and 0 indicates inactive. We denote ’s exposure as , or as for exposure at time . We assume that once a node has activated, it remains active. The critical value is the minimum exposure that will trigger ’s activation. For each , cascades reveal a critical value interval which contains the true critical value . An interval such as indicates that . The maximum of is obtained from ’s exposure at activation. The minimum of is ’s exposure at the most recent update before activation, plus 1. In the case where , there is measurement error denoted .
We say a node “updates” when it checks neighbor activation statuses and makes an activation decision. Nodes are assumed to check all neighbor statuses when updating, which is important for establishing the correct lower bound on .
To briefly recap our example above, assume ’s true critical value . Then, updates at but remains inactive with , and updates and activates at some later with . In this case, and we know . We add one to because updated yet remained inactive. The lowest possible exposure that could trigger activation is therefore the next integer . With the knowledge that , using the EAA rule and taking the maximum of results in measurement error .
2.2 Precise measurement condition
A simple condition determines whether a node’s critical value is precisely measured, or whether the critical value is uncertain within an interval. This condition can be applied to cascades where both updates and activations are observed. If only activations are observed, it is impossible to obtain lower bounds on critical value intervals.
Consider a node which updates at and some later . Assume is inactive at () and active at (), giving critical value interval .
Condition 1.
The critical value for is precisely measured if the upper and lower bounds of are equal, .
Intuitively, updates at but does not activate, and then updates at and activates. In that time, ’s exposure has changed by exactly one. In this case, ’s critical value is , which is the exposure at activation. Intervals that contain more than a point indicate uncertainty about where the true critical value lies. Condition 1 thus provides a quantification of the measurement uncertainty in addition to identifying precisely measured nodes.
An edge case occurs for innovators, or nodes which activate with 0 exposure. If we assume that critical values cannot be negative, this provides a lower bound and innovators are precisely measured. However, if critical values can be negative (e.g. “super eager” to activate as found in latent dependent variable models [20]), then innovators are not precisely measured. In a given case, correctly measuring innovators relies on an assumption about the underlying distribution of critical values.
Empirically, nodes often update many times with equal exposure before activating. This corresponds to cases where factors aside from exposure affect activation. For instance a node may go from 0 exposure to 2 exposure, and have many updates at 2 exposure before adopting. In these cases, we focus on the interval rather than the interval. Even if other factors change to encourage activation at 2 exposure, uncertainty still exists whether those factors would have caused activation at 1 exposure as well.
2.3 Distributional consequences of the EAA rule
The two main approaches to studying social cascades theoretically are probabilistic [26, 28, 34] and threshold [24, 46, 13, 12] approaches. Probabilistic approaches treat each active neighbor as having an independent chance to activate ego. Threshold approaches encode an assumption that nodes require a specific level of reinforcement to activate.
In both cases, there is a distribution of critical values that is of interest to researchers. In the probabilistic case, understanding this distribution allows estimating the transmissibility factor of the cascade. In the threshold case, this distribution is the distribution of thresholds.
While the substantive interpretations differ, opacity creates similar issues for recovering true critical value distributions in both cases. Figure 3 shows consequences of using the EAA rule for probabilistic (A,B), integer threshold (C,D), and fractional threshold (E,F) models. Integer threshold models assume only active neighbors have influence, while fractional threshold models assume that both active and inactive neighbors have influence [13].
In five out of six cases, the EAA rule does not recover the basic critical value distribution shape. For probabilistic models (A,B), the true distribution is geometric, yet the EAA distribution looks vaguely normal. In the fractional threshold cases (E,F), the measurements produced by the EAA rule seem to have no relationship to the underlying distribution. In the case where all nodes have fractional threshold 0.2 except for 5% seed nodes (F), the EAA rule records 10% of nodes as having “threshold” of 1.0. In the one case where the EAA rule roughly replicates the true distribution shape (normal distribution in panel D), it creates a long tail of “high threshold” nodes not present in the true distribution.
These results indicate that opacity presents substantial challenges for estimating the effects of peer influence. Regressionbased approaches often use the count or proportion of active neighbors as an independent variable, which does not reliably replicate true node sensitivity to peer behavior (see SI for further discussion of the implications of opacity for regression models).
2.4 Hashtag cascades
Social media provides an ideal setting to examine the empirical implications of the opacity problem. This section demonstrates that opacity can qualitatively affect conclusions about whether and to what extent social influence/reinforcement facilitates diffusion. For about 2 in 5, different responses to opacity yield opposite conclusions about the effectiveness of social reinforcement.
The effects of peer influence can be plotted using curves [18, 37, 31] of the probability of first activation with exposure , given ever being exposed. Work using this methodology has found that higher exposure levels substantially increase the probability of hashtag adoption, particularly for controversial issues like politics [18, 37]. More recent work has found that hashtags spread more like simple contagion where nodes are immunized after the first exposure [31]. Since complex contagions have different cascade dynamics than simple contagions [12, 8], understanding the effects of reinforcement is consequential for predictions and interventions in online networks.
To that end, we tracked cascades for 50 hashtags (see SI for list and selection procedure) with between 40,000 and 360,000 unique adopters and 41,000 to 1.3 million total usages per tag. A total of 3.2 million users (“active users”) tweeted using one or more of these hashtags. These active users are connected by 45 million bidirected @mention links, which is a common measure of ties on Twitter [37]. An additional 105 million bidirected @mention edges connect active users to inactive users. In these data, active users tweeted a total of 7 billion times.
Computing critical value intervals
To construct critical value intervals for all users, we assume updates (checking alter status) correspond with tweet times. A user is engaged with the platform when tweeting, and may check the tweets of friends in the seconds before or after sending a tweet. A user’s Twitter timeline therefore provides an update record. For each node first adopting each hashtag , there is a first usage time denoted . The time of tweets prior to are denoted in descending order as etc., where is the time of the tweet immediately prior to ’s first usage of .
Iterating through pairs of timestamps in reverse chronological order gives descending time intervals , etc. These time intervals correspond to periods of time when was “not looking”, bookended by updates. We find the most recent time interval where at least one neighbor first used , denoted , and count the total number of neighbor’s first usages of in . If exactly one neighbor adopted in , then by Condition 1 ’s critical value for using is precisely measured. If more then one neighbor activated in , then has an interval of size larger than one. For instance if had 2 neighbors activate in and 4 neighbors activate before , then ’s critical value interval would be , with a width of two.
Empirical measurement rates
Measurement uncertainty affects 21% of activations, although there is substantial variation both by hashtag and exposure level. Note that we exclude critical value 0 nodes from this analysis because measuring these nodes precisely relies on an assumption about the critical value distribution. Nodes which adopt with higher exposure are more likely to suffer from measurement uncertainty (Figure 4). Among nodes that activate with exposure 1, 90.9% are measured precisely. Among nodes that activate with exposure 2, only 65.7% are precisely measured.
Variation also exists in measurement rates between tags. Among nodes that activate with exposure 2, 73.6% are precisely measured among the 13 easiesttomeasure tags, while only 49.1% are precisely measured among the 13 hardesttomeasure tags.
Social reinforcement and p(k) curves
Based on critical value intervals for each node, we construct “lower” and “upper” curves by taking the minima and maxima of intervals, respectively. The lower curve assumes that the EAA rule is maximally wrong. While taking interval minima may at first seem like an unrealistic assumption, it avoids introducing a longtailed error into the data and can therefore have lower error than maxima in some cases. The upper curve is equivalent to using the EAA rule as applied in existing research.
For some tags, upper and lower curves show similar patterns (Figure 5, red lines), while for others upper and lower curves diverge sharply (blue lines). For nearly all hashtags, the upper curve has , corresponding to social reinforcement increasing adoption probability. For 19 out of 50 hashtags, the lower curve reverses this pattern, , suggesting that these tags do not benefit from reinforcement. We refer to these tags as having a “drop” for the lower curve.
Examining characteristics of cascades, we find that tags with high clustering coefficients among adopters tended to have a drop (Figure 5B). Other factors, such as the overall measurement rate of the hashtag or the temporal concentration of adoptions (Gini coefficient of adoptions across days) were poor predictors of a drop.
Taken together, these results indicate that the opacity problem creates more uncertainty about cascade dynamics when adopters are clustered in dense networks. Two similarly plausible assumptions for dealing with opacity produce opposite qualitative results about the effects of social reinforcement for nearly 2 in 5 hashtags. This finding paradoxically makes it particularly difficult to determine the effects of social reinforcement when adopters have high clustering, even though contagion which requires social reinforcement is most likely to spread in clustered networks [13].
2.5 Reducing measurement error with predictive models
In empirical cases, many nodes have uncertain critical values. Simulation provides a way to quantify measurement error since critical values are known. Using simulated cascades with integer thresholds on random graphs, we find that measurement error is substantial in a variety of cases, with the EAA rule having an average rootmeansquared error (RMSE) of 3.1 to 8.1 times baseline in cases we study. We simulate a scenario where nodelevel characteristics contribute to the node threshold. Training a predictive model on the precisely measured nodes using this information reduces this error to 1.15 to 1.35 times baseline.
Simulation details
Integer thresholds are generated by the function , where . The covariate represents some feature of (e.g. age, wealth) that provides information about how much social reinforcement needs before activation. The error variance represents the inherent unpredictability in thresholds and provides a natural baseline for comparing errorreduction methods. An unbiased ordinary least squares (OLS) model trained on the true data will have a rootmeansquare error (RMSE) equal to the error variance, which here is set to 1.
Graphs are generated using a power law with clustering algorithm [27] with clustering parameter 0.1. Graphs have 1000 nodes and mean degree varies between 12, 16, and 20, with 1000 replications for each parameter set. Power law with clustering graphs exhibit high heterogeneity of node degree found in online settings, while also building in clustering that is typical of social networks. See the SI for a description of cascade dynamics.
Simulation results
Mean degree  Nodes  Activations  Precisely measured 

12  1000  720.7  113.4 
16  1000  924.1  43.0 
20  1000  983.1  13.0 
As graph mean degree increases, more nodes activate but fewer are precisely measured (Table 1). For a fixed integer threshold distribution, adding links simulates a more “viral” cascade which spreads further in less time. When cascades spread quickly relative to node update pace, there is a higher chance that a node is “not looking” while multiple neighbors adopt, leading to large threshold intervals and increased measurement uncertainty.
Using the EAA rule produces a substantially different distribution from the true threshold distribution, including introducing a long right tail. The maximum threshold in the true distribution is around 20, but the EAA rule records many nodes with “thresholds” of over 30.
Despite the small number of precisely measured nodes, they have an important advantage: we are certain their threshold is measured without error. This allows estimating a model relating to node characteristic . In these simulations, for precisely measured nodes is often slightly less than 0, indicating selection on the error due to lower threshold nodes being more likely to be measured precisely. Nevertheless, models estimated with the precisely measured nodes (Figure 6, blue) reliably outperform two natural alternatives: using the EAA rule directly (red) and estimating a model to predict thresholds using all active nodes (green). Compared to a baseline RMSE of 1, a model using the precisely measured nodes increases error by 14% to 35%, while using all active nodes increases error by 203% to 507%, and the EAA rule increases error by 310% to 807%. The lowerror predictions using the precisely measured nodes are obtained despite using, on average, only 13 to 113 observations. The other two approaches use, on average, between 721 and 983 observations.
3 Conclusion
Studying the opacity problem highlights the dangers of using exposure at activation to measure the critical level of social influence needed to trigger activation. To our knowledge, we are the first to recognize the potential danger of this problem for research on social influence and network cascades. Even under optimal conditions, the EAA rule inflates measurements of critical values, leading to overestimation of influence and incorrectly diagnosing as social contagions what are actually social allergens.
Although opacity presents substantial challenges, it can be addressed when researchers have access to node update behavior. In these cases, the magnitude of the opacity problem can be quantified at the node level using Condition 1. When covariates exist to predict uncertain critical values, the error introduced by opacity can be reduced dramatically.
Employing this response to opacity has the potential to yield more precise answers to important questions. For instance, models of node critical values can be used to better study complex contagion [13] empirically. Results from such an analysis can then be applied to problems such as influence maximization [28, 29].
Our approach also facilitates refinements of the cascade prediction problem [15], where a model can be trained on active nodes while the contagion is still unfolding. Such a model can be used to identify susceptible nodes on the cascade fringe.
Additional work is needed to refine the method we propose for reducing measurement error. We consider only precisely measured critical values, yet there are many critical values that fall within a narrow interval which may provide useful information. Future work should address how to better incorporate this information, as well as investigating the ability to predict critical values in different domains.
4 Acknowledgements
We thank David Strang, Thomas Davidson, participants of INSNA Sunbelt 2016, and members of the Social Dynamics Lab for helpful discussions and comments. This research has been supported in part by National Science Foundation grant SES1357442 and the Department of Defense Minerva Initiative grant FA95501510036.
5 Appendix
5.1 Simulation details
We simulate cascades using the NetworkX package in Python 3 [25]. All code is available at https://github.com/georgeberry/thresholds. Statistical analyses are done using Scikit Learn [35] in Python 3, and in R.
We use 4 different graph generation routines at various points: BarabàsiAlbert [7], powerlaw with clustering [27], WattsStrogatz [47], and an “atlas” of all small graphs [36]. All four are builtin to the NetworkX package.
Small graphs
NetworkX provides a full enumeration of all small graphs (“graph atlas”). We take all such graphs with between two and four nodes. Then, we filter out graphs which have more than one component, giving a set of all connected graphs with between two and four vertices. Call this set of graphs .
For each , we generate all possible critical value assignments given that the critical value is less than or equal to node degree. For node in graph with degree , has the critical value set . Taking the product of for all gives the set of critical value assignments for the graph .
For each critical value assignment, we simulate cascades by updating nodes randomly and activating nodes immediately if exposure is greater than critical value, . We filter out graphs where at least one node never activates, which occurs when each inactive node has updated yet none activates.
For these set of “admitted” graphs where all nodes eventually activate, we simulate 100 cascades per graph. For each simulated cascade, we record exposure at each node update. This allows applying the exposureatactivation (EAA) rule to determine if nodes are precisely measured or not. If node updates at and some later , and , then the node is precisely measured. If there is no lower bound (e.g. no update before activation) and the exposure at activation is greater than zero, we call the node imprecisely measured. Innovators which adopt with 0 active neighbors are considered precisely measured here. As we discuss in the main text, this assumption about innovators may not be appropriate in all cases.
Distributional simulations
We use a BarabàsiAlbert graph generation process for these simulations. Graphs have 1000 nodes and each node has 6 links to attach preferentially to other nodes. In each case 5% of nodes are “seed” nodes and are either randomly activated (in the probabilistic case) or have threshold of 0 (in the threshold case). Each simulation is repeated 100 times.
For independent cascade model runs, we use two different methods to spread the cascade, which we call “push” and “random updates”.
The “push” version randomly selects an active node and iterates through its neighbors. Each neighbor flips a coin that comes up heads with probability and activates immediately if the coin comes up heads. When a node flips a coin, we consider it to be updating and so record its exposure. We also record the number of coins it has flipped. For instance, if a node flips a coin that comes up tails and then flips a coin that comes up heads, it has critical value two. At activation time, however, it may have more than two active neighbors.
The push model has a quirk not found in other models studied: exposure at the most recent update before activation may not be a lower bound of the true critical value. This occurs because nodes are “compelled” to update by an active node but effectively ignore other active neighbors. The EAA rule still overestimates critical values, however.
The “random updates” model works similarly to the threshold model. Nodes randomly update and flip a number of coins equal to the number of neighbors that have activated since last update time. Exposure is recorded at each update time, plus the outcome of the coin flips. The first “heads” flip is the critical value, and the node activates immediately.
Integer and fractional threshold simulations work by randomly updating nodes and checking if exposure is greater than threshold. Nodes activate immediately if the threshold is crossed. For plotting, fractional thresholds are binned to the nearest 0.05.
Error reduction simulations
Cascade dynamics are simulated by updating inactive nodes in random order, with nodes activating immediately if exposure is greater than or equal to the threshold, . Nodes record exposure before and after activation, allowing identification of precisely measured nodes by applying Condition 1. Cascades run until either all nodes have activated, or each inactive node has updated without activating. This process produces graphs where some nodes remain inactive (nodes with threshold greater than degree), which replicates the common empirical setting where not everyone adopts.
5.2 Twitter analysis
The Twitter data used for this analysis was collected for another project supported by an NSF grant (SES 1226483) between November 2013 and October 2014. Tweets were localized to a country using the method described in [17]. Users in Anglophone countries (US, UK, CA, AU, NZ, SG) were extracted and retweets were filtered out.
We selected users who had used one or more of 50 hashtags with moderatetohigh usage. We extracted the entire timeline for these users.
Hashtag selection
Hashtags were selected for first occurring between 2012 and 2014. We manually examined the top 1000 hashtags by usage among the Anglophone users and selected tags that we believed related to specific events or social media phenomena. Two project members independently examined the top 1000 tags, nominated tags, and then discussed disagreements.
The full list of (lowercased) tags is: obama2012, election2012, kony2012, romney, rippaulwalker, teamobama, bringbackourgirls, trayvonmartin, hurricanesandy, cantbreathe, miley, olympics2014, prayfornewtown, goodbyebreakingbad, governmentshutdown, riprobinwilliams, romneyryan2012, harlemshake, euro2012, marriageequality, benghazi, debate2012, newtown, linsanity, zimmerman, betawards2014, justicefortrayvon, samelove, worldcupfinal, prayersforboston, nobama, ferguson, springbreak2014, drawsomething, nfldraft2014, romney2012, snowden, replaceashowtitlewithtwerk, inaug2013, ivoted, trayvon, ios6, voteobama, jodiarias, windows8, mentionsomebodyyourethankfulfor, sharknado2, gop2012, whatdoesthefoxsay, firstvine.
5.3 Effects of opacity and the EAA rule on estimates of social influence
Our results indicate that models of social influence and peer effects suffer from substantial error due to opacity and the EAA rule.
Here we present the particular problem the EAA rule causes when estimating parametric models of peer effects commonly found in the social sciences. Importantly, even though opacity and the EAA rule cause critical values to be overestimated, this does not necessarily imply a particular type of bias for regression models. Estimates of the effect of social influence may be too large or too small, depending on the particular form bias takes.
This occurs because of two different biases introduced by opacity + the EAA rule: 1) the weak influence effect where nodes which would have activated at low exposure are instead measured as requiring high exposure; and 2) the phantom influence effect where nodes which would have activated with no active neighbors are instead measured as requiring some amount of social influence. The weak influence effect makes influence appear to be weaker than it really is (although, as we show below, it can increase peer effect model coefficients), while the phantom influence effect creates the illusion of influence where none exists. The worst case for the weak influence effect occurs when a node which requires one neighbor to activate is instead measured as requiring all neighbors to activate. The worst case for the phantom influence effect occurs when a node which requires zero neighbors to activate is instead measured as requiring one neighbor to activate.
The particular combination of these two effects may be specific to the dataset. To better understand these effects, we show how measurement error impacts nonparametric probability estimates of activation at some exposure level, and then consider how measurement error affects a simple regression model of social influence similar to models found in the literature.
Assume there is a binary outcome and call the exposure of . ’s exposure at activation may contain error: , where denotes the true critical value and denotes error. If we assume that nodes activate immediately, the error affects the data in a peculiar way: active nodes have measurement error since the EAA rule produces estimated critical values above the true critical value, but the exposure measurements for inactive nodes do not have error.
To see this more clearly, consider the effect of on estimating adoption probability at some particular exposure level . The true adoption probability at this is
(1) 
where denotes the count of nodes whose true critical value is and denotes the count of nodes who were ever exposed. With measurement error, this becomes
(2) 
where indicates the count of nodes whose true critical value plus measurement error is , and indicates the count of nodes which were exposed and remained inactive. This error means that for some low values, the estimated probability will be too small, while for some large the estimated probability will be too large.
We can write down some properties of with this in mind: , , and may be correlated with . Additionally, while . In other words, there is measurement error for active nodes but not for inactive nodes, on the condition that we assume immediate activation.
We can examine further examine the impacts of measurement error on a simple regression model similar to those used in empirical work. Assume that each of ’s active neighbors increases ’s probability to adopt by . If is ’s exposure and is idiosyncratic error, we can then estimate a linear probability model of form
(3) 
First, we examine the weak influence effect. We choose and . Since observations with are active with probability 1, the range of is restricted to . Three different types of error are added to the active nodes only: 1) exponentially distributed error with ; 2) normally distributed error if the realization of the random variable is greater than zero; 3) binomial error where 3 is added to the observation with probability .
While nonparametric estimates of activation probability at give the result above (underestimates at some low , overestimates at some high ), effects on estimated are somewhat more complicated. The types of error we consider combined with our assumed distribution of lead to while . This is counterintuitive: since we underestimate the effectiveness of social influence at low levels, we overestimate the social influence model coefficient. The reason for this can be seen clearly in Figure 7: models on data with error suggest that higher exposures are disproportionately effective at encouraging activation. This increases the slope of while shifting the entire line to the right, leading to a low . This is present in both the linear probability model and fitting the same data with a logistic model.
This result shows that the weak influence effect has two consequences: 1) underestimating activation probability at low while overestimating it at high ; and 2) introducing bias into , which can paradoxically make the coefficient for social influence larger. Note that different assumptions about (e.g. uniformly distributed) can reverse this effect and cause to be too small.
The results of the phantom influence effect are more straightforward and are shown in Table 2. We set for all (therefore ), while retaining the same distribution as above. This means that the the constant . We add the same error to the active observations as above. As predicted, when adding error, we find an effect of influence much larger than zero.
Dependent variable:  
y  
True  Exp error  Norm 0 error  Binom error  
(1)  (2)  (3)  (4)  
0.398  0.196  0.289  0.247  
(0.388, 0.407)  (0.189, 0.204)  (0.280, 0.298)  (0.239, 0.256)  
0  0.199  0.259  0.251  
(0.195, 0.203)  (0.251, 0.268)  (0.245, 0.257) 
This demonstration is not meant to be realistic but rather to give an idea of the effects of the opacity problem plus the EAA rule on estimated coefficients. Where possible, it is helpful to examine estimates of activation probabilities at various exposure levels directly. Both weak influence and phantom influence effects are likely present in empirical data.
5.4 When is opacity not a problem?
The main text focuses on the case where there are no activation delays, and where the researcher does not control node update ordering. In this case, opacity is a common problem even in small graphs.
We can imagine other such processes:

Researcher control over node update ordering

Researcher control over node activation delays

Researcher can observe the activation delay

Researcher can observe when a node’s critical value is exceeded before public activation has taken place
We assume that in all cases, researchers do not control graph structure or the node critical values themselves. We present two results in this section: first, a restatement in greater detail of the finding in the main text that opacity is a problem in the case normally studied by social scientists. Second, we present a case where opacity is not a problem. This requires the researcher to have control over node updates, node activation delays, and to observe when a node’s critical value is exceeded before public activation occurs. This case does not occur in any domain we are aware of, but we present it to give an idea of the challenges posed by the opacity problem.
Notation
Individual nodes in a graph are indexed with , and time periods are indexed with . We use subscripts for individuals and superscripts for times .

—the graph structure composed of sets of nodes and edges . Vertices are labeled .

—the vector of length containing critical values for each vertex, with a value for denoted .

—the node update ordering. , specifies the set of nodes that update simultaneously at time . denotes the time periods where node updates.

—the vector of length containing node activation delays, with corresponding to the activation delay for node .

—the vector of length indicating which nodes have publicly activated at time . If has publicly activated at , then we write , else .

—the vector of length indicating which node critical values have been met at time , but that have not necessarily publicly activated due to an activation delay. We write if ’s critical value has been exceeded at , otherwise .

—the vector of length indicating each node’s exposure at time . In other words, how many active neighbors does have at time .
Assumptions
Assumption 1.
Granularity: The researcher sees every state the contagion process takes on.
Assumption 2.
Saturation: All nodes in will activate eventually if they continue to update.
Assumption 3.
No updates during activation delays: Nodes do not continue to update while already satisfied and waiting to activate.
Definitions
Definition 1.
An element is exogenous when the researcher does not have control over it. and are always exogenous, while and may or may not be exogenous.
Definition 2.
Node ’s critical value is precisely measured if, for some , we have and while . In this case, the node’s critical value .
Note: If we can observe , simply replace references to with throughout. This condition is similar to the equivalent condition in the main text.
This nodelevel condition allows formulating a straightforward processlevel condition.
Definition 3.
A contagion process is measurable if we are guaranteed to precisely measure all critical values, regardless of the values of the exogenous elements.
In other words, if are exogenous, then we must show that regardless of the values of these elements, all nodes are guaranteed to be correctly measured. To prove it is not measurable, we need only provide a counterexample. Note that this means that some may produce all correctly measured nodes, but that this will not be certain.
Theorems
Theorem 1.
If are exogenous, the process is not measurable, even if we can observe and .
Proof.
Since are exogenous, we can pick any values for the counterexample. Assume that , , and , . Let so that updates at time 1, and update at time 2, and updates again at time 3. Let , indicating no lags in public activation.
Then node ’s critical value can be 1 or 2. This process can be seen visually in Figure 1B.
∎
Theorem 2.
A contagion process is measurable if are exogenous, are controlled by the researcher, and the researcher observes and .
Proof.
Intuitively, we can choose node update ordering and activation delays such that at each even time period all nodes update (except those in the queue defined below), and at each odd time period exactly one node publicly activates. We use private satisfaction status to determine when individual node critical values are exceeded apart from public activation .
There is a “cold start” problem here, so we can randomly update nodes one at a time until one activates at critical value zero. If the cascade stalls we repeat this cold start process, and by Assumption 2 above we’re confident that the cascade will start again.
Let be a firstinfirstout queue of satisfied but unactivated nodes, where but for all . We add node to the end of if, at some time , updates and while . At each odd time period, we remove the first element of and activate it publicly. This defines an update ordering where each node updates simultaneously at each time in (unless it is already in ). It also defines a where exactly one node activates publicly at each .
To see that such a process is measurable, consider any . Since updates at every time until it is put in the queue , and we see at each , find . By construction, exactly one node activated publicly at , which implies that ’s true critical value is , ’s exposure at .
∎
We require the ability to construct and as the process is unfolding. This type of centralized control does not exist in social networks, which are decentralized by definition.
We note that there are likely other constructions of and that satisfy the measurability condition. And, as we noted above, there may be other contagion processes not discussed here that allow measurability.
References
 Sinan Aral, Lev Muchnik, and Arun Sundararajan. Distinguishing influencebased contagion from homophilydriven diffusion in dynamic networks. Proceedings of the National Academy of Sciences, 106(51):21544–21549, 2009.
 Sinan Aral and Dylan Walker. Identifying influential and susceptible members of social networks. Science, 337(6092):337–341, 2012.
 Susan Athey, Dean Eckles, and Guido Imbens. Exact Pvalues for Network Interference. arXiv:1506.02084v1, (June), 2015.
 Lars Backstrom. Group formation in large social networks: membership, growth, and evolution. KDD, pages 44–54, 2006.
 Eytan Bakshy, Dean Eckles, Rong Yan, and Itamar Rosenn. Social Influence in Social Advertising: Evidence from Field Experiments. In EC ’12, pages 146–161, 2012.
 Abhijit Banerjee, Arun G. Chandrasekhar, Esther Duflo, and Matthew O. Jackson. The diffusion of microfinance. Science, 341(July):1236498, 2013.
 AlbertLaszlo Barabasi and Reka Albert. Emergence of Scaling in Random Networks. Science, 286(5439):509–512, October 1999.
 Vladimir Barash, Christopher Cameron, and Michael Macy. Critical phenomena in complex contagions. Social Networks, 34(4):451–461, October 2012.
 Javier BorgeHolthoefer, Raquel A. Banos, Sandra GonzalezBailon, and Yamir Moreno. Cascading behaviour in complex sociotechnical networks. Journal of Complex Networks, 1(1):3–24, June 2013.
 Yann Bramoulle, Habiba Djebbari, and Bernard Fortin. Identification of peer effects through social networks. Journal of Econometrics, 150(1):41–55, May 2009.
 Ellsworth Campbell and Marcel SalathÃ©. Complex social contagion makes networks more vulnerable to disease outbreaks. Scientific Reports, 3:1905, 2013.
 Damon Centola, Victor M. Eguiluz, and Michael W. Macy. Cascade dynamics of complex propagation. Physica A: Statistical Mechanics and its Applications, 374(1):449–456, 2007.
 Damon Centola and Michael Macy. Complex Contagions and the Weakness of Long Ties. American Journal of Sociology, 113(3):702–734, 2007.
 Damon M Centola. The Spread of Behavior in an Online Social Network Experiment. Science, 1194(September):1194–1198, 2010.
 Justin Cheng, Lada Adamic, P. Alex Dow, Jon M. Kleinberg, and Jure Leskovec. Can cascades be predicted? WWW, pages 925–936, 2014.
 Nicholas A. Christakis and James H. Fowler. The Spread of Obesity in a Large Social Network Over 32 Years. New England Journal of Medicine, 357(4):370–9, 2007.
 Ryan Compton, David Jurgens, and David Allen. Geotagging One Hundred Million Twitter Accounts with Total Variation Minimization. arXiv Preprint. arXiv1404.7152, 2014.
 David Crandall, Dan Cosley, Daniel Huttenlocher, Jon Kleinberg, and Siddharth Suri. Feedback effects between similarity and social influence in online communities. KDD, page 160, 2008.
 Giacomo De Giorgi, Michele Pellizzari, and Silvia Redaelli. Identification of Social Interactions through Partially Overlapping Peer Groups. American Economic Journal, 2:241–275, 2010.
 Paul J. DiMaggio and Filiz Garip. Network Effects and Social Inequality. Annual Review of Sociology, 38(1):93–118, 2012.
 Dean Eckles, Brian Karrer, and Johan Ugander. Design and analysis of experiments in networks: Reducing bias from interference. arXiv:1404.7530, page 29, 2014.
 Noah E. Friedkin and Eugene C. Johnsen. Social Influence and Opinions. Journal of Mathematical Sociology, 15(34):193–205, 1990.
 Sandra GonzalezBailon, Javier BorgeHolthoefer, Alejandro Rivero, and Yamir Moreno. The Dynamics of Protest Recruitment through an Online Network. Scientific Reports, 1(1), December 2011.
 Mark Granovetter. Threshold models of collective behavior. American Journal of Sociology, 83(6):1420–1443, 1978.
 Aric A. Hagberg, Daniel A. Schult, and Pieter J. Swart. Exploring Network Structure, Dynamics, and Function using NetworkX. SciPy, pages 11–16, 2008.
 Herbert W. Hethcote. The mathematics of infectious diseases. SIAM review, 42(4):599–653, 2000.
 Petter Holme and Beom Jun Kim. Growing scalefree networks with tunable clustering. Physical review E, 65(2):026107, 2002.
 David Kempe, Jon Kleinberg, and Eva Tardos. Maximizing the spread of influence through a social network. KDD, page 137, 2003.
 David Kempe, Jon Kleinberg, and Eva Tardos. Influential Nodes in a Diffusion Model for Social Networks. ICALP, 3580:1127–1138, 2005.
 A.S. Klovdahl, J.J. Potterat, D.E. Woodhouse, J.B. Muth, S.Q. Muth, and W.W. Darrow. Social networks and infectious disease: The Colorado Springs study. Social Science & Medicine, 38(1):79–88, January 1994.
 Kristina Lerman. Information Is Not a Virus, and Other Consequences of Human Cognitive Limits. Future Internet, 8(2):21, May 2016.
 Michael W. Macy. Chains of Cooperation: Threshold Effects in Collective Action. American Sociological Review, 56(6):730–747, 1991.
 Charles F. Manski. Identification of Endogenous Social Effects: The Reflection Problem. Review of Economic Studies, 60(3):531, 1993.
 Romualdo PastorSatorras, Claudio Castellano, Piet Van Mieghem, and Alessandro Vespignani. Epidemic processes in complex networks. Reviews of modern physics, 87(3):925, 2015.
 Fabian Pedregosa, GaÃ«l Varoquaux, Alexandre Gramfort, Vincent Michel, Bertrand Thirion, Olivier Grisel, Mathieu Blondel, Peter Prettenhofer, Ron Weiss, Vincent Dubourg, and others. Scikitlearn: Machine learning in Python. Journal of Machine Learning Research, 12(Oct):2825–2830, 2011.
 Ronald Read, C. and Robin J. Wilson. An Atlas of Graphs. Clarendon Press, 1998.
 Daniel M Romero, Brendan Meeder, and Jon Kleinberg. Differences in the mechanics of information diffusion across topics: idioms, political hashtags, and complex contagion on twitter. WWW, pages 695–704, 2011.
 Greg Ver Steeg, Rumi Ghosh, and Kristina Lerman. What stops social epidemics? arXiv preprint arXiv:1102.1985, 2011.
 David Strang and Sarah A. Soule. Diffusion in Organizations and Social Movements: From Hybrid Corn to Poison Pills. Annual Review of Sociology, 24(1):265–290, 1998.
 Steven H. Strogatz. Exploring complex networks. Nature, 410(6825):268–276, 2001.
 Johan Ugander, Lars Backstrom, Cameron Marlow, and Jon Kleinberg. Structural diversity in social contagion. Proceedings of the National Academy of Sciences, 109(16):5962–5966, April 2012.
 Thomas W. Valente. Network Models of the Diffusion of Innovations. Cresskill New Jersey Hampton Press, 1995.
 Thomas W. Valente. Social network thresholds in the diffusion of innovations. Social Networks, 18(1):69–89, 1996.
 Thomas W. Valente. Network Interventions. Science, 337(6090):49–53, 2012.
 Susan van den Hof, Christine M. A. Meffre, Marina A. E. Conynvan Spaendonck, Frits Woonink, Hester E. de Melker, and Rob S. van Binnendijk. Measles outbreak in a comunity with very low vaccine coverage, the Netherlands. Emerging Infectious Diseases, 7(3):593–597, 2001.
 Duncan J. Watts. A simple model of global cascades on random networks. Proceedings of the National Academy of Sciences, 99(9):5766–5771, 2002.
 Duncan J. Watts and Steven H. Strogatz. Collective dynamics of ’smallworld’ networks. Nature, 393(6684):440–2, June 1998.
 Lilian Weng, Filippo Menczer, and YongYeol Ahn. Virality Prediction and Community Structure in Social Networks. Scientific Reports, 3, August 2013.