Introduction

According to the privacy paradox, the way users share information online is erratic (Barnes 2006): Although being strongly concerned about their privacy, people communicate much personal information online (Taddicken and Jers 2011). However, despite its popularity in both academic research (Philipp K. Masur 2023) and the public press (New York Public Radio 2018), empirical support for the privacy paradox is tenuous (Baruh, Secinti, and Cemalcilar 2017). Research is increasingly building on the privacy calculus model (Dienlin 2023), which states that communication online can be explained by perceived risks and expected benefits (Bol et al. 2018; Krasnova et al. 2010).

Although the privacy calculus has gained momentum in academic research, several important questions remain unanswered. First, the privacy calculus’ empirical foundation needs to be improved (Knijnenburg et al. 2017). To date, much research on the privacy calculus builds on (a) self-reports of behavior (Krasnova et al. 2010), (b) vignette approaches (Bol et al. 2018), or (c) one-shot experiments in the lab (Trepte, Scharkow, and Dienlin 2020). Self-reports are often unreliable measures of online behavior (Parry et al. 2021), vignette approaches have limited external validity, and one-shot experiments cannot analyze effects of sustained use. The first aim is therefore to replicate the privacy calculus model in a more authentic and long-term setting using actual behavioral data. Doing so will show its robustness, and it will help understand if behavior is better explained by the privacy paradox approach or by the privacy calculus model.

Second, the privacy calculus model is criticized for neglecting the actual process of weighing pros and cons, and for over-emphasizing rationality (Knijnenburg et al. 2017). According to critics, simply showing that both concerns and gratifications correlate with communication online does not prove that an explicit weighing process took place. Instead, we would need to find out if people deliberately compare costs and benefits before communicating online. Therefore, I will also analyze the process of comparing pros and cons explicitly, by elaborating on what I will call the the privacy deliberation process (Omarzu 2000). Next, focusing exclusively on costs and benefits might not fully capture how and why users engage online. I, hence, extend the privacy calculus by analyzing how online communication is affected by trust and self-efficacy, two variables less focused on rationality and more oriented toward socio-psychological aspects critical in online contexts (Metzger 2004).

Online behavior is determined by the person itself and their external circumstances (Bazarova and Masur 2020; Trepte, Scharkow, and Dienlin 2020; Spottswood and Hancock 2017). The privacy calculus upholds that much of behavior is rational and self-determined, but how strongly does this process depend on online affordances? It is well known that the affordances of many online services are optimized to elicit as much interaction as possible (Ellison and Vitak 2015; Philipp K. Masur, DiFranzo, and Bazarova 2021), for example via low threshold communication features such as likes, shares, replies, or reactions (Carr, Hayes, and Sumner 2018). Implicit and explicit cues on how a website is used can increase communication (Spottswood and Hancock 2017; Trepte, Scharkow, and Dienlin 2020). How large are the effects of external popularity cues such as like and dislike buttons compared to the internal weighing of pros and cons? And how easily can the mechanisms of the privacy calculus be affected by means of popularity cues?

In conclusion, in this paper we contribute to communication theory by means of (a) replication, (b) falsification, (c) extending the range of existing theory, (d) elucidating mechanisms, and (e) theory comparison (DeAndrea and Holbert 2017).

Replicating the Privacy Calculus

The privacy calculus analyzes why people communicate online. When are we willing to engage in a conversation? It builds on the calculus of behavior (Laufer and Wolfe 1977) and states that people are weighing risks and benefits before actively communicating. Communication is closely related to privacy, which is defined as a “voluntary and temporary withdrawal of a person from the general society” (Westin 1967, 7). People regulate their privacy by deciding what and what not to communicate (Dienlin 2014). Communication is therefore also closely related to self-disclosure. Just as it is impossible to not communicate, communication inherently promotes self-disclosure. In a recent study, communication quantity and the frequency of expressing one’s political opinion was almost indistinguishable (r = .91; AUTHORS). Next to breadth and depth, communication quantity is hence often considered a central dimension of self-disclosure (Omarzu 2000).

Sharing information carries risks, as recipients may reuse it in different contexts, potentially harming the original sender (Philipp K. Masur, DiFranzo, and Bazarova 2021). As a result, people who aim to avoid risks should be inclined to share less information, especially in online contexts where audiences are much larger (Vitak 2012). Indeed, empirical research confirms that people with higher privacy needs and privacy concerns are less likely to communicate online (Krasnova et al. 2010; Philipp K. Masur and Trepte 2021; Philipp K. Masur 2023). This finding is now confirmed by several meta-analyses and reviews (Baruh, Secinti, and Cemalcilar 2017; Dienlin and Sun 2021; Gerber, Gerber, and Volkamer 2018). Similarly, people who are more concerned about their privacy also engage in more privacy protection behaviors (Baruh, Secinti, and Cemalcilar 2017; Stubenvoll et al. 2022).

Even outweighing concerns, the most relevant drivers of online communication are expected gratifications (Dienlin and Metzger 2016; Bol et al. 2018; Kezer, Dienlin, and Baruh 2022). On average, people are happy to trade in parts of their privacy to receive something more valuable in return (Laufer and Wolfe 1977). In online communication, the most important benefits include social support, social capital, entertainment, information-seeking, and self-presentation (Dhir and Tsai 2017; Ellison, Steinfield, and Lampe 2007; Krasnova et al. 2010; Whiting and Williams 2013).

H1: People who are more concerned about their privacy are less likely to communicate actively on a website.

H2: People who obtain more gratifications from using a website are more likely to communicate actively on a website.

Extending the Privacy Calculus

Although privacy calculus implies that people explicitly compare benefits and risks before communicating online, prior research has neglected this aspect (Knijnenburg et al. 2017). Only observing that privacy concerns or expected gratifications and communication online are related does not prove that an explicit weighing process took place (Knijnenburg et al. 2017). Instead, we need to analyze if, and if so by how much, people actively deliberate about their privacy by comparing benefits and risks, and whether doing so influences their willingness to communicate. Self-disclosure theory (Omarzu 2000; Altman 1976) suggests that if the benefits of communication are attractive, deciding whether or not to communicate is a “conscious and deliberative process” (Omarzu 2000, 183). I hence introduce and investigate a novel concept termed privacy deliberation. Privacy deliberation captures the extent to which individuals explicitly compare risks and benefits before communicating with others.

How could deliberating about one’s privacy affect communication? On the one hand, it could reduce subsequent communication. Refraining from communication—the primary means of connecting with others (Altman 1976)—often requires restraint (Omarzu 2000). This is especially true for social media, which are designed to foster communication and participation (Ellison and Vitak 2015; Philipp K. Masur, DiFranzo, and Bazarova 2021). Actively thinking about whether communicating is worthwhile might be the first step not to participate. In addition, actively reflecting about one’s behavior represents a central and Type 2 approach toward decision making, which is often associated with more critical and cautious behavior (Kahneman 2011; Petty and Cacioppo 1986).

On the other hand, deliberating about privacy might also increase communication. The default behavior in online contexts is passively browsing the content, but not actively engaging in communication (Ozimek, Brailovskaia, and Bierhoff 2023). Especially in new contexts without prior experience, actively pondering one’s options might trigger users to leave their default state of passiveness and to become active and involved. In light of the numerous benefits mentioned above, it might make sense to conclude that participation is beneficial, thereby fostering communication (Krasnova et al. 2010).

RQ1: Do people who deliberate more actively about their privacy communicate more or less online?

It is useful to understand the privacy calculus from the perspective of bounded rationality (Simon 1990). Bounded rationality states that “(1) humans are cognitively constrained; (2) these constraints impact decision making; and (3) difficult problems reveal the constraints and highlight their significance.” (Bendor 2015, 1303). It is important to emphasize that although human behavior is considered partly irrational, bounded rationality does not state that it is completely irrational (Gigerenzer, Selten, and Workshop 2002). Instead, rationality needs to be understood as a continuum. And in the context of online privacy, rationality is impeded by information asymmetries, presence bias, intangibility, illusory control, or herding (Acquisti, Brandimarte, and Loewenstein 2020). It follows that to provide a more complete picture, additional factors less focused on rationality but more on socio-psychological aspects should also explain communication.

Two central factors that help us understand online communication are self-efficacy and trust (Hossain and Wigand 2004; Metzger 2004). Privacy violations create psychological distress (Ledbetter 2019). Experiencing online contexts as a safe space that users can sufficiently control is important for engaging in online communication. If users are more familiar, experienced, and knowledgeable in a given online context, they are more likely to navigate that online contexts successfully and to communicate actively (Baruh, Secinti, and Cemalcilar 2017; Krämer and Schäwel 2020; Park 2013). People with more privacy self-efficacy engage more successfully in self-withdrawal (Dienlin and Metzger 2016) and protective online behavior (Boerman, Kruikemeier, and Zuiderveen Borgesius 2021; van Ooijen, Segijn, and Opree 2022). Hence, if users possess more self-efficacy to participate, they should also communicate more.

H3: People are more likely to communicate on a website when their self-efficacy to actively use the website is higher.

In all situations where people lack experience, control, or competence, a central variable to understand behavior is trust (Gefen, Karahanna, and Straub 2003). Trust plays a key role especially in online contexts (Metzger 2004, 2006). Users often cannot control the environment or the way their information is handled (Acquisti, Brandimarte, and Loewenstein 2020; Bräunlich et al. 2020). Trust either captures “specific beliefs dealing primarily with the integrity, benevolence, and ability of another party” (Gefen, Karahanna, and Straub 2003, 55, emphasis added) or a “general belief that another party can be trusted” (Gefen, Karahanna, and Straub 2003, 55, emphasis added). In online contexts, there are different targets of trust, including (a) the information system, (b) the provider, (c) the Internet, and (d) the community of other users (Söllner, Hoffmann, and Leimeister 2016). People who put more trust in the providers of networks, for example, disclose more personal information (Li 2011). To comprehensively capture and understand online communication, trust should hence be included.

H4: People are more likely to communicate on a website when they have greater trust in the provider, the website, and the other users.

Analyzing the Impact of Popularity Cues

How are the privacy calculus and communication affected by the context, the digital infrastructure? How easily can the calculus be manipulated externally? One of the central tools to afford and govern online behavior are popularity cues such as like and dislike buttons (Stsiampkouskaya, Joinson, and Piwek 2023). Popularity cues have been shown to affect behavior (Krämer and Schäwel 2020; Philipp K. Masur, DiFranzo, and Bazarova 2021; Trepte, Scharkow, and Dienlin 2020). For example, online comments that already have several dislikes are more likely to receive further dislikes (Muchnik, Aral, and Taylor 2013). When users disagree with a post, they are more likely to click on a button labeled respect compared to a button labeled like (Stroud, Muddiman, and Scacco 2017).

How and why might popularity cues affect the privacy calculus? In analyzing this question, it makes sense to analyze the cues’ underlying affordances (Ellison and Vitak 2015; Fox and McEwan 2017). Affordances are mental representations of how objects might be used. They emphasize that it is not the objective features that determine behavior, but rather our subjective perceptions (Gibson 2015). Popularity cues such as like and dislike buttons, which are of primary interest in this study, are understood as “paralinguistic digital affordances” (Carr, Hayes, and Sumner 2018, 142), lowering thresholds to partake in online communication.

Popularity cues likely impact the privacy calculus via two underlying theoretical mechanisms (Krämer and Schäwel 2020): First, the mere presence of popularity cues might affect whether people are willing to disclose. Being able to attract likes might afford active communication. The mere option to receive dislikes, conversely, might inhibit communication. Second, actually receiving likes or dislikes might affect behavior, by means of positive reinforcement (likes) or punishment (dislikes) (Skinner 2014). To illustrate, likes are affirmative and embody the positivity bias of social media (Schreurs, Meier, and Vandenbosch 2022). Receiving a like online is similar to receiving a compliment offline (Carr, Hayes, and Sumner 2018; Sumner, Ruge-Jones, and Alcorn 2017). Like buttons afford and emphasize a gain frame (Rosoff, Cui, and John 2013). These gains can be garnered only through active participation. In situations where people can gain immediate positive outcomes, concerns and risks that are more vague and in the future often become less relevant (presence bias, Ainslie and Haslam 1992). Because like buttons emphasize positive outcomes, it is likely that concerns decrease. In situations where there is more to win, people might also more actively deliberate about whether or not to communicate.

Dislikes, instead, represent a punishment, introducing a loss frame. Websites featuring both like and dislike buttons should therefore be more ambivalent compared to websites without popularity cues, fostering privacy deliberation. Privacy concerns should not be reduced anymore: People who need more privacy are also more shy and risk averse (Dienlin and Metzger 2024). Implementing the dislike button might therefore increase privacy concerns, canceling out the positive effects of the like button. At the same time, communication and benefits might still be increased compared to a website without like and dislikes buttons, as online benefits are often considered to outweigh risks (positivity bias, Schreurs, Meier, and Vandenbosch 2022).

H5. Compared to a control group with no like and dislike buttons, people who use a website with like buttons (a) communicate more; (b) obtain more gratifications; (c) are less concerned about their privacy; and (d) deliberate more about whether they should communicate online.

H6. Compared to a control group with no like and dislike buttons, people who use a website with like and dislike buttons (a) communicate more; (b) obtain more gratifications; and (c) deliberate more about whether they should communicate online.

H7. Compared to people who use a website with only like buttons, people who use a website with like and dislike buttons (a) are more concerned about their privacy, and (b) deliberate more about whether they should communicate online.

In conclusion, this leads to an updated and extended model of the privacy calculus, in which the affordances of online contexts affect communication mediated via experienced privacy concerns, expected gratifications, privacy deliberation, self-efficacy, and trust. For an overview of the theoretical model, see Figure @ref(fig:model).

Overview of the extended privacy calculus model.

Overview of the extended privacy calculus model.

Methods

Open Science

This manuscript features a companion website, which includes the data, research materials, analysis scripts, and a reproducible version of this manuscript (see https://XMtRa.github.io/like_dislike). The hypotheses, sample size, research materials, analyses, and exclusion criteria were preregistered (see https://osf.io/a6tzc/?view_only=5d0ef9fe5e1745878cd1b19273cdf859). In some cases, the preregistered approach had to be changed (see companion website). Analyses not preregistered are reported as exploratory analyses.

Procedure

The study was designed as an online field experiment with three different groups. The first experimental group used a website that included like buttons; the second experimental group used an identical website including both like and dislike buttons; and the control group used an identical website without like and dislike buttons. Participants were randomly distributed to one of the three websites in a between-subject design.

The data were collected in Germany. Participants were recruited using the professional panel agency Norstat. As incentive, participants were awarded digital points, to receive special offers from online retailers. Participants had to be over 18 years old and reside in Germany. In a first step, the company sent its panel members an invitation to participate in the study. In this invitation, panel members were asked to participate in a study analyzing the current threat posed by terrorist attacks. Members who agreed to participate were sent the first questionnaire (T1). The questionnaire asked participants about their sociodemographic background, (b) provided more details about the study, and (c) included a registration link for the website, which was introduced as “participation platform”. Afterward, participants were randomly assigned to one of the three websites. After registration was completed, participants were invited (but not obliged) to visit the website and to discuss the topic of the terrorism threat in Germany. Participants could use the website and write comments over the course of one week. Subsequently, participants received a follow-up questionnaire in which the self-reported measures reported below were collected (T2).

Screenshot of the landing page and of communication taking place. Translated into English.Screenshot of the landing page and of communication taking place. Translated into English.

Screenshot of the landing page and of communication taking place. Translated into English.

The online website was programmed based on the open-source software discourse (https://www.discourse.org/). To make sure the website was professional and authentic, several pretests with students from the local university were run. Nine hundred sixty participants created a user account on the website (see below) and used the website actively. Overall, they spent 162 hours online, wrote 1,171 comments, and clicked on 560 popularity cues. All communication was checked, and there were no instances of people providing meaningless text or doubting the experiment. For a screenshot of the landing page and for examples of comments, see Figure @ref(fig:website).

Participants

Sample size was determined using a priori power analyses. The power analyses were based on estimates from the literature. When researching aspects of privacy online (e.g., Baruh, Secinti, and Cemalcilar 2017), effects are often small (i.e., r = .10, Cohen 1992). Hence, the minimum effect size was set to be r = .10. The aim was to be able to detect this effect with a probability of at least 95% (i.e., power = 95%). Using the regular alpha level of 5%, power analyses suggested a minimum sample size of N = 1,077. In the end, I was able to include N = 559 in the analyses (see below), which was significantly lower than the original aim. With this sample size, the study had a power of 77% to find an effect at least as large as r = .10. Sensitivity analyses showed that the current study could still make reliable results (i.e., with a power of 95%) for effects at least as large as r = .14. In conclusion, although not as powerful as planned, the study is still adequately powered to find the small effects reported in the privacy literature (Baruh, Secinti, and Cemalcilar 2017).

A quota sample that matched the German population in terms of age, gender, and federal state was collected. In sum, 1,619 participants completed the survey at T1; 960 participants created a user account on the website; and 982 participants completed the survey at T2. The data were connected using tokens and IP addresses. For technical reasons, the data of several participants could not be matched (for example, because they used different devices for the respective steps). In the end, the data of 590 participants could be matched successfully. Considered unreasonably fast, 29 participants were excluded who finished the questionnaire at T2 in less than three minutes. To detect atypical data and response patterns, Cook’s distance was calculated. I excluded 2 participants with clear response patterns (i.e., straight-lining). The final sample included N = 559 participants. The sample characteristics at T1 and T2 were as follows. T1: age = 45 years, gender = 49% male, college degree = 22%. T2: age = 46 years, gender = 49% male, college degree = 29%. One participant did not report their gender.

Measures

Factor validity was assessed using confirmatory factor analyses (CFA). If the CFAs revealed insufficient fit, malfunctioning items were deleted. All items were measured on bipolar 7-point scales. Answer options were visualized as follows: -3 (strongly disagree), -2 (disagree), -1 (slightly disagree), 0 (neutral), +1 (slightly agree), +2 (agree), +3 (strongly agree). For the analyses, answers were coded from 1 to 7. All items measuring the same variable were presented in randomized order on the same page.

All measures showed high factorial validity. For an overview of the means, standard deviations, factorial validity, and reliability, see Table @ref(tab:CFA). For the variables’ distributions, see Figure @ref(fig:corrplot). For all items and their distributions, see companion website.

(#tab:CFA) Psychometric Properties, Factorial Validity, and Reliability of Measures
m sd chisq df p-value cfi tli rmsea srmr omega ave
Privacy concerns 3.21 1.51 11.04 9 0.27 1.00 1.00 0.02 0.01 0.96 0.80
Gratifications 4.76 1.22 34.03 5 0.00 0.98 0.95 0.10 0.02 0.93 0.74
Privacy deliberation 3.93 1.29 15.55 5 0.01 0.98 0.96 0.06 0.02 0.85 0.53
Self-efficacy 5.21 1.04 3.23 1 0.07 0.99 0.96 0.06 0.01 0.83 0.59
Trust general 5.08 0.94 2.07 1 0.15 1.00 0.99 0.04 0.01 0.87 0.70
Trust specific 5.25 1.12 89.11 24 0.00 0.96 0.94 0.07 0.04 0.93 0.61

Note. omega = Raykov’s composite reliability coefficient omega; avevar = average variance extracted.

 

Privacy concerns were measured with seven items based on Buchanan et al. (2007). One example item was “When using the participation platform, I had concerns about my privacy”. One item was deleted due to poor psychometric properties. The mean was m = 3.21. This value is below the scale’s midpoint of 4, showing that on average people were not strongly concerned about their privacy.

General gratifications were measured with five items based on Sun et al. (2015). One example item was “Using the participation platform has paid off for me”. The mean was m = 4.76, which was above the scale’s midpoint. This shows that on average people considered the website to be beneficial. For exploratory analyses, we also collected a scale measuring specific gratifications, not included here.

Privacy deliberation was measured with five self-designed items. One example item was “While using the participation platform I have weighed the advantages and disadvantages of writing a comment.” The mean lay on the scale’s neutral midpoint (m = 3.93). (For an interpretation, see below.)

Self-efficacy was captured with six self-designed items, which measured whether participants felt that they had sufficient self-efficacy to write a comment on the website. For example, “I felt technically competent enough to write a comment.” Two inverted items were deleted due to poor psychometric properties. People felt self-efficacious to use the website (m = 5.21).

Two types of trust were measured. General trust was operationalized based on Söllner, Hoffmann, and Leimeister (2016), addressing three targets (i.e., provider, website, and other users), measured with one item each. One example item was “The operators of the participation platform seemed trustworthy.” Specific trust was operationalized for the same three targets with three sub-dimensions each (i.e., ability, benevolence/integrity, and reliability), measured with one item each. Example items were “The operators of the participation platform have done a good job” (ability), “The other users had good intentions” (benevolence/integrity), “The website worked well” (reliability). Participants placed a lot of trust in the website, the providers and the other users (trust general: m = 5.08; trust specific: m = 5.25).

Communication was calculated by counting the number of words each participant wrote in a comment. Communication was zero-inflated and heavily skewed: While 58 percent did not communicate at all, the maximum number of words communicated by a single user was 3198 words. On average, participants wrote 77 words.

Data Analysis

As preregistered, all hypotheses and research questions were initially tested using structural equation modeling with latent variables. The influence of the three websites was analyzed using contrast coding. Because the assumption of multivariate normality was violated, I estimated the models using robust maximum likelihood (Kline 2016). As recommended by Kline (2016), to assess global fit I report the model’s \(\chi^2\), RMSEA (90% CI), CFI, and SRMR. To exclude confounding influences, I controlled all variables for age, gender, and education, which have been shown to affect both privacy concerns and online communication (Philipp K. Masur 2023; Tifferet 2019). The preregistered hypotheses were tested with a one-sided significance level of 5%; the research questions were tested with a two-sided 5% significance level using family-wise Bonferroni-Holm correction.

As became apparent when analyzing the data, the preregistered analyses had two major problems. First, communication was zero-inflated and gamma distributed. Although it is possible to analyze non-normal data with structural equation modeling, it is recommended to use analyses that model the variables’ distribution, which can be achieved with Bayesian hurdle models (McElreath 2016). In conclusion, in the exploratory analyses I ran (unstandardized) Bayesian hurdle regression models, modeling the outcome as a zero-inflated gamma distribution using default (flat) priors (chains = 4, iterations = 2,000, warm-up = 1,000, Bürkner 2017). Second, in the preregistered analyses several variables were combined that are theoretically and empirically closely related, leading to multicollinearity (Vanhove 2021). As a remedy, I adopted a causal modeling perspective, controlling only for confounders—in this case, age, gender, and education—but not for mediators (Rohrer 2018). To assess the effects, I tested whether or not the 95% highest density intervals of the average marginal effects excluded zero. If they excluded zero, effects can be considered “significant” (McElreath 2016). I also plotted the distribution of the effects. For more information on the fitted models, see online companion website.

The data were analyzed using R (Version 4.4.0; R Core Team 2018) and the R-packages brms (Version 2.21.0; Bürkner 2017, 2018), lavaan (Version 0.6.18; Rosseel 2012), marginaleffects (Arel-Bundock 2023), papaja (Version 0.1.2; Aust and Barth 2018), pwr (Version 1.3.0; Champely 2018), quanteda (Benoit 2018), semTools (Version 0.5.6; Jorgensen et al. 2018), and tidyverse (Version 2.0.0; Wickham 2017).

Results

Descriptive Analyses

Above diagonal: zero-order correlation matrix; diagonal: density plots for each variable; below diagonal: bivariate scatter plots for zero-order correlations. Solid regression lines represent linear regressions, dotted regression lines represent quadratic regressions. Calculated with the model predicted values for each variable (baseline model).

Above diagonal: zero-order correlation matrix; diagonal: density plots for each variable; below diagonal: bivariate scatter plots for zero-order correlations. Solid regression lines represent linear regressions, dotted regression lines represent quadratic regressions. Calculated with the model predicted values for each variable (baseline model).

I first plotted the bivariate relations of all variables (see Figure @ref(fig:corrplot)). All variables referring to the privacy calculus demonstrated the expected bivariate relationships with communication. For example, people who were more concerned about their privacy disclosed less information (r = -.15). The mean of privacy deliberation was m = 3.93. Altogether, 32% of participants reported having actively deliberated about their privacy.

The bivariate results showed two large correlations: specific trust and gratifications (r = .79), and privacy concerns and privacy deliberation (r = .61). As all six variables were later analyzed within a single multiple regression, problems of multicollinearity might occur.

Preregistered Analyses

First, as preregistered I ran a structural equation model with multiple predictors. The model fit the data okay, \(\chi^2\)(387) = 934.98, < .001, cfi = .94, rmsea = .05, 90% CI [.05, .05], srmr = .05. With regard to H1, privacy concerns did not significantly predict communication (\(\beta\) > -.01, = -0.01, 95% CI [-0.30, 0.28], = -0.07, = .473; one-sided). Regarding H2, results showed that gratifications did not predict communication (\(\beta\) = -.02, = -0.04, 95% CI [-0.19, 0.12], = -0.45, = .327; one-sided). RQ1 similarly revealed that privacy deliberation did not predict communication (\(\beta\) = -.10, = -0.16, 95% CI [-0.35, 0.02], = -1.76, = .078; two-sided). Regarding H3, however, I found that experiencing self-efficacy predicted communication substantially (\(\beta\) = .37, = 0.77, 95% CI [0.49, 1.05], = 5.42, < .001; one-sided). Concerning H4, results showed that trust was not associated with communication (\(\beta\) = -.03, = -0.09, 95% CI [-0.56, 0.38], = -0.36, = .358; one-sided).

However, these results should be treated with caution. I found several signs of multicollinearity, evidenced by the large standard errors or “wrong” and reversed signs of predictors (Vanhove 2021). For example, in the bivariate analysis trust had a positive relation with communication, whereas in the multiple regression the effect was negative—which should make us skeptical.

Next, I analyzed the effects of the popularity cues. It was for example expected that websites with like buttons would lead to more communication, more gratifications, more privacy deliberation, and less privacy concerns. The results showed that the popularity cues had no effects on communication and on the privacy calculus variables.

For an illustration, see Figure @ref(fig:popularitycues). For the detailed results of the specific inference tests using contrasts, see companion website.

Distributions and means with 95\% CIs of the model-predicted values for each variable, separated for the three websites. Control: Website without buttons. Like: Website with like buttons. Like \& Dislike: Website with like and dislike buttons. Values from preregistered SEM.

Distributions and means with 95% CIs of the model-predicted values for each variable, separated for the three websites. Control: Website without buttons. Like: Website with like buttons. Like & Dislike: Website with like and dislike buttons. Values from preregistered SEM.

Exploratory Analyses

As explained above, the preregistered results were problematic. Communication was not normally distributed and the predictors were collinear. I hence updated the analyses, using Bayesian hurdle models controlling only for confounders but not mediators. In predicting communication, I opted for general trust over specific trust, as it exhibited lower levels of empirical and theoretical overlap with gratifications. The updated exploratory analyses showed different results.

Exploratory analyses. Plotted are the average marginal effects of the Bayesian hurdle models. The difference between the control and the like \& disklike group is significant.

Exploratory analyses. Plotted are the average marginal effects of the Bayesian hurdle models. The difference between the control and the like & disklike group is significant.

Hypotheses 1, 2, 3, and 4 were all confirmed: If participants were more concerned about their privacy, they communicated less: With each one-point increase in privacy concerns, on average they wrote 16 words less (95% HDI: -28, -5). If participants expected more gratifications from participation, they communicated more actively: If their expected gratifications increased by one point, on average they also wrote 27 words more (95% HDI: 13, 44). If participants felt more self-efficacious, they communicated much more: If their self-efficacy increased by one point, on average they wrote 73 words more (95% HDI: 56, 94). The relationship was curvilinear, almost exponential: Whereas a change in self-efficacy from 1 to 2 only led to an “increase” of 0 words, a change from 6 to 7 led to an increase of 103 words. Next, if participants experienced more trust in the website, provider, and other users, they also communicated much more: If their trust increased by one point, on average they wrote 31 words more (95% HDI: 14, 54). The first research question asked how privacy deliberation would affect communication. The results revealed a negative effect. The more people deliberated about their privacy, the less they communicated. If privacy deliberation increased by one point, on average they wrote 33 words less (95% HDI: -51, -20).

(#tab:results) Unstandardized average marginal effects of the privacy calculus variables and the popularity cues on communication. Credibility intervals excluding zero are considered significant.
Predictor Estimate LL UL
Privacy calculus
   Privacy concerns -16 -28 -5
   Expected gratifications 27 13 44
   Privacy deliberation -33 -51 -20
   Trust 31 14 54
   Self-efficacy 73 56 94
Experimental conditions
   Like & dislike vs. control -48 -93 -10
   Like vs. control -35 -82 4
   Like vs. like & dislike 13 -19 44

Note. HDI = highest density interval, LL = lower level; UL = upper level. Reported are average marginal effects.

 

I then reanalyzed the effects of the popularity cues on communication. Compared to the control condition without popularity cues, implementing like buttons did not significantly affect communication, b = -35 (95% HDI: -82, 4). But note that the effect was very close to not including the zero. However, implementing both like and dislike buttons affected communication. Contrary to what I expected, implementing both popularity cues decreased communication. If both popularity cues were present, participants on average wrote 48 words less (95% HDI: -93, -10). The introduction of both cues hence led to a 45% decline in number of words that were written.

Finally, I tested if the effect of the popularity cues on communication were potentially mediated by the privacy calculus variables. Results showed that this was not the case. The popularity cues only affected behavior but not the predictors of the privacy calculus model (see online companion website). The results suggest that the effect was either direct or mediated by other variables not included here (Coenen 2022).

Discussion

This study analyzed three questions: Can the privacy calculus can be replicated with behavioral data in an authentic setting? Should the privacy calculus model be extended theoretically? Do like and dislike buttons affect online communication and the privacy calculus? To this end, a preregistered field experiment was conducted, which lasted one week. The privacy calculus model was extended: The privacy deliberation processes was tested explicitly, and trust and self-efficacy were included as predictors.

The preregistered analyses showed that the popularity cues did not affect communication. Only self-efficacy emerged as a significant predictor of online communication. All other variables remained insignificant. However, the preregistered analyses have to be treated with caution. The predictors were collinear, which makes their integration in one single model problematic (Vanhove 2021). In addition, the main variable and outcome of the study, number of communicated words, was zero-inflated and gamma distributed, which requires a different type of analysis. The preregistered analyses using structural equation modeling with multiple predictors were hence problematic.

To address these issues, I conducted Bayesian hurdle-gamma models (see section Data analysis). This updated approach changed the results. People who were more concerned about their privacy wrote fewer words. To illustrate, people who reported being very much concerned posted only 32 words on average, whereas people who reported being not concerned posted 115 words. People who experienced privacy concerns hence differed strongly in their communication behavior from people who were unconcerned. Participants who received more gratifications wrote substantially more words. The effect was even larger, almost twice as large. For each point-increase in gratifications, participants wrote 27 words more. Attaining benefits online is hence a strong predictor of online communication. Together, the results provide further support for the privacy calculus and against the privacy paradox (Baruh, Secinti, and Cemalcilar 2017). Communication online does not seem to be overly illogical. To large extents, it is aligned with respondents’ concerns and benefits.

Results showed that trust and self-efficacy were important drivers of online communication. Participants who placed more trust into the website, the providers, and the other community members communicated more actively. Interestingly, self-efficacy was the strongest of all predictors. Participants who felt more self-efficacious disclosed much more than others. To illustrate, if people reported no self-efficacy, they wrote only 1 word on average. However, when they reported high levels of self-efficacy, they wrote 254 words. This finding further supports the underlying premise of bounded rationality (Simon 1990). Although more rational aspects such as costs and benefits influence behavior, behavior is also determined by more socio-psychological variables such as trust and self-efficacy. The findings are therefore closely aligned with existing theory. For example, the technology acceptance model states that online behavior is most strongly determined by usefulness and ease of use (Venkatesh et al. 2003)—two variables closely related to gratifications and self-efficacy.

The privacy calculus was criticized for not explicitly analyzing the process of weighing pros and cons before disclosing (Knijnenburg et al. 2017). In this study, I thus elaborated on the privacy deliberation process. The results showed that only one third of all participants agreed to have actively weighed the benefits and risks before communicating on the platform. This figure seems comparatively low. Even in new online contexts, the majority of users does not actively deliberate about their online communication, suggesting that online use is to large extents implicit (Acquisti, Brandimarte, and Loewenstein 2020).

Interestingly, and perhaps also somewhat surprisingly, privacy deliberation and privacy concerns were strongly correlated (r = .61). If we are concerned we also think and deliberate more actively about our privacy. And if we are not concerned we do not deliberate. This finding can be aligned with decision theory (Elsbach and Barr 1999): When being concerned, we are in a negative state; and when in a negative state, we judge more critically. At this point, it is still unknown if thinking about privacy increases concerns or, conversely, if growing more concerned about privacy makes us deliberate more carefully.

The updated results showed that implementing both like and dislike buttons decreased communication. This was an unexpected and interesting finding. It suggests that negative feedback, or perhaps even only risks of negative feedback, can stifle communication. The effect was strong: Implementing both like and dislikes cues led to a 45% decrease in number of written words. When compared to the privacy calculus variables, we see that the effect is of similar size: Growing a bit more concerned or receiving less gratifications have a comparable impact on behavior. This finding is aligned with studies reporting strong effects of popularity cues on behavior (Muchnik, Aral, and Taylor 2013).

The negative effects of dislike buttons might help explain why almost all existing successful social network sites have chosen to omit negative popularity cues. At the time of writing, only a handful of websites have (partially) implemented dislike buttons (e.g., youtube, stackexchange, or reddit). Despite the positivity bias of social media (Schreurs, Meier, and Vandenbosch 2022), chances of receiving negative feedback and communication are real, as can be seen by moral outrages or “shit-storms”. Explicit negative popularity cues are low threshold paralinguistic affordances (Carr, Hayes, and Sumner 2018). They likely prime or trigger negative experiences or expectations, thereby stifling communication. Interestingly, however, they did not affect the privacy calculus variables, and no indirect effect was found. Hence, the effect is either transmitted via variables not included here (Coenen 2022), or perhaps subconscious and direct.

Websites only including like buttons had no effect on the number of communicated words. If anything, there was an unexpected (non-significant) trend toward reduced communication. Although one might expect that like buttons, being positive feedback cues, increase communication, it is also plausible that they can decrease communication. Not receiving any likes can be perceived as ostracizing (Schneider et al. 2017). In the context of this study, participants discussed a political topic. Here, not receiving likes might be even more threatening and intimidating than on regular social media, where it is more common to discuss every-day and low-threshold topics. Although like buttons are commonplace in social media, the findings suggest that in specific contexts they inhibit communication.

Limitations

The main implications and results discussed above rest on exploratory analyses not registered a priori. Exploratory analyses are part of the research process, compatible with preregistration, and important for scientific progress. The updated analyses represent and document a learning process, which arguably led to an improved analysis. However, the results should still be considered somewhat preliminary, to be confirmed in subsequent studies.

Whereas the effects of the popularity cues on all variables can be interpreted from a causal perspective (but see below), more caution is needed regarding the effects of the privacy calculus variables on communication. Although the effects were controlled for age, gender, and education, other variables not included here could potentially bias the causal estimates (Coenen 2022). In addition, in order not to reveal the study intention the self-reported measures were collected after the field phase. Demand effects might have led participants to align their answers to their prior behavior. To illustrate, users who communicated more actively might have experienced more self-efficacy afterward. As a result, the coefficients might overestimate the actual effects.

In experiments only the treatment variable should be manipulated, while all others should be held constant (assumption of stable unit treatment, Kline 2016). Being a field experiment, several variables could not be held constant, such as the content of communication by other users, the unfolding communication dynamics, and the characteristics of other users. Future research should repeat the design, preferably using several runs of the same experiment, to further assess generalizability and robustness.

Conclusion

This study provides further support for the privacy calculus model and against the privacy paradox approach. Expected benefits, privacy concerns, deliberating about benefits and risks, trust, and self-efficacy all affected communication. Like and dislike buttons reduced communication significantly. Users can be considered proactive and reasonable. But, similar to everyday offline contexts, they are also affected by the affordances of their environment, and often act implicitly without pondering the consequences of their actions.

References

Acquisti, Alessandro, Laura Brandimarte, and George Loewenstein. 2020. “Secrets and Likes: The Drive for Privacy and the Difficulty of Achieving It in the Digital Age.” Journal of Consumer Psychology 30 (4): 736–58. https://doi.org/10.1002/jcpy.1191.
Ainslie, George, and Nick Haslam. 1992. “Hyperbolic Discounting.” In Choice over Time, 57–92. New York, NY, US: Russell Sage Foundation.
Altman, Irwin. 1976. “Privacy: A Conceptual Analysis.” Environment and Behavior 8 (1): 7–29. https://doi.org/10.1177/001391657600800102.
Arel-Bundock, Vincent. 2023. Marginaleffects: Predictions, Comparisons, Slopes, Marginal Means, and Hypothesis Tests. https://CRAN.R-project.org/package=marginaleffects.
Aust, Frederik, and Marius Barth. 2018. papaja: Create APA Manuscripts with R Markdown. https://github.com/crsh/papaja.
Barnes, Susan B. 2006. “A Privacy Paradox: Social Networking in the United States.” First Monday 11 (9). https://doi.org/10.5210/fm.v11i9.1394.
Baruh, Lemi, Ekin Secinti, and Zeynep Cemalcilar. 2017. “Online Privacy Concerns and Privacy Management: A Meta-Analytical Review.” Journal of Communication 67 (1): 26–53. https://doi.org/10.1111/jcom.12276.
Bazarova, Natalya N, and Philipp K Masur. 2020. “Towards an Integration of Individualistic, Networked, and Institutional Approaches to Online Disclosure and Privacy in a Networked Ecology.” Current Opinion in Psychology 36 (December): 118–23. https://doi.org/10.1016/j.copsyc.2020.05.004.
Bendor, John. 2015. “Bounded Rationality.” In International Encyclopedia of the Social & Behavioral Sciences, 773–76. Elsevier. https://doi.org/10.1016/B978-0-08-097086-8.93012-5.
Benoit, Kenneth. 2018. Quanteda: Quantitative Analysis of Textual Data. https://doi.org/10.5281/zenodo.1004683.
Boerman, Sophie C., Sanne Kruikemeier, and Frederik J. Zuiderveen Borgesius. 2021. “Exploring Motivations for Online Privacy Protection Behavior: Insights from Panel Data.” Communication Research 48 (7): 953–77. https://doi.org/10.1177/0093650218800915.
Bol, Nadine, Tobias Dienlin, Sanne Kruikemeier, Marijn Sax, Sophie C. Boerman, Joanna Strycharz, Natali Helberger, and Claes H. Vreese. 2018. “Understanding the Effects of Personalization as a Privacy Calculus: Analyzing Self-Disclosure Across Health, News, and Commerce Contexts.” Journal of Computer-Mediated Communication 23 (6): 370–88. https://doi.org/10.1093/jcmc/zmy020.
Bräunlich, Katharina, Tobias Dienlin, Johannes Eichenhofer, Paula Helm, Sabine Trepte, Rüdiger Grimm, Sandra Seubert, and Christoph Gusy. 2020. “Linking Loose Ends: An Interdisciplinary Privacy and Communication Model.” New Media & Society, February, 1461444820905045. https://doi.org/10.1177/1461444820905045.
Buchanan, Tom, Carina Paine, Adam N. Joinson, and Ulf-Dietrich Reips. 2007. “Development of Measures of Online Privacy Concern and Protection for Use on the Internet.” Journal of the American Society for Information Science and Technology 58 (2): 157–65. https://doi.org/10.1002/asi.20459.
Bürkner, Paul-Christian. 2017. brms: An R Package for Bayesian Multilevel Models Using Stan.” Journal of Statistical Software 80 (1): 1–28. https://doi.org/10.18637/jss.v080.i01.
———. 2018. “Advanced Bayesian Multilevel Modeling with the R Package brms.” The R Journal 10 (1): 395–411. https://doi.org/10.32614/RJ-2018-017.
Carr, Caleb T., Rebecca A. Hayes, and Erin M. Sumner. 2018. “Predicting a Threshold of Perceived Facebook Post Success via Likes and Reactions: A Test of Explanatory Mechanisms.” Communication Research Reports 35 (2): 141–51. https://doi.org/10.1080/08824096.2017.1409618.
Champely, Stephane. 2018. Pwr: Basic Functions for Power Analysis. https://CRAN.R-project.org/package=pwr.
Coenen, Lennert. 2022. “The Indirect Effect Is Omitted Variable Bias. A Cautionary Note on the Theoretical Interpretation of Products-of-Coefficients in Mediation Analyses.” European Journal of Communication, March, 026732312210822. https://doi.org/10.1177/02673231221082244.
Cohen, Jacob. 1992. “A Power Primer.” Psychological Bulletin 112 (1): 155–59. https://doi.org/10.1037/0033-2909.112.1.155.
DeAndrea, David C., and R. Lance Holbert. 2017. “Increasing Clarity Where It Is Needed Most: Articulating and Evaluating Theoretical Contributions.” Annals of the International Communication Association 41 (2): 168–80. https://doi.org/10.1080/23808985.2017.1304163.
Dhir, Amandeep, and Chin-Chung Tsai. 2017. “Understanding the Relationship Between Intensity and Gratifications of Facebook Use Among Adolescents and Young Adults.” Telematics and Informatics 34 (4): 350–64. https://doi.org/10.1016/j.tele.2016.08.017.
Dienlin, Tobias. 2014. “The Privacy Process Model.” In Medien Und Privatheit, edited by S. Garnett, S. Halft, M. Herz, and J. M. Mönig, 105–22. Passau, Germany: Karl Stutz.
———. 2023. “Privacy Calculus: Theories, Studies, and New Perspectives.” In The Routledge Handbook of Privacy and Social Media, edited by Sabine Trepte and Philipp Masur, 70–79. Routledge.
Dienlin, Tobias, and Miriam Metzger. 2024. “Who Needs Privacy? Exploring the Relations Between Need for Privacy and Personality.” Collabra: Psychology 10 (1): 120402. https://doi.org/10.1525/collabra.120402.
Dienlin, Tobias, and Miriam J. Metzger. 2016. “An Extended Privacy Calculus Model for SNSsAnalyzing Self-Disclosure and Self-Withdrawal in a Representative U.S. Sample.” Journal of Computer-Mediated Communication 21 (5): 368–83. https://doi.org/10.1111/jcc4.12163.
Dienlin, Tobias, and Ye Sun. 2021. “Does the Privacy Paradox Exist? Comment on Yu Et Al.’s (2020) Meta-Analysis.” Meta-Psychology 5 (November). https://doi.org/10.15626/MP.2020.2711.
Ellison, Nicole B., Charles Steinfield, and Cliff Lampe. 2007. “The Benefits of Facebook ‘Friends:’ Social Capital and College Students’ Use of Online Social Network Sites.” Journal of Computer-Mediated Communication 12 (4): 1143–68. https://doi.org/10.1111/j.1083-6101.2007.00367.x.
Ellison, Nicole B., and Jessica Vitak. 2015. “Social Network Site Affordances and Their Relationship to Social Capital Processes.” In The Handbook of the Psychology of Communication Technology, edited by S. Shyam Sundar, v.33:205–27. Handbooks in Communication and Media. Chichester, MA: Wiley Blackwell.
Elsbach, Kimberly D., and Pamela S. Barr. 1999. “The Effects of Mood on Individuals’ Use of Structured Decision Protocols.” Organization Science 10 (2): 181–98. https://doi.org/10.1287/orsc.10.2.181.
Fox, Jesse, and Bree McEwan. 2017. “Distinguishing Technologies for Social Interaction: The Perceived Social Affordances of Communication Channels Scale.” Communication Monographs 9 (January): 1–21. https://doi.org/10.1080/03637751.2017.1332418.
Gefen, David, Elena Karahanna, and Detmar W. Straub. 2003. “Trust and TAM in Online Shopping: An Integrated Model.” MIS Q 27 (1): 5190.
Gerber, Nina, Paul Gerber, and Melanie Volkamer. 2018. “Explaining the Privacy Paradox: A Systematic Review of Literature Investigating Privacy Attitude and Behavior.” Computers & Security 77 (August): 226–61. https://doi.org/10.1016/j.cose.2018.04.002.
Gibson, James J. 2015. The Ecological Approach to Visual Perception. New York, NY: Psychology Press.
Gigerenzer, Gerd, Reinhard Selten, and Dahlem Workshop, eds. 2002. Bounded Rationality: The Adaptive Toolbox. 1st ed. Dahlem Workshop Reports. Cambridge, MA: MIT Press.
Hossain, Liaquat, and Rolf T. Wigand. 2004. ICT Enabled Virtual Collaboration Through Trust.” Journal of Computer-Mediated Communication 10 (1): JCMC1014. https://doi.org/10.1111/j.1083-6101.2004.tb00233.x.
Jorgensen, T. D., Pornprasertmanit, S., Schoemann, A. M., Rosseel, and Y. 2018. semTools: Useful Tools for Structural Equation Modeling. https://CRAN.R-project.org/package=semTools.
Kahneman, Daniel. 2011. Thinking, Fast and Slow. New York: Farrar, Straus and Giroux.
Kezer, Murat, Tobias Dienlin, and Lemi Baruh. 2022. “Getting the Privacy Calculus Right: Analyzing the Relations Between Privacy Concerns, Expected Benefits, and Self-Disclosure Using Response Surface Analysis.” Cyberpsychology: Journal of Psychosocial Research on Cyberspace 16 (4). https://doi.org/10.5817/CP2022-4-1.
Kline, Rex B. 2016. Principles and Practice of Structural Equation Modeling. 4th ed. Methodology in the Social Sciences. New York, NY: The Guilford Press.
Knijnenburg, Bart, Elaine Raybourn, David Cherry, Daricia Wilkinson, Saadhika Sivakumar, and Henry Sloan. 2017. “Death to the Privacy Calculus?” SSRN Electronic Journal, January. https://doi.org/10.2139/ssrn.2923806.
Krämer, Nicole C, and Johanna Schäwel. 2020. “Mastering the Challenge of Balancing Self-Disclosure and Privacy in Social Media.” Current Opinion in Psychology, Privacy and Disclosure, Online and in Social Interactions, 31 (February): 67–71. https://doi.org/10.1016/j.copsyc.2019.08.003.
Krasnova, Hanna, Sarah Spiekermann, Ksenia Koroleva, and Thomas Hildebrand. 2010. “Online Social Networks: Why We Disclose.” Journal of Information Technology 25 (2): 109–25. https://doi.org/10.1057/jit.2010.6.
Laufer, Robert S., and Maxine Wolfe. 1977. “Privacy as a Concept and a Social Issue: A Multidimensional Developmental Theory.” Journal of Social Issues 33 (3): 22–42. https://doi.org/10.1111/j.1540-4560.1977.tb01880.x.
Ledbetter, Andrew M. 2019. “Parent-Child Privacy Boundary Conflict Patterns During the First Year of College: Mediating Family Communication Patterns, Predicting Psychosocial Distress.” Human Communication Research 45 (3): 255–85. https://doi.org/10.1093/hcr/hqy018.
Li, Yuan. 2011. “Empirical Studies on Online Information Privacy Concerns: Literature Review and an Integrative Framework.” Communications of the Association for Information Systems 28 (January): 453–96.
Masur, Philipp K. 2023. “Understanding the Effects of Conceptual and Analytical Choices on ‘Finding’ the Privacy Paradox: A Specification Curve Analysis of Large-Scale Survey Data.” Information, Communication & Society 26 (3): 584–602. https://doi.org/10.1080/1369118X.2021.1963460.
Masur, Philipp K., Dominic DiFranzo, and Natalie N. Bazarova. 2021. “Behavioral Contagion on Social Media: Effects of Social Norms, Design Interventions, and Critical Media Literacy on Self-Disclosure.” Plos One 16 (7): e0254670.
Masur, Philipp K, and Sabine Trepte. 2021. “Transformative or Not? How Privacy Violation Experiences Influence Online Privacy Concerns and Online Information Disclosure.” Human Communication Research 47 (1): 49–74. https://doi.org/10.1093/hcr/hqaa012.
McElreath, Richard. 2016. Statistical Rethinking: A Bayesian Course with Examples in R and Stan. Chapman & Hall/CRC Texts in Statistical Science Series 122. Boca Raton: CRC Press/Taylor & Francis Group.
Metzger, Miriam J. 2004. “Privacy, Trust, and Disclosure: Exploring Barriers to Electronic Commerce.” Journal of Computer-Mediated Communication 9 (4). https://doi.org/10.1111/j.1083-6101.2004.tb00292.x.
———. 2006. “Effects of Site, Vendor, and Consumer Characteristics on Web Site Trust and Disclosure.” Communication Research 33 (3): 155–79. https://doi.org/10.1177/0093650206287076.
Muchnik, Lev, Sinan Aral, and Sean J. Taylor. 2013. “Social Influence Bias: A Randomized Experiment.” Science 341 (6146): 647–51. https://doi.org/10.1126/science.1240466.
New York Public Radio. 2018. “The Privacy Paradox.” https://project.wnyc.org/privacy-paradox/.
Omarzu, Julia. 2000. “A Disclosure Decision Model: Determining How and When Individuals Will Self-Disclose.” Personality and Social Psychology Review 4 (2): 174–85. https://doi.org/10.1207/S15327957PSPR0402_5.
Ozimek, Phillip, Julia Brailovskaia, and Hans-Werner Bierhoff. 2023. “Active and Passive Behavior in Social Media: Validating the Social Media Activity Questionnaire (SMAQ).” Telematics and Informatics Reports 10 (June): 100048. https://doi.org/10.1016/j.teler.2023.100048.
Park, Yong Jin. 2013. “Digital Literacy and Privacy Behavior Online.” Communication Research 40 (2): 215–36. https://doi.org/10.1177/0093650211418338.
Parry, Douglas A., Brittany I. Davidson, Craig J. R. Sewall, Jacob T. Fisher, Hannah Mieczkowski, and Daniel S. Quintana. 2021. “A Systematic Review and Meta-Analysis of Discrepancies Between Logged and Self-Reported Digital Media Use.” Nature Human Behaviour, May, 1–13. https://doi.org/10.1038/s41562-021-01117-5.
Petty, Richard, and John Cacioppo. 1986. Communication and Persuasion: Central and Peripheral Routes to Attitude Change. Springer Series in Social Psychology. New York: Springer-Verlag. https://doi.org/10.1007/978-1-4612-4964-1.
R Core Team. 2018. R: A Language and Environment for Statistical Computing. Vienna, Austria: R Foundation for Statistical Computing. https://www.R-project.org/.
Rohrer, Julia M. 2018. “Thinking Clearly about Correlations and Causation: Graphical Causal Models for Observational Data.” Advances in Methods and Practices in Psychological Science 24 (2): 251524591774562. https://doi.org/10.1177/2515245917745629.
Rosoff, Heather, Jinshu Cui, and Richard S. John. 2013. “Heuristics and Biases in Cyber Security Dilemmas.” Environment Systems and Decisions 33 (4): 517–29. https://doi.org/10.1007/s10669-013-9473-2.
Rosseel, Yves. 2012. lavaan: An R Package for Structural Equation Modeling.” Journal of Statistical Software 48 (2): 1–36. http://www.jstatsoft.org/v48/i02/.
Schneider, Frank M., Britta Zwillich, Melanie J. Bindl, Frederic R. Hopp, Sabine Reich, and Peter Vorderer. 2017. “Social Media Ostracism: The Effects of Being Excluded Online.” Computers in Human Behavior 73: 385–93.
Schreurs, Lara, Adrian Meier, and Laura Vandenbosch. 2022. “Exposure to the Positivity Bias and Adolescents’ Differential Longitudinal Links with Social Comparison, Inspiration and Envy Depending on Social Media Literacy.” Current Psychology, November. https://doi.org/10.1007/s12144-022-03893-3.
Simon, Herbert A. 1990. “Bounded Rationality.” In Utility and Probability, edited by John Eatwell, Murray Milgate, and Peter Newman, 15–18. London: Palgrave Macmillan UK. https://doi.org/10.1007/978-1-349-20568-4_5.
Skinner, B. F. 2014. Science and Human Behavior. Upper Saddle River, NJ: Pearson Education.
Söllner, Matthias, Axel Hoffmann, and Jan Marco Leimeister. 2016. “Why Different Trust Relationships Matter for Information Systems Users.” European Journal of Information Systems 25 (3): 274–87. https://doi.org/10.1057/ejis.2015.17.
Spottswood, Erin L., and Jeffrey T. Hancock. 2017. “Should I Share That?: Prompting Social Norms That Influence Privacy Behaviors on a Social Networking Site.” Journal of Computer-Mediated Communication 1 (2): 26. https://doi.org/10.1111/jcc4.12182.
Stroud, Natalie Jomini, Ashley Muddiman, and Joshua M. Scacco. 2017. “Like, Recommend, or Respect?: Altering Political Behavior in News Comment Sections.” New Media & Society 19 (11): 1727–43. https://doi.org/10.1177/1461444816642420.
Stsiampkouskaya, Kseniya, Adam Joinson, and Lukasz Piwek. 2023. “To Like or Not to Like? An Experimental Study on Relational Closeness, Social Grooming, Reciprocity, and Emotions in Social Media Liking.” Journal of Computer-Mediated Communication 28 (2): zmac036. https://doi.org/10.1093/jcmc/zmac036.
Stubenvoll, Marlis, Alice Binder, Selina Noetzel, Melanie Hirsch, and Jörg Matthes. 2022. “Living Is Easy with Eyes Closed: Avoidance of Targeted Political Advertising in Response to Privacy Concerns, Perceived Personalization, and Overload.” Communication Research, November, 00936502221130840. https://doi.org/10.1177/00936502221130840.
Sumner, Erin M., Luisa Ruge-Jones, and Davis Alcorn. 2017. “A Functional Approach to the Facebook Like Button: An Exploration of Meaning, Interpersonal Functionality, and Potential Alternative Response Buttons.” New Media & Society 20 (4): 1451–69. https://doi.org/10.1177/1461444817697917.
Sun, Yongqiang, Nan Wang, Xiao-Liang Shen, and Jacky Xi Zhang. 2015. “Location Information Disclosure in Location-Based Social Network Services: Privacy Calculus, Benefit Structure, and Gender Differences.” Computers in Human Behavior 52 (January): 278–92. https://doi.org/10.1016/j.chb.2015.06.006.
Taddicken, Monika, and Cornelia Jers. 2011. “The Uses of Privacy Online: Trading a Loss of Privacy for Social Web Gratifications?” In Privacy Online: Perspectives on Privacy and Self-Disclosure in the Social Web, 143–58. Berlin, Germany: Springer.
Tifferet, Sigal. 2019. “Gender Differences in Privacy Tendencies on Social Network Sites: A Meta-Analysis.” Computers in Human Behavior 93 (April): 1–12. https://doi.org/10.1016/j.chb.2018.11.046.
Trepte, Sabine, Michael Scharkow, and Tobias Dienlin. 2020. “The Privacy Calculus Contextualized: The Influence of Affordances.” Computers in Human Behavior 104 (March): 106115. https://doi.org/10.1016/j.chb.2019.08.022.
van Ooijen, Iris, Claire M. Segijn, and Suzanna J. Opree. 2022. “Privacy Cynicism and Its Role in Privacy Decision-Making.” Communication Research, February, 00936502211060984. https://doi.org/10.1177/00936502211060984.
Vanhove, Jan. 2021. “Collinearity Isn’t a Disease That Needs Curing.” Meta-Psychology 5 (April). https://doi.org/10.15626/MP.2021.2548.
Venkatesh, Morris, Davis, and Davis. 2003. “User Acceptance of Information Technology: Toward a Unified View.” MIS Quarterly 27 (3): 425. https://doi.org/10.2307/30036540.
Vitak, Jessica. 2012. “The Impact of Context Collapse and Privacy on Social Network Site Disclosures.” Journal of Broadcasting & Electronic Media 56 (4): 451–70. https://doi.org/10.1080/08838151.2012.732140.
Westin, A. F. 1967. Privacy and Freedom. New York, NY: Atheneum.
Whiting, Anita, and David Williams. 2013. “Why People Use Social Media: A Uses and Gratifications Approach.” Qualitative Market Research: An International Journal 16 (4): 362–69. https://doi.org/10.1108/QMR-06-2013-0041.
Wickham, Hadley. 2017. Tidyverse: Easily Install and Load the ’Tidyverse’. https://CRAN.R-project.org/package=tidyverse.