Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Privacy nudges for disclosure of personal information: A systematic literature review and meta-analysis

  • Athina Ioannou ,

    Roles Conceptualization, Formal analysis, Investigation, Methodology, Visualization, Writing – original draft, Writing – review & editing

    a.ioannou@surrey.ac.uk

    Affiliation Surrey Business School, University of Surrey, Guildford, United Kingdom

  • Iis Tussyadiah,

    Roles Conceptualization, Formal analysis, Funding acquisition, Investigation, Methodology, Project administration, Supervision, Validation, Writing – original draft, Writing – review & editing

    Affiliation School of Hospitality and Tourism Management, University of Surrey, Guildford, United Kingdom

  • Graham Miller,

    Roles Funding acquisition, Project administration, Supervision, Writing – review & editing

    Affiliation School of Hospitality and Tourism Management, University of Surrey, Guildford, United Kingdom

  • Shujun Li,

    Roles Formal analysis, Funding acquisition, Project administration, Supervision, Writing – review & editing

    Affiliation School of Computing, University of Kent, Canterbury, United Kingdom

  • Mario Weick

    Roles Conceptualization, Formal analysis, Funding acquisition, Methodology, Visualization, Writing – original draft, Writing – review & editing

    Affiliation Department of Psychology, Durham University, Durham, United Kingdom

Abstract

Objective

Digital nudging has been mooted as a tool to alter user privacy behavior. However, empirical studies on digital nudging have yielded divergent results: while some studies found nudging to be highly effective, other studies found no such effects. Furthermore, previous studies employed a wide range of digital nudges, making it difficult to discern the effectiveness of digital nudging. To address these issues, we performed a systematic review of empirical studies on digital nudging and information disclosure as a specific privacy behavior.

Method

The search was conducted in five digital libraries and databases: Scopus, Google Scholar, ACM Digital Library, Web of Science, and Science Direct for peer-reviewed papers published in English after 2006, examining the effects of various nudging strategies on disclosure of personal information online.

Results

The review unveiled 78 papers that employed four categories of nudge interventions: presentation, information, defaults, and incentives, either individually or in combination. A meta-analysis on a subset of papers with available data (n = 54) revealed a significant small-to-medium sized effect of the nudge interventions on disclosure (Hedges’ g = 0.32). There was significant variation in the effectiveness of nudging (I2 = 89%), which was partially accounted for by interventions to increase disclosure being more effective than interventions to reduce disclosure. No evidence was found for differences in the effectiveness of nudging with presentation, information, defaults, and incentives interventions.

Conclusion

Identifying ways to nudge users into making more informed and desirable privacy decisions is of significant practical and policy value. There is a growing interest in digital privacy nudges for disclosure of personal information, with most empirical papers focusing on nudging with presentation. Further research is needed to elucidate the relative effectiveness of different intervention strategies and how nudges can confound one another.

1. Introduction

Human decision making is inherently complex. Immersed in digital environments and while performing online activities, individuals are faced daily with numerous privacy and security decisions: configuring visibility in social networking sites, allowing access to sensitive data in mobile apps, clicking or ignoring links embedded in emails, and so forth. Various factors such as heuristics, cognitive and behavioral biases, and incomplete and asymmetric information can affect privacy decisions and behaviors, often leading to deficient and regrettable choices, such as oversharing, that are not aligned with users’ own intentions [1]. Habitual responses and low awareness of ransomware and scams, for example, can result in users disclosing sensitive personal information to unauthorized sources, such as employees giving untrusted third-party sources access to confidential corporate data [2]. These privacy failures impact not only the individuals and organizations directly involved in the collection and use of personal information, but also the economy and society at large due to business disruption, loss of productivity, and other financial and social ramifications. Therefore, finding ways to encourage users to make more informed privacy decisions is of significant practical and policy importance.

Studies have suggested the use of paternalistic interventions or nudges to guide and assist people into changing their security and privacy behaviors in digital environments [1]. A nudge is “any aspect of the choice architecture that alters people’s behavior in a predictable way without forbidding any options or significantly changing their economic incentives” [3](p.6). In other words, nudging can influence decision making allowing people to take any course of action, “without limiting individual choice, but in fact preserving freedom of choice” [1](p.44:11). Existing research on nudging has been conducted mostly in offline environments [4]. However, there is a growing body of literature on the use of digital nudging in the context of individuals’ privacy-related behavior and decisions. Digital nudging refers to the use of interface elements aiming to guide users to make more informed judgements and decisions in digital choice environments [4, 5].

Empirical studies on the effectiveness of digital nudging on user behaviors are not fully conclusive. Whilst some studies have shown that digital nudging can change user behaviors [6] some others have found no such effect [7]. This may not be a complete surprise as there are vast discrepancies in the type of nudges that have been studied. In addition, outcome behaviors differ widely. For example, Halevi, Kuppusamy, Caiazzo, and Memon [8] sought to nudge users to increase information disclosure, while Bravo-Lillo et al. [9] sought to nudge users to reduce the likelihood of installing malicious software.

As this brief discussion highlights, there is an urgent need for a systematic review and meta-analysis of the literature to establish the effectiveness of nudging interventions in changing privacy-related behaviors and decisions (cf. [10]). In addition, the varied nature of nudging interventions calls for a framework to classify nudging strategies.Acquisti et al. [1] provided such a conceptual framework, with a focus on privacy nudging, which is particularly pertinent for the present investigation. They argued for a conceptual distinction between digital nudges relying on information, defaults, incentives, timing, reversibility or presentation techniques, either individually or in combination. These dimensions were suggested to address various pitfalls when users make privacy decisions, including issues associated with cognitive biases and information asymmetry. In the present study, we adopt Acquisti et al.’s [1] framework to establish the effects of digital nudges on privacy decisions in digital environments. We focus on disclosure of personal information as a specific type of privacy decisions. Information disclosure (also known as data disclosure or online disclosure) refers to the act of making personal information accessible to other interested parties. We define personal information as any information that can identify an individual directly or indirectly, such as biographical information, telephone number and email address, workplace data and education information, location data, physical, physiological, genetic, mental, economic, cultural or social data of a person [11]. Through our focus on personal information disclosure, we extend a recent meta-analysis on the effectiveness of nudging in online and offline environments [12]. The present work casts a much wider net than Hummel and Maedche’s [12], which only examined studies citing [3], thus providing a more comprehensive assessment of the literature.

2. Method

This systematic literature review was conducted following the guidelines proposed by PRISMA (2015) as well as by Kitchenham and Charters [13] who have adapted the PRISMA guidelines in the Software Engineering context. The most critical element of this study, the review protocol, was developed first. The protocol was published in a networking portal for scientists, ResearchGate (https://www.researchgate.net/). The study followed the steps outlined in the review protocol, which are discussed in the following subsections.

2.1. Search strategy

The research question addressed in this systematic review is: “What are the effects of various intervention (nudging) strategies on disclosure of personal information online?” The search was conducted in five digital libraries and databases: Scopus, Google Scholar, ACM Digital Library, Web of Science, and Science Direct. All searches were based on title, abstract, and keywords, and took place in September 2020. The search strategy and the combination of keywords used in this study is presented in Table 1. Please note that prod refers to experimental design. Prod is used as a synonym to nudge.

The search strategy depicted in Table 1 was used as a template for the searches in all five databases. As each one of the digital libraries has a different search engine, a preliminary search was conducted on each digital library adapting the search terms to the requirements of the relevant search engine. More details on the exact search terms used in each database can be found in S1 Appendix. Since in the last decade the privacy of individuals has been affected by new technological solutions such as smartphones and Internet of things, papers published before 2006 are excluded because the meaning of privacy might have changed in the years following 2006 [14]. All details of papers that were considered potentially relevant were managed using the reference management tool Mendeley (version 1.19).

2.2. Study selection

Two co-authors (reviewers, hereafter) conducted searches in all five digital libraries. Each reviewer performed an initial screening of their search results based on titles and abstracts. The results of the initial screening were compared between reviewers to check for consistency and discrepancies in the search process.

2.2.1. Inclusion criteria.

Studies evaluating the effects of various intervention (nudging) strategies on the disclosure of personal information in the context of online privacy were of interest in the present review. Thus, the following inclusion criteria were applied:

  • Empirical studies are eligible for inclusion. Empirical studies reporting experimental manipulations and quasi-experimental variations are considered eligible, as well as studies conducted in the laboratory, field, and online.
  • Studies that include technology usage, such as internet, social media, e-commerce websites, mobile phones, and other digital platforms, are considered eligible. There are no age restrictions for participants. Where information is available on individuals’ health status, studies of interest include healthy participants.
  • Studies using one or more of the following intervention strategies (nudges) as mentioned in the work of Acquisti et al. [1] are eligible: information (feedback and education), presentation (framing, ordering, salience and structure), default, timing, reversibility and incentive (priming).
  • Studies that depict the intervention strategies as independent variable(s) and intention/willingness to disclose or share personal information as well as actual disclosure or sharing behavior as dependent variable(s) are of interest. Studies that include antecedents of dependent and independent variable(s) as well as mediators and/or moderators in this relationship are also eligible.

2.2.2. Exclusion criteria.

Papers that meet the following criteria were excluded from the review:

  • Papers published in languages other than English.
  • Theoretical or conceptual publications related to intervention strategies and individual privacy.
  • Papers reporting studies conducted in clinical settings or with clinical samples (e.g., visually impaired individuals) or special populations (e.g., abuse survivors).
  • Papers reporting studies investigating the effect of nudging interventions on outcomes other than information disclosure, such as password creation, selection of secure Wi-Fi, intention towards policy compliance, intention to install software applications, and similar others.

In the event the primary reviewer was unsure regarding the application of inclusion and exclusion criteria in a particular study, the opinion of the second primary reviewer was sought in order to reach a decision. Failure to meet any one of the above eligibility criteria resulted in exclusion from the review and any apparent discrepancies during the selection process was resolved through discussion or, in case no agreement was reached, through consultation of an independent reviewer. The number of excluded papers, including reasons of exclusion for those excluded following review of the full text, was recorded at each stage.

2.3. Study quality assessment

There have been several quality checklists published in extant academic literature, although most of them addressed medical studies. Aiming to rigorously assess the methodological quality of the studies included in this systematic review, we followed the guidelines of Kitchenham and Charters [13] as well as the systematic review of Zhou et al., [15] on the existing quality assessment tools that are being used in systematic reviews in the area of Software Engineering. Both studies recommended a set of questions, derived from the most commonly used checklists and guidelines regarding the design, conduct, analysis, and conclusions of each study included in this systematic review. According to the literature, researchers should review the available list of questions in the context of their own study and select the most appropriate evaluative questions for their study [13]. For this reason, questions from different checklists that have been used in previous research were reviewed and integrated for the purposes of this systematic review. The criteria to evaluate each study were based on the evaluative questions presented in Table 2.

The scoring procedure was Y = 1 and N = 0. Studies could reach a minimum of 0 and maximum 10 points. We established a cut-off of 6 as it represents 60% of all questions in the quality assessment [16]. Papers receiving a score exceeding 6 (>6) were decided to be retained in this systematic review (see S1 Appendix).

2.4. Data extraction

The following data were extracted from each paper: full reference (including name(s) of author(s), year of publication, and publication venue), description of the intervention strategy, dependent variable(s) (e.g., intention to disclose information or actual disclosure behavior), antecedent, moderator, and mediator variables(s), main findings, type of study (e.g., experiment), type of statistical analysis (e.g., regression), type of outcome variable (e.g., dichotomous, continuous), randomness of allocation, and existence of control group. Specifically, for studies that were considered potential candidates for inclusion in a meta-analysis, we recorded effect sizes or any other data that would allow for the calculation of an effect size.

Each study included in the systematic review was reviewed by two co-authors. One acted as the main data extractor, while the second acted as the data checker. The first reviewer was responsible for extracting the data from all studies that were selected for inclusion, while the second reviewer was responsible for checking if the data extracted by the former were correct. Any apparent disagreements during the data extraction process were resolved through discussion and, in case that no agreement could be reached, the consultation of an independent reviewer was sought.

2.5. Statistical analysis

We performed a meta-analysis on papers included in the systematic review, which reported appropriate effect sizes or data that enabled us to calculate effect size statistics. As a result, a further evaluation of the selected studies was performed in order to select studies that fulfil the requirements necessary for a meta-analysis. The selection of papers that would be included in the meta-analysis was based on whether the papers satisfied the following criterion [17]: Papers were considered eligible for meta-analysis only if they reported results including an effect size (Cohen’s d, Odds Ratio, Pearson’s r, Chi squared, eta squared) or provided enough information for the computation of one (means and standard deviations, the percentage of participants in treatment and control groups, F values, regression coefficients). When the reported results did not enable the computation of an effect size, the paper was excluded from meta-analysis [18].

3. Results and discussion

3.1. Search results

Integrating the results derived from all five digital libraries, 2,046 papers were identified before duplicate papers were removed. More specifically, each database yielded the following results: 343 papers from Scopus, 159 papers from Science Direct, 314 papers from ACM Digital Library, 769 papers from Web of Science, and 459 papers from Google Scholar. Moreover, additional papers were identified by scanning reference lists and citations of prominent articles in the investigated subject, a search approach also known as snowballing. It has been suggested that researchers can achieve the best possible coverage of existing literature by using snowballing as a complementary approach to database search [19]. The snowballing process resulted in 106 papers. We screened the title, abstract, and keywords of 2,046 + 106 = 2,152 papers identified through database searches and snowballing. We excluded papers that were deemed irrelevant or duplicates. As a result, a total of 254 papers passed the initial screening and were subject to a more detailed assessment. The full list of papers (n = 254) was then evaluated against a specific set of inclusion and exclusion criteria. In case of similar studies using the same dataset, we retained the one providing more detailed information. After evaluating the inclusion and exclusion criteria, 150 papers were excluded, with 104 papers remaining for full text assessment. The full text screening resulted in 39 papers identified for exclusion, mostly due to the fact that the outcome variable in the study was not related to information disclosure or the employed intervention cannot be classified as a nudge. Other reasons for exclusion include papers conducting surveys instead of experiments, papers investigating information disclosure using factors other than nudging/interventions as well as papers that have published the same results in different outputs. As a result, a total of 78 papers were included in the present systematic review (see Fig 1).

thumbnail
Fig 1. PRISMA flow diagram of information through different stages of review.

https://doi.org/10.1371/journal.pone.0256822.g001

3.2. Study characteristics

Overall, the final selection of papers included in the qualitative review suggests that during the last decade there has been a surge of interest in the effects of nudges on privacy decisions (see Fig 2). More specifically, during the years 2013–2020, there has been a dramatic increase in scientific publications (n = 64), while during the years 2006–2012 the published papers in the area were quite scarce (n = 14), and in 2008 there were no publications at all. As a result, we can infer that the investigation of privacy nudges has grown significantly during the last decade, most likely since 2011–2012, taking into consideration the time delay between investigations and academic publications.

Regarding the type of publications, most papers were published in academic journals (n = 43), while almost half of the included papers were published in conference proceedings (n = 35), thus signifying a balance in the publication outputs of the included papers. Furthermore, regarding the type of participants, most papers (n = 35) used commercial platforms to recruit participants (e.g., Amazon Mechanical Turk), while several papers drew on student samples (n = 15) in their experiments. Regarding the origin of participants, most papers were conducted in the USA (n = 24) and in Europe (n = 12), although a majority of papers did not reveal the location of the experiment or the origin of the recruited participants (n = 30).

Table 3 presents the studies included in the qualitative part of this systematic review (n = 78) along with author(s) and publication year, intervention or nudging strategy, and main findings of the study. More detailed information regarding the papers can be found in S1 Appendix. Overall, it can be inferred that the majority of papers employed factorial between-subjects experimental designs as well as surveys that include scenarios in order to test the effects of nudging on intended or actual information disclosure of individuals.

thumbnail
Table 3. Studies categorized by intervention (nudging) strategy.

https://doi.org/10.1371/journal.pone.0256822.t003

The present systematic review focuses on the effects of intervention strategies on information disclosure of individuals in online contexts. Thus, in order to explicate the effects of different nudging strategies, we grouped studies into four nudging categories according to the dimensions put forward by Acquisti et al. (2017) [1]: information, presentation, defaults, and incentives. Reversibility and timing, which are part of the nudging dimensions suggested by Acquisti et al. (2017) [1], were initially taken into consideration. However, during the search process we identified no papers under the timing dimension, and only one paper that could be classified into reversibility as well as presentation nudging. According to Acquisti et al. (2017), the dimensions of the framework are not mutually exclusive. For purposes of comparability, we classified the paper under presentation nudging, thus covering the four dimensions in the framework. As some papers have used a combination of the aforementioned four categories of nudges, a fifth category (i.e., combination) was added. As seen in Table 3, it is apparent that most papers (47 out of 78 papers) used presentation interventions to influence privacy decisions and disclosure of personal information, followed by combinations of the four interventions (fourteen papers), information (nine papers), incentives (four papers) and defaults (four papers).

3.3. Qualitative synthesis

3.3.1. Nudging with presentation.

Nudging with presentation refers to how information and choices are presented to users using design and delivery concepts such as framing, ordering of alternatives, saliency, and structure of presented information. Examples include reminders, privacy notices and warnings, design features, and visual cues. Forty-one papers were identified employing presentation as a nudging intervention aiming to explore the influence on information disclosure, however, the majority of extant literature has offered mixed and contradictory findings.

Several papers found that the use of warnings can reduce actual information disclosure of users [6, 2025]. Similarly, several studies found that priming, either in the form of privacy-related articles, stories, or videos, was effective at encouraging individuals to limit the sharing of personal information [2628]. Moreover, a number of papers have investigated the effects of visual cues for privacy decision making; results vary depending on the cue being used by each paper, such as anthropomorphic [29] or monetary images [30]. The majority of papers agree that visual cues can increase the propensity of an individual to disclose personal information [2932]. However, one research paper argues that the use of visual cues in the form of reminders about the audience of posts can reduce unintended user disclosures [33].

There is a considerable interest in the effects of different design features on individuals’ willingness to disclose personal information. A number of papers have empirically demonstrated that certain design features, such as variety of customization options [34], high immediacy levels [35], perception of information control [34, 36], exposure to calls of action [37], videos including female narrators combined with vibrant colors and light musical tones [38], a dynamic privacy score [39], norm-shaping design patterns [40], as well as contextual cues amplifying or downplaying privacy concerns [41], can increase the disclosure of personal information. Also, studies have shown that persuasive messages that are more positively framed or include higher argument strength [42] as well as motivational messages can increase the disclosure of personal sensitive information [43]; while initially participants reacted negatively to the use of hyper personalization in advertisements, this did not discourage them from sharing sensitive personal information [44].

Meanwhile, other research has failed to show that presentation nudges, including changes in the design of search engines [45], hard to read fonts [46] and increased transparency in the provision of information regarding data processing [47], have an effect on the amount of information being shared during privacy decision making. Contradictory findings are offered by other studies investigating the effects of design features, such as the use of an anthropomorphic character and the length of security messages embedded in notification messages, showing that they can effectively reduce the amount of personal sensitive information that an online consumer is sharing [48].

Moreover, there has been a growing interest in investigating the effect of social nudges, which refer to social framing and peers’ behavior pressure aimed to influence users’ privacy decision making. Research findings agree that social nudges, revealing that the majority of users’ peers have engaged in a similar behavior such as divulgence of personal information [49, 50], usage of applications [51], or adjustment of privacy settings [52], can increase information disclosure. Furthermore, research has shown that adding an obfuscation option, allowing users to share sensitive data such as physical activity and GPS location in anonymized form, can increase the sharing of personal information [53, 54], while a reversibility option, being able to go back and delete previously entered information before submitting, can reduce individual information disclosure [55]. The use of other social nudges, such as presenting profile images of the audience of a social network or numerical information about the audience of a social network, was shown to positively affect people’s posting behavior resulting in users making ‘better’ choices by posting more messages on a private network rather than in a public one [56]. While information about other users (social norms) displayed by a privacy tool did not enhance the privacy protection behavior of Facebook users [57].

In addition, there has been a surge of interest in investigating the role of privacy policies and notices as effective aids on privacy and disclosure decisions; studies have demonstrated that increases in the permissiveness [58] as well as vagueness [59] of a privacy policy can reduce the willingness of an individual to share his/her personal information, while simplification techniques failed to impact disclosure [7]. Moreover, the impact of different degrees of data protection on intention and actual information disclosure has been examined, with results revealing that both normative and objective factors (objective and relative changes in privacy protection) in privacy notices can influence information disclosure [60]. However, in a series of experiments, Adjerid et al. [61] demonstrated that the impact of privacy notices on information disclosure is variable; in their first study results showed that privacy notices implying lower protection actually decreased disclosure, thus achieving the desired support and protection of users, while in the second study findings revealed that the effect of a riskier privacy notice can be significantly reduced by minimal distractions such as a 15 second delay between the notice and the disclosure decision.

Research has also investigated how fine grained sharing options can affect users’ disclosure decisions, revealing that when finer options are removed users tend to choose the closest option thus increasing their disclosure rather than erring on the safe side and choosing not to share at all [62]. Meanwhile, the display of finer grained information embedded in data requests lowered users’ likelihood to share personal information [63]. Aiming to compare the effects of different interfaces, Vitale et al. [64] found no significant differences in the amount of personal information disclosed when a kiosk or a robot were employed, while Zhu, Carpenter and Kulkarni [65] observed that an interface providing privacy suggestions to users helped them in limiting unnecessary disclosures.

3.3.2. Nudging with information.

Nudging with information aims at educating and creating awareness in users about the risks and benefits associated with privacy and security decisions using effective communication messages. Provision of information includes two approaches: education and feedback. Education includes the provision of information to users before their engagement with the system thus referring to future decisions, while feedback refers to the information provided alongside the use of the system. Nine papers were identified and included in the present systematic review using the provision of information aiming at assisting and supporting privacy and security decision making. Overall, eight papers have used nudging with information methods and only one study has used feedback. Examples of education nudges include privacy policy documents, notifications, and privacy notices. The papers included have emerged during the years 2006–2020, with most of them being published after 2014, thus demonstrating that only recently has there been an increased interest in information interventions to influence disclosure.

Results of the included papers seem to be inconsistent regarding information priming towards changing privacy and security decisions. A number of papers found no evidence that priming nudges such as warnings and data notifications are effective at reducing information disclosure of personal information [6668]. However, recent research found that privacy messages prompting users to think about privacy and security issues induced them to share less personal information [69, 70] while the provision of terms and conditions inhibited information disclosure [71]. In the same vein, research findings indicate that users participating in Facebook training shared less personal information [72], while trustmarks and third party certification are the most effective tools towards gaining users’ trust and consequently affecting their sharing behavior [73]. Regarding the provision of feedback as a nudging intervention, results indicated that users who received feedback were more comfortable towards sharing their location information [74].

3.3.3. Nudging with defaults.

Nudging with defaults is defined as intervention aiming to influence users’ privacy decisions by setting default options that best serve and align with users’ privacy needs and expectations [1]. Four papers were identified in the present systematic review employing the use of defaults as nudging strategy towards influencing privacy decisions. The papers were published between 2006–2015, demonstrating that during the last decade the use of defaults has received attention and becoming an area of interest in the privacy decision making field. Results of the papers show that the use of traditional auto completion tools, usually found in web forms, could cause significantly more information being disclosed [75], while restrictive default privacy settings helped users to share less information on online social networks [76]. Also, the use of an opt-in frame, comparing to an opt-out frame, for the provision of consent proved to be a more effective strategy towards protecting users’ privacy [77] while pre-selected options within a choice frame may increase user participation in online activities such as sign up [78].

3.3.4. Nudging with incentives.

Nudging with incentives aims at motivating users to behave according to their stated preferences. This nudging intervention can take the form of either rewards or punishments. Four papers using incentives as the main nudging strategy were included in the final selection of the present systematic review. The publication years span from 2010 until 2018; all four papers focus on investigating the effects of monetary incentives on privacy decisions in general information disclosure contexts [79, 80] as well as in more specific contexts such as sharing biometric data [8] and granting Facebook access [81]. Research has shown that the higher the amount of the monetary incentive to encourage disclosure, the higher the percentage of people who shared their information [81]. However, other authors argue that, in contrast with the conventional view, most people are not open to sharing their personal data for any feasible reward [8]. Consistent with this view, some studies found thatfinancial rewards were not effective at motivating users’ disclosure [80] or that monetary rewards can actually undermine information dislcosure intentions [79].

It is noteworthy that there is a widespread confusion and considerable debate regarding the relation between nudges and incentives and how nudges can be properly distinguished from other interventions [82]. Thaler and Sunstein (2008) defined a nudge as “any aspect of the choice architecture that alters people’s behavior in a predictable way without forbidding any options or significantly changing their economic incentives” [3](p.6). Thus, following this definition, incentives may not be classified as nudges. However, Thaler and Sunstein also introduced the acronym NUDGES as an easy reminder of the six principles for good choice architecture, with the first principle referring to incentives. According to Thaler & Sunstein (2008), the six principles for good choice architecture can be viewed as the acronym NUDGES: iNcentives, Understand Mappings, Defaults, Give Feedback, Expect error, and Structure complex choices. Saghai (2013) amended the definition of nudges by clarifying what it meant by preservation of freedom of choice and by elaborating the importance of substantial noncontrol in nudges: “A nudges B when A makes it more likely that B will φ, primarily by triggering B’s shallow cognitive process, while A’s influence preserves B’s choice-set and is substantially noncontrolling (i.e., preserves B’s freedom of choice)” [83](p. 491). Saghai (2013) further suggested that while nudges are substantially noncontrolling, incentives can be substantially controlling or substantially noncontrolling, depending on the valuation (magnitude) of its benefits.

As previously mentioned, the present work follows the classification of nudging for privacy behavior grounded on the seminal work of Acquisti et al. (2017) [1]. In their work, Acquisti et al. (2017) [1] argued that rewards (and punishments) can be used as tools to overcome cognitive and behavioral biases, such as hyperbolic discounting, that affect privacy decision making. All four studies included in the present review offered monetary incentives to participants, either as discount coupon or compensation, ranging from $0.20 to $50, in exchange for users’ personal information. Their results indicate that incentives can both increase and decrease disclosure (or have no effect). Following Saghai’s (2013) [83], the incentives used in these studies may be considered nudges as their effect on disclosure behavior is not explained by the magnitude of the incentives. Overall, the present work does not seek resolve theoretical debates centered on how to distinguish different (nudging or non-nudging) interventions. Rather, we take a pragmatic approach drawing on an existing classification framework [1] to provide a comprehensive overview of the literature, leaving it for future research to revisit and potentially refine the classification of nudges for privacy decision making.

3.3.5. Combination.

A number of papers have used nudges belonging to more than one of the previously mentioned categories (i.e., presentation, information, incentives, and defaults). These are presented below.

There is a small but growing body of academic literature focusing on comparing the effects of rewards, mainly monetary incentives, with presentation nudges, mostly including design features. Most research agrees that the combination of design features with monetary incentives constitute an effective strategy towards enhancing information disclosure. More specifically, papers have demonstrated that embedding social design features within enterprise social systems combined with the provision of monetary incentives positively influenced users’ attitudes to share personal information [84], while the use of nudging messages combined with monetary incentives resulted in increased social sharing behavior [85]. Aiming to evaluate the role of privacy assurances as well as the risk-benefit trade off in the context of consumer information disclosure, findings have shown that privacy statements/notices along with monetary rewards can motivate people to share more personal information [86] as well as provide accurate identifiable information [87]. More recently, emerging literature demonstrates that the use of rewards and safety cues has varying effects on privacy decision making as privacy assurances, presented with safety cues, are more effective over monetary rewards in encouraging information disclosure [88]. Moreover, it has been empirically demonstrated that mandatory fields within a web form could decrease voluntary sharing of personal information for the remaining optional fields, while the provision of monetary incentives for the same fields could increase the rates of disclosure [89].

Research has investigated the effects of information and incentive nudges aiming to understand their impact on disclosure outcomes. Results have shown that among different groups of nudges, including standard privacy policies, customer empowerment, blockchain supported privacy policy and monetization, participants shared similar amounts of personal information [90]. Further it has been demonstrated that a mix of social and financial incentives, such as feedback and cash incentives, were effective strategies in motivating users towards information disclosure [91]. Moreover, investigating the combined effects of trust and compensation in the form of an excerpt describing the e-commerce company and several types of benefits, respectively, Premazzi et al. [92] revealed that in low privacy conditions incentives improve information disclosure, while in high privacy conditions compensation decreases actual sharing of information. Control over one’s data make people more comfortable with data sharing, while monetary compensation actually decreased people’s willingness to share their data with an unknown third party data broker [93].

There has been increased interest in the experimental investigation of the role of default and presentation nudges aiming to influence privacy decision making and sharing of personal information. It has been empirically demonstrated that one’s profile information in an social network that is set up under a shared-by-default setting, comparing to a private-by-default setting, could increase sharing of personal information while the granularity of categories had no effect on users’ sharing tendency [94]. Moreover, evidence has been observed that the opt-in default settings as well as social consensus constitute powerful mechanisms in increasing information sharing [95]. While in their study examining the potential impact of privacy nudges being tailored to users’ personality traits, Warberg, Acquisti, and Sicker [96] used a variety of nudges such as opt-in/opt-out, framing and social norms, demonstrating that it is very difficult to tailor nudges to users’ characteristics.

At last, the present review identified only one study focusing on investigating the impact of information and presentation nudges together on privacy decision making. It has been shown that justification messages could lower users’ disclosure rates, while the order of data requests could increase the disclosure of data that is requested first and decrease the disclosure of data that is requested later [97].

3.4. Meta-analysis

A meta-analysis was conducted using RevMan 5.3 [98] to quantify the effects of nudges on information disclosure decisions. The meta-analysis aims to combine all effect sizes derived from individual studies, resulting in an estimate of the overall effect size regarding the outcome in question (i.e., information disclosure). All 78 papers were initially considered for a meta-analysis. However, 24 papers failed to report sufficient statistical information required for the computation of effect sizes and thus were excluded [99]. As a result, 54 papers were included in the meta-analysis. Nine papers reported more than one experimental study, and 23 studies employed more than one intervention, resulting in a total of 68 studies with 118 effects. Table 4 presents the specific nudges used in the included studies (total number of nudges = 84). All classifications and coding of effect sizes were performed by one co-author and two independent research assistants. Disagreements were resolved either through discussion or by consulting a third coder. Please note that we denote with n the number of papers, i the number of studies, and k the number of effect sizes.

thumbnail
Table 4. Nudges in 54 papers included in the meta-analysis.

https://doi.org/10.1371/journal.pone.0256822.t004

Effect sizes were expressed as standardized mean differences (Hedges’ adjusted g), which can be interpreted similar to Cohen’s d, but include an adjustment for small sample bias. Since the true effect size in the population stands to differ between studies employing different interventions, we opted for a random effects model using an inverse-variance method to combine results from different studies. We classified interventions as those seeking to decrease disclosure (57.3%) and those seeking to increase disclosure (40.8%), based on predictions put forward by the original author(s) (one study encompassing two interventions did not report any predictions). To enable a synthesis of all interventions, positive effect sizes denote predicted differences between a treatment and control group (i.e., an increase in disclosure when an increase was predicted, or a reduction in disclosure when a reduction was predicted), and negative effect sizes denote unpredicted differences (i.e., an increase in disclosure when a reduction was predicted, or a reduction in disclosure when an increase was predicted). Since the original author(s)’s predictions may not reflect true differences in the population, we supplemented our coding with the coining method described by Fanelli, Costas, and Ioannidis [100]. In particular, we re-coded unpredicted effects into predicted effects (i.e., we multiplied effects by -1) when an effect size exceeded a conservative 1.65*SE threshold and thus had a likelihood of occurrence of p = .05 (one-sided). This adjustment affected four out of 118 effect sizes (3.4%). In further sensitivity analyses, we repeated all primary analyses using absolute effect sizes (a more liberal approach to coining, see ([100]).

As shown in the forest plot depicted in Fig 3, nudging interventions had a small-to-medium sized overall effect on disclosure, Hedges’ g = 0.32 [0.25, 0.38]. Effect sizes were heterogeneous, I2 = 89%, as anticipated, and re-affirming our decision to employ a random effects model. When looking at different types of nudging interventions, the strongest effect was observed for incentive and default interventions, Hedges’ g = 0.42 [0.14, 0.70] and 0.41 [0.24, 0.59], respectively. Meanwhile, the weakest effect was observed for information interventions, Hedges’ g = 0.18 [0.06, 0.31], followed by presentation interventions, Hedges’ g = 0.33 [0.24, 0.42]. Effect sizes do not differ significantly between subgroups, χ2(3) = 6.19, p = .10. Furthermore, pairwise comparisons (k = 6) keeping the global error rate at 5% also revealed no significant differences between subgroups. Thus, we refrain from drawing any conclusions regarding the relative strengths of the different interventions (presentation vs. incentive vs. information vs. default).

thumbnail
Fig 3. Forest plot.

Note: Where applicable, the number of participants was adjusted by dividing the total number of participants by the number of measures and/or interventions administered to the same group of participants (denoted (i) to (vi); see ([10]). This adjustment does not affect the total number of participants. Note that, while study-level confidence intervals are wider when adjustments are made, estimates of central tendency (i.e., standardised mean difference) are unaffected.

https://doi.org/10.1371/journal.pone.0256822.g003

Secondary analysis. Overall, interventions to increase disclosure had a stronger impact on disclosure when compared to interventions to decrease disclosure, Hedges’ g = 0.33 [0.25, 0.40] and 0.20 [0.13, 0.27], respectively; test for subgroup differences: χ2(1) = 6.04, p = .01.

Sensitivity analysis. As outlined above, we repeated all analyses, this time employing a more liberal approach to coining using absolute effect sizes [100]. This sensitivity analysis yielded the same conclusions, with an overall effect size of Hedges’ g = 0.35 [0.29, 0.41] across all interventions.

Finally, we also examined funnel plots to gauge the presence of potential reporting bias. As shown in Fig 4 and Fig 5A–5D, there was some evidence for reporting bias for all nudging strategies. However, the number of studies using incentive, information, and default nudges was relatively low, which limits the conclusions that can be drawn from the funnel plots. On the other hand, the number of studies reporting presentation nudging interventions was noticeably larger. Here, the funnel plot points to the presence of several studies reporting unexpectedly large effects. Two effects in particular appear to be outliers, with standardized mean differences that exceed the standard error more than fifteen-fold. Effect sizes remain significantly heterogeneous when the two outliers are removed, I2 = 81%. This is consistent with the observation that the nature of the presentation interventions differs widely between studies (see Table 3). Outliers aside, differing interventions may explain the distribution of effect sizes observed for presentation interventions [101]. Crucially, removing large effects relative to the standard error from the meta-analysis (including the two outliers) had a limited impact on the overall effectiveness of presentation interventions, which remained significant. Similarly, removing outliers did not impact the conclusions derived from the primary and secondary analyses.

thumbnail
Fig 5.

a. Funnel plot for presentation nudging. b. Funnel plot for incentive nudging. c. Funnel plot for information nudging. d. Funnel plot for information nudging.

https://doi.org/10.1371/journal.pone.0256822.g005

Discussion. There is a growing body of research investigating the impact of presentation nudges on users’ information disclosure. In our meta-analytic review, the majority of effects (84 out of 118) represent presentation nudging, while the rest of the three categories (iincentives = 5; iinformation = 20; idefaults = 9) account for less than 30% of the total effect sizes. Presentation nudges ranged from warnings and design features, to social nudges and visual cues. Even within those sub-types, presentation nudges vary. For example, studies examining design features have experimented with information layout, different fonts, anthropomorphic characters, and so forth. It might be that different sub-categories of presentation nudges show differential effects on information disclosure. Future research studies should aim at exploring in more depth the impact of different presentation nudges. For example, in the case of visual anthropomorphic cues, studies may consider race and gender as well as different contexts such as a doctor in healthcare, a salesperson in an e-commerce website or a teacher in an online educational platform in order to explain potential differences in findings in existing studies.

Results of the meta-analysis show that in absolute terms incentives had the strongest effect on users’ information disclosure.This is consistent with the wider literature on behavioral change. For example, previous meta analytic research indicates that financial incentives are effective in promoting changes in behavior such as encourage pro-environmental behavior (e.g., recycling) [102] and healthy behavior change (e.g., smoking) [103]. However, it should be noted that the number of studies in this category of interventions was relatively small (i = 5), and as such conclusions should be drawn with caution. A separate meta-analysis excluding papers with incentives has been conducted, which shows no significant changes in the results and hence no changes to the conclusions of this study (see S1 Appendix).

Our findings reveal that although the effects of information nudges were significant, they were (in absolute terms) the smallest out of all nudging strategies. This is consistent with the studies on health behaviors, which found that information campaigns were often either ineffective at changing behaviors [104], or the campaigns backfired promoting undesirable outcomes [105]. According to Thaler (2018) [106], a behavioural intervention that does not encourage a behavior in the individual’s best interest is called a ‘sludge’. In this review, all selected papers aimed at examining the effect of nudges on disclosure. However, their results sometimes were unexpected.

Furthermore, regarding the effect of default nudges, the results of the meta-analysis revealed a medium-sized effect. These results dovetail recent meta-analytic evidence attesting to the effectiveness of default nudges across a range of outcomes [12]. However, since the number of studies examining information disclosure more specifically remains relatively small (i = 9), some caution is warranted, and further studies are needed to ascertain the effectiveness of default nudges on information disclosure.

Finally, our secondary analysis revealed that interventions aimed at increasing disclosure had a stronger impact on disclosure when compared to interventions aimed at decreasing disclosure. In other words, nudges may be more effective in motivating people to share their personal information rather than discouraging people to withhold such information. This asymmetry may be explained by the presence of a floor effect: if participants are somewhat reluctant at baseline to disclose personal information, then there is more ‘headroom’ for subsequent interventions to increase (vs. reduce) disclosure. Future studies are needed to further explore the potential differential effects of nudging aimed at increasing vs. decreasing information disclosure.

4. Discussion

4.1. Implications

Identifying ways to nudge users into making more informed and desirable privacy decisions is of significant practical and policy value [1]. The main contribution of this systematic meta-analytic review is the comprehensive evidence gathering of the applications and effectiveness of a variety of privacy nudging strategies to influence disclosure of personal information in digital environments. The systematic review of empirical research publications from 2006, which included 69 papers, revealed a growing interest in digital privacy nudges for disclosure of personal information while also unveiling inconsistencies in research findings. In general, three broad conclusions could be derived from the results of the qualitative synthesis and meta-analysis: (1) most empirical papers have focused on nudging with presentation to influence users’ privacy decisions, (2) a meta-analysis of 54 papers revealed that nudging strategies can effectively influence information disclosure, (3) further research is needed to elucidate the relative effectiveness of different intervention strategies and how nudges can confound one another.

Firstly, most papers included in this systematic review experimented with presentation nudges to influence disclosure of information. While some of these strategies are closely related to other nudging dimensions (e.g., notices and warnings are related to nudging with information), it is apparent that other nudging dimensions have yet to be tested widely in academic research. The different nudging dimensions are designed to address specific users’ biases and limitations in privacy decision making. Nudging with presentation aims at reducing users’ cognitive load, addressing such biases as framing effect, optimism bias, and overconfidence. Some of these biases and limitations can also be addressed by information nudges, to reduce information asymmetry [1]. Other biases, such as status quo bias and habit, are better addressed by default nudges, reducing users’ efforts by configuring user interface according to expectation. Therefore, it can be suggested that existing privacy nudging studies thus far have provided more support on ways to reduce users’ cognitive load, but less on ways to reduce efforts and increase motivation.

Secondly, evidence on the effects of nudging strategies on disclosure of personal information has been rather mixed and inconsistent in existing literature. However, our meta-analysis of 54 papers testing a variety of nudges demonstrated that overall nudging strategies were effective in influencing disclosure of personal information. These findings align with a recent meta-analysis [12] investigating the effectiveness of a wide range of nudging strategies (e.g., reminders, feedback, simplifications) in various contexts such as health, energy, policymaking, and privacy. However, in contrast to this previous work, we were able to examine the impact of different digital nudging strategies on personal information disclosure, showing that, while all nudging strategies had a statistically significant effect on disclosure, incentive and default interventions appeared to be particularly effective. However, it should be noted that the number of studies in the latter categories was relatively low. As a result, more primary research on privacy nudges is essential, along with new evidence syntheses to establish the impact of different nudges on information disclosure.

Thirdly, researchers have tested different combinations of multiple nudging strategies simultaneously. This base of research offers important conceptual and practical contributions as the research design allows for the comparison and identification of how one nudge can be used to amplify (or reduce) the effect of another. Practically speaking, using combinations of nudges may offer a “one-two punch” to influence disclosure behavior. From a theoretical point of view, when users are bound by a set of limitations, elucidating the extent to which different limitations (e.g., biases) play a role in privacy decision making (e.g., which will most likely lead to decision heuristics) is important. Research utilizing combination of nudges, designed to leverage or mitigate different users’ limitations, will allow for identifying most persistent behavioral problems in privacy decision and thus prioritizing relevant nudging strategies to alleviate them.

This study offers relevant practical and policy implications on a range of intervention strategies that can be used to nudge users into making responsible decisions when sharing personal information. For practitioners wishing to influence disclosure behaviors, this study showed that nudging to increase disclosure may be more effective than nudging to decrease disclosure. In addition, this study revealed which strategies are most likely to be effective in motivating users to share information, such as nudging with incentives (e.g., monetary rewards), defaults (e.g., opting-out option), and (some) presentation options (e.g., calls to action), while highlighting that caution may be advisable with strategies that seek to change behavior by educating users (i.e., information nudges). Importantly, this study contributes to a better understanding of ways to influence information disclosure in digital environments.

4.2. Limitations

Despite its contributions, this study has several limitations, which mainly come from variation in research design amongst the studies included in the systematic review. There has been a wide range of interventions strategies used in the studies. While these strategies can be grouped into various categories (i.e., nudging dimensions), some of them are not completely comparable. For example, different studies experimented with different presentations of privacy warnings, such as warnings with source, framing of warnings, and warning dialogues. The contexts of data disclosure are also varied. Furthermore, studies use different types of outcome variables representing disclosure; some studies used dichotomous variables, continuous variables while others use multiple disclosure outcomes. Also, some of the papers conducting experimental studies have deployed relatively small sample sizes, as a result, generalized conclusions should be drawn with caution. Future research is essential in order to evaluate the effectiveness of a wider range of privacy nudging strategies aiming to identify tried and tested practice for effective privacy nudging.

4.3. Future work

Based on the findings and due to the limitations of this systematic review, the following can be suggested to guide the direction of future research in this area. In order to alleviate issues with variations in study designs, future studies should endeavor to experiment using the same intervention approaches in different contexts (e.g., e-commerce vs online health environment), using different interfaces (e.g., embodied vs disembodied agents), and/or with different groups of participants (e.g., older adults). Having more empirical research on specific nudging approaches (e.g., through replication studies) will allow for more rigorous testing of consistency in nudging effects on disclosure. This can also be achieved by ensuring consistent use of outcome variables in experiments with nudging by encouraging the use of standardized, validated measures of disclosure behavior. Furthermore, it is important for future research to focus on conducting experiments testing actual information disclosure, instead of behavioral intentions, in real settings in order to provide more robust and generalizable results. Most importantly, the results of this systematic review indicate that the quest for best practices in privacy nudging should continue. For example, empirical studies experimenting with defaults as nudging strategies are relatively scant as most research has focused on presentation nudges. Also, reversibility and timing nudges have been largely neglected from existing empirical research. Overall, it is important for future research to experiment with a richer variety of interventions in order to inform best practices of nudging for privacy. Lastly, future research should focus on the refinement of the privacy nudging framework provided by Acquisti et al. [1]. Our review has uncovered significant variation in the category of presentation nudging. Within-category differences make it difficult to understand and identify the potential varying impact of different presentation nudges (e.g., visual, social, and linguistic cues) on information disclosure.

4.4. Conclusion

Digital nudging has been mooted as a tool to alter user privacy behavior. The present work offers comprehensive evidence on the influence of various nudging strategies on the disclosure of personal information in digital environments. The systematic review of 78 papers showed that the majority of empirical research has focused on presentation nudges to influence users’ privacy decisions; while the meta-analysis of 54 papers revealed that interventions aiming to increase disclosure is more effective than those aiming to decrease disclosure. Further research is needed to continue the quest for best practices in nudging.

References

  1. 1. Acquisti A, Adjerid I, Balebako RH, Brandimarte L, Cranor LF, Komanduri S, et al. Nudges for Privacy and Security: Understanding and Assisting Users’ Choices Online. ACM Comput Surv. 2017;50.
  2. 2. Proofpoint Security Awareness Training. State Of The Phish. 2019.
  3. 3. Thaler RH, Sunstein CR. Libertarian Paternalism. Am Econ Rev. 2003;93: 175–179. Available: http://www.jstor.org/stable/3132220
  4. 4. Schneider BYC, Weinmann M. Digital Nudging: Guiding Online User Choices through Interface Design. Commun ACM. 2018;61: 67–73.
  5. 5. Weinmann M, Schneider C, Brocke J vom. Digital Nudging. Bus Inf Syst Eng. 2016;58: 433–436.
  6. 6. Krol K, Preibusch S. Control versus Effort in Privacy Warnings for Webforms. Proc 2016 ACM Work Priv Electron Soc—WPES’16. 2016; 13–23.
  7. 7. Ben-shahar O, Chilton A. Simplification Of Privacy Disclosures: An experimental Test. J Legal Stud. 2016;45: 1–17. Available: https://doi.org/10.1086/688405
  8. 8. Halevi T, Kuppusamy TK, Caiazzo M, Memon N. Investigating users’ readiness to trade-off biometric fingerprint data. 2015 IEEE International Conference on Identity, Security and Behavior Analysis, ISBA 2015. 2015.
  9. 9. Bravo-Lillo C, Komanduri S, Cranor LF, Reeder RW, Sleeper M, Downs J, et al. Your Attention Please: Designing Security-decision UIs to make genuine risks harder to ignore. SOUPS ‘13 Proc Ninth Symp Usable Priv Secur. 2013; 6:1–6:12.
  10. 10. Higgins J, Green S. Cochrane Handbook for Systematic Reviews of Interventions. Higgins JP, Green S, editors. Chichester, UK: John Wiley & Sons, Ltd; 2011. https://doi.org/10.1002/9780470712184
  11. 11. European Parliament and Council of European Union (2016) Regulation (EU) 2016/679. Available: https://eur-lex.europa.eu/legal-content/EN/TXT/HTML/?uri=CELEX:32016R0679&from=EN
  12. 12. Hummel D, Maedche A. How effective is nudging? A quantitative review on the effect sizes and limits of empirical nudging studies. J Behav Exp Econ. 2019;80: 47–58.
  13. 13. Kitchenham B, Charters S. Guidelines for performing Systematic Literature Reviews in Software Engineering (EBSE 2007–001). Tech Rep. 2007.
  14. 14. Gerber N, Gerber P, Volkamer M. Explaining the privacy paradox: A systematic review of literature investigating privacy attitude and behavior. Comput Secur. 2018;77: 226–261.
  15. 15. Zhou Y, Zhang H, Huang X, Yang S, Babar MA, Tang H. Quality assessment of systematic reviews in software engineering: a tertiary study. Proceedings of the 19th International Conference on Evaluation and Assessment in Software Engineering. 2015. p. 14.
  16. 16. Carvalho L de F, Pianowski G. Pathological personality traits assessment using Facebook: Systematic review and meta-analyses. Comput Human Behav. 2017;71: 307–317.
  17. 17. Wu J, Lederer A. A Meta-Analysis of the Role of Environment-Based Voluntariness in Information Technology Acceptance. MIS Q. 2009;33: 419–432.
  18. 18. Tifferet S. Gender differences in privacy tendencies on social network sites: A meta-analysis. Comput Human Behav. 2019;93: 1–12.
  19. 19. Wohlin C. Guidelines for snowballing in systematic literature studies and a replication in software engineering. Proc 18th Int Conf Eval Assess Softw Eng—EASE ‘14. 2014; 1–10.
  20. 20. Carpenter S, Shreeves M, Brown P, Zhu F, Zeng M. Designing Warnings to Reduce Identity Disclosure. Int J Hum Comput Interact. 2018;34: 1077–1084.
  21. 21. Carpenter S, Zhu F, Kolimi S. Reducing online identity disclosure using warnings. Appl Ergon. 2014;45: 1337–1342. pmid:24161300
  22. 22. Carpenter S, Zhu F, Zeng M, Shreeves M. Expert Sources in Warnings May Reduce the Extent of Identity Disclosure in Cyber Contexts. Int J Hum Comput Interact. 2017;33: 215–228.
  23. 23. Larose R, Rifon N. Promoting i-Safety: Effects of Privacy Warnings and Privacy Seals on Risk Assessment and Online Privacy Behavior. J Consum Aff. 2007;42: 127–149. pmid:26276123
  24. 24. Samat S, Acquisti A. Format vs. Content: The Impact of Risk and Presentation on Disclosure Decisions. Symposium on Usable Privacy and Security (SOUPS), Santa Clara, California. 2017. Available: https://www.usenix.org/conference/soups2017/technical-sessions/presentation/samat-disclosure
  25. 25. Wang W. Building a Contemplative IS Workforce through Promoting Mindfulness in IS Design–A Case Study. Twenty-first Americas Conference on Information Systems, Puerto Rico, 2015. 2015. pp. 1–11.
  26. 26. Babula E, Mrzygłód U, Poszewiecki A. Consumers’ need of privacy protection–Experimental results. Econ Sociol. 2017;10: 74–86.
  27. 27. Gambino A, Kim J, Sundar SS, Ge J, Rosson MB. User Disbelief in Privacy Paradox: Heuristics that determine disclosure. Proceedings of the 2016 CHI Conference Extended Abstracts on Human Factors in Computing Systems—CHI EA ‘16. 2016. pp. 2837–2843.
  28. 28. Nosko A, Wood E, Kenney M, Archer K, De Pasquale D, Molema S, et al. Examining priming and gender as a means to reduce risk in a social networking context: Can stories change disclosure and privacy setting use when personal profiles are constructed? Comput Human Behav. 2012;28: 2067–2074.
  29. 29. Monteleone S, Bavel R Van, Rodríguez-Priego N, Esposito G. Nudges to Privacy Behaviour: Exploring an Alternative Approach to Privacy Notices. Luxembourg, Luxembourg: Publications Office of the European Union; 2015. https://doi.org/10.2791/142795
  30. 30. Mukherjee S, Manjaly JA, Nargundkar M. Money makes you reveal more: Consequences of monetary cues on preferential disclosure of personal information. Front Psychol. 2013;4: 1–5. pmid:23382719
  31. 31. Kim J, Gambino A, Sundar S, Rosson M, Aritajati C, Ge J, et al. Interface Cues to Promote Disclosure and Build Community: An Experimental Test of Crowd and Connectivity Cues in an Online Sexual Health Forum. Proc ACM Human-Computer Interact. 2018;2: 1–18.
  32. 32. Sah YJ, Peng W. Effects of visual and linguistic anthropomorphic cues on social perception, self-awareness, and information disclosure in a health website. Comput Human Behav. 2015;45: 392–401.
  33. 33. Wang Y, Leon PG, Acquisti A, Cranor LF, Forget A, Sadeh N. A field trial of privacy nudges for facebook. Proceedings of the 32nd annual ACM conference on Human factors in computing systems—CHI ‘14. 2014. pp. 2367–2376.
  34. 34. Mothersbaugh DL, Foxx WK, Beatty SE, Wang S. Disclosure Antecedents in an Online Service Context: The Role of Sensitivity of Information. J Serv Res. 2012;15: 76–98.
  35. 35. Lee D, Larose R. The Impact of Personalized Social Cues of Immediacy on Consumers ‘ Information Disclosure: Cyberpsychol Behav Soc Netw. 2011;14: 337–343. pmid:21204695
  36. 36. Brandimarte L, Acquisti A, Loewenstein G. Misplaced Confidences: Privacy and the Control Paradox. Soc Psychol Personal Sci. 2013;4: 340–347.
  37. 37. Ilany Tzur N, Zalmanson L, Oestreicher-Singer G. The Dark Side of User Participation—The Effect of Calls to Action on Trust and Information Revelation. ICIS 2016. 2016. pp. 1–15.
  38. 38. Keith MJ, Fredericksen JT, Reeves KS, Babb J. Optimizing Privacy Policy Videos to Mitigate the Privacy Policy Paradox. Proceedings of the 51st Hawaii International Conference on System Sciences. 2018. pp. 3698–3707.
  39. 39. Hughes-Roberts T, Furnell S. Privacy as a secondary goal problem: an experiment examining control. Proc Eighth Int Symp Hum Asp Inf Secur Assur (HAISA 2014). 2015;23: 382–393.
  40. 40. Chang D, Krupka EL, Adar E, Acquisti A. Engineering Information Disclosure: Norm Shaping Designs. Proceedings of the 2016 CHI Conference on Human Factors in Computing Systems—CHI ‘16. 2016. pp. 587–597.
  41. 41. John LK, Acquisti A, Loewenstein G. Strangers on a Plane: Context-Dependent Willingness to Divulge Sensitive Information. J Consum Res. 2011;37: 858–873.
  42. 42. Becker M, Matt C, Hess T. It’s not just about the product: How persuasive communication affects the disclosure of personal health information. Data Base Adv Inf Syst. 2020;51: 37–50.
  43. 43. Rudnicka A, Cox AL, Gould SJJ. Why do you need this? Selective disclosure of data among citizen scientists. Conf Hum Factors Comput Syst—Proc. 2019; 1–11.
  44. 44. Hanson J, Wei M, Veys S, Kugler M, Strahilevitz L, Ur B. Taking Data Out of Context to Hyper-Personalize Ads. CHI 2020, Honolulu, HI, USA. 2020. pp. 1–13.
  45. 45. Rodríguez-Priego N, van Bavel R, Monteleone S. The disconnection between privacy notices and information disclosure: an online experiment. Econ Polit. 2016;33: 433–461.
  46. 46. Balebako R, Péer E, Brandimarte L, Cranor L, Acquisti A. Is it the Typeset or the Type of Statistics? Disfluent Font does not Reduce Self-disclosure. Proceedings of the LASER 2013 (LASER 2013). 2013. pp. 1–11. Available: https://www.usenix.org/laser2013/program/balebako
  47. 47. Betzing JH, Tietz M, vom Brocke J, Becker J. The impact of transparency on mobile privacy decision making. Electron Mark. 2020;30: 607–625.
  48. 48. Rodríguez-Priego N, Van Bavel R. The Effect of Warning Messages on Secure Behaviour Online: results from a lab experiment. JRC Tech Reports. Joint Research Centre (Seville site).; 2016. pmid:27478588
  49. 49. Acquisti A, John LK, Loewenstein G. The Impact of Relative Standards on the Propensity to Disclose. J Mark Res. 2012;49: 160–174.
  50. 50. Zhang B, Xu H. Privacy Nudges for Mobile Applications: Effects on the Creepiness Emotion and Privacy Attitudes. Proceedings of the 19th ACM Conference on Computer-Supported Cooperative Work & Social Computing—CSCW ‘16. 2016. pp. 1674–1688.
  51. 51. Kroschke M, Steiner M. The Influence of Social Cues on Users’ Information Disclosure Intentions–The Case of Mobile Apps. ICIS 2017 Proc. 2017; 1–17.
  52. 52. Spottswood EL, Hancock JT. Should I Share That? Prompting Social Norms That Influence Privacy Behaviors on a Social Networking Site. J Comput Commun. 2017;22: 55–70.
  53. 53. Yeh C, Wang Y-S, Lin S-J, Tseng TH, Lin H-H, Shih Y-W, et al. What drives internet users’ willingness to provide personal information? Online Inf Rev. 2018;42: 923–939.
  54. 54. Wang J, Wang N, Jin H. Context Matters? How Adding the Obfuscation Option Affects End Users’ Data Disclosure Decisions. Proceedings of the 21st International Conference on Intelligent User Interfaces—IUI ‘16. 2016. pp. 299–304.
  55. 55. Peer E, Acquisti A. The impact of reversibility on the decision to disclose personal information. J Consum Mark. 2016;33: 428–436.
  56. 56. Alemany J, del Val E, Alberola J, García-Fornes A. Enhancing the privacy risk awareness of teenagers in online social networks through soft-paternalism mechanisms. Int J Hum Comput Stud. 2019;129: 27–40.
  57. 57. Meier Y, Schäwel J, Kyewski E, Krämer NC. Applying Protection Motivation Theory to Predict Facebook Users’ Withdrawal and Disclosure Intentions. ACM Int Conf Proceeding Ser. 2020; 21–29.
  58. 58. Gerlach J, Widjaja T, Buxmann P. Handle with care: How online social network providers’ privacy policies impact users’ information sharing behavior. J Strateg Inf Syst. 2015;24: 33–43.
  59. 59. Bhatia J, Breaux TD, Reidenberg JR, Norton TB. A Theory of Vagueness and Privacy Risk Perception. 2016 IEEE 24th International Requirements Engineering Conference. 2016. pp. 26–35.
  60. 60. Adjerid I, Peer E, Acquisti A. Beyond the Privacy Paradox: Objective versus Relative Risk in Privacy Decision Making. MIS Q. 2018;42: 465–488.
  61. 61. Adjerid I, Acquisti A, Brandimarte L, Loewenstein G. Sleights of Privacy: Framing, Disclosures, and the Limits of Transparency. Proceedings of the Ninth Symposium on Usable Privacy and Security—SOUPS ‘13. 2013. p. 1.
  62. 62. Knijnenburg BP, Kobsa A, Jin H. Preference-based Location Sharing: Are More Privacy Options Really Better? Chi ‘13. 2013. pp. 2667–2676.
  63. 63. Eling N, Rasthofer S, Kolhagen M, Bodden E, Buxmann P. Investigating users’ reaction to fine-grained data requests: A market experiment. Proceedings of the Annual Hawaii International Conference on System Sciences. 2016. pp. 3666–3675.
  64. 64. Vitale J, Tonkin M, Ojha S, Williams M-A. Be More Transparent and Users Will Like You:A Robot Privacy and User Experience Design Experiment. Proceedings of the 2018 ACM/IEEE International Conference on Human-Robot Interaction—HRI ‘18. 2018. pp. 379–387.
  65. 65. Zhu F, Carpenter S, Kulkarni A. Understanding identity exposure in pervasive computing environments. Pervasive Mob Comput. 2012;8: 777–794.
  66. 66. Junger M, Montoya L, Overink FJ. Priming and warnings are not effective to prevent social engineering attacks. Comput Human Behav. 2017;66: 75–87.
  67. 67. Feri F, Giannetti C, Jentzsch N. Disclosure of personal information under risk of privacy shocks. J Econ Behav Organ. 2016;123: 138–148.
  68. 68. Zhang B, Wu M, Kang H, Go E, Sundar SS. Effects of security warnings and instant gratification cues on attitudes toward mobile websites. Proceedings of the 32nd annual ACM conference on Human factors in computing systems—CHI ‘14. 2014. pp. 111–114.
  69. 69. Mamonov S, Benbunan-Fich R. The impact of information security threat awareness on privacy-protective behaviors. Comput Human Behav. 2018;83: 32–44.
  70. 70. Marreiros H, Tonin M, Vlassopoulos M, Schraefel MC. “Now that you mention it”: A survey experiment on information, inattention and online privacy. J Econ Behav Organ. 2017;140: 1–17.
  71. 71. Molina MD, Shyam Sundar S, Gambino A. Online privacy in public places: How do location, terms and conditions and VPN influence disclosure? Conf Hum Factors Comput Syst—Proc. 2019; 1–6.
  72. 72. Smith KH, Méndez Mediavilla FA, White GL. The Impact of Online Training on Facebook Privacy. J Comput Inf Syst. 2018;58: 244–252.
  73. 73. Aiken KD, Boush DM. Trustmarks, objective-source ratings, and implied investments in advertising: Investigating online trust and the context-specific nature of internet signals. J Acad Mark Sci. 2006;34: 308–323.
  74. 74. Tsai J, Kelley P, Drielsma P, Cranor L, Hong J, Sadeh N. Who’s Viewed You? The Impact of Feedback in a mobile location Sharing System. CHI 2009. 2009. pp. 901–913. pmid:21593138
  75. 75. Knijnenburg BP, Kobsa A, Jin H. Counteracting the negative effect of form auto-completion on the privacy calculus. Thirty Fourth International Conference on Information Systems,. 2013. pp. 1–21. Available: https://aisel.aisnet.org
  76. 76. Tschersich M. Comparing the configuration of privacy settings on social network sites based on different default options. Proceedings of the 48th Hawaii International Conference on System Sciences. 2015. pp. 3453–3462.
  77. 77. Baek YM, Bae Y, Jeong I, Kim E, Rhee JW. Changing the default setting for information privacy protection: What and whose personal information can be better protected? Soc Sci J. 2014;51: 523–533.
  78. 78. Lai Y-L, Hui K-L. Internet opt-in and opt-out: Investigating the Roles of Frames, Defaults and Privacy Concerns. Proc 2006 ACM SIGMIS CPR Conf Comput Pers Res Forty four years Comput Pers Res Achiev challenges Futur—SIGMIS CPR ‘06. 2006; 253.
  79. 79. Lu Y, Ou CXJ, Angelopoulos S. Exploring the effect of monetary incentives on user behavior in Online Sharing Platforms. Proceedings of the 51st Hawaii International Conference on System Sciences. 2018. pp. 3437–3444. http://hdl.handle.net/10125/50324
  80. 80. Li H, Sarathy R, Xu H. Understanding situational online information disclosure as a privacy calculus. J Comput Inf Syst. 2010;51: 62–71.
  81. 81. Steinfeld N. Trading with privacy: The price of personal information. Online Inf Rev. 2015;39: 923–938.
  82. 82. Hansen PG. The definition of nudge and libertarian paternalism: Does the hand fit the glove? Eur J Risk Regul. 2016;7: 155–174.
  83. 83. Saghai Y. Salvaging the concept of nudge. J Med Ethics. 2013;39: 487–493. pmid:23427215
  84. 84. Mettler T, Winter R. Are business users social? A design experiment exploring information sharing in enterprise social systems. J Inf Technol. 2016;31: 101–114.
  85. 85. Huang N, Hong Y, Chen P-Y, Wu S-Y. Digital Nudging for Online Social Sharing: Evidence from A Randomized Field Experiment. Proceedings of the 51st Hawaii International Conference on System Sciences | 2018. 2018. pp. 1483–1491.
  86. 86. Hui K, Teo H, Lee S. The Value of Privacy Assurance: An Exploratory Field Experiment. MIS Q. 2007;31: 19–33.
  87. 87. Xie E, Teo HH, Wan W. Volunteering personal information on the internet: Effects of reputation, privacy notices, and rewards on online consumer behavior. Mark Lett. 2006;17: 61–74.
  88. 88. Gabisch JA, Milne GR. Self-disclosure on the web: Rewards, safety cues, and the moderating role of regulatory focus. J Res Interact Mark. 2013;7: 140–158.
  89. 89. Preibusch S, Krol K, Beresford AR. The Privacy Economics of Voluntary Over-disclosure in Web Forms. Econ Inf Secur Priv. 2013; 183–209.
  90. 90. Frey RM, Bühler P, Gerdes A, Hardjono T, Fuchs KL, Ilic A. The Effect of a Blockchain-Supported, Privacy- Preserving System on Disclosure of Personal Data. Network Computing and Applications (NCA), 2017 IEEE 16th International Symposium. 2017. pp. 5–9. https://doi.org/10.1109/NCA.2017.8171385
  91. 91. Hutton L, Henderson T, Kapadia A. “Here I am, now pay me!”: Privacy concerns in incentivised location-sharing systems. WiSec 2014—Proceedings of the 7th ACM Conference on Security and Privacy in Wireless and Mobile Networks. 2014. pp. 81–86. https://doi.org/10.1145/2627393.2627416
  92. 92. Premazzi K, Castaldo S, Grosso M, Raman P, Brudvig S, Hofacker CF. Customer Information Sharing with E-Vendors: The Roles of Incentives and Trust. Int J Electron Commer. 2010;14: 63–91.
  93. 93. Weydert V, Desmet P, Lancelot-Miltgen C. Convincing consumers to share personal data: double-edged effect of offering money. J Consum Mark. 2020;37: 1–9.
  94. 94. Knijnenburg B, Kobsa A. Increasing sharing tendency without reducing satisfaction: finding the best privacy-settings user interface for social networks. Thirty Fifth International Conference on Information Systems,. 2016. pp. 1–8. Available: https://aisel.aisnet.org
  95. 95. Craciun G. Choice defaults and social consensus effects on online information sharing: The moderating role of regulatory focus. Comput Human Behav. 2018;88: 89–102.
  96. 96. Warberg L, Acquisti A, Sicker D. Can privacy nudges be tailored to individuals’ decision making and personality traits? Proc ACM Conf Comput Commun Secur. 2019; 175–197.
  97. 97. Knijnenburg BP, Kobsa A. Making decisions about privacy: Information disclosure in context-aware recommender systems. ACM Trans Interact Intell Syst. 2013;3: 1–33.
  98. 98. Review Manager (RevMan). The Cochrane Collaboration; 2020. p. Version 5.4.
  99. 99. Marino C, Gini G, Vieno A, Spada MM. A comprehensive meta-analysis on Problematic Facebook Use. Comput Human Behav. 2018;83: 262–277. pmid:29024900
  100. 100. Fanelli D, Costas R, Ioannidis JPA. Meta-assessment of bias in science. Proc Natl Acad Sci U S A. 2017;114: 3714–3719. pmid:28320937
  101. 101. Sterne JAC, Sutton AJ, Ioannidis JPA, Terrin N, Jones DR, Lau J, et al. Recommendations for examining and interpreting funnel plot asymmetry in meta-analyses of randomised controlled trials. BMJ. 2011;343. pmid:21784880
  102. 102. Maki A, Burns RJ, Ha L, Rothman AJ. Paying people to protect the environment: A meta-analysis of financial incentive interventions to promote proenvironmental behaviors. J Environ Psychol. 2016;47: 242–255.
  103. 103. Giles EL, Robalino S, McColl E, Sniehotta FF, Adams J. The effectiveness of financial incentives for health behaviour change: Systematic review and meta-analysis. PLoS ONE. Public Library of Science; 2014. pmid:24618584
  104. 104. Hollands GJ, French DP, Griffin SJ, Prevost AT, Sutton S, King S, et al. The impact of communicating genetic risks of disease on riskreducing health behaviour: Systematic review with meta-analysis. BMJ. 2016;352. pmid:26979548
  105. 105. Capewell S, Graham H. Will Cardiovascular Disease Prevention Widen Health Inequalities? PLoS Med. 2010;7: e1000320. pmid:20811492
  106. 106. Thaler RH. Nudge, not sludge. Science. 2018361: 431. pmid:30072515