Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Misleading graphs in context: Less misleading than expected

Abstract

Misleading graphs are a source of misinformation that worry many experts. Especially people with a low graph literacy are thought to be persuaded by graphs that misrepresent the underlying data. But we know little about how people interpret misleading graphs and how these graphs influence their opinions. In this study we focus on the effect of truncating the y-axis for a line chart which exaggerates an upgoing trend. In a randomized controlled trial, we showed participants either a normal or a misleading chart, and we did so in two different contexts. After they had seen the graphs, we asked participants their opinion on the trend and to give an estimation of the increase. Finally we measured their graph literacy. Our results show that context is the only significant factor in opinion-forming; the misleading graph and graph literacy had no effect. None of these factors had a significant impact on estimations for the increase. These results show that people might be less susceptible to misleading graphs than we thought and that context has more impact than a misleading y-axis.

Introduction

Graphs can misrepresent the underlying data in many ways, and this worries experts in different fields, from health communication [1] to data visualization [2].

Misleading graphs were omnipresent in the media during the covid pandemic, both in (social) media and governmental communications [3, 4]. Earlier it was shown that even in the scientific journal ‘Science’ 30% of all graphs contained errors [5], and a later study showed that 31% of the graphs in the Journal of American Medicine were ambiguous [6].

There are many ways for a graph to be ’wrong’. For example, there can be inconsistencies within a graph, such as a y-axis label contradicting the title or the y-axis being flipped [7]. There have been many attempts to categorize misleading graphs. For instance, [8] considers seven types of distortion where the first type is manipulating scale ratios: for instance, by truncating a y-axis such that it does not start at zero. A recent overview [2] also includes this error, along with many others. This study will focus on one specific form of misleading graphs: line charts that truncate the y-axis. We focus on line charts as these are some of the most commonly used types of visualization [9].

Graph literacy

Graph literacy is the individual’s ability to read, process, and comprehend data visualisations [10]. In general, people interpret graphs in a few steps [11]. At first, they identify the visual features, such as the direction or colour of a line. Secondly, they interpret the relations that the visual features represent. Lastly, they match these interpretations with the labelled variables.

Individuals with higher graph literacy can process graphs with greater ease and have an expected higher comprehension of the graph’s content [10]. Eye-tracking studies confirmed this effect and moreover found that people with low graph literacy have an overreliance on spatial-to-conceptual mappings, whilst people with high graph literacy spend more time looking at features such as the numbers on the scales and the axis labels [7].

There are multiple tests available to determine the level of graph literacy of an individual. For example, the objective graph literacy scale [12] is a test with 13 items. It is a reliable indicator of an individual’s graph literacy level, but it takes around ten minutes to complete, which is not manageable for some studies. The four-item short graph literacy test scale (SGL) [13] is a shortened version of the objective graph literacy scale with still sufficiently good psychometric properties.

The interpretation of misleading graphs

A few studies have studied the interpretation of conflicted graphs or other types of misleading graphs, which usually covered only bar charts. For instance, [14] researched bar graphs’ interpretation when shifting the y-axis, but with a sample size of only nine participants. [15] studied tilted bar charts and found that these seem to be interpreted similarly to the original bar charts. An older study that used different types of misleading bar charts showed that these improperly designed charts can influence decision making [16]. [10, 17] studied the interpretation of graphs with 3D effects and concluded that in general 2D bar charts result in higher comprehension than 3D bar charts.

Experiments showed that individuals with high graph literacy have a more accurate interpretation of self-conflicting graphs [7]. [18] found that bar charts with a truncated y-axis were found to be significantly less credible than proper graphs, but the information reproduction of skewed graphs was significantly higher.

Graphs in context

In real-life situations, graphs are used in a context, and we know that prior knowledge about a subject influences the reader’s interpretation of graphs [19]. The study by [18] also looked at graphs in different contexts: both deceptive and non-biased graphs were shown in a general news and a political news setting to communication students. In both settings, they found that ‘a student sample was effectively unfazed by a deceptive graphic’. They recommend that their study should be replicated in a random, representative sample.

It is known that people who do not have a strong opinion on a subject are more susceptible to persuasion using charts [20]. Therefore we decided to do a follow-up study to [18] with a random, representative sample and graphs in a made-up context that people would not have existing strong opinions on. Furthermore, we focus on line graphs since they are understudied in the current literature.

Research question and hypothesis

In this paper, we study the interpretation of line graphs. Specifically, we study whether and how truncating the y-axis and the context the graph is presented in influence the interpretation. We check whether this influence is moderated by graph literacy. Our expectation is that (i) non-shifted line graphs are better interpreted than shifted line graphs, (ii) that the context influences the judgement, and (iii) that these relations are stronger in participants with low graph literacy.

Methods

Design

The current study uses a quantitative survey-based research design with four distinct surveys. The surveys can be separated on two between-person factors, namely, context and graphical representation.

The contextualisation presents the participants a narrative that focuses on the fictional ’Bluebeak’, a non-native bird in Denmark. One story about the ’Bluebeak’ is that the bird is endangered, whereas the other story focuses on the disruption the ’Bluebeak’ causes to Denmark’s eco-system. Both narratives are provided in S1 Appendix.

The graphical design element accompanies the ’Bluebeak’ contextualisation, where one group of each contextualisation gets a different graph presented. The first possible graph being a line graph at which the y-axis starts at zero, whilst the other group gets a line graph that starts two standard deviations below the mean (in line with the suggestion by [14]). However, apart from the scaling of the vertical axis, both graphical representations are the same (see Fig 1).

Population & sample assignment

Data collection is done through Prolific (www.prolific.co), which uses representative samples from adult US citizens to fill out surveys and compensate them for their time. Prolific randomly assigns each participant to one of the four groups. The questionnaire (see S1 Appendix) is filled in unsupervised and online through Prolific.

Participants provided written informed consent. The Ethics Committee Psychology of the University of Groningen has approved this study (PSY-1920-S-0441).

Measurement

Variables.

The study uses two grouping factors, contextualisation and graphical representation. Additionally, we ask the participants to make a judgement call about the displayed situation, with scores ranging from 1 (very bad) to 5 (very good).

Participants are asked to estimate the proportional increase of the number of ’Bluebeaks’ in the previously presented graph to determine which type of graph the participants give the most accurate proportional increase estimation. Graph literacy is measured using the SGL scale (see: Instruments) to determine whether varying levels of graph literacy influence their judgement and estimation of Denmark’s ’Bluebeak’ situation.

Finally, a set of demographics are included to have options for control and evaluations of sample balance. The demographics are age, gender, and educational level. Age is measured on a categorical scale ranging from 1 (18–25) to 7 (66 +), and is assumed to be continuous within the analysis. Gender has options male, female and other. Education level is measured on a 7-point scale.

Instruments.

The short graph literacy scale consists of four items with different types of graphs displayed [13]. For instance, a pie chart is displayed and the reader is asked to assess the size of a slice of the pie. The possible scores on the short graph literacy scale range from 0 to 4, where zero indicates low graph literacy and a score of four indicates high graph literacy.

Data analysis plan.

Descriptive statistics for all variables are provided. Outliers are detected and removed based on two criteria. To avoid non-serious participants, all those with a participation duration exceeding the mean duration plus three times the standard deviation are excluded. Furthermore, outliers are detected using the MAD (median absolute deviation) on the percentage estimation variable since this is the sole variable that took manual input from the participant. For the MAD, the threshold of 2.5 is chosen, which is classified as a mildly conservative criterion [21].

The relation between axis type, context, judgement and interpretation is analysed via multiple linear regression. Two sets of models are built, one using judgment and the other using estimated proportional increase as the dependent variable. The percentage variable is centred around the correct answer; this way, the estimates received by the model are interpretable as the amount which the given estimates diverge from the correct answer.

For both sets of models, we make versions with (i) the main effects of axis-type and context, (ii) as (i) but with the inclusion of SGL as moderation variable, (iii) as (ii) but including the interaction between axis type and context. In all three models, age, education level, and gender will be included as covariates.

Power and sample size.

A posthoc power analysis is conducted for 247 participants, at which the multiple regression with α = .05 can discover slopes with small effects f2 = .05 with a power of .968.

Results

Descriptive statistics

In total, 313 people started with the questionnaire. A total of 66 participants were excluded due to the following reasons: not completing the questionnaire (13), not providing consent (1), exceeding our participation duration threshold (5), not specifying gender (6) and providing scores exceeding the MAD-threshold (41). Data for 247 participants were included. Descriptive statistics are provided in Tables 1 and 2 for these n = 247 participants.

thumbnail
Table 1. Descriptive statistics (mean score and standard deviation; the percentage for ‘male’) of the demographic characteristics.

https://doi.org/10.1371/journal.pone.0265823.t001

None of the bivariate Pearson correlations between pairs of the variables in Table 2 is significant. (r(judgment, percentage) = -0.002, t(245) = -.0.026, p = .979, 95% CI [-0.12; 0.123]; r(judgment, SGL) = 0.059, t(245) = 0.929, p = .353, 95%CI [-0.066; 0.183]; r(percentage, SGL) = -0.000, t(245) = -0.005, p = .996, 95%CI [-0.125; 0.124]), indicating no direct linear relations.

Linear regression

Two multiple regression models are fitted to assess the possible association between variables, one with judgment and the other with the centred percentage estimate as the dependent variable. For both models, the underlying statistical assumptions have been checked, and there were no reasons to abandon the choice for multiple linear regression. The results are given in Tables 3 and 4.

thumbnail
Table 3. Model summary for the multiple linear regression model with judgment as the dependent variable.

https://doi.org/10.1371/journal.pone.0265823.t003

thumbnail
Table 4. Model summary for the multiple linear regression model with percentage as the dependent variable.

https://doi.org/10.1371/journal.pone.0265823.t004

As predictor variables, the model has the type of graph (shifted vs non-shifted), the type of context (invasive vs endangered), the score on the short graph literacy (SGL) test and its interaction with the type of graph, and the demographic variables education level, age and gender.

When looking at the model with judgment as the dependent variable, context clearly is significant, with the average score in the endangered group 1.86 points higher than the ecosystem group (95% CI [1.64, 2.07], p < .001). None of the other variables is significant. When SGL is not included in the model, graph type is significant (with shifted graphs scoring, on average, 0.25 higher than normal graphs (95% CI [.03, .46], p = .023), but this effect vanishes when graph literacy is included.

A similar picture arises when looking at the model with the percentage estimate centred around the correct answer (approximately 16%) as the dependent variable. Here none of the variables is significant. (The age group 46–55 differs (p = .039) from the reference group, but this effect is non-significant when correcting for multiple comparisons.) When SGL is excluded from the model, the variable graph does become significant, with shifted graphs scoring 9.12 percentage points higher than non-shifted graphs (95% CI [5.86, 12.39], p < .001) but also here, the effect vanishes when SGL is included in the model.

Conclusion & discussion

This study looks at how a truncated y-axis, context, graph literacy and demographics influence the judgment and interpretation of graphs.

When participants are required to make a judgement call based on a displayed graph, it becomes apparent that the story itself matters more in the decision-making process than the shape of the graph’s y-axis. This is in contrast to the conclusions of [22], who reported that distorted graphs influence both judgement and comprehension. The deviating results could be due to different choices in the operationalisation of the studies. For instance, [22] studied bar graphs, whereas we focused on line graphs.

For the interpretation of the graph, truncating the y-axis does have a significant impact on percentage estimates, which is in line with [14], who claimed that the y-axis of a line graph should start at least 1.5 standard deviations to be interpreted correctly. However, we see that the influence of the truncated y-axis vanishes when we include graph literacy as a variable, which confirms that people with a low graph literacy have more problems with interpreting misleading graphs [7, 10, 23].

All in all, our randomized experiment on a representative sample of the US population confirms the hypothesis by [18] that people are unfazed by misleading charts—except for people with a low graph literacy when it comes to making estimations.

One limitation of our study is that some of the participants did not seem to be used to working with percentages and gave unreasonably high estimates (for instance, one participant estimated 3,000% for an increase that was, in reality, 19.4%). An explanation for these very high estimates could be that the upper bound of the line graphs was 3,000. Maybe these participants thought the exercise was to state to which value the number of ‘bluebeaks’ had increased. These outliers did not influence our conclusions, and we kept them in our data set, but it might be good to realise that there might be a deeper misunderstanding hidden between the variables we measure.

Another limitation is that in our study large groups of participants had the same level of graph literacy. Even though the Short Graph Literacy test graphs are very basic graphs, 75% percent of the participants answered only 1 or 2 questions correctly. This either means that many people do not know how to read and interpret simple graphs, which is in line with [24]. Furthermore it is in line with [25] who found statistical literacy in general, to be low, especially in the US population. It could also mean that the wording or topics of the exercises make them more difficult than they are supposed to be. Only a quarter of the participants demonstrated a decent graph literacy level. The Short Graph Literacy test [13] might be too short to separate between different levels of graph literacy. This might explain why graph literacy did not influence the judgement in our study. For follow-up studies, we would recommend using a more extensive graph literacy test.

For future research on this subject, it would also be interesting to see which kinds of contextualization lead to under or overestimating the correct answer. Furthermore, it might be good to see if these results hold for other graph types, such as bar charts.

Finally, we want to emphasize that although truncating the y-axis seems to have less influence on people’s judgement and understanding than previously thought, it is in no way an excuse to make or use misleading graphs.

References

  1. 1. Okan Y, Garcia-Retamero R, Cokely ET, Maldonado A. Individual Differences in Graph Literacy: Overcoming Denominator Neglect in Risk Comprehension. Journal of Behavioral Decision Making. 2012;25(4):390–401.
  2. 2. Cairo A. How Charts Lie: Getting Smarter about Visual Information. 1st edition. New York: Norton & Co.; 2019. 256 p.
  3. 3. Engledowl C, Weiland T. Data (Mis)representation and COVID-19: Leveraging Misleading Data Visualizations For Developing Statistical Literacy Across Grades 6–16. Journal of Statistics and Data Science Education. 2021;29(2):160–4.
  4. 4. Kwon ON, Han C, Lee C, Lee K, Kim K, Jo G, et al. Graphs in the COVID-19 news: A mathematics audit of newspapers in Korea. Educational Studies in Mathematics [Internet]. 2021. pmid:34934226
  5. 5. Cleveland WS. Graphs in Scientific Publications. The American Statistician. 1984;38(4):261.
  6. 6. Cooper RJ, Schriger DL, Close RJ. Graph literacy: The quality of graphs in a large-circulation journal. Annals of Emergency Medicine. 2002;40(3):317–22. pmid:12192357
  7. 7. Okan Y, Galesic M, Garcia-Retamero R. How People with Low and High Graph Literacy Process Health Graphs: Evidence from Eye-tracking. Journal of Behavioral Decision Making. 2016;29(2–3):271–94.
  8. 8. Bryan J. Seven Types of Distortion: A Taxonomy of Manipulative Techniques used in Charts and Graphs. Journal of Technical Writing and Communication. 1995;25(2):127–79.
  9. 9. Keizer J, Luz CF, Sinha B, Gemert-Pijnen L, Albers CJ, Beerlage-de Jong N, et al. The Visual Dictionary of Antimicrobial Stewardship, Infection Control, and Institutional Surveillance. Frontiers in Microbiology. 2021;12:743939. pmid:34777290
  10. 10. Duesbery L, Werblow J, Yovanoff P. Graph literacy Moderates the Interaction of Decorative Dimensionality and Cognitive Demand in Computer-Based Graph Comprehension. Journal of Educational Computing Research. 2011;45(1):75–93.
  11. 11. Shah P, Mayer RE, Hegarty M. Graphs as aids to knowledge construction: Signaling techniques for guiding the process of graph comprehension. Journal of Educational Psychology. 1999;91(4):690–702.
  12. 12. Galesic M, Garcia-Retamero R. Graph Literacy. Medical Decision Making. 2010;31(3):444–57.
  13. 13. Okan Y, Janssen E, Galesic M, Waters EA. Using the Short Graph Literacy Scale to Predict Precursors of Health Behavior Change. Medical Decision Making. 2019;39(3):183–95. pmid:30845893
  14. 14. Witt JK. Graph Construction. Meta-Psychology [Internet]. 2019;2018(895). Available from:
  15. 15. Okan Y, Garcia-Retamero R, G M., Cokely ET. When Higher Bars Are Not Larger Quantities: On Individual Differences in the Use of Spatial Information in Graph Comprehension. Spatial Cognition & Computation. 2012;12(2–3):195–218.
  16. 16. Arunachalam V, Pei BKW, Steinbart PJ. Impression Management with Graphs: Effects on Choices. Journal of Information Systems. 2002;16(2):183–202.
  17. 17. Spence I. Visual psychophysics of simple graphical elements. Journal of Experimental Psychology: Human Perception and Performance. 1990;16(4):683–92. pmid:2148585
  18. 18. Geidner N, Cameron J. Readers perceive deceptive graphics as less credible. Newspaper Research Journal. 2017;38(4):473–83.
  19. 19. Shah P. Graph Comprehension: The Role of Format, Content and Individual Differences. In: Andersom M, Meyer B, Olivier P, editors. Diagrammatic Representation and Reasoning [Internet]. Springer; 2002. https://doi.org/10.1007/978-1-4471-0109-3_10.
  20. 20. Pandey AV, Manivannan A, Nov O, Satterthwaite M, Bertini E. The Persuasive Power of Data Visualization. IEEE Transactions on Visualization and Computer Graphics. 2014 Dec;20(12):2211–20. pmid:26356935
  21. 21. Miller J. Short Report: Reaction Time Analysis with Outlier Exclusion: Bias Varies with Sample Size. The Quarterly Journal of Experimental Psychology Section A. 1991;43(4):907–12.
  22. 22. Woller-Carter MM, Okan Y, Cokely ET, Garcia-Retamero R. Communicating and Distorting Risks with Graphs: An Eye-Tracking Study. Proceedings of the Human Factors and Ergonomics Society Annual Meeting. 2012;56(1):1723–7.
  23. 23. Raschke RL, Steinbart PJ. Mitigating the Effects of Misleading Graphs on Decisions by Educating Users about the Principles of Graph Design. Journal of Information Systems. 2008;22(2):23–52.
  24. 24. Börner K, Maltese A, Balliet RN, Heimlich J. Investigating aspects of data visualisation literacy using 20 information visualisations and 273 science museum visitors. Information Visualization. 2016;15(3):198–213.
  25. 25. Gigerenzer G. Risk literacy. In: Brockman J, editor. This will make you smarter: New scientific concepts to improve your thinking. New York: Harper Perennial; 2012. p. 259–61.