Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Planning and performance in teams: A Bayesian meta-analytic structural equation modeling approach

  • Udo Konradt ,

    Roles Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Project administration, Writing – original draft, Writing – review & editing

    konradt@psychologie.uni-kiel.de

    Affiliation Institute of Psychology, Kiel University, Kiel, Germany

  • Alexander Nath,

    Roles Formal analysis, Methodology, Software, Writing – original draft, Writing – review & editing

    Affiliation Institute of Psychology, Kiel University, Kiel, Germany

  • Martina Oldeweme

    Roles Conceptualization, Data curation, Methodology, Project administration, Supervision, Writing – original draft, Writing – review & editing

    Affiliation Institute of Psychology, Kiel University, Kiel, Germany

Abstract

We meta-analyzed the relationship between team planning and performance moderated by task, team, context, and methodological factors. For testing our hypothesized model, we used a meta-analytic structural equation modeling approach. Based on K = 33 independent samples (N = 1,885 teams), a mixed-effects model indicated a non‐zero moderate positive effect size (ρ = .31, 95% CI [.20, .42]). Methodological quality, generally rated as adequate, was unrelated to effect size. Sensitivity analyses suggest that effect sizes were robust to exclusion of any individual study and publication bias. The statistical power of the studies was generally low and significantly moderated the relationship, with a large positive relationship for studies with high-powered (k = 42, ρ = .40, 95% CI [.27, .54]) and a smaller, significant relationship for low-powered studies (k = 54, ρ = .16, 95% CI [.01, .30]). The effect size was robust and generally not qualified by a large number of moderators, but was more pronounced for less interdependent tasks, less specialized team members, and assessment of quality rather than quantity of planning. Latent class analysis revealed no qualitatively different subgroups within populations. We recommend large‐scale collaboration to overcome several methodological weaknesses of the current literature, which is severely underpowered, potentially biased by self-reporting data, and lacks long-term follow-ups.

Introduction

Planning is considered to be an indispensable and essential prerequisite for individuals, teams, and organizations to take purposeful action and to achieve success [1]. In organizational psychology and organizational behavior, team planning generally refers to a group-level process, which is essential for accomplishing tasks, coordinating effort, and for general effectiveness [2]. Among the various benefits of planning are that it can help teams to establish and synchronize goals [3], improve their use of resources, schedule work tasks, and set appropriate deadlines [4]. In addition, planning serves to decide how to measure success and what milestones to set, which have an important function in monitoring and feedback according to several psychological theories of learning, motivation, and leadership. Consequently, theories of team planning (e.g., [5, 6]) and qualitative reviews (e.g., [1, 7]) in the organizational psychology literature suggest that team planning has positive effects on team performance. However, despite its theoretical and practical relevance for team processes and outcomes, we cannot draw robust and broad conclusions. Thus, a meta-analysis is warranted that provides an unbiased summary of the effect of team planning on team performance (Fig 1).

Existing theory and research on team processes suggest that the strength of planning’s effect on performance is conditional on several moderators that affect how teams act, including task, team, and context factors (e.g., [8]; for a review, see [9]). Understanding these conditions is essential to refine existing planning theory and is also important to inform practice and policy in organizations. Thus, the second aim of our research is to provide insights into the moderating role of these boundary factors.

With our study, we also aim to expand the focus on the positive effects of team planning, currently dominant in theory and research, to provide a more differentiated view that allows for the possibility that planning can have negative effects or a null effect on performance in teams. In the context of entrepreneurship research, the effects of planning at the organizational level were found to be heterogeneous, indicating no uniform direction [10, 11]. Studies in other contexts also provide further evidence that there can indeed be negative relationships between team planning and team performance (e.g., [12, 13]). Consequently, the third aim of this study is to advance understanding of team planning by identifying classes of positive and non-positive effects and specifying the conditions under which they occur.

In the present study, we systematically review the literature to explore how planning affects performance in teams and to examine whether there are systematic differences in this relationship. We combined a reasonable number of primary studies (96 effect sizes), involving a sizeable number of teams (N = 1,885). We used Bayesian meta-analytic structural equation modeling (MASEM) to improve the statistical power and estimate a true effect more precisely. Drawing on team effectiveness frameworks, we also contribute to the existing literature by analyzing a broad range of moderators that might attenuate or reverse the relationship between planning and performance. Specifically, by aggregating across small studies, which are typical of research in this area, we achieve adequate statistical power to examine factors that moderate the effect of team planning that have never before been addressed in any primary study. We also consider several methodological factors of studies to assess the reliability of the results.

Theory and hypotheses

Planning in teams

A review of the organizational literature revealed that team planning has been defined in various ways [14]. In the broadest sense, it refers to team processes—mechanisms by which team inputs are transformed into team outcomes [6]—directed at generating “a proposed sequence of actions intended to accomplish a set goal” ([15], p. 488). Marks and colleagues [16], who divide team processes into transitions, action, and interpersonal processes, regard planning as a transition process, referring to an activity that guides a team’s accomplishment of a goal. To be more specific, team planning involves the decision-making processes of specifying goals and determining the course of action, setting out rules and procedures for handling critical situations, allocating resources and assigning roles, and monitoring progress toward the goal [6, 17, 18]. Moreover, planning is described as “the active and conscious mental simulation of future action sequences intended to direct behavior and optimize the attainment of certain outcomes” [1, p. 214], In this meta-analysis, we define team planning as an interactive transition process taking place among team members, which is aimed at generating, evaluating, and selecting future activities to achieve the team’s goal or objective.

Team planning has been conceptualized in several ways that differ in terms of their focus on timing, function, or content (e.g., [5]). A prevalent conceptualization is that of Weingart [18], who conceptualizes team planning in terms of timing by distinguishing between pre-planning, which takes place before work on the task begins, and in-process planning, which takes place while it is underway. Similarly, Marks and colleagues [16] distinguish between reactive adjustment, where plans are adjusted while working on the task, deliberate planning, in which all the actions needed to complete the task are determined in advance, and contingency planning in which alternative plans are developed when the anticipated conditions change. Despite different foci, all definitions address the core elements of generating, evaluating, and selecting future activities relevant to the team’s goal or objective.

Planning in teams and performance

Transition processes such as team planning are considered to be a fundamental element in accomplishing tasks and team functioning, and essential for team effectiveness [1, 7, 16, 19]. In organizational psychology and organizational behavior, team effectiveness and team performance are broadly conceptualized in terms of tangible outputs from the team and reactions by team members [20]. Team performance refers to the extent to which teams fulfill the requirements of their job effectively, as identified formally in their job description [21], in terms of quality, functionality, and reliability [22]. Meta-analytic evidence suggests moderate positive effect sizes between transition processes and team performance [23]. Similarly, reflexivity theory [19] and team self-regulation theory (cf. [24]) propose that team planning increases team performance, as it leads to more effective problem solving and enhances coordination. Mumford and colleagues [1] also argue that planning promotes learning and enhances motivation in teams, both of which improve performance. George and colleagues’ research [25] provided empirical evidence for this contention at the organizational level. We therefore offer the following hypothesis:

  1. Hypothesis 1: Planning in teams is positively related to performance.

Moderators

Theories and frameworks of teamwork suggest that a range of factors either directly affect or moderate the relationship between transition processes in teams and performance. However, it remains unclear whether these factors play a significant role in the relationship between team planning and performance. In integrative and ‘meta’ frameworks of team effectiveness (e.g., [8, 26, 27]) moderators are divided into three categories: task, team, and context factors. For each of these categories, research on team planning has suggested that several variables may be relevant. Within frameworks focusing on team effectiveness, factors specific to the task include the type of task, task interdependence, time constraints, and task complexity, because all of these factors are assumed to be important determinants of the relationship between team processes and performance [19, 2830].

Team factors that influence the planning process may include authority, team size, team familiarity, skill level, and skill differentiation [23, 3134]. Context factors that can affect the planning process may include knowledge distribution, separate planning sessions, familiarity with the task, the type of planning (i.e., pre-action vs. in-action planning), and guided planning [18, 3538]. The relevant moderators are summarized in Table 1, together with definitions and corresponding literature.

Finally, we also consider whether effect sizes are contingent on methodological moderators such as the type of study, dimension of planning, study setting, and the measures used for both planning and performance (see Table 1). Previous research on teams has indicated that effect sizes are strongly dependent on whether the study is correlational or experimental [39] and conducted in real-life or hypothetical settings [40]. We also examine whether the scales used to measure planning and performance influence the effect sizes in terms of what dimension of planning is used [41] and also what measure [42]. In addition, we examine the methodological quality of each study, using as our criteria the sampling and representativeness (four items), the statistical analyses (two items), reliability and validity of the measures (two items), and the design and fidelity (two items, see S1 Appendix). Thus, we address the following research question:

  1. Research Question 1: Does the association between team planning and team performance vary according to the (a) task factors (type of task, task interdependence, time constraints, and task complexity), (b) team factors (authority, team size, development stage, skill level, and skill differentiation), (c) context factors (knowledge distribution, planning session, task familiarity, and treatment), and (d) methodological factors (type of study, dimension of planning, study setting, measure of planning, and measure of performance)?

Unobserved heterogeneity in the effects of planning on performance

Although team planning has been suggested to have a positive effect on task completion, team processes, emergent states, and performance [3, 7, 19], researchers have found there to be substantial heterogeneity in team planning effects and have also shown planning to have some detrimental effects (e.g., [12, 13]). This suggests unobserved heterogeneity in the sample, where different patterns of effect sizes may belong to different subpopulations and in classes not captured by observable characteristics [43]. Describing these classes and examining how they come about could advance theory and inform practice.

One approach to uncovering these different subpopulations is to consider the risks and potential flaws associated with team planning. Montoya and colleagues [44] outlined five major problems which teams encounter when planning, including a tendency to discuss shared rather than unique information, a bias towards pre-discussion preferences, unevenly distributed participation and contribution to the discussion, excessive cohesiveness, which leads to groupthink, and avoidance of team planning. Those tendencies are regarded as natural team behaviors, which limit their planning capabilities, deplete their resources, and thus make them less effective. This is consistent with resource allocation theory (e.g., [45]), which posits that planning requires time and cognitive resources and might be a distraction, preventing a team from focusing on actions, processes, and task performance [14, 46]. Likewise, Mumford and colleagues [1] argued that team planning can be a demanding and intensive activity; people must therefore be able and willing to invest their limited resources in it. Also, a detrimental effect might be particularly large if the task has a high routine component and follows an algorithm or well-known rules. In fact, the time-consuming and resource-intensive nature of planning is often seen to be at odds with enhancing performance; planning is regarded as a distraction from goal-directed actions and as a reason why resources become depleted. Lei and colleagues [12] provided empirical evidence of this; they showed that planning activities initially increased team performance, but too much planning had adverse effects on task work and subsequent performance. This implies that different patterns of relationships might be expected. When reviewing team planning research in entrepreneurship settings, Gielnik et al. [11] concluded that there were marked differences in effect sizes found in primary studies, suggesting that effect sizes do not follow a specific pattern. In light of this, we expect our sample to show two distinct classes of effect sizes: one with positive effect sizes and another with non-positive effect sizes. Therefore, we hypothesize that:

  1. Hypothesis 2: In terms of the relationship between planning and team performance, there are two classes of studies that differ in their mean effect size: One group shows a positive effect size and the other group shows a non-positive effect size.

Method

Literature search

The data were initially collected in October through December of 2017 and were augmented by a second collection period in March/April 2021. The data were collected in four ways including published as well as unpublished studies and datasets to minimize publication bias [47]. First, we conducted a systematic keyword search in six electronic databases (PsycInfo, Scopus/ScienceDirect, EBSCO, APA, PubPsych, ERIC). To account for unpublished data and grey literature we additionally searched the databases Google Scholar and SSRN. In sum, this search covered articles, dissertations, books, reports, theses, unpublished datasets, and preprints. Titles, abstracts, and keywords were searched for the following terms: team planning OR planning in teams OR group planning OR planning in groups. We considered studies published during the last 50 years. Second, we checked the studies included in meta-analyses of team processes as well as those included in references in primary studies [2, 23]. Third, we searched for unpublished data via mailing lists and contacted researchers likely to have worked in this area. This search yielded a total of 1,888 matches.

Inclusion and exclusion criteria

Within the scope of a preliminary screening, we excluded duplicates and studies conducted in clinical or educational settings. The latter were excluded because the cognitive abilities required for planning evolve over the human lifespan [48]. The remaining 863 studies were deemed to be eligible if they (a) included a quantitative measure of team planning, (b) included a quantitative measure of team performance, and (c) reported the zero-order correlation, or if the size and direction of the zero-order correlation could be computed from the information presented. Team planning did not have to be designated as such, but had to comply with the definition given above. Therefore, we included the study by Maynard and colleagues [49] and coded preparation activities as planning. Studies that looked at team plans and their consequences but not at the planning process itself (e.g., [50]) were not included (see Fig 2 for the PRISMA flow diagram showing inclusion and exclusion criteria for our meta-analysis). 33 studies with 96 effect sizes were eligible for inclusion in the meta-analysis.

thumbnail
Fig 2. PEISMA flow diagram for inclusion and exclusion criteria for the present meta-analysis.

https://doi.org/10.1371/journal.pone.0279933.g002

Coding procedure

Studies were coded for measures of team planning and team performance, the zero-order correlation between these two, and their respective reliabilities. When no reliability measure was reported we used the mean across all the reliabilities documented within that study as an estimate. For moderator analysis, we coded relevant task factors, team factors, context factors, and methodology factors (for details see Table 1). The coding scheme also included methodological quality (for details, see next section). Three trained reviewers (other than the authors) used a standardized coding format to extract data from studies and coded a third of the studies separately and in duplicate. The resulting average interrater reliability, as indicated by Cohen’s kappa coefficient, was κ = .86, generally regarded as “almost perfect” [51]. Any disagreements were resolved by consensus.

Methodological quality and statistical power assessment.

We evaluated the methodological quality of the studies using 10 scored items, which were then combined to give a summary score (see S2 Appendix). On a scale from 1 (low) to 31 (high), the methodological quality of the studies ranged between 16.00 and 25.00, with a mean score of 21.04 (median = 21.00). This indicates that they were, on average, of medium methodological quality. We also assessed the post-hoc power of each study using G*Power [52].

Statistical methods

We applied meta-analytic structural equation modeling (MASEM) using Mplus 8.4 [53]. MASEM combines the strengths of meta-analysis and structural equation modeling, including Bayesian estimation, allowing the modeling of error variance, evaluation of model fit, and examination of a complex set of theoretical hypotheses [54]. We conducted a one-stage three-level random effects model, which allowed each study to have its own population effect size [54]. This type of analysis therefore allowed us to account for nested effect sizes. We thereby estimated ‘regularized’ true study effects that are adjusted by within-study variability, showing better the true variation across studies and resulting in smaller and more uniform credible intervals compared to the sample mean [55]. In multilevel models it is assumed that the random effects are uncorrelated with the regressors. Violating this assumption creates endogeneity, which is a major threat to internal validity [56]. Following suggestions made by Antonakis and colleagues [56], we avoided this problem by adding interactions between the predictor and its cluster mean at level 1.

We used a Bayesian framework because of its advantages over conventional random-effects meta-analysis [57] and because it makes the parameter estimates and standard errors more reliable when fitting multilevel MASEM models with few higher level units (e.g., [58, 59]). We used standard settings for the prior distributions, which resulted in a flat, non-informative prior distribution. We employed four Markov chain Monte Carlo chains and set the number of iterations to 10,000. Of the 10,000 samples, every 20th sample was stored and used for posterior inference to ensure model stability and avoid autocorrelation.

Model fit was indicated by the deviance information criterion (DIC) [60], which is a Bayesian generalization of the Akaike information criterion (AIC) and Bayesian information criterion (BIC) in ML, with low values indicating a better model fit. A potential scale reduction (PSR) criterion [61] (Gelman & Rubin, 1992) was checked, with a value close to 1 indicating that the between-chain variation is small relative to the within-chain variation and providing evidence of convergence. Finally, to evaluate the convergence behavior of the Markov chains, we checked trace plots to ensure there were no trends, spikes, or other irregularities. With all our models, we found there were only negligible irregularities or no irregularities at all. 90% credibility intervals that exclude zero indicate statistically significant one-sided relationships.

Estimation of overall effect size.

Following Cheung [54], we estimated the average correlation for team planning and team performance at the within level in the three-level MASEM approach. This approach takes account of effect sizes nested within studies. We hereby corrected effect sizes for unreliability. We then transformed the effect sizes by dividing them by the known sampling variance of estimates, referred to as the “transformed variables approach” ([54] p. 314), which leads to a known variance of one. The levels were then specified as proposed by Cheung (2015) [54] as Level 2 (unique labels for effect sizes) and Level 3 (unique label for every study to provide information on how effect sizes are nested). The random slope was defined at the within level (Level 1). This leads to the following specification,

Level 1: ,

Level 2: ,

Level 3: , where yij is the transformed effect size, the transformation variable, eij the transformed residual, λij the true effect size for the ith effect size in the jth study, fj the true effect size in the jth study, residual terms for respective levels, and β0 the average population effect ([54]; see also [62]). To characterize average effect sizes, we used the criteria outlined by Bosco and colleagues [63] for gauging small (below 0.09), medium (between 0.09 and 0.37), and large (greater than 0.37) effect sizes.

Identification of heterogeneity.

Between-study heterogeneity was assessed in three ways. First, to determine whether there was substantial between-study variability in effect sizes that was not due to sampling error, we computed Cochran’s Q. Second, we computed , the heterogeneity variances at levels 2 and 3 respectively (i.e. and respectively). Third, I2 was computed to examine whether the heterogeneity was substantial. Low, moderate, and high levels of heterogeneity in observed effect sizes that are not due to sampling error are indicated by I2 levels of 25%, 50%, and 75% respectively [64].

Hypothesis testing

Consistent with the literature [65, 66] we used latent meta-regression within the three-level MASEM approach instead of subgroup analysis to clarify the relationship between multiple moderators and effect size. As moderators might be statistically confounded, multivariate techniques allow one to control for spurious influences [67]. The three-level random-effects meta-analysis was therefore extended to a mixed-effects model [54]. We then undertook a stepwise test of Research Questions 1a–d; to arrive at the most robust set of moderators, we used a series of latent multiple meta-regression analyses to identify significant predictors first in each category, then across categories using only those moderators which were initially significant. Specifically, the stepwise approach was applied to (a) reduce capitalization on chance and avoid an inflated Type I error rate, which can occur in separate linear regression analyses [68], (b) identify multicollinearity among subsets of correlates moderators, which can lead to inflated standard errors and biased estimates, and (c) identify a small number of moderators with the most explanatory power for our outcome. In step 1 we looked for multicollinearity among variables by examining bivariate zero-order correlations and we excluded any variables that correlated above the critical limit of .5 [69]. As a result, two variables (team familiarity and guided planning) were omitted from subsequent analyses. In addition, one moderator (knowledge distribution) was omitted due to the lack of variance reported in reviewed studies. In step 2 we performed four separate multiple regressions, with task factors (type of task, task interdependence, time constraints, and task complexity), team factors (authority, team size, skill level, and skill differentiation), context factors (planning session, task familiarity, and type of planning), and methodological factors (type of study, dimension of planning, study setting, type of planning data, and type of performance data) as the moderators. In step 3 we analyzed significant moderators of the step 2 analyses jointly in a multiple moderator model to examine the specific contribution of the most relevant moderators.

To test Hypothesis 2, we employed a latent class analysis (LCA) [70], applying a latent mixture modeling approach using the robust maximum likelihood estimator in Mplus 8.4. A latent mixture modeling approach offers many advantages over traditional cluster techniques, including that it enables rigorous statistical tests to be undertaken to assess model fit, provides formal criteria for making decisions about the adequate number of clusters, and allows different distributional forms to be used. We tested a series of unconditional LCA models, starting with a one-class model and then successive models with one additional class. We estimated the 10 best of 1,000 random starts with 40 iterations and checked that the results were sensitive to the number of random starts for the k class model. For class enumeration, the criteria we used were parsimony, substantive consideration, and fit indices. The information criteria were the AIC, consistent AIC (CAIC), BIC, and sample-size-adjusted BIC (SABIC) and likelihood ratio tests included the bootstrap likelihood ratio test (BLRT) and Lo–Mendell–Rubin likelihood ratio test (LMR-LRT). Classes that represented less than 5% of the sample were considered unacceptable (cf. [71]). We also checked the entropy values, which represent the quality of classification of individuals into latent classes. Entropy values over 0.8 indicate a good separation of the latent classes. The final model was evaluated in terms of its parsimony, theoretical justification, and interpretability [71, 72].

Statistical outliers.

Based on our theoretical underpinnings which we addressed in the derivation of Hypothesis 2, we regard influential cases that appear to deviate from the overall patterns in data not primarily as statistical outliers but as substantial effects. Following the roadmap recommended by Gibbert and colleagues [73], we therefore did not remove influential cases but kept them in the analysis. However, as statistical outliers can cause serious problems in statistical analyses, we followed the rules for identifying influential cases in LCA modeling. Consistent with recommended best practice [74], we checked statistical outliers that had either a very small class to be extracted [75] or that did not make substantive sense in terms of parameter estimates [76]. There were no statistical outliers.

Statistical power.

The post-hoc power of the three-level meta-analysis was assessed using Monte Carlo simulations, with estimates from our data as the population parameter values [59]. The robustness of the power estimates was checked using different seeds and 1,000 repetitions for each seed. The post-hoc power ranged between .986 and .990, which indicates excellent power to find moderate effect sizes (cf. [63]).

Sensitivity analysis.

We explored the robustness of the observed outcomes by determining whether the main effect results were biased, due to the use of non-informational priors [77]. For this purpose, we performed supplementary analyses using vague priors for respective parameters—that is, normally distributed priors for location parameters with an expected value of zero and a variance of 1010 and inverse gamma priors for variances. The parameters of the inverse gamma distribution were set to 0.001, 0.01, and 0.1. In addition, findings from meta-analyses might also be biased due to the inclusion of low-quality and/or underpowered studies. Therefore, we assessed both the methodological quality and the post-hoc achieved power of each study (median = 0.42, range: 0.05 to 1.00). As proposed by Coyne and colleagues [78], we examined whether the findings were robust when studies with less than 50% power were excluded from the analysis. Finally, we used the methodological quality and statistical power of studies as moderating variables of the overall effect size to provide statistical evidence of possible bias.

Results

Fig 3 displays the cumulative forest plot of the 80 effect sizes included in the meta-analysis, sorted from the largest to the smallest samples. Inspection of the overall effect sizes and their CIs indicated a very small drift of about 0.03, indicating no relationship between study size and effect size (cf. [79]). Hypothesis 1 predicted that team planning would be positively related to performance. Controlling for cluster means of Level 1 variables we found a positive significant relationship between team planning and performance (ρ = .31, PSD = 0.06, 95% CI [.20, .42], p < .001, k = 96 nested in K = 33 studies), supporting our hypothesis. The cluster mean of the Level 1 variables had no significant effect on the relationship (B = -0.28, PSD = 31.80, ns.), which indicated that there was no endogeneity bias in our model. Substantial heterogeneity was observed, as indicated by a significant Cochran’s Q (Q = 355.02, p < .001), a low to moderate I2 (I2 = 73.20%, p < .001), and significant heterogeneity variances at Level 2 and Level 3 ( = .024, p = .003; = .024, p < .001). Hence, the findings suggest that variance across primary studies was not only due to sampling error, providing a rationale for further moderator and latent class analyses.

thumbnail
Fig 3. Cumulative forest plot arranged by study size, from highest to lowest.

https://doi.org/10.1371/journal.pone.0279933.g003

Bivariate correlations of moderators are shown in Table 2. A stepwise meta-regression procedure was used to address Research Questions 1a–d to determine whether the association between team planning and performance varies according to task, team, context, and methodological factors (Table 3). Moderator analyses revealed that task interdependence (B = -0.23, PSD = 0.11, p = .02), skill differentiation (B = -0.29, PSD = 0.12, p = .01) as well as dimension of planning (B = -0.18, PSD = 0.06, p < .01) moderated the overall. This indicated that the effects were significantly larger when task interdependence was moderate compared to high (there were no studies with low/absent task interdependence), skill differentiation was low compared to moderate skill differentiation (there were no studies with high skill differentiation), and when the quality of team planning was assessed compared to the quantity of planning. A summary of study results regarding Hypothesis 1 and Research Question 1a-d can be found in Fig 4.

thumbnail
Fig 4. Summary of study results.

Note a 1 = low, 2 = medium, 3 = high. b 1 = low, 2 = medium. c 1 = qualitative, 2 = quantitative. *p < .05. **p < .01.

https://doi.org/10.1371/journal.pone.0279933.g004

thumbnail
Table 2. Assessment of multicollinearity: Correlations between moderators.

https://doi.org/10.1371/journal.pone.0279933.t002

thumbnail
Table 3. Results of the Bayesian stepwise meta-regression.

https://doi.org/10.1371/journal.pone.0279933.t003

Hypothesis 2 predicted there would be two classes with distinct effect sizes in terms of the relationship between planning and performance in teams. As shown in Table 4, the two-class model outperformed the one-class model, as indicated by the information criteria and BLRT. SABIC, the most effective model selection index for class enumeration in growth mixture models and best the index to use when the sample size is small [80], also indicated a better model fit. However, the two-class model produced a small class (k = 4), which violates the assumption of there being a substantial class size [71]. Thus, Hypothesis 2 was rejected. This result indicates that there is no substantial variation in the relationships between planning and performance. Rather, we can assume a homogeneous class of relationships in the data, which supports our previous finding of a moderately sized positive association between team planning and performance.

Sensitivity analyses

We performed sensitivity analyses to explore the robustness of the meta-analytic estimates. First, the previous sensitivity analyses indicated no bias resulting from the use of non-informational priors. Differences between the estimates were found only in the third decimal, suggesting that the results were stable when using non-informational rather than vague informational priors. To check whether the relationship found was robust when we accounted for methodological quality and achieved power of primary studies, we undertook two meta regressions. The findings suggested that the relationship was not biased by the quality of primary studies (B = -0.02, PSD = 0.01, ns.) but was biased significantly by achieved power (B = 0.58, PSD = 0.06, p < .001), as elevated relationships were found in studies with higher power. To clarify this, we tested whether the findings were robust when only primary studies with a power higher than 0.5 (k = 42) were used in the analysis. This led to a larger mean effect size point estimate and CI (ρ = .40, PSD = 0.07, 95% CI [.27, .54], p < .001) in studies with higher achieved power. In contrast, we found a smaller, significant effect size in studies with a power less than or equal to 0.5 (k = 54, PSD = 0.07, 95% CI [.01, .30], p = .02). Thus, the magnitude of the effect size was sensitive to power limitations, and the smaller effect size estimate for the full set of studies constituted the lower point estimator of the true effect.

Furthermore, as Table 3 indicates, Study Setting and Type of Study do not influence or bias the results of the meta-analysis in a multivariate setting. To check if this holds true for the univariate cases of Study Setting and Type of Study as well as Sample we conducted univariate subgroup analyses revealing no significant moderation for all three variables.

Discussion

The aim of this study was to meta-analyze studies on the effectiveness of team planning for team performance and to identify moderator variables that affect the effect sizes obtained. In addition, while most of the literature to date has focused on the positive effects of planning on performance, we also examined samples with null association between planning and performance.

Our findings provide clear evidence of a moderately sized positive association between team planning and performance, which corroborates existing theoretical and empirical literature [e.g., 5–7]. There are three ways in which these findings are substantiated. First, by using a mixed linear model to estimate true study effects, we accounted for within-study variability. These regularized model estimates [55] provide a better indication of the true variation across studies and result in smaller and more uniform credible intervals compared to the sample mean. Second, the sensitivity analyses and publication bias testing revealed that our findings were generally robust, except in the case of power. Because the effect size estimate across studies with adequate power led to elevated effects, our general finding represents the lower boundary of the estimate. Third, the methodological quality of studies did not substantially alter the result. However, as the parameter estimates were not inflated by low-power studies, the effect size estimate for the full set of studies represents a rather conservative point estimate of the true effect.

The results of this study also suggest that the effect size is only marginally affected by the boundary conditions. Apart from task interdependence, skill differentiation, and dimension of planning (quality vs. quantity), none of the other moderators discussed in substantive theory [8, 26, 27] affected the mean effect sizes. This justifies the conclusion that our result is generalizable across different tasks, team settings, and contexts. Our finding—that the dimension of planning moderated the effect sizes and was associated with larger effect sizes when quality rather than quantity of planning was assessed—is in line with the notion that the sheer amount of planning does not necessarily lead to higher performance levels if the quality of planning is poor or the plan itself is unsuitable [40] A few studies of transition processes also demonstrated the importance of the quality, as opposed to the quantity, of team processes (e.g., [81, 82]). For example, Otte and colleagues [81] revealed that the amount of team reflection was negatively related to performance improvement and the quality of team reflection was positively related to it when being considered simultaneously. Likewise, Santos and colleagues [82] found that when a team’s mental model lacks quality, a highly similar mental model may be disruptive to team functioning. Furthermore, we found that task interdependence negatively moderated effect sizes, meaning that moderate levels of task interdependence were associated with a greater relationship between team planning and performance than high levels of task interdependence. However, there were no studies coded with a 1 = “low”, meaning that the moderator was dichotomous. Hence, we cannot draw any conclusions about the relationship between team planning and performance when task interdependence is low or even absent. The findings together with past research [83] might suggest a quadratic association, meaning that at least a moderate task interdependence is needed for planning to be effective at the first glance, but very high levels of task interdependence might be detrimental to the functioning of team planning. Hence, findings should be taken with caution and further research on team planning in situations where task interdependence is low as well as the investigation of possible quadratic or curvilinear associations is needed.

In addition, our results showed that when the level of team members’ specialized knowledge or functional capacity was low, effect sizes were higher [34]. These results can be explained by the fact that members in teams characterized by low skill differentiation do not have specific roles associated with their abilities. Consequently, those teams must determine their specific and distinctive roles, determine objectives and tasks. However, there were no studies with high levels of skill differentiation, hence we advocate future research to investigate the relationship in those situations.

There are several possible reasons for the marginal influence of moderators. First, the moderating effects found in previous studies are based on univariate analyses, which, when the moderators overlap empirically, can result in parameter bias [67]. Because we controlled for spurious influences, our results are more conclusive with respect to potential statistical confounding. Also, several primary studies were underpowered, which could have led to an increased risk of Type I error and inflated effect-size estimates [84]. This risk is particularly present in team studies, where the sample size shrinks because team member responses are aggregated to the team level. As we aggregate across studies, our results at the meta-analytic level provide the largest statistical power of true moderator effects being the first one to reach an adequate level [85]. One must also keep in mind that our comparisons are non-experimental, which bears the risk that our results may be confounded by other variables. Finally, the heterogeneity of many moderator variables was restricted, often limited to very few categories, and only a few (or no) studies were available, which might be a threat to validity [79]. Although a common practice in meta-analysis, one must be cautious of generalizing findings unless there is sufficient causal evidence to support the results.

We also demonstrated that all effect sizes have true positive effects. This finding contradicts the results of previous research (e.g., [12, 13]) and confirms our general result. Nevertheless, taking into account heterogeneity in the variance of unobserved effects could open up a new avenue for research.

Finally, the current review also enables us to identify critical gaps in the literature. Very few of the studies (e.g., [86, 87]). included in this review examined temporal aspects of change in team planning. Adopting a temporal lens in research and looking at the duration, frequency, timing, and sequence of planning processes [88, 89] would improve the literature substantially. We thus advocate studies with strong designs, such as intensive longitudinal designs [90] involving 20 or more waves of data collection, as this may help to overcome this limitation and provide a better understanding of the team-specific dynamics (cf. [91]). It is also noteworthy that only a few studies differentiated between the quality and quantity of team planning. As mentioned above we found a moderating effect of the dimension of planning (assessment of quality or quantity), which might indicate a similar functioning as the aforementioned transition processes. However, it is not fully understood how these dimensions are related. In this meta-analysis we were not able to analyze the quantity and quality of team planning simultaneously, like, for example, Otte and colleagues [81] did with team reflection. Future research should thus expand on the current findings by assessing how the quality and quantity of planning in combination affect team outcomes (see also [92]).

There are several research and managerial implications from the results of this study. The positive effect of team planning on performance was found to be robust with respect to nearly all moderator variables, indicating general importance of planning to improve performance. This applies in particular when team members do not have specific and distinctive roles. This suggests that managers should encourage their teams to plan and create a supportive environment by establishing guidelines for effective planning, promote appropriate communication methods, and make use of project management tools. Furthermore, as the findings for the dimension of planning suggest, the quality of planning rather than the sheer amount of planning should be emphasized within planning activities. The robust effect size estimator provides valid information to calculate the economic benefits and cost of interventions in teams that are seeking to improve their planning [93]. Finally, our findings provide useful information for researchers to select appropriate sample sizes for future studies and determine realistic prior distribution in Bayesian models, which outperform noninformative, weakly informed, and diffuse priors in model parameter estimates [57].

Limitations and future directions

This study and its results have some limitations, which suggest potential directions for future research. The main limitations of this meta-analysis pertain to the methodological limitations of primary studies. First, only a relatively small number of empirical studies were available, and the methodological quality in terms of the lack of randomized controlled trials was not convincing, which limits the validity and generalizability of our results. Another limitation is that the findings might be biased by the power achieved in primary studies [78]. Less than half of the effect sizes (k = 42) analyzed were from studies that had achieved adequate post-hoc power. We found that studies that were adequately powered showed both a larger positive effect and a narrower confidence interval compared to the main effect. For this reason, we advocate future research to increase study power not only by referring to the omnipresent call for larger sample sizes. However, as applied researchers are compelled to limit their sample sizes for economic and other reasons, alternative ways of increasing power are to use within-subject or longitudinal research designs, extreme groups, planned missing designs, or potential covariates in experimental studies (see [94], for a review). Another way to increase power is to decrease standard errors by adding more explanatory variables (i.e., covariates) that reduce error variance in the criteria.

Conclusion

With our study, we aimed to examine the extent of the association between team planning and team performance and to expand the rather narrow focus of much research on the positive effects of planning, enabling us to provide a more nuanced view. Based on the data available, we found an overall positive medium effect size, which was only marginally affected by a wide range of contexts and settings.

Supporting information

S1 Appendix. Descriptive statistics of studies used in the meta-analysis [2, 4, 5, 12, 13, 18, 24, 36, 41, 49, 83, 87, 95113].

https://doi.org/10.1371/journal.pone.0279933.s002

(DOCX)

S2 Appendix. Checklist for the assessment of the methodological quality of the studies reviewed.

https://doi.org/10.1371/journal.pone.0279933.s003

(DOCX)

Acknowledgments

Declarations: Availability of data and materials: The dataset used for statistical analysis is available from the OSF database via https://osf.io/jdywa/?view_only=bf57cf3a284e42b1a514da2d97ae584c. Additional materials are available from the corresponding author on reasonable request.

Author note: We greatly appreciate the assistance of Alina Christin Böckmann, Anna-Lena Immke, and Jes Schulte am Hülse with data acquisition and data processing as well as Sabrina Krys for help with the manuscript.

References

  1. 1. Mumford M. D., Schultz R. A., & van Doorn J. R. (2001). Performance in planning: Processes, requirements, and errors. Review of General Psychology, 5, 213–240. https://doi.org/10.1037/1089-2680.5.3.213
  2. 2. DeChurch L. A., & Haas C. D. (2008). Examining team planning through an episodic lens: Effects of deliberate, contingency, and reactive planning on team effectiveness. Small Group Research, 39, 542–568. https://doi.org/10.1177/1046496408320048
  3. 3. Stout R., Cannon‐Bowers J., Salas E., & Milanovich D. M. (1999). Planning, shared mental models, and coordinated performance: An empirical link is established. Human Factors, 41, 61–71. https://doi.org/10.1518/001872099779577273
  4. 4. Janicik G. A., & Bartel C. A. (2003). Talking about time: Effects of temporal planning and time awareness norms on group coordination and performance. Group Dynamics: Theory, Research, and Practice, 7, 122–134. https://doi.org/10.1037/1089-2699.7.2.122
  5. 5. Fisher D. M. (2014). Distinguishing between taskwork and teamwork planning in teams: Relations with coordination and interpersonal processes. Journal of Applied Psychology, 99, 423–436. pmid:24188388
  6. 6. Ilgen D. R., Hollenbeck J. R., Johnson M., & Jundt D. (2005). Teams in organizations: From input-process-output models to IMOI models. Annual Review of Psychology, 56, 517–543. pmid:15709945
  7. 7. Mumford M. D., Mecca J. T., & Watts L. L (2015). Planning processes: Relevant cognitive operations. In Mumford M. D. & Frese M. (Eds.), The Psychology of Planning in Organizations (pp. 9–30). Routledge.
  8. 8. Campion M. A., Medsker G. J., & Higgs A. C. (1993). Relations between work group characteristics and effectiveness: Implications for designing effective work groups. Personnel Psychology, 46, 823–847. https://doi.org/10.1111/j.1744-6570.1993.tb01571.x
  9. 9. Salas E., Stagl K. C., Burke C. S., & Goodwin G. F. (2007). Fostering team effectiveness in organizations: Toward an integrative theoretical framework. In Shuart B., Spaulding W., & Poland J. (Eds.), Nebraska Symposium on Motivation: Vol. 52. Modeling Complex Systems (pp. 185–243). University of Nebraska Press. https://doi.org/10.2307/j.ctt1djmh0r.9
  10. 10. Brinckmann J., Grichnik D., & Kapsa D. (2010). Should entrepreneurs plan or just storm the castle? A meta-analysis on contextual factors impacting the business planning–performance relationship in small firms. Journal of Business Venturing, 25, 24–40. https://doi.org/10.1016/j.jbusvent.2008.10.007
  11. 11. Gielnik M. M., Frese M., & Stark M. S. (2015). Planning and entrepreneurship. In Mumford M. D. & Frese M. (Eds.), The Psychology of Planning in Organizations: Research and Applications (pp. 289–311). Routledge. https://doi.org/10.4324/9780203105894
  12. 12. Lei Z., Waller M. J., Hagen J., & Kaplan S. (2016). Team adaptiveness in dynamic contexts: Contextualizing the roles of interaction patterns and in-process planning. Group & Organization Management, 41, 491–525. https://doi.org/10.1177/1059601115615246
  13. 13. Waller M. J. (1999). The timing of adaptive group responses to nonroutine events. Academy of Management Journal, 42, 127–137. https://doi.org/10.5465/257088
  14. 14. Frese M., Mumford M. D., & Gibson C. (2015). Organizational planning: The psychology of performance introduction. In Mumford M. D. & Frese M. (Eds.), The Psychology of Planning in Organizations (pp. 17–24). Routledge.
  15. 15. Rosen M. A., & Dietz A. S. (2017). Team performance measurement. In Salas E, Rico R., & Passmore J. (Eds.), The Wiley Blackwell Handbook of the Psychology of Team Working and Collaborative Processes (pp. 479–502). Wiley Blackwell.
  16. 16. Marks M. A., Mathieu J. E., & Zaccaro S. J. (2001). A temporally based framework and taxonomy of team processes. Academy of Management Review, 26, 356–376. https://doi.org/10.2307/259182
  17. 17. Rousseau V., & Aubé C. (2010). Team self-managing behaviors and team effectiveness: The moderating effect of task routineness. Group & Organization Management, 35, 751–781. http://dx.doi.org/10.1177/1059601110390835
  18. 18. Weingart L. (1992). Impact of group goals, task component complexity, effort, and planning on group performance. Journal of Applied Psychology, 77, 803–809. https://doi.org/10.1037/0021-9010.77.5.682
  19. 19. West M. A. (2000). Reflexivity, revolution, and innovation in work teams. In Beyerlein M. M., Johnson D., & Beyerlein S. T. (Eds.), Product Development Teams (Vol. 5, pp. 1–29). JAI Press.
  20. 20. Mathieu J. E., & Gilson L. L. (2012). Criteria issues and team effectiveness. In Kozlowski S. W. J. (Ed.), The Oxford Handbook of Organizational Psychology, Volume 2. Oxford University Press. https://doi.org/10.1093/oxfordhb/9780199928286.013.0027
  21. 21. Borman W. C., & Motowidlo S. M. (1993). Expanding the criterion domain to include elements of contextual performance. In Schmitt N. & Borman W. C. (Eds.), Personnel Selection in Organizations (pp. 71–98). Jossey-Bass.
  22. 22. Ancona D. G., & Caldwell D. F. (1992). Bridging the boundary: External activity and performance in organizational teams. Administrative Science Quarterly, 634–665. https://doi.org/10.2307/2393475
  23. 23. LePine J. A., Piccolo R. F., Jackson C. L., Mathieu J. E., & Saul J. R. (2008). A meta‐analysis of teamwork processes: Tests of a multidimensional model and relationships with team effectiveness criteria. Personnel Psychology, 61, 273–307. https://doi.org/10.1111/j.1744-6570.2008.00114.x
  24. 24. Mehta A., Feild H., Armenakis A., & Mehta N. (2009). Team goal orientation and team performance: The mediating role of team planning. Journal of Management, 35, 1026–1046. https://doi.org/10.1177/0149206308326773
  25. 25. George B., Walker R., Monster J. (2019). Does strategic planning improve organizational performance? A meta‐analysis. Public Administration Review, 79, 810–819. https://doi.org/10.1111/puar.13104
  26. 26. Dinh J. V., & Salas E. (2017). Factors that influence teamwork. The Wiley Blackwell Handbook of the Psychology of Team Working and Collaborative Processes, 13–41. https://doi.org/10.1002/9781118909997.ch2
  27. 27. Tannenbaum S. I., Beard R. L., & Salas E. (1992). Team building and its influence on team effectiveness: An examination of conceptual and empirical developments. Advances in Psychology, 82, 117–153. http://dx.doi.org/10.1016/S0166-4115(08)62601-1
  28. 28. McGrath J. E. (1984). Groups: Interaction and Performance (Vol. 14). Prentice-Hall.
  29. 29. Guzzo R. A., & Shea G. P. (1992). Group performance and intergroup relations in organizations. In Dunnette M. D. & Hough L. (Eds.), Encyclopedia of Group Processes & Intergroup Relations (2nd ed., pp. 269–313). Annual Reviews.
  30. 30. Shure G. H., Rogers M. S., Larsen I. M., & Tassone J. (1962). Group planning and task effectiveness. Sociometry, 25, 263–282. https://doi.org/10.2307/2786128
  31. 31. Langfred C. W. (2005). Autonomy and performance in teams: The multilevel moderating effect of task interdependence. Journal of Management, 31, 513–529. https://doi.org/10.1177/0149206304272190
  32. 32. Harrison D. A., Mohammed S., McGrath J. E., Florey A. T., & Vanderstoep S. W. (2003). Time matters in team performance: Effects of member familiarity, entrainment, and task discontinuity on speed and quality. Personnel Psychology, 56, 633–669. https://doi.org/10.1111/j.1744-6570.2003.tb00753.x
  33. 33. Mathieu J. E., Heffner T. S., Goodwin G. F., Salas E., & Cannon-Bowers J. A. (2000). The influence of shared mental models on team process and performance. Journal of Applied Psychology, 85, 273–283. pmid:10783543
  34. 34. Lee S. M., Koopman J., Hollenbeck J. R., Wang L. C., & Lanaj K. (2015). The team descriptive index (TDI): A multidimensional scaling approach for team description. Academy of Management Discoveries, 1, 91–116. https://doi.org/10.5465/amd.2013.0001
  35. 35. Stasser G., & Titus W. (1985). Pooling of unshared information in group decision making: Biased information sampling during discussion. Journal of Personality and Social Psychology, 48, 1467–1478. https://doi.org/10.1037/0022-3514.48.6.1467
  36. 36. Chen G., Thomas B., & Wallace J. C. (2005). A multilevel examination of the relationships among training outcomes, mediating regulatory processes, and adaptive performance. Journal of Applied Psychology, 90, 827–841. pmid:16162057
  37. 37. Earley P. C., Lee C., & Hanson L. A. (1990). Joint moderating effects of job experience and task component complexity: Relations among goal setting, task strategies, and performance. Journal of Organizational Behavior, 11, 3–15. https://doi.org/10.1002/job.4030110104
  38. 38. Smith-Jentsch K. A., Cannon-Bowers J. A., Tannenbaum S. I., & Salas E. (2008). Guided team self-correction: Impacts on team mental models, processes, and effectiveness. Small Group Research, 39, 303–327. https://doi.org/10.1177/1046496408317794
  39. 39. Mitchell G. (2012). Revisiting truth or triviality: The external validity of research in the psychological laboratory. Perspectives on Psychological Science, 7, 109–117. pmid:26168439
  40. 40. Tannenbaum S. I., & Cerasoli C. P. (2013). Do team and individual debriefs enhance performance? A meta-analysis. Human Factors, 55, 231–245. pmid:23516804
  41. 41. Smith K. G., Locke E. A., & Barry D. (1990). Goal setting, planning, and organizational performance: An experimental simulation. Organizational Behavior and Human Decision Processes, 46, 118–134. https://doi.org/10.1016/0749-5978(90)90025-5
  42. 42. Andersson D., Rankin A., & Diptee D. (2017). Approaches to team performance assessment: a comparison of self-assessment reports and behavioral observer scales. Cognition, Technology & Work, 19, 517–528. https://doi.org/10.1007/s10111-017-0428-0
  43. 43. McLachlan G., & Peel D. (2000). Finite Mixture Models. Wiley.
  44. 44. Montoya A. C., Carter D. R., Martin J., & DeChurch L. A. (2015). The five perils of team planning: Regularities and remedies. In Mumford M. D. (Ed.), The Psychology of Planning in Organizations: Research and Applications (pp. 166–185). Taylor and Francis Inc. https://doi.org/10.4324/9780203105894
  45. 45. Norman D. A., & Bobrow D. G. (1975). On data-limited and resource-limited processes. Cognitive Psychology, 7, 44–64. https://doi.org/10.1016/0010-0285(75)90004-3
  46. 46. Frese M., Krauss S. I., Keith N., Escher S., Grabarkiewicz R., Luneng S. T., et al. (2007). Business owners’ action planning and its relationship to business success in three African countries. Journal of Applied Psychology, 92, 1481–1498. pmid:18020791
  47. 47. Cooper H. (2003). Psychological Bulletin: Editorial. Psychological Bulletin, 129, 3–9. https://psycnet.apa.org/doi/10.1037/0033-2909.129.1.3
  48. 48. Haith M. M. (1997). The development of future thinking as essential for the emergence of skill in planning. In Friedman S. L. & Kofsky Scholnick E. (Eds.), The Developmental Psychology of Planning: Why, How, and When do We Plan (pp. 25–42). Psychology Press.
  49. 49. Maynard M. T., Mathieu J. E., Rapp T. L., & Gilson L. L. (2012). Something(s) old and something(s) new: Modeling drivers of global virtual team effectiveness. Journal of Organizational Behavior, 33, 342–365. https://doi.org/10.1002/job.1772
  50. 50. Thürmer J. L., Wieber F., & Gollwitzer P. M. (2017). Planning and performance in small groups: Collective implementation intentions enhance group goal striving. Frontiers in Psychology, 8, 603. pmid:28469592
  51. 51. Landis J. R., & Koch G. G. (1977). The measurement of observer agreement for categorical data. Biometrics, 33, 159–174. pmid:843571
  52. 52. Faul F., Erdfelder E., Lang A. G., & Buchner A. (2007). G* Power 3: A flexible statistical power analysis program for the social, behavioral, and biomedical sciences. Behavior Research Methods, 39, 175–191. pmid:17695343
  53. 53. Muthén L. K., & Muthén B. O. (1998–2017). Mplus User’s Guide. Eighth Edition. Muthén & Muthén.
  54. 54. Cheung M. W. L. (2015). Meta-Analysis: A Structural Equation Modeling Approach. John Wiley & Sons. https://doi.org/10.1037/a0032968
  55. 55. Casella G., & Berger R. L. (2002). Statistical Inference. Duxbury.
  56. 56. Antonakis J., Bastardoz N., & Rönkkö M. (2019). On ignoring the random effects assumption in multilevel models: Review, critique, and recommendations. Organizational Research Methods. Advance online publication. https://doi.org/10.1177/1094428119877457
  57. 57. Turner R., & Higgins J. (2019). Bayesian meta-analysis. In Cooper H., Hedges L. &, Valentine J. (Eds.), The Handbook of Research Synthesis and Meta-Analysis (pp. 299–314). Russell Sage Foundation.
  58. 58. Hox J. J., van de Schoot R., & Matthijsse S. (2012). How few countries will do? Comparative survey analysis from a Bayesian perspective. Survey Research Methods, 6, 87–93. https://doi.org/10.18148/srm/2012.v6i2.5033
  59. 59. Muthén B., & Asparouhov T. (2012). Bayesian structural equation modeling: A more flexible representation of substantive theory. Psychological Methods, 17, 313–335. pmid:22962886
  60. 60. Spiegelhalter D. J., Best N. G., Carlin B. P., & Van Der Linde A. (2002). Bayesian measures of model complexity and fit. Journal of the Royal Statistical Society: Series b (statistical methodology), 64, 583–639. http://dx.doi.org/10.1111/1467-9868.00353
  61. 61. Gelman A. and Rubin D. B. (1992). Inference from iterative simulation using multiple sequences. Statistical Science, 7, 457–472. http://dx.doi.org/10.1214/ss/1177011136
  62. 62. Assink M., & Wibbelink C. J. M. (2016). Fitting three-level meta-analytic models in R: A step-by-step tutorial. The Quantitative Methods for Psychology, 12, 154–74. https://doi.org/10.20982/tqmp.12.3.p154
  63. 63. Bosco F. A., Aguinis H., Singh K., Field J. G., & Pierce C. A. (2015). Correlational effect size benchmarks. Journal of Applied Psychology, 100, 431–449. pmid:25314367
  64. 64. Borenstein M., Higgins J. P., Hedges L. V., & Rothstein H. R. (2017). Basics of meta‐analysis: I2 is not an absolute measure of heterogeneity. Research Synthesis Methods, 8, 5–18. https://doi.org/10.1002/jrsm.1230
  65. 65. Gonzalez-Mulé E., & Aguinis H. (2018). Advancing theory by assessing boundary conditions with metaregression: A critical review and best-practice recommendations. Journal of Management, 44, 2246–2273. https://doi.org/10.1177/0149206317710723
  66. 66. Steel P. D., & Kammeyer-Mueller J. D. (2002). Comparing meta-analytic moderator estimation techniques under realistic conditions. Journal of Applied Psychology, 87, 96–111. pmid:11916220
  67. 67. Raudenbush S. W. (1994). Random effects models. In Cooper H. & Hedges L. V. (Eds.), Handbook of Research Synthesis (pp. 301–321). Sage.
  68. 68. Hambleton R. K., & Jones R. W. (1994). Item parameter estimation errors and their influence on test information functions. Applied Measurement in Education, 7, 171–186.
  69. 69. Bijmolt T. H., Van Heerde H. J., & Pieters R. G. (2005). New empirical generalizations on the determinants of price elasticity. Journal of Marketing Research, 42, 141–156. https://doi.org/10.1509%2Fjmkr.42.2.141.62296
  70. 70. Collins L. M., & Lanza S. T. (2010). Latent Class and Latent Transition Analysis: With Applications in the Social, Behavioral, and Health Sciences. John Wiley & Sons.
  71. 71. Marsh H. W., Muthén B., Asparouhov T., Lüdtke O., Robitzsch A., Morin A. J. S., et al. (2009). Exploratory structural equation modeling, integrating CFA and EFA: Application to students’ evaluations of university teaching. Structural Equation Modeling, 16, 439–476. http://dx.doi.org/10.1080/10705510903008220
  72. 72. Bauer D. J., & Curran P. J. (2003). Overextraction of latent trajectory classes: Much ado about nothing? Reply to Rindskopf (2003), Muthén (2003), and Cudeck and Henly (2003). Psychological Methods, 8, 384–393.https://doi.org/10.1037/1082-989X.8.3.384
  73. 73. Gibbert M., Nair L. B., Weiss M., & Hoegl M. (2020). Using outliers for theory building. Organizational Research Methods, 1–10. https://doi.org/10.1177/1094428119898877
  74. 74. Aguinis H., Gottfredson R. K., & Joo H. (2013). Best-practice recommendations for defining, identifying, and handling outliers. Organizational Research Methods, 16, 270–301. https://doi.org/10.1177/1094428112470848
  75. 75. Aitkin M., & Wilson G. (1980). Mixture models, outliers and the EM algorithm. Technometrics, 22, 325–331. https://doi.org/10.2307/1268316
  76. 76. Jorgensen M. (1990). Influence-based diagnostics for finite mixture models. Biometrics, 46, 1047–1058. https://doi.org/10.2307/2532447
  77. 77. Van Erp S., Mulder J., & Oberski D. L. (2018). Prior sensitivity analysis in default Bayesian structural equation modeling. Psychological Methods, 23, 363–388. pmid:29172613
  78. 78. Coyne J. C., Thombs B. D., & Hagedoorn M. (2010). Ain’t necessarily so: Review and critique of recent meta-analyses of behavioral medicine interventions in health psychology. Health Psychology, 29, 107–116. pmid:20230082
  79. 79. Schmidt F. L., & Hunter J. E. (2015). Methods of Meta-Analysis. Correcting Error and Bias in Research Findings. 3rd Edn. Sage.
  80. 80. Chen Q., Luo W., Palardy G. J., Glaman R., & McEnturff A. (2017). The efficacy of common fit indices for enumerating classes in growth mixture models when nested data structure is ignored: A Monte Carlo study. Sage Open, 7, 1–19. https://doi.org/10.1177/2158244017700459
  81. 81. Otte K.-P., Konradt U., & Oldeweme M. (2018). Effective team reflection: The role of quality and quantity. Small Group Research, 49, 739–766. https://doi.org/10.1177/1046496418804898
  82. 82. Santos C. M., Uitdewilligen S., & Passos A. M. (2015). A temporal common ground for learning: The moderating effect of shared mental models on the relation between team learning behaviours and performance improvement. European Journal of Work and Organizational Psychology, 24, 710–725.
  83. 83. Saavedra R., Earley P. C., & Van Dyne L. (1993). Complex interdependence in task-performing groups. Journal of Applied Psychology, 78, 61–72.
  84. 84. Christley R. M. (2010). Power and error: increased risk of false positive results in underpowered studies. The Open Epidemiology Journal, 3, 16–19.
  85. 85. Lipsey M. W. (2003). Those confounded moderators in meta-analysis: Good, bad, and ugly. The Annals of the American Academy of Political and Social Science, 587, 69–81. https://doi.org/10.1177%2F0002716202250791
  86. 86. Mohammed S., Hamilton K., & Lim A. (2009). The incorporation of time in team research: Past, current, and future. In Salas E., Goodwin G. F., & Burke C. S. (Eds.), Team Effectiveness in Complex Organizations: Cross-Disciplinary Perspectives and Approaches (Society for Industrial and Organizational Psychology Frontier Series). Erlbaum. https://doi.org/10.4324/9780203889312
  87. 87. Woolley A. W. (2009). Means vs. ends: Implications of process and outcome focus for team adaptation and performance. Organization Science, 20, 500–515. a. http://dx.doi.org/10.1287/orsc.1080.038
  88. 88. Aguinis H., & Bakker R. M. (2020). Time is of the essence: Improving the conceptualization and measurement of time. Human Resource Management Review, Advance online publication. https://doi.org/10.1016/j.hrmr.2020.100763
  89. 89. Oldeweme M., Konradt U., & Brede M. (2021). The rhythm of teamwork: Discovering a complex temporal pattern of team processes. Group Dynamics: Theory, Research, and Practice. Advance online publication. https://doi.org/10.1037/gdn0000175
  90. 90. Howard M. C., & Hoffman M. E. (2018). Variable-centered, person-centered, and person-specific approaches: Where theory meets the method. Organizational Research Methods, 21, 846–876. https://doi.org/10.1177/1094428117744021
  91. 91. Collins L. M. (2006). Analysis of longitudinal data: The integration of theoretical model, temporal design, and statistical model. Annual Review of Psychology, 57, 505–528. pmid:16318605
  92. 92. Konradt U., Otte K.-P., Schippers M. C., & Steenfatt C. (2016). Reflexivity in teams: A review and new perspectives. The Journal of Psychology: Interdisciplinary and Applied, 150(2), 153–174. https://doi.org/10.1080/00223980.2015.1050977
  93. 93. Kingston G. (2001). Cost benefit analysis in theory and practice. Australian Economic Review, 34, 478–478. https://doi.org/10.1111/1467-8462.00217
  94. 94. Ledgerwood A. (2019). New developments in research methods. In Finkel E. J. & Baumeister R. F. (Eds.), Advanced Social Psychology, Second Edition. Oxford University Press.
  95. 95. DeShon R. P., Kozlowski S. W. J., Schmidt A. M., Milner K. R., & Wiechmann D. (2004). A multilevel, multiple goal model of feedback effects on the regulation of individual and team performance. Journal of Applied Psychology, 89, 1035–1056. https://psycnet.apa.org/doi/10.1037/0021-9010.89.6.1035
  96. 96. Dirks K. T. (1999). The effects of interpersonal trust on work group performance. Journal of Applied Psychology, 84, 445–455. pmid:10380424
  97. 97. Earley C. P., & Mosakowski E. (2000). Creating hybrid team cultures: An empirical test of transnational team functioning. Academy of Management Journal, 43, 26–49. https://doi.org/10.5465/1556384
  98. 98. Gevers J. M., van Eerde W., & Rutte C. G. (2001). Time pressure, potency, and progress in project groups. European Journal of Work and Organizational Psychology, 10, 205–221. https://doi.org/10.1080/13594320143000636
  99. 99. Jehn K. A., & Shah P. P. (1997). Interpersonal relationships and task performance: An examination of mediation processes in friendship and acquaintance groups. Journal of Personality and Social Psychology, 72, 775–790. https://doi.org/10.1037/0022-3514.72.4.775
  100. 100. Kauffeld S., & Lehmann-Willenbrock N. (2012). Meetings matter. Effects of team meetings on team and organizational success. Small Group Research, 43, 130–158. https://doi.org/10.1177/1046496411429599
  101. 101. Larson N. L., McLarnon M. J. W., & O’Neill T. A. (2020). Challenging the “static” quo: Trajectories of engagement in team processes toward a deadline. Journal of Applied Psychology, 105(10), 1145–1163. pmid:31985237
  102. 102. Lin C. P., Liu M. L., Joe S. W., & Tsai Y. H. (2017). Predicting top management approval and team performance in technology industry. Personnel Review. 46, 46–67. http://dx.doi.org/10.1108/PR-01-2015-0007
  103. 103. Mathieu J. E., & Rapp T. L. (2009). Laying the foundation for successful team performance trajectories: The roles of team charters and performance strategies. Journal of Applied Psychology, 94, 90–103. pmid:19186898
  104. 104. Mathieu J. E., & Schulze W. (2006). The influence of team knowledge and formal plans on episodic team process-performance relationships. Academy of Management Journal, 49, 605–619. https://doi.org/10.5465/amj.2006.21794678
  105. 105. Müller A. (2009). Zielklärung und Handlungsplanung beim Problemlösen in Gruppen [Goal-clarification and action-planning during problem-solving in groups]. Psychologie des Alltagshandelns, 2, 2–13.
  106. 106. Oldeweme M., & Konradt U. (2020). Team Planning: Development and Validation of Long-Form, Short-Form and Single Item Scales. [Manuscript submitted for publication]. Institute of Psychology, Kiel University.
  107. 107. Pearsall M. J., & Venkataramani V. (2015). Overcoming asymmetric goals in teams: The interactive roles of team learning orientation and team identification. Journal of Applied Psychology, 100(3), 735–748. pmid:25384202
  108. 108. Rapp A., Ahearne M., Mathieu J., & Rapp T. (2010). Managing sales teams in a virtual environment. International Journal of Research in Marketing, 27, 213–224. http://dx.doi.org/10.1016/j.ijresmar.2010.02.003
  109. 109. Siegel Christian J., Pearsall M. J., Christian M. S., & Ellis A. P. (2014). Exploring the benefits and boundaries of transactive memory systems in adapting to team member loss. Group Dynamics: Theory, Research, and Practice, 18, 69–86. http://dx.doi.org/10.1037/a0035161
  110. 110. Simons T., Pelled L. H., & Smith K. A. (1999). Making use of difference: Diversity, debate, and decision comprehensiveness in top management teams. Academy of Management Journal, 42, 662–673. http://dx.doi.org/10.5465/256987
  111. 111. Tasa K., & Whyte G. (2005). Collective efficacy and vigilant problem solving in group decision making: A non-linear model. Organizational Behavior and Human Decision Processes, 96, 119–129. https://doi.org/10.1016/j.obhdp.2005.01.002
  112. 112. Van der Kleij R., Lijkwan J. T., Rasker P. C., & De Dreu C. K. (2009). Effects of time pressure and communication environment on team processes and outcomes in dyadic planning. International Journal of Human-Computer Studies, 67, 411–423. https://doi.org/10.1016/j.ijhcs.2008.11.005
  113. 113. Weldon E., Jehn K. A., & Pradhan P. (1991). Processes that mediate the relationship between a group goal and improved group performance. Journal of Personality and Social Psychology, 61, 555–569. pmid:1960648