Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

The initial development and validation of the Social Adaptability Skills Questionnaire: SASQ

Abstract

Changing clubs over the course of an athletic career may not always be easy, and this has raised questions about how these changes affect career development. However, few studies have focused on the process of adapting to a new club and the factors that lead to success or failure. To address this gap in the literature, we aimed to develop and provide the initial validation of a questionnaire designed to assess athletes’ social adaptability skills (SAS). To do so, we conducted four studies, from the initial development stage to the final validation stage. In the first phase, we generated questionnaire items with clear content and face validity. The second phase explored the factor structure and reliability of the Social Adaptability Skills Questionnaire (SASQ). This was carried out with 543 young athletes in talent development through exploratory factor analysis (EFA), which was validated with confirmatory factor analysis (CFA). The EFA yielded a 17-item, four-factor structure with good internal reliability (⍺ = 0.876). The CFA revealed that the model fit indices were acceptable (RMSEA = 0.06, CFI = 0.809, TLI = 0.844, and GFI = 0.926). In addition, Latent Class Analysis (LCA) was applied to determine the predictive validity of SASQ resulting into identification of three classes (low achievers, average achievers, and high achievers) with four discriminating dimensions (coach, teammates, family, and club). The SASQ appears to be a promising psychometric instrument of potential usefulness for education and program reviews in applied settings and a measurement tool in talent development research.

Introduction

It is unusual today for athletes in team sports to spend their entire careers in the same club. In 2019, the CIES Observatory [1] reported that the number of international transfers in basketball alone rose 34% between the 2010/2011 and 2018/2019 seasons. The effects of these club-to-club moves on players’ equilibrium and their sporting lives are not well known, although recurring individual stories (e.g., Mercato transfers in soccer, NBA basketball drafts, and major hockey league transfers) and a recent study have underlined distinct periods of adaptation to the new environment [2]. Although professional athletes may show adequate resources for adaptation as they move from club to club over the course of their careers, this may not be the case for talented young athletes preparing to enter the professional world, as they may lack some of the basic skills and psychosocial resources (example: positive thinking, goal setting skills, self-confidence skills) that are needed to ensure successful moves [2]. However, in talent development research programs [e.g., 35] and especially those focused on team sports − where club changes are usual [2] − this issue has not been taken into account. We thus sought to develop a valid instrument to assess the psychosocial skills of young players that will be useful for future positive moves. We assumed that this instrument could also be used to design individualized programs to develop these skills in training centres for professional team sports.

The notion that young talented players also need to be equipped with psychosocial skills before they enter the professional world has thus been identified as a major concern. For example, Hauw and colleagues e [6] used the MacAdams identity multilayer model and identified three types of resources available to players during their transitions: (i) level 1: dispositional traits, which are general dispositions that can facilitate adaptations during club-to-club changes; (ii) level 2: personal action constructs (PACs), which draw on skills like personal strivings; and (iii) level 3: narratives, which are the life stories constituted from experience to give meaning to oneself. Regarding club changes, level 1 refers to the more or less high level of the personality disposition to new social environments (openness to experience, agreeableness, emotional stability, extraversion, and conscientiousness) [6]. The development of this level may require deep psychological work with young team athletes. Level 2 of the MacAdams model thus may be the right target for preparing players for future club changes. Level 3 refers to the life stories around the experience of club changes and thus might be used by sports psychologists when working with new professional athletes arriving in a club after experiencing other changes in their careers.

In the development of talented young athletes, studies have identified general competences such as self-regulation, goal setting, performance assessment, resilience, grit, and self-awareness as important for success [79]. MacNamara and colleagues drew up a list of ten key skills, which they referred to as the psychological characteristics of developing excellence (PCDEs), with positive or adaptive characteristics such as goal setting, self-organization, commitment, imagery use, self-awareness and others [10]. However, all these psychosocial skills are “situated resources”, meaning they are attuned to specific and typical situations like training, competing, or relationships with coach or parents [11]. The need to identify the specific skills for successful club changes thus emerged.

Acculturation and integration studies indicate that most youths in host countries and cities face challenges during transitions, and this has prompted calls to facilitate their adaptation and ensure their full inclusion [12]. For example, these youths may need to learn the customs and language of the receiving country so that they can communicate and socialize to achieve full social inclusion [13]. Social skills are not only key factors in achieving inclusion but also a concern on the sports field, especially with young players. Therefore, instruments like the Social Skills Questionnaire (CHASO III) [13] and the Social Skills Questionnaire for College Students (SSQ-U) [14] have been used to assess these skills. However, these questionnaires are general or specific to college and not well situated for sports, especially club changes. For athletes, some of the elements of the situation that provide them affordances in the sporting environment are coaches and teammates’ specific behaviours, as reported in several studies [2, 15, 16]. We therefore saw a need for a specific instrument that is situated, short, and quickly and inexpensively applied to measure the skills deployed during club change.

Research to date has reported that athletes’ psychosocial skills influence their development in a variety of ways, including supporting (e.g., athletes’ stability and harmony), constraining (e.g., undermining the well-being of players), accelerating (i.e., increasing performance success and satisfaction), or modifying how they handle problematic situations [2]. As part of their investigation into this problem, Owiti and Hauw [2] identified a range of psychosocial factors, known as social adaptability skills (SAS), that contributed to successful outcomes. These SAS structures were organized into four factors (i.e., coach, teammates, family/friends, and club). The sub-themes in relation to the coach were: obeying orders, language barrier, inequality/favouritism, reduced play time, difficulty combining sports and other life domains, pressure to perform, lack of structured training, and low-level competition. Sub-themes related to teammates included: respect, language barrier, and negative perception of others. This was followed by sub-themes related to family/friends, which included: climate change, distance, language and culture shock barrier. Last, club included the following sub-themes: poor facilities and logistics and lack of support. These results indicate the events of critical situations that can be encountered during club changes. Aspiring athletes must be prepared to deal with these situations in order to maximize their development opportunities (e.g., adapting during the club change), endure setbacks (e.g., adapting to coaching styles), and navigating key transitions (e.g., living away from family and friends) [2].

It is generally expected that young athletes possess the required psychosocial skills for sports participation and the pursuit of excellence. Yet there is a need for talent development programs to also consider the skills necessary for changing clubs and adapting to new environments. Having identified the key skills needed to adapt to different situations during club to club change, the next step is to develop a psychometrically sound measure of these skills with an ultimate goal of designing and testing a learning program for these athletes. The aim of the current research was thus to develop and validate a questionnaire to assess the level of SAS possessed by young athletes. Such an instrument is needed to provide regular feedback and refinement and to identify areas that require immediate attention, improvement, or maintenance. As a result of these findings, training programs could be formulated to improve athletes’ SAS performance by identifying the challenges they face and including specific training to help enhance those weaknesses. Finally, the validation of this questionnaire would contribute significantly to the international research community focussed with talent development [2, 15].

To summarize, our aim was to outline the initial development and validation of a questionnaire designed to assess the SAS of young athletes. We thus designed four studies from the initial development stage up to the final validation stage.

Study 1

The purpose of this first study was to generate the questionnaire items for the development of a psychometric measure of SAS that would be useful to athletes during club changes. Similarly to previous studies, the initial items were developed in accordance with the recommendations for developing new measurement scales [e.g., 17, 18].

Method

The items were developed in several stages and elite basketball players were recruited for the first stage since this population changes clubs frequently over the course of their career progression [1].

Participants.

After receiving approval from the Research Ethics Commission of the University of Lausanne (Project number: E_SSP_122021_00002) the participants gave their written informed consent. The study involved a convenience sample of 20 European and American professional basketball players (age range: 20–36 years; mean: 26.05, SD: 4.12). Seventeen were male and three were female. In total, 12 players had attended European basketball academies in their respective countries and had performed at the highest national levels (e.g., Swiss Basketball League and National Basketball League of France). The eight others had gone through North American academies (e.g., in the USA & Canada) playing in one of the highest leagues (e.g., the National Basketball Association G League and Women’s National Basketball Association).

Data collection.

Individual interviews were conducted using a semi-structured interview guide. We sought to help the players recall their personal experiences during club changes. As part of the interview, we worked first on creating a timeline of the changes, which required us to break down the flow of time into periods to keep a semiotic trace of the adaptability experience.

The interview guide incorporated all of the changes from club to club to facilitate our in-depth understanding of the dynamics of personal experience within the context of each club’s environment. All interviews were recorded digitally in their entirety and lasted from 30 to 90 minutes.

Data analysis.

Of the 20 participant interviews, we chose only episodes that displayed the problematic situations faced by the players during club changes. Identifying a problematic situation involved looking for disturbances in the transition that prevented players from performing at their best [2]. Data were analysed qualitatively using inductive content analysis. A constant comparative method was employed, which involved the development of additional codes, identification of emerging themes within the data, and continuous comparison of the codes (e.g.,, between participants or between time periods). In accordance with the criterion of data saturation [1922], data collection and analysis were discontinued when the categories upon which the theory was built no longer producing new insights.

Results.

Four major themes consistently emerged from this analysis and comprised problematic meaningful experiences with (i) the coach, (ii) the teammates, (iii) being away from family and friends, and (iv) the club.

We therefore based our initial item generation on the four emerging themes and this resulted in a preliminary list of 55 items representing the 15 SAS sub-themes earlier mentioned.

Item justification.

Expert panel. An independent panel of six experts (PhD and postdoc students in sports psychology) reviewed the initial list of 55 items [4, 23, 24]. In addition to their expertise in psychological research, applied talent development, and coaching, all of the experts were familiar with the SASQ’s aims and rationale. Each expert rated the content relevance and representativeness of each item on a scale from 1: not at all relevant to 5: completely relevant. Items that were rated 4: relevant or less were discussed by the whole panel [4]. At the end of this stage, some of these items were marked for deletion. A first panel review resulted in the rewording of several items (N = 41) due to grammatical errors and comprehension issues. A total of ten items were deleted and five additional items were added. Furthermore, the five items that were added were part of the interviews and not from the expert panel. A second expert panel was constituted and asked to follow the same procedure, and this resulted in the rewording of several items (N = 25) and deletion of several others (N = 14). As a result of this stage, 36 items were included in the SASQ.

Cognitive interviews.

During the third stage, cognitive interviews were conducted to check for misunderstandings, unclear questions, inconsistencies, and inappropriate options [4, 25]. A written informed consent was obtained from all participants (and parents/guardians when participants were under 16 years of age). Six young athletes in development were interviewed individually, reflecting the intended target population of the questionnaire (there were two participants in each age group: 12–14 years, 15–17 years, and 18–21 years). Two of the participants competed at a national level, whereas the remaining four competed at a regional level in their respective sport (football N = 2; basketball N = 2; handball N = 1, and ice hockey N = 1).

In each item, comments were directly entered under the question to summarize the findings. A complete review of the comments was conducted using Conrad and Blair’s [25] classification system. This process resulted in 34 items, of which two items were omitted due to lexical problems [26].

Pilot test.

The fourth stage was a pilot test conducted using the 34-item version of the SASQ to examine the comprehensibility of the questionnaire and the ease of overall administration [4, 27, 28].

Participants.

The 34-item list was pilot-tested with a sample of athletes (N = 21, 12 males, 9 females; age range: 12–21 years) competing in different sports (basketball, N = 9), (football, N = 6) (rugby, N = 3), and (handball, N = 3). Each athlete competed at a representative level (e.g., regional and national levels) in their respective sports. In order to ensure that the SASQ was suited to the range of sports and age groups, a purposefully stratified sample was selected.

Procedure.

After receiving approval from the Research Ethics Commission of the University of Lausanne (Project number: E_SSP_122021_00002) and informed written consent from all participants (and parents/guardians when participants were under 16 years of age). The pilot version of the SASQ contained 34 items using a 6-point Likert scale with a similarity response format from 1: very unlike me to 6: very like me as suggested by Lei Chang [4, 29]. The 34-item list contained a mixture of negatively (N = 13) and positively (N = 21) worded questions to minimize the danger of acquiescence bias [4]. The pilot version took between 12 and 15 minutes to complete. After the item analysis, participants were encouraged to write comments next to any problem items. Debriefing was conducted after each item on the questionnaire was completed and participants were encouraged to provide feedback.

Data analysis and results.

Following the pilot test data analysis, the SASQ still contained 34 items, ensuring that at least two items per sub-theme were retained [30]. As a result of the first stage, the elite athletes played different roles, which contributed to the retention of all items, since they suggested themes focusing on their challenges during club to club changes. Young athletes were able to identify incomprehensible or difficult questions through the pilot test study. A revised questionnaire structure was sent to two members of the expert panel with the request that they comment on the face validity, content validity, comprehensibility, and comprehensiveness of the SASQ in light of the proposed target population [4]. This was done to ensure that the 34 items adequately covered the questionnaire constructs. The questionnaire did not require any further modifications after this stage.

Study 2

In study 2, we performed exploratory factor analysis (EFA) to examine the possibility of reducing the large number of SASQ variables into manageable size while retaining as much of the original information as possible [31, 32].

Method

Sample size.

The Kaiser-Meyer-Olkin (KMO) test of sampling adequacy and the Bartlett sphericity test were used to assess the quality of the correlations in order to proceed (or not) with the EFA [33, 34].

Participants.

A total of 543 participants (mean age: 16.67, SD: 2.42, range: 12–21 years) completed the questionnaire (see Table 1). There is no consensus regarding sample size calculation using either EFA or CFA [35]. Following considerable discussion within the literature, the current study followed the recommendation of DeVellis [35] indicating participant item ratio of 10:1. Therefore, since SASQ contained 34 items, we required between 340–510 to fulfil the empirical rule.

Participants were purposefully sampled in line with their characteristics and ability to adapt in the environmental sporting conditions for which the SASQ was designed. By virtue of their commitment and selection to formally established development environments (e.g., regional or national, and in academy or not in academy groups), they had been recruited as junior athletes with the potential to become senior elites.

Procedure.

Approval was received from the Research Ethics Commission of the University of Lausanne (Project number: E_SSP_122021_00002) and written informed consent was obtained from all participants. Self-administered paper-pencil SASQ questionnaires were distributed to the 543 participants at their respective clubs between March and May 2022. As participation was anonymous, the athletes were given coded names based on their initials and age and their parents’ initials (mother then father).

Data analysis.

The Statistical Package for Social Sciences (SPSS version 27) was used to examine the factor structure of the SASQ with statistical significance set at p < 0.05. Prior to the analysis, scores of negatively worded questions were reversed. Descriptive statistics were performed on the 34 items with means, standard deviation, measures of kurtosis and skewness, and the analysis of missing values. Participants could decide not to answer certain questions. For the 34 items, 0.24% of the values were missing. Normally, it is considered inconsequential when the missing data percentage is below 5% [36]. Therefore, we used an exclude cases listwise method for the descriptive statistics.

In addition to allowing significant items to be retained and interpreted, the emerging factor structure provided insight into the latent factors underpinning the SASQ. The specific approach utilized principal axis factoring (PAF) extraction. An oblique with direct oblimin rotation was selected to improve the interpretation of the data since the factors were likely to be correlated.

The criteria used to determine the number of factors to be extracted included the scree plot [30, 37]. Kaiser’s criterion, which advocates for retention of factors with an eigenvalue (EV) ≥ 1.0 was applied since low unit values reflect instability in the factor [38]. Further consideration was given to the communalities and factor loadings [34, 37, 39].

Results.

We followed a combination of approaches that preserve the content validity and psychometric properties of a questionnaire [34, 3740], and our results indicated that items with high factor loadings were maintained, with the expert views also taken into consideration. It should be noted that we first ran an EFA on the 34 items (S1 Raw data), and due to the exploratory nature of the analysis, a total of 12 items were deleted at this stage, leaving the SASQ with only 22 items. Nine of the items were dropped due to low communalities, below 0.3 (items Q1, Q2, Q4, Q6, Q14, Q15, Q17, Q19, and Q21), and an additional three items (Q20, Q24, Q27) were deleted after in-depth discussions. Therefore, the following results are based on the remaining 22 items (Table 2).

The response distribution in Table 2 indicates that the respondents scored high points on all the items (i.e., very like me rather than very unlike me) with the lowest score for item 5 (mean: 4.26, SD: 1.38) and the highest score for item 8 (mean: 5.73, SD: 0.97). The skewness and kurtosis coefficients were also examined. Items 11 and 12 did not follow the normal distribution and as such, a Box-cox normality plot was applied to find a transformation that approximately normalised the data. This process was then verified by computing the correlation coefficient of a normal probability plot.

The KMO static value was 0.910, which is above the minimum criterion of 0.6. Also, all the KMO values for individual items were greater than 0.77. The Bartlett’s test of sphericity had a Chi-square approximation of 4451.76 and was significant at p < 0.001. Based on the KMO and Bartlett’s sphericity tests, the EFA method was suitable for use in the current research [30, 38, 39].

Studies have suggested that after performing the extraction, the communalities must be greater than 0.30 for good validity to be assumed [30, 38, 39, 41]. The SASQ met most of this criterion, except for items 1 (0.277), 5 (0.261), 10 (0.254), 16 (0.256), and 22 (0.211), which were then eliminated, leaving 17 items. The scree plot revealed that four factors should be retained, consistent with the previous findings [2]. Further, the 22 initial items were grouped into four factors and rotated factor loads varied between 0.310 and 0.843 (Table 3). The four factors explained 64.83% of the variance in the 22-item version.

thumbnail
Table 3. Grouping of the 22 items into four factors with rotated factor loads.

https://doi.org/10.1371/journal.pone.0281971.t003

To determine whether the final four-factor solution was adequate, the determinants of the 22-item correlation matrix and reproduced matrices were used to evaluate model fit [42]. The reproduced correlations were also close to the original correlations, and values in the residual correlation matrix were small, indicating good model fit (Table 4).

Reliability analysis

Internal consistency.

The general reliability static was α = 0.876, indicating good reliability [43, 44]. The correlations between each item and the total questionnaire score were reliable (above 0.3) except for items 5 (0.282) and 22 (0.279), which were finally dropped. The Cronbach alpha values, if deleted, were also calculated and were expected not to exceed 0.876 [31]. This result re-confirmed the deletion of item 5, which had a reliability of 0.877. To ensure that the factor structure had not been affected by deleting item 5, the factor analysis was run again.

Study 3

The four-factor, 17-item model identified in the EFA was examined using confirmatory factor analysis (CFA), with the intention to configure the final instrument (Fig 1). As a result of this second-order analysis, the hierarchical structure of the instrument and the relationships between variables in the EFA were validated [30, 45]. In addition, acknowledging that predictive validity is one of the most important forms of validity, we similarly evaluated the association between SASQ items using Latent Class Analysis (LCA) method. The combination of these two methodologies constituted an exploratory sequential mixed methodology design approach to integrate social adaptability skills during athletes change from club to club. Similar mixed methodology has been previously applied with success in determining the internal and predictive validity of psychometric instruments [46].

Confirmatory factor analysis (CFA)

Participants.

The participants and data set in study 2 were again used to test the factor structure of the SASQ. The CFA thus used a total of 543 participants (mean age: 16.67, SD: 2.42, range: 12–21 years). We followed the recommendations by DeVellis [35] of participant item ratio of 10:1 in estimating the sample size.

Data analysis.

The analysis was conducted using AMOS version 27. The CFA is presented in route diagrams where the circles represent latent variables and the squares represent the observed variables [4345]. Two-headed arrows indicate covariance between the four latent variables, and single-headed arrows indicate the assumed direction of influence [47]. Within the literature, the most used approximation indices include [34, 48, 49]; (a) the absolute indices referring to the degree to which the covariances implied in the model match the observed covariances through which the free parameters are estimated, and (b) incremental indices which refers to the degree to which the model is superior to the alternative model, usually where no covariances between variables are specified- null or independent model.

Despite reporting both the absolute and incremental indices in the current study, the CFA model was validated using the absolute indices (i.e., normalized Chi-Square- X2/df and Root Mean Squared Error of Approximation-RMSEA) since an optimal fit is indicated by values close to zero. Normalized Chi-Square (X2) values < 3.0 indicate reasonable adjustment [50, 51] with a value of 5.0 being the minimum acceptable [52]. As concerns RMSEA, values of ≤ 0.06 indicate an adequacy of the model [5256], but normally the most used cut-off values are [42, 49, 53]; ≤ 0.05 good fit, ≤ 0.08 acceptable fit, ≤ 0.10 indicate a mediocre fit, and > 0.10 a poor (un acceptable fit). Factor weights (FL; factor loading) were also computed with values equal to or greater (≥ 0.5) being accepted [30]. Ideally, according to Hair [38] accepted values should be over 0.7.

Results.

The Chi-square value (X2) was 327.96 (p = 0.000) (Table 5). However, it has been argued that X2 static value shows high sensitivity to larger sample sizes like ours [47]. Therefore, a normalized Chi-Square (X2/df) value was reported since it reduced the test sensitivity to the sample size and model complexity. In the current study, the RMSEA value was 90% CI < 0.06, indicating model adequacy.

As concerns the incremental fit, the comparative fit index (CFI), the Tucker-Lewis index (TLI), and the goodness-of-fit index (GFI) values indicated a good fit to the model (CFI = 0.809, TLI = 0.844, and GFI = 0.926, respectively). The CFI, TLI, and GFI have a range of 0 to 1, with values indicating greater validity as they approach 1 [30].

Last, the latent variables for the four subscales of the SAS had nine, three, three, and two items each (Fig 1). According to the route diagram, all items show acceptable factor loads (standardized values) ranging from 0.66 to 2.74, except for teammates factor (Q12R) which had a value of 0.31. Nine elements of the coach construct had factorial loads between 1.00 and 2.74, as did the three elements of the teammates construct, which had factorial loads between 0.31 and 1.00. The three elements of the family construct ranged from 0.95 to 1.13, and the last two elements of the club construct ranged from 0.66 to 1.00.

Latent class analysis

Participants and data analysis.

The same data of 543 participants used in CFA was reused to classify the Social Adaptability Skills (SAS) into meaningful groups using latent class analysis (LCA) [57]. According to Nylund, Asparouhov and Muthen [58], LCA models are used to, “uncover unobserved heterogeneity in a population and to find substantively meaningful groups of people that are similar in their responses to measured variables” (p. 536). The current study settled on LCA because it transcends other conventional parametric analyses by relaxing the assumptions of homogeneity of variance, skewness, normal distribution, and linearity that cannot be analysed using the same conventional parametric clustering techniques [59]. The analysis was performed on Latent Gold® 6.0 [59] with the maximum likelihood method for parameter estimation. The LCA in the present study comprised: (a) building and classification of a cluster model (b) bootstrapping to confirm optimum fit model, and (c) predicting the precision of the optimal model using classification statistics.

The optimal model was determined by comparing the fit statistics of three models [60] namely: (i) Akaike Information Criterion (AIC): a measure of relative fit or quality of the model of the data, which considers model parsimony (the number of parameters). AIC is estimated based on log-likelihood squared (L2) and LL. In either case, lower AIC indices indicate a better fit [61], (ii) Consistent AIC (CAIC) based on LL, and (iii) Bayesian Information Criterion (BIC) based on LL: A better fit is indicated by low BIC indices when comparing different models. Furthermore, BIC imposes a stronger penalty on the number of parameters than AIC. Therefore, when AIC and other fit statistics differ, the former is usually prioritized when assessing model fit [61].

It should be noted that, during the analysis, the competing models were nested with the optimal model in order to preserve the same number of latent classes and variables on each model. An examination of the statistical significance (p value) of the conditional bootstrapping analysis determined the amount of improvement in fit. If the p value generated in conditional bootstrapping is statistically significant (p < 0.5), then the competing model provides a statistically significant improvement in model fit, even though it might compromise inferior AIC, BIC, and CAIC [62, 63]. Finally, in estimating the effect size, an entropy R2 was computed using the sample size, posterior probabilities, and a number of classes [63]. According to Kaplan and Keller [64], high entropy R2 values indicate high precision in classification and the prediction of the emerging latent classes.

Results

Latent class analysis.

Fig 2 illustrates conditional probabilities of the three classes on a profile graph. On the graph, four significant variables are displayed on the horizontal axis against probability values on the vertical axis, making it easier to observe and interpret the comparison between latent classes. From the graph, class three obtained the highest rating on each corresponding dimensions (coach, teammates, family, and club) maximum score of six on all items. Class 1 obtained the lowest rating on each corresponding dimension’s score while class two obtained average ratings an each corresponding dimension.

Following the LCA analysis, the SASQ 17- items that significantly contributed to discriminating participants were determined.. Table 6 demonstrates the loadings indicating “how well” the indicator is explained by the model [63, p. 113], and mean scores of the item across the three groups, alongside their p values estimated based on Wald statistics. The loading indices are the magnitude of the contribution of the variable to measuring the construct. All items had higher loadings (> 0.5) except for item Q22R which reported a loading of 0.489. The three groups of participants were differentiated based on the SASQ 17-item questions. For example, high achievers had the highest mean (75.047) for item Q5 followed by average achievers (54.711) and finally low achievers (53.655), meaning that the response of high achievers as regards item Q5 was significantly higher than responses generated by participants in average and low achievers group. Similarly, response concerning item Q25 reported that high achievers (M = 17.229) were significantly higher than those of average and low achievers groups (M1,2 = 12.339, 11.119) (p < 0.05).

thumbnail
Table 6. Classification of the SASQ items in three class model.

https://doi.org/10.1371/journal.pone.0281971.t006

Four models were fitted to the data and as demonstrated in Table 7, a three class model differentiating between three different groups (low achievers, average achievers, and high achievers) had an optimal fit to the data (BIC = 4172.964; AIC = 3437.369; CAIC = 3642.426). Although the model had a higher AIC index than that of a the four-class model, however, since it had lower BIC and CAIC, a smaller number of parameters to estimate (161; more parsimonious), and a significant conditional bootstrapping p value, there was evidence for its optimal fit. Finally, the effect sizes computed by the entropy R2 index were 0.8658 (two class), 0.9184 (three class), and 0.9738 (four class), which translated into 87%, 91%, and 97% respectively of accuracy in classifying the participants into three latent classes and the prediction of their SASQ test scores.

Internal reliability.

Once the CFA and LTA validated the SASQ factorial structure with 17 items (see S1 Appendix), a reliability analysis to measure the internal consistency was performed using Cronbach’s alpha (α). Nevertheless, its limitations are well known [65], including the assumptions of uncorrelated errors, tau-equivalence and normality. Therefore, the current study analysed the internal consistency using composite reliability, which provided a better option (Table 8). Raykov’s formula was used to calculate the composite reliability [66]. An acceptable threshold for composite reliability falls between 0.6 to 0.7 but should not be above 0.95 [38]. This means that the underlying SASQ factors presented acceptable internal consistency and can be used reliably in future research or real-world applications.

Test re-test reliability.

In order to evaluate the reliability, a test-retest analysis was carried out [30, 6769], with the same subject measured at two different times (always in conditions of similar application) [33, 67, 69].

Participants.

One hundred and seventy-four athletes (males N = 94 and females N = 80) performing their sport at the regional (N = 120) or national (N = 54) level participated in this study. The mean age was 17 years (age range: 12–21; SD = 1.79). All participated in team sports (football N = 48, 27.6%; basketball N = 67, 38.5%; volleyball N = 25, 14.4%; rugby N = 10, 5.7%; and unihockey N = 24, 13.8%). Initially, 189 participants took part in the first administration of the SASQ; however, only 174 participants participated in both administrations. Since we had 17 items, the sample size (N = 174) met the empirical rule of the 10:1 participant-items ratio recommended in the literature [30, 35].

Procedure.

An approval was granted by the Lausanne University Research Ethics Commission (Project number: E_SSP_122021_00002) for the study. The procedures were explained to participants, and they subsequently gave their written consent, along with their parents or guardians if they were under 16 years of age. Between August and October 2022, paper-pencil SASQ questionnaires were distributed to the participants. At the first administration, participants were assigned a code name based on their initials, age, and the initials of their parents (mother and father). Scale reliability was assessed by completing the questionnaire a second time, with code names to ensure correspondence between results at time 1 (T1) and time 2 (T2). According to Bonnet’s guidelines [70] the sample size to obtain a 95% CI for the intraclass correlation coefficient (ICC) (p) with a desired width of 0.2 for two repetitions is greater than 159 if p is ≥ 0.6. The current study recruited 174 participants which was well above the recommendation. The SASQ instrument was administered twice at 4-week intervals.

Data analysis.

Descriptive statistics were performed on the 17 items with means, standard deviations, measures of skewness and kurtosis, and analysis of missing values. In the test-retest reliability analysis, we used the ICC (two-way mixed effects, absolute agreement) and a paired sample t-test to compare the factors between measurement times. Koo and colleague [44], defined ICC values less than 0.5 indicated low reliability; between 0.5 and 0.75, moderate; between 0.75 and 0.90, good; and above 0.90, excellent. All data were analysed using SPSS (version 27).

Results.

Scale reliability was supported by test-retest data (Table 9) (S1 Dataset).

thumbnail
Table 9. Descriptive statistics, internal reliability and intraclass correlation (N = 174).

https://doi.org/10.1371/journal.pone.0281971.t009

In general, participants presented moderate to high scores on the factors at T1 (mean: 4.14–5.17) and T2 (mean: 4.13–5.33), with corresponding similar standard deviations: T1 (SD = 0.15–0.79) and T2 SD = 0.06–0.76, respectively. According to the ICC scores, the coach factor showed “excellent reliability” between T1 and T2 and teammates and family factors showed “good reliability”. However, the ICC scores for the club factor indicated “moderate reliability”. Cronbach’s alpha varied minimally between T1 and T2 on all the four factors. A paired t-test indicated no significant difference between the two administrations.

Discriminant validity analysis

According to Hair, discriminant validity ensures that a construct measure is empirically unique and represents phenomena of interest that other measures in a structural equation model do not capture [71]. Essentially, discriminant validity requires that test results are not highly correlated with measures from which they are supposed to differ [72]. In the current study, we propose the heterotrait-monotrait ratio of correlations (HTMT) as a new approach to assess discriminant validity in variance-based SEM [73]. This is due to the fact that recent research suggests that the routinely used Fornell-Larcker criterion is not effective under certain circumstances of discriminant validity testing [74].

Since there are two ways of using the HTMT to assess discriminant validity: (i) as a criterion or (ii) as a statistical test, the former involves comparing it to a predefined threshold with the exact threshold levels of the HTMT being debatable [43, 75]. We therefore chose to use the statistical test to assess the HTMT. This was carried out by applying a bootstrapping procedure to construct confidence intervals for the HTMT in order to test the null hypothesis (H0: HTMT ≥ 1) against the alternative hypothesis (H1: HTMT < 1) . A confidence interval (CI) containing the value one (i.e., H0 holds) indicates a lack of discriminant validity. Conversely, if the value one falls outside the interval’s range, this suggests that the two constructs are empirically distinct. The advantage of using CI is that they provide more information by highlighting the direction and magnitude of the difference, or, if the hypothesis is not rejected, the power of the procedure can be assessed by the width of the interval [76].

Results.

The computations yields values between (0.661 and 0.849) (Table 10). Upon comparing if the CI result value (one) falls outside the interval range, it can be reported that all the four constructs are empirically distinct.

Convergent validity.

In order to confirm the extent to which the measures capture the common SASQ constructs, a convergent validity analysis was performed. It involved calculating the average factor loading for all the measures in each SASQ construct (i.e., the sum of the squared loadings divided by the number of measures). Convergent validities above r = .70 are recommended, whereas those below r = .50 should be avoided [77]. However, research evidence suggests that actual levels of convergent validity in psychology research still vary widely [77]. Therefore, without more specific guidance, researchers reach logically inconsistent conclusions, arguing that a convergent validity as low as r = .28 [e.g.,78] indicates the measures converge, whereas others report convergent analysis as high as r = .75 [e.g., 79] signalling high convergence.

Results.

Results in Table 11 indicate that the measures of coach and teammates constructs (0.732 and 0.761) respectively reported high convergent validity. In contrast, both the measures of family and club constructs (0.657 and 0.631) respectively reported average or modest convergent validity.

Study 4

In this fourth study, we standardized the SASQ values by generating scores that ranked the measured factors in order to increase the reliability and generalizability.

Participants

The participants and dataset in study 2 were again used to standardize the SASQ scores. The standardization analysis thus used 543 participants (mean age: 16.67, SD: 2.42, range: 12–21 years).

Method.

Once descriptive analyses were performed on the raw scores, focus was put on standardizing the SASQ scores. This was done using the R statistical software package, with the raw scores transformed into T-scores to calculate the median scores for each of the four SASQ dimensions (S2 Dataset). Once the median score was established, three distribution groups were formed: low achievers, average achievers, and high achievers (Fig 3). Scores ≤ 40% designated low achievers and scores ≥ 60% designated high achievers. For the average achievers, a score was established by determining the average between the low and high achievers, resulting in a 50% score representation. For ease of interpretation, high achievers were described as athletes who were able to deploy the necessary SAS to deal with problematic experiences during club changes. In contrast, low achievers were described as athletes who did not have these skills. Average achievers were those athletes who had some but not all of these SAS.

thumbnail
Fig 3. Group distribution/ranks of each of the four SASQ dimensions.

https://doi.org/10.1371/journal.pone.0281971.g003

Results.

For the coach dimension, most of the scores indicated low achievers, with 40% of the total score. In contrast, average and high achievers respectively represented 38% and 22%. In general, these results indicated that most of the athletes encountered problematic situations concerning the coaching dimension. For the teammates dimension, the low achiever group represented the highest percentage of the total score with 49%, followed by high achievers at 28% and average achievers at 23%. The club dimension indicated that average achievers were the majority, with 42% of the total score as opposed to 33% and 25% for the low and high achievers, respectively. A different situation was observed for the family dimension in that low achievers represented 36% of the total score, while the average and high achievers each represented 32%. In general, apart from the club dimension, for which the average achievers were the majority, low achievers were the majority in the remaining three dimensions, clearly indicating that within the general French-speaking Swiss youth athlete population, most do not possess or deploy the SAS skills during problematic club changes.

An example. Alex, a footballer, obtained the following scores: coach: 20, teammates: 15, family: 9, and club:10, for a total SASQ score of 54. Using Table 12, Alex is a low achiever for the score on the coach dimension, part of 40% of the general French-speaking Swiss population. Thus, 60% of this general population had higher scores than Alex. The scores on the teammates dimension indicate that he is a high achiever and has higher scores than 59% of our general study population. The score of 9 on the family dimension presents Alex as an average achiever, having outscored 40% of our general study population while another 40% had higher scores than he did. Last, with a score of 10 on the club dimension, he was a high achiever, with a higher score than 59% of our general study population. Overall, Alex’s total percentage SASQ score was 53%.

An examination of each of the three group means found that most individuals rated on SASQ were classified as low achievers on all of the three dimensions (i.e., coach teammates, and family) except on club dimension which was represented mostly by average achievers (Table 13). The results above indicate that, in general, most participants lacked the possession and deployment of SAS and that only a few individuals possessed and deployed the necessary SAS.

Discussion

The aim of the current study was to outline the initial development and validation of a questionnaire designed to assess the SAS levels of young athletes in talent development. In addition to providing regular feedback and refinement, SAQ instrument was also expected to identify areas that needed immediate attention, improvement or maintenance.

The current study demonstrated acceptable internal consistency when compared with previous studies and reference instruments, such as the Psychological Characteristics of Developing Excellence (PCDEQ) questionnaire [4] or the Talent Development Environment Questionnaire (TDEQ) [5]. More specifically, given that the data from EFA analysis presented a general reliability of α = 0.876 and the reliability for each of item ranging from α = 0.865 to α = 0.875, SASQ should be considered a reliable instrument. Compared to the reliability scores of other instruments that measure social skills, example: the Social Skills Questionnaire (CHASO III) [13] with α = 0.70, the Social Skills Questionnaire for College Students (SSQ-U) [14] with α = 0.70, or the Social Skills Questionnaire for Traumatic Brain Injury (SSQ-TBI) [80] with α = 0.80. The SASQ provides reliability in that, it can measure the situated social skills that athletes possess and, therefore, it can be relied upon by researchers.

According to previous research [2], the 17 items spread over four factors were also satisfactory. Since the objective of the current study was to develop and validate an easily understandable instrument that measures relevant dimensions of the problematic experiences encountered by athletes during club changes, a 17 item instrument was deemed important. The fact that the SAQ instrument retained the previously proposed four-factor model increases its value. Compared to long, complicated scales that have been applied in previous studies (e.g., TDEQ with 68 items [5], Social Skills Questionnaire (CHASO III) with 76 items [13], Social Skills Inventory with 90 items [81], and the Social Skills Questionnaire for College Students (SSQ-U) with 38 items [14], shorter psychometric instruments are preferable [82]).

The structural equation modelling (SEM) results indicated that, in general, the SASQ model was adequate for validating the constructs. The EFA enabled us to reduce the number of SASQ items from 34 to 17 with four dimensions by selecting items that represented better psychometric properties. Doing so resulted in adequate levels of correlation, which indicates excellent validity, as confirmed through CFA yielding four stable factors. In addition, the construct validity of the SASQ agreed with the findings of Owiti and Hauw [2], who also reported a four-factor model related to coach, teammates, family and friends, and club. In this sense, the dimensions of the resulting scale encompass all possible levels of the social adaptability skills that athletes require during club changes.

Regarding the model fit of the CFA, we considered various criteria in validating the fit indices. According to various research [30, 31, 54, 55, 83], RMSEA values greater than 0.1 indicate poor performance, between 0.08 and 0.10, acceptable, and less than 0.05 indicate acceptable performance. The current study reported a value of 0.06 indicating an acceptable fit indices. The TLI value of 0.844 in the current study was a little below the criterion since an acceptable value should be over 0.9 for a good fit [5052, 5456]. The Chi-square result was significant, and it has been suggested that this value is sensitive to the sample size. It was therefore normalized and had an acceptable result of below 3 [47, 48, 84]. Based on these indices, this sample fits a four-factor model reasonably well.

Accordingly, in terms of factor loading, factor weights are usually considered significant when equal to or greater than 0.5 (FL ≥ 0.50) [31]. However, some authors suggest that 0.30 is the minimum requirement for interpreting a sample size ≥ 350 [34, 36, 38]. The CFA revealed that 15 loadings were greater than 0.30 (range: 0.30 to 0.62) and only two loadings were below 0.30 (range: 0.26 to 0.29). Low factor loadings suggest that the items are correlated, thus not measuring well-separated latent concepts. We offer two possible explanations as regards the low factor loadings; first, in a real world setting, it is extremely difficult to find latent factors that make up one concept being absolutely independent. As for the case of SASQ which measures the problematic experiences of players during club changes, these experiences may be closely associated and therefore expressed in the same way by the participants. Second, since the participants formed a homogeneous group, as all were aspiring athletes, it is possible that the distinguishing experiences of a specific difficulty were not expressed.

The computational yield for the discriminant validity in the current study reported the following results: (Coach and Teammates = 0.733; Coach and Family = 0.676; Coach and Club = 0.742; Family and Teammates = 0.849; Club and Family = 0.661; Teammates and Club = 0.766). Comparing these results with the threshold values as defined in HTMT [77], suggested that discriminant validity had been established. Further test of convergent validity which reflects the extent to which measures capture common constructs reported the following results: (Coach = 0.732; Teammates = 0.761; Family = 0.657; Club = 0.631). Convergent validity range (r = .70 and above) being recommended whereas range (r = 0.5 and below) is to be rejected [77]. Despite the favourable convergent validity results reported in the current study, research evidence suggest that actual levels of convergent validity still vary widely [77].

It should be noted that good practice dictates a minimum of three items per factor [52, 85]. Three of the SASQ factors consisted of at least three items per factor, although the club factor had only two items. In retaining the club factor with only two items, we present the following arguments: First, there was a strong theoretical and practical reason linked to the observation that future athletes are already self-motivated to pursue their careers despite problematic challenges concerning the club. Second, our decision is supported by previous research in the psychology literature suggesting that constructs that do not have a wide domain or those that are not conceptualized as multidimensional may present single-item measures [8689]. Third, a factor with two items is only considered reliable when the items are highly correlated (r = 0.70) but fairly uncorrelated with the other items, as was the case in the current study [90].

Concerns have been raised about considering psychometric measures as non-interdependent scales because the reliability of single-item scales can be legitimately questioned [91]. This suggestion extends to the SASQ-17 item psychometric measure which cannot be considered as uncorrelated [92]. However, since validity implies reliability [93, 94] the choice of the most relevant SASQ items were based on pragmatic grounds including their predictive value for relevant outcomes. It is therefore essential to emphasize the importance of external validity when choosing the most suitable SASQ scores, some of which should be correlated or not with each other. Adequate SASQ scores should not be selected to internal validity evidence alone [46].

The results of the LCA for predicting SAS disposition and deployment by athletes replicated findings of the CFA underlying the benefits of combining two methodologies. In applying the predictive validity of the SASQ items, we confirmed the existence of three specific participant profiles that could be differentially related to SAS during club to club change. Results of the LCA suggested that the SASQ could discriminate participants into three characteristic profiles (example: low achievers, average achievers, and high achievers). In line with the LCA approach, athletes could be clustered into meaningful and relatively homogenous classes on the basis of the SASQ-17 item scores.

According to our findings, categories established on the basis of the SASQ 17-items were associated with distinct outcomes such as a participant being either in a low, average, or high achiever group. This highlighted the fact that it was possible to perform SASQ scores based predictions through this process of group categorization. In this case SASQ data could be collected as part of routine psychosocial skills training within talent and development programs. Indeed and in line with the results of the present study, SASQ items could be used as a first step to identify specific subgroups when athletes change from club to club. In a second step, specific profile membership and its association with important variables could be used to inform the coach and club decisions.

The LCA also reported optimal fit for SASQ items based on BIC, AIC, and CAIC results, this was a confirmation of similar results of “goodness fit” reported in CFA. However, it should be noted that the LCA reported a non-significant result on item Q22R (I always feel paralysed whenever my trainer/coach puts pressure on me to reach my sporting goals) on its inability to discriminate participants as concerns difficulties encountered with coaches. However, this was not the case with the CFA model.

At this point, we have to admit that despite equipping young athletes with these psychosocial skills, it does not necessarily guarantee that they will reach elite level competition, since there exists other variables that could influence the possibility of achieving success. However, their absence may prevent these athletes in overcoming obstacles along their success pathway [3, 10].

Limitations

Although we developed a valid and reliable instrument to measure social adaptability skills during club changes, our work has some limitations that must be kept in mind when using the instrument. First, the measure was constructed based on data from six team sports, which makes generalizability a limitation. Therefore, before using SASQ in other sports, evidence of the validity and reliability of the instrument in those sports is needed. In addition, our sample was drawn from the French-speaking part of Switzerland, which points to the culture-based approach of our standardization of the SASQ. Further research might focus on multicultural validation of the SASQ.

Second, considering that the participants were nested within teams, multilevel analysis could have also been applied. The this type of analysis might have provided more accurate parameter estimates, standard errors, and associated tests of significance, as well as allowing us to examine variance by sport type. In future, researchers should use multilevel modelling to examine SASQ data and also determine whether there are systematic differences between sport types and the four factors.

Third, due to the general interest, we have to admit that the LCA reported Q22R as having low factor loading, however, the CFA confirmed the same Q22R as having contributed to the factor structure of SASQ.

Implication for practice

According to the current findings, although new, SASQ instrument appears to be useful for researchers, coaches, and clubs alike. Individuals working with athletes can use the instrument to predict whether young athletes have the necessary social adaptability skills to handle club changes and, based on the results, they can design intervention and inclusion programs for them. During such interventions, the constructs of the SASQ can be measured before and after the club change in order to evaluate any changes in these behavioural precursors.

For the young athletes to develop the SAS necessary to overcome future obstacles, their immediate environment needs to be carefully managed. It is also important to integrate education programs designed to evaluate and develop these SAS, whether they take the form of formal or informal psychosocial training within talent identification and development programs.

Professional players may also benefit from the findings of the current study. In order to facilitate smooth transitions between clubs, they could be encouraged to assess whether they possess the necessary skills. SASQ has the potential to benefit coaches by providing insight that can enhance players’ transition from one club to another and focusing on development. Professional sports clubs and high-performance organizations could also use the questionnaire for identifying major adaptation barriers faced by incoming players. Last, the findings can also be used by sports clubs to maintain and enhance existing resources that help players with transitions.

References

  1. 1. CIES Observatory (2019). International Basketball Migration Report. Retrieved from http://www.fiba.basketball/documents/ibmr2019.pdf (accessed June 16, 2022).
  2. 2. Owiti S., & Hauw D. (2021). The problematic experience of players’ mutations between clubs: Discovering the social adaptability skills required. Frontiers in Sports and Active Living, 3, 591438. https://doi.org/10.3389/fspor.2021.591438 pmid:34957394
  3. 3. Collins D., & MacNamara Á. (2012). The rocky road to the top: Why talent needs trauma. Sports Medicine, 42(11), 907–914. https://doi.org/10.1007/BF03262302 pmid:23013519
  4. 4. MacNamara Á., & Collins D. (2011). Development and initial validation of the psychological characteristics of developing excellence questionnaire. Journal of Sports Sciences, 29(12), 1273–1286. https://doi.org/10.1080/02640414.2011.589468 pmid:21812724
  5. 5. Martindale R. J. J., Collins D., Wang J. C. K., McNeill M., Lee K. S., Sproule J., et al. (2010). Development of the talent development environment questionnaire for sport. Journal of Sports Sciences, 28(11), 1209–1221. https://doi.org/10.1080/02640414.2010.495993 pmid:20694933
  6. 6. Hauw D., Gesbert V., Crettaz von Roten F., & Rolland J.-P. (2022). A multilayer approach for assessing the psychological needs of aspiring soccer players: Implications for overseeing talent development. Professional Psychology: Research and Practice, 53(2), 133–142. https://doi.org/10.1037/pro0000438
  7. 7. Abbott A., & Collins D. (2004). Eliminating the dichotomy between theory and practice in talent identification and development: Considering the role of psychology. Journal of Sports Sciences, 22(5), 395–408. https://doi.org/10.1080/02640410410001675324 pmid:15160593
  8. 8. Toering T., Elferink-Gemser M., Jordet G., Jorna C., Pepping G.-J., & Visscher C. (2011). Self-regulation of practice behavior among elite youth soccer players: An exploratory observation study. Journal of Applied Sport Psychology, 23(1), 110–128. https://doi.org/10.1080/10413200.2010.534544
  9. 9. Van Yperen N. W. (2009). Why some make it and others do not: Identifying psychological factors that predict career success in professional adult soccer. The Sport Psychologist, 23(3), 317–329. https://doi.org/10.1123/tsp.23.3.317
  10. 10. MacNamara Á., Button A., & Collins D. (2010). The role of psychological characteristics in facilitating the pathway to elite performance part 1: Identifying mental skills and behaviors. The Sport Psychologist, 24(1), 52–73. https://doi.org/10.1123/tsp.24.1.52
  11. 11. Gesbert V., & Hauw D. (2019). Commentary: Interpersonal coordination in soccer: interpreting literature to enhance the representativeness of task design, from dyads to teams. Frontiers in Psychology, 10, 1093. https://doi.org/10.3389/fpsyg.2019.01093 pmid:31139125
  12. 12. Bexell S. M., Decker Sparks J. L., Tejada J., & Rechkemmer A. (2019). An analysis of inclusion gaps in sustainable development themes: Findings from a review of recent social work literature. International Social Work, 62(2), 864–876. https://doi.org/10.1177/0020872818755860
  13. 13. Caballo V. E., & Salazar I. C. (2017). Development and Validation of a New Instrument for the Evaluation of the Social Skills: The" Social Skills Questionnaire"(Ssq). Behavioral Psychology-Psicologia Conductual, 25(1), 5–24.
  14. 14. Morán V. E., Olaz F. O., & Del Prette Z. A. P. (2015). Social skills questionnaire for Argentinean college students (Ssq-u) development and validation. The Spanish Journal of Psychology, 18, E95. https://doi.org/10.1017/sjp.2015.92 pmid:26610605
  15. 15. Owiti S., Hauw D., & Collins D. (2020). Applying a multilayer construct of social adaptability skills within talent development. Frontiers in Psychology, 10, 3006. https://doi.org/10.3389/fpsyg.2019.03006 pmid:32010031
  16. 16. Owiti S., Bersier T., & Hauw D. (2021). Individual differences in professional sport narrative experience during basketball players club mutation. Heliyon, 7(9), e08015. https://doi.org/10.1016/j.heliyon.2021.e08015 pmid:34585019
  17. 17. American Educational Research Association, American Psychological Association, National Council on Measurement in Education, & Joint Committee on Standards for Educational and Psychological Testing (U.S.) (Eds.). (1999). Standards for educational and psychological testing. American Educational Research Association.
  18. 18. Zervas Y., Stavrou N. A., & Psychountaki M. (2007). Development and validation of the self-talk questionnaire (S-tq) for sports. Journal of Applied Sport Psychology, 19(2), 142–159.
  19. 19. Corbin J. M., & Strauss A. L. (2015). Basics of qualitative research: Techniques and procedures for developing grounded theory (4th ed.). Los Angeles: SAGE.
  20. 20. Holt N. L., & Tamminen K. A. (2010). Moving forward with grounded theory in sport and exercise psychology. Psychology of Sport & Exercise, 11(6), 419–422.
  21. 21. Strauss A., & Corbin J. (1998). Basics of qualitative research techniques. Thousand Oaks, CA: Sage publications.
  22. 22. Weed M. (2009). Research quality considerations for grounded theory research in sport & exercise psychology. Psychology of Sport & Exercise, 10(5), 502–510.
  23. 23. Fraenkel J. R., & Wallen N. E. (2009). How to design and evaluate research in education (7th ed). McGraw-Hill.
  24. 24. Wiersma L. D. (2001). Conceptualization and development of the sources of enjoyment in youth sport questionnaire. Measurement in Physical Education and Exercise Science, 5(3), 153–177. https://doi.org/10.1207/S15327841MPEE0503_3
  25. 25. Conrad, F., & Blair, J. (1996, August). From impressions to data: Increasing the objectivity of cognitive interviews. In Proceedings of the section on survey research methods, annual meetings of the American Statistical Association (Vol. 1, No. 10). Alexandria: American Statistical Association.
  26. 26. Dillman D. A. (2000). Mail and Internet surveys: The tailored design method. New York: Wiley.
  27. 27. Czaja R., Blair J., & Blair E. (2014). Designing surveys: A guide to decisions and procedures (Third edition). SAGE.
  28. 28. Collins D. (2003). Pretesting survey instruments: an overview of cognitive methods. Quality of life research, 12(3), 229–238. pmid:12769135
  29. 29. Chang Lei. (1994). A psychometric evaluation of 4-point and 6-point Likert-type scales in relation to reliability and validity. Applied Psychological Measurement, 18(3), 205–215. https://doi.org/10.1177/014662169401800302
  30. 30. Cid L., Monteiro D., Teixeira D. S., Evmenenko A., Andrade A., Bento T., et al. (2022). Assessment in sport and exercise psychology: Considerations and recommendations for translation and validation of questionnaires. Frontiers in Psychology, 13, 806176. https://doi.org/10.3389/fpsyg.2022.806176 pmid:35360588
  31. 31. Field A. P. (2020). Discovering statistics using IBM SPSS statistics (Fourth edition. South Asian adaptation). SAGE Publications.
  32. 32. Wiktorowicz, J. (2016). Exploratory factor analysis in the measurement of the competencies of older people / Eksploracyjna analiza czynnikowa w ocenie kompetencji osób starszych. Ekonometria, 4. https://doi.org/10.15611/ekt.2016.4.03
  33. 33. Hill M., & Hill A. (2005). Investigacao por Questionario, 2nd Edn. Lisbon: Edicoes Silabo.
  34. 34. Worthington R. L., & Whittaker T. A. (2006). Scale development research: A content analysis and recommendations for best practices. The Counseling Psychologist, 34(6), 806–838. https://doi.org/10.1177/0011000006288127
  35. 35. DeVellis R. F. (2017). Scale development: Theory and applications (Fourth edition). SAGE.
  36. 36. Schafer J. L. (1999). Multiple imputation: A primer. Statistical Methods in Medical Research, 8(1), 3–15. https://doi.org/10.1177/096228029900800102 pmid:10347857
  37. 37. Kahn J. H. (2006). Factor analysis in counseling psychology research, training, and practice: Principles, advances, and applications. The Counseling Psychologist, 34(5), 684–718. https://doi.org/10.1177/0011000006286347
  38. 38. Hair J. F. (Ed.). (2017). A primer on partial least squares structural equation modeling (Pls-sem) (Second edition). Sage.
  39. 39. Tabachnick B. G., Fidell L. S., & Ullman J. B. (2019). Using multivariate statistics (Seventh edition). Pearson.
  40. 40. Goetz A. R., Cougle J. R., & Lee H.-J. (2013). Revisiting the factor structure of the 12-item Disgust Propensity and Sensitivity Scale–Revised: Evidence for a third component. Personality and Individual Differences, 55(5), 579–584. https://doi.org/10.1016/j.paid.2013.04.029
  41. 41. Siembida E. J., Moss K., Kadan-Lottick N., & Bellizzi K. M. (2018). The patient–provider relationship in adolescent oncology: Anexploratory factor analysis of a thirteen-item self-report measure. Journal of Adolescent Health, 63(4), 509–512. https://doi.org/10.1016/j.jadohealth.2018.04.016 pmid:30131286
  42. 42. Pett M. A., Lackey N. R., & Sullivan J. J. (2003). Making sense of factor analysis: The use of factor analysis for instrument development in health care research. Sage Pub.
  43. 43. Kline R. B. (2016). Principles and practice of structural equation modeling (Fourth edition). The Guilford Press.
  44. 44. Koo T. K., & Li M. Y. (2016). A guideline of selecting and reporting intraclass correlation coefficients for reliability research. Journal of Chiropractic Medicine, 15(2), 155–163. https://doi.org/10.1016/j.jcm.2016.02.012 pmid:27330520
  45. 45. Abrahim S., Mir B. A., Suhara H., Mohamed F. A., & Sato M. (2019). Structural equation modeling and confirmatory factor analysis of social media use and education. International Journal of Educational Technology in Higher Education, 16(1), 32. https://doi.org/10.1186/s41239-019-0157-y
  46. 46. Golay P., Basterrechea L., Conus P., & Bonsack C. (2016). Internal and predictive validity of the french health of the nation outcome scales: Need for future directions. PLOS ONE, 11(8), e0160360. https://doi.org/10.1371/journal.pone.0160360 pmid:27483015
  47. 47. Rodrigues F., Hair J. F., Neiva H. P., Teixeira D. S., Cid L., & Monteiro D. (2019). The basic psychological need satisfaction and frustration scale in exercise (Bpnsfs-e): Validity, reliability, and gender invariance in Portuguese exercisers. Perceptual and Motor Skills, 126(5), 949–972. https://doi.org/10.1177/0031512519863188 pmid:31345112
  48. 48. Hoyle R. H. (Ed.). (2013). Structural equation modeling: Concepts, issues, and applications. SAGE Publications, Inc.
  49. 49. Brown T. A. (2015). Confirmatory factor analysis for applied research (Second edition). The Guilford Press
  50. 50. Arbuckle J. (2013). IBM SPSS Amo 22 User’s Guide. Armonk, NY: IBM
  51. 51. Biddle S. J. H., Markland D., Gilbourne D., Chatzisarantis N. L. D., & Sparkes A. C. (2001). Research methods in sport and exercise psychology: Quantitative and qualitative issues. Journal of Sports Sciences, 19(10), 777–809. https://doi.org/10.1080/026404101317015438 pmid:11561674
  52. 52. Hu L., & Bentler P. M. (1999). Cutoff criteria for fit indexes in covariance structure analysis: Conventional criteria versus new alternatives. Structural Equation Modeling: A Multidisciplinary Journal, 6(1), 1–55. https://doi.org/10.1080/10705519909540118
  53. 53. Byrne B. M. (2016). Structural equation modeling with Amos: Basic concepts, applications, and programming (Third edition). Routledge, Taylor & Francis Group
  54. 54. Peugh J., & Feldon D. F. (2020). “How well does your structural equation model fit your data?”: Is marcoulides and yuan’s equivalence test the answer? CBE—Life Sciences Education, 19(3), es5. https://doi.org/10.1187/cbe.20-01-0016 pmid:32510273
  55. 55. Yuan K.-H., Chan W., Marcoulides G. A., & Bentler P. M. (2016). Assessing structural equation models by equivalence testing with adjusted fit indexes. Structural Equation Modeling: A Multidisciplinary Journal, 23(3), 319–330. https://doi.org/10.1080/10705511.2015.1065414
  56. 56. Marsh H. W., Hau K.-T., & Wen Z. (2004). In search of golden rules: Comment on hypothesis-testing approaches to setting cutoff values for fit indexes and dangers in overgeneralizing hu and bentler’s (1999) findings. Structural Equation Modeling: A Multidisciplinary Journal, 11(3), 320–341. https://doi.org/10.1207/s15328007sem1103_2
  57. 57. Magidson J., & Vermunt J. K. (2004). Latent class models. In D. Kaplan (Ed.), The Sage handbook of quantitative methodology for the social sciences (pp. 175–198). Thousand Oaks, CA: Sage Publications.
  58. 58. Nylund K. L., Asparouhov T., & Muthén B. O. (2007). Deciding on the number of classes in latent class analysis and growth mixture modeling: A monte carlo simulation study. Structural Equation Modeling: A Multidisciplinary Journal, 14(4), 535–569. https://doi.org/10.1080/10705510701575396.
  59. 59. Magidson J., & Vermunt J. K. (2016). Latent Gold [Computer software]. Belmont, MA: Statistical Innovations.
  60. 60. Fraley C. (1998). How many clusters? Which clustering method? Answers via model-based cluster analysis. The Computer Journal, 41(8), 578–588. https://doi.org/10.1093/comjnl/41.8.578.
  61. 61. Vrieze S. I. (2012). Model selection and psychological theory: A discussion of the differences between the Akaike information criterion (Aic) and the Bayesian information criterion (Bic). Psychological Methods, 17(2), 228–243. https://doi.org/10.1037/a0027127 pmid:22309957
  62. 62. Booth J., Hall P., & Wood A. (1992). Bootstrap estimation of conditional distributions. The Annals of Statistics, 20(3). https://doi.org/10.1214/aos/1176348786.
  63. 63. Vermunt J. K., & Magidson J. (2005). Technical guide for Latent GOLD Choice 4.0: Basic and advanced. Belmont, MA: Statistical Innovations Inc.
  64. 64. Kaplan D., & Keller B. (2011). A note on cluster effects in latent class analysis. Structural Equation Modeling: A Multidisciplinary Journal, 18(4), 525–536. https://doi.org/10.1080/10705511.2011.607071.
  65. 65. Yang Yanyun, & Green, S. B. (2011). Coefficient alpha: A reliability coefficient for the 21st century? Journal of Psychoeducational Assessment, 29(4), 377–392. https://doi.org/10.1177/0734282911406668
  66. 66. Raykov T. (1997). Estimation of composite reliability for congeneric measures. Applied Psychological Measurement, 21(2), 173–184. https://doi.org/10.1177/01466216970212006
  67. 67. Nideffer R. M., & Sagal M.-S. C. (2001). Assessment in sport psychology. Fitness Information Technology.
  68. 68. Buckworth J., Dishman R., O’Connor P., & Tomporowski P. (2013). Exercise Psychology. Champaign, IL: Human Kinetics.
  69. 69. Vallerand R. J. (1989). Vers une méthodologie de validation trans-culturelle de questionnaires psychologiques: Implications pour la recherche en langue française. Canadian Psychology / Psychologie Canadienne, 30(4), 662–680. https://doi.org/10.1037/h0079856
  70. 70. Bonett D. G. (2002). Sample size requirements for estimating intraclass correlations with desired precision. Statistics in Medicine, 21(9), 1331–1335. https://doi.org/10.1002/sim.1108 pmid:12111881
  71. 71. Hair J. F. (Ed.). (2010). Multivariate data analysis (7th ed). Prentice Hall.
  72. 72. Campbell D. T. (1960). Recommendations for APA test standards regarding construct, trait, or discriminant validity. American Psychologist, 15(8), 546–553. https://doi.org/10.1037/h0048255.
  73. 73. Henseler J., Ringle C. M., & Sarstedt M. (2015). A new criterion for assessing discriminant validity in variance-based structural equation modeling. Journal of the Academy of Marketing Science, 43(1), 115–135. https://doi.org/10.1007/s11747-014-0403-8.
  74. 74. Henseler J., Dijkstra T. K., Sarstedt M., Ringle C. M., Diamantopoulos A., Straub D. W., et al. (2014). Common beliefs and reality about pls: Comments on Rönkkö and Evermann (2013). Organizational Research Methods, 17(2), 182–209. https://doi.org/10.1177/1094428114526928
  75. 75. Clark L. A., & Watson D. (1995). Constructing validity: Basic issues in objective scale development. Psychological Assessment, 7(3), 309–319. https://doi.org/10.1037/1040-3590.7.3.309
  76. 76. Shaffer J. P. (1995). Multiple hypothesis testing. Annual Review of Psychology, 46(1), 561–584. https://doi.org/10.1146/annurev.ps.46.020195.003021
  77. 77. Carlson K. D., & Herdman A. O. (2012). Understanding the impact of convergent validity on research results. Organizational Research Methods, 15(1), 17–32. https://doi.org/10.1177/1094428110392383
  78. 78. Larraza-Kintana M., Wiseman R. M., Gomez-Mejia L. R., & Welbourne T. M. (2007). Disentangling compensation and employment risks using the behavioral agency model. Strategic Management Journal, 28(10), 1001–1019. https://doi.org/10.1002/smj.624
  79. 79. Podsakoff P. M., MacKenzie S. B., & Podsakoff N. P. (Eds.). (2018). The Oxford handbook of organizational citizenship behavior. Oxford University Press
  80. 80. Francis H. M., Osborne-Crowley K., & McDonald S. (2017). Validity and reliability of a questionnaire to assess social skills in traumatic brain injury: A preliminary study. Brain Injury, 31(3), 336–343. https://doi.org/10.1080/02699052.2016.1250954 pmid:28102714
  81. 81. Oldmeadow J. A., Quinn S., & Kowert R. (2013). Attachment style, social skills, and Facebook use amongst adults. Computers in Human Behavior, 29(3), 1142–1149. https://doi.org/10.1016/j.chb.2012.10.006
  82. 82. Liu E., & Blunden H. (2020). Cultural diversity and sensitivity in public estate renewals: Evidence from an Australian longitudinal study. Housing Studies, 35(6), 1004–1024. https://doi.org/10.1080/02673037.2019.1644296
  83. 83. Fabrigar L. R., Wegener D. T., MacCallum R. C., & Strahan E. J. (1999). Evaluating the use of exploratory factor analysis in psychological research. Psychological Methods, 4(3), 272–299. https://doi.org/10.1037/1082-989X.4.3.272
  84. 84. Bentler P. M. (1990). Comparative fit indexes in structural models. Psychological Bulletin, 107(2), 238–246. https://doi.org/10.1037/0033-2909.107.2.238 pmid:2320703
  85. 85. Raubenheimer J. (2004). An item selection procedure to maximise scale reliability and validity. SA Journal of Industrial Psychology, 30(4). https://doi.org/10.4102/sajip.v30i4.168
  86. 86. Bergkvist L., & Rossiter J. R. (2007). The Predictive Validity of Multiple-Item Versus Single-Item Measures of the Same Constructs. Journal of Marketing Research, 44(2), 175–184.
  87. 87. Drolet A. L., & Morrison D. G. (2001). Do we really need multiple-item measures in service research? Journal of Service Research, 3(3), 196–204. https://doi.org/10.1177/109467050133001
  88. 88. Gosling S. D., Rentfrow P. J., & Swann W.B. Jr. (2003). A Very Brief Measure of the Big Five Personality Domains. Journal of Research in Personality, 37, 504–528.
  89. 89. Wanous J. P., Reichers A. E., & Hudy M. J. (1997). Overall job satisfaction: How good are single-item measures? Journal of Applied Psychology, 82(2), 247–252. https://doi.org/10.1037/0021-9010.82.2.247 pmid:9109282
  90. 90. Yong A.G. & Pearce S. (2013) A Beginner’s Guide to Factor Analysis: Focusing on Exploratory Factor Analysis. Tutorials in Quantitative Methods for Psychology, 9, 79–94. https://doi.org/10.1080/10413200601185156
  91. 91. Speak B. L., Hay P., & Muncer S. J. (2015). HoNOS–their utility for payment by results in mental health. International Journal of Health Care Quality Assurance, 28(2), 115–128. https://doi.org/10.1108/IJHCQA-08-2013-0096 pmid:26335165
  92. 92. Muncer S. J., & Speak B. (2016). Mokken scale analysis and confirmatory factor analysis of the Health of the Nation Outcome Scales. Personality and Individual Differences, 94, 272–276. https://doi.org/10.1016/j.paid.2016.01.051.
  93. 93. Furr R. M., & Bacharach V. R. (2014). Psychometrics: An introduction (Second edition). SAGE.
  94. 94. Gavin H. (2008). Understanding research methods and statistics in psychology. SAGE Publications.