Validation of a socioemotional battery for online higher education applicants
-
Anabel de la Rosa GómezUniversidad Nacional Autónoma de México (México)
-
Pablo D. ValenciaUniversidad Nacional Autónoma de México (México)
-
Indira Ochoa CarrascoUniversidad Nacional Autónoma de México (México)
-
Alma Cristina Razo RuvalcabaUniversidad Nacional Autónoma de México (México)
How to cite: De la Rosa Gómez, A., Valencia, P. D., Ochoa Carrasco, I., & Razo-Ruvalcaba, A. C. Validation of a socioemotional battery for online higher education applicants. Revista Fuentes, 28(1), 14–27. https://doi.org/10.12795/revistafuentes.2026.28828
Abstract
Introduction: This study aimed to validate an institutional socioemotional battery for online bachelor’s degree applicants in Mexico, given the importance of these variables for academic success. Method: Two samples participated: Sample 1 (N=5272) for exploratory analysis and Sample 2 (N=2227) for confirmatory analysis. A 73-item self-report battery was initially administered. Data analysis involved item screening, exploratory graph analysis, and exploratory factor analysis on Sample 1, followed by confirmatory factor analysis on Sample 2 to validate the structure and assess reliability. Results: The final 30-item battery yielded seven stable dimensions: academic self-efficacy, family support, growth expectations, emotional expression, expressive suppression, emotional dysregulation, and adaptive emotional regulation. The model fit was excellent in both samples, with strong factor loadings, distinct inter-factor correlations, and high reliability. Most scales exhibited skewed distributions. Discussion: The identified dimensions align with literature on online academic success. The reduced, validated instrument enhances measurement efficiency and provides crucial early information on socioemotional factors. The combined use of network psychometric and traditional factor analytical methods strengthens the findings. Conclusions: This battery is a valid tool for assessing socioemotional characteristics in online learning, enabling early identification of risks and informing pedagogical strategies.
Keywords
academic success, emotional regulation, factor analysis, higher education, higher education, online education, psychometric validation, socioemotional variables.
1. Introduction
Recent literature has identified several indicators associated with academic success in online courses. Among these, academic self-efficacy stands out as a significant predictor of student performance and as a mediator of motivation, learning, and overall achievement (Gao et al., 2021; Lim et al., 2021; Neroni et al., 2022; Panigrahi et al., 2021; Yokoyama, 2024; Yunusa & Umar, 2021). For students in online and virtual environments, higher self-efficacy promotes better learning practices and retention (Kundu, 2020; Lehikko, 2021). It is also linked to the successful completion of educational programs, particularly at higher and graduate levels (Mejri & Borawski, 2023).
Social support is another crucial factor for student engagement and positive academic outcomes (Matteucci & Soncini, 2021; Ozer, 2024). Online students often juggle work, family, and studies, which can affect their ability to stay engaged and complete courses. Recent studies highlight that family involvement in education positively influences student engagement (Ferraces Otero et al., 2020; Yang et al., 2023). Conversely, a lack of this support can lead to students dropping out due to family responsibilities (Hernández González & Blackford, 2022; Hernández Ortiz et al., 2023). In contrast, students who receive support from both their family (partners and parents) and educational environments (teachers and peers) are more likely to engage and achieve academic success in online, hybrid, or face-to-face programs (Borup et al., 2020; Gao et al., 2021; Mensah et al., 2024).
Regarding the personal and work expectations of online students, research shows they are more likely to feel satisfied and continue their studies if their expectations align with their experiences. These expectations include course flexibility, social interaction, teacher support, and the alignment of the course with their personal and educational goals. Students seek to enjoy course content, face challenges that promote learning new things, and achieve rewards such as degrees, personal growth, and professional development (Henry, 2020; Landrum et al., 2021). Those who understand the time and effort required for a program are more likely to persist and complete more courses (James, 2022).
Emotional regulation may play a significant role in students’ academic achievement. Studies indicate that academic achievement emotions are influenced by a student’s level of emotional regulation (Harley et al., 2019; Nadeem et al., 2023; Pekrun et al., 2023). For instance, emotional dysregulation is associated with lower academic engagement (De Neve et al., 2023). Positive emotions (e.g., pride, confidence, joy) resulting from good grades can improve future grades, whereas negative emotions (e.g., frustration, anxiety, sadness) stemming from poor grades tend to reduce performance, potentially creating a negative cycle. While inadequate emotional regulation can affect both well-being and academic performance, effective emotional regulation fosters them (Harley et al., 2019). Strategies like cognitive reappraisal can boost well-being and help students manage their emotions better, whereas emotional suppression might be detrimental in the long run (Beaumont et al., 2023).
The literature on dropout in Distance Education indicates that it is influenced by a combination of factors related to the instructional modality and the learner’s contextual conditions. These include challenges in tracking students’ academic trajectories and insufficient instructional and organizational support (Escanés et al., 2014), as well as socioeconomic constraints such as associated costs and difficulties in maintaining enrollment. Academic, personal, and sociocultural factors also play a role, particularly among engineering and social science students (Bardales et al., 2025). Additionally, work schedules, family responsibilities, and gaps between expectations and the actual university experience hinder academic integration (Heredia Ramos, 2024). At the individual level, low motivation, difficulties adapting to online learning, limited self-regulation, and lack of time for autonomous learning significantly increase dropout risk (Delgado & Miranda, 2021; Heredia Ramos, 2024). In this regard, it is important to emphasize that many socioemotional variables associated with success in digital learning environments—such as intrinsic motivation, self-efficacy, self-regulated learning, sense of belonging, and perceived social and instructional support— can also operate as risk factors when they are absent or weakened.
Considering the importance of socioemotional variables for understanding students’ experiences in digital learning environments, it is important to consider how these dimensions have been assessed in empirical research. These variables and indicators have primarily been measured using questionnaires designed by participating institutions or by adapting instruments such as the MSLQ (Motivated Strategies for Learning Questionnaire), TOOLS (Test of Online Learning Success), ERQ (Emotion Regulation Questionnaire), Social and Emotional Competencies Questionnaire (SEC Q) and Socioemotional Competencies Scale (ECSE), bot in particular, having been applied in Latin American contexts comparable to of this study. Some studies relied on self-reporting through online surveys, while a smaller number conducted interviews and focus groups with students. In this context, the propaedeutic program offered by National Autonomous University of Mexico (Universidad Nacional Autónoma de México, UNAM) to its applicants and incoming students for distance learning degrees has incorporated several key indicators, including academic self-efficacy, social support, professional development expectations, and emotional regulation. Previous studies have utilized data from this program to examine the online learning skills of these applicants (Hernández Gutiérrez & Enríquez Vázquez, 2022). However, despite the administration of a battery of questions on the socioemotional characteristics of this population at the program’s outset, the authors are unaware of any examination of the psychometric properties of this battery to date.
In summary, understanding applicants’ readiness for online learning (Hassan Abuhassna et al., 2022) and assessing their intellectual and socioemotional capabilities are essential for designing effective learning experiences. Analyzing these dimensions not only facilitates adaptation to this modality but also helps anticipate and mitigate emotional barriers (Engin, 2017; Li, 2024; Yu et al., 2022). By integrating the assessment of these socioemotional variables into admission and follow-up processes, we contribute to successfully addressing the challenges of digital education, fostering retention and academic performance, and strengthening equity by providing timely support to those with greater needs. Therefore, the present instrumental study aimed to: (1) analyze the internal structure (dimensionality) of this battery of socioemotional variables, as provided in the supplementary materials (Supplementary Table 1, available via the OSF link), within the context of a propaedeutic course for applicants to a distance learning degree program; (2) estimate the reliability of the resulting scales; and (3) examine how these variables are distributed within the target population.
2. Method
2.1. Participants
The study involved two distinct samples. Sample 1, utilized for exploratory analyses, consisted of 5272 individuals who were in the process of applying to an online bachelor’s degree program, with a mean age of 31.33 years (SD = 9.98). The majority were women (67.1%), with men comprising 32.9%. In terms of marital status, most were single (65.7%), followed by those married or cohabiting (31.2%), and a smaller proportion divorced, separated, or widowed (2.9%). A substantial majority (57.6%) reported having no children. Regarding educational attainment, 58.0% had completed high school, 21.1% had incomplete undergraduate studies, 15.7% held an undergraduate degree, and 5.2% had some level of graduate studies. Employment was reported by 76.3% of this sample.
Sample 2, designated for confirmatory purposes, comprised 2227 individuals who had successfully navigated the university application process and had already been selected, with a mean age of 29.80 years (SD=9.44). This sample was composed of 53.0% women, 45.9% men, and 1.1% who preferred not to disclose their sex. Similar to Sample 1, the largest group was single (67.9%), with 26.5% married or cohabiting, and 5.7% divorced, separated, or widowed. A higher percentage (73.3%) in Sample 2 did not have children. Educational backgrounds in Sample 2 showed 40.7% with high school as their highest level, 29.6% with incomplete undergraduate studies, 20.8% with an undergraduate degree, and 8.9% with some graduate studies. Employment rates were comparable, with 74.6% of Sample 2 participants being employed.
2.2. Measure
Battery of Academic and Socioemotional Skills. The initial pool of items consisted of 73 self-report items, all responded to on a 7-point Likert-type scale. Its content included items related to academic self-efficacy, social support, emotional regulation, among other constructs. The first 18 items, designed to assess self-efficacy, ranged from 1 («not at all effective») to 7 («very effective»). The remaining 55 items were measured on a scale from 1 («Strongly disagree») to 7 («Strongly agree»). It is important to note that, while this comprehensive battery of items has been widely used, its psychometric properties had not been previously studied. The original pool of items can be found in Supplementary Table 1.
2.3. Procedure
Data for Sample 1 were collected between May and June 2024. Data for Sample 2 were collected from the program conducted between December 2024 and January 2025. Both data collection periods occurred within the context of the propaedeutic program that National Autonomous University of Mexico (Universidad Nacional Autónoma de México, UNAM) offers free of charge to applicants for its online degree programs.
A key distinction between the two cohorts considered for this study is their target populations: the first cohort included all applicants, while the second cohort comprised only those applicants who were selected in the admission process. At the time of this study, participation in the propaedeutic program was not mandatory, meaning all participants in this study did so voluntarily. The battery of questions that is the focus of this research was administered at the beginning of the program, before the official commencement of its three modules.
2.4. Data Analysis
Data analyses were conducted separately for the two independent subsamples. The first sample was used for item-level screening and exploratory procedures, while the second sample served as a holdout for confirmatory testing of the final measurement model.
In the first step, item-level descriptive statistics were examined to identify items with problematic distributional properties. Specifically, items with absolute zero-centered kurtosis values equal to or greater than 7 were considered excessively non-normal and were excluded from further analyses (Bandalos & Finney, 2019). Next, an exploratory graph analysis (EGA; Golino & Epskamp, 2017) was conducted on the remaining items to identify the underlying dimensional structure. A bootstrap procedure was applied to assess the stability of the item assignments to dimensions. Items were considered unstable and subsequently removed if they were assigned to the same dimension in less than 65% of the bootstrap replications (Christensen & Golino, 2021).
Within each dimension identified by the EGA, a two-step exploratory factor analytic approach was implemented to refine the item pool. The first step involved applying the Gulliksen’s Pool (G-pool) procedure, a diagnostic strategy designed to flag items with potential psychometric weaknesses based on two criteria: item extremeness and suboptimal discrimination indices (Ferrando et al., 2023). Items flagged by the G-pool were not automatically removed; rather, each flagged item was evaluated in context, and decisions regarding item retention or deletion were made based on both statistical output and theoretical relevance. The second step consisted of fitting unidimensional exploratory factor analysis (EFA) models within each community of items, using unweighted least squares estimation with Pearson correlations. In each model, items with loadings below .50 were iteratively removed—one at a time—until all remaining items had loadings greater than or equal to .50 and Horn’s parallel analysis supported the extraction of a single factor.
After this initial refinement, the retained items from all dimensions were pooled and subjected to a second EGA (with bootstrapping) and a new EFA (with parallel analysis) to verify the convergence of the factor structures across both methods. These analyses served to ensure that the dimensional configuration observed initially was consistent and replicable. A further round of item refinement was carried out based on the results of the multidimensional EFA. Items with loadings below .50 (or with loadings ≥.32 in more than one factor) were again removed, given that each factor retained at least three items. Subsequently, the expected residual correlation direct change (EREC) procedure was used to identify pairs of items (i.e., doublets) with residual dependencies that could compromise model fit (Ferrando et al., 2022). Starting with the pair showing the highest residual correlation, each doublet was evaluated for potential redundancy. If one of the items in the pair had substantially higher kurtosis or a noticeably lower factor loading than the other, it was considered for deletion. When neither criterion clearly favored one item over the other, the decision was made based on theoretical considerations. This EREC procedure was combined with the factor loading threshold of .50 to ensure that only psychometrically robust items were retained. Next, to enhance the practical utility of the resulting scale and mitigate participant fatigue while maintaining robust psychometric properties, items were also deleted in some dimensions (based on the .50 loading threshold) to ensure that all dimensions comprised a maximum of five items.
To make sure that the refined set of items continued to support the same structural configuration, a final EGA with bootstrapping was conducted. The structure identified was then subjected to a restricted factor analysis, sometimes referred to as a “confirmatory” factor analysis (CFA), but in this context used as an extension of the exploratory phase. This restricted model imposed zero loadings on all cross-loadings, allowing each item to load on only one factor, but did not assume that the structure was final or definitive (Ferrando & Lorenzo-Seva, 2000). The model was estimated using robust maximum likelihood (MLR), and model fit was assessed using conventional indices: the comparative fit index (CFI), Tucker-Lewis index (TLI), root-mean-square error of approximation (RMSEA), and standardized root-mean-square residual (SRMR). Robust versions of CFI, TLI, and RMSEA were also computed to account for potential non-normality (Brosseau-Liard et al., 2012; Brosseau-Liard & Savalei, 2014). Following standard benchmarks, values of CFI and TLI above .95, RMSEA below .06, and SRMR below .08 were interpreted as indicative of good model fit (Hu & Bentler, 1999). When necessary, model refinement was guided by modification indices and theoretical considerations. At the end of this exploratory phase, a final EGA was modeled to make sure that the item-dimension structure remained consistent.
Once the structure was finalized in Sample 1, it was evaluated in Sample 2 through a CFA using the MLR estimator. This step served to assess the replicability of the model in an independent sample. Additionally, internal consistency reliability for each factor was estimated using the omega coefficient. Measurement invariance was assessed using multi-group CFA. We examined invariance across sex (male vs. female) and age. For the latter, participants were grouped into late adolescence (17–24), early adulthood (25–34), and established adulthood (35–72) to balance sample sizes across developmental stages. We tested configural, metric, scalar, and strict invariance hierarchically. Model fit was evaluated using the Satorra-Bentler scaled χ² difference test and changes in fit indices. Given the sensitivity of χ² to large sample sizes, invariance was primarily determined by the change in CFI (ΔCFI), such that a decrease in CFI exceeding .010 was considered evidence of non-invariance (Cheung & Rensvold, 2002; Chen, 2007).
Finally, the distributional properties of each scale were visually examined with violin plots of their standardized scores. All analyses were conducted using specialized software and R packages. Specifically, for the EGA, the EGAnet package (version 2.3.0), implemented in R, was utilized. EFA was performed using the FACTOR software (version 12.06.07). Finally, for both CFA and the estimation of internal consistency reliability, the lavaan (version 0.6-19) and semTools (version 0.5-7) R packages were employed.
3. Results
3.1. Initial Screening
Initial inspection of the items revealed that the absolute kurtosis values for items 31 («I generally have a cordial relationship with my professors»), 41 («I will succeed in my chosen university career»), and 46 («Achieving a university degree will motivate me to continue my preparation») were very high (Supplementary Table 2). Consequently, these items were excluded from further analyses.
3.2. Exploratory Phase
The initial EGA identified eight item communities, which demonstrated high stability (Supplementary Figure 1). Preliminary item refinement was conducted using the G-pool procedure and a series of unidimensional EFAs. After obtaining robust unidimensional solutions (λ≥.50), a new EGA and EFA were simultaneously performed with all refined items, and both techniques converged on the same structure. However, in the EFA, one dimension (containing items related to social relationships) exhibited λ≥.50 for only two of its items. Consequently, this dimension was discarded, and the analysis was re-run. Following a new series of analyses, items were progressively eliminated due to low factor loadings, high residual correlations with other items, or to reduce scale length. This process yielded a new, nearly final solution with seven clearly defined dimensions (Supplementary Table 3). Upon examining the item content, the dimensions were named as follows: (1) academic self-efficacy, (2) family support, (3) growth expectations, (4) emotional expression, (5) expressive suppression, (6) emotional dysregulation, and (7) adaptive emotional regulation.
In the final phase of item refinement, a restricted/confirmatory factor analysis (CFA) was performed to test the derived structure. The resulting model demonstrated an acceptable fit (CFI = .95, Robust CFI = .95, TLI = .94, Robust TLI = .94, RMSEA = .04, Robust RMSEA = .04, SRMR = .03). Despite this acceptable fit, modification indices suggested potential model improvement through the inclusion of a residual correlation between items 21 (“My parents or guardians are committed to my education”) and 22 (“My parents or guardians are interested in knowing how I am doing in school”). Given that this correlation could be explained by the similar phrasing of the items, item 21 was eliminated due to its lower factor loading. No additional modifications were made, and this final model exhibited a very good fit (CFI = .96, Robust CFI = .96, TLI = .96, Robust TLI = .96, RMSEA = .03, Robust RMSEA = .04, SRMR = .03). This final selection of items was also subjected to a final EGA, yielding a highly stable solution identical to that found with factorial methods (Figure 1).

Figure 1. Exploratory Graph Analysis with the Refined Set of Items
3.3. Confirmatory Factor Analysis & Internal Consistency Reliability
Finally, the CFA performed on Sample 2 showed that the model fit the data very well: CFI = .96, Robust CFI = .96, TLI = .95, Robust TLI = .96, RMSEA = .03, Robust RMSEA = .04, SRMR = .03. As shown in Table 1, all factor loadings (except for item 71 “I feel nervous when taking an exam”.) were greater than .50, and no items with complex loadings were identified.
Table 1.
Factor Loadings of the Final Factorial Solution
| Item | Self-efficacy | Family support | Growth expectations | Emotional expression | Expressive suppression | Emotional dysregulation | Adaptive emotion regulation |
| 1 | .75 | ||||||
| 3 | .79 | ||||||
| 4 | .84 | ||||||
| 5 | .73 | ||||||
| 7 | .65 | ||||||
| 22 | .78 | ||||||
| 23 | .80 | ||||||
| 24 | .82 | ||||||
| 26 | .85 | ||||||
| 35 | .73 | ||||||
| 36 | .81 | ||||||
| 37 | .62 | ||||||
| 38 | .85 | ||||||
| 42 | .72 | ||||||
| 47 | .85 | ||||||
| 49 | .82 | ||||||
| 51 | .89 | ||||||
| 52 | .73 | ||||||
| 53 | .84 | ||||||
| 54 | .73 | ||||||
| 57 | .56 | ||||||
| 65 | .76 | ||||||
| 66 | .81 | ||||||
| 69 | .69 | ||||||
| 71 | .43 | ||||||
| 58 | .53 | ||||||
| 59 | .70 | ||||||
| 61 | .66 | ||||||
| 62 | .76 | ||||||
| 63 | .81 |
Regarding the inter-factor correlations, no instances suggesting redundancy between any two dimensions were observed (Table 2). The highest correlation was found between adaptive emotional regulation and academic self-efficacy. Furthermore, expressive suppression consistently showed low correlations with the other factors, except for an inverse correlation with emotional expression and a direct correlation with emotional dysregulation.
Table 2
Correlations Between Latent Variables of the Final Factorial Solution
| 1 | 2 | 3 | 4 | 5 | 6 | 7 | |
| 1. Self-efficacy | — | ||||||
| 2. Family support | .20 | — | |||||
| 3. Growth expectations | .37 | .25 | — | ||||
| 4. Emotional expression | .37 | .17 | .25 | — | |||
| 5. Expressive suppression | -.07 | -.02 | .04 | -.43 | — | ||
| 6. Emotional dysregulation | -.34 | -.05 | -.06 | -.22 | .31 | — | |
| 7. Adaptive emotion regulation | .59 | .20 | .36 | .44 | -.04 | -.40 | — |
Fit indices for the invariance models are presented in Table 3. Regarding age, the model demonstrated scalar (strong) invariance, as the decrease in fit from the metric to the scalar model was within acceptable limits (ΔCFI = -.005). However, strict invariance was not supported (ΔCFI = -.033). Regarding sex, the model demonstrated strict invariance. Despite significant χ² differences attributable to sample size, the changes in CFI never exceeded the -.010 threshold across all nested models. This indicates that factor structure, loadings, intercepts, and residual variances are equivalent between males and females.
Table 3
Fit Indices for Measurement Invariance Models by Age and Sex
| Model | χ² | df | CFI | RMSEA | Δχ² | Δdf | ΔCFI |
| Invariance by Age | |||||||
| Configural | 2144.9 | 1152 | .962 | .037 | — | — | — |
| Metric | 2208.45 | 1198 | .961 | .037 | 65.59* | 46 | -.001 |
| Scalar | 2380.92 | 1244 | .956 | .038 | 188.41*** | 46 | -.005 |
| Strict | 3235.54 | 1304 | .924 | .050 | 570.46*** | 60 | -.033 |
| Invariance by Sex | |||||||
| Configural | 1723.48 | 768 | .962 | .037 | — | — | — |
| Metric | 1748.61 | 791 | .962 | .037 | 27.52 | 23 | 0 |
| Scalar | 2008.5 | 814 | .953 | .040 | 306.32*** | 23 | -.009 |
| Strict | 2039.94 | 844 | .952 | .040 | 48.33* | 30 | -.001 |
| Note. ΔCFI values ≥ -0.010 indicate invariance is maintained. *p < .05, ***p < .001. | |||||||
3.5. Distributional Properties
As displayed in Figure 2, most scales had skewed distributions. Adaptive emotion regulation, emotional expression, family support, growth expectations, and self-efficacy were negatively skewed, indicating that most responses were concentrated at the upper tail of the distribution. On the other hand, emotional dysregulation was positively skewed, thus showing the opposite pattern. Finally, only expressive suppression was distributed in an approximately symmetric manner.

4. Discussion
In the present study, a 30-item battery was empirically derived to measure seven relevant socioemotional characteristics in applicants to online bachelor’s degree programs. The analyses were conducted using large applicant samples, which provided robustness and reliability to the findings. Through both exploratory and confirmatory analyses, a clear and coherent structure was identified among the evaluated dimensions. Analysis of the score distributions for the battery’s scales revealed a marked asymmetry in most scales, suggesting skewed response patterns at both extremes. This phenomenon may reflect common characteristics of the applicant population concerning their socioemotional skills. The only scale with an approximately symmetrical distribution was expressive suppression, indicating a higher proportion of responses clustered around the average scale value.
The findings from this study are consistent with recent literature on academic success in online education, which has highlighted variables such as academic self-efficacy, family support, growth expectations, and emotional regulation as key factors for student success at higher education levels (Gao et al., 2021; Mejri & Borawski, 2023; Neroni et al., 2022; Panigrahi et al., 2021). These same dimensions were identified in the battery, thereby reinforcing its conceptual soundness. The scales measure self-efficacy and family support, reflecting the importance of these constructs for understanding student responses and, consequently, for implementing retention and academic performance strategies that can support student engagement in online education (Borup et al., 2020; Kundu, 2020; Lehikko, 2021).
Furthermore, the instrument’s inclusion of growth expectations aligns with studies linking satisfaction and persistence in online programs to the perception of personal and professional development (Henry, 2020; Landrum et al., 2021). Examining other proposals such as the Social and Emotional Competencies Questionnaire (SEC-Q) analyzed by Figueroa-Varela et al. (2023) and the Socioemotional Competencies Scale (ECSE) reviewed by Vizcaíno-Escobar et al. (2025) further highlights the distinct contribution of this battery. Unlike these instruments, the present measure incorporates dimensions directly tied to the academic context—such as academic self-efficacy and growth expectations—and explicitly considers family support, enabling the assessment of social and personal resources relevant to student performance. Likewise, the differentiation between emotional expression and suppression offers a more nuanced understanding of emotional regulation in learning environments, a feature not addressed in the SEC-Q or ECSE.Similarly, the inclusion of scales on emotional suppression, regulation, and dysregulation strengthens the instrument’s capacity to capture the influence of emotions on learning, an aspect that is really important given the role of emotional management in sustaining well-being, performance, and academic engagement (Beaumont et al., 2023; Harley et al., 2019). Taken together, these elements position the battery not only as a validated psychometric tool but also as an instrument aligned with contemporary theoretical trends in distance education, providing context-specific insights into the socioemotional dynamics of university students.
A significant contribution of this research was the reduction of the number of items from 73 to 30, which preserved the instrument’s quality without compromising its psychometric properties. This shorter version facilitates its application in other contexts and enhances measurement efficiency. Another important contribution is its responsiveness to the literature and theorists in distance education (Hassan Abuhassna et al., 2022; Li, 2024; Yu et al., 2022), who emphasize the importance of gathering early information on socioemotional and contextual variables that can influence students’ adaptation and academic performance in this modality. Another significant methodological strength of this research lies in the integration of both network psychometric (i.e., EGA) and traditional factor analytical methods. This combined approach provides a more robust and comprehensive understanding of the dimensionality of the measure under study, combining the strengths of two prominent paradigms in modern measurement theory (Borsboom et al., 2022). The convergence of findings from these distinct analytical frameworks enhances our confidence in the identified dimensionality and reliability of the scales.
In relation to the social relationships dimension, its removal was based on psychometric considerations related to the heterogeneity of the items originally included. Although all items referred to relevant social agents in students’ lives—such as family members, peers, and teachers—they encompassed a mix of distinct constructs (e.g., self-efficacy, support, perception, expectations), which prevented the formation of a theoretically coherent and statistically stable dimension. Theoretically, this decision acknowledges that social relationships in educational contexts constitute a complex domain that cannot be adequately captured through a set of conceptually mixed indicators. Therefore, for future versions of the instrument, we propose re-specifying the social dimension by incorporating more homogeneous items anchored in clearly defined sub-constructs, such as social support, perceived availability of support, expectations from significant others, or interpersonal climate. This approach will enable a more precise assessment of the social environment’s contribution to the phenomenon under study, thereby strengthening the theoretical and psychometric validity of the instrument.
Regarding the limitations of this study, the cross-sectional design does not allow for the observation of temporal changes in constructs such as academic self-efficacy, social support, or emotional regulation evolve in online education. Additionally, information collected through self-report instruments may introduce biases related to social desirability or self-reflection errors, as well as potential effects inherent to this response format, which could affect the accuracy and consistency of student responses. Furthermore, given that the sample is specifically focused on a propaedeutic program at a particular institution, caution should be exercised when attempting to generalize the findings to other institutions with different characteristics and contexts.
Based on these considerations, future research should examine the predictive validity of the constructs analyzed in relation to outcomes such as academic performance or dropout. It would also be valuable to conduct multigroup invariance analyses to determine whether the measurements remain stable across different student profiles. Additionally, future studies could extend the social dimension that was excluded in this research in order to explore more deeply the role of social support in digital learning environments, as well as incorporate test–retest designs to evaluate the temporal stability of the measurements.
These findings provide evidence supporting the battery as a valid and effective tool for assessing key socioemotional variables in online learning modalities. Having a validated version of this battery for students about to commence an online course enables an accurate and adequate assessment of the emotional characteristics that influence academic performance in online and distance learning, while considering the cultural, educational, and social particularities of the country. This facilitates the early identification of risks such as academic lag and dropout. It also strengthens pedagogical decision-making and contributes to the design of institutional strategies for student support. Moreover, the validation of this tool opens the door for educational research that will allow for a deeper understanding of the conditions, strengths, and needs of students in virtual learning environments. Future studies could examine whether the evaluated dimensions act as significant predictors of academic success in distance modalities, which would facilitate the development of longitudinal and cross-sectional studies aimed at designing strategies for prevention and early intervention against risks such as academic lag and dropout.
5. Conclusions
The results of this study support the psychometric validity of a 30-item battery designed to assess seven key socioemotional dimensions in applicants to online bachelor’s degree programs. The use of robust analytical methods, including exploratory and confirmatory factor analyses combined with network psychometrics, allowed for the identification of a clear, coherent, and reliable structure. The reduction from 73 to 30 items without compromising measurement quality represents a significant contribution, enhancing the instrument’s practical application in real educational settings. The identified dimensions—such as academic self-efficacy, family support, growth expectations, and emotional regulation—are consistent with recent literature on academic success in virtual learning environments, reinforcing the conceptual soundness and relevance of the battery.
Notably, the battery should be understood as an initial diagnostic screening tool, rather than a high-stakes or exclusionary decision mechanism. Its responsible use requires complementing the socioemotional results with academic and psychoeducational indicators to ensure a more comprehensive interpretation of students’ readiness for online learning and to prevent potential biases associated with single-source assessments. Likewise, the socioemotional dimensions included in the instrument provide a valuable lens for understanding students’ initial profiles and for identifying early risk factors such as academic lag or dropout. Their relevance to the processes of adaptation and persistence in online education strengthens the broader contribution of this study to the field. Furthermore, the validation of this battery opens opportunities for future research aimed at exploring the predictive role of these dimensions in academic performance, as well as for developing longitudinal and cross-institutional studies that deepen the understanding of socioemotional readiness in digital learning environments.
Altogether, this study contributes to strengthening more inclusive, evidence-informed, and context-sensitive approaches to online education, highlighting the importance of socioemotional indicators as part of a broader institutional strategy for supporting student success.
CRedIT
ADLRG, ACRR: Conceptualization and Methodology; Writing – Original Draft; PDV: Data Curation; Formal Analysis; Writing – Original Draft; ADLRG, IOC: Conceptualization; Writing – Review & Editing; Project Administration; Resources.
AI Usage Statement
During the preparation of this work, the authors used Gemini in order to proofread the manuscript and improve its readability and language. After using this tool, the authors reviewed and edited the content as needed and took full responsibility for the content of the publication.
Data availability statement
Supplementary materials can be accessed at https://osf.io/dtepy/?view_only=00b36d83cc784db19ae6d74e97aeb24a
Funding and Acknowledgements
The authors would like to express their gratitude to Coordination for Open University and Digital Education (Coordinación de Universidad Abierta y Educación Digital, CUAED) for their valuable support.
No external funding was received for this work.
Knowledge Transfer
The validated battery from this study can be used by higher education institutions offering online programs to assess applicants’ socioemotional skills early on. Its reduced 30-item version facilitates application in admission or diagnostic processes, enabling the identification of academic risks and supporting pedagogical decisions aimed at improving student retention.
Ethical Considerations
The data used in this study were collected from institutional records and program activities. All participants provided informed consent for the use of their responses for research purposes. The study was conducted in accordance with institutional guidelines, ensuring confidentiality and the ethical handling of all data.
Referencias
Bandalos, D. L., & Finney, S. J. (2019). Factor analysis. Exploratory and confirmatory. In G. R. Hancock, L. M. Stapleton, & R. O. Mueller (Eds.), The reviewer’s guide to quantitative methods in the social sciences (2nd ed., pp. 98–122). Routledge.
Bardales, E. S., Carrasco Rituay, A. M., Marín, Y. R., Caro, O. C., Fernández, M. T., & Santos, R. C. (2025). Determinants of academic desertion: A case study in a Peruvian university. Power and Education, 0(0). https://doi.org/10.1177/17577438241312617
Beaumont, J., Putwain, D. W., Gallard, D., Malone, E., Marsh, H. W., & Pekrun, R. (2023). Students’ emotion regulation and school-related well-being: Longitudinal models juxtaposing between- and within-person perspectives. Journal of Educational Psychology, 115(7), 932–950. https://doi.org/10.1037/edu0000800
Borsboom, D., Cramer, A. O. J., Fried, E. I., Isvoranu, A.-M., Robinaugh, D. J., Dalege, J., & Van Der Maas, H. L. J. (2022). Network perspectives. In A.-M. Isvoranu, S. Epskamp, L. J. Waldorp, & D. Borsboom, Network Psychometrics with R (1st ed., pp. 9–27). Routledge. https://doi.org/10.4324/9781003111238-2
Borup, J., Graham, C. R., West, R. E., Archambault, L., & Spring, K. J. (2020). Academic Communities of Engagement: An expansive lens for examining support structures in blended and online learning. Educational Technology Research and Development, 68(2), 807–832. https://doi.org/10.1007/s11423-020-09744-x
Brosseau-Liard, P. E., & Savalei, V. (2014). Adjusting incremental fit indices for nonnormality. Multivariate Behavioral Research, 49(5), 460–470. https://doi.org/10.1080/00273171.2014.933697
Brosseau-Liard, P. E., Savalei, V., & Li, L. (2012). An investigation of the sample performance of two nonnormality corrections for RMSEA. Multivariate Behavioral Research, 47(6), 904–930. https://doi.org/10.1080/00273171.2012.715252
Chen, F. F. (2007). Sensitivity of goodness of fit indexes to lack of measurement invariance. Structural Equation Modeling, 14(3), 464–504. https://doi.org/10.1080/10705510701301834
Cheung, G. W., & Rensvold, R. B. (2002). Evaluating goodness-of-fit indexes for testing measurement invariance. Structural Equation Modeling, 9(2), 233–255. https://doi.org/10.1207/S15328007SEM0902_5
Christensen, A. P., & Golino, H. (2021). Estimating the stability of psychological dimensions via bootstrap exploratory graph analysis: A Monte Carlo simulation and tutorial. Psych, 3(3), 479–500. https://doi.org/10.3390/psych3030032
De Neve, D., Bronstein, M. V., Leroy, A., Truyts, A., & Everaert, J. (2023). Emotion regulation in the classroom: A network approach to model relations among emotion regulation difficulties, engagement to learn, and relationships with peers and teachers. Journal of Youth and Adolescence, 52(2), 273–286. https://doi.org/10.1007/s10964-022-01678-2
Delgado Celis, Z. y Miranda Díaz, G. (2021). Factores asociados a la deserción y rezago en una licenciatura en línea. En Miranda Díaz, G. Y. Delgado Celis, Z. (Coordinadores), Educación Mediada por Tecnología (pp. 3-28). Educación y Cultura Libre. https://chat.iztacala.unam.mx/sites/chat.iztacala.unam.mx/files/2021-07/EducacionMediadaTecnologia_0.pdf#page=10
Engin, M. (2017). Analysis of students’ online learning readiness based on their emotional intelligence level. Universal Journal of Educational Research, 5(12A), 32–40. https://doi.org/10.13189/ujer.2017.051306
Escanés, G., Herrero, V., Merlino, A., & Ayllón, S. (2014). Deserción en educación a distancia: factores asociados a la elección de modalidad como desencadenantes del abandono universitario. Virtualidad, Educación Y Ciencia, 5(9), 45-55. https://doi.org/10.60020/1853-6530.v5.n9.9549
Ferraces Otero, M. J., Lorenzo Moledo, M., Godás Otero, A., & Santos Rego, M. A. (2020). Students’ mediator variables in the relationship between family involvement and academic performance: Effects of the styles of involvement. Psicología Educativa, 27(1), 85–92. https://doi.org/10.5093/psed2020a19
Ferrando, P. J., Hernandez-Dorado, A., & Lorenzo-Seva, U. (2022). Detecting correlated residuals in exploratory factor analysis: New proposals and a comparison of procedures. Structural Equation Modeling: A Multidisciplinary Journal, 29(4), 630–638. https://doi.org/10.1080/10705511.2021.2004543
Ferrando, P. J., & Lorenzo-Seva, U. (2000). Unrestricted versus restricted factor analysis of multidimensional test items: Some aspects of the problem and some suggestions. Psicológica, 21(3), 301–323. https://bit.ly/40RAVNc
Ferrando, P. J., Lorenzo-Seva, U., & Bargalló-Escrivà, M. T. (2023). Gulliksen’s pool: A quick tool for preliminary detection of problematic items in item factor analysis. PLoS ONE, 18(8), e0290611. https://doi.org/10.1371/journal.pone.0290611
Figueroa-Varela, M., & Rodríguez Vera, M. (2023). Análisis de las propiedades psicométricas del Cuestionario de Competencias Socioemocionales SEC-Q en estudiantes universitarios chilenos. Revista Latinoamericana de Psicología, 55, 10-17. https://doi.org/10.14349/rlp.2023.v55.2
Gao, H., Ou, Y., Zhang, Z., Ni, M., Zhou, X., & Liao, L. (2021). The relationship between family support and e-learning engagement in college students: The mediating role of e-learning normative consciousness and behaviors and self-efficacy. Frontiers in Psychology, 12, 573779. https://doi.org/10.3389/fpsyg.2021.573779
Golino, H. F., & Epskamp, S. (2017). Exploratory graph analysis: A new approach for estimating the number of dimensions in psychological research. PLoS ONE, 12(6), e0174035. https://doi.org/10.1371/journal.pone.0174035
Harley, J. M., Pekrun, R., Taxer, J. L., & Gross, J. J. (2019). Emotion regulation in achievement situations: An integrated model. Educational Psychologist, 54(2), 106–126. https://doi.org/10.1080/00461520.2019.1587297
Hassan Abuhassna, Freed Awae, Kawthar Bayoumi, Diaya Uddeen Alzitawi, Ahmed H Alsharif, & Noraffandy Yahaya. (2022). Understanding online learning readiness among university students: A bibliometric analysis. International Journal of Interactive Mobile Technologies (iJIM), 16(13), 81–94. https://doi.org/10.3991/ijim.v16i13.30605
Heredia Ramos, M. A. (2024). Una aproximación al fenómeno de la deserción estudiantil en la Universidad Nacional Abierta ya Distancia: retos para la construcción de un modelo de permanencia. Hallazgos, 21(41), 311-342. https://doi.org/10.15332/2422409x.9603
Henry, M. (2020). Online student expectations: A multifaceted, student-centred understanding of online education. Student Success, 11(2), 91–98. https://doi.org/10.5204/ssj.1678
Hernández González, C. A., & Blackford, B. J. (2022). Engagement as antecedent of academic achievement and the moderating impact of work-family-school inter-role conflict for online graduate students. The International Journal of Management Education, 20(3), 100676. https://doi.org/10.1016/j.ijme.2022.100676
Hernández Gutiérrez, M., & Enríquez Vázquez, L. (2022). Habilidades del futuro para los estudiantes del SUAyED [Skills for the future in students at SUAyED]. Revista Digital Universitaria, 23(6). https://doi.org/10.22201/cuaieed.16076079e.2022.23.6.2
Hernández Ortiz, S. E., Edel Navarro, R., & Esquivel Gámez, I. (2023). Agentes de la educación en línea asociados con la permanencia del estudiante [Online educational agents associated with student retention]. Educación Superior, 34, 133–157. https://doi.org/10.56918/es.2022.i34.pp133-157
Hu, L., & Bentler, P. M. (1999). Cutoff criteria for fit indexes in covariance structure analysis: Conventional criteria versus new alternatives. Structural Equation Modeling: A Multidisciplinary Journal, 6(1), 1–55. https://doi.org/10.1080/10705519909540118
James, J. L. (2022). Students as stakeholders: Understanding expectations can increase student retention. Journal of College Student Retention: Research, Theory & Practice, 24(1), 20–42. https://doi.org/10.1177/1521025119898844
Kundu, A. (2020). Toward a framework for strengthening participants’ self-efficacy in online education. Asian Association of Open Universities Journal, 15(3), 351–370. https://doi.org/10.1108/AAOUJ-06-2020-0039
Landrum, B., Bannister, J., Garza, G., & Rhame, S. (2021). A class of one: Students’ satisfaction with online learning. Journal of Education for Business, 96(2), 82–88. https://doi.org/10.1080/08832323.2020.1757592
Lehikko, A. (2021). Measuring self-efficacy in immersive virtual learning environments: A systematic literature review. Journal of Interactive Learning Research, 32(2), 125–146. https://doi.org/10.70725/672759jhzbkw
Li, S. (2024). Learning analytics enhanced online learning support. Routledge.
Lim, J. R. N., Rosenthal, S., Sim, Y. J. M., Lim, Z.-Y., & Oh, K. R. (2021). Making online learning more satisfying: The effects of online-learning self-efficacy, social presence and content structure. Technology, Pedagogy and Education, 30(4), 543–556. https://doi.org/10.1080/1475939X.2021.1934102
Matteucci, M. C., & Soncini, A. (2021). Self-efficacy and psychological well-being in a sample of Italian university students with and without Specific Learning Disorder. Research in Developmental Disabilities, 110, 103858. https://doi.org/10.1016/j.ridd.2021.103858
Mejri, S., & Borawski, S. (2023). Predictors of persistence and success in online education. International Journal on E-Learning, 22(3), 239–257. https://doi.org/10.70725/807249yyafpy
Mensah, C., Kugbonu, M. A., Appietu, M. E., Nti, G. A., & Forson, M. A. (2024). Social support, computer self-efficacy, online learning engagement and satisfaction among undergraduate hospitality students. Cogent Education, 11(1), 2335803. https://doi.org/10.1080/2331186X.2024.2335803
Nadeem, A., Umer, F., & Anwar, M. J. (2023). Emotion regulation as predictor of academic performance in university students. Journal of Professional & Applied Psychology, 4(1), 20–33. https://doi.org/10.52053/jpap.v4i1.157
Neroni, J., Meijs, C., Kirschner, P. A., Xu, K. M., & De Groot, R. H. M. (2022). Academic self-efficacy, self-esteem, and grit in higher online education: Consistency of interests predicts academic success. Social Psychology of Education, 25(4), 951–975. https://doi.org/10.1007/s11218-022-09696-5
Ozer, S. (2024). Social support, self‐efficacy, self‐esteem, and well‐being during COVID-19 lockdown: A two‐wave study of Danish students. Scandinavian Journal of Psychology, 65(1), 42–52. https://doi.org/10.1111/sjop.12952
Panigrahi, R., Srivastava, P. R., & Panigrahi, P. K. (2021). Effectiveness of e-learning: The mediating role of student engagement on perceived learning effectiveness. Information Technology & People, 34(7), 1840–1862. https://doi.org/10.1108/ITP-07-2019-0380
Pekrun, R., Marsh, H. W., Suessenbach, F., Frenzel, A. C., & Goetz, T. (2023). School grades and students’ emotions: Longitudinal models of within-person reciprocal effects. Learning and Instruction, 83, 101626. https://doi.org/10.1016/j.learninstruc.2022.101626
Vizcaino-Escobar, Annia Esther, Chou-Ramírez, Leysi, Fernández-Castillo, Evelyn, Klimenko, Olena, & Otero-Ramos, Idania. (2025). Estudio exploratorio de validez y confiabilidad de la Escala de Competencias Socioemocionales en estudiantes universitarios cubanos. Revista de estudios y experiencias en educación, 24(54), 123-140. https://dx.doi.org/10.21703/rexe.v24i54.2780
Yang, D., Chen, P., Wang, K., Li, Z., Zhang, C., & Huang, R. (2023). Parental involvement and student engagement: A review of the literature. Sustainability, 15(7), 5859. https://doi.org/10.3390/su15075859
Yokoyama, S. (2024). Impact of academic self-efficacy on online learning outcomes: A recent literature review. EXCLI Journal, 23. https://doi.org/10.17179/EXCLI2024-7502
Yu, J., Huang, C., He, T., Wang, X., & Zhang, L. (2022). Investigating students’ emotional self-efficacy profiles and their relations to self-regulation, motivation, and academic performance in online learning contexts: A person-centered approach. Education and Information Technologies, 27(8), 11715–11740. https://doi.org/10.1007/s10639-022-11099-0
Yunusa, A. A., & Umar, I. N. (2021). A scoping review of Critical Predictive Factors (CPFs) of satisfaction and perceived learning outcomes in E-learning environments. Education and Information Technologies, 26(1), 1223–1270. https://doi.org/10.1007/s10639-020-10286-1