Skip to main content

ORIGINAL RESEARCH article

Front. Psychol., 23 November 2020
Sec. Educational Psychology

Development and Psychometric Properties of the Questionnaire for Assessing Educational Podcasts (QAEP)

  • Department of Psychobiology and Behavioral Sciences Methodology, University of Málaga, Málaga, Spain

The aim of this research was to develop and validate the Questionnaire for Assessing Educational Podcasts (QAEP), an instrument designed to gather students’ views about four dimensions of educational podcasts: access and use, design and structure, content adequacy, and value as an aid to learning. In study 1 we gathered validity evidence based on test content by asking a panel of experts to rate the clarity and relevance of items. Study 2 examined the psychometric properties of the QAEP, including confirmatory factor analysis with cross-validation to test the factor structure of the questionnaire, as well as item and reliability analysis. The results from study 1 showed that the experts considered the items to be clearly worded and relevant in terms of their content. The results from study 2 showed a factor structure consistent with the underlying dimensions, as well as configural and metric invariance across groups. The item analysis and internal consistency for scores on each factor and for total scores were also satisfactory. The scores obtained on the QAEP provide teachers with direct student feedback and highlight those aspects that need to be enhanced in order to improve the teaching/learning process.

Introduction

The incorporation of information and communication technologies (ICT) into the educational field over the last decade has produced important changes in the teaching/learning process. Among these technologies, podcasts have become increasingly popular. Podcasts are digital media files comprising audio and/or video which can be automatically downloaded from the web to devices such as smartphones, PCs or MP3/4 players (O’Bannon et al., 2011; Alarcón et al., 2017; O’Connor and Andrews, 2018). Their ease of use without restrictions of time and place, coupled with rapid and free availability for most portable devices, makes them a useful tool for promoting cooperative and self-directed learning (Association of College and Research Libraries, 2000; Evans, 2008; Heilesen, 2010; Hill and Nelson, 2011; Reychav and Wu, 2015; Hargett, 2018; O’Connor et al., 2020a, b), especially in higher education (Vajoczki et al., 2010; Alarcón et al., 2017).

Research shows that podcasts have been used in a wide variety of ways in higher education (McGarr, 2009; Chester et al., 2011; Van Zanten et al., 2012; Popova et al., 2014). These include: (a) recording of face-to-face lectures that can then be used as substitutes for traditional classes or as a supplementary form of content review (Gosper et al., 2007; Lightbody et al., 2007; William and Fardon, 2007; McGarr, 2009; McKenzie, 2008; Van Zanten, 2008; Han and Klein, 2019); (b) creative or student-generated podcasts, which have attracted particular attention among researchers as they give students an active role in the learning process, helping to develop competences such as critical thinking (Frydenberg, 2008), collaborative knowledge (Lee et al., 2008), and teamwork and technological skills (Cane and Cashmore, 2008); (c) tutorials (Tynan and Colbran, 2006), showing the steps involved in a specific activity, and glossaries of key terms (Lightbody et al., 2007), which introduce key concepts of a subject; and (d) short 3–5 min podcasts, which are becoming very popular as a way of summarizing a lecture or presenting basic concepts (Lee and Chan, 2007; Abdous et al., 2012; Van Zanten et al., 2012), and which are usually used as complementary material (Bell et al., 2007; Alarcón et al., 2017; Han and Klein, 2019; Matulewicz et al., 2020).

The empirical evidence about the effectiveness and benefits of podcasting in the educational field is wide ranging. Some researchers have focused on the impact that the use of this technology has on cognitive and affective variables related to intellectual and emotional experiences, such as learning, academic performance, comprehension and anxiety, among others (Khechine et al., 2013). The results indicate that students report an improvement in their learning, better academic performance and comprehension of content, less anxiety, and greater commitment to study (Lee and Chan, 2007; Morris, 2010; Hill et al., 2012; Kennedy and Thomas, 2012; Popova et al., 2014; Rockhill et al., 2019).

Student satisfaction with podcasts is another indicator used to assess the effectiveness of this tool (Vajoczki et al., 2010). Student satisfaction refers to the “favourability of a student’s subjective evaluation of the various outcomes and experiences associated with education” (Elliott and Shin, 2002: 198). In order to evaluate this variable, Alarcón et al. (2017) designed the Student Satisfaction with Educational Podcasts Questionnaire (SSEPQ). This questionnaire consists of 10 Likert-type items, each with four response options, which assess satisfaction with regard to perceived content adequacy, ease of use, usefulness and benefits to learning from the podcast. The construct validity analysis indicated a single factor, such that the total score on the questionnaire provides a general index of students’ satisfaction with podcasts. As part of the validation study the authors applied the questionnaire to psychology undergraduates and found that they were highly satisfied with the use of short 3–5 min podcasts in a Methods and Statistics course (Alarcón et al., 2017).

The main advantage of the SSEPQ is its brevity and its ability to provide teachers with direct and general feedback from students. However, it does not consider specific aspects related to technical features, content adequacy or the design and structure of podcasts. Some authors have emphasized the importance of evaluating the information available through media, but have also pointed out the lack of agreement regarding quality indicators for assessing podcasts (Paterson et al., 2015; Singh et al., 2016; Kaahwa et al., 2019). Although studies have identified a variety of indicators, related mainly to credibility, content or design (Lin et al., 2015; Paterson et al., 2015; Thoma et al., 2015), these indicators were designed for external peer or expert assessment, rather than for evaluation by podcast users, such as students. Consequently, there is need for a questionnaire that assesses the educational podcasts used by students, and which could provide teachers with useful information for improving their learning tools. The purpose of this study was to develop and to examine the psychometric properties of the Questionnaire for Assessing Educational Podcasts (QAEP), an instrument that considers four dimensions of these podcasts: access and use, design and structure, content adequacy, and value as an aid to learning. The research involved two studies. In study 1 we gathered validity evidence based on test content (content validity) by asking a panel of experts to rate the clarity and relevance of items. Study 2 involved a comprehensive psychometric analysis of the QAEP, including confirmatory factor analysis (CFA) with cross-validation to examine the factor structure of the questionnaire, as well as calculation of corrected item-factor correlations and corrected item-total correlations and reliability testing. Finally, we analyzed and interpreted the scores obtained when administering the questionnaire to a sample of psychology undergraduates.

Study 1: Design of the Questionnaire and Validity Evidence Based on Test Content

Design of the Questionnaire

We began by conducting focus groups with students in order to determine the aspects of educational podcasts that should be assessed. This process identified four main dimensions to assess: access and use, design and structure, content adequacy, and value as an aid to learning. Next, we drew up a battery of items as indicators of each dimension, taking into account the rules for writing items (Muñiz and Fonseca-Pedrero, 2019) such as representativeness, relevance, specificity, clarity, brevity, simplicity, and comprehensibility. These items were then submitted to further focus group discussion, this time with a panel of experts who were involved in teaching innovation projects and who had experience of teaching or supporting teaching on measurement and research methods courses for undergraduates.

The initial questionnaire consisted of 24 items with a 4-point Likert-type response format (1: strongly disagree, 4: strongly agree), covering the above mentioned dimensions of educational podcasts: (a) access and use (5 items), referring to the ease in locating and accessing podcasts, as well as their use on different devices and in different places (e.g., It was easy to access the podcasts; I was able to view the podcasts on different devices); (b) Design and structure (6 items), that is, whether the display format (audio, video, and design), the synchronization between audio and video and the length of the podcasts are adequate (e.g., The length of the podcasts is appropriate for understanding their content; The audio and video are properly synchronized); (c) Content adequacy (6 items), which refers to whether the content is presented clearly and whether the information is accurate and adequately represents the topic being studied (e.g., The content of the podcasts is relevant to the subject; The examples used in the podcasts are appropriate); and (d) Value as an aid to learning (7 items), that is, whether the podcasts help to improve understanding and reinforce content, increase students’ motivation to study and encourage independent learning (e.g., The podcasts were a good aid to learning about the subject; The podcasts gave me a better understanding of the subject content). Higher scores on these factors indicate greater agreement with the dimension being assessed, that is, that the podcasts have been easy to use and access, have an adequate structure and are well designed, have adequate content and are useful for learning, respectively.

Procedure

The analysis of validity evidence based on test content was focused on the domain relevance, which refers to the extent to which each item on a test is relevant to the targeted domain (Sireci and Faulkner-Bond, 2014). This requires a panel of expert judges, who rate the relevance of the items according to established criteria (Osterlind, 1989). Here we used the blind protocol, in which the judges are given the items and the domains without any indications of which item is meant to be matched with which domain (Hambleton, 1980; Dunn et al., 1999). The judges rated the relevance of each item to each domain on a 5-point scale (1: low degree of relevance; 5: high degree of relevance). The protocol also incorporated a 5-point scale in order to rate the clarity of the item, with higher scores indicating greater clarity.

Six judges (4 men and 2 women) assessed the relevance of the items to the domains. They were aged between 34 and 55 years (M = 44, SD = 8.25) and had between 11 and 32 years of professional experience. All six were experts in educational innovation and podcasting and had experience of teaching in higher education. Three judges were lecturers in the Faculty of Education, while the other three were from the Faculty of Psychology and were also experts in measurement and research methodology. None of the authors of the present paper participated as a judge and none of the judges was involved in the course in which the podcasts were used.

The protocol was sent via e-mail to the panel of experts. In order to provide a context for their task, the protocol contained a definition of the test domains, as well as instructions for completing the task. The experts were also asked to provide data on age, gender, professional profile, and years of experience.

Data Analysis

We calculated the means for item clarity, considering as adequate those items with a score of 3 or more (out of 5).

In order to assess the validity evidence based on test content, we calculated the mean scores for item relevance, along with Aiken’s V index (Aiken, 1980, 1985) and its 95% confidence interval. The V index summarizes the ratings of item content relevance obtained from a panel of experts and ranges from 0 (disagreement) to 1 (perfect agreement). We considered a V index associated with the theoretical dimension of 0.70 as satisfactory (Charter, 2003). Penfield and Giacobbi (2004) proposed the calculation of confidence intervals as a means of testing the null hypothesis that V is equal to the pre-established cut-off point. Based on this criterion, items were considered to have an adequate degree of relevance if the V index was above this cut-off and the 95% confidence interval did not include the value 0.70.

Data analysis was performed with SPSS v24. The program created by Merino and Livia (2009) was used to compute the confidence intervals of the V index.

Results

Means for item clarity were above 3.30 in all cases, indicating that the experts considered them to be clearly worded. However, four items had a V index with a 95% confidence interval that included the value 0.70: It was easy to see the podcasts, V = 0.75 [0.55–0.88]; The content of the podcasts is presented in a logical order, V = 0.71 [0.51–0.85]; The podcasts provide clear information about the topic in question, V = 0.79 [0.59–0.91]; and The podcasts complement the face-to-face lectures, V = 0.75 [0.55–0.88]. These four items were therefore eliminated. The remaining 20 items yielded a V index above 0.70 for the theoretical dimension and the confidence interval did not include this value. The mean relevance rating was also above 4 for all these items, indicating that the experts considered them to be relevant to the theoretical dimension. In addition, all these items had a V index below 0.40 on the dimensions to which they did not belong. These 20 items formed the final version of the QAEP. The results are shown in Table 1.

TABLE 1
www.frontiersin.org

Table 1. Means for clarity and relevance, and values of the V index and its 95% confidence interval.

Study 2. Psychometric Analysis of the QAEP

We proceeded to obtain validity evidence based on the instrument’s internal structure. First, we tested the proposed factor structure (construct validity) of four first-order factors and one second-order factor. Subsequently, we carried out an analysis of items and of the reliability of the factors and total score. Finally, we performed descriptive analyses of QAEP scores.

Materials and Methods

Sample

Participants were 245 students (68 males and 177 females) aged between 18 and 54 years (M = 21.22, SD = 6.32) who were enrolled in the Research Methods and Statistics course that is offered during the first year of the Degree in Psychology at the University of Malaga (Spain). They all used the podcasts at least once during the academic year. In order to analyze the factor structure the total sample was split into two randomized sub-samples: the first (the calibration sample) consisted of 136 individuals (35 males and 101 females) with an average age of 22.57 years (SD = 7.72), while the second (the validation sample) comprised 109 individuals (33 males and 76 females) with an average of 22.06 years (SD = 7.17).

Instrument

The QAEP described in study 1 was administered. We expected the QAEP to show four first-order factors, in accordance with its theoretical dimensions (access and use, design and structure, content adequacy, and value as an aid to learning), and one second-order factor that subsumes these factors and which supports the use of a total score for the assessment of educational podcasts.

For this second study we used 11 educational podcasts related to different topics covered by the aforementioned Research Methods and Statistics course. The podcasts were designed by the authors and created in audio/video format using Microsoft PowerPoint, Audacity® and Camtasia® software. Each podcast presented theoretical and practical content related to the main topics covered by the course syllabus (e.g., introduction to statistical inference, Type I and Type II error, parametric and non-parametric tests, etc.), the purpose being to provide complementary material. The length of the podcasts was 3–5 min, beginning with a short summary of the contents, followed by a step-by-step guide to performing statistical analysis with IBM SPSS and how to interpret the results obtained.

Educational podcasts were uploaded to the course’s virtual campus and were available to students throughout the semester. They were also used in practical classes during the course, thus ensuring that all participants engaged with them at least once during the academic year.

Procedure

The QAEP was administered to students on the day of the final exam of the Research Methods and Statistics course. Participants were asked to provide basic personal data (code number, age, and gender) and were informed that all data were anonymous and would be used exclusively for research purposes.

Data Analysis

The internal structure of the QAEP (construct validity) was analyzed by means of CFA. A cross-validation strategy was employed using the two randomized samples described above. In the calibration sample (n = 136) we tested a model comprising four first-order factors and one second-order factor. To verify the factor structure underlying the QAEP we then checked the fit of this model in the validation sample (n = 109), applying several measures of covariance structure equivalence. Configural invariance was examined to establish whether the number of factors and factor-loading patterns were the same across groups, constraining the factor structure to be equal across the two groups (configural model). We then examined metric invariance in order to test equality with respect to the first-order and second-order factor loadings across groups. This analysis was carried out in two steps: in the first, the model was tested by constraining all first-order factor loadings to be equal, while in the second the model was tested by constraining second-order factor loadings to be equal across groups.

Confirmatory factor analyses were performed via structural equation modeling, using the EQS 6.3 software package (Bentler, 2006) with the maximum likelihood and robust estimation methods and the polychoric correlation matrix of items. The Satorra-Bentler chi-square (χS-B2) was computed with the following goodness-of-fit indices (Bentler, 2006): the non-normed fit index (NNFI; Bentler and Bonett, 1980), the comparative fit index (CFI; Bentler, 1990), the root mean square error of approximation (RMSEA; Browne and Cudeck, 1993; Steiger, 2000) and the 90% confidence interval for the RMSEA. Values of the NNFI and CFI close to or greater than 0.95 are indicative of a good fit (Hu and Bentler, 1999). Values of the RMSEA less than 0.08 indicate a reasonable fit (Browne and Cudeck, 1993) and those less than 0.06 represent a good fit (Hu and Bentler, 1999). Because the chi-square test to compare the fit of the nested models is sensitive to sample size, the configural and metric invariance was assessed by comparing the CFI values, as recommended by Cheung and Rensvold (2002). It is considered that the constraints are tenable if the decrease in CFI is less than or equal to 0.01 between the most constrained model and the configural model.

Next, we carried out item analyses by computing corrected item-factor correlations and corrected item-total correlations. The score of each respective item was eliminated when computing the corresponding corrected correlation. Values greater than 0.30 are considered satisfactory (De Vaus, 2002). We also obtained evidence of reliability (internal consistency) by calculating Cronbach’s alpha coefficients for scores on each factor and the total score on the QAEP.

Finally, a descriptive analysis was carried out in order to obtain mean scores for each dimension of the QAEP, thus providing a measure of students’ views about the podcasts. The factor score was calculated by summing scores for the items that load on each factor, while the total score was calculated as the sum of all 20 item responses. Item, reliability and descriptive analyses were all computed using IBM SPSS v24.

Results

The factor structure of the QAEP based on four first-order factors and one second-order factor was tested in the calibration sample and the validation sample. The results showed a good fit in both samples. We then calculated the goodness-of-fit indices related to the test for multigroup configural invariance, all of which indicated a good fit. The goodness-of-fit indices related to the equality of the first-order and second-order factor loadings were also satisfactory. Furthermore, the CFI showed no decrease from the configural model to the model with first-order and second-order factor loadings constrained to be equal across groups. These results indicate that the QAEP has a stable structure across groups. Table 2 shows the fit indices for the cross-validation strategy and Table 3 the factor loadings, all of which are statistically significant.

TABLE 2
www.frontiersin.org

Table 2. Fit indices for the second-order factor model of the QAEP for the calibration and validation samples and measurement invariance tests across samples.

TABLE 3
www.frontiersin.org

Table 3. Factor loadings, corrected item-factor correlations, item-total correlations and Cronbach’s alpha coefficient for each factor.

Table 3 also shows the corrected item-factor correlations and corrected item-total correlations, with all values being above 0.30. The Cronbach’s alpha coefficients for scores on each factor were above 0.70, and the value of alpha for the total score was 0.92.

Having verified that the QAEP had adequate psychometric properties, we then performed a descriptive analysis in order to obtain a measure of students’ views about the educational podcasts used in the Research Methods and Statistics course. Table 4 shows the descriptive statistics obtained. Means for all items were above 3 (out of 4).

TABLE 4
www.frontiersin.org

Table 4. Mean (M) and standard deviation (SD) for items, factors and total score of the QAEP.

Discussion

This paper describes the design and initial evaluation of the Questionnaire for Assessing Educational Podcasts (QAEP) (study 1), followed by an exhaustive analysis of its psychometric properties (study 2). The questionnaire comprises 20 items covering four dimensions of educational podcasts: access and use, design and structure, content adequacy, and value as an aid to learning.

Validity evidence based on test content was obtained by asking a panel of experts to assess the clarity and relevance of items. To this end we used the blind protocol, such that the judges did not know which items were matched to which theoretical domain (Hambleton, 1980; Dunn et al., 1999). This approach is considered superior to the non-blind protocol. Based on the results we eliminated four items that did not fulfill the inclusion criterion. The remaining 20 items that formed the QAEP were consistent with the theoretical dimension and also relevant in terms of their content. This indicates that the items of the QAEP adequately represent the dimensions underlying the questionnaire. The experts also considered that these items were clearly worded.

In accordance with the proposed theoretical structure we then used CFA and a cross-validation strategy to test a model based on four first-order factors and one second-order factor. The results showed that the proposed structure fitted the data adequately and that the QAEP had a stable structure across groups, with configural and metric invariance. The second-order factor supports the use of a total score as a measure of students’ views about educational podcasts.

Finally, we conducted a reliability analysis. Cronbach’s alpha coefficients for scores on each factor were above 0.70, and the value of alpha for the total score was 0.92. The corrected item-factor correlations and corrected item-total correlations were also satisfactory.

Having confirmed that the QAEP shows adequate psychometric properties, we then obtained descriptive data for each of its dimensions, administering the questionnaire to a sample of psychology undergraduates. Regarding the dimension access and use, the mean score of 14.33 (out of 16) indicates that students considered the podcasts to have been easy to use and access, and that they could use them in different places and on different devices. This reflects a recognized advantage of podcasts as educational tools, namely the possibility of viewing them as often as is wished and on any device (Evans, 2008; Heilesen, 2010; Vajoczki et al., 2010; Hill and Nelson, 2011; Reychav and Wu, 2015; Alarcón et al., 2017). In relation to the dimension design and structure, the mean score of 17.47 (out of 20) indicates that students were positive about the length of the podcasts and the display format (audio, video, and design). The highest rating (mean of 14.60 out of 16) was obtained for the content adequacy dimension. This indicates that the podcasts were considered to be clear in their presentation and that the information they contained was accurate and provided an adequate summary of the topics addressed. Finally, the mean score of 24.16 (out of 28) on the value as an aid to learning dimension suggests that students felt that the podcasts had facilitated their learning of the subject, reinforcing their understanding, increasing their motivation to study and encouraging independent learning. The mean total score of 70.56 (out of 80) likewise shows that the students were very positive about the educational podcasts.

The limitations of this study are that all the participants were psychology undergraduates on a Research Methods and Statistics course and that the evaluation was based on just 11 podcasts used during this course. This may restrict the generalizability of results. Future studies should therefore analyze the applicability of the QAEP to different courses and disciplines (e.g., engineering, medicine, etc.), as well as to different educational levels such as baccalaureate or secondary school.

Conclusion

The QAEP is a short and easy-to-administer questionnaire for exploring students’ views about educational podcasts, specifically as regards ease of access and use, structure and design, content adequacy, and value as an aid to learning. The scores obtained provide teachers with direct student feedback and highlight those aspects that need to be enhanced in order to improve the teaching/learning process.

Data Availability Statement

All datasets presented in this study are included in the article/supplementary material.

Ethics Statement

The studies involving human participants were reviewed and approved by the Experimentation Ethics Committee of University of Malaga. The patients/participants provided their written informed consent to participate in this study.

Author Contributions

All authors listed have made a substantial, direct and intellectual contribution to the work, and approved it for publication.

Funding

The research reported in this paper was supported by a Teaching Innovation Project (PIE17-012), funded by the University of Malaga.

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

Abdous, M., Facer, B. R., and Yen, C. (2012). Academic effectiveness of podcasting: a comparative study of integrated versus supplemental use of podcasting in second language classes. Comput. Educ. 58, 43–52. doi: 10.1016/j.compedu.2011.08.021

CrossRef Full Text | Google Scholar

Aiken, L. (1980). Content validity and reliability of single items or questionnaire. Educ. Psychol. Meas. 40, 955–959. doi: 10.1177/001316448004000419

CrossRef Full Text | Google Scholar

Aiken, L. (1985). Three coefficients for analyzing the reliability and validity of ratings. Educ. Psychol. Meas. 45, 131–142. doi: 10.1177/0013164485451012

CrossRef Full Text | Google Scholar

Alarcón, R., Blanca, M. J., and Bendayan, R. (2017). Student satisfaction with educational podcasts questionnaire. Escritos de Psicología 10, 126–133. doi: 10.5231/psy.writ.2017.14032

CrossRef Full Text | Google Scholar

Association of College and Research Libraries (2000). Information Literacy Competency Standards for Higher Education. Available online at: http://www.ala.org/ala/acrl/acrlstandards/informationliteracycompetency.htm (accessed January 20, 2020).

Google Scholar

Bell, T. M., Cockburn, A., Wingkvist, A., and Green, R. (2007). “Podcasts as a supplement in tertiary education: an experiment with two Computer Science courses,” in Conference on Mobile Learning Technologies and Applications, (Auckland: Massey University), 70–77.

Google Scholar

Bentler, P. M. (1990). Comparative fit indexes in structural models. Psychol. Bull. 107, 238–246. doi: 10.1037/0033-2909.107.2.238

PubMed Abstract | CrossRef Full Text | Google Scholar

Bentler, P. M. (2006). EQS 6 Structural Equations Program Manual. Encino, CA: Multivariate Software, Inc.

Google Scholar

Bentler, P. M., and Bonett, D. G. (1980). Significance tests and goodness of fit in the analysis of covariance structures. Psychol. Bull. 88, 588–606. doi: 10.1037/0033-2909.88.3.588

CrossRef Full Text | Google Scholar

Browne, M. W., and Cudeck, R. (1993). “Alternative ways of assessing model fit,” in Testing Structural Equation Models, eds K. Bollen and J. Long (Newbury Park, CA: Sage), 136–162.

Google Scholar

Cane, C., and Cashmore, A. (2008). “Students’ podcasts as learning tools,” in Podcasting for Learning in Universities, eds G. Salmon and P. Edirisingha (Maidenhead: Open University Press), 146–152.

Google Scholar

Charter, R. A. (2003). A breakdown of reliability coefficients by test type and reliability method, and the clinical implications of low reliability. J. Gen. Psychol. 130, 290–304. doi: 10.1080/00221300309601160

PubMed Abstract | CrossRef Full Text | Google Scholar

Chester, A., Buntine, A., Hammond, K., and Atkinson, L. (2011). Podcasting in education: student attitudes, behaviour and self-efficacy. Educ. Technol. Soc. 14, 236–247.

Google Scholar

Cheung, G. W., and Rensvold, R. B. (2002). Evaluating goodness-of-fit indexes for testing measurement invariance. Struct. Equat. Model. 9, 233–255. doi: 10.1207/S15328007SEM0902_5

CrossRef Full Text | Google Scholar

De Vaus, D. (2002). Analyzing Social Science Data: 50 Key Problems in Data Analysis. Thousand Oaks, CA: Sage.

Google Scholar

Dunn, J. G. H., Bouffard, M., and Rogers, W. T. (1999). Assessing item content-relevance in sport psychology scale-construction research: issues and recommendations. Meas. Phys. Educ. Exer. Sci. 3, 15–36. doi: 10.1207/s15327841mpee0301_2

CrossRef Full Text | Google Scholar

Elliott, K. M., and Shin, D. (2002). Student satisfaction: an alternative approach to assessing this important concept. J. High. Educ. Pol. Manag. 24, 197–209. doi: 10.1080/1360080022000013518

CrossRef Full Text | Google Scholar

Evans, C. (2008). The effectiveness of m-learning in the form of podcast revision lectures in higher education. Comput. Educ. 50, 491–498. doi: 10.1016/j.compedu.2007.09.016

CrossRef Full Text | Google Scholar

Frydenberg, M. (2008). Principles and pedagogy: the two Ps of podcasting in the information technology classroom. Inform. Syst. Educ. J. 6, 3–11.

Google Scholar

Gosper, M., McNeill, M., Woo, K., Phillips, R., Preston, G., and Green, D. (2007). Web-Based Lecture Recording Technologies: Do Students Learn From Them? Paper presented at Educause Australasia 2007, Melbourne. Available online at: http://www.cpd.mq.edu.au/teaching/wblt/dissemination.htm (accessed April 15, 2020).

Google Scholar

Hambleton, R. K. (1980). “Test score validity and standard-setting methods,” in Criterion-Referenced Measurement, ed. R. A. Berk (Baltimore, MD: Johns Hopkins Press), 80–123.

Google Scholar

Han, E., and Klein, K. C. (2019). Pre-class learning methods for flipped classrooms. Am. J. Pharm. Educ. 83, 40–49. doi: 10.5688/ajpe6922

PubMed Abstract | CrossRef Full Text | Google Scholar

Hargett, J. L. (2018). Podcasting in nursing education: using commercially prepared podcasts to spark learning. Teach. Learn. Nurs. 13, 55–57. doi: 10.1016/j.teln.2017.08.003

CrossRef Full Text | Google Scholar

Heilesen, S. B. (2010). What is the academic efficacy of podcasting? Comput. Educ. 55, 1063–1068. doi: 10.1016/j.compedu.2010.05.002

CrossRef Full Text | Google Scholar

Hill, J., Nelson, A., France, D., and Woodland, W. (2012). Integrating podcast technology effectively into student learning: a reflexive examination. J. Geogr. High. Educ. 36, 437–454. doi: 10.1080/03098265.2011.641171

CrossRef Full Text | Google Scholar

Hill, J. L., and Nelson, A. (2011). New technology, new pedagogy? Employing video podcasts in learning and teaching about exotic ecosystems. Environ. Educ. Res. 17, 393–408. doi: 10.1080/13504622.2010.545873

CrossRef Full Text | Google Scholar

Hu, L., and Bentler, P. M. (1999). Cutoff criteria for the indexes in covariance structure analysis: conventional criteria versus new alternatives. Struct. Equat. Model. 6, 1–55. doi: 10.1080/10705519909540118

CrossRef Full Text | Google Scholar

Kaahwa, M., Zhu, C., and Muhumuza, M. (2019). Satisfaction of distance education university students with the use of audio media as a medium of instruction. The case of Mountains of the Moon University, Uganda. Int. J. Educ. Pedagog. Sci. 13, 737–750. doi: 10.5281/ZENODO.3298920

CrossRef Full Text | Google Scholar

Kennedy, M. J., and Thomas, C. N. (2012). Effects of content acquisition podcasts to develop preservice teachers’ knowledge of positive behavioral interventions and supports. Exceptionality 20, 1–19. doi: 10.1080/09362835.2011.611088

CrossRef Full Text | Google Scholar

Khechine, H., Lakhal, S., and Pascot, D. (2013). University students’ perception of the pedagogical use of podcasts: a case study of an online information system course. J. Educ. Train. Stud. 1, 136–151. doi: 10.11114/jets.v1i2.139

CrossRef Full Text | Google Scholar

Lee, M. J. W., and Chan, A. (2007). Reducing the effects of isolation and promoting inclusivity for distance learners through podcasting. TOJDE 8, 85–104.

Google Scholar

Lee, M. J. W., McLoughlin, C., and Chan, A. (2008). Talk the talk: Learner-generated podcasts as catalysts for knowledge creation. Br. J. Educ. Technol. 39, 501–521. doi: 10.1111/j.1467-8535.2007.00746.x

CrossRef Full Text | Google Scholar

Lightbody, G., McCullagh, P., Hughes, J., and Hutchison, M. (2007). “The use of audio podcasts to enhance the delivery of a computer networks course,” in Paper presented at the Higher Education Academy’s 8th Annual Information and Computer Science (HEA-ICS) Conference, (Hampshire: University of Southampton).

Google Scholar

Lin, M., Thoma, B., Trueger, N. S., Ankel, F., Sherbino, J., and Chan, T. (2015). Quality indicators for blogs and podcasts used in medical education: modified Delphi consensus recommendations by an international cohort of health professions educators. Postgrad. Med. J. 91, 546–560. doi: 10.1136/postgradmedj-2014-133230

PubMed Abstract | CrossRef Full Text | Google Scholar

Matulewicz, A. T., Hammond, V., Patterson, J. A., Frankart, L. M., and Donohoe, K. L. (2020). Utilizing widely available podcasts to create a reflection activity for pharmacy students. Curr. Pharm. Teach. Learn. 12, 1215–1223. doi: 10.1016/j.cptl.2020.05.003

CrossRef Full Text | Google Scholar

McGarr, O. (2009). A review of podcasting in higher education: its influence on the traditional lecture. Aust. J. Educ. Technol. 25, 309–321. doi: 10.14742/ajet.1136

CrossRef Full Text | Google Scholar

McKenzie, W. A. (2008). “Where are audio recordings of lectures in the new educational technology landscape?,” in Hello! Where are You in the Landscape of educational technology? Proceedings of ASCILITE, (Melbourne, VIC: ASCILITE), 628–632.

Google Scholar

Merino, C., and Livia, J. (2009). Intervalos de confianza asimétricos para el índice la validez de contenido: un programa Visual Basic para la V de Aiken [Asymmetric confidence intervals for the content validity index: a Visual Basic program for Aiken’s V]. Anales de Psicología 25, 169–172.

Google Scholar

Morris, N. (2010). Podcasts and mobile assessment enhance student learning experience and academic performance. Biosci. Educ. 16, 1–7. doi: 10.3108/beej.16.1

PubMed Abstract | CrossRef Full Text | Google Scholar

Muñiz, J., and Fonseca-Pedrero, E. (2019). Diez pasos para la construcción de un test [A ten-step guide to test construction]. Psicothema 31, 7–16. doi: 10.7334/psicothema2018.291

PubMed Abstract | CrossRef Full Text | Google Scholar

O’Bannon, B. W., Lubke, J. K., Beard, J. L., and Britt, V. (2011). Using podcasts to replace lecture: effects on student achievement. Comput. Educ. 57, 1885–1892. doi: 10.1016/j.compedu.2011.04.001

CrossRef Full Text | Google Scholar

O’Connor, S., and Andrews, T. (2018). Smartphones and mobile applications (apps) in clinical nursing education: a student perspective. Nurse Educ. Today 69, 172–178. doi: 10.1016/j.nedt.2018.07.013

PubMed Abstract | CrossRef Full Text | Google Scholar

O’Connor, S., Chu, C. H., Thilo, F., Lee, J. J., Mather, C., and Topaz, M. (2020a). Professionalism in a digital and mobile world: a way forward for nursing. J. Adv. Nurs. Virtual Issue 76, 4–6. doi: 10.1111/jan.14224

PubMed Abstract | CrossRef Full Text | Google Scholar

O’Connor, S., Daly, C. S., MacArthur, J., Borglin, G., and Booth, R. G. (2020b). Podcasting in nursing and midwifery education: an integrative review. Nurse Educ. Pract. 47:102827. doi: 10.1016/j.nepr.2020.102827

PubMed Abstract | CrossRef Full Text | Google Scholar

Osterlind, S. J. (1989). Constructing Test Items. Boston, MA: Kluwer Academic Publishers.

Google Scholar

Paterson, Q., Thoma, B., Milne, W. K., Lin, M., and Chan, T. M. (2015). A systematic review and qualitative analysis to determine quality indicators for health professions education blogs and podcasts. J. Grad. Med. Educ. 7, 549–554. doi: 10.4300/JGME-D-14-00728.1

PubMed Abstract | CrossRef Full Text | Google Scholar

Penfield, R. D., and Giacobbi, P. R. (2004). Applying a score confidence interval to Aiken’s item content-relevance index. Meas. Phys. Educ. Exer. Sci. 8, 213–225. doi: 10.1207/s15327841mpee0804_3

CrossRef Full Text | Google Scholar

Popova, A., Kirschner, P. A., and Joiner, R. (2014). Effects of primer podcasts on stimulating learning from lectures: how do students engage? Br. J. Educ. Technol. 45, 330–339. doi: 10.1111/bjet.12023

CrossRef Full Text | Google Scholar

Reychav, I., and Wu, W. (2015). Mobile collaborative learning: the role of individual learning in groups through text and video content delivery in tablets. Comput. Hum. Behav. 50, 520–534. doi: 10.1016/j.chb.2015.04.019

CrossRef Full Text | Google Scholar

Rockhill, C., Pastore, D., and Johnston, D. (2019). The effectiveness of podcasts in sport management education. J. Hosp. Leisure Sport Tour. Educ. 25:100211. doi: 10.1016/j.jhlste.2019.100211

CrossRef Full Text | Google Scholar

Singh, D., Alam, F., and Matava, C. (2016). A critical analysis of anesthesiology podcasts: identifying determinants of success. JMIR Med. Educ. 2:e14. doi: 10.2196/mededu.5950

PubMed Abstract | CrossRef Full Text | Google Scholar

Sireci, S., and Faulkner-Bond, M. (2014). Validity evidence based on test content. Psicothema 26, 100–107. doi: 10.7334/psicothema2013.256

PubMed Abstract | CrossRef Full Text | Google Scholar

Steiger, J. H. (2000). Point estimation, hypothesis testing, and interval estimation using the RMSEA: some comments and a reply to Hayduk and Glaser. Struct. Equat. Model. 7, 149–162. doi: 10.1207/S15328007SEM0702_1

CrossRef Full Text | Google Scholar

Thoma, B., Chan, T. M., Paterson, Q. S., Milne, W. K., Sanders, J. L., and Lin, M. (2015). Emergency medicine and critical care blogs and podcasts: establishing an international consensus on quality. Ann. Emerg. Med. 66, 396–402. doi: 10.1016/j.annemergmed.2015.03.002

PubMed Abstract | CrossRef Full Text | Google Scholar

Tynan, B., and Colbran, S. (2006). “Podcasting, student learning and expectations,” in Proceedings of the ASCILITE Conference, eds L. Markauskaite, P. Goodyear, and P. Reimann (Sydney, NSW: ASCILITE), 825–832.

Google Scholar

Vajoczki, S., Watt, S., Marquis, N., and Holshausen, K. (2010). Podcasts: are they an effective tool to enhance student learning? a case study from McMaster University, Hamilton Canada. J. Educ. Multimedia Hypermedia 19, 349–362.

Google Scholar

Van Zanten, R. (2008). “The value of lecture podcasting for distance and on-campus students,” in Proceedings of the ASCILITE Conference, (Melbourne, VIC: ASCILITE), 1066–1070.

Google Scholar

Van Zanten, R., Somogyi, S., and Curro, G. (2012). Purpose and preference in educational podcasting. Br. J. Educ. Technol. 43, 130–138.

Google Scholar

William, J., and Fardon, M. (2007). “Perpetual connectivity: lecture recordings and portable media players,” in Proceedings of ASCILITE, eds R. J. Atkinson and C. McBeath (Singapore: ASCILITE), 1084–1092.

Google Scholar

Keywords: podcasts, cross-validation, reliability analysis, validity evidence, feedback

Citation: Alarcón R and Blanca MJ (2020) Development and Psychometric Properties of the Questionnaire for Assessing Educational Podcasts (QAEP). Front. Psychol. 11:579454. doi: 10.3389/fpsyg.2020.579454

Received: 02 July 2020; Accepted: 14 September 2020;
Published: 23 November 2020.

Edited by:

Michael S. Dempsey, Boston University, United States

Reviewed by:

Dick Ng’ambi, University of Cape Town, South Africa
Juan De Dios Benítez Sillero, University of Córdoba, Spain

Copyright © 2020 Alarcón and Blanca. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Rafael Alarcón, ralarcon@uma.es

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.