Statistical profiles of highly-rated learning objects
Introduction
Learning objects (LOs) are often defined as digital entities that can be used and reused in the process of learning and education, and are considered by many as the cornerstones for the widespread development and adoption of e-learning initiatives. Several initiatives and proposals for LO quality evaluation have been discussed in the last years (Díaz et al., 2002, Kay and Knaack, 2009, Nesbit et al., 2002, Nesbit et al., 2003, Vargo et al., 2003, Williams, 2000) nevertheless, there is still no consensus on what constitutes a good quality LO, neither which is the best way of conducting the process of evaluation. In part, this can be attributed to the heterogeneous and multi-faceted nature of these resources. As they can differ in several aspects (size, granularity, technology used, type, metadata standard, instructional design, duration, etc.) (Churchill, 2007), it is reasonable to assume that the quality criteria and the ways of measurement them will also differ accordingly to these many aspects. Moreover, the different evaluation approaches also reflect the many particular contexts of usage for the learning objects, as each one of them usually measures quality from the perspective of “a given repository, a country or a community of users” (Vuorikari, Manouselis, & Duval, 2008). In any case, the continuous growing of educational resources on the internet has turned impractical to rely only on human effort to classify good quality learning materials, and has raised the concern about the development of new automated techniques and tools that could be used to complement the existing approaches in order to relieve manual work. The actual abundance of resources inside repositories (Ochoa & Duval, 2009) and the availability of contextual evaluations in some of them have opened the possibility of seeking for intrinsic metrics of learning objects that could be used as indicators of quality. This means to say that learning objects could be “mined” and quantitative measures of good and not-good resources could be compared in order to discover intrinsic attributes associated with quality, thus allowing the creation of statistical profiles of good and poor resources that could serve as the basis for quality prediction. In fact, such approach was previously successfully applied to automatically analyze the usability of websites by Ivory and Hearst (2002b). It is known that learning object quality can be considered a more complex construct than usability as the latter is included in existing instruments as LORI (Nesbit et al., 2003) as one out of the several attributes considered, and that we cannot take for granted that the same correlations found by them are still applicable to ratings of learning objects (even though it may be hypothesized that the former affects the latter to some extent). So a first step in finding statistical profiles for highly-rated learning objects is exploring evidence on potential intrinsic measures which contribute to the classification of learning object quality, taking as a point of departure some of the ones that were identified for usability and others also found in related literature. This was initially done in the specific context of learning objects by García-Barriocanal & Sicilia (2009), where the authors preliminarily explored statistical profiles of highly-rated learning objects referenced on MERLOT repository1. In that work, the authors contrasted four basic metrics (number of links, size in bytes, number of images and number of personal collections; the last one as a factor of contrast) against the main categories of disciplines available in MERLOT (Arts, Business, Education, Humanities, Mathematics and Statistics, Science and Technology, and Social Sciences) and have found initial (but still unclear) evidence that the number of images is normally associated with the ratings of a learning object, and could consequently be considered as a possible intrinsic measure that could be used to assess quality.
Even though automated analysis cannot replace traditional inspection techniques, it carries the potential of offering an inexpensive and time saving mechanism to a priori explore the quality of materials, therefore complementing other existing approaches. This paper aims to offer the very first foundations for the development of such tool by contrasting intrinsic metrics of highly-rated and poorly-rated learning objects stored in MERLOT and identifying which metrics are mostly associated with rated resources in the context of this repository. Such metrics can further serve as possible input variables to be used inside the tool. The deployment of such automated tool would certainly improve the general quality of the services provided by the repository regarding the processes of searching, selecting and recommending good quality materials. Contributors could, for instance, benefit of such new feature by evaluating beforehand the quality of their resources, which would allow their improvement through the use of the quality metrics referenced by the tool. We believe this would positively affect their intention to contribute to the repository with new resources. Moreover, it is known that many resources included by teachers inside their virtual courses are links to external websites (González-Videgaray, Hernández-Zamora, & del-Río-Martínez, 2009), such tool would allow educators to have a complementary perspective of quality of the resources before adding them into their courses.
In here, we extend the work of García-Barriocanal & Sicilia (2009) in a number of ways. First, a much larger number of metrics are used in the analysis. Second, in the previous work the metrics were computed only for the main page2 of the resources, whereas in here we computed all internal websites up to a 2 level depth from the root node. Third, as MERLOT also classifies the materials according to their type (such as: Animation, Simulation, Drill, etc), and these different types normally present distinct features according to the literature (Churchill, 2007), we have also performed an analysis to contrast the metrics for the materials regarding this classification. And fourth, we initially tested the employment of some of these metrics in the composition of different Linear Discrimant models in order to verify their accuracy in the process of discriminating LOs regarding their quality.
The rest of this paper is structured as follows. Section 2 describes existing previous work regarding automated evaluation of LOs and some potential intrinsic measures that can be automatically extracted from the resources. Section 3 describes the data and the methodology applied for this study, as well as the statistical profiles encountered so far. Section 4 briefly explores the development of models to predict quality based on those profiles. Finally, section 5 presents the conclusions and limitations of this study, as well as some open possibilities for future work.
Section snippets
Characterizing quality and automated assessment
As mentioned before, assessing quality of learning resources is a difficult and complex task that often revolve around multiple and different aspects that must be observed. For instance, in the context of digital libraries, Custard and Sumner (2005) claim that concerns about quality are mainly related to issues of: 1) accuracy of content, 2) appropriateness to intended audience, 3) effective design, and 4) completeness of metadata documentation. In the specific field of learning multimedia
Quantitative and measurable aspects of learning objects
To the best of our knowledge, there is no empirical evidence of intrinsic metrics that are indicators of LOs’ quality, however there are some works in adjacent fields which can serve as a source of inspiration. For instance, empirical evidence of quality indicators has been found by Custard and Sumner (2005) in the field of educational digital libraries. In that work, the authors identified and computed 16 metrics for quality and trained a support vector machine model to assess resources
Predicting learning object classification
We used Linear Discriminant Analysis (LDA) to build models in order to distinguish good from not-good resources, good from average resources, and good from poor resources for the Science and Technology discipline intersected with the Simulation material type in the context of peer-reviews thresholds. This method is suitable to classify objects into one or more groups based on features that describe the objects. In order to build these models we used the 13 intrinsic metrics16
Conclusions and Outlook
In this paper we analyzed 35 measures of 1765 learning objects refereed by the MERLOT repository, and developed statistical profiles for these materials taking their associated ratings as a baseline for quality comparison. The study has presented significant contributions that can further lead to the development of contextualized models for the automated evaluation of learning objects quality inside repositories
The first important discovery is the confirmation of preliminary findings of
Acknowledgements
The results presented in this project have been partially funded by Carolina Foundation through its Mobility Program for Public Brazilian Professors, by the University of Alcalá and the CAM (Comunidad de Madrid), as part of project MARIA (code CCG08-UAH/TIC-4178) and by the Spanish Ministry of Science and Innovation through projects MAPSEL (code TIN2009-14164-C04-01) and MAVSEL (code TIN2010-21715-C02-01).
References (44)
- et al.
The design of Cloudworks: applying social networking practice to foster the exchange of learning and teaching ideas and designs
Computers & Education
(2010) - et al.
Learning objects in theory and practice: A vision from Mexican University teachers
Computers & Education
(2009) - et al.
E-Learning personalization based on hybrid recommendation strategy and learning style identification
Computers & Education
(2011) - et al.
Students’ perceptions of teaching technologies, application of technologies, and academic performance
Computers & Education
(2009) - et al.
Automatically characterizing resource quality for educational digital libraries
- et al.
Focused crawling for downloading learning objects – An Architectural perspective
Interdisciplinary Journal of E-Learning and Learning Objects
(2009) Size matters: word count as a measure of quality on wikipedia
Developing and sustaining a national learning-object sharing network: a social capital theory perspective
Project MERLOT: Bringing peer review to web-based educational resources
Journal of Technology and Teacher Education
(2006)- et al.
Analyzing associations between the different ratings dimensions of the MERLOT repository
Interdisciplinary Journal of E-Learning and Learning Objects
(2011)
Empirical analysis of Errors on human-generated learning objects metadata
Towards a useful classification of learning objects
Educational Technology Research and Development
Using machine learning to support quality Judgments
D-Lib Magazine
Evaluation of hypermedia educational systems: criteria and Imperfect measures
LearnRank: towards a real quality measure for learning
Learning and teaching styles in Engineering education
Journal of Engineering Education
Preliminary explorations on the statistical profiles of highly-rated learning objects
Exposure and support of Latent social networks among learning object repository users
Journal of the Universal Computer Science
Evaluating collaborative filtering recommender systems
ACM Transactions on Information and System Security
Statistics and causal Inference
Journal of the American Statistical Association
Improving Web site design
IEEE Internet Computing
Statistical profiles of highly-rated web sites
Cited by (33)
Would you use them? A qualitative study on teachers' assessments of open educational resources in higher education
2022, Internet and Higher EducationCitation Excerpt :These tools focus either on the evaluation of resources in online repositories, or on rubrics that offer teachers guidelines. Previous studies have offered analyses on, for example, the extent to which the selection of high-quality resources from online repositories could be supported by evaluative metadata (Abramovich & Schunn, 2012), peer reviews and user comments (Cechinel & Sánchez-Alonso, 2011; Clements & Pawlowski, 2012; Kelty, Burrus, & Baraniuk, 2008), automated analysis (Başaran, 2016; Cechinel, Sánchez-Alonso, & García-Barriocanal, 2011), or usage data (Kurilovas et al., 2011). Other studies focused on the importance of quality assurance in OER repositories, by providing quality indicators for designing effective repositories (Atenas & Havemann, 2014; Atenas, Havemann, & Priego, 2014; Clements, Pawlowski, & Manouselis, 2015).
A systematic review of design and technology components of educational digital resources
2018, Computers and EducationA Learning Quality Metadata approach: Automatic quality assessment of virtual training from metadata
2016, Computer Standards and InterfacesCitation Excerpt :Moreover, Bethard et al. [47] studied the automatic characterization of quality LO in educational digital libraries, measuring aspects such as the existence of sponsor, the clear identification of age range, and how well organized is the resource to help achieving the learning goals. LOR have also been subjected to analysis of LO quality: an example is the case of Merlot where Cechinel, Sánchez-Alonso, and García [10] conducted a study of the intrinsic characteristics of highly valued LO in this repository. All these measurable aspects of quality of LO can be classified into two groups: quantitative and qualitative.
Open educational resources repositories literature review – Towards a comprehensive quality approaches framework
2015, Computers in Human BehaviorCitation Excerpt :Adding social and collaborative features has been a recent trend of LORs to facilitate wider user engagement (Monge, Ovelar, & Azpeitia, 2008; Sánchez-Alonso, Sicilia, García-Barriocanal, Pagés-Arévalo, & Lezcano, 2011). According to previous studies (Attwell, 2005; Barton, Currier, & Hey, 2003; Clements & Pawlowski, 2012) quality of OER plays a significant role in the success of the open content repositories (LOR) (Cechinel, Sánchez-Alonso, & García-Barriocanal, 2011; Tate & Hoshek, 2009). Therefore, it is vital to study LORs quality approaches (Clements, Pawlowski, & Manouselis, 2014) in a systematic way.
Evaluating collaborative filtering recommendations inside large learning object repositories
2013, Information Processing and ManagementCitation Excerpt :At last, MERLOT also allows users to bookmark their preferred materials in Personal Collections as a way of providing personalization according to users’ individual interests (Sicilia, Sánchez-Alonso, García-Barriocanal, & Rodriguez-Garcia, 2009). Previous works have identified the information stored in the Personal Collections as potential indicators of quality inside the MERLOT repository (Cechinel, Sánchez-Alonso, & García-Barriocanal, 2011; García-Barriocanal & Sicilia, 2009). For the context of our study, it is possible to identify two potential datasets that contain user’s preferences and that are suitable for the implementation of collaborative filtering recommendations: (1) the ratings given by users, and (2) the presence of resources in users Personal Collections.
A systematic literature review on educational recommender systems for teaching and learning: research trends, limitations and opportunities
2023, Education and Information Technologies