Next Article in Journal
Peer Assessment: Channels of Operation
Next Article in Special Issue
The Effect of Digital Device Usage on Student Academic Performance: A Case Study
Previous Article in Journal
Investigating Pre-Service Biology Teachers’ Diagnostic Competences: Relationships between Professional Knowledge, Diagnostic Activities, and Diagnostic Accuracy
Previous Article in Special Issue
Use of Mobile Applications in Developing Reading Comprehension in Second Language Acquisition—A Review Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Random Controlled Trial to Examine the Efficacy of Blank Slate: A Novel Spaced Retrieval Tool with Real-Time Learning Analytics

1
Frank H. Netter MD School of Medicine, Quinnipiac University, Hamden, CT 06518, USA
2
English Language Institute, Pace University, New York, NY 10038, USA
3
Jackson Institute for Global Affairs, Yale University, New Haven, CT 06520, USA
4
Blank Slate Technologies LLC, Arlington, VA 22203, USA
*
Author to whom correspondence should be addressed.
Submission received: 8 February 2021 / Revised: 20 February 2021 / Accepted: 23 February 2021 / Published: 25 February 2021
(This article belongs to the Special Issue Using Technology in Higher Education—Series 1)

Abstract

:
Learner-centered coaching and feedback are relevant to various educational contexts. Spaced retrieval enhances long-term knowledge retention. We examined the efficacy of Blank Slate, a novel spaced retrieval software application, to promote learning and prevent forgetting, while gathering and analyzing data in the background about learners’ performance. A total of 93 students from 6 universities in the United States were assigned randomly to control, sequential or algorithm conditions. Participants watched a video on the Republic of Georgia before taking a 60 multiple-choice-question assessment. Sequential (non-spaced retrieval) and algorithm (spaced retrieval) groups had access to Blank Slate and 60 digital cards. The algorithm group reviewed subsets of cards daily based on previous individual performance. The sequential group reviewed all 60 cards daily. All 93 participants were re-assessed 4 weeks later. Sequential and algorithm groups were significantly different from the control group but not from each other with regard to after and delta scores. Blank Slate prevented anticipated forgetting; authentic learning improvement and retention happened instead, with spaced retrieval incurring one-third of the time investment experienced by non-spaced retrieval. Embedded analytics allowed for real-time monitoring of learning progress that could form the basis of helpful feedback to learners for self-directed learning and educators for coaching.

1. Introduction

The human experience of forgetting after initial learning is near ubiquitous [1,2,3,4]. It is difficult and effortful for us to assimilate extensive knowledge into long-term memory. For example, health professionals must learn a large volume of information that will be applied months or years later in caring for patients in various clinical circumstances. Regrettably, after initial learning and assessment, knowledge retention by these learners in many instances decays and information is forgotten [5,6,7,8]. Human memory and power of recall deteriorates rapidly if we do not reinforce what we have learned.
Spaced retrieval practice involves bringing information from long-term memory back into working memory and enhances retention and knowledge application in both child and adult learners [9,10,11,12]. It is a learning technique that incorporates intervals of time between reviews of previously learned material. This is performed in order to exploit the phenomenon whereby humans more easily remember or learn items when they are studied recurrently over a long time span [9,13]. The principle is useful in many circumstances, but especially so where a learner must acquire a large amount of information and retain it indefinitely in memory (e.g., a second language, medical pharmacology, first responder training and regulations, operational risk and compliance training). Most of us have experienced this in how we learned multiplication tables in mathematics (e.g., 6 × 7 = 42, 6 × 8 = 48, 6 × 9 = 54); we practiced them at intervals again and again until the answers were automatic. In short, educational encounters that are distributed and repeated over time result in more efficient learning and improved retention [14].
Increasing attention is being paid to digital technologies that can enhance existing practices in education, teaching, and learning [15,16,17]. As a consequence, technology enhanced learning has emerged as a term that describes the use of hardware devices or software applications to support and help learning beyond what might otherwise be readily achieved [18]. This is consistent with the growing desire in higher education settings for learner-centered coaching and helpful feedback to complement self-regulated learning (i.e., self-monitored activities, practices, and behaviors that learners engage in to pursue academic mastery [19,20,21,22,23,24,25].
These three contexts prompted Blank Slate Technologies LLC to develop a novel internet-based, spaced retrieval software application with embedded learning analytics called Blank Slate [26]. This was performed with the hope of helping learners keep critical information front of mind (i.e., sustain recurrent successful memory retrievals long term) and for their Total Knowledge Analytics™ platform to allow educators or organizational leaders see who knows what in real time.
The potential implications of this study for teaching and learning practice center around: learners themselves; learners in one-on-one teaching relationships with educators; and learners as part of a learning community (e.g., school class, university course, and work training program). For example, physical therapy, physicians, and sports medicine trainees are among those required to learn the muscles of the human posterior shoulder (i.e., supraspinatus, infraspinatus, teres minor, teres major, subscapularis, deltoid, trapezius, rhomboid major, rhomboid minor, latissimus dorsi, and levator scapulae; Figure 1A). This is an information-dense task that requires learning specific muscle names coupled to their visual identification. It is also something that many people do not find easy and struggle to do without making errors or simply forgetting. Blank Slate provides a means for educators to create 10 digital cards that incorporate the image represented in Figure 1B, each asking learners to name one by one, card by card, muscles A–J. Additionally, educators could create 10 more digital cards using Figure 1B that instead name each specific muscle one by one, card by card, and ask learners to identify the muscle’s position A–J on the human back and shoulder. This gives learners practice with two related but distinct retrieval pathways: (1) given the visual position of the muscle, retrieve its name from long-term to working memory, and (2) given the name of the muscle, retrieve its correct location on the back/shoulder from long-term to working memory. This is an example of how spaced retrieval helps to reinforce schema formation by solidifying cognitive frameworks individual learners form when interacting with the material [27]. Rather than learners testing themselves daily (non-spaced retrieval) with these digital cards, Blank Slate has learners test themselves on fewer days that are spaced out across time (spaced retrieval) (Figure 1C). After each question, learners rate the card as easy to remember, hard to remember, or that they had forgotten the information it asked for. Learners would see all 20 cards on day 1. However, thereafter, Blank Slate’s learning analytics algorithm would use each individual learner’s easy, hard, and forgot ratings to shuffle the cards and present harder cards (i.e., those that they need more practice with) at shorter time intervals and easier cards (i.e., those that they need less practice with) at longer time intervals. This is personalized for each individual learner. Blank Slate’s learning analytics can show individual students and educators where they have knowledge gaps and learning deficits; allowing them to adapt what or perhaps how they study, to receive additional teaching or be directed to supplemental resources, or to use learning progress tracking (i.e., seeing forget/hard cards become easier and eventually effortless to recall) to boost learner confidence and raise perceptions of self-efficacy. Lastly, the learning analytics provide data on whole learning cohorts, which can guide educators in what or how they teach (e.g., to support just-in-time teaching [28] or flipped classroom educational strategies [29,30,31,32]). If a large group of learners are struggling with the rotator cuff muscles (i.e., supraspinatus, infraspinatus, teres minor, and subscapularis), educators may adapt their teaching materials and methods in the present (or in future iterations of the class) to focus on these muscles in particular (e.g., by incorporating a case of a patient with a rotator cuff muscle tear from doing bench press exercises poorly in the gym, and having learners speculate over which everyday movements may be impeded and why).
Our research objective was to examine the efficacy of Blank Slate to (1) offset normal human forgetting; (2) unobtrusively monitor learner progress; and (3) create a detailed data record, computationally analyzed to display helpful feedback on individual learner performance. We hypothesized three outcomes: (i) the control group, having no access to Blank Slate, would experience forgetting as measured by post-intervention assessment; (ii) the sequential and algorithm groups would experience increased knowledge acquisition and retention as measured by post-intervention assessment, and (iii) the algorithm group would accrue the same knowledge acquisition and retention increases as the sequential group but would spend less time in total interacting with the Blank Slate application.

2. Materials and Methods

2.1. Participants

This randomized controlled trial lasted four weeks and included a total of 93 students from six different universities or colleges (Boston University, MA, USA; Campbell University, NC, USA; Chemeketa Community College, OR, USA; Pace University, NY, USA; Quinnipiac University, CT, USA; and Radford University, VA, USA) in the United States recruited through social media or email invitations. Approximately half were medical students, one-quarter community college students, and the remainder undergraduate students. Prior to participation, informed consent was obtained from all students and this study was approved by Quinnipiac University’s Institutional Review Board.

2.2. Study Design

A three-arm parallel-groups pre-test-post-test study design with a 2:2:1 block randomization scheme was used to assign the 93 participants to either a control group (n = 19), a sequential group (n = 37), or an algorithm group (n = 37) (Figure 2). The distribution of medical students, community college students, and undergraduate students was similar across the three groups. This parallel-groups design had an overall power (i.e., overall F test) of 0.93 to detect differences between groups at the trial’s post-test conclusion. This power analysis assumed that sequential (non-spaced) and algorithm (spaced) retrieval would have a large effect compared to no review (Cohen’s d = 0.8).

2.3. Intervention

All participants first viewed a 23 min teaching video on the Republic of Georgia and then immediately afterwards took a 60 multiple-choice-question (MCQ) quiz (pre-intervention). The teaching video was created by author J.M. using the Central Intelligence Agency’s world factbook on Georgia [33] and was hosted by author D.M. via Zoom video conferencing (Zoom Video Communications, Inc; San Jose, CA, USA). Georgia was chosen in order to minimize prior knowledge as a confounder. The 60 MCQs were created by authors D.M., R.F., J.M., and M.T., and mapped to either the remember, understand, apply, or analyze tiers of Bloom’s cognitive taxonomy [34], and were double Downing & Haladyna (2004)’s recommendation of 30 items for a test to adequately sample domain knowledge (i.e., knowledge of a specific, specialized field; in this case, Georgia) [35,36]. The questions asked about Georgia’s geography, people and culture, government, economy, transportation, military, transnational issues, history, current affairs, education, and tourism (Table S1). Next, the sequential and algorithm groups were given access for four weeks to a Blank Slate account (Blank Slate Technologies LLC; Arlington, VA, USA) with 60 digital cards representing information tested in the MCQ quiz. The algorithm group experienced the authentic Blank Slate spaced retrieval algorithm, which selects the next card to be viewed based on previous individual performance; they reviewed however many cards were presented to them by the algorithm on a daily basis. The sequential group reviewed all 60 cards, presented in the same sequence, every day; the spaced retrieval algorithm was disabled for this group. The control group had no access to Blank Slate. After four weeks, all participants took the same quiz again (post-intervention).

2.4. Data Analysis

Descriptive statistics include means with standard deviations for quantitative variables and frequencies with percentages for qualitative variables. An analysis of variance with Bonferroni-corrected post-hoc comparisons was used to compare pre- and post-intervention quiz scores between the three assigned groups. Paired t-tests were used to compare pre- to post-intervention scores within the same groups. Unpaired t-tests were used to compare mean total review time between the sequential and algorithm groups. These were chosen because the sample size was large enough (n = 74 for the 2 group comparisons) for the central limit theorem to apply, which states the sampling distribution of the sample means approximates normality and thus the t-test for group comparisons is acceptable. Moreover, Levene’s test for equality of variance indicated no significant departure from the equal variance assumption and the equal number of students per group makes the t-test robust against test assumptions. To alleviate any potential concerns, non-parametric Mann–Whitney U tests were also conducted and the results were the same as the t-tests. Line graphs were used to display the daily averages for the Blank Slate account data collected from the sequential and algorithm groups over the four-week intervention period. For analyses, the app data were aggregated across days for each student resulting in an average value for each student and independent samples t-tests were performed to compare the sequential and algorithm groups. Analyses were conducted in SPSS v26 (IBM SPSS Statistics, Armonk, NY, USA) and the alpha level for statistical significance was set at 0.05.

3. Results

3.1. Pre- and Post-Intervention Quiz Scores

Figure 3 shows the pre- and post-intervention quiz scores for the control (no retrieval), sequential (non-spaced retrieval) and algorithm (spaced retrieval) groups. There were no significant difference between the pre-intervention scores (p = 0.247); which were 40 (±6), 42 (±7), and 43 (±8) questions correct for control, sequential, algorithm groups respectively. At post-intervention there was a statistically significant group difference between the scores (p < 0.001); which were 34 (±5), 59 (±1), and 58 (±2) questions correct for control, sequential, algorithm groups respectively. The control post-intervention mean score was significantly lower than the sequential (p < 0.01) and algorithm (p < 0.01) groups; but the sequential group did not differ from the algorithm group (p = 0.274). Pairwise contrasts revealed the control group’s post-intervention mean score of 34 (±5) was significantly lower than their pre-intervention score of 40 (±6) (p < 0.001). The sequential group’s post-intervention mean score of 59 (±1) was significantly higher than their pre-intervention score of 42 (±7) (p < 0.001). The algorithm group’s post-intervention mean score of 57 (±2) was significantly higher than their pre-intervention score of 43 (±8) (p < 0.001).
Figure 4 shows a scatterplot of each individual participant’s relative delta (i.e., the difference between their pre- and post-intervention score) by group. The range of delta scores was −1 to −15, 6 to 29, and 0 to 33 for the control, sequential, and algorithm groups, respectively.
Figure 5 shows the mean number of questions answered correctly by each group at the start and end of the four-week intervention, as well as their mean review time. The control group’s mean number of questions answered correctly decreased by 6 (±4). The sequential group’s mean number of questions answered correctly increased by 18 (±7). The algorithm group’s mean number of questions answered correctly increased by 14 (±8). The sequential group spent a mean of 96 (±46) mins in total reviewing the 60 digital cards presented to them by Blank Slate; in contrast, the algorithm group spent a mean of 33 (±11) mins. This ~66% reduction of mean review time was significantly different (p = 0.025).

3.2. Comparison of Retrieval-Related Usage Characteristics

Figure 6 shows the mean number of digital card-prompted retrieval attempts each day by group. There were two students in the sequential group who repeatedly cycled through the deck of digital cards again and again on the first day (over 10 times) and this inflated the mean. Excluding the first day, the sequential group had a mean of 62.7 (±17.2) retrieval attempts per day compared to 16.3 (±11.2) for the algorithm group, which was consistent with the non-spaced retrieval and spaced retrieval underlying design for these two arms of the trial. This amounted to almost four times as many required retrieval attempts per day compared to the algorithm group; a difference that was significant and very large (p < 0.001, d = 3.27, Table 1).
The sequential group successfully retrieved from long-term memory a slightly greater amount of information relevant to the 60 digital Blank Slate cards than the algorithm group; but this difference of 90.7% (±15.7) versus 88.4% (±6.4) was not significant and the size of the difference was small (p = 0.704, d = 0.21, Table 1).
Figure 7 shows the mean time in seconds spent per retrieval attempt by the sequential and algorithm groups. There was no difference between groups for the first day of Blank Slate use when both groups were expected to review all 60 cards. However, by the second day, there was a noticeable difference, where the algorithm group was spending more time retrieving information pertaining to the digital cards presented to them. Across days, the algorithm group spent over 50% longer (6.7 ± 2.5 s vs 4.2 ± 1.7 s) retrieving information per card and this difference was significant and large (p < 0.001, d = −1.14, Table 1). Although the sequential group took less time reviewing each question, this did not compensate for the more questions they reviewed and thus overall the sequential group spent more time each day reviewing cards (Figure 8 and Table 1). The algorithm group was more efficient and spent less than half as much time each day reviewing cards (1.8 ± 2.1 min vs 4.2 ± 1.6 min) and the difference was significant and large (p < 0.001, d = 1.23, Table 1).
Table 2 and Figure 9 shows the distribution of digital card-prompted retrieval attempts coded as easy (i.e., successful retrieval from memory was effortless), hard (i.e., successful retrieval from memory was difficult), or forgot (i.e., retrieval from memory was unsuccessful) by participants in the sequential and algorithm groups. These data help explain why the algorithm group exhibited more retrieval time per card (Figure 6). As the ‘All Viewings’ row of Table 2 and first column of Figure 9 shows, during the four weeks of use, the sequential group reviewed a higher proportion of digital cards that prompted retrieval attempts they deemed easy, while the algorithm group reviewed a higher proportion that prompted retrieval attempts they deemed hard or as retrieval failures (i.e., forgot). The Blank Slate spaced retrieval algorithm prioritizes the selection of digital cards coded as hard or forgot by individual users for subsequent retrieval practice. It is unsurprising that digital cards coded previously as hard or forgot would prompt retrieval attempts that take longer when they are encountered again.
The first time a digital card is viewed, there is no information on its retrieval difficulty level and so the distribution of easy, hard, and forgot is similar between the sequential and algorithm groups (‘1st Viewing’ row of Table 2). By the third viewing, the sequential group is encountering proportionally more easy questions (‘3rd Viewing’ row of Table 2). By the sixth viewing, the algorithm group is being presented with proportionally more hard and forgot cards (‘6th Viewing’ row of Table 2). By the tenth viewing, there is a large and clear difference, with the algorithm group being over four times as likely to be presented with a hard card and over twice as likely to be presented with a previously forgotten card (‘10th Viewing’ row of Table 2). This cannot be explained as better performance by the sequential group since there was no difference in the percentage of successful retrievals (Table 1).

3.3. Learner Analytics

Blank Slate’s Total Knowledge AnalyticsTM platform provided a rich array of computationally analyzed data reports pertaining to each individual learner and the 74 participants (from the sequential and algorithm groups combined) as a small population. Illustrative excerpts for selected learner analytics reports are shown; learner names and email addresses have been replaced with pseudonyms to maintain their anonymity.
Figure 10 shows leaderboard rankings organized by either (A) user participation (i.e., the cumulative number of sessions a user has logged into so as to interact with Blank Slate, and (B) user accuracy (i.e., the latest proportion of successful retrieval attempts expressed as a percentage. These reports can display leaderboard statistics for a given week, month or for the entire history of Blank Slate use.
Figure 11 displays the six participants who had the worst retrieval performance. The ranking may be reversed to display participants in order of best-to-worst performance. This screenshot is from the earliest days of the four-week trial when ‘[email protected]’ was the participant who struggled the most to retrieve the information prompted by the digital cards.
Figure 12 shows (in yellow) the amount of time learner ‘[email protected]’ spent engaging in retrieval practice each day from October 6, 2020 to November 1, 2020. Figure 12 also shows (in green) this participant’s learning progress over time. On October 6, their retrieval from memory attempts were ~60% successful. This climbed to 100% success on October 22, October 24, and October 29 through November 1, 2020.
Figure 13 displays ‘[email protected]’’s coding of the 60 digital cards that prompted spaced retrieval as easy (green), hard (orange), and forgot (red) using a pie chart to indicate proportional breakdown over the four-week trial. This participant experienced 323 retrieval attempts—194 they coded as easy (60.1%), 100 as hard (30.9%), and 29 as forgot (i.e., cannot recall) (9.0%).
Figure 14 displays a selection of the 60 digital cards that ‘[email protected]’ coded as easy, hard, or forgot (i.e., cannot recall). These data could form the basis for tailored feedback to the learner (for self-regulated learning) or educator (for coaching).
Figure 15 shows the bottom 8 items from a ranked organization of the 60 digital cards for ‘[email protected]’. They are ranked from hardest to easiest to retrieve the relevant information solicited from memory.

4. Discussion

The assumption that teaching results in durable learning is foundational to many of our educational activities. Unfortunately, information is often lost soon after it is learned [37]. According to the trace theory of memory, neurochemical signaling encoding information evokes synaptic pattern alterations in the brain reflective of a memory ‘trace’. Information in short-term working memory lasts several seconds and if it is not rehearsed the neurochemical signaling quickly fades, leaving the synaptic reorganizations that were newly forming weak and unstable [37]. Learning is fundamentally about retrieving; we learn in order to later remember and apply what we have recalled. Practicing retrieval while you study is thought to strengthen the neurosynaptic connections that are the basis of memory traces, enriching and improving the learning process [38]. Indeed, there has been an exponential rise in the number of papers published on retrieval practice in education journals between 1991 and 2015 [9]. Alongside this, technology enhanced learning has become well positioned as an interface between digital technology and higher education teaching to flatten the human forgetting curve by facilitating spaced retrieval practice. We investigated whether Blank Slate was able to prevent forgetting and instead promote learning, while gathering and analyzing data in the background that could give insights into learners’ performance. The data collected support our three hypothesized outcomes.
As can be seen in Figure 3, Figure 4 and Figure 5, Blank Slate prevented the forgetting experienced by the control group participants. Not only that, Blank Slate facilitated further knowledge acquisition and long-term retention through retrieval practice. This is consistent with research in the cognitive and educational psychology field which has shown that time spent retrieving information is much better than simply reading or reviewing the information for the equivalent amount of time [39]. People may consider this counterintuitive because their experience of studying was to repeatedly re-read or re-write information, or highlight a text, or make carefully arranged and colorful study notes. Repeatedly retrieving actually works much better for long-term retention. Retrieval practice also leads to better transfer to new forms of questions than does repeated reading, reviewing, or re-writing. For example, a practice question might be about sonar in submarines and how bouncing sounds off objects in the water localizes those objects. Then the question on the final exam might be about sonar in bats—the same principle is at work, but in a different context [40,41]. Furthermore, retrieval practice leads to greater learning than educational techniques like concept mapping that are intended to nurture deep comprehension [10]. Blank Slate’s algorithm produced authentic spaced retrieval in the algorithm group. Figure 6 shows (after some initial enthusiasm in the first two days by participants in the sequential group that led them to have >60 retrieval attempts) that the sequential group interacted with all 60 digital cards across the four-week trial. Even as proportionally these retrievals became easier as Table 2 shows: easy was 57.7% (1st viewing), 83.6% (3rd viewing), 88.0% (6th viewing), and 91.4% (10th viewing). For the algorithm group, Figure 6 shows a much lower daily retrieval burden, with some spikes at days 5, 9, and 22, and even then those spikes were smaller than the baseline of the sequential group. To sum up, participants in the algorithm (spaced retrieval) group experienced the same learning improvements as the sequential (non-spaced retrieval) group with fewer retrieval attempts and less time consumed. Time is precious. Therefore, another advantage of spaced retrieval over non-spaced daily practice is that the time freed up can be used for other activities.
In today’s environment, learning generally occurs in one of two contexts: synchronous or asynchronous. Students who take part in synchronous learning (i.e., happening with others at the same time) have to reserve time and commit to a specific meeting in order to attend live teaching sessions or online courses in real time. This may not be ideal for those who already have busy or compressed schedules. Asynchronous learning (i.e., happening independently at many different times) on the other hand can occur even when the student or teacher are not contemporaneous. Students will typically complete learning activities on their own and merely use the internet as a support tool rather than venturing online solely for interactive classes. With technology enhanced learning, we now not only have the means to make asynchronous learning resources available, but also to unobtrusively monitor their use [42]. Blank Slate is an example of a software application that supports asynchronous learning and facilitates our ability to weave in background assessment in real time: we can ask people what they know, ask them to commit to short answers, then generate learning data both for individuals and whole groups of learners based on their responses. Such, digitally-enhanced assessments were defined by the International Summit on Information Technology in Education (EDUsummIT) 2013 as those that “integrate: (1) an authentic learning experience involving digital media with (2) embedded continuous unobtrusive measures of performance, learning and knowledge, (i.e., ‘stealth assessment’) which (3) creates a highly detailed data record that can be computationally analyzed and displayed so that (4) learners and teachers can immediately utilize the information to improve learning.” [43]. Unobtrusive assessment is seamlessly woven into the fabric of Blank Slate’s digital environment. Its learner analytics platform represents a subtle, yet powerful process by which learner performance data are continuously gathered during the course of learning to support inferences made about the level of progress towards content mastery [44]. Blank Slate’s convergence of asynchronous technologies that embed unobtrusive assessments/analytics with spaced learning motifs may represent an exciting opportunity to further competency-based learning.

4.1. Implications for Teaching and Learning Practice

This investigation adds to the growing body of knowledge about spaced retrieval practice (for a comprehensive review see Karpicke, 2017 [9]). It reproduced, by means of Blank Slate, learning benefits ascribed to spaced retrieval in a sample of Higher Education learners located across the United States. It further contributes the verification of Blank Slate as a novel spaced retrieval software application with embedded real-time learner analytics. Blank Slate presents an opportunity for spaced retrieval practice and externally-sourced feedback to be brought into close proximity via a shared digital platform. This is significant because feedback coupled to retrieval practice dramatically amplifies knowledge retention improvements [4,45,46]. Indeed, giving learners feedback to guide their progress has always been viewed as beneficial to learning; however, the need to combine opportunities for learner reflection and coaching with feedback to help learners achieve anticipated outcomes has been under stressed [47]. That said, unguided learner reflection often lacks fidelity [48,49]. Eva and Regehr noted that traditional self-assessment by learners usually takes the form of a “personal, unguided reflection on performance for the purposes of generating an individually-derived summary of one’s own level of knowledge, skill, and understanding in a particular area.” [50]. In other words, individuals intuitively see themselves are the best source of information and look inward to generate an assessment of their own knowledge and abilities. Blank Slate’s learner analytics reports (see Figure 10, Figure 11, Figure 12, Figure 13, Figure 14 and Figure 15 for examples) represent an external source of information to inform learner reflection, self-directed learning processes, and feedback shared by educators. It provides a digital means to direct our attention to trustworthy feedback and to share it in a way that may improve its acceptance by the recipients [51]. In a learner-centered coaching approach, teachers could use available performance data generated by Blank Slate to contribute specific feedback that is relevant to fostering learners’ continued momentum towards competence and then mastery. In this way, it may also help educators shift practically away from using assessments only to gather evidence “of” learning (i.e., at the end of a course or period of studying) and towards incorporating assessments “for” learning as a teaching strategy [8].

4.2. Strengths, Limitations, and Future Studies

Strengths of this investigation included the robust randomized controlled trial design and multi-institutional sample; however, this study’s generalizability may be limited to university and college populations. The modest sample size is a major limitation; however, we believed it to be prudent to conduct a small study with sufficient power to establish that Blank Slate produces the benefits of spaced retrieval documented in the literature with one of its potential audiences (i.e., higher education learners) before any larger, follow-up studies are pursued. Another limitation is that we could only observe the learner analytics data and not act on it to provide feedback or coaching. To do so would have introduced a source of confounding that could have skewed interpretation of the spaced retrieval data. Future studies should focus on examining (i) the efficacy of Blank Slate with a more diverse representation of humans, and (ii) the effects of Blank Slate’s learning analytics being used as source of feedback for self-regulated learning and learning coupled to educator-led feedback or coaching opportunities.

5. Conclusions

Our findings indicate that Blank Slate prevented forgetting and instead promoted learning and knowledge retention in a time-efficient manner. There is a need for technology-assisted identification of learning gaps that serve as a source of data input to guide individualized, formative feedback, and mentoring or coaching opportunities. Blank Slate as a technology enhanced learning tool integrated: (i) spaced retrieval learning experiences involving digital media with (ii) embedded continuous unobtrusive measures of performance, learning and knowledge, which (iii) created a highly detailed data record that was computationally analyzed and displayed so that (iv) stakeholders could utilize the information for feedback and coaching. Blank Slate, as an internet-based software application that can operate on multiple devices, is positioned to contribute to emerging digital infrastructure that supports education and learning.

Supplementary Materials

The following are available online at https://0-www-mdpi-com.brum.beds.ac.uk/2227-7102/11/3/90/s1, Table S1: List of the 60 author-created multiple-choice questions used for pre-test and post-test assessment.

Author Contributions

Conceptualization, D.M. and R.F.; data curation, D.M.; formal analysis, D.M. and R.F.; funding acquisition, M.T.; investigation, D.M.; methodology, D.M., R.F. and J.M.; project administration, D.M. and R.F.; resources, D.M., R.F., J.M. and M.T.; software, M.T.; supervision, D.M.; visualization, D.M. and R.F.; writing—original draft, D.M. and R.F.; writing—review and editing, D.M., R.F., J.M. and M.T. All authors have read and agreed to the published version of this manuscript.

Funding

Funding for this project was provided by Blank Slate Technologies LLC.

Institutional Review Board Statement

This study was conducted according to the guidelines of the Declaration of Helsinki and approved by the Human Experimentation Committee/Institutional Review Board of Quinnipiac University (#10620; 15 September 2020).

Informed Consent Statement

Informed consent was obtained from all subjects involved in this study.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available in order to maintain participant confidentiality per IRB protocol.

Conflicts of Interest

D.M., R.F. and J.M. declare no conflicts of interest. M.T. declares a conflict of interest on the basis that he is the founder and CEO of Blank Slate Technologies LLC. Blank Slate Technologies LLC as the funder of this study had no role in the design of this study; in the collection, analyses, or interpretation of data; in the writing of this manuscript; or in the decision to publish the results. All authors confirm that the reported research results and their interpretation are free from inappropriate influences, manipulation, or suppression.

References

  1. Ebbinghaus, H. Memory: A Contribution to Experimental Psychology; Dover Publications, Inc: New York, NY, USA, 1964. [Google Scholar]
  2. Baddeley, A. Your Memory: A User’s Guide; Firefly Books: Buffalo, NY, USA, 2004. [Google Scholar]
  3. Cepeda, N.J.; Vul, E.; Rohrer, D.; Wixted, J.T.; Pashler, H. Spacing effects in learning: A temporal ridgeline of optimal retention. Psychol. Sci. 2008, 19, 1095–1102. [Google Scholar] [CrossRef]
  4. Larsen, D.P. Planning Education for Long-Term Retention: The Cognitive Science and Implementation of Retrieval Practice. Semin. Neurol. 2008, 38, 449–456. [Google Scholar] [CrossRef] [Green Version]
  5. McKenna, S.P.; Glendon, A.I. Occupational first aid training: Decay in cardiopulmonary resuscitation (CPR) skills. J. Occup. Psychol. 1985, 58, 109–117. [Google Scholar] [CrossRef]
  6. Kerfoot, B.P.; Baker, H.; Jackson, T.L.; Hulbert, W.C.; Federman, D.D.; Oates, R.D.; DeWolf, W.C. A multi-institutional randomized controlled trial of adjuvant web-based teaching to medical students. Acad. Med. 2006, 81, 224–230. [Google Scholar] [CrossRef] [PubMed]
  7. Larsen, D.P.; Butler, A.C.; Lawson, A.L.; Roediger, H.L. The importance of seeing the patient: Test-enhanced learning with standardized patients and written tests improves clinical application of knowledge. Adv. Health Sci. Educ. 2013, 18, 409–425. [Google Scholar] [CrossRef]
  8. Green, M.L.; Moeller, J.J.; Spak, J.M. Test-enhanced learning in health professions education: A systematic review: BEME Guide No. 48. Med. Teach. 2018, 40, 337–350. [Google Scholar] [CrossRef] [PubMed]
  9. Karpicke, J.D. Retrieval-based learning: A decade of progress. In Cognitive Psychology of Memory, Learning and Memory: A Comprehensive Reference; Wixted, J.T., Ed.; Academic Press: Cambridge, MA, USA, 2017; Volume 2, pp. 487–514. [Google Scholar] [CrossRef]
  10. Karpicke, J.D.; Blunt, J.R. Retrieval Practice Produces More Learning than Elaborative Studying with Concept Mapping. Science 2011, 331, 772–775. [Google Scholar] [CrossRef] [PubMed]
  11. Fazio, L.K.; Marsh, E.J. Retrieval-Based Learning in Children. Curr. Dir. Psychol. Sci. 2019, 28, 111–116. [Google Scholar] [CrossRef]
  12. Leonard, L.B.; Deevy, P. Retrieval Practice and Word Learning in Children with Specific Language Impairment and Their Typically Developing Peers. J. Speech Lang. Hear. Res. 2000, 63, 3252–3262. [Google Scholar] [CrossRef] [PubMed]
  13. Karpicke, J.D.; Roediger, H.L. Expanding retrieval practice promotes short-term retention, but equally spaced retrieval enhances long-term retention. J. Exp. Psychol. Learn. Mem. Cogn. 2007, 33, 704–719. [Google Scholar] [CrossRef] [Green Version]
  14. Toppino, T.C.; Gerbier, E. About practice: Repetition, spacing, and abstraction. Psychol. Learn Motivat. 2014, 60, 113–189. [Google Scholar] [CrossRef]
  15. Li, Q. Student and Teacher Views About Technology: A Tale of Two Cities? J. Res. Technol. Educ. 2007, 39, 377–397. [Google Scholar] [CrossRef]
  16. Tamim, R.M.; Bernard, R.M.; Borokhovski, E.; Abrami, P.C.; Schmid, R.F. What Forty Years of Research Says About the Impact of Technology on Learning: A Second-Order Meta-Analysis and Validation Study. Rev. Educ. Res. 2011, 81, 4–28. [Google Scholar] [CrossRef] [Green Version]
  17. Passey, D. Inclusive Technology Enhanced Learning: Overcoming Cognitive, Physical, Emotional and Geographic Challenges; Routledge: New York, NY, USA, 2013. [Google Scholar]
  18. Sen, A.; Leong, C.K.C. Technology-Enhanced Learning. In Encyclopedia of Education and Information Technologies; Tatnall, A., Ed.; Springer: Cham, Switzerland, 2020. [Google Scholar] [CrossRef]
  19. Zimmerman, B.J. Investigating self-regulation and motivation: Historical background, methodological developments, and future prospects. Am. Educ. Res. J. 2008, 45, 166–183. [Google Scholar] [CrossRef]
  20. Cleary, T.J.; Durning, S.J.; Gruppen, L.D.; Hemmer, P.A.; Artino, A.R. Self-regulated learning in medical education. In Oxford Textbook of Medical Education; Walsh, E., Ed.; Oxford University Press: Oxford, UK, 2013. [Google Scholar]
  21. Gandomkar, R.; Mirzazadeh, A.; Jalili, M.; Yazdani, K.; Fata, L.; Sandars, J. Self-regulated learning processes of medical students during an academic learning task. Med. Educ. 2016, 50, 1065–1074. [Google Scholar] [CrossRef] [PubMed]
  22. Simpson, D.; Marcdante, K.; Souza, K.H.; Anderson, A.; Holmboe, E. Job Roles of the 2025 Medical Educator. J. Grad. Med. Educ. 2018, 10, 243–246. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. Zheng, B.; Ward, A.; Stanulis, R. Self-regulated learning in a competency-based and flipped learning environment: Learning strategies across achievement levels and years. Med. Educ. Online 2020, 25, 1686949. [Google Scholar] [CrossRef]
  24. Panadero, E. A Review of Self-regulated Learning: Six Models and Four Directions for Research. Front. Psychol. 2017, 8, 422. [Google Scholar] [CrossRef] [PubMed]
  25. Cutrer, W.B.; Miller, B.; Pusic, M.V.; Mejicano, G.; Mangrulkar, R.S.; Gruppen, L.D.; Hawkins, R.E.; Skochelak, S.E.; Moore, D.E. Fostering the Development of Master Adaptive Learners: A Conceptual Model to Guide Skill Acquisition in Medical Education. Acad. Med. 2017, 92, 70–75. [Google Scholar] [CrossRef]
  26. Blank Slate Technologies LLC. Available online: https://blankslatetech.co (accessed on 8 February 2021).
  27. Pumilia, C.A.; Lessans, S.; Harris, D. An Evidence-Based Guide for Medical Students: How to Optimize the Use of Expanded-Retrieval Platforms. Cureus 2020, 12, e10372. [Google Scholar] [CrossRef]
  28. Novak, G.; Patterson, E.T.; Gavrin, A.D.; Christian, W. Just-In-Time Teaching: Blending Active Learning with Web Technology; Prentice Hall: Upper Saddle River, NJ, USA, 1999. [Google Scholar]
  29. Muzyk, A.; Fuller, J.S.; Jiroutek, M.R.; Grochowski, C.O. Implementation of a flipped classroom model to teach psychopharmacotherapy to third-year doctor of pharmacy (PharmD) students. Pharm Educ. 2015, 15, 44–53. [Google Scholar]
  30. Martinelli, S.M.; Chen, F.; Mcevoy, M.D.; Zvara, D.A.; Schell, R.M. Utilization of the flipped classroom in anesthesiology graduate medical education: An initial survey of faculty beliefs and practices about active learning. J. Educ. Perioper. Med. 2018, 20, E617. [Google Scholar]
  31. Han, E.; Klein, K.C. Pre-class learning methods for flipped classrooms. AJPE 2019, 83, 6922. [Google Scholar] [CrossRef] [PubMed]
  32. Ding, C.; Wang, Q.; Zou, J.; Zhu, K. Implementation of flipped classroom combined with case- and team-based learning in residency training. Adv. Physiol. Educ. 2021, 45, 77–83. [Google Scholar] [CrossRef]
  33. Central Intelligence Agency. The World Factbook: George. 2009. Available online: http://teacherlink.ed.usu.edu/tlresources/reference/factbook/geos/gg.html (accessed on 23 January 2021).
  34. Anderson, L.W.; Krathwohl, D.R.; Bloom, B.S. A Taxonomy for Learning, Teaching, and Assessing: A Revision of Bloom’s Taxonomy of Educational Objectives, Complete ed.; Longman: New York, NY, USA, 2001. [Google Scholar]
  35. Downing, S.M.; Haladyna, T.M. Validity threats: Overcoming interference with proposed interpretations of assessment data. Med. Educ. 2004, 38, 327–333. [Google Scholar] [CrossRef] [PubMed]
  36. Hjørland, B.; Albrechtsen, H. Toward A New Horizon in Information Science: Domain Analysis. J. Am. Soc. Inform. Sci. 1995, 46, 400–425. [Google Scholar] [CrossRef]
  37. Murre, J.M.; Dros, J. Replication and analysis of Ebbinghaus’ forgetting curve. PLoS ONE 2015, 10, e0120644. [Google Scholar] [CrossRef]
  38. McKeown, D.; Mercer, T.; Bugajska, K.; Duffy, P.; Barker, E. The visual nonverbal memory trace is fragile when actively maintained, but endures passively for tens of seconds. Mem. Cogn. 2019, 48, 212–225. [Google Scholar] [CrossRef] [Green Version]
  39. Roediger, H.L.; Karpicke, J.D. Test-enhanced learning: Taking memory tests improves long-term retention. Psychol. Sci. 2006, 17, 249–255. [Google Scholar] [CrossRef] [PubMed]
  40. Karpicke, J.D. Retrieval-based learning: Active retrieval promotes meaningful learning. Curr. Dir. Psychol. Sci. 2012, 21, 157–163. [Google Scholar] [CrossRef]
  41. McDaniel, M.A. Rediscovering transfer as a central concept. In Science of Memory: Concepts; Roediger, H.L., Dudai, Y., Fitzpatrick, S.M., Eds.; Oxford University Press: New York, NY, USA, 2007; pp. 355–360. [Google Scholar]
  42. Georgiadis, K.; Van Lankveld, G.; Bahreini, K.; Westera, W. On The Robustness of Stealth Assessment. IEEE Trans. Games 2020, 1. [Google Scholar] [CrossRef]
  43. Webb, M.; Gibson, D. Technology enhanced assessment in complex collaborative settings. Educ. Inf. Technol. 2015, 20, 675–695. [Google Scholar] [CrossRef] [Green Version]
  44. Shute, V.J.; Ventura, M.; Bauer, M.I.; Zapata-Rivera, D. Melding the power of serious games and embedded assessment to monitor and foster learning: Flow and grow. In Serious Games: Mechanisms and Effects; Ritterfeld, U., Cody, M., Vorderer, P., Eds.; Routledge, Taylor and Francis: Mahwah, NJ, USA, 2009; pp. 295–321. [Google Scholar]
  45. Butler, A.C.; Karpicke, J.D.; Roediger, H.L. Correcting a metacognitive error: Feedback increases retention of low-confidence correct responses. J. Exp. Psychol. Learn. Mem. Cogn. 2008, 34, 918–928. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  46. Rowland, C.A. The effect of testing versus restudy on retention: A meta-analytic review of the testing effect. Psychol. Bull. 2014, 140, 1432–1463. [Google Scholar] [CrossRef] [PubMed]
  47. Krackov, S.K.; Pohl, H.S.; Peters, A.S.; Sargeant, J.M. Feedback, reflection and coaching: A new model. In A Practical Guide for Medical Teachers; Dent, J.A., Harden, R.M., Hunt, D., Eds.; Elsevier Ltd: Edinburgh, UK, 2017; pp. 281–288. [Google Scholar]
  48. Gordon, M.J. A review of the validity and accuracy of self-assessments in health professions training. Acad. Med. 1991, 66, 762–769. [Google Scholar] [CrossRef]
  49. Davis, D.A.; Mazmanian, P.E.; Fordis, M.; Van Harrison, R.; Thorpe, K.E.; Perrier, L. Accuracy of Physician Self-assessment Compared with Observed Measures of Competence: A Systematic Review. JAMA 2006, 296, 1094–1102. [Google Scholar] [CrossRef]
  50. Eva, K.W.; Regehr, G. “I’ll never play professional football” and other fallacies of self-assessment. J. Contin. Educ. Heal. Prof. 2008, 28, 14–19. [Google Scholar] [CrossRef] [PubMed]
  51. Eva, K.W.; Regehr, G. Self-Assessment in the Health Professions: A Reformulation and Research Agenda. Acad. Med. 2005, 80, S46–S54. [Google Scholar] [CrossRef]
Figure 1. Drawing of the human posterior shoulder and back with muscles (A) identified and (B) de-identified. (C) compares example spaced versus non-spaced retrieval motifs across time.
Figure 1. Drawing of the human posterior shoulder and back with muscles (A) identified and (B) de-identified. (C) compares example spaced versus non-spaced retrieval motifs across time.
Education 11 00090 g001
Figure 2. The randomized controlled trial’s parallel-groups study design.
Figure 2. The randomized controlled trial’s parallel-groups study design.
Education 11 00090 g002
Figure 3. Mean (±standard deviation) pre- and post-intervention quiz scores for the control, sequential, and algorithm groups; *** = p < 0.001.
Figure 3. Mean (±standard deviation) pre- and post-intervention quiz scores for the control, sequential, and algorithm groups; *** = p < 0.001.
Education 11 00090 g003
Figure 4. Scatterplot of each individual participant’s relative delta (i.e., the difference between their pre- and post-intervention score) by group.
Figure 4. Scatterplot of each individual participant’s relative delta (i.e., the difference between their pre- and post-intervention score) by group.
Education 11 00090 g004
Figure 5. Mean number of questions answered correctly by each group at the start and end of the four-week intervention, and the mean total review time for the sequential and algorithm groups.
Figure 5. Mean number of questions answered correctly by each group at the start and end of the four-week intervention, and the mean total review time for the sequential and algorithm groups.
Education 11 00090 g005
Figure 6. Mean (±standard error of the mean) number of digital card-prompted retrieval attempts each day by the sequential and algorithm groups.
Figure 6. Mean (±standard error of the mean) number of digital card-prompted retrieval attempts each day by the sequential and algorithm groups.
Education 11 00090 g006
Figure 7. Mean (±standard error of the mean) time in seconds spent per retrieval attempt each day by the sequential and algorithm groups.
Figure 7. Mean (±standard error of the mean) time in seconds spent per retrieval attempt each day by the sequential and algorithm groups.
Education 11 00090 g007
Figure 8. Mean (±standard error of the mean) time in minutes spent reviewing Blank Slate’s digital cards and retrieving information each day by the sequential and algorithm groups.
Figure 8. Mean (±standard error of the mean) time in minutes spent reviewing Blank Slate’s digital cards and retrieving information each day by the sequential and algorithm groups.
Education 11 00090 g008
Figure 9. Percentage proportion of retrieval attempts coded as easy, hard, or forgot by participants in the (A) sequential and (B) algorithm groups for all viewings then their first, third, sixth, and tenth viewings.
Figure 9. Percentage proportion of retrieval attempts coded as easy, hard, or forgot by participants in the (A) sequential and (B) algorithm groups for all viewings then their first, third, sixth, and tenth viewings.
Education 11 00090 g009
Figure 10. Blank Slate learner analytics—a screenshot of the leaderboard of anonymized participants ranked by (A) participation and by (B) accuracy.
Figure 10. Blank Slate learner analytics—a screenshot of the leaderboard of anonymized participants ranked by (A) participation and by (B) accuracy.
Education 11 00090 g010
Figure 11. Blank Slate learner analytics—a subset of anonymized participants ranked by worst to best performance.
Figure 11. Blank Slate learner analytics—a subset of anonymized participants ranked by worst to best performance.
Education 11 00090 g011
Figure 12. Blank Slate learner analytics—a single anonymized participant’s total time spent interacting with Blank Slate (yellow) each day of the four-week trial and their percentage of successful retrievals from memory (green). Lefthand y axis: percentage of successful retrievals from memory. Righthand y axis: time in minutes and seconds spent interacting with Blank Slate. X axis: days from October 6 2020 to November 1 2020 using the U.S. date format (mm-dd); for example, 10/7 = October 7 2020.
Figure 12. Blank Slate learner analytics—a single anonymized participant’s total time spent interacting with Blank Slate (yellow) each day of the four-week trial and their percentage of successful retrievals from memory (green). Lefthand y axis: percentage of successful retrievals from memory. Righthand y axis: time in minutes and seconds spent interacting with Blank Slate. X axis: days from October 6 2020 to November 1 2020 using the U.S. date format (mm-dd); for example, 10/7 = October 7 2020.
Education 11 00090 g012
Figure 13. Blank Slate learner analytics—a single anonymized participant’s coding of the 60 digital cards that prompt spaced retrieval as easy (green), hard (orange), or forgot (i.e., cannot recall) (red).
Figure 13. Blank Slate learner analytics—a single anonymized participant’s coding of the 60 digital cards that prompt spaced retrieval as easy (green), hard (orange), or forgot (i.e., cannot recall) (red).
Education 11 00090 g013
Figure 14. Blank Slate learner analytics—an excerpt of specific digital cards that a single anonymized participant’s coded as easy (green), hard (orange), or forgot (i.e., cannot recall) (red).
Figure 14. Blank Slate learner analytics—an excerpt of specific digital cards that a single anonymized participant’s coded as easy (green), hard (orange), or forgot (i.e., cannot recall) (red).
Education 11 00090 g014
Figure 15. Blank Slate learner analytics—an excerpt of the specific digital cards that a single anonymized participant found to be difficult to retrieve ranked by hardest to easiest.
Figure 15. Blank Slate learner analytics—an excerpt of the specific digital cards that a single anonymized participant found to be difficult to retrieve ranked by hardest to easiest.
Education 11 00090 g015
Table 1. Mean (±standard deviation) of Blank Slate outcomes by group; * excludes study day 1.
Table 1. Mean (±standard deviation) of Blank Slate outcomes by group; * excludes study day 1.
Outcome SequentialAlgorithmp-ValueCohen’s d
Retrieval attempts per day *62.7 (±17.2)16.3 (±11.2)<0.0013.27
Percentage of successful retrievals from memory 90.7 (±15.7)88.4 (±6.4)0.7040.21
Seconds per retrieval attempt4.3 (±1.7)6.7 (±2.5)<0.001−1.14
Minutes per day using Blank Slate * 4.2 (±1.6)1.8 (±2.1)<0.0011.23
Table 2. Total number of digital card viewings and percentage of retrieval attempts coded as easy, hard, or forgot by participants in the sequential and algorithm groups for their first, third, sixth, and tenth viewings.
Table 2. Total number of digital card viewings and percentage of retrieval attempts coded as easy, hard, or forgot by participants in the sequential and algorithm groups for their first, third, sixth, and tenth viewings.
EasyHardForgot
All Viewings
Sequential
 
44,053 (87.8%)
 
1596 (3.2%)
 
4512 (9.0%)
Algorithm7286 (72.4%)1034 (10.3%)1743 (17.3%)
1st Viewing
Sequential
Algorithm
 
57.7%
61.6%
 
9.6%
7.7%
 
32.7%
30.6%
3rd Viewing
Sequential
Algorithm
 
83.6%
79.9%
 
8.5%
7.3%
 
7.9%
12.8%
6th Viewing
Sequential
Algorithm
 
88.0%
76.0%
 
3.6%
12.1%
 
8.4%
11.9%
10th Viewing
Sequential
Algorithm
 
91.4%
75.1%
 
2.3%
10.7%
 
6.3%
14.2%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

McHugh, D.; Feinn, R.; McIlvenna, J.; Trevithick, M. A Random Controlled Trial to Examine the Efficacy of Blank Slate: A Novel Spaced Retrieval Tool with Real-Time Learning Analytics. Educ. Sci. 2021, 11, 90. https://0-doi-org.brum.beds.ac.uk/10.3390/educsci11030090

AMA Style

McHugh D, Feinn R, McIlvenna J, Trevithick M. A Random Controlled Trial to Examine the Efficacy of Blank Slate: A Novel Spaced Retrieval Tool with Real-Time Learning Analytics. Education Sciences. 2021; 11(3):90. https://0-doi-org.brum.beds.ac.uk/10.3390/educsci11030090

Chicago/Turabian Style

McHugh, Douglas, Richard Feinn, Jeff McIlvenna, and Matt Trevithick. 2021. "A Random Controlled Trial to Examine the Efficacy of Blank Slate: A Novel Spaced Retrieval Tool with Real-Time Learning Analytics" Education Sciences 11, no. 3: 90. https://0-doi-org.brum.beds.ac.uk/10.3390/educsci11030090

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop