Next Article in Journal
A Sustainable Transport System—The MMQUAL Model of Shared Micromobility Service Quality Assessment
Previous Article in Journal
Nature Conservation in Sustainability
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

SeisTutor: A Custom-Tailored Intelligent Tutoring System and Sustainable Education

1
Department of Computer Science and Engineering, CMR Institute of Technology Kandlakoya, Hyderabad 501401, Telangana, India
2
School of Computing, DIT University, Dehradun 248009, Uttarakhand, India
3
Department of Computer Science, BITS Pilani, Dubai Campus, Dubai P.O. Box 345055, United Arab Emirates
4
Department of Business Management, College of Business Administration, Kingdom University, Riffa 3903, Bahrain
*
Author to whom correspondence should be addressed.
Sustainability 2022, 14(7), 4167; https://0-doi-org.brum.beds.ac.uk/10.3390/su14074167
Submission received: 15 February 2022 / Revised: 26 March 2022 / Accepted: 29 March 2022 / Published: 31 March 2022

Abstract

:
Education is the cornerstone of improving people’s lives and achieving global sustainability. Intelligent systems assist sustainable education with various benefits, including recommending a personalized learning environment to learners. The classroom learning environment facilitates human tutors to interact with every learner and obtain the opportunity to understand the learner’s psychology and then provide learning material (access learner previous knowledge and well-align the learning material as per learner requirement) to them accordingly. Implementing this cognitive intelligence in Intelligent Tutoring System is quite tricky. This research focused on mimicking human tutor cognitive intelligence in the computer-aided system of offering an exclusive curriculum or quality education for sustainable learners. The prime focus of this research article was to evaluate the proposed SeisTutor using Kirkpatrick four-phase evaluation model. The experimental results depict the enhanced learning gained through intelligence incorporated SeisTutor against the intelligence absence, as demonstrated.

1. Introduction

Quality education is a core element of sustainable development, intended to ensure equitable and inclusive quality education and encourage opportunities for lifelong learning for all. Research claims that education is critical to livelihood and the world’s long-term viability [1,2]. New digital technologies are transforming education in both informal and formal learning environments. This overarching goal is linked to several general issues influencing education in the digital era. Some of the significant developments and their impact on education can be stated as follows: (a) educational objectives; (b) learning environments and ecosystems; (c) learning and teaching processes; and (d) educational policy and governance.
According to the Sustainable Development Goal, equality (male and female) to access excellent technical learning and training practices is anticipated by 2030 [3]. Education is the cornerstone of improving people’s lives and achieving global sustainability [1,2]. Furthermore, equitable and inclusive education equips people to devise innovative solutions for their day-to-day lives. Incorporating digital inclusion and good quality educational practice enables the learners to deliver motivation, knowledge, and opportunities; offer skill training to employ SDG explication; and resolve the SDG challenges [4,5].
Human, cultural, ethical, and ecological principles are described as sustainable education development. It offers good teaching practices to societies that aid the growth of communities, institutions, organizations, and the competitive environment. Thus, it enhances the economic and social life [6,7]. Intelligent systems assist sustainable education with a wide variety of benefits, including recommending a personalized learning environment to learners [8].
The value of utilizing automated systems over long-duration teaching practices is becoming more apparent. Incorporating Artificial intelligence and expert systems in education helps gauge real-world scenarios. E-learning, technology, applications, e-teaching principles, and sustainable development are the most vital parts of a sustainable e-learning environment. The principles of intelligent learning or e-learning comprise the framework, models, and theories, whereas the principle of intelligent tutoring or e-teaching comprises a program, curriculum, and pedagogy. Good teaching practices and evaluation models, collaborative, and custom-tailored or personalized learning approaches work together to determine and recommend a custom-tailored pedagogy or tutoring strategy [9].
Personalization emphasizes self-development and the academic realm. The learners’ self-growth depends on the skill and knowledge gained through online courses and internet communities. A custom-tailored or personalized learning environment encompasses learning tools, services, and application or system that recommends learning per learners’ needs based on learner capabilities, a low-cost tutoring environment, and a personalized learner profile [10].
In order to effectively meet this learner-centric requirement, an intelligent tutoring system can offer personalized learning experiences to learners. The intelligent tutoring system (ITS) is an artificial intelligence technique that offers the learner exclusive learning material, aligned and gathered as per the learner’s grasping ability and preferred media of learning.
SeisTutor was developed for learning “Seismic Data Interpretation” as a subject domain of this research. This research aims to provide pastoral care to learners by cost-effectively offering one-to-one, customized learning material. SeisTutor offers a personalized learning environment. It brings personalization in identifying tutoring strategy (based on pretest (learning style test and domain knowledge test)), exclusive curriculum design, and observation of learner psychological state of mind during learning sessions. The feature of personalization is built on many aspects, such as accuracy of predicting tutoring strategy (based on pretest), curriculum design, and psychological parameters. The only technique to determine the performance of the personalization facility of SeisTutor is to appraise the system in actual circumstances (learners who are learning the course material). Appraise acts as a critical element for quality assurance because it enables the learner to provide valuable feedback on the learning experience and on learning content, which further helps to understand the learner’s perspective and makes the learning better.
The assessment of SeisTutor was accompanied in the 2018–2019 scholastic year. The objective of the evaluation was to examine how effectively SeisTutor personalized itself to fulfill the learner’s needs and whether SeisTutor helps enhance the learning gain for learning the “Seismic Data Interpretation” domain. In order to accomplish this, the following program was outlined:
  • Perform a schematic literature review on how to evaluate the efficacy of learning;
  • Generate an assessment for determining the efficacy of learning;
  • Perform analysis to identify how efficiently learners learn with SeisTutor.
This article is organized into five sections. It begins with a schematic literature review on the evaluation of learning with ITS (Section 2). Further, Section 3 introduces the architecture of SeisTutor in detail, which includes the general description and mathematical justification of its functionalities. Section 4 illustrates the experimental analysis performed to evaluate the effectiveness of SeisTutor systematically. Section 5 elaborates on the comparative analysis with current work, and in Section 6, we finally conclude the implication of this work.

2. Background and Preliminaries

In this section, comprehensive coverage of the development of a sustainable, intelligent tutoring system (ITS) from traditional computer-aided instruction (CAI) to artificial intelligence techniques embedded tutoring systems and the evaluation prototype are discussed.
The academic realm provides a framework for considering long-term sustainable education [11]. Incorporating AI and expert systems in education plays a vital role in offering learning (in the form of blended, online, virtualization, visualization, etc.) at work, home, coffee shops, museums, city streets, and parks. Another trend is shifting our perceptions of context and spaces of learning. Finally, it considers the potential of digitally mediated learning to be accessible anytime and anywhere [12]. Here the teaching environment expanded temporally and spatially. Ubiquitous learning abandons the discrepancy between formal and informal learning. A planned curriculum and competent facilitators play a role in keeping individuals actively engaged; however, the learning is more inclined to need, purpose, and context [13]. These necessitate a paradigm shift from “curriculum-based” to “problem-based” learning, requiring a rethinking of learning motivations, processes, and content. One of the most remarkable features of providing learners with online learning possibilities is that they have significantly more autonomy and choice in pursuing paths of study that are fascinating and meaningful to them. Learning evolves from a recipient, passive model to a more self-directed, collaborative, and active model [14,15]. Teachers can collaborate with domain experts and technology to develop a learning environment that appropriately recognizes learner challenges and recommend an adaptive instructional strategy that can help understand the concept effectively. The goal of a “flipped” learning environment is not to replace teachers; instead, it is to automate the learning processes; enable the teachers to focus their talents, energies, experience, and creativity on tutoring.
Computer-aided instruction (CAI) systems store the learning material, which the learner further uses in different ways (representation) [16]. These systems have several limitations. These systems mainly focused on quantitative education (teacher-centric) rather than qualitative nature (learner-centric) education. The CAI systems adopt a highly primitive tutoring strategy, which leads to less interaction between the CAI tutor and the learner. Further advancement led to the origin of the field of Intelligent Computer-Aided Instruction (ICAI) or Intelligent Tutoring System (ITS) [17].
In 1950, The first ITS was technologized as CAI [18]. McDonald, Woolf developed intelligence embedded computer-aided instruction, which delivers learning material by establishing effective interaction with the learner. Thus, these systems emulate the cognitive intelligence of humans only to a small extent, i.e., provide necessary guidance based on the learner’s action.
ITS, developed by [19], creates questions on vocabulary and arithmetic but cannot adapt and model learner needs. In advancement to this, several adaptive systems were developed by the researchers [20,21,22]. Advancement in the learning systems led to significant modifications in the architecture of ITS. Further improvements in Pedagogy Model led to adapting the learning material to learner competency. Thus, incorporating AI techniques shifted to fine-tune pedagogical recommendation and learner feedback.
Physics is a well-documented subject domain. An intelligent tutoring system developed by the Andes comprises physics as the knowledge domain [23,24]. The Andes incorporates an intelligent feature of determining performance parameters, predicting learners’ subsequent actions, and identifying and recommending a suitable strategy. For decision-making, they utilize a Bayesian network.
In order to adjudge previous learner knowledge, pretest plays an important role. ITS developed by [25] offers a personalized learning environment by offering learning material to the learner based on previous knowledge about the course. The pathfinder technique was used to determine the previous knowledge of the learner. ITS developed by [26] proposed a solution for determining the learners’ curriculum by utilizing a profile framework.
Further advancement in this field uses a data mining technique to mine the meaningful learning path for the learner. This system tracks learner activity during learning and recommends the most suitable learning path. ITS developed by [27] incorporated two methodologies: determining the learning path and recommending the learning path. Initially, the system utilizes the apriori algorithm to generate an initial course coverage plan; they used course-concept analysis to determine the association among the course-concept and then adjudged the preferable course coverage plan.
Another [28] proposal uses a fuzzy rule base technique to determine the association between the list of materials and learner requirements based on web navigation. In the recent development in technologies, some concepts of ontologies, genetic-based algorithms, and artificial neural networks are used to recommend a suitable course coverage plan [29,30].
ITS developed by [31] utilizes the nature-inspired algorithm to adjudge the custom-tailored learning path. They used two critical parameters for fitness function; one is difficulty level and other associations between the course concepts. Another research work on e-learning systems makes use of the nature-inspired algorithm for determining optimal course coverage plans based on the incorrect response on the pretest [32,33].
In [34], the author used ACO techniques to recommend an adaptive learning path by considering the learner’s learning style. The authors of [35] utilized self-organizing techniques to recommend optimal course coverage plans to the learner. Similar practices can be applied through a probabilistic technique. In which nodes indicate the pedagogy items and edges indicate the hypertext links (preferred probabilities), and learners act as ants who have to traverse all the nodes [36,37,38] (Skinner. et al.; Brusilovsky. et al., 1996; Karampiperis. et al., 2004; Dhabi. et al., 2006; Bert van den Berg. et al., 2005).
Evaluating and validating an ITS is a challenging task due to the lack of standard agreement and procedure. The primarily used prototype for evaluating the training program of ITS is the prototype established by Donald Kirkpatrick [39,40]. Many researchers revised this prototype, but its basic architecture is still the same [41,42].
Kirkpatrick’s prototype comprises four stages of evaluation, shown in Figure 1 and briefly discussed in Figure 2.
Computing results is the best way to quantify the effectiveness of any learning program, but it is challenging to conduct. The authors of Ref. [48] provided a statement regarding the evaluation that not all learning programs focus on impacting the learning performance of a learner; instead, they can perform for a purpose. Other researcher uses the Kirkpatrick model as a base model. The author of Ref. [49] introduced the fifth level of the Kirkpatrick model named Return of Investment (ROI), which is used to measure the effectiveness of learning or investment.
Kirkpatrick’s model is considered a base model for judging the efficacy of a training program by Ref. [48]. Their suggestion is to set the initial objective (from the learning program) and then monitor the fulfillment of objectives after the learning program. The authors of Ref. [50] disapprove of Kirkpatrick’s model, giving the following reason:
  • Offline Test (Written Test) lacks validity and reliability in quantifying knowledge, skill, and attitude (KSA);
  • The 100% response rate is idealistic;
  • Control groups are not feasible in the learning program context.
Furthermore, educational organizations recommend considering the merits and demerits of various evaluation prototypes and methodologies to build an organizational-specific evaluation prototype that fulfills their requirements [51]. Furthermore, it was proposed that the evaluation prototype prioritizes both education practices and findings [52]. Valid, reliable, inexpensive, and acceptable is the feature of an ideal evaluation prototype. Moreover, the evaluation prototype may encompass quantitative, objective, subjective, and qualitative methods. Thus, evaluation results are advantageous for determining the learning attainments of the learning program [53,54,55,56].
The conclusion drawn from the literature recommends that the learning program include three levels of Kirkpatrick’s prototype. Kirkpatrick Level 3 and level 4 are very challenging to observe in an educational learning program [57] because level 1 and level 2 can be quantified during an ongoing learning session. However, levels 3 and level 4 require post-assessment analysis. Level 4 needs the rigorous observation of the inference of the learning program. For evaluation, there is no articulated framework. As per [57], continuous feedback and instruction help the learner achieve the best learning skills [58,59,60,61,62].

3. Proposed Prototype

This section illustrates the SeisTutor architecture and briefly discusses functionality incorporated in SeisTutor. SeisTutor is an ITS explicitly designed for the “Seismic Data Interpretation” subject domain. SeisTutor recommends learning contents as per learner performance in pretest (prior knowledge assessment test). In addition, SeisTutor keeps track of the learner’s behavior (psychological state) during the entire learning session and conducts a test to determine the degree of understanding of the topic. SeisTutor is adaptive, i.e., content and link-level [63]. Adaptive content level indicates that the learners with different performances in the pretest (prior knowledge test) obtain different learning material. Before indulging learners in the learning session, SeisTutor enables learners to go through a pretest. There are two kinds of assessments performed in pretest: a learning style test and a prior knowledge test. The current focus of this research is on the prior knowledge test. SeisTutor observes the learner’s performance during the tests and aligns the learning material accordingly.
For link elimination, ideally, the curriculum covers all the links of sub-topics to learn—however, the links eliminated are not in the determined curriculum for the learner. SeisTutor continually observes and stores the learner’s action, performance, and behavior. This information is further utilized for making intelligent strategic plans for recommendation and evaluation work.

3.1. SEISTUTOR Architecture

SeisTutor follows all the guidelines for implementing an intelligent tutoring system. The critical functional model of an ITS is the domain model, learner model, pedagogy model, and learner interface. SeisTutor architecture comprises a learner interface, pedagogy model, learner model, and domain model [64,65,66]. The following subsection briefly illustrates the various components involved in SeisTutor.

3.1.1. Domain Model

This model is described as a cluster of concepts. Here concept terminologies indicate the single topic. Other terminologies are used in the different research papers, such as knowledge element, object, learning outcome, and attribute. In the current context, concepts possess a prerequisite relationship to each other. Each concept is further segregated into learning units. SeisTutor utilizes unit variant technologies to attain content-level adaptation. Content level adaptation indicates that the system has alternative units and recommends the learning units based on learner grasping and learning style [67,68,69].

3.1.2. Pedagogy Model

This model consists of various rules and logic that build a knowledge infrastructure essential for adapting the learning materials as per the learner’s:

Curriculum Planner

The planner generates a curriculum as a sequence of learning units covered during the learning session [64];

Learning Assessment

Learning assessment is the process of determining the learner’s learning process. The accuracy of this adjudging model acts as a critical factor that affects the adaptation practice;

Understanding Assessment

The understanding assessment identifies the learner’s degree of understanding of the concept. SeisTutor makes any decision by referring to all the models, such as determining the curriculum for the learner and offering the learning content as per the learner’s grasping level and learning style.

3.1.3. Learner Model

SeisTutor captures three characteristics of a learner. This model captures the learner’s activity during the learning session and stores and updates the learner’s information for making the decision. This information is fruitful for the system to adapt based on learner characteristics (grasping level and learning style).

Learner Demographic Information

Learner demographic information includes learner’s basic details, such as name, learner username, and email id. This information is used to create the learner profile and collected when the first-time learner signup with SeisTutor.

Psychological State

It recognizes the learner’s emotions during the tutoring session. Recognizing the psychological state of mind is essential because it helps determine how far the learner liked the learning contents during the learning session. SeisTutor determines six emotions, i.e., smile, sad, neutral, surprise, fear, and anger.
This psychological state recognition module becomes triggered as soon as the learner starts the learning session, as shown in Figure 3. Initially, participants belong to both the studies and undergo the initial assessment phase (Pretest); after the pretest, their learner profile and learning style are determined. The I2A2 learning style prototype is utilized to ascertain the learning style [67]. Here I2A2 is the name of the learning style, i.e., imagistic (learning through observing), intuitive (learning through interpreting written words), acoustic (learning through listening), active (learning through action). Based on tutoring strategy (study 2 (learner study without customized learning) and study 1 (learner study with customized learning)) and course coverage plan, the suitable mode of learning (study 1 (learner study with customized learning)) is chosen by the learner and pedagogy model and offers customized content through learner interface. Thus, SeisTutor examines the participant’s attainment in the pretest and predicts the individualized tutoring strategy (study 2 (learner study without customized learning) and study 1 (learner study with customized learning)) and custom-tailored curriculum for the applicants (study 1 (learner study with customized learning)).
As illustrated in Figure 3, the CNN-based emotion recognition module is instantiated as soon as the learner begins the learning session. This emotion recognition module takes the snap of the learner via webcam, which acts as an input to the CNN-based emotion recognition module. This module determines the learner’s psychological (emotion) state, which is gauged for future analysis (phase 1: Evaluation of Reaction). The gathering psychological (emotion) state is repeated until the learner completes all the learning contents (topics) associated with all the weeks.

Learner Performance

It evaluates the learner’s performance by quantifying their learning by organizing quizzes and tests. Learner performance was quantified based on one parameter, i.e., the number of correct responses. As soon as learners finish the learning content of each week, they have to complete a quiz and understanding test. One quiz is associated with every week (shown in Figure 4). Each quiz contains five questions, and each question contains only one hint. Hints appear to the learner based on the learner’s request to seek help to solve the question. SeisTutor asks the learner to summarize the learned concepts in understanding the test. Based on the user entered information, dictionary-based sentimental analysis is performed. The result gives a score out of 100, which tells the learner’s overall understanding of the concepts.
In order to determine the overall learning gain, SeisTutor has pretest and post-test assessment scores. The average learning gain is computed (using Equation (1)).
L e a r n i n g _ G a i n = P o s t T e s t _ S c o r e L P r e T e s t _ S c o r e L

Learning History

It keeps track of the learner’s activity, such as login time, total time spent, etc., and its interaction details during the entire tutoring session. SeisTutor utilizes this information to make necessary action, which further helps to make the whole learning process effective.
In order to start tutoring with SeisTutor, learners need to register themselves first (see Figure 5). After registration, SeisTutor creates a learner account and instructs the learner to complete a pretest. Pretests are comprised of two tests:
  • Prior Knowledge Test;
  • Learning Style Test.

Prior Knowledge Test

The prior or Previous Knowledge test is the preliminary test used to identify the domain’s elementary knowledge to know the learner’s initial learning level. This test is responsible for analyzing the learner’s knowledge about the domain. The domain model examined the test result and categorized the learner into three learning profiles, i.e., “beginner, intermediate and expert” (see Figure 5). A further outcome of this result is also responsible for determining the curriculum, which is exclusively designed for the learner. This test comprises twenty questions, and all the questions are verified by the domain expert of seismic data interpretation.

Learning Style Test

This test is responsible for determining preferred media for learning. The research has noticed that learner performance is gradually increased if learning material is provided as per preferred learning media (see Figure 5). Thus, this test analyses the learn-er-preferred media. The learner model examines the test result and categorizes the learner into four learning styles, i.e., “Imagistic, Acoustic, Intuitive, and Active”.
Based on these two tests, the pedagogy model determines the tutoring strategy (the result of learning profile and learning style). A tutoring session then begins with the determined tutoring strategy. As soon as learning begins, psychological features are triggered, capturing learner emotions and saving the results in the database for future reference. After completion of every week, the learner is tested. Based on the test result, learner performance and degree of understanding are computed.

3.1.4. Learner Interface Model

This model helps to establish communication between the learning system and the learner via multiple visualization tools (audio, video, textual, navigation, etc.).

4. Evaluation of SeisTutor

After tutoring, post-tutoring assessments of learners are performed by SeisTutor. This section depicts the analysis used to determine the impact of both the studies, i.e., study 1 (learner study with customized learning) and study 2 (learner study without customized learning) learning methods with SeisTutor. Study 1 (learner study with customized learning) exercises learning practices with features such as personalized curriculum design, recognition of the psychological state of the learner during the learning process, and learner’s degree of understanding of taught concepts. When it is not practiced, it is characterized as study 2 (learner study without customized learning). An experimental comparison of both the study groups was tabulated (see Figure 6 and Table 1). These variances help to identify the discrepancy in the learning experiences.

4.1. Experimental Design and Methodology

The SeisTutor evaluation is the fundamental piece of the development of this framework. In order to quantify the adequacy and effectiveness of SeisTutor, assessment tests were conducted. The SeisTutor was tested on a selected population of students, teachers, and both (teachers and students) from an anonymous university. A total of 60 learners was volunteered in the evaluation process. Based on their compliance in the participation, a compliance agreement form was issued that demonstrates essential details related to the assessment process. It is requisite for each applicant to give their approval for participation in the assessment process. Applicants were haphazardly assigned one of the groups. Thirty-two applicants were in study 1 (learner study with customized learning), and the remaining were in study 2 (learner study without customized learning).
Out of 60 learners, 30% are learners pursuing graduation, 17% are graduates, 35% are post-graduates, and the remaining minimum qualifications are doctorate (Ph.D.).
Twelve percent of learners were in the 18–20 age group, 18% of learners were in the 20–22 and 22–24 age group, 5% of learners were in the 24–28 age group, 12% of learners were in the 28–32 age group, 22% of learners were in the 32–34, and the remaining were in thh above 34 age group (shown in Figure 7. SeisTutor is explicitly created for the “Seismic-Data-Interpretation” domain. As a result, it is intended to be used by participants or learners belonging to the petroleum engineering and exploration domain. Thus, to quantify the effectiveness of SeisTutor, undergraduate learners (B.Tech/B.E. Petroleum engineering), teachers (Petroleum Engineering dept), and others (government exploration industry) are taken into consideration (refer to Figure 7).
The learner underwent a pretest as soon as they were registered with SeisTutor. Their learning style and grasping level (learning level) were adjudged as 28 learners, as mentioned above, were in study 1 (learner study with customized learning). Based on their responses in the pretest, the custom-tailored curriculum was determined, which is realigned and reorganized from the domain (content) capsule. SeisTutor examines every learner involved in study 1 (learner study with customized learning), identifies their psychological state (emotions) during the learning session, and quantifies the degree of understanding of the concepts. The remaining learners in study 2 (learner study with customized learning) follow the standard curriculum for learning, i.e., contents in the same sequence (irrespective of their prior knowledge about the domain). Thus, all the learners follow the same learning path. The point to underline is that their pretest performance is not used for exclusive learning path recommendations. The learning session begins weekly for both the study groups and subsequent post-assessment tests.

4.2. Data Preparation

Before analysis, obtained data underwent a data screening phase. In this phase, the elimination of missing values and data normalization was performed. For deducing conclusion about the effectiveness of learning through SeisTutor, learner’s performances, i.e., prior knowledge test (pretest), understanding test, psychological state results, and quiz test (post-test) during learning, were taken into consideration. SPSS version 25 was used for accomplished analysis.

4.3. Min Max Normalization

The learner classification parameters (prior knowledge test (0–5), post-assessment test (0–5), learning gain (0–5), learner emotion (0–100)) of each learner were used and normalized using the methods called Min–Max. It converts a value of Y = y 1 , y 2 , y 3 , , y n , and converges in the range of [A,B]. The formula for score standardization is specified below, where A is the lowest range; B is the highest range. In our case [A,B] is [0, 10];
Z = y i L o w e s t   v a l u e   i n   Y   H i g h e s t   v a l u e   i n   Y L o w e s t   v a l u e   i n   Y   × B A + A

5. Result and Discussions

The evaluation of SeisTutor was performed using the Kirkpatrick evaluation model. As discussed in Section 2, the Kirkpatrick evaluation prototype comprises of four-phase shown in Figure 1 and Figure 2.

5.1. Kirkpatrick Phase 1: Evaluation of Reaction

Evaluation of reaction as its name indicates the reaction (emotion) of the learner during learning or how far the learner likes the learning content and teaching process, i.e., pedagogy. SeisTutor incorporates an emotion recognition module and an open-end questionnaire (Learner Feedback).
The Min–Max normalization utilizes to maintain the uniformity, which converges the original values in the scope of [0, 10].
As mentioned above in Table 1, the psychological state of the learner was determined only for the applicants involved in study 1 (learner study with customized learning). Thus, their descriptive states are shown in Figure 8. From the stats shown in Figure 9, the average mean score percentage among 28 applicants is 44% for the emotion happy, 40% for the emotion neutral, 36% for the emotion angry, 32% for the emotion surprise, 30% for the emotion fear, and 24% for the emotion sad. Thus, one can deduce with confidence that, on average, learners are happy with the learning content and teaching process, i.e., pedagogy.

5.2. Kirkpatrick Phase 2: Evaluation of Learning

Evaluation of learning, as its name indicates, is how effectively learners grasp the learning content. SeisTutor conducts a small quiz and degree of the understanding test to adjudge the learner’s overall learning.
The average learning gain of applicants involved in study 1 (learner study with customized learning) is 22%, and in study 2 (learner study without customized learning), it is 12%. Thus, it is concluded that if learning material is offered as per the learner’s inclination with an exclusively designed curriculum based on the learner’s previous knowledge, then the proposed SeisTutor succeeds in enhancing the learner’s curiosity and interest, which indirectly enhance the overall learning gain (refer Figure 9 and Figure 10).
These tests were performed on learning gain and degree of understanding of study 1 (learner study with customized learning) through SeisTutor. Figure 10 illustrates the progressive learning gain of 22% discovered among learners, those applicants who participated in study 1 (learner study with customized learning) SeisTutor. Furthermore, this information (statistical) was used in the analysis, i.e., Bivariate Pearson Correlation.
Understanding tests were designed and conducted only for study 1 (learner study with customized learning) because it strongly proves the effectiveness of learning gain achieved by intelligent features incorporated SeisTutor. Here, the correlation of learning gain with itself is one because a variable or parameter is perfectly interrelated. The Pearson correlation of learning gain with a degree of understanding is 0.484, and the significance is two-tailed, i.e., the p-value is less than 0.01 (refer to Table 2 and Table 3). Thus, one can confidently say that Learning gain and degree of understanding are statistically significant linear relationships. (see Figure 11).

5.3. Kirkpatrick Phase 3: Evaluation of Behaviour

Evaluation of behavior is quite tuff to quantify. In order to measure this, SeisTutor collects feedback from the applicants. As the learner completes all the learning concepts every week, SeisTutor requests the learner’s feedback. In this section, a conclusion from the learner feedback is drawn.
Learners who were the impeccable part of this evaluation of SeisTutor had a good perception of the system, and their feedbacks were very encouraging. It was shown in their reactions whether they would like to recommend the SeisTutor to others who need to take this study. Around 93% of the learners showed that they would recommend it to others, out of which 48% showed strong agreement, and the remaining 45% agreed on a recommendation as well (see Table 4 and Figure 12). The overall satisfaction with SeisTutor was around 93%, out of which 45% were strongly satisfied, and 48% were satisfied as well. It was also observed that learners’ studies became productive with the SeisTutor.
Few questions were asked on the impact of the intelligent features provided by SeisTutor and were collected and summarized in Table 4 As some intelligent features are not provided for study 2 (learner study without customized learning) applicants, 28 effective feedback was taken into consideration from study 1 (learner study with customized learning) participants.
Most of the participants were happy with the tutoring strategy provided by the system, with 86% satisfaction, which includes 46% who were satisfied and 40% who were strongly satisfied. Eighty-five percent of participants felt that learning from their own learning experience made them perform better, with 40% who were strongly satisfied and 45% were satisfied. The participants were happy with the recommended exclusive curriculum by the system with 85% satisfaction, which includes 35% who were satisfied and 50% who were strongly satisfied. Most of the participants were happy with the recommended custom-tailored curriculum provided by the system, with 85% satisfaction, which includes 39% who were satisfied and 46% strongly satisfied. Ninety-two percent of participants agreed that the understanding test each week corresponds to the lesson taught, 39% strongly agreed, and 53% also agreed. Lastly, 82% of students agreed with the psychological parameter accurately determined with SeisTutor, with 39% who strongly agreed to this and 43% who were satisfied (see Table 5 and Figure 12).
The overall impact of the support provided by SeisTutor on the learning process was assessed through the learner’s feedback questionnaire answered by 60 participants (see Table 6 and Figure 12). The analyzed results showed that 87% of the students were happy with SeisTutor supports, with 47% satisfied and 40% strongly satisfied. In addition, 78% of the students were happy with the system navigation support to find the needed information, with 43% satisfied and 35% strongly satisfied. Lastly, out of 80% of the students, 42% had a strong agreement, and 38% agreed that the SeisTutor pre-learning procedure was beneficial for learning.
The usefulness of the lesson components such as lesson explanations, revisions, presented quizzes, and the question hints in the learning process were evaluated in Table 7. The questionnaire feedback results show that 85% of students were happy with the content explained by SeisTutor, with 47% satisfied and 38% strongly satisfied. Moreover, 78% of students showed interest and agreed that the tutoring resources were adequate, with 35% strongly satisfied and 43% satisfied. It is clear that the quizzes and hints were realistic and focused on the learning contents provided by SeisTutor.
The impact of the interactive graphical user interface, content organization, and design features of SeisTutor in the learning process was evaluated through the learner questionnaire responses described in Table 8. The questionnaire results revealed that 77% of the learners were 40% strongly satisfied and 37% satisfied with the interactive GUI and content organization of SeisTutor. Eighty percent of the learners were 44% strongly satisfied and 36% satisfied with SeisTutor to compel and support to complete the quizzes and lessons. Finally, students were happy with the account setup process with the system, which maintains the learners learning progress, grades, and basic account information.
The learner’s overall evaluation of the SeisTutor showed that 82% of learners agreed that tutoring should begin based on the learner profile considering their learning style and previous knowledge. Most of the students were unaware of their learning style, and about 80% of students never knew about it. Most of the students liked the artificial intelligence features such as automatic selection of the tutoring strategies, dynamically assessing the learner attainment, and flipping the tutoring plan or strategy.
The learner’s feedback questionnaire responses were retrieved and analyzed in an accessible fashion. Some learners put their suggestions to improve the productivity of SeisTutor. Most of the suggestions were general and related to the improvement of the system, and few were pessimistic regarding improvement of the quality of learning contents, improving the quality of the video lessons, and hints provided by the system. At last, through the overall evaluation of SeisTutor, 87% of learners agreed that they improved their learning performance and outcomes.

5.4. Kirkpatrick Phase 4: Evaluation of Results

Evaluation of results illustrates the overall impact of learning on the learner. In order to quantify the effectiveness of learning, paired wise sample T-test was performed on existing information, i.e., pretest and post-test results of participants involved in both the studies (study 1 (learner study with customized learning) and study 2 (learner study without customized learning)). Here two cases are taken into consideration.
Case_1: 
A Paired_Sampled_T-Test performed on study 1 (learner study with customized learning) consists of intelligent features depicted in Table 1;
Hypothesis_Case_1.0: 
Let the participants involved in study 1 (learner study with customized learning) have similar pretest and post-test mean scores;
Hypothesis_Case_1.1: 
Let the participants involved in study 1 (learner study with customized learning) not have similar pretest and post-test mean scores;
Case_2: 
A Paired-Sampled-T-Test performed on study 2 (learner study without customized learning), which consist of feature depicted in Table 1;
Hypothesis_Case_2.0: 
Let the participants involved in study 2 (learner study without customized learning) have similar pretest and post-test mean scores;
Hypothesis_Case_2.1: 
Let the participants involved in study 2 (learner study without customized learning) not have similar pretest and post-test mean scores.
The calculated T value ( T S t a t s ,) for study 1 (learner study with customized learning) is 11.410, p < 0.01 (see Table 9 and Table 10. On average post-test scores were 2.21786 points higher than pretest scores. Here the calculated T S t a t s is greater that T c r i t i c a l , thus hypothesis 1.0 is rejected. From Table 9 and Table 10 one can deduce with confidence that there is a significant difference between pretest and post-test scores. The calculated T value ( T S t a t s ,) for study 2 (learner study without customized learning) is 5.312, p < 0.01 (see Table 11 and Table 12). On average, post-test scores were 1.24719 points higher than pretest scores. Here the calculated T S t a t s , is greater than T c r i t i c a l , thus hypothesis 2.0 is rejected. From Table 11 and Table 12, one can deduce with confidence that there is a significant difference between pretest and post-test scores.
Both the studies reject the null hypothesis, which means both the studies provide effective training. However, this research aims to identify which study has a high impact on enhancing learning gain. In conclusion, the aim of both the studies was compared. Study 1 (learner study with customized learning) is higher than Study 2 (learner study without customized learning). Thus, study 1 (learner study with customized learning) has a significant difference in the post tutoring and pretest scores and provides more effective training than study 2 (learner study without customized learning).
From the research, it was observed that quality education plays a vital role in providing sustainable development of society and its need. Quality education includes a personalized learning environment and teaching process.
Learners of all ages are empowered with the knowledge and skills to confront the linked global challenges that we face, including climate change, environmental degradation, and the loss of biodiversity. Education for sustainable development empowers learners of all ages.
An intelligent tutoring system named SeisTutor was developed to accomplish the objective of personalized learning, which observes the learner’s preferences, grasping level, and knowledge level and recommends a personalized curriculum (learning environment) and personalized teaching process. This analysis concludes that the intelligent incorporated SeisTutor used in study 1 (learner study with customized learning) outperforms the SeisTutor used in study 2 (learner study without customized learning) in providing custom-tailored intended curriculum, identifying learner sentiments while learning, and computing the learner’s overall degree of knowledge that meets the learner’s sustainable needs. Thus, SeisTutor used in study 1 can provide effective quality education for sustainability.
Practicing sustainability empowers humans to construct knowledge, explore values, and develop an appreciation of the environment and its relationship to their worlds. This lays the foundations for sustainable education.

6. Conclusions

This article demonstrates the proposed sustainable personalized intelligent tutoring system prototype, named as SeisTutor. From the research, it was noted that a learner receives repetitive learning content, which indirectly disorients the learner. Thus, to address this issue, a bug model was utilized, which analyzes the bugs and recommends the custom-tailored curriculum to the learner. This technique helps to bring empathy to ITS. SeisTutor is not a passive tutor; it also analyses the learner’s behavior, i.e., the psychological state of the learner during learning, which helps to understand the learner’s experience with SeisTutor (about learning content). The experimental results revealed that SeisTutor utilized by participants in study 1 (learner study with customized learning) provided a customized learning sequence or path of learning material that endorses effective learning. Experimental analysis reveals the effective learning gain (when the learner receives the custom-tailored sequenced learning material) of 44.34% compared to SeisTutor used in study 2 (learner study without customized learning) (not sequenced learning material). In order to evaluate the overall effectiveness of SeisTutor, the Kirkpatrick four-phase evaluation model was utilized. The analysis reveals that the participants involved in study 1 (learner study with customized learning) attain, while study 2 (learner study without customized learning) attains study 1 (learner study with customized learning) provides effective learning.

Author Contributions

Conceptualization, N.S. and V.K.G.; methodology, N.S.; software, A.K.M.; validation, A.K.M. and R.K.M.; formal analysis, N.S.; investigation, V.K.G.; resources, N.N.; data curation, R.K.M. and N.N.; writing—original draft preparation, N.S.; writing—review and editing, N.S. and V.K.G.; visualization, A.K.M.; supervision, R.K.M. and V.K.G.; project administration, R.K.M. and N.N.; funding acquisition, N.N. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ghosn-Chelala, M. Exploring sustainable learning and practice of digital citizenship: Education and place-based challenges. Educ. Citizsh. Soc. Justice 2018, 14, 40–56. [Google Scholar] [CrossRef]
  2. Dybach, I. Institutional aspects of educational quality management in higher educational establishments. Econ. Dev. 2019, 18, 33–43. [Google Scholar] [CrossRef]
  3. Boeren, E. Understanding Sustainable Development Goal (SDG) 4 on “quality education” from micro, meso and macro perspectives. Int. Rev. Educ. 2019, 65, 277–294. [Google Scholar] [CrossRef] [Green Version]
  4. Guijarro, F.; Poyatos, J. Designing a Sustainable Development Goal Index through a Goal Programming Model: The Case of EU-28 Countries. Sustainability 2018, 10, 3167. [Google Scholar] [CrossRef] [Green Version]
  5. Hanemann, U. Examining the application of the lifelong learning principle to the literacy target in the fourth Sustainable Development Goal (SDG 4). Int. Rev. Educ. 2019, 65, 251–275. [Google Scholar] [CrossRef]
  6. Bush, T. Research on educational leadership and management. Educ. Manag. Adm. Leadersh. 2018, 46, 359–361. [Google Scholar] [CrossRef] [Green Version]
  7. Singh, S.K. The human side of management. Int. J. Educ. Manag. 2019, 33, 2–4. [Google Scholar] [CrossRef] [Green Version]
  8. Siemens, G. Connectivism: A Learning Theory for the Digital Age. 2004. Available online: www.elearnspace.org/Articles/connectivs.htm (accessed on 29 May 2021).
  9. Beetham, H.; Sharpe, R. Rethinking Pedagogy for Digital Age: Designing and Delivering E-Learning; Routledge: Abingdon, UK, 2007. [Google Scholar]
  10. Samah, N.A.; Yahaya, N.; Ali, M.B. Individual differences in online personalized learning Environment. Educ. Res. Rev. 2011, 6, 516–521. [Google Scholar]
  11. De Castell, S.; Droumeva, M.; Jenson, J. Building as interface: Sustainable educational ecologies. Medien. Z. Für Theor. Und Prax. Der Medien. 2014, 24, 75–93. [Google Scholar] [CrossRef] [Green Version]
  12. Burbules, N.C. Meanings of “ubiquitous learning”. In Ubiquitous Learning; Cope, B., Kalantzis, M., Eds.; University of Illinois Press: Urbana, IL, USA, 2009; pp. 15–20. [Google Scholar]
  13. Sharples, M.; De Roock, R.; Ferguson, R.; Gaved, M.; Herodotou, C.; Koh, E.; Kukulska-Hulme, A.; Looi, C.; McAndrew, P.; Rienties, B.; et al. Innovating Pedagogy 2016: Open University Innovation Report 5; Keynes, M., Ed.; Institute of Educational Technology, Open University: Edinburgh, Scotland, 2016. [Google Scholar]
  14. Burbules, N.C. El aprendizaje ubicuo: Nuevos contextos, nuevos procesos. Rev. Entramados Educ. Y Soc. 2014, 1, 131–135. [Google Scholar]
  15. Burbules, N.C. Ubiquitous learning and the future of teaching. In Teacher Education in a Transnational World; Bruno-Jofre, R., Johnston, S., Eds.; University of Toronto Press: Toronto, ON, Canada, 2014; pp. 177–187. [Google Scholar]
  16. Lawler, R.W.; Yazdani, M. (Eds.) Artificial Intelligence and Education: Learning Environments and Tutoring Systems; Intellect: Bristol, UK, 1987. [Google Scholar]
  17. Nwana, H.S. Intelligent tutoring systems: An overview. Artif. Intell. Rev. 1990, 4, 251–277. [Google Scholar] [CrossRef]
  18. Skinner, B.F. Teaching Machines. Science 1961, 205, 90–106. [Google Scholar]
  19. Uhr, L. Teaching machine programs that generate problems as a function of interaction with students. In Proceedings of the 1969 24th National Conference, New York, NY, USA, 26–28 August 1969; pp. 125–134. [Google Scholar]
  20. Sleeman, D.; Brown, J.S. Introduction: Intelligent Tutoring Systems; Intelligent Tutoring Systems, Ed.; Academic Press: Cambridge, MA, USA, 1982; pp. 1–11. [Google Scholar]
  21. Woods, P.; Hartley, J.R. Some learning models for arithmetic tasks and their use in computer based learning. Br. J. Educ. Psychol. 1971, 41, 38–48. [Google Scholar] [CrossRef]
  22. Suppes, P. On using computers to individualize instruction. In The Computer in American Education; Association for Educational Data Systems: Washington, DC, USA, 1967; pp. 11–24. [Google Scholar]
  23. Gertner, A.S.; VanLehn, K. Andes: A coached problem solving environment for physics. In Proceedings of the 5th International Conference, Berlin, Germany, 8–11 June 2022; pp. 133–142. [Google Scholar]
  24. Conati, C.; Gertner, A.; Vanlehn, K.; Druzdzel, M.J. Online student modeling for coached problem solving using Bayesian networks. In User Modeling; Springer: Berlin/Heidelberg, Germany, 1997; pp. 231–242. [Google Scholar]
  25. Chen, L.H. Enhancement of student learning performance using personalized diagnosis and remedial learning system. Comput. Educ. 2011, 56, 289–299. [Google Scholar] [CrossRef]
  26. Haoran, X.; Di, Z.; Fu, L.W.; Tak-Lam, W.; Yanghui, R.; Simon, H.W. Discover Learning Path for Group Users: A ProÞle-based Approach. Neurocomputing 2017, 254, 59–70. [Google Scholar]
  27. Hsieh, T.-C.; Wang, T.-I. A mining based approach on discovering courses pattern for constructing suitable learning path. Expert Syst. Appl. 2010, 37, 4156–4167. [Google Scholar] [CrossRef]
  28. Li, Y.; Huang, R. Dynamic Composition of Curriculum for Personalized E-Learning; Knowledge Science & Engineering Institute, Beijing Normal University: Beijing, China, 2006. [Google Scholar]
  29. Karampiperis, P.; Sampson, D. Adaptive Instructional Planning using Ontologies. In Proceedings of the IEEE International Conference on Advanced Learning Technologies (ICALT’04), Joensuu, Finland, 30 August–1 September 2004. [Google Scholar]
  30. Dron, J. Achieving Self-Organisation in Network-Based Learning Environments. Ph.D. Thesis, University of Brighton, Brighton, UK, 2002. [Google Scholar]
  31. Chen, C.M. Intelligent web-based learning system with personalized learning path guidance. Comput. Educ. 2008, 51, 787–814. [Google Scholar] [CrossRef]
  32. Agbonifo, O.C.; Obolo, O.A. Genetic Algorithm-Based Curriculum Sequencing Model for Personalised E-Learning System. Int. J. Educ. Manag. Eng. 2018, 5, 27–35. [Google Scholar] [CrossRef]
  33. Seridi, H.; Sari, T.; Sellami, M. Adaptive Instructional Planning in Intelligent Learning Systems. In Proceedings of the IEEE International Conference on Advanced Learning Technologies (ICALT’06), Kerkrade, The Netherlands, 5–7 July 2006. [Google Scholar]
  34. Hong, C.-M.; Chen, C.-M.; Chang, M.-H.; Chen, S.-C. Intelligent Web-based tutoring System with Personalized Learning Path Guidance. In Proceedings of the IEEE International Conference on Advanced Learning Technologies (ICALT’07), Niigata, Japan, 18–20 July 2007. [Google Scholar]
  35. Bonabeau, E.; Dorigo, M.; Theraulaz, G. Swarm Intelligence: From Natural to Artificial Systems; Oxford University Press: Oxford, UK, 1999. [Google Scholar]
  36. Brusilovsky, P.; Schwarz, E.; Weber, G. ELMART: An intelligent Tutoring System on World Wide Web. In Third International Conference on Intelligent Tutoring Systems ITS-96; Frasson, C., Gauthier, G., Lesgold, A., Eds.; Springer: Berlin, Germany, 1996; pp. 261–269. [Google Scholar]
  37. Dahbi, A.; Elkamoun, N.; Berraissoul, A. Adaptation and Optimisation of Pedagogical Paths by Ant’s Algorithm. In Proceedings of the IEEE Information and Communication Technology (ICTTA’06), Damascus, Syria, 24–28 April 2006. [Google Scholar]
  38. Van den Berg, B.; Van Es, R.; Tattersall, C.; Janssen, J.; Manderveld, J.; Brouns, F.; Kurves, H.; Koper, R. Swarm-based Sequencing recommendations in E-Learning. In Proceedings of the 2005 5th International Conference on Intelligent Systems Design and Applications (ISDA’05), Warsaw, Poland, 8–10 September 2005. [Google Scholar]
  39. Kirkpatrick, D.L. Techniques for evaluating training programs. Train. Dev. J. 1979, 33, 78–92. [Google Scholar]
  40. Kirkpatrick, D. Evaluating Training Programs; Berrett-Koehler: San Francisco, CA, USA, 1994. [Google Scholar]
  41. Bates, R. A critical analysis of evaluation practice: The Kirkpatrick model and the principle of beneficence. Eval. Program Plan. 2004, 27, 341–347. [Google Scholar] [CrossRef]
  42. Endres, G.J.; Kleiner, B.H. How to measure management training and development effectiveness. J. Eur. Ind. Train. 1990, 14, 3–7. [Google Scholar] [CrossRef]
  43. Kirkpatrick, D.L.; Kirkpatrick, J.D. The Transfer of Learning to Behavior: Using the Four Levels to Improve Performance; Berrett Koehler Publication: San Francisco, CA, USA, 2005. [Google Scholar]
  44. Kirkpatrick, D.L.; Kirkpatrick, J.D. Evaluating Training Programs: The Four Levels, 3rd ed.; Berrett Koehler Publication: San Francisco, CA, USA, 2006. [Google Scholar]
  45. Kirkpatrick, D.L.; Kirkpatrick, J.D. Implementing the Four Levels; Berrett Koehler Publication: San Francisco, CA, USA, 2007. [Google Scholar]
  46. Knowles, M.S.; Holton, E.F.; Swanson, R.A. The Adult Learner, 5th ed.; Gulf Publishing Company: Houston, TX, USA, 1998. [Google Scholar]
  47. Ehlers, U.D.; Schneckenberg, D. Changing Cultures in Higher Education: Moving Ahead to Future Learning; Springer: Berlin/Heidelberg, Germany, 2010. [Google Scholar]
  48. McEvoy, G.M.; Buller, P.F. Five uneasy pieces in the training evaluation puzzle. Train. Dev. J. 1990, 44, 39–42. [Google Scholar]
  49. Phillips, J.J. A rational approach to evaluating training programs including calculating ROI. J. Lend. Credit. Risk Manag. 1997, 79, 43–50. [Google Scholar]
  50. Reeves, T.C.; Hedberg, J.G. Interactive Learning Systems Evaluation; Educational Technology: Englewood Cliffs, NJ, USA, 2003. [Google Scholar]
  51. Fitzpatrick, J.L.; Sanders, J.R.; Worthen, B.R. Program Evaluation: Alternative Approaches and Practical Guidelines, 3rd ed.; Allyn and Bacon: Boston, MA, USA, 2004. [Google Scholar]
  52. Musal, B.; Taskiran, C.; Gursel, Y.; Ozan, S.; Timbil, S.; Velipasaoglu, S. An example of program evaluation project in undergraduate medical education. Educ. Health 2008, 21, 113. [Google Scholar]
  53. Morrison, J. ABC of learning and teaching in medicine: Evaluation. Br. Med. J. 2003, 326, 385–387. [Google Scholar] [CrossRef]
  54. Harden, A.; Peersman, G.; Oliver, S.; Mauthner, M.; Oakley, A. A systematic review of the effectiveness of health promotion interventions in the workplace. Occup. Med. 1999, 49, 540–548. [Google Scholar] [CrossRef] [Green Version]
  55. Steinert, Y.; Mann, K.; Centeno, A.; Dolmans, D.; Spencer, J.; Gelula, M.; Prideaux, D. A systematic review of faculty development initiatives designed to improve teaching effectiveness in medical education: BEME Guide No. 8. Med. Teach. 2006, 28, 497–526. [Google Scholar] [CrossRef]
  56. Malki, A.A.; Al-Bareeq, J.M.; Al-Halili, A. Evaluation of research writing workshop. Bahrain Med. Bull. 2003, 25, 127–130. [Google Scholar]
  57. Praslova, L. Adaptation of Kirkpatrick’s four level model of training criteria to assessment of learning outcomes and program evaluation in Higher Education. Educ. Assess. Eval. Account. 2010, 22, 215–225. [Google Scholar] [CrossRef]
  58. Bitchener, J.; Young, S.; Cameron, D. The effect of different types of corrective feedback on ESL student writing. J. Second Lang. Writ. 2005, 14, 191–205. [Google Scholar] [CrossRef]
  59. Storch, N.; Tapper, J. Discipline specific academic writing: What content teachers comment on. High. Educ. Res. Dev. 2000, 19, 337–356. [Google Scholar] [CrossRef]
  60. Polio, C.; Fleck, C.; Leder, N. “If I Only Had More Time:” ESL Learners’ Changes in Linguistic Accuracy on Essay. J. Second. Lang. Writ. 1998, 7, 43–68. [Google Scholar] [CrossRef]
  61. Elliot, N.; Klobucar, A. Automated essay evaluation and the teaching of writing. In Handbook of Automated Essay Evaluation: Current Applications and New Directions; Shermis, M.D., Burstein, J., Eds.; Routledge: London, UK, 2013; pp. 16–35. [Google Scholar]
  62. Aryadoust, V.; Mehran, P.; Alizadeh, P. Validating a computer-assisted language learning attitude instrument used in Iranian EFL context: An evidence-based approach. Comput. Assist. Lang. Learn. J. 2016, 29, 561–595. [Google Scholar] [CrossRef]
  63. Kirkpatrick, J.; Kayser-Kirkpatrick, W. The Kirkpatrick Four Levels: A Fresh Look after 55 Years; Kirkpatrick Partners: Ocean City, NJ, USA, 2014. [Google Scholar]
  64. Singh, N.; Ahuja, N.J.; Kumar, A. A Novel Architecture for Learner-Centric Curriculum Sequencing in Adaptive Intelligent Tutoring System. J. Cases Inf. Technol. (JCIT) 2018, 20, 1–20. [Google Scholar] [CrossRef] [Green Version]
  65. Singh, N.; Ahuja, N.J. Bug Model Based Intelligent Recommender System with Exclusive Curriculum Sequencing for Learner-Centric Tutoring. Int. J. Web-Based Learn. Teach. Technol. (IJWLTT) 2019, 14, 1–25. [Google Scholar] [CrossRef] [Green Version]
  66. Singh, N.; Ahuja, N.J. Implementation and Evaluation of Intelligence Incorporated Tutoring System. Int. J. Innov. Technol. Explor. Eng. (IJITEE) 2019, 8, 4548–4558. [Google Scholar]
  67. Kumar, A.; Ahuja, N.J. Assessment of Learning Style of Learner using I 2A 2 Learning Style Model. Int. J. Innov. Technol. Explor. Eng. (IJITEE) 2019, 8, 154–159. [Google Scholar]
  68. Singh, N.; Ahuja, N.J. Empirical Analysis of Explicating the Tacit Knowledge Background, Challenges and Experimental findings. Int. J. Innov. Technol. Explor. Eng. (IJITEE) 2019, 8, 4559–4568. [Google Scholar]
  69. Singh, N.; Kumar, A.; Ahuja, N.J. Implementation and Evaluation of Personalized Intelligent Tutoring System. Int. J. Innov. Technol. Explor. Eng. (IJITEE) 2019, 8, 46–55. [Google Scholar]
Figure 1. Kirkpatrick’s four stages of evaluation.
Figure 1. Kirkpatrick’s four stages of evaluation.
Sustainability 14 04167 g001
Figure 2. Brief illustration of Kirkpatrick’s four phases or stages of assessment. Evaluation of Reaction [43,44,45], Evaluation of Learning [42,46,47], Evaluation of Results [45].
Figure 2. Brief illustration of Kirkpatrick’s four phases or stages of assessment. Evaluation of Reaction [43,44,45], Evaluation of Learning [42,46,47], Evaluation of Results [45].
Sustainability 14 04167 g002
Figure 3. Flow Diagram of CNN-based Emotion Recognition module.
Figure 3. Flow Diagram of CNN-based Emotion Recognition module.
Sustainability 14 04167 g003
Figure 4. Quiz and Understanding Test Representation.
Figure 4. Quiz and Understanding Test Representation.
Sustainability 14 04167 g004
Figure 5. Flow diagram of learner characteristics model.
Figure 5. Flow diagram of learner characteristics model.
Sustainability 14 04167 g005
Figure 6. Flow of accomplishing the evaluation process.
Figure 6. Flow of accomplishing the evaluation process.
Sustainability 14 04167 g006
Figure 7. Learner Demographic Characteristics.
Figure 7. Learner Demographic Characteristics.
Sustainability 14 04167 g007
Figure 8. Descriptive statistics of psychological parameter of learner for study 1 (learner study with customized learning).
Figure 8. Descriptive statistics of psychological parameter of learner for study 1 (learner study with customized learning).
Sustainability 14 04167 g008
Figure 9. Learning gain of study 1 (learner study with customized learning) and study 2 (learner study without customized learning).
Figure 9. Learning gain of study 1 (learner study with customized learning) and study 2 (learner study without customized learning).
Sustainability 14 04167 g009
Figure 10. Learner’s Learning Gain.
Figure 10. Learner’s Learning Gain.
Sustainability 14 04167 g010
Figure 11. Linear relationship with learning gain and degree of understanding.
Figure 11. Linear relationship with learning gain and degree of understanding.
Sustainability 14 04167 g011
Figure 12. Evaluation of learner behavior based on learner feedback.
Figure 12. Evaluation of learner behavior based on learner feedback.
Sustainability 14 04167 g012
Table 1. Feature of study 1 (learner study with customized learning) and study 2 (learner study without customized learning) evaluation groups.
Table 1. Feature of study 1 (learner study with customized learning) and study 2 (learner study without customized learning) evaluation groups.
Study 2 (Learner Study without Customized Learning)Study 1 (Learner Study with Customized Learning)
Personalized Tutoring contentsOffer learning content (similar curriculum) based on tutoring strategy.Offer personalized learning content (different curricula) based on the determined prior knowledge level and tutoring strategy [64,65,66].
Psychological State trackingPsychological state of the learner is not tracked during an ongoing learning session.Determine psychological state of the learner during an ongoing learning session.
Degree of Understanding computationLearner’s understanding of the concept is not adjudged.Quantify learner’s understanding of the concept.
Table 2. Average mean score of learning gain and degree of understanding.
Table 2. Average mean score of learning gain and degree of understanding.
Study 1 (Learner Study with Customized Learning)
Parameters
No. of
Participants
(n)
Learning Gain
Mean (/10)Standard DeviationMean%
Learning Gain282.21701.0279522%
Degree of Understanding282.54671.3120125%
Table 3. Correlation matrix between learning gain and degree of understanding.
Table 3. Correlation matrix between learning gain and degree of understanding.
Parameters Learning GainDegree of Understandings
Learning GainPearson_correlation10.484 **
Sig._(2-tailed) 0.009
N2828
Degree of
Understanding
Pearson_correlation0.484 **1
Sig._(2-tailed)0.009
N2828
** Correlation is significant at the 0.01 level (2-tailed).
Table 4. Learner feedback on effectiveness of SeisTutor.
Table 4. Learner feedback on effectiveness of SeisTutor.
QuestionsDegreeApplicant
Evaluation
Group
Strongly SatisfySatisfyNeutralDissatisfyStrongly Dissatisfy
System EffectivenessWhat is your overall level of satisfaction with SeisTutor?2723631Study 1
Study 2
Learning through this tutoring system (SeisTutor) was easy.2726511Study 1
Study 2
Did you feel that you were achieving learning outcomes?3021630Study 1
Study 2
I would recommend a course through SeisTutor with no instructor help2924340Study 1
Study 2
Would you recommend SeisTutor to individual who needs to take another course?2527530Study 1
Study 2
Did SeisTutor support you to make your study productive?2827302Study 1
Study 2
How well does this system deliver on your learning intentions?3121521Study 1
Study 2
Table 5. Learner Feedback on Adaptivity of SeisTutor.
Table 5. Learner Feedback on Adaptivity of SeisTutor.
QuestionsDegreeApplicant
Evaluation
Group
Strongly SatisfySatisfyNeutralDissatisfyStrongly Dissatisfy
Adaptivity/PersonalizationDid SeisTutor satisfy you with dynamic creation of your learning profile?2721714Study 1
Study 2
Were you convenient and satisfied with the tutoring strategy presented to you by SeisTutor?2419962Study 1
Study 2
The information provided by SeisTutor is at a level that you understand.29171202Study 1
Study 2
The tutoring session was at the right level of difficulty for me.2623920Study 1
Study 2
As a learner, did you feel that your learning style was appropriately judged? 2925321Study 1
Study 2
Once tutoring began, and you were tutored, were your learning preferences sufficiently satisfied?2424921Study 1
Study 2
Did the experience of learning by your own learning preference make you perform better?2421672Study 1
Study 2
Based on your prior subject knowledge, has SeisTutor accurately determined an exclusive curriculum for you?148231Study 1
How satisfied are you with the exclusively determined curriculum?137431Study 1
As a learner, did you feel learning material enabled you to improve your ability to formulate and analyze the problem?1014130Study 1
Are you satisfied with the sequencing of learning content?1409320Study 1
Does sequencing of learning material relate to your previous knowledge? (Give Rating)1211230Study 1
Does the learning content formulated under various learning levels and styles satisfactorily justify itself? (Give Rating)177202Study 1
Has this learning session been successful in improving your knowledge in the subject domain? (Give Rating)1211230Study 1
Did this learning material fulfill your expectations?1113211Study 1
Does the understanding test at the end of each week correspond to the lessons taught?1113202Study 1
SeisTutor compels and supports me to complete the quizzes, understanding tests, and lessons. 1312201Study 1
The post-tutoring evaluation system (weekly understanding) as it exists is:1410121Study 1
How do you rate the sequence of the lessons in the course?188002Study 1
Has SeisTutor accurately determined your psychological (emotional) state during the tutoring sessions? (Give Rating)117550Study 1
Do you feel recognition of emotion during ongoing tutoring is indicative of empathy for the system?1312201Study 1
Is the course content relevant and well organized?1410121Study 1
Table 6. Learner feedback on SeisTutor ongoing learning support.
Table 6. Learner feedback on SeisTutor ongoing learning support.
QuestionsDegreeApplicant
Evaluation Group
Strongly SatisfySatisfyNeutralDissatisfyStrongly Dissatisfy
How are you satisfied with the system support?24171162Study 1
Study 2
The system navigation support enabled finding the needed information easily.21179112Study 1
Study 2
Was the pre-learning procedure available in SeisTutor helpful to you?2517693Study 1
Study 2
Were you able to understand the language used to explain the lessons in SeisTutor?3321600Study 1
Study 2
The tutoring was flexible to meet my learning requirements.3021720Study 1
Study 2
Table 7. Learner Feedbacks on learning material, quizzes and overall SeisTutor support.
Table 7. Learner Feedbacks on learning material, quizzes and overall SeisTutor support.
QuestionsStrongly SatisfySatisfyNeutralDissatisfyStrongly DissatisfyApplicant
Evaluation Group
SeisTutor explained the content correctly.2325381Study 1
Study 2
SeisTutor made the course as interesting as possible.3119910Study 1
Study 2
The tutoring resources were adequate.2119794Study 1
Study 2
The presentation of the course content stimulated my interest during the learning session.3224211Study 1
Study 2
The course content is relevant and well organized.2925321Study 1
Study 2
SeisTutor supported me in understanding the content, which I found confusing?2726610Study 1
Study 2
Did the quiz at the end of each week correspond to the lessons taught?2827311Study 1
Study 2
The question-wise hints were helpful.2726601Study 1
Study 2
Did the SeisTutor react decidedly to your necessities?26211111Study 1
Study 2
Was the learning provided sufficiently to take the quiz?3618420Study 1
Study 2
During ongoing tutoring, assessments are a fair test of my knowledge and learning preferences.3221520Study 1
Study 2
Table 8. Learner Feedback on learning material presentation and on overall SeisTutor’s look and feel.
Table 8. Learner Feedback on learning material presentation and on overall SeisTutor’s look and feel.
QuestionsStrongly SatisfySatisfyNeutralDissatisfyStrongly
Dissatisfy
Applicant Evaluation Group
How satisfied are you with the look and feel (user interface design) of this system?3219810Study 1
Study 2
How satisfied are you with the account setup experience of this system?3121530Study 1
Study 2
How pleasing is the color scheme used in this system?28211100Study 1
Study 2
How user-friendly is this system? Give a rating2721822Study 1
Study 2
SeisTutor compels and supports me to complete the quizzes and lessons.2527611Study 1
Study 2
How satisfied are you with the organization/customization of contents feature of the system?2926401Study 1
Study 2
Table 9. Statistical results of paired sample T-test of study 1 (learner study with customized learning).
Table 9. Statistical results of paired sample T-test of study 1 (learner study with customized learning).
Comparison_ItemLearning-Mode
MeanNStd._DeviationStd._Error_Mean
Post-test_of_Study1_Applicants3.9375280.394550.07456
Pretest_of_Study1_Applicants1.7196280.997400.18849
Table 10. Paired sample T-test results of study 1 (learner study with customized learning).
Table 10. Paired sample T-test results of study 1 (learner study with customized learning).
MeanStd._DeviationStd._Error_Mean95% Confidence Interval_of_the DifferenceTdfSig (2 Tailed)
LowerUpper
Pair_1: Post-test_of_Study_1–Pretest_of_Study 1 (learner study with customized learning)2.217861.028560.194381.819022.6166911.410270.000
Table 11. Statistical results of paired sample T-test of study 2 (learner study without customized learning).
Table 11. Statistical results of paired sample T-test of study 2 (learner study without customized learning).
Comparison_ItemLearning-Mode
MeanNStd._DeviationStd._Error_Mean
Post-test_of_Study2_Applicants3.6525320.589150.10415
Pretest_of_Study2_Applicants2.4053321.395650.24672
Table 12. Paired sample T-test results of study 2 (learner study without customized learning).
Table 12. Paired sample T-test results of study 2 (learner study without customized learning).
MeanStd._DeviationStd._Error_Mean95% Confidence Interval_of_the DifferenceTdfSig (2 Tailed)
LowerUpper
Pair_1: Post-test_of_Study_2–Pretest_of_Study_21.247191.328040.234770.768381.726005.312310.041
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Singh, N.; Gunjan, V.K.; Mishra, A.K.; Mishra, R.K.; Nawaz, N. SeisTutor: A Custom-Tailored Intelligent Tutoring System and Sustainable Education. Sustainability 2022, 14, 4167. https://0-doi-org.brum.beds.ac.uk/10.3390/su14074167

AMA Style

Singh N, Gunjan VK, Mishra AK, Mishra RK, Nawaz N. SeisTutor: A Custom-Tailored Intelligent Tutoring System and Sustainable Education. Sustainability. 2022; 14(7):4167. https://0-doi-org.brum.beds.ac.uk/10.3390/su14074167

Chicago/Turabian Style

Singh, Ninni, Vinit Kumar Gunjan, Amit Kumar Mishra, Ram Krishn Mishra, and Nishad Nawaz. 2022. "SeisTutor: A Custom-Tailored Intelligent Tutoring System and Sustainable Education" Sustainability 14, no. 7: 4167. https://0-doi-org.brum.beds.ac.uk/10.3390/su14074167

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop