Next Article in Journal
Outdoor Adventure Education: Trends and New Directions—Introduction to a Special Collection of Research
Next Article in Special Issue
Strategies for Flipped Learning in the Health Professions Education in South Korea and Their Effects: A Systematic Review
Previous Article in Journal
Cloud-Based ICME Software Training
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

The Continuous Intention to Use E-Learning, from Two Different Perspectives

by
Rana Saeed Al-Maroof
1,
Khadija Alhumaid
2 and
Said Salloum
3,*
1
English Language & Linguistics, Al-Buraimi University College, Al-Buraimi 512, Oman
2
College of Education, Zayed University, Abu Dhabi, UAE
3
Research Institute of Sciences and Engineering (RISE), University of Sharjah, Sharjah, UAE
*
Author to whom correspondence should be addressed.
Submission received: 29 November 2020 / Revised: 16 December 2020 / Accepted: 21 December 2020 / Published: 25 December 2020
(This article belongs to the Special Issue Assessment and Evaluation in Higher Education)

Abstract

:
During the recent vast growth of digitalization, e-learning methods have become the most influential phenomenon at higher educational institutions. E-learning adoption has proved able to shift educational circumstances from the traditional face-to-face teaching environment to a flexible and sharable type of education. An online survey was conducted, consisting of 30 teachers and 342 students in one of the universities in the United Arab Emirates. The results show that teachers’ and students’ perceived technology self-efficacy (TSE), ease of use (PEOU), and usefulness (PU) are the main factors directly affecting the continuous intention to use technology. Instructors’ technological pedagogical content knowledge (TPACK) and perceived organizational support (POS) positively affect the intention to use the technology, whereas students’ controlled motivation (CTRLM) has a greater influence on their intention to use the technology, due to the type of intrinsic and extrinsic motivation that they have and which they can develop throughout the process of learning. The findings support the given hypotheses. In addition, they provide empirical evidence of a relationship between perceived organizational support and perceived pedagogical content knowledge. In fact, they are considered the key factors that support the use of technology continuously.

1. Introduction

Teachers and students may perceive the importance of e-learning differently. Teachers usually focus on the importance of training and support that may enhance the effective use of e-learning platforms, whereas the perceived usefulness and ease of use are influential factors from the students’ perspective. The differences in their perspectives stem from the fact that their roles are different. Students usually receive the product through the e-learning system and can get all the different advantages that the system may offer; thus, they act as the consumers of the product [1]. On the other hand, teachers are the providers of the educational product, as they provide learners with the content and synthesize the given information in simple and concise language [1,2].
Past studies have focused on the importance of e-learning and its implementation all over the world; some have focused on the continuous use of e-learning [3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19] and some on the effect of either teachers’ or students’ attitudes towards e-learning [20,21,22,23,24,25,26]. In other words, no studies have put forward the implementation of two models that focus on how the perceived interactivity of education technology influences teachers’ and students’ perceptions and urges them to continue using the technology. This study assumes that some factors affecting teachers’ intentions to use e-learning platforms continually are different from those affecting students’ intentions to use the same platforms continuously. Therefore, this study proposes two different models that tackle both teachers’ and students’ continuous intentions to use technology. The two models will focus on a certain predictive power that has a more direct relationship with the teachers’ and students’ perceptions regarding the continued use of the technology. For instance, one of the factors that contributes to the teachers’ continuous use of the e-learning platform is the support they get from the university to enhance the use of the e-learning environment. On the other hand, one of the crucial factors for the students’ perception is the controlled motivation that embraces certain intrinsic and extrinsic factors. It is worth mentioning that all previous studies have focused on students’ perspectives on any technology-based technique. The acceptance of technology from a student’s perspective has been dealt with in many papers, such as [27,28,29]. The fact that this study focuses on the effect of the same e-learning tool from a teacher’s perspective separately is what sets this paper apart from other previous studies.
The objective of this study was to propose a theoretical framework that could be validated later through a proposed model that predicates the continuous intention to use e-learning among students at public universities in Dubai. There are numerous examples of literature related to technology acceptance [25,30,31,32,33,34] and continuous intention [35,36,37,38] that have been reviewed to identify the most common factors affecting the continuous intention to use the e-learning platform. The main concentration has been on theories that have been proven to have great predictive power in understanding users’ perceptions and on theories that help to explain the importance of continuous use from two different perspectives. Hence, the Technological Pedagogical Content Knowledge (TPACK) was initiated by [39], Technology Acceptance Model (TAM) acceptance theory by [40], Social Cognitive Theory by [41], Perceived Organizational Support (POS) by [42], and Motivational Theory (MT) by [43,44]. The main factors that have been derived from these theories are perceived use and perceived usefulness, computer self-efficacy, controlled motivation, and so forth. The table below (Table 1) summarizes the studies that have tackled the continuous use of e-learning platforms.
As seen in the previous table showing the studies in the existing literature, much research has been conducted focusing on students and/or teachers within one proposed model. Nevertheless, searching for the predictive power behind both teachers’ and students’ intentions by proposing different variables is still neglected. To our knowledge, no research has examined the continued-use intention (CU) of teachers and instructors using e-learning platforms in higher education. Without knowledge of teachers’ and students’ CU, it is impossible to enhance e-learning in the Gulf area or to support its programs, systems, or administrative policies in terms of helping to sustain the e-learning platform.

2. Theoretical Framework and Hypotheses

The proposed framework has certain factors that can make the intention to use e-learning more measurable from two different perspectives. TPACK and POS are crucial elements that usually guide the teaching and learning environments from the teachers’ perspectives. On the other hand, controlled motivation (CTRLM) is a factor that combines students’ intrinsic and extrinsic motivations. Nevertheless, certain factors are equally important to both teachers and students, such as technology self-efficacy (TSE), perceived usefulness (PU), and perceived ease of use (PEOU).

2.1. Technological Pedagogical Content Knowledge (TPACK)

Teachers’ knowledge cannot be tackled easily, as it is a complex concept that has many embedded elements [39]. The most important element is pedagogical content knowledge (PCK), which has been the domain of study for many researchers and practitioners. Its importance stems from the fact that it comprises both the content and the pedagogy that can explain how a particular topic is organized and how it is represented to the learners [53].
Since its emergence, TPACK has become a must since all teachers want to have a full understanding of the relationship between pedagogy and technology. TPACK refers to the type of technological pedagogical knowledge that teachers need to organize and present the intended teaching material effectively [39]. TPACK is one of the influential factors affecting teachers’ perspectives. It refers to technological pedagogical content knowledge which includes: TCK (technological content knowledge), TPK (technological pedagogical knowledge), and PCK (pedagogical content knowledge) [54]. Self-assessment surveys and performance-based assessments are the basic instruments for evaluating TPACK [55,56].
The framework that comprises TPACK can be explained as combing different elements, such as content knowledge (CK), that highlight the teacher’s knowledge of the subject matter. It includes knowledge of different types, such as knowledge of the theory, discipline, psychological aspects, historical aspects, and so forth [57]. The other two closely related elements are the pedagogy knowledge (PK), which is closely related to the teacher’s knowledge concerning methodology, process, and practice, and the pedagogical content knowledge, which is concerned with how teachers interpret and tailor the teaching material to suit certain pedagogical aims and purposes. This implies that the difference between the former element and the latter is the fact that the latter is related to methodology, assessment, and teaching style knowledge which can be used differently based on students’ prior knowledge. Other elements are related to technology, as it comprises technology knowledge, technological content knowledge, and technological pedagogical knowledge. They are related to the ability of users to use technology to accomplish different tasks. The technological content knowledge is related to how technology can affect teaching material and vice versa. The final element is technological pedagogical knowledge, which has to do with the constraints that technology may impose on teaching material. This stems from the fact that certain technology is not developed for educational purposes and should be accustomed to suit educational purposes [57,58]. Accordingly, the following hypothesis can be formed:
Hypothesis 1 (H1).
TPACK will positively affect teachers’ CU in the e-learning environment.

2.2. Technology Self-Efficacy (TSE)

Self-efficacy is an effective factor that can reflect how students’ own belief in their abilities to use technology affects their acceptance of the learning environment. Therefore, self-efficacy and learning are two factors that can, interactively and dynamically, affect each other [59]. Self-efficacy in the e-learning environment is considered an intrinsic motivator as far as continuous intention is concerned. It usually refers to the degree of confidence that users have in making use of technology [60]. Technology self-efficacy is usually identified as the ability to use technology without facing any crucial problems. It embraces two subdivisions: the estimation of result (users’ estimations about their own input) and estimation efficacy (users’ estimations in achieving the final result) [61,62,63]. Within the environment of e-learning, self-efficacy is highly connected to users’ own beliefs regarding technology. Some believe that using technology is tremendously easy and achievable, while others may share a contradictory belief, as they may face problems in their ability to learn the appropriate way of using technology [64]. This simply implies that whenever users have a high level of technological self-belief, they may perceive the whole system properly; hence, they will be able to continue using the technology in a positive way. Accordingly, the following hypothesis may be formulated:
Hypothesis 2 (H2).
Technology Self-efficacy will positively affect teachers’ and students’ CU in the e-learning environment.

2.3. Technology Acceptance Model (TAM)

A review of recent studies has shown that certain variables are crucial to understanding the reasons behind the continuous intention to use e-learning. Regarding Davis’s TAM [65], it has been proven that PU and PEOU are the most influential factors in users’ continuous-use intentions. Interestingly, PU is more effective than PEOU when one wants to deal with the use of technology [60]. This study focuses only on two constructs within the TAM theory, which have proven effective in investigating the continuous use of technology; these are perceived usefulness and ease of use. [65] adopted the view that the perceived usefulness (PU) and the perceived ease of use (PEOU) of technology form the baseline for examining individuals’ usage intentions. PU is defined as the degree to which a person believes that using a technological system supports user performance, whereas PEOU tends to refer to the degree to which a person believes that use will be free of effort. Due to the fact that PEOU has proved to be of great significance only during the early-acceptance stage of technology use [54,65], PEOU may not directly affect teachers’ and students’ CU in the e-learning environment. Hence, we hypothesize the following:
Hypothesis 3 (H3).
The level of PU will positively affect teachers’ and students’ CU.
Hypothesis 4 (H4).
The level of PEOU will not affect teachers’ and students’ CU.

2.4. Perceived Organizational Support (POS)

Organizational support theory has a close relation with how users perceive organizational feedback regarding the use of technology. Users’ perceptions may vary in accordance with an organization’s rewards, fairness, and supervisor support [42]. To put it differently, when users had a positive attitude towards their organization, they were more willing to pursue their intention to use its technology and vice versa. The organization has a crucial role in enhancing the use of technology by motivating users internally at the organizational level [44,66]. Related literature is guided by the fact that organizational support for use of technology has a high impact on teachers’ and students’ CU to adhere to computer technology, particularly in technical support [67,68,69]. In this respect, teachers and students may have different subjective perspectives of the role of educational institutions (colleges and universities) in creating a motivational atmosphere regarding the continuous use of technology. Hence, it is hypothesized that if teachers and students had a positive perception of organizational support (POS), they would support the continuous usage of the technology. Thus, the following hypothesis may be formed:
Hypothesis 5 (H5).
The level of POS support will positively affect teachers’ CU.

2.5. Controlled Motivation (CTRLM)

According to [43], students’ intrinsic and extrinsic motivations can be dealt with in terms of a hieratical model in which three factors play an influential role: contextual, situational, and global factors. One of the motivations that have a massive effect on user perception is called Controlled Motivation (CTRLM), which refers to a source of negative perception that is illustrated by the pressure that students may be under, both internally and externally. This type of pressure may lead to maladaptive outcomes, which are, in turn, illustrated by a combination of negative effects, perceived incompetence, and dissatisfaction [44]. A more updated view regarding controlled motivation is given by [44], who proposed two different types within CTRLM: introjected and external regulation. Introjected regulation has a close relationship with an individual’s behavioral engagement, such as obligation, avoidance of guilt, ego-enhancement, and internal rewards. On the other hand, external regulation affects behavior engagement, including compliance, external rewards, and avoidance of punishments.
Hypothesis 6 (H6).
Controlled Motivation negatively affects students’ continuous intention to use e-learning platforms.

2.6. The Proposed Research Models

Based on the review of previous studies, it has been noticed that most of the studies on continuous intention have focused on one model that has factors that may be crucial to teachers but not students [60,70]. Therefore, this study attempted to build two models that can meet both teachers’ and students’ continuous intentions to meet e-learning demands. The proposed research models rely on these hypotheses, as illustrated in Figure 1 and Figure 2.

3. Methodology

3.1. Participants

The participants (n = 372) were classified into two categories—teachers and students—in accordance with the two proposed models. This two-level selection was based on certain influential factors. First, participants were identified as having sufficient experience in using e-learning platforms. Second, two different surveys were prepared and distributed to the two targeted groups of participants. The two surveys may have had similar and shared questions, but the teachers’ survey may have had additional questions that the students’ survey may have lacked. Responses were received from 372 participants. Therefore, the total number of teacher participants in this study was 30. These were college instructors with a nearly equal female-to-male gender ratio. On the other hand, the total number of student participants was 342. These were students at The British University in Dubai (BUiD).

3.2. Data Collection

During the winter semester of 2019/2020, the data was collected through online surveys from individuals studying at the British University in Dubai (BUiD) from 15 January to 20 February 2020. The aggregated response rate was 93%; 400 questionnaires were circulated, out of which 372 were answered by respondents. This means that 372 questionnaires were filled out correctly and found to be useful, while 28 were rejected because of missing values. The prospective sample size was 306 respondents with respect to a population of 1500. Thus, the sample size of 372 correct responses was suitable, according to [71], because—bearing in mind the required sample size—the sample size of 372 is a higher figure. Thus, this sample size could be reviewed using structural equation modeling [72] to verify the hypotheses. It must be noted that hypotheses were based on the current theories and were adjusted to the e-learning context. In order to assess the measurement model, the researchers used structural equation modeling (SEM) [72]. Further treatment was performed using a final path model.

3.3. Students’ Personal Information/Demographic Data

The assessment of personal/demographic data is covered in Table 2. The percentage of males was 53%, while for females it was 47%. A total of 33% of students had ages ranging from 18 to 29 years, while 67% of respondents were aged over 29. In terms of academic background, 39% were students from the Faculty of Engineering and IT, 35% were from the Faculty of Education and 26% belonged to the Faculty of Business and Law. The majority of respondents came from sophisticated families and held university degrees; 49% of participants had bachelor’s degrees, 42% had master’s degrees, and 9% had a doctoral degree. When the respondents were ready to volunteer and were easily approachable, the purposive sampling approach was used as per [3]. This sample was created by students coming from different faculties, with different ages, enrolling in diverse programs at different levels. Moreover, with the aid of IBM SPSS Statistics ver. 23, the demographic data was evaluated. Table 2 depicts the complete demographic data of the respondents.

3.4. Study Instrument

The survey instrument used to validate the hypothesis was determined in this research. The survey, consisting of 30 items, was used for the measurement of seven constructs in the questionnaire. Table 3 depicts the sources of the constructs. The questions from prior studies were modified in order to enhance the appropriateness of the research.

3.5. Pilot Study for the Questionnaire

A pilot study was conducted to check the reliability of the questionnaire items. Approximately 40 students and teachers were chosen on a random basis from the given population to establish the pilot study. The sample size was set based on 10% of the aggregated sample size of this study (400 students and teachers) and thus adhered strictly to the research criteria. Cronbach’s alpha test was utilized for the computation of internal reliability through IBM SPSS Statistics ver. 23, in order to judge the outcomes of the pilot study. Thus, the appropriate findings were shown for the measurement items. A value of 0.7 was taken to be an acceptable value for the reliability coefficient, considering the model for social science research [14]. Table 4 and Table 5 show the Cronbach’s alpha values for the seven measurement scales for teachers and students.

3.6. Survey Structure

The questionnaire survey given to students and teachers had two sections. Within the first part, personal data was given to gather information about students and teachers. The second section had a group of questions related to the main factors of the proposed models. The teachers’ questionnaire had six sub-sections coinciding with the six factors proposed in the model. Similarly, the students’ questionnaire had five sub-sections related to the five factors proposed in the model. With the help of the five-point Likert Scale, the 42 items were evaluated. The scales included the following: (1) strongly disagree, (2) disagree, (3) neutral, (4) agree, and (5) strongly agree.

4. Findings and Discussion

4.1. Data Analysis

Along with the help of SmartPLS V.3.2.7 software, the partial least squares-structural equation modeling (PLS-SEM) was utilized to conduct the data analysis in this research [15]. The assessment approach had two steps of a structural model and a measurement model allowed to study the collected data [16]. There were various reasons for choosing PLS-SEM in the study. First, as the research is an extension of a current theory, PLS-SEM was considered the best option [17]. Second, the complex models within exploratory research can be effectively tackled with the help of PLS-SEM [18]. Third, PLS-SEM analyzes a complete model as a single unit, so there is no need to divide it [19]. Lastly, PLS-SEM provides concurrent analysis for measurement, as well as a structural model, leading to more accurate calculations [20].

4.2. Convergent Validity

In order to review the measurement model, it was suggested by [16] that construct reliability—including composite reliability (CR), Dijkstra–Henseler’s rho (pA), and Cronbach’s alpha (CA) and validity (including convergent and discriminant validity)—must be considered. Cronbach’s alpha (CA) has values between 0.782 and 0.895, as Table 6 and Table 7 show, in order to determine construct reliability. These statistics are higher than the threshold value of 0.7 [78]. According to Table 6 and Table 7, the outcomes also show that the composite reliability (CR) has values from 0.796 to 0.882; these values are evidently bigger than the recommended value of 0.7 [79]. As an alternative, the construct reliability must be appraised by researchers by means of the Dijkstra–Henseler’s rho (pA) reliability coefficient [80]. Like CA and CR, the reliability coefficient ρA must show 0.70 or higher in exploratory studies and values of more than 0.80 or 0.90 for further stages of research [78,81,82]. The reliability coefficient ρA of each measurement construct is above 0.70 according to Table 6 and Table 7. According to these outcomes, the construct reliability is verified and all the constructs were considered to be accurate.
Convergent validity can be measured by testing the average variance extracted (AVE) as well as the factor loading [16]. Table 6 and Table 7 suggest that all values of factor loadings exceeded the threshold value of 0.7. Moreover, Table 6 and Table 7 show that the values obtained for the AVE were higher than the threshold value of 0.5, ranging from 0.509 to 0.718. Depending on the expected results, the convergent reliability can be obtained for all the constructs.

4.3. Discriminant Validity

The two criteria that were suggested should be measured to obtain the measurement of discriminant validity were the Fornell–Larcker measure and the Heterotrait–Monotrait ratio (HTMT) [16]. As per the findings of Table 8 and Table 9, these needs have been verified by the Fornell–Larker criterion as each AVE value, together with its square root, exceeds the value of the correlation of AVE with other constructs [83].
Table 10 and Table 11 show the outcomes of the HTMT ratio, indicating that the threshold value of 0.85 is bigger than the values of other constructs [27], and hence confirming the HTMT ratio. These outcomes play a role in the evaluation of the discriminant validity. The outcomes of the analysis indicated a smooth and simple assessment of the measurement model in terms of the model’s validity and reliability. To conclude, it can be said that the collected data was appropriate for additionally evaluating the structural model.

4.4. Model Fit

The subsequently mentioned fit measures are ensured by SmartPLS: the standard root mean square residual (SRMR), exact fit criteria, Euclidean distance (d_ULS), geodesic distance (d_G), Chi-square, Normed Fit Index (NFI), and RMS Theta show the model fit in PLS-SEM [84]. The difference between experimental correlations and the correlation matrix inferred from model [85] are indicated by SRMR, and values smaller than 0.08 are assumed to serve as good model-fit measures [86]. The NFI values that are higher than 0.90 point out a good model fit [87]. The NFI is a ratio of the Chi-square value of the proposed model to the null model (also known as the benchmark model) [88]. The NFI increases with larger parameters and therefore, the NPI is not suggested as a model-fit pointer [85]. The discrepancy between the empirical covariance matrix and the covariance matrix, inferred from the composite factor model, is indicated by the metrics of squared Euclidean distance (d_ULS) and the geodesic distance (d_G) [80,85]. The RMS Theta helps in the measurement of the degree of outer model residuals correlation and is appropriate for reflective models only [88]. The nearer the RMS Theta value is to zero, the more superior the PLS-SEM model, and their values of less than 0.12, are assumed to be a good fit, with anything other than this suggesting an absence of fit [89]. The saturated model evaluates the correlation between all constructs, as recommended by [29], while the approximate model takes all the effects and model structure into consideration. The RMS Theta value was 0.073 in Model A and 0.073 in Model B, as given in Table 12 and Table 13, which gives an idea that the specific goodness-of-fit for the PLS-SEM model was big enough to prove global PLS model validity.

4.5. Hypotheses Testing Using PLS-SEM

The interdependence between different theoretical constructs related to the structural model was studied by using a combination of the structural equation model with maximum-likelihood estimation and SmartPLS [38,39]. This indicates the procedure of analysis of the proposed hypothesis. About 83% and 71% variance were found within the continuous intention to use the e-learning platform, as shown in Table 14 and Table 15, which indicates a high predictive power of Models A and B [37]. For all the proposed hypotheses, outcomes of the PLS-SEM technique provided the beta (β) values, t-values, and p-values, which have been stated in Table 16 and Table 17. It is evident that each and every hypothesis is supported by all the researchers. The empirical data backs hypotheses H1, H2, H3, H4, H5, and H6 on the basis of the analyzed data. The standardized path coefficients and path significances are demonstrated in Figure 3 and Figure 4.
In Model A, technological pedagogical content knowledge (TPACK), technology self-efficacy (TSE), perceived ease of use (PEOU), perceived usefulness (PU), and perceived organizational support (POS) have significant effects on continuous intention to use the e-learning platform (CU) ((β = 0.336, p < 0.001), (β = 0.426, p < 0.05), β = 0.589, p < 0.05), (β = 0.625, p < 0.05) and (β = 0.553, p < 0.001), respectively); hence, H1, H2, H3, H4, and H5 are supported.
In Model B, technology self-efficacy (TSE), perceived ease of use (PEOU), perceived usefulness (PU), and controlled motivation (CTRLM) have significant effects on continuous intention to use the e-learning platform (CU) ((β = 0.290, p < 0.001), (β = 0.357, p < 0.05), β = 0.465, p < 0.05) and (β = 0.243, p < 0.05), respectively); hence, H2, H3, H4, and H6 are supported.

5. Discussion and Conclusions

This study proposed two unique CU models that took into consideration factors that affect both instructors’ and students’ attitudes. The two models can be theoretically extended to enhance other technology-supported educational environments and instructional processes. The first research model of instructors’ CU was proposed, taking into consideration certain social cognitive theory along with personal, behavioral, and environmental elements that are closely related to instructors’ CU. In general, the results of SEM analysis has supported all the proposed hypothesis. From a practical perspective, this study has proven that POS is the most influential factor that affects instructors’ CU of e-learning platforms.
Ref. [90,91,92] seem to agree with the current conclusion in stating that POS could motivate its staff members, leading to an upgrade of the organization. However, a study by [93] placed an emphasis on staff members’ personalities and readiness to change. This implies that the lack of organizational support may have negative consequences. One of the studies by [60] has proposed that when instructors feel that there is no adequate organizational support, they are less likely to continuously use the technology, especially in an educational atmosphere where instructors are supposed to implement various in-class pedagogical changes to facilitate a better learning environment for the students.
In fact, POS is not the only factor that affects instructors’ CU, but rather instructors’ TPACK is another key factor that affects CU. Most of the previous research has proven that the organizational support may affect users’ motivation to use the technology. A study by [94] put emphasis on the effect of PACK in facilitating the e-learning process by both teachers and adult learners. This seems to be in line with the results obtained from this study where TPACK affected, to great extent, the teachers’ performance. It is assumed that whenever teachers’ content, technological, and pedagogical knowledge is high, it implies that his or her ability to change the teaching material to suit the newly used technology will be more practical and effective. Obviously, teachers are more motivated when there is reliable technical support and IT staff that can facilitate the process of establishing new computer-supported knowledge [66,68].
Similarly, CTRLM has an effect on students’ CU. CTRLM, along with technology self-efficacy (TSE), are the key factors that have a great impact on students’ CU. The study has proven that CTRLM is connected deeply with the willingness to use the e-learning teaching platform. The higher the motivation is, the more effective results are obtained. Previous studies have tackled the effect of CTRLM on students’ performance and have proven that there are both pedagogical and non-pedagogical elements that affect motivation [95,96,97]. These studies have indicated that technology development has placed a positive effect on motivation, as it urges students to get engaged in the new learning platform. This makes motivation very high and involvement in classes even higher, increasing the students’ willingness to learn, and thus, using the technology continuously [98,99].

5.1. Practical Implications

According to the study outcomes, salient factors were noticed in determining users’ acceptance of the e-learning system and technological pedagogical content knowledge, technology self-efficacy, perceived ease of use, perceived usefulness, perceived organizational support, and controlled motivation were significant. Hence, upon implementing a new e-learning system, faculty members must be informed about the system’s features and its technical issues, as well as its usefulness, so that they feel self-assured and can gain insight into the system. To increase faculty use of the e-learning system, faculty members are of the view that universities should deliver workshops, extensive training, and awareness programs on the system’s features, benefits, and usage [100]. In addition, a national survey was conducted, in which 57% of faculty members said that they could become more productive [100] if the use of e-learning technology was communicated thoroughly to them in their courses. Faculty members in this study also reported that, if they were aware of the positive impact of such technology on student learning, they would be inspired to learn and use the e-learning system. Hence, extended online help and periodic training programs for the e-learning system should be offered by the universities to ensure increased use of the system and to increase the faculty’s self-efficacy. As a result, this overall phenomenon would help faculty members to obtain practical exposure, acquire better skills and become more proficient in using the e-learning system; as a result of this, their use of the system would be enhanced. In accordance with our findings, there was a weak influence on faculty attitudes toward the e-learning system because of facilitating conditions. Consequently, to ensure the smooth running of the e-learning system, attention must be given by universities to ensuring technical support and reliable network access. Moreover, online and face-to-face support and guidance should be provided by the universities for faculty members to ensure that members have positive attitudes toward the e-learning system and, consequently, that they become capable of extended use of the system [65,101].

5.2. Limitations and Further Research

This study has the following limitations. Only one university in the United Arab Emirates was considered for studying the impact of factors on the adoption of e-learning systems; this is the key limitation. The applicability and pertinence of this study would have been enhanced if more institutes and universities in the United Arab Emirates had been taken into consideration. With further analysis and insight into the e-learning system, the researchers could have better understood the factors that influence such a system. In addition, the participation of a limited number of students (372) is another limitation. According to [45], a survey questionnaire method was used for data collection. For appropriate and improved outcomes, an enhanced instrument is being sought and more institutes will be taken into account from other regions, such as the Arab Gulf, including Kuwait, Saudi Arabia, and Bahrain. Furthermore, invitations to join the study will be sent to more students, and researchers will conduct focus groups and interviews for suitable results. Moreover, we are seeking to implement the e-learning system in specific Arab universities that have contributed to the research.

Author Contributions

Conceptualization, R.S.A.-M., K.A. and S.S.; methodology, R.S.A.-M., K.A. and S.S.; software, S.S.; validation, S.S.; formal analysis, R.S.A.-M., K.A. and S.S.; investigation, R.S.A and K.A.; resources, K.A.; data curation, S.S.; writing—original draft preparation, R.S.A.-M.; writing—review and editing, K.A. and S.S.; visualization, R.S.A.-M., K.A. and S.S.; supervision, R.S.A.-M., K.A. and S.S.; project administration, R.S.A.-M., K.A. and S.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wagner, N.; Hassanein, K.; Head, M. Who is responsible for e-learning success in higher education? A stakeholders’ analysis. J. Educ. Technol. Soc. 2008, 11, 26–36. [Google Scholar]
  2. Kituyi, G.; Tusubira, I. A Framework for the Integration of E-learning in Higher Education Institutions in Developing Countries. Int. J. Educ. Dev. Using ICT 2013, 9, 19–36. [Google Scholar]
  3. Luo, N.; Zhang, M.; Qi, D. Effects of Different Interactions on Students’ Sense of Community in E-learning Environment. Comput. Educ. 2017, 115, 153–160. [Google Scholar] [CrossRef]
  4. Hong, J.-C.; Tai, K.-H.; Hwang, M.-Y.; Kuo, Y.-C.; Chen, J.-S. Internet Cognitive Failure Relevant to Users’ Satisfaction with Content and Interface Design to Reflect Continuance Intention to Use a Government E-learning System. Comput. Human Behav. 2017, 66, 353–362. [Google Scholar] [CrossRef]
  5. Alghizzawi, M.; Habes, M.; Salloum, S.A.; Ghani, M.A.; Mhamdi, C.; Shaalan, K. The Effect of Social Media Usage on Students’E-learning Acceptance in Higher Education: A case study from the United Arab Emirates. Int. J. Inf. Technol. Lang. Stud. 2019, 3, 13–26. [Google Scholar]
  6. Alshurideh, M.; al Kurdi, B.; Salloum, S.A. Examining the Main Mobile Learning System Drivers’ Effects: A Mix Empirical Examination of Both the Expectation-Confirmation Model (ECM) and the Technology Acceptance Model (TAM). In Proceedings of the 5th International Conference on Advanced Intelligent Systems and Informatics (AISI’19), Cairo, Egypt, 26–28 October 2019; Volume 1058. [Google Scholar]
  7. Salloum, S.A.; Al-Emran, M.; Habes, M.; Alghizzawi, M.; Ghani, M.A.; Shaalan, K. Understanding the Impact of Social Media Practices on E-Learning Systems Acceptance. In Proceedings of the 5th International Conference on Advanced Intelligent Systems and Informatics (AISI’19), Cairo, Egypt, 26–28 October 2019; Volume 1058. [Google Scholar]
  8. Alshurideh, M.; Salloum, S.A.; al Kurdi, B.; Monem, A.A.; Shaalan, K. Understanding the quality determinants that influence the intention to use the mobile learning platforms: A practical study. Int. J. Interact. Mob. Technol. 2019, 13, 157–183. [Google Scholar] [CrossRef]
  9. Al-Emran, M.; Arpaci, I.; Salloum, S.A. An empirical examination of continuous intention to use m-learning: An integrated model. Educ. Inf. Technol. 2020, 25, 2899–2918. [Google Scholar] [CrossRef]
  10. Salloum, S.A.; Shaalan, K. Investigating Students’ Acceptance of E-Learning System in Higher Educational Environments in the UAE: Applying the Extended Technology Acceptance Model (TAM); The British University in Dubai: Dubai, UAE, 2018. [Google Scholar]
  11. Salloum, S.A.; Alhamad, A.Q.M.; Al-Emran, M.; Monem, A.A.; Shaalan, K. Exploring Students’ Acceptance of E-Learning Through the Development of a Comprehensive Technology Acceptance Model. IEEE Access 2019, 7, 128445–128462. [Google Scholar] [CrossRef]
  12. Muqtadiroh, F.A.; Nisafani, A.S.; Saraswati, R.M.; Herdiyanti, A. Analysis of User Resistance Towards Adopting E-Learning. Procedia Comput. Sci. 2019, 161, 123–132. [Google Scholar] [CrossRef]
  13. Tawafak, R.M.; Romli, A.B.T.; Arshah, R.b.A.; Malik, S.I. Framework design of university communication model (UCOM) to enhance continuous intentions in teaching and e-learning process. Educ. Inf. Technol. 2019, 25, 817–843. [Google Scholar] [CrossRef]
  14. al Kurdi, B.; Alshurideh, M.; Salloum, S.A. Investigating a theoretical framework for e-learning technology acceptance. Int. J. Electr. Comput. Eng. 2020, 10, 6484–6496. [Google Scholar] [CrossRef]
  15. Al-Maroof, R.S.; Salloum, S.A. An Integrated Model of Continuous Intention to Use of Google Classroom. In Recent Advances in Intelligent Systems and Smart Applications; Springer: Cham, Switzerland, 2020; Volume 295. [Google Scholar]
  16. Al-Maroof, R.A.; Arpaci, I.; Al-Emran, M.; Salloum, S.A.; Shaalan, K. Examining the Acceptance of WhatsApp Stickers Through Machine Learning Algorithms. In Recent Advances in Intelligent Systems and Smart Applications; Springer: Cham, Switzerland, 2020; Volume 295. [Google Scholar]
  17. Al-Maroof, R.S.; Salloum, S.A.; AlHamadand, A.Q.M.; Shaalan, K. A Unified Model for the Use and Acceptance of Stickers in Social Media Messaging. In International Conference on Advanced Intelligent Systems and Informatics; Springer: Cham, Switzerland, 2019; pp. 370–381. [Google Scholar]
  18. Al Kurdi, B.; Alshurideh, M.; Salloum, S.A.; Obeidat, Z.M.; Al-dweeri, R.M. An Empirical Investigation into Examination of Factors Influencing University Students’ Behavior towards Elearning Acceptance Using SEM Approach. Int. J. Interact. Mob. Technol. 2020, 14, 19–41. [Google Scholar] [CrossRef] [Green Version]
  19. Al-Maroof, R.S.; Salloum, S.A.; Al Hamadand, A.Q.; Shaalan, K. Understanding an Extension Technology Acceptance Model of Google Translation: A Multi-Cultural Study in United Arab Emirates. Int. J. Interact. Mob. Technol. 2020, 14, 157–178. [Google Scholar] [CrossRef] [Green Version]
  20. Demetriadis, S.; Barbas, A.; Molohides, A.; Palaigeorgiou, G.; Psillos, D.; Vlahavas, I.; Tsoukalas, I.; Pombortsis, A. ‘Cultures in negotiation’: Teachers’ acceptance/resistance attitudes considering the infusion of technology into schools. Comput. Educ. 2003, 41, 19–37. [Google Scholar] [CrossRef] [Green Version]
  21. Buckenmeyer, J. Revisiting teacher adoption of technology: Research implications and recommendations for successful full technology integration. Coll. Teach. Methods Styles J. 2008, 4, 7–10. [Google Scholar] [CrossRef]
  22. Aldunate, R.; Nussbaum, M. Teacher adoption of technology. Comput. Human Behav. 2013, 29, 519–524. [Google Scholar] [CrossRef]
  23. Eyyam, R.; Yaratan, H.S. Impact of use of technology in mathematics lessons on student achievement and attitudes. Soc. Behav. Personal. Int. J. 2014, 42, 31S–42S. [Google Scholar] [CrossRef]
  24. Nelson, M.J.; Hawk, N.A. The impact of field experiences on prospective preservice teachers’ technology integration beliefs and intentions. Teach. Teach. Educ. 2020, 89, 103006. [Google Scholar] [CrossRef]
  25. Scherer, R.; Siddiq, F.; Tondeur, J. All the same or different? Revisiting measures of teachers’ technology acceptance. Comput. Educ. 2020, 143, 103656. [Google Scholar] [CrossRef]
  26. Raes, A.; Depaepe, F. A longitudinal study to understand students’ acceptance of technological reform. When experiences exceed expectations. Educ. Inf. Technol. 2020, 25, 533–552. [Google Scholar] [CrossRef]
  27. Salloum, S.A.; Al-Emran, M.; Shaalan, K.; Tarhini, A. Factors affecting the E-learning acceptance: A case study from UAE. Educ. Inf. Technol. 2019, 24, 509–530. [Google Scholar] [CrossRef]
  28. Cheng, Y. Antecedents and consequences of e-learning acceptance. Inf. Syst. J. 2011, 21, 269–299. [Google Scholar] [CrossRef]
  29. Al-Gahtani, S.S. Empirical investigation of e-learning acceptance and assimilation: A structural equation model. Appl. Comput. Inform. 2016, 12, 27–50. [Google Scholar] [CrossRef] [Green Version]
  30. Sunny, S.; Patrick, L.; Rob, L. Impact of cultural values on technology acceptance and technology readiness. Int. J. Hosp. Manag. 2019, 77, 89–96. [Google Scholar] [CrossRef]
  31. Jacobs, J.V.; Hettinger, L.J.; Huang, Y.H.; Jeffries, S.; Lesch, M.F.; Simmons, L.A.; Verma, S.K.; Willetts, J.L. Employee acceptance of wearable technology in the workplace. Appl. Ergon. 2019, 78, 148–156. [Google Scholar] [CrossRef]
  32. Scherer, R.; Siddiq, F.; Tondeur, J. The technology acceptance model (TAM): A meta-analytic structural equation modeling approach to explaining teachers’ adoption of digital technology in education. Comput. Educ. 2019, 128, 13–35. [Google Scholar] [CrossRef]
  33. Estriegana, R.; Medina-Merodio, J.-A.; Barchino, R. Student acceptance of virtual laboratory and practical work: An extension of the technology acceptance model. Comput. Educ. 2019, 135, 1–14. [Google Scholar] [CrossRef]
  34. Taherdoost, H. A review of technology acceptance and adoption models and theories. Procedia Manuf. 2018, 22, 960–967. [Google Scholar] [CrossRef]
  35. Dehghani, M.; Kim, K.J.; Dangelico, R.M. Will smartwatches last? Factors contributing to intention to keep using smart wearable technology. Telemat. Inform. 2018, 35, 480–490. [Google Scholar] [CrossRef]
  36. Obal, M. What drives post-adoption usage? Investigating the negative and positive antecedents of disruptive technology continuous adoption intentions. Ind. Mark. Manag. 2017, 63, 42–52. [Google Scholar] [CrossRef]
  37. Choi, B.K.; Yeo, W.-D.; Won, D. The implication of ANT (Actor-Network-Theory) methodology for R&D policy in open innovation paradigm. Knowl. Manag. Res. Pract. 2018, 16, 315–326. [Google Scholar]
  38. Ifinedo, P. Examining students’ intention to continue using blogs for learning: Perspectives from technology acceptance, motivational, and social-cognitive frameworks. Comput. Human Behav. 2017, 72, 189–199. [Google Scholar] [CrossRef]
  39. Mishra, P.; Koehler, M.J. Technological pedagogical content knowledge: A framework for teacher knowledge. Teach. Coll. Rec. 2006, 108, 1017–1054. [Google Scholar] [CrossRef]
  40. Davis, F.D. Perceived usefulness, perceived ease of use, and user acceptance of information technology. MIS Q. 1989, 13, 319–340. [Google Scholar] [CrossRef] [Green Version]
  41. Bandura, A. Social Foundations of Thought and Action. In The Health Psychology Reader; SAGE: Englewood Cliffs, NJ, USA, 1986. [Google Scholar]
  42. Eisenberger, R.; Huntington, R.; Hutchison, S.; Sowa, D. Perceived organizational support. J. Appl. Psychol. 1986, 71, 500. [Google Scholar] [CrossRef]
  43. Vallerand, R.J. Toward a hierarchical model of intrinsic and extrinsic motivation. In Advances in Experimental Social Psychology; Elsevier: Amsterdam, The Netherlands, 1997; Volume 29, pp. 271–360. [Google Scholar]
  44. Ryan, R.M.; Deci, E.L. Self-determination theory and the facilitation of intrinsic motivation, social development, and well-being. Am. Psychol. 2000, 55, 68. [Google Scholar] [CrossRef]
  45. Al-Emran, M.; Salloum, S.A. Students’ Attitudes Towards the Use of Mobile Technologies in e-Evaluation. Int. J. Interact. Mob. Technol. 2017, 11, 195–202. [Google Scholar] [CrossRef]
  46. Sabah, N.M. Motivation factors and barriers to the continuous use of blended learning approach using Moodle: Students’ perceptions and individual differences. Behav. Inf. Technol. 2020, 39, 875–898. [Google Scholar] [CrossRef]
  47. Al-Busaidi, K.A. An empirical investigation linking learners’ adoption of blended learning to their intention of full e-learning. Behav. Inf. Technol. 2013, 32, 1168–1176. [Google Scholar] [CrossRef]
  48. Al-Busaidi, K.A.; Al-Shihi, H. Key factors to instructors’ satisfaction of learning management systems in blended learning. J. Comput. High. Educ. 2012, 24, 18–39. [Google Scholar] [CrossRef]
  49. Hwang, I.-H.; Tsai, S.-J.; Yu, C.-C.; Lin, C.-H. An empirical study on the factors affecting continuous usage intention of double reinforcement interactive e-portfolio learning system. In Proceedings of the 2011 6th IEEE Joint International Information Technology and Artificial Intelligence Conference, Chongqing, China, 20–22 August 2011; Volume 1, pp. 246–249. [Google Scholar]
  50. Lin, K.-M. e-Learning continuance intention: Moderating effects of user e-learning experience. Comput. Educ. 2011, 56, 515–526. [Google Scholar] [CrossRef]
  51. Ho, C.-H. Continuance intention of e-learning platform: Toward an integrated model. Int. J. Electron. Bus. Manag. 2010, 8, 206. [Google Scholar]
  52. Lee, M.-C. Explaining and predicting users’ continuance intention toward e-learning: An extension of the expectation–confirmation model. Comput. Educ. 2010, 54, 506–516. [Google Scholar] [CrossRef]
  53. Shulman, L. Knowledge and teaching: Foundations of the new reform. Harv. Educ. Rev. 1987, 57, 1–23. [Google Scholar] [CrossRef]
  54. Schmidt, D.A.; Baran, E.; Thompson, A.D.; Mishra, P.; Koehler, M.J.; Shin, T.S. Technological pedagogical content knowledge (TPACK) the development and validation of an assessment instrument for preservice teachers. J. Res. Technol. Educ. 2009, 42, 123–149. [Google Scholar] [CrossRef]
  55. Chai, C.S.; Koh, J.H.L.; Tsai, C.-C.; Tan, L.L.W. Modeling primary school pre-service teachers’ Technological Pedagogical Content Knowledge (TPACK) for meaningful learning with information and communication technology (ICT). Comput. Educ. 2011, 57, 1184–1193. [Google Scholar] [CrossRef]
  56. Oner, D. A virtual internship for developing technological pedagogical content knowledge. Australas. J. Educ. Technol. 2020, 36, 27–42. [Google Scholar] [CrossRef]
  57. Koehler, M.; Mishra, P. What is technological pedagogical content knowledge (TPACK)? Contemp. Issues Technol. Teach. Educ. 2009, 9, 60–70. [Google Scholar] [CrossRef] [Green Version]
  58. Koehler, M.J.; Shin, T.S.; Mishra, P. How do we measure TPACK? Let me count the ways. In Educational Technology, Teacher Knowledge, and Classroom Impact: A Research Handbook on Frameworks and Approaches; IGI Global: Hershey, PA, USA, 2012; pp. 16–31. [Google Scholar]
  59. Lu, Y.-L.; Lien, C.-J. Are they learning or playing? Students’ perception traits and their learning self-efficacy in a game-based learning environment. J. Educ. Comput. Res. 2020, 57, 1879–1909. [Google Scholar] [CrossRef]
  60. Cai, J.; Yang, H.H.; Gong, D.; MacLeod, J.; Zhu, S. Understanding the continued use of flipped classroom instruction: A personal beliefs model in Chinese higher education. J. Comput. High. Educ. 2019, 31, 137–155. [Google Scholar] [CrossRef]
  61. Venkatesh, V. Determinants of perceived ease of use: Integrating control, intrinsic motivation, and emotion into the technology acceptance model. Inf. Syst. Res. 2000, 11, 342–365. [Google Scholar] [CrossRef] [Green Version]
  62. Chen, K.; Chen, J.V.; Yen, D.C. Dimensions of self-efficacy in the study of smart phone acceptance. Comput. Stand. Interfaces 2011, 33, 422–431. [Google Scholar] [CrossRef]
  63. Cebeci, U.; Ertug, A.; Turkcan, H. Exploring the determinants of intention to use self-checkout systems in super market chain and its application. Manag. Sci. Lett. 2020, 10, 1027–1036. [Google Scholar] [CrossRef]
  64. Bailey, A.A.; Pentina, I.; Mishra, A.S.; Mimoun, M.S.B. Mobile payments adoption by US consumers: An extended TAM. Int. J. Retail Distrib. Manag. 2017, 45, 626–640. [Google Scholar] [CrossRef]
  65. Fathema, N.; Shannon, D.; Ross, M. Expanding the Technology Acceptance Model (TAM) to Examine Faculty Use of Learning Management Systems (LMSs) In Higher Education Institutions. J. Online Learn. Teach. 2015, 11, 210–232. [Google Scholar]
  66. Lam, S.; Cheng, R.W.; Choy, H.C. School support and teacher motivation to implement project-based learning. Learn. Instr. 2010, 20, 487–497. [Google Scholar] [CrossRef] [Green Version]
  67. Kurup, V.; Hersey, D. The changing landscape of anesthesia education: Is Flipped Classroom the answer? Curr. Opin. Anesthesiol. 2013, 26, 726–731. [Google Scholar] [CrossRef]
  68. Missildine, K.; Fountain, R.; Summers, L.; Gosselin, K. Flipping the classroom to improve student performance and satisfaction. J. Nurs. Educ. 2013, 52, 597–599. [Google Scholar] [CrossRef] [Green Version]
  69. Bøe, T.; Gulbrandsen, B.; Sørebø, Ø. How to stimulate the continued use of ICT in higher education: Integrating information systems continuance theory and agency theory. Comput. Human Behav. 2015, 50, 375–384. [Google Scholar] [CrossRef]
  70. Bölen, M.C. Exploring the determinants of users’ continuance intention in smartwatches. Technol. Soc. 2020, 60, 101209. [Google Scholar] [CrossRef]
  71. Krejcie, R.V.; Morgan, D.W. Determining sample size for research activities. Educ. Psychol. Meas. 1970, 30, 607–610. [Google Scholar] [CrossRef]
  72. Chuan, C.L.; Penyelidikan, J. Sample size estimation using Krejcie and Morgan and Cohen statistical power analysis: A comparison. J. Penyelid. IPBL 2006, 7, 78–86. [Google Scholar]
  73. Bhattacherjee, A. Understanding information systems continuance: An expectation-confirmation model. MIS Q. 2001, 25, 351–370. [Google Scholar] [CrossRef]
  74. Almahamid, S.; Rub, F.A. Factors that determine continuance intention to use e-learning system: An empirical investigation. Int. Conf. Telecommun. Tech. Appl. Proc. CSIT 2011, 5, 242–246. [Google Scholar]
  75. Thiruchelvi, A.; Koteeswari, S. A conceptual framework of employees’ continuance intention to use e-learning system. Asian J. Res. Bus. Econ. Manag. 2013, 3, 14–20. [Google Scholar]
  76. Grandgenett, N.F. Perhaps a matter of imagination: TPCK in mathematics education. In Handbook of Technological Pedagogical Content Knowledge (TPCK) for Educators; Routledge: London, UK, 2008; p. 145166. [Google Scholar]
  77. Compeau, D.R.; Higgins, C.A. Computer self-efficacy: Development of a measure and initial test. MIS Q. 1995, 19, 189–211. [Google Scholar] [CrossRef] [Green Version]
  78. Nunnally, J.C.; Bernstein, I.H. Psychometric Theory; McGraw-Hill: New York, NY, USA, 1994. [Google Scholar]
  79. Kline, R.B. Principles and Practice of Structural Equation Modeling; Guilford Publications: New York, NY, USA, 2015. [Google Scholar]
  80. Dijkstra, T.K.; Henseler, J. Consistent and asymptotically normal PLS estimators for linear structural equations. Comput. Stat. Data Anal. 2015, 81, 10–23. [Google Scholar] [CrossRef] [Green Version]
  81. Hair, J.F.; Ringle, C.M.; Sarstedt, M. PLS-SEM: Indeed a silver bullet. J. Mark. Theory Pract. 2011, 19, 139–152. [Google Scholar] [CrossRef]
  82. Henseler, J.; Ringle, C.M.; Sinkovics, R.R. The use of partial least squares path modeling in international marketing. In New Challenges to International Marketing; Emerald Group Publishing Limited: Bingley, UK, 2009; pp. 277–319. [Google Scholar]
  83. Fornell, C.; Larcker, D.F. Evaluating Structural Equation Models With Unobservable Variables and Measurement Error. J. Mark. Res. 1981, 18, 39–50. [Google Scholar] [CrossRef]
  84. Trial, D. Model Fit. Available online: https://www.modelfit.com/ (accessed on 20 December 2020).
  85. Hair, J.F., Jr.; Hult, G.T.M.; Ringle, C.; Sarstedt, M. A Primer on Partial Least Squares Structural Equation Modeling (PLS-SEM); Sage Publications: New York, NY, USA, 2016. [Google Scholar]
  86. Hu, L.; Bentler, P.M. Fit indices in covariance structure modeling: Sensitivity to underparameterized model misspecification. Psychol. Methods 1998, 3, 424. [Google Scholar] [CrossRef]
  87. Bentler, P.M.; Bonett, D.G. Significance tests and goodness of fit in the analysis of covariance structures. Psychol. Bull. 1980, 88, 588. [Google Scholar] [CrossRef]
  88. Lohmöller, J.B. Latent Variable Path Modeling with Partial Least Squares; Physica: Heidelberg, Germany, 1989. [Google Scholar]
  89. Henseler, J.; Dijkstra, T.K.; Sarstedt, M.; Ringle, C.M.; Diamantopoulos, A.; Straub, D.W.; Ketchen, D.J., Jr.; Hair, J.F.; Hult, G.T.M.; Calantone, R.J. Common beliefs and reality about PLS: Comments on Rönkkö and Evermann (2013). Organ. Res. Methods 2014, 17, 182–209. [Google Scholar] [CrossRef] [Green Version]
  90. Baranik, L.E.; Roling, E.A.; Eby, L.T. Why does mentoring work? The role of perceived organizational support. J. Vocat. Behav. 2010, 76, 366–373. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  91. Aselage, J.; Eisenberger, R. Perceived organizational support and psychological contracts: A theoretical integration. J. Organ. Behav. Int. J. Ind. Occup. Organ. Psychol. Behav. 2003, 24, 491–509. [Google Scholar] [CrossRef]
  92. Stamper, C.L.; Johlke, M.C. The impact of perceived organizational support on the relationship between boundary spanner role stress and work outcomes. J. Manag. 2003, 29, 569–588. [Google Scholar]
  93. Mitchell, J.I.; Gagné, M.; Beaudry, A.; Dyer, L. The role of perceived organizational support, distributive justice and motivation in reactions to new information technology. Comput. Human Behav. 2012, 28, 729–738. [Google Scholar] [CrossRef] [Green Version]
  94. Alsofyani, M.M.; Aris, B.b.; Eynon, R.; Majid, N.A. A preliminary evaluation of short blended online training workshop for TPACK development using technology acceptance model. Turkish Online J. Educ. Technol. 2012, 11, 20–32. [Google Scholar]
  95. Siahaan, A.U.; Aji, S.B.; Antoni, C.; Handayani, Y. Online Social Learning Platform vs E-Learning for Higher Vocational Education in Purpose to English Test Preparation. In Proceedings of the 7th International Conference on English Language and Teaching (ICOELT 2019), Padang, Indonesia, 4–5 November 2019; pp. 76–82. [Google Scholar]
  96. Heidig, S.; Clarebout, G. Do pedagogical agents make a difference to student motivation and learning? Educ. Res. Rev. 2011, 6, 27–54. [Google Scholar] [CrossRef]
  97. Corbalan, G.; Kester, L.; van Merriënboer, J.J.G. Towards a personalized task selection model with shared instructional control. Instr. Sci. 2006, 34, 399–422. [Google Scholar] [CrossRef] [Green Version]
  98. Zaric, N.; Lukarov, V.; Schroder, U. A Fundamental Study for Gamification Design: Exploring Learning Tendencies’ Effects. Int. J. Serious Games 2020, 7, 3–25. [Google Scholar] [CrossRef]
  99. Hanif, M. Students’ Self-Regulated Learning in Iconic Mobile Learning System in English Cross-Disciplined Program. Anatol. J. Educ. 2020, 5, 121–130. [Google Scholar] [CrossRef]
  100. Fathema, N.; Sutton, K.L. Factors influencing faculty members’ Learning Management Systems adoption behavior: An analysis using the Technology Acceptance Model. Int. J. Trends Econ. Manag. Technol. 2013, 2, 20–28. [Google Scholar]
  101. Panda, S.; Mishra, S. E-Learning in a Mega Open University: Faculty attitude, barriers and motivators. EMI Educ. Media Int. 2007, 44, 323–338. [Google Scholar] [CrossRef]
Figure 1. The e-learning technology model adopted for teachers. Note: H1–5 = Hypotheses 1–5.
Figure 1. The e-learning technology model adopted for teachers. Note: H1–5 = Hypotheses 1–5.
Education 11 00006 g001
Figure 2. The e-learning technology model adopted for students.
Figure 2. The e-learning technology model adopted for students.
Education 11 00006 g002
Figure 3. Path coefficient of the model (significant at ** p <= 0.01, * p < 0.05) (Model A).
Figure 3. Path coefficient of the model (significant at ** p <= 0.01, * p < 0.05) (Model A).
Education 11 00006 g003
Figure 4. Path coefficient of the model (significant at ** p < = 0.01, * p < 0.05) (Model B).
Figure 4. Path coefficient of the model (significant at ** p < = 0.01, * p < 0.05) (Model B).
Education 11 00006 g004
Table 1. Most relevant studies of e-learning platforms in different sectors.
Table 1. Most relevant studies of e-learning platforms in different sectors.
Authors/ReferenceTarget PopulationObjective/GoalModels Adopted
[45]StudentsTo explain the f-variables that affect continued use of m-learning.TAM, Theory of Planned Behavior (TPB), and Expectation Confirmation Model (ECM).
[46]StudentsTo examine students’ continuous use of blended learning, with reference to behavioral attitudes, motivations, and barriers.TAM, TPB and self-determination theory (SDT).
[47]StudentsTo make a connection between learners’ adoption and satisfaction with LMS in blended learning in relation to certain learners’ personal characteristics in terms of continuous use of the e-learning environment.TAM and satisfaction factor (SAT).
[48]InstructorsTo examine the influential factors which may contribute to instructors’ satisfaction with LMS use in a blended learning atmosphere.LMS, system and instructors’ characteristics that are derived from well-established factors.
[49]StudentsTo investigate students’ behavior of continuance intentions to use the double reinforcement interactive e-portfolio learning system. TAM and IS continuance post-acceptance model (IS-TAM).
[50]LearnersTo investigate the basic determinants behind the continuous intention to use e-learning. TAM and Negative Critical Incident (NCI).
[51]People chosen randomly through a high-traffic websiteTo investigate the motivational factors that affect the synthesized model that is composed of a combination of TAM, ECM, COGM and SDM.TAM, ECM and cognitive model (COGM).
[52]Technology usersTo investigate and predict the main reason behind users’ intentions to continue using e-learning.ECM, TAM, and, TPB.
Table 2. Demographic data of the respondents.
Table 2. Demographic data of the respondents.
CriterionFactorFrequencyPercentage
GenderFemale17547%
Male19753%
AgeBetween 18 and 2912233%
Between 30 and 399826%
Between 40 and 498824%
Between 50 and 596417%
FacultiesFaculty of Engineering and IT14539%
Faculty of Education12935%
Faculty of Business and Law9826%
Education qualificationBachelor18249%
Master15742%
Doctorate339%
Table 3. Constructs and their sources.
Table 3. Constructs and their sources.
ConstructsNumber of ItemsSource
CU2[73,74,75]
CTRLM5[43]
TPACK4[39,76]
TSE7[41,77]
PEOU3[40]
PU4[40]
POS5[42]
Note: TPACK = Technological pedagogical content knowledge; TSE = Technology self-efficacy; PEOU = Perceived ease of use; PU = Perceived usefulness; POS = Perceived organizational support; CTRLM = Controlled motivation; CU = Continuous intention to use e-learning platform.
Table 4. Cronbach’s alpha values for the pilot study (Cronbach’s alpha ≥ 0.70) for teachers (Model A).
Table 4. Cronbach’s alpha values for the pilot study (Cronbach’s alpha ≥ 0.70) for teachers (Model A).
ConstructsCronbach’s Alpha
CU0.756
TPACK0.779
TSE0.864
PEOU0.889
PU0.734
POS0.852
Note: TPACK = Technological pedagogical content knowledge; TSE = Technology self-efficacy; PEOU = Perceived ease of use; PU = Perceived usefulness; POS = Perceived organizational support; CU = Continuous intention to use e-learning platform.
Table 5. Cronbach’s alpha values for the pilot study (Cronbach’s alpha ≥ 0.70) for students (Model B).
Table 5. Cronbach’s alpha values for the pilot study (Cronbach’s alpha ≥ 0.70) for students (Model B).
ConstructsCronbach’s Alpha
CU0.872
CTRLM0.881
TSE0.798
PEOU0.736
PU0.797
Note: TSE = Technology self-efficacy; PEOU = Perceived ease of use; PU = Perceived usefulness; CTRLM = Controlled motivation; CU = Continuous intention to use e-learning platform.
Table 6. Convergent validity results that ensure acceptable values (Factor loading, Cronbach’s alpha, composite reliability (CR), Dijkstra–Henseler’s rho (pA) ≥ 0.70 and average variance extracted (AVE) > 0.5) (Model A).
Table 6. Convergent validity results that ensure acceptable values (Factor loading, Cronbach’s alpha, composite reliability (CR), Dijkstra–Henseler’s rho (pA) ≥ 0.70 and average variance extracted (AVE) > 0.5) (Model A).
ConstructsItemsFactor
Loading
Cronbach’s AlphaCRpAAVE
Technology Self-EfficacyTSE10.7750.8740.7990.8320.536
TSE20.736
TSE30.820
TSE40.901
TSE50.756
TSE60.723
TSE70.797
Technological Pedagogical Content KnowledgeTPACK 10.7110.8290.8820.7910.552
TPACK 20.869
TPACK 30.909
TPACK 40.790
Perceived Ease of UsePEOU10.8290.8440.8120.8170.661
PEOU20.847
PEOU30.746
Perceived UsefulnessPU10.7340.8160.8280.8250.623
PU20.766
PU30.889
PU40.850
Perceived Organizational SupportPOS10.7290.8630.8140.8830.718
POS20.848
POS30.758
POS40.819
POS50.878
Continuous intention to use e-learning platformCU10.7960.8150.8760.8980.673
CU20.801
Table 7. Convergent validity results that ensure acceptable values (Factor loading, Cronbach’s alpha, composite reliability, Dijkstra–Henseler’s rho ≥ 0.70 & AVE > 0.5) (Model B).
Table 7. Convergent validity results that ensure acceptable values (Factor loading, Cronbach’s alpha, composite reliability, Dijkstra–Henseler’s rho ≥ 0.70 & AVE > 0.5) (Model B).
ConstructsItemsFactor
Loading
Cronbach’s AlphaCRPAAVE
Technology Self-EfficacyTSE10.726
TSE20.8260.7820.8330.8230.705
TSE30.710
TSE40.868
TSE50.746
TSE60.733
Perceived Ease of UsePEOU10.7630.8950.8000.8360.559
PEOU20.890
PEOU30.849
Perceived UsefulnessPU10.7930.8560.8790.8080.696
PU20.709
PU30.873
PU40.821
Controlled MotivationCTRL10.8320.8050.7960.8070.700
CTRL20.802
CTRL30.875
CTRL40.810
CTRL50.796
Continuous intention to use e-learning platformCU10.7250.8780.8180.8160.509
CU20.878
Table 8. Fornell–Larcker Scale (Model A).
Table 8. Fornell–Larcker Scale (Model A).
TSETPACKPEOUPUPOSCU
TSE0.876
TPACK0.1650.845
PEOU0.1250.2530.802
PU0.5690.4870.5580.790
POS0.1870.2020.2910.1150.787
CU0.3690.1980.3780.3830.1780.803
Note: TPACK = Technological pedagogical content knowledge; TSE = Technology self-efficacy; PEOU = Perceived ease of use; PU = Perceived usefulness; POS = Perceived organizational support; CU = Continuous intention to use e-learning platform.
Table 9. Fornell–Larcker Scale (Model B).
Table 9. Fornell–Larcker Scale (Model B).
TSEPEOUPUCTRLCU
TSE0.768
PEOU0.3680.801
PU0.2670.2290.887
CTRL0.6490.4920.3990.844
CU0.4220.3270.3020.1880.870
Note: TSE = Technology self-efficacy; PEOU = Perceived ease of use; PU = Perceived usefulness; CTRLM = Controlled motivation; CU = Continuous intention to use e-learning platform.
Table 10. Heterotrait–Monotrait Ratio (HTMT) (Model A).
Table 10. Heterotrait–Monotrait Ratio (HTMT) (Model A).
TSETPACKPEOUPUPOSCU
TSE
TPACK0.560
PEOU0.1360.487
PU0.2660.3630.556
POS0.2960.2000.2700.544
CU0.3890.6350.3780.6380.555
Note: TPACK = Technological pedagogical content knowledge; TSE = Technology self-efficacy; PEOU = Perceived ease of use; PU = Perceived usefulness; POS = Perceived organizational support; CU = Continuous intention to use e-learning platform.
Table 11. Heterotrait–Monotrait Ratio (HTMT) (Model B).
Table 11. Heterotrait–Monotrait Ratio (HTMT) (Model B).
TSEPEOUPUCTRLCU
TSE
PEOU0.232
PU0.5060.436
CTRL0.3920.4570.503
CU0.6970.6090.2100.264
Note TSE = Technology self-efficacy; PEOU = Perceived ease of use; PU = Perceived usefulness; CTRLM = Controlled motivation; CU = Continuous intention to use e-learning platform.
Table 12. Model fit indicators (Model A).
Table 12. Model fit indicators (Model A).
Complete Model
Saturated ModelEstimated Mod
SRMR0.0310.041
d_ULS0.7863.216
d_G0.5650.535
Chi-Square466.736473.348
NFI0.6240.627
RMS Theta0.073
Table 13. Model fit indicators (Model B).
Table 13. Model fit indicators (Model B).
Complete Model
Saturated ModelEstimated Mod
SRMR0.0120.031
d_ULS0.6052.317
d_G0.5160.506
Chi-Square461.646472.347
NFI0.6330.642
RMS Theta0.061
Table 14. R2 of the endogenous latent variables (Model A).
Table 14. R2 of the endogenous latent variables (Model A).
ConstructsR2Results
Continuous intention to use e-learning platform0.832High
Table 15. R2 of the endogenous latent variables (Model B).
Table 15. R2 of the endogenous latent variables (Model B).
ConstructsR2Results
Continuous intention to use e-learning platform0.709High
Table 16. Hypotheses testing of the research model (significant at ** p < = 0.01, * p < 0.05) (Model A).
Table 16. Hypotheses testing of the research model (significant at ** p < = 0.01, * p < 0.05) (Model A).
HRelationshipPatht-Valuep-ValueDirectionDecision
H1TPACK -> CU0.33612.2230.001PositiveSupported **
H2TSE -> CU0.4265.2690.026PositiveSupported *
H3PU -> CU0.5896.7160.018PositiveSupported *
H4PEOU -> CU0.6255.5840.023PositiveSupported *
H5POS -> CU0.55316.1080.000PositiveSupported **
Note: TPACK = Technological pedagogical content knowledge; TSE = Technology self-efficacy; PEOU = Perceived ease of use; PU = Perceived usefulness; POS = Perceived organizational support; CU = Continuous intention to use e-learning platform.
Table 17. Hypotheses testing of the research model (significant at ** p < = 0.01, * p < 0.05) (Model B).
Table 17. Hypotheses testing of the research model (significant at ** p < = 0.01, * p < 0.05) (Model B).
HRelationshipPatht-Valuep-ValueDirectionDecision
H2TSE -> CU0.29014.5780.000PositiveSupported **
H3PEOU -> CU0.3573.1160.043PositiveSupported *
H4PU -> CU0.4652.6460.035PositiveSupported *
H6CTRLM -> CU0.2434.3610.033PositiveSupported *
Note: TSE = Technology self-efficacy; PEOU = Perceived ease of use; PU = Perceived usefulness; CTRLM = Controlled motivation; CU = Continuous intention to use e-learning platform.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Saeed Al-Maroof, R.; Alhumaid, K.; Salloum, S. The Continuous Intention to Use E-Learning, from Two Different Perspectives. Educ. Sci. 2021, 11, 6. https://0-doi-org.brum.beds.ac.uk/10.3390/educsci11010006

AMA Style

Saeed Al-Maroof R, Alhumaid K, Salloum S. The Continuous Intention to Use E-Learning, from Two Different Perspectives. Education Sciences. 2021; 11(1):6. https://0-doi-org.brum.beds.ac.uk/10.3390/educsci11010006

Chicago/Turabian Style

Saeed Al-Maroof, Rana, Khadija Alhumaid, and Said Salloum. 2021. "The Continuous Intention to Use E-Learning, from Two Different Perspectives" Education Sciences 11, no. 1: 6. https://0-doi-org.brum.beds.ac.uk/10.3390/educsci11010006

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop