Next Article in Journal
Statistic-Driven Proton Transfer Affecting Nanoscopic Organization in an Ethylammonium Nitrate Ionic Liquid and 1,4-Diaminobutane Binary Mixture: A Steamy Pizza Model
Previous Article in Journal
Robust Hybrid Beamforming Scheme for Millimeter-Wave Massive-MIMO 5G Wireless Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Decision Support System Using Text Mining Based Grey Relational Method for the Evaluation of Written Exams

by
Mehmet Erkan Yuksel
1,* and
Huseyin Fidan
2
1
Department of Computer Engineering, Mehmet Akif Ersoy University, Burdur 15030, Turkey
2
Department of Industrial Engineering, Mehmet Akif Ersoy University, Burdur 15030, Turkey
*
Author to whom correspondence should be addressed.
Submission received: 4 October 2019 / Revised: 7 November 2019 / Accepted: 15 November 2019 / Published: 19 November 2019

Abstract

:
Grey relational analysis (GRA) is a part of the Grey system theory (GST). It is appropriate for solving problems with complicated interrelationships between multiple factors/parameters and variables. It solves multiple-criteria decision-making problems by combining the entire range of performance attribute values being considered for every alternative into one single value. Thus, the main problem is reduced to a single-objective decision-making problem. In this study, we developed a decision support system for the evaluation of written exams with the help of GRA using contextual text mining techniques. The answers obtained from the written exam with the participation of 50 students in a computer laboratory and the answer key prepared by the instructor constituted the data set of the study. A symmetrical perspective allows us to perform relational analysis between the students’ answers and the instructor’s answer key in order to contribute to the measurement and evaluation. Text mining methods and GRA were applied to the data set through the decision support system employing the SQL Server database management system, C#, and Java programming languages. According to the results, we demonstrated that the exam papers are successfully ranked and graded based on the word similarities in the answer key.

1. Introduction

The concept of measurement, which aims to determine and grade differences, is defined as the observation of characteristics and the expression of observation results [1]. It is accepted as a critical component of the education system and contains various techniques to define the level of education status and differences [2]. Between these techniques used to determine the levels of success, measurement tools such as tests, essay exams, written exams, and oral exams are available. Written exams are considered as essential tools when it comes to measuring the original and creative thinking power of students, written expression skills, subject evaluation, and application of knowledge and skills [2,3,4]. However, the validity and reliability of written exams are reduced due to some factors such as non-objective assessment, not to pay attention and care to the scoring criteria, making unnecessary verbiage in answers, scoring difficulties, and time-consuming situations [2,3,4,5].
In recent years, increasing attention to artificial intelligence has encouraged the progress of data mining and analytics in the field of education. Data mining is a knowledge-intensive task concerned with deriving higher-level insights from data. It is defined as ‘‘the non-trivial process of identifying valid, novel, potentially useful, and ultimately understandable patterns in data’’ [6,7] or ‘‘the analysis of large observational datasets to find unsuspected relationships and to summarize the data in novel ways that are both understandable and useful to the data owner’’ [8]. It is the process of discovering patterns (new aspects, hidden knowledge) and extracting useful information from large data sets or databases through the methods used in artificial intelligence, machine learning, statistics, and database systems to help decision-makers to make correct decisions [9,10].
Data mining specialized in the field of education is called Educational Data Mining (EDM). EDM refers to the methods and tools used to obtain information from educational records (i.e., exam results, online logs to track learning behaviors) and then analyzes the information to formulate outcomes. It is concerned with developing methods (i.e., statistics, correlation analysis, regression, online analytical processes, social network analysis, cluster analysis, and text mining) to explore the unique and increasingly large-scale data obtained from educational environments and utilizing those methods to better understand students and the environments in which they learn [11,12,13]. Since data mining techniques are usually used to evaluate training models and tools, EDM is seen as knowledge-intensive work. In the literature, classification, clustering, and estimation techniques have been applied, and various applications have been developed by using structured data. In particular, with the development of internet-based technologies, unstructured data analysis (e.g., documents, text files, media files, images, emails, social media data, and so on) is not encountered much in the field of educational technologies [14,15,16]. However, the fact that the vast majority of educational data is in textual format provides significant analysis opportunities for the research. Text mining, where unstructured data analysis can be performed, is an analysis technique based on data mining of textual data. It is carried out by Information Retrieval (IR) and Natural Language Processing (NLP) methods and claims that a document can be represented by the number of words in the texts [17,18,19].
Text mining, which analyzes emails, online reviews, documents, tweets, sentences, and survey responses, enables applications such as clustering, classification, and summarization according to similarity levels of data. The similarity analysis of documents and text data in excess amounts is particularly of interest in the literature [20,21,22]. In our study, the main goal is to evaluate written exam answers by using text mining in order to contribute to the correct decision of the instructor. We developed a two-stage decision support system that employs the SQL Server database management system, C#, and Java programming languages to rate the exam results. Since the insufficiencies of the Term Frequency-Inverse Document Frequency (TF-IDF) method in short text analysis [22,23,24], we took advantage of Grey relational analysis (GRA), which gives effective results in case of incomplete information. In this context, we constructed a decision matrix by using the Bag of Words (BoW) and Vector Space Model (VSM). Then, we applied a two-stage GRA to the decision matrix. According to the experimental results, we observed that the level of similarity between the exam answers and the answer key is determined successfully by the decision support system we developed.
The remainder of this paper is organized as follows: Section 2 presents a literature review about the evaluation of exams, data mining, text mining, and EDM; Section 3 focuses on the GRA concept; Section 4 describes the proposed method; Section 5 summarizes the findings of the analysis; and Section 6 presents discussions and suggestions.

2. Background

2.1. The Evaluation of Exams

The concept of measurement, which is defined as expressing the result of observations in terms of quality and quantity, is a research topic aimed at determining the differences [1]. The fact that individuals and objects have different characteristics makes it necessary to measure these characteristics [25]. According to [26], the concept of measurement has an approach involving the identification of differences. In other words, by measuring, differences between objects can be quantitatively determined and graded according to the properties possessed by the objects. In this way, measurement methods can be seen as a decision-making process. The concept of assessment, which is used to determine the intended learning outcomes in education levels, rather than a successful or unsuccessful evaluation form today, points to a level of learning-oriented approach [27]. According to [28], when deciding on the level of learning, the interests, motivation, preparation, participation, perseverance, and achievement of the students should be considered. Various measuring instruments or methods are used to evaluate the achievement between the mentioned elements. Commonly used measuring instruments are oral exams, multiple-choice exams (tests), and written exams [29]. Oral exams have low validity and reliability. They cannot go beyond measuring isolated information. The emotional and mental state of the student affects the test result. The appearance of the student (i.e., dressing style) also affects the evaluation result. For these reasons, oral exams are a measurement tool that results in high inconsistencies. In multiple-choice exams, the student chooses the most appropriate answer to the question. However, these exams reduce expression skills, decision-making skills, and interpretation power, and cannot measure mental skills. It is not seen as a robust measurement tool for such reasons. Contrary to oral exams and multiple-choice exams, there is a consensus in the literature that written exams offer effective results in measurement and assessment, although they are laborious in practice [1,2,3,4,5]. The advantages and disadvantages of these exam types are presented in Table 1, Table 2, and Table 3, respectively.

2.1.1. Written Exams

When it comes to putting together ideas, organizing, evaluating, and applying knowledge and skills, the most appropriate method for the measurement is written exams. Therefore, written exams, which aim to find out the student’s original and creative thinking power, are a beneficial technique for measuring writing skills and are quite a widely applied method for measurement at all levels of the education system starting from primary school [3,4,5]. Written exams, which are also called classical exams and open-ended exams, are created with answers by the students in a handwritten or an electronic environment. The answers to the written exams, which reflect the students’ knowledge and thinking skills about the subject, are usually short texts [25,26,27].
Written exams, which are suitable for measuring high-level mental skills, have some advantages over the other exam types. The most important feature of the written exams is the freedom to answer. The student is free to decide on how to approach the questions, which practical information to use, how to arrange the answer, and how much emphasis is placed on every aspect of the answer. Therefore, the written exams enable students to have the ability to produce, integrate, and express ideas and to develop skills to state their thoughts in written form. Furthermore, unlike other types of tests, these exams increase the validity of results because of the chance factor and lack of possibility of copying, providing a more accurate determination of the outcomes and performance [26,27,28,29].
Besides the advantages of written exams, there are some restrictions. According to educators, exam scoring is the most significant limitation in the evaluation of exams [1,2,3]. The scoring process can be influenced by the paper layout, imperfections, in-class behavior, and students’ family characteristics (i.e., interests and attitudes). Therefore, different scores can be shown on the same answer sheet. Consequently, there may be some situations in which results distort the measuring success and reduce reliability [3,4,5,25]. In [4], the authors emphasized that evaluators should pay attention to the learning outcomes and evaluate answers carefully. In this context, the evaluation process was described as “three blind men and an elephant”. On the other hand, students who do not know the exact answer can write answers consisting of unnecessary statements to show that they have knowledge. In this case, it is challenging to identify information to be scored, and the educator loses the objectivity; thus, reducing the validity and the reliability of the exam. Finally, the excessive course load on the students can extend the period in which the educator evaluates the learning success, and reduce the efficiency of exams through these results [2,3,4,28,29,30].
Because of the positive effects of results obtained from written exams on the measurement and evaluation, we focused on the analysis of these exams. From this point of view, we developed a decision support system as a solution to the problems experienced in scoring the written exams.

2.2. Data Mining

With the increasing growth of data used in many application areas (e.g., bioinformatics, business intelligence, finance, retail, health, telecommunication, Web search engines, social media, and digital libraries), data mining meets the needs for flexible, scalable, and efficient data analysis in today’s information age. Data mining (also known as knowledge discovery from data) can be considered as a natural evolution of information technology and a confluence of several disciplines (i.e., machine learning, statistics, information retrieval, pattern recognition, and bioinformatics) and application fields. It is the process of discovering interesting patterns (that represent knowledge and are novel, potentially useful, and easily understandable by humans) from large amounts of data [31,32]. Data mining, as a knowledge discovery process, usually includes an iterative sequence of the following steps: data cleaning, data integration, data selection, data transformation, pattern discovery, pattern evaluation, and knowledge presentation. It can be performed by any kind of data (e.g., transactional data, data warehouse data, database data, advanced data types that have versatile forms, structures and rather different semantic meanings, such as Web data, engineering design data, data streams, hypertext and multi-media data, graph and networked data, spatial and spatiotemporal data, time-related or sequence data) as long as the data are meaningful for the target application [33,34].
There are several techniques to specify the kinds of patterns to be discovered in data mining tasks. In general, such data mining tasks are classified into two major groups: descriptive and predictive. Descriptive tasks characterize the properties of data in a target dataset. Predictive tasks carry out induction on the available data to make predictions [17]. The main techniques used in data mining tasks are classification and regression; characterization and discrimination; cluster analysis; the mining of frequent patterns, associations, and correlations; association rules; outlier detection; sequence analysis; time series analysis; sentiment analysis; social network analysis; prediction; and text mining [31,32]. As a highly application-driven discipline, data mining incorporates many techniques from the other research areas such as artificial intelligence, machine learning, database systems, data warehouses, information retrieval, high-performance computing, statistics, pattern recognition, and visualization [33,34]. Consequently, the interdisciplinary nature of data mining development significantly contributes to the success of data mining and its extensive application domains.

2.3. Text Mining

The analysis of structural data is possible with data mining approaches. However, the vast majority of data we encounter in our daily lives, such as documents, texts, and pictures, are composed of unstructured data. In the analysis of these data, the insufficiency of data mining approaches requires the use of analyses involving textual operations and language constructs. In this context, text mining can be defined as the process to obtain confidential information from textual data [35]. The first study on the analysis of textual data was done by [36]. The authors, aiming to extract keywords based on NLP from project descriptions recorded in the database in text form, have performed a cluster analysis of the documents by using those keywords. The process of converting textual data into a structured form was first described by [37]. In this study, the textual data in documents were converted into a structured form using IR, and then meaningful concepts were determined by using the text categorization approach of [36]. In recent years, the use of text mining, which is also an interdisciplinary research area related to the IR, the NLP, and the Named Entity Recognition (NER), has increased rapidly and become the focus of interest in almost all application domains.
To be able to analyze the textual data, first of all, they must be represented numerically. For the numerical representation of a document, the Bag of Words (BoW), the Binary model, and the Vector Space Model (VSM) representations are employed [18]. In the BoW method, the number of words in the document is determined, and the document can, therefore, be represented numerically [38,39]. In the Binary model, a set of terms in the document is referenced, and the values “0” and “1” are encoded to determine whether or not a specific term is related to the document; “1” is used in the case of the term is in the text, otherwise “0” is used. The VSM is used to vectorize the word counts in the text [40]. In other words, it can be considered as a matrix based on BoW, which shows the word frequencies in the document. If more than one document is compared, each word is weighted depending on the frequency numbers in other documents. In general, the TF-IDF method is used for the weighting operation.
There are two techniques used in text mining: contextual analysis and lexical analysis [31,32]. The contextual analysis is aimed at determining words and word associations based on the frequency values of the words in the text. The BoW and VSM methods are employed in contextual analysis to represent the documents. Lexical analysis is carried out in the context of NLP for the meaningful examination of texts [39,41]. Within the meaningful analyses, applications such as opinion mining, sentiment analysis, evaluation mining, and thought mining are performed and used to extract social, political, commercial, and cultural meanings from textual data. Besides, the main methods used in text mining include clustering, classification, and summarization [19,42,43].
  • Clustering: It is the method to organize similar contents from unstructured data sources such as documents, news, images, paragraphs, sentences, comments, or terms in order to enhance retrieval and support browsing. It is the process to group the contents based on fuzzy information, such as words or word phrases in a set of documents. The similarity is computed using a similarity function. Measurement methods such as Cosine distance, Manhattan distance, and Euclidean distance are used for the clustering process. Besides, there is a wide variety of different clustering algorithms, such as hierarchical algorithms, partitioning algorithms, and standard parametric modeling-based methods. These algorithms are grouped along different dimensions based either on the underlying methodology of the algorithm, leading to agglomerative or partitional approaches, or on the structure of the final solution, leading to hierarchical or non-hierarchical solutions [35,38,43].
  • Classification: It is a data mining approach to the grouping of the data according to the specified characteristics. It is carried out in two stages as learning and classification. In the first stage, a part of the data set is used for training purposes to determine how data characteristics will be classified. In the second stage, all datasets are classified by these rules. Supervised machine learning algorithms are used for classification methods. These algorithms include decision trees, I Bayes, nearest neighbor, classification and regression trees, support vector machines, and genetic algorithms [31,35,43].
  • Summarization: It is used for documents and aims to determine the meanings, words, and phrases that can represent a document. It is carried out based on the language-specific rules that textual data belongs to. NLP approaches, which are one of the complex processes in terms of computer systems, are used in applications such as speech recognition, language translation, automated response systems, text summarization, and sentiment analysis [31,35,43].

2.4. Educational Data Mining (EDM)

EDM is a research area in which education system issues (e.g., learning models, programs, activities, teaching methods and materials, resources, and vocational qualifications of teachers) are studied based on data mining. It aims to develop education models and techniques and analyze data related to the attitudes and behaviors of students, teachers, administrators, teaching methods and materials [13]. In the literature, we observed that data mining techniques such as clustering, classification, and prediction are typically used in EDM studies [16,17]. With EDM applications, educators, administrators, and researchers make effective decisions by using tools such as student modeling, performance assessment, educational outcome prediction, course activity analysis, management activity analysis, and learning behaviors analysis [44]. In this context, EDM analyses that employ statistical methods and machine learning techniques are decision support systems performed with educational data.
The fact that the educational data consists of multiple factors and that all data cannot be transferred to the electronic environment creates problems in educational analysis processes [45]. Therefore, many educators, managers, and policymakers have hesitated to use data mining methods. In recent years, the transfer of historical data to the electronic media and the more regular recording of data have attracted the attention of researchers to the field of data mining. Besides, the experience is essential in acquiring general and special field vocational competences in education. In this context, it is indisputable that the analysis of data based on past experiences will significantly contribute to the education process. However, the inclusion of historical data in the analysis poses more challenging processes for EDM. Traditional data mining techniques used in EDM are inadequate in complex data analysis. Thus, it is inevitable that EDM analyses make use of artificial intelligence techniques, such as machine learning, NLP, expert systems, deep learning, neural networks, and big data. The analysis of educational data (the evaluation of relations, patterns, and associations based on the historical data) through these techniques will enable more precise assessments related to education and ensure that future decisions will be made more accurately by the administrators, supervisors, and teachers.

3. Grey Relational Analysis (GRA)

The Grey system theory is used to solve problems with a small sample and missing information and produce results in the case of uncertainty and incomplete information [46]. While the exact unknown information in the Grey system theory is represented by black, precisely known information is represented by white, and the incomplete information between these two endpoints is represented by grey [47,48,49].
GRA is a relational ranking, classification, and decision-making technique developed by using the Grey system theory [47]. With this method, it is aimed to determine the degree of relations between the reference series and each factor in a grey system. The relationship level between the factors is called a Grey relational degree. GRA is used to make the most appropriate selection based on various criteria between the options. It aims at the selection of the best alternatives in a multi-criteria decision-making process. It can be regarded as an optimization method. The calculation to perform the comparison and sorting between alternatives is realized in six steps: the creation of decision matrix, comparison matrix, normalization matrix, absolute value table, grey relational coefficient matrix, and grey relational degrees [50,51].
Step 1: The first operation in GRA is to create a decision matrix consisting of alternatives a and decision criteria c. Equation (2) shows the decision matrix X . X a represents each alternative, and X a c represents the criteria value of each alternative. The alternative number in the decision matrix is denoted by m , and the criteria number is denoted by n .
a = { 1 , 2 , 3 , . m } ,   c = { 1 , 2 , 3 , . n }
X a c = [ X 11 X 12 X 21 X 22 X m 1 X m 2 X 1 n X 2 n X m n ]
Step 2: Based on the values in the decision matrix, the reference series is added to the matrix, and the comparison matrix is obtained. The value of the reference series is selected based on the maximum or minimum values of the decision criteria in the alternatives, as well as the optimal criterion value, which the ideal alternative can obtain. Comparison matrix X 0 is
X 0 = [ X 01 X 02 X 11 X 12 X m 1 X m 2 X 0 n X 1 n X m n ]
Step 3: The process of reducing a datum to small intervals is called normalization. After normalization, it becomes possible to compare the series with each other. Thus, the normalization prevents the analysis from being adversely affected, depending on the discrete values [50,51,52]. In GRA, the normalization varies according to whether the criterion values in the series are maximum or minimum. If the maximum value of the series contributes positively to the decision (utility-based), Equation (4) is used. If the minimum value of the series contributes positively to the decision (cost-based), Equation (5) is used. If there is a positive contribution of a desired optimal value to the decision (optimal based), Equation (6) is used to normalize the matrix X 0 . In this case, the normalization matrix X is obtained by Equation (7)
X a c =   X a c m i n X a c maxX a c m i n X a c
X a c =   m a x X a c X a c maxX a c m i n X a c
X a c =   | X a c X 0 c | maxX a c X 0 c
X = [ X 01 X 02 X 11 X 12 X m 1 X m 2 X 0 n X 1 n X m n ]
Step 4: Equation (8) indicates the absolute subtraction of the reference series and the alternatives. The absolute value table is created by taking the absolute difference of each alternative series in X and the reference series and obtained by Equation (9). For example, the absolute value of the second alternative third criterion is calculated as Δ 23 = | X 03 X 23 | .
Δ a c =   | X 0 c   X a c |
Δ = [ Δ 01 Δ 02 Δ 11 Δ 12 Δ m 1 Δ m 2 Δ 0 n Δ 1 n Δ m n ]
Step 5: The absolute value table ( Δ ) in Equation (9) is used to calculate the grey relational coefficients indicating the degree of closeness of the decision matrix to the comparison matrix. The grey relational coefficients are calculated by Equation (12).
Δ m a x =   m a x a m a x c Δ a c
Δ m i n =   m i n a m i n c Δ a c
γ a c =   Δ m i n +   ρ Δ m a x Δ a c +   ρ Δ m a x
In Equation (12), ρ denotes the distinguishing coefficient and ρ [ 0 , 1 ] . The purpose of ρ is to expand or compress the range of coefficients. In cases where there is a large difference in data, ρ is chosen close to 0 otherwise close to 1. In the literature, it is stated that ρ does not affect the ranking and is generally taken as ρ = 0.5 [53].
Step 6: Grey relational degrees determining the relationship between the reference series and alternatives can be calculated in two ways. If the decision criteria have equal significance, Equation (13) is used to calculate the degrees. In the other case, Equation (14) is used. The value w a c denotes the weight of the decision criteria in Equation (14). The value δ a calculated by the arithmetic mean of γ a c is close to 1 indicates that the relationship with the reference series is high.
δ a =   1 n c = 1 n γ a c
δ a =   j = 1 n w a c γ a c

4. Proposed Method

Text mining is a complex process consisting of four steps (Figure 1), in which textual data is converted into structured data and analyzed by data mining techniques.
In this study, we developed a decision support system based on text mining and GRA to evaluate the written exams. Figure 2 depicts our system model. At the university level, 50 students in the same class took an examination, and the instructor evaluated the answer sheets. In this context, we analyzed the relationship level between the students’ written exams and the answer key.
In our system, we used the C# programming language for data collection, data preparation, text processing, and GRA operations. We also found word roots using Java programming language and performed data operations through the Microsoft SQL Server database management system. The database tables used in the analysis process are shown in Figure 3.

4.1. Data Collection

Students taking the exam were announced a week before the exam date. On a volunteer basis, 50 students participated in the exam in the computer laboratory. Five questions were asked to the students, and the answer sheets were received in an electronic format (as a Word document). The answer key for the first question is shown in Appendix A (a,b). In order to be able to describe the GRA-based system model used in this research to the reader as short and understandable, we created a sample table by considering five student answers and explained the GRA calculations according to this table. The answers given by the five students are presented in Appendix A (c). The answers, the scores, and the answer key were saved into the database with the application programming interface we developed.

4.2. Data Preprocessing

The efficiency of text mining technique is directly related to data preprocessing. Data preprocessing influences the health of the analysis results. It consists of three steps, which are applied, respectively: tokenization, filtering, and stemming. In the tokenization phase, which is performed by text processing functions of the software developed, the marks such as point, comma, and exclamation are extracted from the text firstly. Then, the words in the text are broken up, and each word is stored in the database. Those steps in the tokenization phase are repeated for each answer of all students and the answer key. The second phase used in data preprocessing is filtering, which is NLP-based and includes removing prepositions and conjunction words. This process is carried out by deleting the words from the database through the programming language. Some of the deleted words in filtering process are as follows in Turkish: fakat (but), ancak (only), bile (even), ile (with), çünkü (because), eğer (if), ve (and), veya (or), etc. The final phase in data preprocessing is the stemming. As a process of finding word roots, stemming is used to extract the prefixes and the suffixes that change the meaning of the words. In this study, we employed the Zemberek 0.11.1, which is the Java-based NLP library designed for performing Turkish NLP operations, to find the root of the words. Roots of the words stored in the database were identified by the Java application we developed. After removing the prefixes and the suffixes, the roots found were recorded in the database again.

4.3. Feature Extraction

Converting data to a structured form is a transformation process to apply data mining techniques to the text data. BoW, binary, and VSM methods are typically used to extract the word counts of the text, and these values are considered as features of the document. In other words, the data conversion task is the feature extraction of the document (textual data) at the same time [54,55].

4.3.1. Bag of Words (BoW)

The BoW approach is used to convert the document into a structured form. In other words, it assumes that the number of words in the text document can represent the document. In our study, we generated two BoW series (BoW1: the word frequency number of answer key; BoW2: the word frequency number of student answers) for each answer. In the BoW generation process given by Algorithm 1, BoW values of both answer key and student answers were placed in the BoW array. A value of 0 for the variable “answer” in the algorithm indicated the answer key BoW.
Algorithm 1. BoW Generation
  • string Answers[i, j];  // i: questions, j: given answers
  • string BoW[i, j, p];   // p: number of words in the answer(i, j)
  •  for(int question = 0; question < i; question++){
  •   for(int answer = 0; answer < j; answer++){
  •   string given_answer = A[i, j];
  •   string[] splitted_answer = given_answer.Split(‘ ’);
  •    for (int place = 0; place < splitted_answer.Count; place++){
  •    BoW[question, answer, place] = splitted_answer[place];
  •    }
  •  }
  • }

4.3.2. Vector Space Model (VSM)

VSM is the most common method to represent textual data numerically. In VSM, each document is represented by a vector created using the frequency count of the items. In document comparisons, the weight values of the words in documents to be compared can be determined relative to the Term Frequency (TF) or the Term Frequency-Inverse Document Frequency (TF-IDF). The TF denotes the number of terms in the document. The TF-IDF is a weighting method of the terms in the document collection [19]. It is possible to compare the multi-documents using the TF-IDF. For the term weight calculation, we used the TF-IDF method as follows:
q ( w ) = f D ( w ) l o g N f D ( w )
In Equation (15), q ( w ) denotes the weight q of the term w . The value of q ( w ) is directly proportional to the number of w in document d , f d ( w ) , and inversely proportional the number of w in other documents D . N is the number of all documents, and f D ( w ) indicates the number of documents in which the related term is located. This approach, which is used in text mining applications effectively, is not a measure of similarity; however, it indicates the importance of a term in documents. Document similarities can be determined using the similarity measures such as Cosines and Euclidean [55,56].
Besides the advantages of simple calculation and allowing a comparison of multiple documents, there are some limitations related to TF-IDF. Insufficient information, defined as the sparsity in short texts, is one of those limitations. Although the TF-IDF provides healthy results in the analysis of long text data, it may not give effective results in the analysis of short text data. On the other hand, the evaluation of a text according to the weight gradations of the terms makes the term with a higher weight degree more important than others. Therefore, the TF-IDF indicates the importance of the frequency-based term in the text [20,56]. According to [21], although this method can be used to separate terms, it is not sufficient to distinguish between documents. Because of the shortcomings of the TF-IDF, it can be problematic when we consider short texts where terms with a low weight scale and terms with higher weight have an equal prefix. Therefore, the use of TF-IDF may not be an appropriate choice when written exams consist of short texts. Furthermore, exam evaluations should be measured with the similarity to the answer key to be referenced, not to other answers. In other words, the evaluation of exam questions should be done by the similarity to the answer key of all answers, and not be weighted according to other answers. In this context, it is suitable to analyze exam answers represented by the VSM with GRA, which is used in case of insufficient information so that written exam evaluations can be carried out by text mining.

4.4. Analysis

The structural text converted into a digital vector is analyzed in detail by employing clustering, classification, and summarization. Moreover, statistical methods and machine learning techniques in data mining tasks can also be used as analysis methods.

5. Experimental Results

5.1. BoW Findings

The number of frequencies of the word roots in the answer key was determined, and the bag of words, B o W 1 , representing the answer key was created for each question. Thus, the bag of words for n questions in the answer key can be defined as
B o W 1 = { B o W 11 , B o W 12 , , B o W 1 n }
In the same way, B o W 2 , was created by representing the answers of m students for the questions and can be defined as
B o W 2 = { B o W 211 , B o W 212 , , B o W 2 m n }
For example, B o W 11 , which is the first question answer key bag of words and B o W 211 , which is the first question first student bag of words are shown in Figure 4. Words are in Turkish. (For all students, see Appendix B (a)).

5.2. VSM Findings

The roots of words and numbers that match the answer key need to be specified for the analysis. The VSM table showing word distribution based on the answers of the first question responded by the five students is presented in Table 4. The VSM table, created by the answers of all students to the first question, is given in Appendix B (b).

5.3. Grey Relational Degree Findings

As illustrated in Figure 5, there is a two-stage decision-making process in our analysis. The first stage is to identify the level of relationship between the answers responded by the student and the words in the answer key for each question. The level of relationship between B o W 1 and B o W 2 was determined based on the result of the first step using VSM prepared for each question.
In the first stage, VSM, used as a decision matrix in the GRA, has a reference series which is added according to the utility-based values of each criterion among alternatives. Normalization was carried out by applying Equations (4) and (6). If the number of frequencies of the words in B o W 1 is less than B o W 2 , Equation (4) is used. If B o W 1 is greater than or equal to B o W 2 , Equation (6) is used for normalization. The absolute differences of the series obtained from the normalization process of the VSM with reference series were calculated, and then the Grey relational coefficients were calculated. ρ = 0.5 was chosen in accordance with the literature in the calculation of the coefficients. The decision matrix formed by the first step is presented in Appendix C (a), the normalization matrix is given in Appendix C (b), the absolute differences are shown in Appendix C (c), and the Grey relational coefficients are given in Appendix C (d). Equation (13) is used for the calculation of Grey relational degrees with the assumption that the words in each question have equal importance. The Grey relational degrees showing the answer key relational level of the answers responded by the students for each question are presented in Table 5.
Table 5 shows the relationship between B o W 1 and B o W 2 for the questions answered by each student and requires a second decision-making process. It is used to rank the exam paper of each student in relation to the answer key. In this context, it was used as a decision matrix for the second stage GRA. The reference series was created by maximum values within the alternative degree values for each question. Since the higher degree of alternatives shows the maximum scoring value, based on the utility-based approach, normalization is carried out by Equation (4), and the absolute differences table is created. The Grey relational coefficients are calculated with Equation (12) by ρ = 0.5 . The Grey relational coefficients of the second stage decision-making process were presented in Appendix D (a). Since each question has equal weight, the Grey relational degrees presented in Table 6 were calculated by Equation (13). Table 6 shows the relevance level of student answers to the answer key. For example, according to the table, the answer closest to the answer key for the first question belongs to student-1 and student-3. In similar, student-5 has the best answer to question 2. If all the questions are considered, the highest score should be given to the student-5.
The analysis results that contain the instructor’s scoring and the Grey relational degrees of the students’ answers are presented in Table 7. (See Appendix D (b) for all students).

6. Discussion and Suggestions

In this study, we developed a decision support system using text mining and Grey system theory. The exam evaluations, considered as multi-criteria decision-making problems, are based on the student answers and word similarities in the answer key. No study has been found on the use of Grey system theory (which is rarely studied in text mining literature) for the evaluation of the written exams. In this regard, our study is a pioneer in the field of text mining and educational data mining.
It should be noted that the study does not provide a scoring system based on the results. However, it realizes an evaluation based on the rating. In this respect, it is a decision support system that helps the decision-maker to evaluate the student exam results. When the results, presented in Table 7, are compared, it is seen that the analysis results made with the classical method are close to each other. The decision support system developed in this way contributes to the assessment of distance learning exams, both in the classical written exams and online, and facilitates the control of exam scores.
Since the study is based on the similarity of the word roots, the preprocessing part of the text mining analysis becomes an important task. At this stage, we observed that the Zemberek library has some inadequacies, such as the shortening of some vocabulary roots. However, since the stemming process was applied to both the answer key and the answer sheets by using the same way, this inadequacy was assumed not to affect the analysis results. Furthermore, since the study is based on contextual text mining, semantic changes in word roots were neglected in the stemming process.
We have observed that there are some limitations related to the root-finding processes involved in the preprocessing. We have accepted that the stemming is applied to both the answer key and student answers in the same way and that the analysis is not affected since the study is contextual. However, in lexical analysis, it cannot be assumed that the situation will not have an analytical effect. In this context, since the development of NLP tools will gain importance, more effective results can be obtained for the lexical text mining studies.
In the future study, the renovation of the research using the lexical text mining method and realizing semantic analyses using artificial intelligence will contribute positively to the literature. However, the inadequacies in NLP implementations in the used language limit the semantic approaches. Particularly phrases, sentences, expressions composed of more than one word, and the shortcomings in the idioms make the lexical text mining analysis difficult. Turkish NLP libraries, which are currently being developed to increase the efficiency and capacity of the libraries such as Zemberek, ITU Turkish NLP, and Tspell, will have a vital role for text mining. Besides, the fact that the data set used in the study is in Turkish does not mean that the proposed model cannot be applied in other languages. On the contrary, thanks to the powerful processing capabilities of English NLP libraries, our model can achieve more successful results in English data sets.

7. Conclusions

In this paper, we presented a new novel approach to written exam evaluation. We developed a decision support system employing text mining and GRA to help decision-makers to evaluate the written exams. We demonstrated that how GRA could be used in two-stage decision systems. We carried out the analysis with the assumption that the word and the problem have equal weight. It has been determined in the application process that even if the factors affecting the decision have different effect weights, it can be performed with GRA easily. Therefore, it will be an exciting research topic to analyze the decision-making process that has differently weighted criteria by using GRA.
The assessment of written exam questions should be based on a referenced answer key. The words used to answer the question should be used to determine word similarity in the answer key prepared for each question. Thus, the evaluation of the written exams is a multi-criteria decision process. Besides, BoW, TF, and TF-IDF methods used in contextual text mining studies can be insufficient. Therefore, this study reveals that a rating can be made based on the word similarities in the written exams. In this context, by employing two-stage GRA, we have demonstrated that Grey relational degrees obtained by word similarities can be used to determine the similarity between the exam papers and the answer key.
In our study, text data in the exam papers of 50 students were structured by text mining methods, and first-order GRA was applied to the decision matrix created by VSM. The ranking of the student exam papers was carried out by applying the second-order GRA so that this method was applied to cover all questions. The decision support system using the GRA similarity values determines the range in which the student assessment score should be. Instead of evaluating as successful or unsuccessful, the success level of a student is determined relationally based on the success level of the other students by taking advantage of the GRA. Consequently, we demonstrated that the exam papers were evaluated using grey relational degrees. Furthermore, we compared the ranking using GRA with the instructor’s ranking, and we observed that there were no significant differences between these two assessments.

Author Contributions

The authors contributed equally to this work.

Funding

This research received no external funding.

Acknowledgments

The authors are grateful to the Editor of the Journal and anonymous referees for their detailed and valuable comments.

Conflicts of Interest

The authors declare that they have no conflict of interest regarding the publication of this research article.

Appendix A

(a) The questions prepared by the instructor (our application software)
Symmetry 11 01426 i001
(b) The answer key of the first question
Symmetry 11 01426 i002
(c) Student answers for the first question
Symmetry 11 01426 i003

Appendix B

(a)  B o W 2 for the first question * (* Student whose B o W 2 is null set were excluded from the table. S: Student, Q: Question.)
SQWords (in both Turkish and English)
11zihin_1& (mental_1&)
31sosyal_1& (social_1&)
61organizma_1&döl_1&başla_1& (organism_1&seed_1&start_1&)
71başla_1&duygu_1& (start_1&emotion_1&)
101döl_1&başla_1&sosyal_1&yön_1& (seed_1&start_1&social_1&direction_1&)
131organizma_1&döl_1&başla_1&beden_1&zihin_1&sosyal_1&yön_1&geçir_1&sürekli_1&düz_1&değ_1&(organism_1&seed_1&start_1&body_1&mental_1&social_1&direction_1&pass_1&continuous_1&straight_1&touch_1&)
141organizma_1&döl_1&başla_1&beden_1&zihin_1&duygu_1&sosyal_1&yön_1&geçir_1&sürekli_1&düz_2&ilerle_1&(organism_1&seed_1&start_1&body_1&mental_1&emotion_1&social_1&direction_1&pass_1&continuous_1&straight_2&move_1&)
151döl_1&başla_1&sürekli_1& (seed_1&start_1&continuous_1&)
181zihin_1& (mental_1&)
191sosyal_1&yön_2& (social_1&direction_2&)
201beden_1& (body_1&)
211sosyal_1& (social_1&)
241başla_1&zihin_1&sosyal_1& (start_1&mental_1&social_1&)
251döl_1&başla_1& (seed_1&start_1&)
261sürekli_1&düz_1&ilerle_1& (continuous_1&straight_1&move_1&)
271organizma_1&döl_1&başla_1&sosyal_1&yön_1&geçir_1&sürekli_1&düz_1&değ_1&(organism_1&seed_1&start_1&social_1&direction_1&pass_1&continuous_1&straight_1&touch_1&
291geçir_1& (pass_1&)
311organizma_1&döl_1&başla_1&zihin_1&sosyal_1&yön_1&(organism_1&seed_1&start_1&mental_1&sosyal_1&yön_1&)
321beden_1&zihin_1&sosyal_1&yön_1&geçir_1&sürekli_1&düz_1&değ_1&(body_1&mental_1&social_1&direction_1&pass_1&continuous_1&straight_1&touch_1&)
331döl_1&başla_1&düz_1&ilerle_1& (seed_1&start_1& straight_1&move_1&)
341başla_1&duygu_1&sosyal_1& (start_1&emotion_1&social_1&)
371başla_1& (start_1&)
391başla_1& (start_1&)
401sosyal_1&yön_1& (social_1&direction_1&)
431sürekli_1& (continuous_1&)
441başla_1& (start_1&)
451sürekli_1& (continuous_1&)
461başla_2&duygu_1&sosyal_1& (start_2&emotion_1&social_1&)
471duygu_1&sosyal_1&yön_1& (emotion_1&social_1&direction_1&)
481döl_1&başla_1&sürekli_1&değ_1& (seed_1&start_1&continuous_1&touch_1&)
491sosyal_1&yön_2& (social_1& direction_2&)
501başla_1&duygu_1&sosyal_1&yön_1& (start_1&emotion_1&social_1&direction_1&)
(b)  V S M 1 table (VSM for the first question)
Words (in both Turkish and English)
QSOrganizma
(organism)
Döl
(seed)
Başla
(start)
Beden
(body)
Zihin
(mental)
Duygu
(emotion)
Sosyal
(social)
Yön
(direction)
Geçir
(pass)
Sürekli
(continuous)
Düz
(straight)
Değ
(touch)
Ilerle
(move)
10 *1111111111111
110000100000000
120000000000000
130000001000000
140000000000000
150000000000000
161110000000000
170010010000000
180000000000000
190000000000000
1100110001100000
1110000000000000
1120000000000000
1131111101111110
1141111111112201
1150110000001000
1160000000000000
1170000000000000
1180000100000000
1190000001200000
1200001000000000
1210000001000000
1220000000000000
1230000000000000
1240010101000000
1250110000000000
1260000000001101
1271110001111110
1280000000000000
1290000000010000
1300000000000000
1311110101100000
1320001101111110
1330110000000101
1340010011000000
1350000000000000
1360000000000000
1370010000000000
1380000000000000
1390010000000000
1400000001100000
1410000000000000
1420000000000000
1430000000001000
1440010000000000
1450000000001000
1460020011000000
1470000011100000
1480110000001010
1490000001200000
1500010011100000

Appendix C

(a) First order GRA decision matrix (first question 5 students)
QuestionStudentWords
1Answer Key1111111111111
1Reference1121111212211
110000100000000
120000000000000
130000001000000
140000000000000
150000000000000
(b) First order GRA normalization matrix (first question 5 students)
QuestionStudentWords
1Reference1111111111111
110010100101100
120010000101100
130010001101100
140010000101100
150010000101100
(c) First order GRA absolute differences (first question 5 students)
QuestionStudentWords
1Reference0000000000000
110010100101100
120010000101100
130010001101100
140010000101100
150010000101100
(d) First order GRA coefficients
QuestionStudentWords
1Reference0000000000000
11110.33310.333110.33310.3330.33311
12110.33311110.33310.3330.33311
13110.3331110.3330.33310.3330.33311
14110.33311110.33310.3330.33311
15110.33311110.33310.3330.33311

Appendix D

(a) Second order GRA coefficients and degrees (S: Student, Q: Question.)
SQ-1Q-2Q-3Q-4Q-5Degrees
10.41170.41890.52410.38740.34220.4169
20.36840.63050.45170.3670.38590.4407
30.41170.44060.58630.41180.36920.4439
40.36840.60720.36920.43670.39130.4346
50.368410.42360.40110.35230.5091
60.41170.40060.47450.36970.37780.4069
70.36840.45210.50290.37410.40450.4204
80.36840.7090.42640.40370.38220.4579
90.36840.41060.60040.39640.3770.4306
100.36840.727510.39230.41330.5803
110.36840.43620.42360.36950.45920.4114
120.36840.45450.61590.38460.34880.4344
130.53850.65350.34460.40220.480.4838
1410.6210.5699110.8382
150.33330.74040.34860.38640.39130.44
160.36840.47570.61590.36410.40590.446
170.36840.48120.50290.39910.35360.421
180.41170.52240.51320.38540.34880.4363
190.41170.52240.52210.420.35050.4453
200.41170.38120.44020.39660.38180.4023
210.41170.45240.40080.41230.33330.4021
220.36840.42950.46680.36230.37110.3996
230.36840.42470.48910.39780.3810.4122
240.41170.42980.45310.41230.3790.4172
250.36840.60780.48130.4430.44330.4688
260.33330.52240.35350.41230.39870.404
270.41170.73120.33330.43930.35230.4536
280.36840.38120.35660.37410.38630.3733
290.41170.59490.42520.40840.40450.4489
300.36840.44230.56270.41230.3790.4329
310.46660.39540.38670.33330.43740.4039
320.46660.38260.35130.40510.36730.3946
330.36840.6160.60040.39440.39130.4741
340.41170.64440.45310.38560.43480.4659
350.36840.42780.57280.37920.53890.4574
360.36840.43290.38050.39680.39130.394
370.33330.46830.54040.39640.39130.4259
380.36840.6160.36270.39210.39130.4261
390.33330.52240.53580.44920.40680.4495
400.36840.41120.61590.41230.39130.4398
410.36840.63050.53780.41960.38180.4676
420.36840.46470.43190.42390.39130.416
430.33330.46470.43820.42270.39130.41
440.33330.46470.48330.37310.39130.4091
450.33330.48610.39280.41230.4220.4093
460.46660.39740.49290.40130.39130.4299
470.41170.52910.52410.41230.3550.4464
480.36840.5490.44670.41230.42450.4402
490.41170.38120.39520.42560.39870.4025
500.36840.33330.40970.40130.39130.3808
(b) Comparison table of evaluator scores and GRA results (S: Student, Q: Question.)
Evaluator ScoresGRA Results
SQ-1Q-2Q-3Q-4Q-5TotalRankDegreesRankSuggestion
15850220330.416933do not change
20101015035250.440718increase the score
31512100037240.443917increase the score
4501515035260.434623increase the score
50201213206580.50913increase the score
610810151053130.406940decrease the score
71515201526770.420431decrease the score
80010151035270.45799increase the score
9102000232290.430626increase the score
10202020202010010.58032decrease the score
110152015207060.411436decrease the score
120150101550140.434424decrease the score
1315202015158530.48384decrease the score
1420201215208720.83821increase the score
151010551040160.4420decrease the score
16015200540170.44615increase the score
17050005480.42130increase the score
1800515525320.436322increase the score
1910000010450.445316increase the score
205805018370.402344decrease the score
21152200239210.402145decrease the score
220155101040180.399646decrease the score
23080008460.412235increase the score
240850013420.417232increase the score
2510101515156590.46886increase the score
261001001030310.40441decrease the score
2720202015108540.453611decrease the score
28108510033280.373350decrease the score
2901808531300.448913increase the score
30010100020340.432925increase the score
3118102020158350.403942decrease the score
321815105250150.394647decrease the score
3310151020055110.47415increase the score
3410410201054120.46598increase the score
351000101040190.457410increase the score
3605100015400.39448decrease the score
3710200012430.425929increase the score
3801800018380.426128increase the score
3900105520350.449512increase the score
40050005490.439821increase the score
4101200012440.46767increase the score
42060006470.41634increase the score
4306140020360.4137decrease the score
44050005500.409139increase the score
450102001040200.409338decrease the score
46101855038220.429927decrease the score
470850215410.446414increase the score
4820151501060100.440219decrease the score
4908100018390.402543decrease the score
50108155038230.380849decrease the score

References

  1. Turgut, M.F.; Baykul, Y. Eğitimde Ölçme ve Değerlendirme; Pegem Akademi: Ankara, Turkey, 2012. [Google Scholar]
  2. Dikli, S. Assessment at a distance: Traditional vs. alternative assessments. Turk. Online J. Educ. Technol. 2003, 2, 13–19. [Google Scholar]
  3. Yılmaz, A. Ölçme ve Değerlendirme; Pegem Akademi: Ankara, Turkey, 2009. [Google Scholar]
  4. Miller, M.D.; Linn, R.L.; Gronlund, N.E. Measurement and Assessment in Teaching; Pearson: Boston, MA, USA, 2013. [Google Scholar]
  5. Fraenkel, J.R.; Wallen, N.E. How to Design and Evaluate Research in Education; McGraw-Hill: New York, NY, USA, 2005. [Google Scholar]
  6. Kurgan, L.A.; Musilek, P. A survey of knowledge discovery and data mining process models. Knowl. Eng. Rev. 2006, 21, 1–24. [Google Scholar] [CrossRef]
  7. Kahraman, C.; Onar, S.C. Intelligent Techniques in Engineering Management, Theory and Applications; Springer: Cham, Switzerland, 2015. [Google Scholar]
  8. Nenonen, N. Analysing factors related to slipping, stumbling, and falling accidents at work: Application of data mining methods to Finnish occupational accidents and diseases statistics database. Appl. Ergon. 2013, 44, 215–224. [Google Scholar] [CrossRef] [PubMed]
  9. Gurbuz, F.; Ozbakir, L.; Yapici, H. Classification rule discovery for the aviation incidents resulted in fatality. Knowl. Based Syst. 2009, 22, 622–632. [Google Scholar] [CrossRef]
  10. Yuksel, M.E. Agent-based evacuation modeling with multiple exits using NeuroEvolution of Augmenting Topologies. Adv. Eng. Inform. 2018, 35, 30–55. [Google Scholar] [CrossRef]
  11. Ji, H.; Park, K.; Jo, J.; Lim, H. Mining students activities from a computer supported collaborative learning system based on peer to peer network. Peer Peer Netw. Appl. 2016, 9, 465–476. [Google Scholar] [CrossRef]
  12. Sailesh, S.B.; Lu, K.J.; Aali, M.A. Profiling Students on Their Course-Taking Patterns in Higher Educational Institutions (HEIs). In Proceedings of the International Conference on Information Science, Kochi, India, 12–13 August 2016. [Google Scholar]
  13. Scheuer, O.; McLaren, B.M. Educational Data Mining, In the Encyclopedia of the Sciences of Learning; Springer: Boston, MC, USA, 2011. [Google Scholar]
  14. Tsai, C.Y.; Tsai, M.H. A Dynamic Web Service Based Data Mining Process System. In Proceedings of the 5th International Conference on Computer and Information Technology, Shanghai, China, 21–23 September 2005. [Google Scholar]
  15. Rygielski, C.; Wang, J.C.; Yen, D.C. Data mining techniques for customer relationship management. Technol. Soc. 2002, 24, 483–502. [Google Scholar] [CrossRef]
  16. Ramakrishnan, R.; Gehrke, J. Database Management Systems; McGraw-Hill Education: New York, NY, USA, 2003. [Google Scholar]
  17. Dunham, M.H. Data Mining: Introductory and Advanced Topics; Prentice-Hall: New Jersey River, NJ, USA, 2003. [Google Scholar]
  18. Harish, B.S.; Guru, D.S.; Manjunath, S. Representation and classification of text documents: A brief review. Int. J. Comput. Appl. Spec. Issue Recent Trends Image Process. Pattern Recognit. 2010, 2, 110–119. [Google Scholar]
  19. Allahyari, M.; Pouriyeh, S.; Assefi, M.; Safaei, S.; Trippe, E.D.; Gutierrez, J.B.; Kochut, K. A brief survey of text mining: Classification, clustering and extraction techniques. arXiv 2017, arXiv:1707.02919. [Google Scholar]
  20. Pan, W.; Zhong, E.; Yang, Q. Transfer learning for text mining. In Mining Text Data; Aggarwal, C., Zhai, C., Eds.; Springer: Boston, MA, USA, 2012; pp. 223–257. [Google Scholar]
  21. Korde, V.; Mahender, C.N. Text classification and classifiers: A survey. Int. J. Artif. Intell. Appl. 2012, 3, 85–99. [Google Scholar]
  22. Xia, T.; Chai, Y. An improvement to TF-IDF: Term distribution based term weight algorithm. J. Softw. 2011, 6, 413–420. [Google Scholar] [CrossRef]
  23. Tasci, S.; Gungor, T. Comparison of text feature selection policies and using an adaptive framework. Expert Syst. Appl. 2013, 40, 4871–4886. [Google Scholar] [CrossRef]
  24. Chen, K.; Zhang, Z.; Long, J.; Zhang, H. Turning from TF-IDF to TF-IGM for term weighting in text classification. Expert Syst. Appl. 2016, 66, 245–260. [Google Scholar] [CrossRef]
  25. Cakan, M. Eğitimde Ölçme ve Değerlendirme; Pegem Akademi: Ankara, Turkey, 2008. [Google Scholar]
  26. Atilgan, H.; Kan, A.; Dogan, N. Eğitimde Ölçme ve Değerlendirme; Ani Yayincilik: Ankara, Turkey, 2009. [Google Scholar]
  27. Stiggins, R. Assessment Manifesto: A Call for the Development of Balanced Assessment Systems; ETS Assessment Training Institute: Portland, OR, USA, 2008. [Google Scholar]
  28. Gecit, Y. Eğitimde Ölçme ve Değerlendirme; Nobel Yayincilik: Ankara, Turkey, 2012. [Google Scholar]
  29. Tekindal, S. Okullarda Ölçme ve Değerlendirme Yöntemleri; Nobel Yayincilik: Ankara, Turkey, 2014. [Google Scholar]
  30. Ozcelik, D.A. Okullarda Ölçme ve Değerlendirme Öğretmen El Kitabı; Pegem Akademi: Ankara, Turkey, 2010. [Google Scholar]
  31. Han, J.; Kamber, M.; Pei, J. Data Mining Concepts and Techniques; Morgan Kaufmann: Waltham, MC, USA, 2012. [Google Scholar]
  32. Witten, I.H.; Frank, E.; Hall, M.A.; Pal, C.J. Data Mining: Practical Machine Learning Tools and Techniques; Morgan Kaufmann: Waltham, MC, USA, 2017. [Google Scholar]
  33. Ozkan, Y. Veri Madenciliği Yöntemleri; Papatya Bilim: Istanbul, Turkey, 2016. [Google Scholar]
  34. Zheng, Y. Trajectory data mining: An overview. ACM Trans. Intell. Syst. Technol. 2015, 6, 29. [Google Scholar] [CrossRef]
  35. Berry, M.W. Survey of Text Mining: Clustering, Classification, and Retrieval; Springer: New York, NY, USA, 2004. [Google Scholar]
  36. Hebrail, G.; Marsais, J. Experiments of textual data analysis at Electricite de France. In New Approaches in Classification and Data Analysis. Studies in Classification, Data Analysis, and Knowledge Organization; Diday, E., Lechevallier, Y., Schader, M., Bertrand, P., Burtschy, B., Eds.; Springer: Heidelberg, Germany, 1994; pp. 569–576. [Google Scholar]
  37. Feldman, R.; Dagan, I. Knowledge Discovery in Textual Databases. In Proceedings of the First International Conference on Knowledge Discovery and Data Mining, Montreal, QC, Canada, 20–21 August 1995. [Google Scholar]
  38. Weiss, S.M.; Indurkhya, N.; Zhang, T. Fundamentals of Predictive Text Mining; Springer: London, UK, 2015. [Google Scholar]
  39. Beliga, S.; Mestrovic, A.; Ipsic, M.S. An overview of graph based keyword extraction methods and approaches. J. Inf. Organ. Sci. 2015, 39, 1–20. [Google Scholar]
  40. Zhang, W.; Yoshida, T.; Tang, X. Text classification based on multi-word with support vector machine. Knowl. Based Syst. 2008, 21, 879–886. [Google Scholar] [CrossRef]
  41. Ravi, K.; Ravi, R. A survey on opinion mining and sentiment analysis: Tasks, approaches and applications. Knowl. Based Syst. 2015, 89, 14–46. [Google Scholar] [CrossRef]
  42. Montoyo, A.; Barco, P.M.; Balahur, A. Subjectivity and sentiment analysis: An overview of the current state of the area and envisaged developments. Decis. Support Syst. 2012, 53, 675–679. [Google Scholar] [CrossRef]
  43. Aggarwal, C.C.; Zhai, C.X. Mining Test Data; Springer: New York, NY, USA, 2012. [Google Scholar]
  44. Romero, C.; Ventura, S.; Pechenizkiy, M.; Baker, R.S. Handbook of Educational Data Mining; CRC Press, Taylor & Francis Group: Boca Raton, FL, USA, 2010. [Google Scholar]
  45. Baker, R.S. Educational data mining: An advance for intelligent systems in education. IEEE Intell. Syst. 2014, 29, 78–82. [Google Scholar] [CrossRef]
  46. Liu, S.; Lin, Y. Grey Information Theory and Practical Applications; Springer: New York, NY, USA, 2006. [Google Scholar]
  47. Deng, J.L. Control problems of grey systems. Syst. Control Lett. 1982, 1, 288–294. [Google Scholar]
  48. Liu, S.; Forrest, J.; Yang, Y. A brief introduction to grey systems theory. Grey Syst. Theory Appl. 2012, 2, 89–104. [Google Scholar] [CrossRef]
  49. Yildirim, B.F. Gri ilişkisel analiz. In Çok Kriterli Karar Verme Yöntemleri; Yildirim, B.F., Onder, E., Eds.; Dora Yayincilik: Bursa, Turkey, 2015; pp. 229–236. [Google Scholar]
  50. Fidan, H. Association rules-based grey relational approach for e-commerce recommender system. In Machine Learning Techniques for Improved Business Analytics; Kumar, G.D., Ed.; IGI Global: Hershey, PA, USA, 2019; pp. 64–93. [Google Scholar]
  51. Fidan, H. Grey relational classification of consumers’ textual evaluations in e-commerce. J. Theor. Appl. Electron. Commer. Res. 2020, 15, 48–65. [Google Scholar] [CrossRef]
  52. Fidan, H. Grey relational clustering analysis of e-commerce customers loyalty. Online Acad. J. Inf. Technol. 2018, 9, 163–182. [Google Scholar]
  53. Ertugrul, I.; Oztas, T.; Ozcil, A.; Oztas, G.Z. Grey relational analysis approach in academic performance comparison of university: A case study of Turkish universities. Eur. Sci. J. 2016, 128–139. [Google Scholar]
  54. Tsai, C.H.; Chang, C.L.; Chen, L. Applying grey relational analysis to the vendor evaluation model. Int. J. Comput. Internet Manag. 2003, 11, 45–53. [Google Scholar]
  55. Turney, P.D.; Pantel, P. From frequency to meaning: Vector space models of semantics. J. Artif. Intell. Res. 2010, 37, 141–188. [Google Scholar] [CrossRef]
  56. Ropero, J.; Gomez, A.; Carrasco, A.; Leon, C.; Luque, J. Term weighting for information retrieval using fuzzy logic. In Fuzzy Logic-Algorithms, Techniques and Implementations; Elmer, P.D., Ed.; Intech: London, UK, 2012; pp. 173–192. [Google Scholar]
Figure 1. Text mining process.
Figure 1. Text mining process.
Symmetry 11 01426 g001
Figure 2. System model. GRA, Grey Relational Analysis; BoW, Bag of Words; VSM, Vector Space Model.
Figure 2. System model. GRA, Grey Relational Analysis; BoW, Bag of Words; VSM, Vector Space Model.
Symmetry 11 01426 g002
Figure 3. Database tables.
Figure 3. Database tables.
Symmetry 11 01426 g003
Figure 4. BoW1 and BoW2 for first question first student.
Figure 4. BoW1 and BoW2 for first question first student.
Symmetry 11 01426 g004
Figure 5. Two-stage decision-making process.
Figure 5. Two-stage decision-making process.
Symmetry 11 01426 g005
Table 1. The advantages and disadvantages of oral exams.
Table 1. The advantages and disadvantages of oral exams.
AdvantagesDisadvantages
  • The most appropriate method for measuring verbal expression skills.
  • The preparation of the questions is easy and takes less time.
  • It allows observing how the student has reached an answer.
  • When the number of students is very small, their knowledge level and academic trust can be determined in more detail.
  • Students are not likely to copy in the exam.
  • Instant feedback is received.
  • There is no standardization.
  • Factors such as personality traits, momentary emotions, speech effectiveness, and oral expression skills are mixed into the scoring process and affect the performance of the exam.
  • Since the application of the exam in crowded classes is laborious and time-consuming, very few questions are asked. This situation significantly reduces the content validity. Besides, it is compulsory to find different questions for students.
  • There is a continuous interaction between the student and the teacher, which may affect the qualifications of the exam. The teacher’s approach to the student may affect the student’s performance.
  • Since the scoring is done with a general impression, its objectivity is low.
Table 2. The advantages and disadvantages of multiple-choice exams.
Table 2. The advantages and disadvantages of multiple-choice exams.
AdvantagesDisadvantages
  • It is easy to apply and score.
  • All kinds of information can be measured.
  • To answer the question takes a short time.
  • It is suitable to be used in exams for groups with high participation.
  • It is easily applicable at all levels and stages of education.
  • Reliability and content validity are high since it is possible to ask many questions in the exam.
  • When the answers are not marked on the exam paper, the exam paper can be used repeatedly.
  • The exam results are objective since they do not vary from evaluator to evaluator.
  • It provides a variety of statistical applications. Statistical procedures can be carried out with data obtained from the exam.
  • There is a chance success.
  • It does not improve the ability of expression.
  • Most of the test time is spent on reading choices and finding the right answer.
  • It measures knowledge and remembering and is limited in measuring information that is at the level of synthesis and evaluation.
  • The forming, organizing, and writing of the questions requires expertise and experience.
  • It is not used much in the measurement of advanced behaviors.
Table 3. The advantages and disadvantages of written exams.
Table 3. The advantages and disadvantages of written exams.
AdvantagesDisadvantages
  • Since they are easy to prepare, they are useful tools.
  • Well-known exam types by the teachers.
  • Suitable for measuring high-level behaviors in the knowledge, comprehension, analysis, synthesis, application, and evaluation stages of cognitive learning.
  • Suitable for measuring and developing behaviors that cannot be easily measured by other tools such as creative thinking, critical thinking.
  • The most appropriate tools to measure the skills of using the language in writing.
  • Improves the written expressions of the students.
  • Allows the students to make more reasonings.
  • Students have the independence of answering.
  • Exam stress is less.
  • Lack of luck increases reliability.
  • Printing and copying are easy.
  • They are subjective measurement tools. The number of questions is small. Thus, reliability and content validity are low.
  • Scoring validity is low and open to evaluator bias. Different evaluators may give different grades to the same exam paper.
  • Suitable for flexible answers. Validity and reliability may be reduced if the behaviors such as expression power, composition ability, writing beauty, and paper layout are involved in the scoring.
  • Application and scoring are tiring and time-consuming.
  • Since there are many time and energy that students spend writing their answers, it can be thought that the time and energy allocated to thinking, analyzing, evaluating, and organizing are limited.
Table 4. VSM table for the first question.
Table 4. VSM table for the first question.
QuestionStudentWords (in Both Turkish and English)
Zihin (Mental)Sosyal (Social)Duygu (Emotion)
Answer key10111
11100
12000
13010
14000
15000
Table 5. First order GRA degrees for the five students.
Table 5. First order GRA degrees for the five students.
StudentQuestion-1Question-2Question-3Question-4Question-5
10.74360.89210.73570.83170.9022
20.79490.7890.75320.84230.8903
30.74360.87690.72410.82040.8945
40.79490.79680.78150.81020.889
50.79490.71360.76160.82520.8992
Table 6. Second order GRA degrees for the five students.
Table 6. Second order GRA degrees for the five students.
StudentQuestion-1Question-2Question-3Question-4Question-5Degrees
10.41170.41890.52410.38740.34220.4169
20.36840.63050.45170.3670.38590.4407
30.41170.44060.58630.41180.36920.4439
40.36840.60720.36920.43670.39130.4346
50.368410.42360.40110.35230.5091
Table 7. Comparison of evaluator scores and GRA degrees for the five students.
Table 7. Comparison of evaluator scores and GRA degrees for the five students.
Evaluator ScoresAnalysis Results
StudentQuestion-1Question-2Question-3Question-4Question-5TotalGRA Degrees
158502200.4169
201010150350.4407
315121000370.4439
45015150350.4346
5020121320650.5091

Share and Cite

MDPI and ACS Style

Yuksel, M.E.; Fidan, H. A Decision Support System Using Text Mining Based Grey Relational Method for the Evaluation of Written Exams. Symmetry 2019, 11, 1426. https://0-doi-org.brum.beds.ac.uk/10.3390/sym11111426

AMA Style

Yuksel ME, Fidan H. A Decision Support System Using Text Mining Based Grey Relational Method for the Evaluation of Written Exams. Symmetry. 2019; 11(11):1426. https://0-doi-org.brum.beds.ac.uk/10.3390/sym11111426

Chicago/Turabian Style

Yuksel, Mehmet Erkan, and Huseyin Fidan. 2019. "A Decision Support System Using Text Mining Based Grey Relational Method for the Evaluation of Written Exams" Symmetry 11, no. 11: 1426. https://0-doi-org.brum.beds.ac.uk/10.3390/sym11111426

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop