Next Article in Journal
Polygenic Adaptation in a Population of Finite Size
Next Article in Special Issue
Exploring the Possibility of a Recovery of Physics Process Properties from a Neural Network Model
Previous Article in Journal
Project Management Monitoring Based on Expected Duration Entropy
Previous Article in Special Issue
Optic Disc Segmentation Using Attention-Based U-Net and the Improved Cross-Entropy Convolutional Neural Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

From Knowledge Transmission to Knowledge Construction: A Step towards Human-Like Active Learning

1
Department of Information Systems and Technologies, Samara National Research University, Moskovskoe Shosse 34, 443086 Samara, Russia
2
Division of Electronics, Ruđer Bošković Institute, Bijenička cesta 54, 10000 Zagreb, Croatia
*
Author to whom correspondence should be addressed.
Submission received: 22 July 2020 / Revised: 15 August 2020 / Accepted: 16 August 2020 / Published: 18 August 2020
(This article belongs to the Special Issue Human-Centric AI: The Symbiosis of Human and Artificial Intelligence)

Abstract

:
Machines usually employ a guess-and-check strategy to analyze data: they take the data, make a guess, check the answer, adjust it with regard to the correct one if necessary, and try again on a new data set. An active learning environment guarantees better performance while training on less, but carefully chosen, data which reduces the costs of both annotating and analyzing large data sets. This issue becomes even more critical for deep learning applications. Human-like active learning integrates a variety of strategies and instructional models chosen by a teacher to contribute to learners’ knowledge, while machine active learning strategies lack versatile tools for shifting the focus of instruction away from knowledge transmission to learners’ knowledge construction. We approach this gap by considering an active learning environment in an educational setting. We propose a new strategy that measures the information capacity of data using the information function from the four-parameter logistic item response theory (4PL IRT). We compared the proposed strategy with the most common active learning strategies—Least Confidence and Entropy Sampling. The results of computational experiments showed that the Information Capacity strategy shares similar behavior but provides a more flexible framework for building transparent knowledge models in deep learning.

1. Introduction

The passive learning technique normally requires an enormous amount of labeled data that has to provide the correct answers (see Figure 1). An active learning environment guarantees better performance while training on less, but carefully chosen, data which reduces the costs of both annotating and analyzing large data sets [1,2,3,4,5,6,7,8,9,10]. In uncertainty sampling, which has been reported to be successful in numerous scenarios and settings [11,12], a machine requests instances which cause uncertainty. This leads to the optimal leveraging of both new and existing data [13].
The process of querying the information imitates a classroom instructional method that actively engages learners in the learning process [14,15,16]. They replace or adapt their knowledge and understanding based on prior knowledge in response to learning opportunities provided by a teacher. This contrasts with a model of instruction whereby knowledge is transmitted from the teacher to learners, which typically presents passive learning. Active learning in an educational setting integrates a variety of strategies and instructional models chosen by a teacher to contribute to learners’ knowledge [17].
Hence, machine active learning strategies are still expected to be more versatile and self-sustaining. In particular, deep neural networks demonstrate remarkable performance on particular supervised learning tasks but are not good at telling when they are not sure while working in an active learning environment. The output from the softmax layer usually tends to be overconfident. Besides, deep neural networks have grown so complex that it seems practically impossible to follow their decision-making process [18].
In this study, we intend to inspect humans and machines reasoning processes [19,20,21,22,23] in order to understand how machines make predictions in an active learning environment. Rather than improving performance, we explored whether we can explain how machines come to decisions by imitating human-like reasoning in multiple-choice testing [24,25,26,27,28,29]. We suggest a new uncertainty sampling strategy based on the four-parameter logistic item response theory (4PL IRT) [24] we call Information Capacity. The strategy guarantees the performance similar to the most common uncertainty sampling techniques—Least Confidence and Entropy Sampling—but allows creating more transparent knowledge models in deep learning.
In deep neural networks, we have little visibility into the understanding of how models come to conclusions. This happens because we do not know how learning is supposed to work. While training a model, we iterate with better data, better configurations, better algorithms, and more computational power, although we have little knowledge why that model converges slowly and generalizes poorly. As a result, we do not have much control over rebuilding that model—it is not transparent [18,30,31].
Information Capacity brings with it a new interpretation of learning processes to enlighten “black-box” models. In contrast to Least Confidence and Entropy Sampling, the proposed strategy relies on neural network architectures to model learners’ behavior, where neurons or network weights of network classifiers are considered to be a group of learners with different proficiency in classifying learning items. Information Capacity ensures more flexible deep architectures with explainable and controllable learning behavior, not restricted to connectionist models.

Related Work

  • Deep active learning. Active learning of deep neural models has hardly been considered to date. The prominent related studies report minimizing test errors and computational efforts [32,33,34,35,36], taking some directions towards interpretability in deep learning [37]. This study approaches another major issue within the context of transparency—a lack of reasoning in deep neural models.
  • IRT-based deep learning. Item response theory has been successfully used in solving machine and deep learning problems [38,39,40,41]. They mostly focus on improving generalization ability through optimizing the parameters of IRT models. Rather than optimizing hyperparameters  [42] via IRT model-fitting, we aimed to find meaningful interpretations of deep networks reasoning with learning behaviors.
  • Meta active learning. The reported studies mostly focused on increasing the accuracy of classification with adaptive optimization schemes [1,5,43]. Instead, we intend to simplify an active learning process by integrating a set of evolving learning behaviors into learning models while improving their transparency.

2. Results

2.1. Design of Experiments

We built a SGD-based CNN classifier with two convolutional layers with a ReLU activation and one dropout layer ending with a softmax layer in PyTorch. The first convolutional layer filters the 1 × 10 input image with the square kernel of size 5. The second convolutional layer takes as input the pooled output of the first convolutional layer with a stride of 2 pixels and the square kernel of size 5. An SGD optimizer with learning rate 0.01 and momentum 0.5 was trained on n epoch = 10 with n batch = 64 and tested with n batch = 1000 .
We tested the CNN model on the MNIST and Fashion MNIST datasets. From each dataset we randomly took m train = 10,000 examples for training and m test = 10,000 examples for testing. The active learning environment was created with three labeled pool | L | = { 100 , 500 , 1000 } with fifty rounds n round = 50 and hundred queried examples | L S | = 100 .
The proposed Information Capacity strategy was implemented in line with the two baseline algorithms—Least Confidence and Entropy Sampling. Each experiment was repeated n run = 10 in order to produce statistically significant estimates.

2.2. Analysis of Experiments

The parameters of the proposed strategy in training the model defined clearly interpretable behavior of learners during multiple-choice testing. The learners (network weights) guessed correctly with the probability a i = 0.1 on the item (labeled example) i of the difficulty β i = 4 . We assumed that there was no penalty for guessing announced. The item discrimination parameter α i = 0.25 reflects how well an item discriminates among the learners located at different points θ j along the continuum. These values for parameters are chosen to minimize the maximum of the information capacity of the items in L but, at the same time, avoid possible inaccuracies caused by machine precision when the informativeness measure values are approaching zero and become imperceptible for different classes.
Implemented guessing behavior reflects “noise” in information. Therefore, a nonzero a i reduced the amount information available for locating learners on the θ continuum. In addition, answering the item i, the learners with locations at θ j did not have a success probability equal to 1 but b i = 0.9 due to partial forgetting. The locations θ j < β i present lower level learners, while the locations θ j β i describe higher level learners. The given values for the parameters α i , β i , a i , b i define the behavior of learners responding to the items in accord with the item information function (see Section 3) that presents the amount of information each item provides.
The experiments confirmed that Information Capacity with pre-defined learning behavior can represent the baseline active learning strategies (see Figure 2 and Figure 3). The values of accuracy on testing over rounds m e a n ± s t d are given in Table 1 and Table 2.
The similarity in learning behavior for different subsets of the MNIST and Fashion MNIST datasets pointed to the conclusion that Information Capacity relies on neural network architectures to model learners’ behavior. It can be explained by the fact that decisions on classification tasks are made at the output layer of a network, but depended on weights (learners) which were set at hidden layers. With increasing amount of labeled pool | L | the similarities between the accuracy curves for different strategies become stronger (see Table 1 and Table 2).
We applied a one-sided Wilcoxon test [44] with Bonferroni correction [45,46] for each round to confirm a lack of statistically significant differences between the three strategies in the accuracy values on testing. Since the p-value for each round turned out to be close to 1, we have a sufficient reason to accept the null hypothesis. Consequently, the similarities between the accuracy curves in Figure 2 and Figure 3 are statistically significant.

2.3. Discussion

We took Least Confidence and Entropy Sampling for comparison for two reasons. First, these active learning strategies are used as baseline sampling techniques for more complex approaches adopted in deep active learning. Second, Information Capacity shares some similarity with them—it finds y i which range over all possible labels (Entropy Sampling) with the least information capacity (Least Confidence).
As progress on improving performance in deep learning has come at the cost of transparency, we find this approach particularly beneficial. Information Capacity allows learners to exhibit different learning behaviors with regard to the IRT hyperparameters. In the experiments, they were chosen in a certain way to rule out the reasoning behind the Least Confidence and Entropy Sampling strategies. In particular, we modeled uncertainty with a group of learners, who adopted both guessing and forgetting strategies ( a i > 0 and b i < 1 ) to classify “hard” items ( β i > 2 ). In addition, it was difficult to assess how strong or weak the learners were ( θ j < β i or θ j > β i ) because the value of discrimination factor was low α i < 0.5 . No penalty for guessing p = 0 delivered less predictable behavioral observations.
As we have seen, Least Confidence and Entropy Sampling can be interpreted by the scenario in which each learner (neuron) in a neural network shares the same behavior. Considering the complexity of deep networks, these backbone strategies seem limited. For increasing the transparency of deep learning process, different combinations of the IRT parameters can be used to construct a variety of educational scenarios and learning strategies with strong or weak learners including learning in groups [47,48,49].
The analysis of different neural network architectures with regard to learning behaviors is beyond the scope of this study. However, we hope that our presentation of neural networks will encourage further research exploring novel neural networks building groups of learners with learning behavior which is not limited to gradient-based methods and primitive connectionist models.

3. Materials and Methods

3.1. Problem Statement

Let X be a feature space and Y be a label space. Let P ( X , Y ) be an unknown underlying distribution, where X X , Y Y . We use labeled training set S m = ( x i , y i ) of m labeled training samples to select a prediction function f F , f : X Y so as to minimize the risk R ( f ) = E ( X , Y ) [ ( f ( x ) , y ) ] , where ( · ) R + is a given loss function. For any labeled set L (training and testing), the empirical risk over L is given by:
r ^ L ( f ) = 1 | L | i = 1 | L | ( f ( x i ) , y i ) .
In a pool-based setting [7,33], an active learner chooses examples from a set U = m L of unlabeled samples according to a query function S. Query functions often select points based on information inferred from the current model f s , the existing training set | L | , and the current pool | U | . The aim is to accurately train the model for a given number of labeled points | L S | .
We consider a class B of learning behaviors during testing, where each behavior B B represents a hypothesis class containing all learners f s B , where s defines a set of parameters in a testing framework for making behavioral observations.

3.2. Testing Framework

We are interested in measuring classification proficiencies of a group of learners (neurons or network weights). Although it seems impossible to directly observe the level of proficiency (working knowledge), we can infer its existence through behavioral observations in a classroom. The learners are given an instrument containing several items (labeled examples) i.e., multiple-choice tests [27,28,50,51,52]. The responses to this instrument constitute the behavioral observations.
Item Response Theory [24,28,53,54,55,56] suggests a variety of models to assess the distance between the learner and the item locations as it clearly defines the learner’s correct response. This means that items located toward the right side have difficulty β . They require a learner to have greater proficiency θ to correctly answer items located on the right side than items located on the left side. In general, items located below 0 are “easy” while items above 0 are “hard”.
In this study, we focused on the four-parameter logistic item response theory (4PL IRT) model which can be presented as [24]:
p ( y i j = 1 | θ j , α i , β i , a i , b i ) = a i + b i a i 1 + exp ( α i ( θ j β i ) ) ,
where p ( y i j = 1 | θ j , α i , β i , a i , b i ) is the probability of providing the correct response y i j = 1 to an item i by a learner j with the location (ability) θ j . From the definition (1) we can see that the rate of success mainly depends on the relationship between the item’s parameters and learners’ proficiency.

3.3. Information Capacity

So far, we considered the estimation of a learner’s location from its uncertainty. Let us now take the opposite side and define a query strategy S.
The instrument’s items—labeled examples—contain a certain amount of information that can be used for estimating the learner location parameters. We assume that each item contributes information to reduce the uncertainty about a learner’s location independent of the other items of the instrument. The amount of information items provide can be presented using the Fisher information as [24,57,58]:
S ( θ ) = E 2 θ 2 ln L = i = 1 m p i 2 p i ( 1 p i ) = 1 σ e 2 ( θ ^ | θ )
where σ e 2 ( θ ^ | θ ) is the asymptotic variance error of the estimate θ . The log likelihood function for a learner j’s response vector is equal to:
ln L ( y i j | θ j , α i , β i , a i , b i ) = i = 1 m ( y i j ln ( p i ) + ( 1 y i j ) ln ( 1 p i ) ) ,
where p i p ( y i j = 1 | θ j , α i , β i , a i , b i ) . The items’ capacity is defined as the maximum of the information function S ( θ ) max .
The definition (2) can be rewritten with regard to (1) in the explicit form as:
S ( θ ; α i , β i , a i , b i ) = i = 1 m α i 2 ( p i a i ) 2 ( 1 p i ) ( b i a i ) ( p i a i ) ( 1 b i ) 2 p i ( 1 p i ) ( b i a i ) 2 ( 1 a i ) 2 .
The detailed derivation of the Equation (4) is given in Appendix A. Figure 4 illustrates the projections of the information function with fixed values α = 0.25 , β = 4 , a = 0.1 , b = 0.9 described in Section 2.2.
A new pool-based strategy—we refer to as Information Capacity—suggests estimating the information capacity for unlabelled instances based on the definition (4). Figure 5 depicts the proposed pool-based strategy for measuring the items capacity S with regard to the items’ difficulty β i , learners’ locations θ j , strategies a i and b i , and penalty announcement p in a classroom. The strategy is aimed at “moving” learners along the difficulty axis while keeping high values for the capacity axis. The learners query the examples with the lowest information capacity. For clarity, the algorithm represents the proposed pool-based active learning with the Information Capacity strategy (see Algorithm 1). The differences in implementation in comparison with the traditional active learning framework are highlighted in blue.
Algorithm 1 Pool-based active learning with Information Capacity strategy
1:
procedureInformationCapacity( m train , m test )    
2:
    Initialize a labeled training set L;  
3:
    Initialize an unlabeled training pool U = m train L ;  
4:
    Initialize a learning behavior of learners B with regard to a set of parameters α , β , a , b ;  
5:
    Train a group of learners on the labeled set L;  
6:
    Measure performance of the group of learners on the test set m test ;  
7:
    Initialize several rounds n round and several queried examples | L S | ;   
8:
    for r o u n d n round do 
9:
        Estimate the probabilities with regard to (1) based on the learning behavior B;  
10:
        Sort the unlabeled items in U according to (4) based on the probabilities from the step 9;  
11:
        Query the items L S with the smallest of the maximum capacity S in a r o u n d ;  
12:
         L L L S ;  
13:
         U U \ L S ;  
14:
        Retrain a group of learners on the labeled set L;  
15:
        Measure performance of the group of learners on the test set m test ;  
16:
    end for  
17:
    return The performance of the learners with the interpretation of their learning behavior.  

4. Conclusions

We present Information Capacity, which is an uncertainty sampling strategy that effectively integrates human- and machine-reasoning processes. The strategy allows embedding into the models different learning behaviors with regard to the parameters of the 4PL IRT model. The experiments on the MNIST and Fashion MNIST datasets with the same CNN model indicate that Information Capacity performs similarly to Least Confidence and Entropy Sampling but brings more transparency into a deep learning process.
We considered the neurons or network weights of the CNN classifier at the last hidden layer as a group of learners with different proficiency in classifying learning items, i.e., images. The pre-defined parameters of the Information Capacity strategy defined their learning behavior: the learners had a success probability b i = 0.9 due to partial forgetting while they guessed correctly with the probability a i = 0.1 on the item i of the difficulty β i = 4 , which discriminated the learners with the factor α i = 0.25 .
The equivalence of the parameters α i , β i , a i , b i for different subsets of the MNIST and Fashion MNIST datasets revealed that the model architecture greatly influences learning behavior. As a direction for further research, we suggest modeling learning behaviors with different network architectures. While keeping equally good performance due to the similarity between different strategies, it seems desirable to optimize neural network architectures and learning processes.
The code used for empirical evaluation is available at https://github.com/yukinoi/human-like-active-learning.

Author Contributions

All coauthors contributed equally to the paper. Conceptualization, methodology and investigation, I.K., T.L. and T.Š.; software, I.K.; validation; T.L. and T.Š.; formal analysis, I.K. and T.L.; writing—original draft, I.K.; writing—review and editing, T.L. and T.Š.; T.Š.: supervision and project administration. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Centre of Excellence project “DATACROSS”, co-financed by the Croatian Government and the European Union through the European Regional Development Fund—the Competitiveness and Cohesion Operational Programme (KK.01.1.1.01.0009) and the Ministry of Education and Science of the Russian Federation (grant No. 074-U01).

Acknowledgments

The authors would like to thank the reviewers for valuable comments and suggestions.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Derivation of the Equation (4)

Let us denote p i p ( y i j = 1 | θ j , α i , β i , a i , b i ) . The derivative of the model (1) can be defined as
p i = α i ( b i a i ) exp ( α i ( θ j β i ) ) 1 + exp ( α i ( θ j β i ) ) 2 .
Extracting the definition (1) from (A1) gives
p i = α i ( p i a i ) exp ( α i ( θ j β i ) ) 1 + exp ( α i ( θ j β i ) ) .
Let us now define the probability of getting an incorrect response:
1 p i = ( 1 a i ) exp ( α i ( θ j β i ) ) 1 + exp ( α i ( θ j β i ) ) 1 b i 1 + exp ( α i ( θ j β i ) ) .
From the Equation (A3), it follows that
exp ( α i ( θ j β i ) ) 1 + exp ( α i ( θ j β i ) ) = ( 1 p i ) ( b i a i ) ( p i a i ) ( 1 b i ) ( b i a i ) ( 1 a i ) .
Taking into account (A4), we can rewrite (A1) as follows:
p i = α i ( b i a i ) ( 1 p i ) ( b i a i ) ( p i a i ) ( 1 b i ) ( b i a i ) ( 1 a i ) .
Substituting (A5) into the definition (2) immediately gives (4).

References

  1. Bachman, P.; Sordoni, A.; Trischler, A. Learning algorithms for active learning. In Proceedings of the 34th International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017. [Google Scholar]
  2. Fazakis, N.; Kanas, V.G.; Aridas, C.K.; Karlos, S.; Kotsiantis, S. Combination of active learning and semi-supervised learning under a self-training scheme. Entropy 2019, 21, 988. [Google Scholar] [CrossRef] [Green Version]
  3. Hsu, W.-N.; Lin, H.-T. Active learning by learning. In Proceedings of the Twenty-Ninth AAAI Conference on Artificial Intelligence, Austin, TX, USA, 25–30 January 2015. [Google Scholar]
  4. Huang, S.; Jin, R.; Zhou, Z. Active learning by querying informative and representative examples. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 36, 1936–1949. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Konyushkova, K.; Raphael, S.; Fua, P. Learning active learning from data. In Proceedings of the Conference on Neural Information Processing Systems (NIPS), Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  6. Ramirez-Loaiza, M.E.; Sharma, M.; Kumar, G.; Bilgic, M. Active learning: An empirical study of common baselines. Data Min. Knowl. Discov. 2017, 31, 287–313. [Google Scholar] [CrossRef]
  7. Settles, B. Active Learning Literature Survey. In Computer Sciences Technical Report; University of Wisconsin–Madison: Madison, WI, USA, 2009. [Google Scholar]
  8. Sourati, J.; Akcakaya, M.; Dy, J.G.; Leen, T.K.; Erdogmus, D. Classification Active Learning Based on Mutual Information. Entropy 2016, 18, 51. [Google Scholar] [CrossRef]
  9. Sourati, J.; Akcakaya, M.; Leen, T.K.; Erdogmus, D.; Dy, J.G. A probabilistic active learning algorithm based on Fisher information ratio. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 2023–2029. [Google Scholar] [CrossRef]
  10. Wu, J.; Sheng, V.S.; Zhang, J.; Li, H.; Dadakova, T.; Swisher, C.L.; Cui, Z.; Zhao, P. Multi-label active learning algorithms for image classification: Overview and future promise. ACM Comput. Surv. 2020, 53, 1–35. [Google Scholar] [CrossRef] [Green Version]
  11. Joshi, A.J.; Porikli, F.; Papanikolopoulos, N. Multi-class active learning for image classification. In Proceedings of the Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009. [Google Scholar]
  12. Yang, Y.; Ma, Z.; Nie, F.; Chang, X.; Hauptmann, A.G. Multi-Class active learning by uncertainty sampling with diversity maximization. Int. J. Comput. Vis. 2015, 113, 113–127. [Google Scholar] [CrossRef]
  13. Hanneke, S. Activized Learning: Transforming passive to active with improved label complexity. J. Mach. Learn. Res. 2012, 13, 1469–1587. [Google Scholar]
  14. Bonwell, C.; Eison, J. Active Learning: Creating Excitement in the Classroom; Jossey-Bass: San Francisco, CA, USA, 1991. [Google Scholar]
  15. Cook, B.R.; Babon, A. Active learning through online quizzes: Better learning and less (busy) work. J. Geogr. High. Educ. 2017, 41, 24–38. [Google Scholar] [CrossRef]
  16. Prince, M. Does Active Learning Work? A review of the research. J. Eng. Educ. 2004, 93, 223–231. [Google Scholar] [CrossRef]
  17. Aubrey, K.; Riley, A. Understanding and Using Educational Theories; Sage Publications Ltd.: Newbury Park, CA, USA, 2015. [Google Scholar]
  18. Mascharka, D.; Tran, P.; Soklaski, R.; Majumdar, A. Transparency by design: Closing the gap between performance and interpretability in visual reasoning. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  19. Castro, R.; Kalish, C.; Nowak, R.; Qian, R.; Rogers, T.; Zhu, X. Human active learning. In Proceedings of the Twenty-Second Annual Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 8–11 December 2008. [Google Scholar]
  20. Lake, B.M.; Salakhutdinov, R.; Tenenbaum, J.B. Human-level concept learning through probabilistic program induction. Science 2015, 350, 1332–1338. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  21. Lake, B.M.; Ullman, T.D.; Tenenbaum, J.B.; Gershman, S.J. Building machines that learn and think like people. Behav. Brain Sci. 2017, 40, e253. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  22. Mastorakis, G. Human-like machine learning: Limitations and suggestions. arXiv 2018, arXiv:1811.06052v1. [Google Scholar]
  23. Wilson, R.C.; Shenhav, A.; Straccia, M.; Cohen, J.D. The Eighty Five Percent Rule for optimal learning. Nat. Commun. 2019, 10, 1–9. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. De Ayala, R.J. The Theory and Practice of Item Response Theory (Methodology in the Social Sciences); The Guilford Press: New York, NY, USA, 2009. [Google Scholar]
  25. Gierl, M.J.; Bulut, O.; Guo, Q.; Zhang, X. Developing, analyzing, and using distractors for multiple-choice tests in education: A comprehensive review. Rev. Educ. Res. 2017, 87, 1082–1116. [Google Scholar] [CrossRef]
  26. Hakel, M.D. (Ed.) Beyond Multiple Choice: Evaluating Alternatives to Traditional Testing for Selection; Lawrence Erlbaum Associates Publishers: Mahwah, NJ, USA, 1998. [Google Scholar]
  27. Lee, C.J. The test taker’s fallacy: How students guess answers on multiple-choice tests. Behav. Decis. Mak. 2019, 32, 140–151. [Google Scholar] [CrossRef]
  28. Lord, F.M. Applications of Item Response Theory to Practical Testing Problems; Erlbaum: Hillsdale, NJ, USA, 1980. [Google Scholar]
  29. Thissen, D.; Steinberg, L.; Fitzpatrick, A.R. Multiple-choice models: The distractors are also part of the item. J. Educ. Meas. 1989, 26, 161–176. [Google Scholar] [CrossRef]
  30. Mittelstadt, B.; Russell, C.; Wachter, S. Explaining explanations in AI. In Proceedings of the Fairness, Accountability, and Transparency (FAT*), Atlanta, GA, USA, 29–31 January 2019; pp. 279–288. [Google Scholar]
  31. Rudin, C. Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nat. Mach. Intell. 2019, 1, 206–215. [Google Scholar] [CrossRef] [Green Version]
  32. Gal, Y.; Islam, R.; Ghahramani, Z. Deep Bayesian active learning with image data. In Proceedings of the 34th International Conference on MachineLearning, Sydney, Australia, 6–11 August 2017. [Google Scholar]
  33. Geifman, Y.; El-Yaniv, R. Deep active learning with a neural architecture search. In Proceedings of the 33rd Conference on Neural Information Processing Systems (NeurIPS 2019), Vancouver, BC, Canada, 8–14 December 2019. [Google Scholar]
  34. Ouali, Y.; Hudelot, C.; Tami, M. An overview of deep semi-supervised learning. arXiv 2020, arXiv:2006.05278v1. [Google Scholar]
  35. Sener, O.; Savarese, S. Active learning for convolutional neural networks: A core-set approach. In Proceedings of the 6th International Conference on Learning Representations (ICLR 2018), Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  36. Wang, D.; Shang, Y. A new active labeling method for deep learning. In Proceedings of the International Joint Conference on Neural Networks (IJCNN), Beijing, China, 6–11 July 2014; pp. 112–119. [Google Scholar]
  37. Budd, S.; Robinson, E.C.; Kainz, B. A survey on active learning and human-in-the-loop deep learning for medical image analysis. arXiv 2019, arXiv:1910.02923v1. [Google Scholar]
  38. Chen, Y.; Filho, T.S.; Prudencio, R.B.C.; Diethe, T.; Flach, P. β3-IRT: A new item response model and its applications. arXiv 2019, arXiv:1903.04016. Available online: https://arxiv.org/abs/1903.04016 (accessed on 3 June 2019).
  39. Martinez-Plumed, F.; Prudencio, R.B.C.; Martinez-Uso, A.; Hernandez-Orallo, J. Item response theory in AI: Analysing machine learning classifiers at the instance Level. Artif. Intell. 2019, 271, 18–42. [Google Scholar] [CrossRef]
  40. Whitehill, J.; Ruvolo, P.; Wu, T.; Bergsma, J.; Movellan, J. Whose vote should count more: Optimal integration of labels from labelers of unknown expertise. In Proceedings of the Advances in Neural Information Processing Systems 22 (NIPS 2009), Vancouver, BC, Canada, 7–10 December 2009. [Google Scholar]
  41. Yeung, C.K. Deep-IRT: Make deep learning based knowledge tracing explainable using item response theory. arXiv 2019, arXiv:1904.11738. [Google Scholar]
  42. Lalor, J.P.; Wu, H.; Yu, H. CIFT: Crowd-informed fine-tuning to improve machine learning ability. arXiv 2017, arXiv:1702.08563v2. [Google Scholar]
  43. Ravi, S.; Larochelle, H. Meta-Learning for Batch Mode Active Learning. 2018. Available online: https://openreview.net/forum?id=r1PsGFJPz (accessed on 4 June 2018).
  44. Wilcoxon, F. Individual comparisons by ranking methods. Biom. Bull. 1945, 1, 80–83. [Google Scholar] [CrossRef]
  45. Dunn, O.J. Multiple comparisons among means. J. Am. Stat. Assoc. 1961, 56, 52–64. [Google Scholar] [CrossRef]
  46. Perneger, T.V. What’s wrong with Bonferroni adjustments. BMJ 1998, 316, 1236–1238. [Google Scholar] [CrossRef]
  47. Kulikovskikh, I.M.; Prokhorov, S.A.; Suchkova, S.A. Promoting collaborative learning through regulation of guessing in clickers. Comput. Hum. Behav. 2017, 75, 81–91. [Google Scholar] [CrossRef]
  48. Le, H.; Janssen, J.; Wubbels, T. Collaborative learning practices: Teacher and student perceived obstacles to effective student collaboration. Camb. J. Educ. 2018, 48, 103–122. [Google Scholar] [CrossRef] [Green Version]
  49. Sawyer, J.; Obeid, R. Cooperative and collaborative learning: Getting the best of both words. In How We Teach Now: The GSTA Guide to Student-Centered Teaching; Obeid, R., Schwartz, A., Shane-Simpson, C., Brooks, P.J., Eds.; Society for the Teaching of Psychology: Washington, DC, USA, 2018; pp. 163–177. [Google Scholar]
  50. Liu, C.W.; Wang, W.C. Unfolding IRT models for Likert-type items with a don’t know option. Appl. Psychol. Meas. 2016, 49, 517–533. [Google Scholar] [CrossRef]
  51. Liu, C.W.; Wang, W.C. A general unfolding IRT model for multiple response styles. Appl. Psychol. Meas. 2019, 43, 195–210. [Google Scholar] [CrossRef] [PubMed]
  52. Sideridis, G.; Tsaousis, I.; Harbi, K.A. Improving measures via examining the behavior of distractors in multiple-choice tests: Assessment and remediation. Educ. Psychol. Meas. 2016, 77, 82–103. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  53. Bonifay, W. Multidimensional Item Response Theory (Quantitative Applications in the Social Sciences); SAGE Publications: Newbury Park, CA, USA, 2020. [Google Scholar]
  54. DeMars, C.E. “Guessing” parameter estimates for multidimensional Item Response Theory models. Educ. Psychol. Meas. 2007, 67, 433–446. [Google Scholar] [CrossRef]
  55. Gin, B.; Sim, N.; Skrondal, A.; Rabe-Hesketh, S. A dyadic IRT model. arXiv 2019, arXiv:1906.01100v1. [Google Scholar]
  56. Reckase, M.D. Multidimensional Item Response Theory; Springer: New York, NY, USA, 2009. [Google Scholar]
  57. Frieden, B.R. Science from Fisher Information: A Unification; Cambridge University Press: Cambridge, UK, 2004. [Google Scholar]
  58. Lehmann, E.L.; Casella, G. Theory of Point Estimation, 2nd ed.; Springer: New York, NY, USA, 1998. [Google Scholar]
Figure 1. The concepts of learning environments: Passive vs. Active.
Figure 1. The concepts of learning environments: Passive vs. Active.
Entropy 22 00906 g001
Figure 2. The accuracy curves on MNIST for different numbers of labeled pool: (a) | L | = 100 . (b) | L | = 500 . (c) | L | = 1000 .
Figure 2. The accuracy curves on MNIST for different numbers of labeled pool: (a) | L | = 100 . (b) | L | = 500 . (c) | L | = 1000 .
Entropy 22 00906 g002
Figure 3. The accuracy curves on Fashion MNIST for different numbers of labeled pool: (a) | L | = 100 . (b) | L | = 500 . (c) | L | = 1000 .
Figure 3. The accuracy curves on Fashion MNIST for different numbers of labeled pool: (a) | L | = 100 . (b) | L | = 500 . (c) | L | = 1000 .
Entropy 22 00906 g003
Figure 4. The projections of the information function on different planes: (a) α -plane with α = [ 0 , 5 ] . (b) β -plane with β = [ 0 , 5 ] . (c) a-plane with a = [ 0 , 1 ] . (d) b-plane with b = [ 0 , 1 ] .
Figure 4. The projections of the information function on different planes: (a) α -plane with α = [ 0 , 5 ] . (b) β -plane with β = [ 0 , 5 ] . (c) a-plane with a = [ 0 , 1 ] . (d) b-plane with b = [ 0 , 1 ] .
Entropy 22 00906 g004
Figure 5. The proposed “classroom” strategy.
Figure 5. The proposed “classroom” strategy.
Entropy 22 00906 g005
Table 1. The values of accuracy on testing over rounds m e a n ± s t d on MNIST.
Table 1. The values of accuracy on testing over rounds m e a n ± s t d on MNIST.
Strategy | L | = 100 | L | = 500 | L | = 1000
Information Capacity 0.8135 ± 0.0254 0.8714 ± 0.0279 0.9153 ± 0.0117
Entropy Sampling 0.8071 ± 0.0259 0.8688 ± 0.0264 0.9133 ± 0.0141
Least Confidence 0.8182 ± 0.0276 0.8727 ± 0.0275 0.9156 ± 0.0126
Table 2. The values of accuracy on testing over rounds m e a n ± s t d on Fashion MNIST.
Table 2. The values of accuracy on testing over rounds m e a n ± s t d on Fashion MNIST.
Strategy | L | = 100 | L | = 500 | L | = 1000
Information Capacity 0.5646 ± 0.0399 0.6279 ± 0.0397 0.6779 ± 0.0245
Entropy Sampling 0.5499 ± 0.0362 0.6252 ± 0.0401 0.678 ± 0.0254
Least Confidence 0.5615 ± 0.0387 0.6273 ± 0.04 0.677 ± 0.0255

Share and Cite

MDPI and ACS Style

Kulikovskikh, I.; Lipic, T.; Šmuc, T. From Knowledge Transmission to Knowledge Construction: A Step towards Human-Like Active Learning. Entropy 2020, 22, 906. https://0-doi-org.brum.beds.ac.uk/10.3390/e22080906

AMA Style

Kulikovskikh I, Lipic T, Šmuc T. From Knowledge Transmission to Knowledge Construction: A Step towards Human-Like Active Learning. Entropy. 2020; 22(8):906. https://0-doi-org.brum.beds.ac.uk/10.3390/e22080906

Chicago/Turabian Style

Kulikovskikh, Ilona, Tomislav Lipic, and Tomislav Šmuc. 2020. "From Knowledge Transmission to Knowledge Construction: A Step towards Human-Like Active Learning" Entropy 22, no. 8: 906. https://0-doi-org.brum.beds.ac.uk/10.3390/e22080906

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop