Next Article in Journal
A Secure CDM-Based Data Analysis Platform (SCAP) in Multi-Centered Distributed Setting
Next Article in Special Issue
A Hybrid Recommender System for HCI Design Pattern Recommendations
Previous Article in Journal
Effect of Substrate-Thickness on Voltage Responsivity of MEMS-Based ZnO Pyroelectric Infrared Sensors
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Study of a Gain Based Approach for Query Aspects in Recall Oriented Tasks

by
Giorgio Maria Di Nunzio
*,† and
Guglielmo Faggioli
Department of Information Engineering, University of Padova, Via Gradenigo 6/b, 35131 Padova, Italy
*
Author to whom correspondence should be addressed.
Authors contributed equally to this work.
Submission received: 8 September 2021 / Revised: 24 September 2021 / Accepted: 26 September 2021 / Published: 29 September 2021
(This article belongs to the Special Issue Human and Artificial Intelligence)

Abstract

:
Evidence-based healthcare integrates the best research evidence with clinical expertise in order to make decisions based on the best practices available. In this context, the task of collecting all the relevant information, a recall oriented task, in order to take the right decision within a reasonable time frame has become an important issue. In this paper, we investigate the problem of building effective Consumer Health Search (CHS) systems that use query variations to achieve high recall and fulfill the information needs of health consumers. In particular, we study an intent-aware gain metric used to estimate the amount of missing information and make a prediction about the achievable recall for each query reformulation during a search session. We evaluate and propose alternative formulations of this metric using standard test collections of the CLEF 2018 eHealth Evaluation Lab CHS.

1. Introduction

The study of the query representation in Information Retrieval has driven a lot of interest in recent years [1,2,3,4,5,6,7]. Several works in the past [8,9,10] showed the positive effect on the retrieval results of fusing runs retrieved with human-made multiple formulations of the same information need. Recent studies have shown how query reformulations automatically extracted from query logs can be as effective as those manually created by users [11]. Furthermore, the performance of a system can greatly improve when the “right” formulation of an information need is selected [4,5]. One of the main challenges in this research area is being able to suggest the best performing query (or queries) among the possible variations [4,5,12,13,14]. For example, Thomas et al. [4] observed that, the most prominent effect in predicting the performance of a query formulation is due to the information need and not to the “query wording”. In this sense, query performance predictors actually predict the complexity of the information need, rather than the one the query itself. Zendel et al. [5] pursue a slightly different task. Following the literature on reference lists [15,16] they try to predict the performance for a query using information about queries representing the same information need. Benham et al. [3] define a fusion approach for multiple query formulations based on the concept of “topic centroid”, which describes the information need as combination of its formulations. Dang et al. [12] address also the problem of improving the ranking results through a query formulation selection phase. Note that, Dang et al. [12] show how they are often capable of putting the best query in the first two positions (not only the first one), a further evidence of the complexity of the task.
A use case of query performance prediction is the systematic compilation of literature review. In fact, systematic reviews are scientific investigations that use strategies to include a comprehensive search of all potentially relevant articles. As time and resources are limited for compiling a systematic review, limits to the search are needed: for example, one may want to estimate how far the horizon of the search should be (i.e., all possible cases/documents that could exist in the literature) in order to stop before the resources are finished [17]. Scells et al. [13] apply several state-of-the-art Query Performance Predictors to select the best query in the Systematic Reviews domain. They show how current Query Performance Prediction approaches perform poorly on this specific task. International evaluation campaigns have organized labs in order to study this problem in terms of the evaluation, through controlled simulation, of methods designed to achieve very high recall [18,19]. The CLEF initiative (http://www.clef-initiative.eu, accessed on 15 February 2021) has promoted the eHealth track since 2013 and, the CLEF 2018 eHealth Evaluation Lab Consumer Health Search (CHS) task [20] investigated the problem of building search engines that are robust to query variations to support information needs of health consumers.
In this paper, we study an alternative formulation of the intent-aware metric proposed by Umemoto et al. [21], in which the authors analyze a metric to estimate the amount of missing information for each query reformulation during a search session. Note that in [21] the authors do not propose an approach capable of predicting the recall of different formulations. Nevertheless, our perception is that, their approach can be easily adapted with good results also to the predictive task. In our case, our research goal is to understand whether a gain based approach can be used to predict the relative importance of each reformulation in terms of recall performance, in the context of Consumer Health Search where users need support for medical information needs.
In this sense, with respect to [21], our contribution is two-fold:
  • we show that it is possible to apply the GAIN measure proposed in [21] to obtain a recall predictor over a set of formulations for the same topic;
  • we furthermore show how to improve the results of such predictor by exploiting also the information obtained through the various formulations.
The paper is organized as follows: in Section 2, we present the original gain metric, while in Section 3 we define our alternative version to predict the performance in a recall-oriented fashion. In Section 4, we discuss the experimental analysis and results; while in Section 5 we give our final remarks.

2. A GAIN-Based Approach

In Umemoto et al. [21], define the intent-aware gain metric and the requirements that it should satisfy. They identify the following properties: importance, documents relevant to a central aspect of the search topic produce higher gain than those relevant to a peripheral one; relevance, highly relevant documents produce higher gain than partially relevant ones; novelty, documents relevant to an unexplored aspect produce higher gain than those relevant to a fully explored aspect.
The set of aspects A t of a topic t is estimated through the process described in [22]: first, a set of subtopics S t is mined given a topic t; then, the subtopics are grouped into a set of clusters C t . These clusters are regarded as the “facets” (We use facets instead of aspects to not repeat the same term that will be use to identify the most representative subtopic.) of t. The most representative subtopic s is chosen from each cluster as formulation of the topic aspect a using the formula a = a r g m a x s C t Imp t ( s ) , where the importance of a subtopic s is defined as:
Imp t ( s ) = d D s N d D t N 1 Rank t ( d )
D s N and D t N denote the sets of the top N retrieved documents for a subtopic s and the topic t, respectively, and Rank t ( d ) is the rank of the document d in the ranked list for t.
It is crucial to stress that the definition of importance, and the following definition of gain, derives from the assumption that there is a known “reference” topic t that describes completely the information need. For such topic t the retrieved documents can be different compared to the ones observed for a query which represents just one aspect a of the topic. In Figure 1, we show an example of a number of subtopics found for a topic t and grouped into three clusters, each one with a representative aspect.
The Intent-Aware Gain is defined for a set of documents D as:
Gain-IA t ( D ) = a A t P ( a | t ) · Gain t , a ( D )
which is a sort of expected value of the gain across the different aspects. P ( a | t ) is the probability that an aspect a is important to the topic t, and Gain t , a ( D ) is the gain that can be obtained by the aspect a from the documents D. The importance probability for an aspect of a topic is computed as:
P ( a | t ) = Imp t ( a ) a A t Imp t ( a )
while the gain which measures how the documents D retrieved for a query contribute to increment the information relative to a specific aspect of the topic is:
Gain t , a ( D ) = 1 d D ( 1 Rel t , a ( d ) )
This last part that is required to compute the Intent-Aware Gain contains the term Rel t , a ( d ) which is the relevance degree of a document d with respect to an aspect a, estimated as follows:
Rel t , a ( d ) = s C a Imp t ( s ) · Rel s ( d ) s C a Imp t ( s )
where C a C t is the cluster of subtopics belonging to the aspect a, and R e l s ( d ) is the relevance degree of a document d to a subtopic s estimated as Rel s ( d ) = 1 / Rank s ( d ) .

3. A Gain for Query Reformulations

Our initial hypothesis in this work is that: (a) we have one information need expressed with different query reformulations, and (b) the topic t is unknown. In particular, given an information need i and its set of reformulations V i , we assume that each reformulation q V i is able to ‘reveal’ different facets of i. Consequently, we need to redefine the expression of the gain of Equation (4) as:
Gain i , q ( D ) = 1 d D ( 1 Rel i , q ( d ) )
where i is the information need and q is a specific (re)formulation.
The main difference with the original approach, apart from changing variable names, is the fact that (i) we do not have a ‘reference’ topic t that describes completely the information need i, and (ii) we have one single cluster of query reformulations, or variants, V i . For these reasons, we also need an alternative definition of relevance that adapts to our case study:
Rel i , q ( d ) = s V i Imp q ( s ) Rel s ( d ) s V i Imp q ( s )
where the relevance of d, retrieved by the query variant q of the information need i, is computed as the weighted average of the relevance of d with respect to all the alternative reformulations in V i . The two terms Imp q ( s ) e Rel s ( d ) remain unaltered compared to the previous definitions:
Imp q ( s ) = d D s N D q N 1 Rank q ( d ) , Rel s ( d ) = 1 Rank s ( d )

3.1. A Similarity Matrix for Recall Prediction

In the proposed context, we can think of an ‘optimal’ query as the one capable of combining all the diverse facets of the information need it represents. In order to estimate which query reformulation q is the closest to the unknown optimal one, we propose the following procedure:
1
we define D q as the set of documents retrieved by q;
2
D i = q V i D q as the set of all documents retrieved by at least one reformulation q;
3
R R | V i | × | D i | as the matrix of rankings for the information need i where each row corresponds to a specific reformulation and each column to a document. The value of an element r k , d of R is defined as | D q | ρ q , d where ρ q , d is the rank of document d retrieved by q. R is at the end normalized with norm l 2 .
At this point, we want to build a similarity matrix to predict the impact in terms of recall that each reformulation will have on the retrieval. We compute the cosine similarity between each pair of rows in R , obtaining a symmetric matrix S where each row (or column) represents how a reformulation is similar to the others. We use the sum the k-th row (or column) of S to predict the importance of the k-th query; then, we order the query reformulations in decreasing order where greater values indicate a higher probability of retrieving more relevant documents. This measure describes how close each query is to the ideal “centroid” query that perfectly describes the topic.

4. Experiments and Analysis

In this section, we describe the analysis of our experiments. In particular, we want to compare the performance in terms of predicted recall among: (i) the gain defined in Equation (6), (ii) an alternative definition that mitigates some arithmetical issues, (iii) and the similarity matrix. To the best of our knowledge, this is the first effort in predicting the recall for the systematic reviews task, when multiple formulations are considered. Therefore, we are not able to directly compare it with an approach explicitly thought for such task. We thus compare our solution with traditional QPP strategies. Furthermore, we use the techniques presented in Umemoto et al. [21] as baselines.

4.1. Test Collection and Retrieval Model

The CLEF 2018 eHealth Evaluation Lab Consumer Health Search (CHS) task [20] investigated the problem of retrieving Web pages to support information needs of health consumers that are confronted with a health problem or a medical condition. One subtask (i.e., subtask 3) of this lab is aimed to foster research into building search systems that are robust to query variations (https://github.com/CLEFeHealth/CLEFeHealth2018IRtask, accessed on 15 February 2021).
Queries There are 50 information need for which we have 7 query reformulation for a total of 350 queries: the original 50 queries issued by the general public augmented with 6 query variations issued individually by 6 research students with no medical knowledge (The queries and the process to obtain them are described in http://www.khresmoi.eu/assets/Deliverables/WP7/KhresmoiD73.pdf, accessed on 15 February 2021).
Collection The collection contains 5,535,120 Web pages and it was created by compiling Web pages of selected domains acquired from the CommonCrawl [20].
Relevance Assessments For each information need, the organizers of the task provided about 500 documents assessed for a total of 25,000 topic-document pairs.
Retrieval Model The index provided by the organizers of the task, an ElasticSearch index version 5.1.1, comes with a standard BM25 model with parameters b = 0.75 and k1 = 1.2 (https://sites.google.com/view/clef-ehealth-2018/task-3-consumer-health-search, accessed on 15 February 2021).
Notice That, among the queries of the CLEF 2018 eHealth CHS collection, the two identified by ids 160006 and 164007 will not retrieve any document in common with the other variants of the same information need (at least for N 1000 ). This is because the text of query 160006 is “nan”, while query 164007 has a typo “pros and cons spirculina”, instead of spirulina, a type of algae. We stress on this aspect since, for those queries, it will not be possible to compute the value of the gain by definition, since the intersection of their ranked list with the ones for other formulations of the same topic will be empty.

4.2. Using Traditional Query Performance Predictors Applied to Recall Prediction for Systematic Reviews

To have a better grasp on the peculiarities of the problem, we first try to apply traditional techniques of Query Performance Prediction (QPP) to our specific setting. We aim at showing that, traditional QPP techniques fail to correctly order formulations when (i) the recall is the key performance indicator; (ii) we sort formulations of the same topic and not queries representing different topics. Showing this, is a further evidence of the importance of using appropriate tools, such as the gain as described in Section 2 to correctly tackle the problem. More in detail, we select a set of very well-know QPP models, in order to determine whether they can be satisfactory applied to the prediction of the recall and can be used with the documents and queries that we have at hand. Traditionally, Query Preformance Predictors are divided into two macro-categories, according to the information they exploit to formulate the prediction: Pre-retrieval predictors and Post-retireval Predictors. Pre-retrieval predictors analyze query and corpus statistics prior to retrieval [14,23,24,25,26,27,28] and post-retrieval predictors that also analyze the retrieval results [15,29,30,31,32,33,34,35,36]. Even though Pre-retrieval predictors have the advantage of being faster, since they do not need to retrieve the documents for a certain run, post-retrieval predictors typically perform better. Table 1 reports the predictors that we include in our analyses and a brief description of how they work. It is important to notice that, as for many QPP models, the models that we selected do not actually predict the performance measure. They associate a score to each of the queries, which is expected to correlate with the performance measure, but is on a different scale and cannot be used directly as estimate of the performance.
The traditional strategy to evaluate how good a query performance predictor is, consists in computing a traditional retrieval performance measure, such as Average Precision (AP), for each of the query, and determine how much such measure correlates with the prediction scores computed by the QPP model [23,24,25,26,28,32,33,34,35,38,41,42,43]. Notice that, there are two main aspects that might impair traditional QPP models in our specific setting:
  • Remember that we are in the setting of the systematic reviews. Therefore, it is by far more important to retrieve as many as possible relevant documents, rather than putting them in the first positions. Therefore, we are not interested in estimating the AP, which is a precision based measure, but our aim is to predict which query will have the best recall;
  • We do not compare queries meant for different information needs, which is the typical evaluation scenario for QPP models.
On the other hand, we aim at understanding which one, among a set of queries representing the same information need, achieve the best result.
To determine whether we are impaired by the first problem, we first apply the traditional QPP considering only the default formulation of each topic, and we compare whether the predictors are capable of correctly determining the inter-topic performance. More in detail, with this first experiment, we are interested in understanding whether the baseline predictors are capable of predicting which topic will have the best recall, using a single formulation for each of them. Table 2 reports the result of such analysis.
We can observe that, by looking at Table 2, the results are in line with previous similar experiments in the literature, such as [5,44]. Almost all the predictors are able to achieve a significant correlation with the recall (with level α = 0.01 ). Two noticeable exceptions are represented by nqc and smv: traditionally, they are considered among the best predictors, but in this specific scenario they fail, with correlations not statistically different from 0. Our hypothesis is that, while pre-retrieval predictors tend to be estimators of the recall base of a query, and therefore tend to correlate with the recall itself, post-retrieval predictors tend to compute their predictors based on the scores that the retrieval model assigns to the top-ranked documents. In this sense, post-retrieval predictors are “top-heavy”: they focus on the upper part of the ranked list of documents. This behaviour favours predicting the performance for top-heavy measures, such as Average Precision or nDCG. Instead, our task consists in predicting the recall, given a long list of documents. It is not unlikely that the upper part of the list of retrieved documents is saturated with relevant ones; nevertheless, we are more interested in being sure that every relevant document has been considered, rather than saying whether the top part of the ranked list contains relevant documents.
We now switch the focus from predicting the performance across topics, to predict the performance within topics. Instead of comparing the performance that the standard formulation is expected to achieve for each topic, we try to sort different formulations for the same topic, according to the predicted performance. Table 3 reports the results of our analysis.
Compared to the results observed in Table 2, the performance achieved by traditional predictors for the “within”-topics prediction, is extremely lower, with very few cases of significantly positive correlation between the predicted and observed recall. Note that, even though we agree with [13] on the fact that predicting the best query among a series of formulations of a topic is a hard task, we end up with diametrically opposite conclusions. Scells et al. [13] observed severe flaws in traditional QPP techniques when predicting the performance across topics. On the other hand, they found the task of predicting the performance within topics (which they refer to as Query Variation Performance Prediction (QVPP)) to be easier, achieving higher (although still very low) results. What we observe here, is diametrically opposite: we found the worst results when predicting results within topics, and performance in line with previous literature for the predictions across topics. A possible explanation for this phenomenon is that we use the traditional QPP models for a different task compared to Scells et al. [13]. In fact, our aim is to predict the recall, while Scells et al. [13] aim at predicting the Average Precision. As a final remark, we want to point out that, Zendel et al. [45] recently showed how the “QVPP” is a harder task, compared to traditional QPP, confirming in this sense our findings.

4.3. Analysis of the Results

Given what we observed in Section 4.2, we are interested in understanding whether the GAIN-based proposed by [21] (cfr. Equation (4)) can overcome the problems in this specific setting shown by traditional QPP models. The results are shown in Figure 2a,d,h. Each figure is divided into two parts: top, we show the distribution of values of the GAIN (or similarity), ordered increasingly, for each query reformulation (350 in total); bottom, we plot for each topic (50 topics) the value of the correlation Kendall τ between the query reformulations ordered by decreasing GAIN (or similarity) and the reformulations ordered by decreasing true recall. The blue dots indicate a statistically significant correlation greater (or lower) than zero, while black dots the topics for which it is not possible to compute the correlation.

4.3.1. Saturated GAIN Distribution

In Figure 2a,d,h, we show that the value of the gain saturates to 1 for most query reformulation. This is more evident when we increase the number of documents N of Equation (6) from N = 100 up to N = 10,000. This behavior, due to the importance in Equation (1) that multiplies N numbers less than one, makes the GAIN not useful to discriminate the different query variants of an information need, since every variant will have gain equal to 1. In addition, when all the reformulations have the same gain, it is impossible to compute the Kendall τ correlation to predict the performance (black dots with correlation value 0 in the figure). Being not saturated is not by itself a desirable feature for the gain measure. Nevertheless, the faster the gain saturates, the harder it is to discriminate between different formulations. In this sense, a GAIN measure capable of spreading better the options in the entire domain is preferable.

4.3.2. Alternative GAIN Definition

In order to mitigate the aforementioned problems, we propose an alternative definition of the gain of Equation (6) substituting the product with an average:
G A I N i , q ( D ) = 1 d D ( 1 Rel i , q ( d ) ) | D |
The results of this new formulation are shown in Figure 2b,e,g. The distribution of the gain is more spread across all the reformulations and does not saturate to one. There is also a more stable prediction of the performances for each topic: the number of statistically significant predictions of the recall of the reformulation is between 17 and 19, from N = 100 and N = 10,000; in addition, the number of negative correlations (wrong predictions of performance) decreases. This indicates (as we may expect) that with more information (more documents, greater N) we can predict better the order of importance, in terms of recall, of each reformulation.

4.3.3. Using Similarity Matrix for Recall Prediction

In Figure 2c,f,i, we show the ability to predict the performance of a query reformulation using the correlation between the similarity-based approach presented in Section 3.1. The values of the Similarity are spread and do not saturate to the maximum value of the sum of a row of S (in our experiments equal to 7). By increasing the number N of documents, we improve the capability to predict the performance of the query reformulation; in particular, there are no statistically significant negative correlation and the total number of negative correlations decreases from N = 100 to N = 10,000.
Besides the qualitative aspects, Table 4 reports also the numerical performance comparison between the GAIN as proposed by [21], its version which employs the mean, and the similarity-based gain.

4.3.4. Final Remarks

In this last section of the analysis of the results, we want to briefly summarize our findings. As a remainder, we want to point out that, the GAIN measure proposed by [21], was originally used to estimate the missing information that the user could have gained, by using different subtopic formulations, showed in a user-interface. Although such task shares similar aspects with the one of predicting the recall, they are not fully overlapping. Our main contributions in this paper are:
  • First, adapting an already established technique to a different task. In this sense, to the best of our knowledge, this is the first effort in adapting the GAIN measure proposed by Umemoto et al. [21] to the query formulation recall prediction task.
  • Secondly, its “mean” version, which we refer to as “Mean Gain”, is observed here for the first time, as a better adaptation of [21] to the predictive task.
  • Finally, the Similarity-based Gain is a completely new contribution of this manuscript, which exploits similar elements to the gain measure proposed by Umemoto et al. [21].
Table 4 shows that the similarity based gain has the overall best performance both compared to other gain based measures and traditional predictors (cfr. Table 3). Interestingly, while the original gain worsen with the increase of the cutoff (as observed both in Table 2 and Table 3), both the mean based and the similarity one tend to improve their performance when the cutoff increase. The original gain suffers of the “saturated gain”, as reported in Section 4.3.1, while our proposal (both mean and similarity) improve as new relevant information is added.

5. Conclusions and Future Work

In this paper, we have presented a study that evaluates different definitions of the GAIN of a reformulation for an information need. We adapted the definition of gain proposed by Umemoto et al. [21] to the context of Consumer Health Search, and we used a standard test collection to evaluate our hypotheses: can we use the gain metric to predict the performance of each reformulation? Is there a better formulation that can produce an order of the importance of each reformulation in terms of recall?
We found that for recall based tasks where the number of documents to retrieve may be large, N > 100 , the original definition of GAIN saturates quickly to 1. We proposed an alternative definition that mitigates this problem, and we also presented a similarity based approach that tries to capture the ‘optimal’ query reformulation among all the available formulations of an information need. The analysis of the results confirms that our approach significantly improves the prediction of the order of importance of each reformulation in terms of recall.
In conclusion, the proposed technique is meant to help practitioners in tackling the systematic reviewing task. In this sense, our technique is not meant for a general-purpose query suggestion strategy (more in line with the model proposed by Umemoto et al. [21]). The described approach is meant to boost queries with higher recall, in a context where the practitioner is, in a sense, forced to explore a large number of documents. Assuming that a practitioner needs to review all the documents about a specific topic, it is vital to reduce as possible time spent reviewing documents. Therefore, our technique can help in determining, among a series of queries for the very same information need, which one is more likely to return the most relevant documents. Being able to explore first more promising queries, can greatly speed up the systematic reviewing process. Among our future work, we plan to further investigate the robustness and generalizability of the approach. In particular, we plan to include new collections, topics and formulations. We plan to investigate the news domain through multiple formulations available in the UQV100 collection [1]. We are currently investigating the possibility to smooth the contribution of each reformulation in the similarity matrix S with a locality parameter w. This parameter can be used as an exponent for each element of S and decide whether to get reformulations closer, w < 1 , or push them far away, w > 1 , to create sub-clusters of reformulations and obtain a better prediction.

Author Contributions

Conceptualization, G.M.D.N. and G.F.; methodology, G.M.D.N. and G.F.; software, G.M.D.N. and G.F.; validation, G.M.D.N. and G.F.; formal analysis, G.M.D.N. and G.F.; investigation, G.M.D.N. and G.F.; resources, G.M.D.N. and G.F.; data curation, G.M.D.N. and G.F.; writing—original draft preparation, G.M.D.N. and G.F.; writing—review and editing, G.M.D.N. and G.F.; visualization, G.M.D.N. and G.F.; supervision, G.M.D.N. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The code used is publicly available at: https://github.com/guglielmof/A-Study-of-a-Gain-Based-Approach the data can be found at: https://github.com/CLEFeHealth/CLEFeHealth2018IRtask and the index is available at: https://sites.google.com/view/clef-ehealth-2018/task-3-consumer-health-search accessed on 8 September 2021.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Bailey, P.; Moffat, A.; Scholer, F.; Thomas, P. UQV100: A test collection with query variability. In Proceedings of the 39th International ACM SIGIR Conference on Research and Development in Information Retrieval, Pisa, Italy, 17–21 July 2016; Association for Computing Machinery: New York, NY, USA, 2016; pp. 725–728. [Google Scholar] [CrossRef]
  2. Bailey, P.; Moffat, A.; Scholer, F.; Thomas, P. Retrieval consistency in the presence of query variations. In Proceedings of the 40th International ACM SIGIR Conference on Research and Development in Information Retrieval, Shinjuku, Japan, 7–11 August 2017; pp. 395–404. [Google Scholar]
  3. Benham, R.; Mackenzie, J.; Moffat, A.; Culpepper, J.S. Boosting Search Performance Using Query Variations. ACM Trans. Inf. Syst. 2019, 37, 1–25. [Google Scholar] [CrossRef] [Green Version]
  4. Thomas, P.; Scholer, F.; Bailey, P.; Moffat, A. Tasks, queries, and rankers in pre-retrieval performance prediction. In Proceedings of the 22nd Australasian Document Computing Symposium, Brisbane, QLD, Australia, 7–8 December 2017; Association for Computing Machinery: New York, NY, USA, 2017. [Google Scholar] [CrossRef]
  5. Zendel, O.; Shtok, A.; Raiber, F.; Kurland, O.; Culpepper, J.S. Information needs, queries, and query performance prediction. In Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval, Paris, France, 21–25 July 2019; pp. 395–404. [Google Scholar]
  6. Culpepper, J.S.; Faggioli, G.; Ferro, N.; Kurland, O. Topic Difficulty: Collection and Query Formulation Effects. ACM Trans. Inf. Syst. 2021, 40. [Google Scholar] [CrossRef]
  7. Culpepper, J.S.; Faggioli, G.; Ferro, N.; Kurland, O. Do hard topics exist? A statistical analysis. In Proceedings of the 11th Italian Information Retrieval Workshop, Bari, Italy, 13–15 September 2021. [Google Scholar]
  8. Burgin, R. Variations in relevance judgments and the evaluation of retrieval performance. Inf. Process. Manag. 1992, 28, 619–627. [Google Scholar] [CrossRef]
  9. Sheldon, D.; Shokouhi, M.; Szummer, M.; Craswell, N. LambdaMerge: Merging the results of query reformulations. In Proceedings of the Fourth ACM International Conference on Web Search and Data Mining, Hong Kong, China, 9–12 February 2011; pp. 795–804. [Google Scholar]
  10. Belkin, N.; Kantor, P.; Fox, E.; Shaw, J. Combining the evidence of multiple query representations for information retrieval. Inf. Process. Manag. 1995, 31, 431–448. [Google Scholar] [CrossRef]
  11. Liu, B.; Craswell, N.; Lu, X.; Kurland, O.; Culpepper, J.S. A comparative analysis of human and automatic query variants. In Proceedings of the 2019 ACM SIGIR International Conference on Theory of Information Retrieval, Santa Clara, CA, USA, 2–5 October 2019; Association for Computing Machinery: New York, NY, USA, 2019; pp. 47–50. [Google Scholar] [CrossRef]
  12. Dang, V.; Bendersky, M.; Croft, W.B. Learning to rank query reformulations. In Proceedings of the 33rd International ACM SIGIR Conference on Research and Development in Information Retrieval, Geneva, Switzerland, 19–23 July 2010; Association for Computing Machinery: New York, NY, USA, 2010; pp. 807–808. [Google Scholar] [CrossRef] [Green Version]
  13. Scells, H.; Azzopardi, L.; Zuccon, G.; Koopman, B. Query variation performance prediction for systematic reviews. In Proceedings of the 41st International ACM SIGIR Conference on Research & Development in Information Retrieval, Ann Arbor, MI, USA, 8–12 July 2018; Association for Computing Machinery: New York, NY, USA, 2018; pp. 1089–1092. [Google Scholar] [CrossRef] [Green Version]
  14. Faggioli, G.; Marchesin, S. What makes a query semantically hard? In Proceedings of the 2nd International Conference on Design of Experimental Search & Information REtrieval Systems, 2021, DESIRES ’21, Padua, Italy, 11–15 July 2021. [Google Scholar]
  15. Shtok, A.; Kurland, O.; Carmel, D. Query Performance Prediction Using Reference Lists. ACM Trans. Inf. Syst. 2016, 34. [Google Scholar] [CrossRef]
  16. Roitman, H. An enhanced approach to query performance prediction using reference lists. In Proceedings of the 40th International ACM SIGIR Conference on Research and Development in Information Retrieval, Tokyo, Japan, 7–11 August 2017; Association for Computing Machinery: New York, NY, USA, 2017; pp. 869–872. [Google Scholar] [CrossRef]
  17. Kastner, M.; Straus, S.; Goldsmith, C.H. Estimating the horizon of articles to decide when to stop searching in systematic reviews: An example using a systematic review of RCTs evaluating osteoporosis clinical decision support tools. AMIA Annu. Symp. Proc. 2007, 2007, 389–393. [Google Scholar]
  18. Roegiest, A.; Cormack, G.V.; Grossman, M.R.; Clarke, C.L. TREC 2015 total recall track overview. In Proceedings of the Twenty-Fourth TREC 2015, Gaithersburg, MD, USA, 17–20 November 2015. [Google Scholar]
  19. Grossman, M.R.; Cormack, G.V.; Roegiest, A. TREC 2016 total recall track overview. In Proceedings of the Twenty-Fifth TREC 2016, Gaithersburg, MD, USA, 15–18 November 2016. [Google Scholar]
  20. Jimmy, J.; Zuccon, G.; Palotti, J.R.M.; Goeuriot, L.; Kelly, L. Overview of the CLEF 2018 consumer health search task. In Proceedings of the Working Notes of CLEF 2018—Conference and Labs of the Evaluation Forum, Avignon, France, 10–14 September 2018; Cappellato, L., Ferro, N., Nie, J., Soulier, L., Eds.; CEUR-WS.org 2018; Volume 2125, Available online: http://repository.ubaya.ac.id/37282/ (accessed on 7 September 2021).
  21. Umemoto, K.; Yamamoto, T.; Tanaka, K. ScentBar: A Query suggestion interface visualizing the amount of missed relevant information for intrinsically diverse search. In Proceedings of the 39th International ACM SIGIR Conference on Research and Development in Information Retrieval, Pisa, Italy, 17–21 July 2016; Association for Computing Machinery: New York, NY, USA, 2016; pp. 405–414. [Google Scholar] [CrossRef]
  22. Tsukuda, K.; Sakai, T.; Dou, Z.; Tanaka, K. Estimating intent types for search result diversification. In Information Retrieval Technology; Banchs, R.E., Silvestri, F., Liu, T.Y., Zhang, M., Gao, S., Lang, J., Eds.; Springer: Berlin/Heidelberg, Germany, 2013; pp. 25–37. [Google Scholar]
  23. Cronen-Townsend, S.; Zhou, Y.; Croft, W.B. Predicting query performance. In Proceedings of the 25th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, Tampere, Finland, 11–15 August 2002; pp. 299–306. [Google Scholar]
  24. Hauff, C.; Hiemstra, D.; de Jong, F. A survey of pre-retrieval query performance predictors. In Proceedings of the 17th ACM Conference on Information and Knowledge Management, Napa Valley, CA, USA, 26–30 October 2008; pp. 1419–1420. [Google Scholar]
  25. He, B.; Ounis, I. Inferring query performance using pre-retrieval predictors. In Proceedings of the International Symposium on String Processing and Information Retrieval, Padova, Italy, 5–8 October 2004; pp. 43–54. [Google Scholar]
  26. Mothe, J.; Tanguy, L. Linguistic features to predict query difficulty. In Proceedings of the ACM Conference on Research and Development in Information Retrieval, SIGIR, Predicting Query Difficulty-Methods and Applications Workshop, Salvador, Brazil, 15–19 August 2005; pp. 7–10. [Google Scholar]
  27. Scholer, F.; Williams, H.E.; Turpin, A. Query Association Surrogates for Web Search. J. Assoc. Inf. Sci. Technol. 2004, 55, 637–650. [Google Scholar] [CrossRef]
  28. Zhao, Y.; Scholer, F.; Tsegay, Y. Effective pre-retrieval query performance prediction using similarity and variability evidence. In Proceedings of the European Conference on Information Retrieval, Glasgow, UK, 30 March–3 April 2008; pp. 52–64. [Google Scholar]
  29. Aslam, J.A.; Pavlu, V. Query hardness estimation using jensen-shannon divergence among multiple scoring functions. In Proceedings of the European Conference on Information Retrieval, Lisbon, Portugal, 14–17 April 2007; pp. 198–209. [Google Scholar]
  30. Roitman, H. Query performance prediction using passage information. In Proceedings of the 41st International ACM SIGIR Conference on Research & Development in Information Retrieval, Ann Arbor, MI, USA, 8–12 July 2018; pp. 893–896. [Google Scholar]
  31. Zamani, H.; Croft, W.B.; Culpepper, J.S. Neural query performance prediction using weak supervision from multiple signals. In Proceedings of the 41st International ACM SIGIR Conference on Research & Development in Information Retrieval, Ann Arbor, MI, USA, 8–12 July 2018; pp. 105–114. [Google Scholar]
  32. Zhou, Y.; Croft, W.B. Ranking robustness: A novel framework to predict query performance. In Proceedings of the 15th ACM International Conference on Information and Knowledge Management, Kansas City, VA, USA, 6–11 November 2006; pp. 567–574. [Google Scholar]
  33. Carmel, D.; Yom-Tov, E.; Darlow, A.; Pelleg, D. What makes a query difficult? In Proceedings of the 29th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, Seattle, DC, USA, 6–11 August 2006; pp. 390–397. [Google Scholar]
  34. Cummins, R. Document Score Distribution Models for Query Performance Inference and Prediction. ACM Trans. Inf. Syst. 2014, 32, 2:1–2:28. [Google Scholar] [CrossRef]
  35. Diaz, F. Performance prediction using spatial autocorrelation. In Proceedings of the 30th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, Amsterdam, The Netherlands, 23–27 July 2007; pp. 583–590. [Google Scholar]
  36. Amati, G.; Carpineto, C.; Romano, G. Query difficulty, robustness, and selective application of query expansion. In Proceedings of the European Conference on Information Retrieval, Lisbon, Portugal, 14–17 April 2004; pp. 127–137. [Google Scholar]
  37. Cronen-Townsend, S.; Zhou, Y.; Croft, W.B. A Language Modeling Framework for Selective Query Expansion; Technical Report; Center for Intelligent Information Retrieval, University of Massachusetts: Amherst, MA, USA, 2004. [Google Scholar]
  38. Zhou, Y.; Croft, W.B. Query performance prediction in web search environments. In Proceedings of the 30th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, Amsterdam, The Netherlands, 23–27 July 2007; pp. 543–550. [Google Scholar]
  39. Shtok, A.; Kurland, O.; Carmel, D.; Raiber, F.; Markovits, G. Predicting Query Performance by Query-Drift Estimation. ACM Trans. Inf. Syst. 2012, 30, 1–35. [Google Scholar] [CrossRef] [Green Version]
  40. Tao, Y.; Wu, S. Query performance prediction by considering score magnitude and variance together. In Proceedings of the 23rd ACM International Conference on Conference on Information and Knowledge Management, Shanghai, China, 3–7 November 2014; pp. 1891–1894. [Google Scholar]
  41. Faggioli, G. Enabling performance prediction in information retrieval evaluation. In Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval, Virtual Event Canada, 11–15 July 2021; Association for Computing Machinery: New York, NY, USA, 2021; p. 2701. [Google Scholar] [CrossRef]
  42. Chifu, A.G.; Laporte, L.; Mothe, J.; Ullah, M.Z. Query performance prediction focused on summarized letor features. In Proceedings of the 41st International ACM SIGIR Conference on Research & Development in Information Retrieval, Ann Arbor, MI, USA, 8–12 July 2018; pp. 1177–1180. [Google Scholar]
  43. Shtok, A.; Kurland, O.; Carmel, D. Using statistical decision theory and relevance models for query-performance prediction. In Proceedings of the 33rd International ACM SIGIR Conference on Research and Development in Information Retrieval, Geneva, Switzerland, 19–23 July 2010; pp. 259–266. [Google Scholar]
  44. Faggioli, G.; Zendel, O.; Culpepper, J.S.; Ferro, N.; Scholer, F. An enhanced evaluation framework for query performance prediction. In Advances in Information Retrieval; Hiemstra, D., Moens, M.F., Mothe, J., Perego, R., Potthast, M., Sebastiani, F., Eds.; Springer International Publishing: Cham, Switzerland, 2021; pp. 115–129. [Google Scholar]
  45. Zendel, O.; Culpepper, J.S.; Scholer, F. Is query performance prediction with multiple query variations harder than topic performance prediction? In Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval, Virtual Event Canada, 11–15 July 2021; Association for Computing Machinery: New York, NY, USA, 2021; pp. 1713–1717. [Google Scholar]
Figure 1. An example of clusters of subtopics and aspects.
Figure 1. An example of clusters of subtopics and aspects.
Applsci 11 09075 g001
Figure 2. The three rows show the results for the measures computed at N = 100, 1000, and 10,000, respectively. The three columns represent results of the GAIN as proposed in [21], the mean aggregation and the similarity-based aggregation strategy. Each subfigure shows (top) the distribution of the GAIN, ordered increasingly, of the 350 queries and (bottom) correlation between the reformulations ordered by predicted GAIN (or similarity) and the reformulations ordered by the true recall.
Figure 2. The three rows show the results for the measures computed at N = 100, 1000, and 10,000, respectively. The three columns represent results of the GAIN as proposed in [21], the mean aggregation and the similarity-based aggregation strategy. Each subfigure shows (top) the distribution of the GAIN, ordered increasingly, of the 350 queries and (bottom) correlation between the reformulations ordered by predicted GAIN (or similarity) and the reformulations ordered by the true recall.
Applsci 11 09075 g002aApplsci 11 09075 g002b
Table 1. Pre- and Post-retrieval predictive baseline models considered.
Table 1. Pre- and Post-retrieval predictive baseline models considered.
TypePredictorDescription
pre-retrievalmax-idf [27]It considers the maximum value of the idf (inverse document frequency) over the query terms
mean-idf [37]It computes the mean value of the idf over the query terms
std-idf [37]It uses the standard deviation of the idf over the query terms
sum-scq [28]Measures similarity based on cf.idf to the corpus, summed over the query terms.
mean-scq [28]It relies on the same value of sum-scq, but it normalizes it with the length of the query
max-scq [28]It relies on the same value of sum-scq, but considers only the maximum value
post-retrievalwig [38]Standard deviation of the top documents scores in the retrieval list.
nqc [39]Difference between the mean retrieval score of the top documents, scaled by the score of the entire corpus
smv [40]It computes the prediction considering the standard deviation of the retrieval scores
Table 2. Kendall’s τ correlation observed between recall and prediction scores for both pre- and post-retrieval traditional predictors, if we compare the default formulations of different topics. Results are in line with correlation values previously observed in other scenarios. The symbol indicates that the correlation is statistically greater than 0 at level α = 0.05 , while the indicates a significance level of 0.01 , the absence of any symbol indicates that results cannot be deemed statistically greater than 0. We compute the Kendall’s τ correlation at different cutoff levels of the ranked lists (100, 1000, and 10,000).
Table 2. Kendall’s τ correlation observed between recall and prediction scores for both pre- and post-retrieval traditional predictors, if we compare the default formulations of different topics. Results are in line with correlation values previously observed in other scenarios. The symbol indicates that the correlation is statistically greater than 0 at level α = 0.05 , while the indicates a significance level of 0.01 , the absence of any symbol indicates that results cannot be deemed statistically greater than 0. We compute the Kendall’s τ correlation at different cutoff levels of the ranked lists (100, 1000, and 10,000).
TypePredictorKendall’s τ
100 1000 10,000
pre-retrievalmax-idf0.3185 0.3260 0.2875
mean-idf0.2996 0.3218 0.2555
std-idf0.2947 0.2989 0.2343
sum-scq0.2637 0.2581 0.1739
mean-scq0.3479 0.3652 0.3299
max-scq0.3502 0.3724 0.2833
post-retrievalwig0.3029 0.3218 0.2882
nqc0.2865 0.19110.1135
smv0.17970.13320.0229
Table 3. Performance achieved by traditional predictors, applied to our specific case. Each predictor has been used to predict the performance of the different formulations. We report the mean score and standard deviation of the correlation computed over the different topics. We also report the first quartile, third quartile and number of topics (over the 50 available) for which the correlation between the predicted and observed recalls for their (re)formulations is significantly greater than 0.
Table 3. Performance achieved by traditional predictors, applied to our specific case. Each predictor has been used to predict the performance of the different formulations. We report the mean score and standard deviation of the correlation computed over the different topics. We also report the first quartile, third quartile and number of topics (over the 50 available) for which the correlation between the predicted and observed recalls for their (re)formulations is significantly greater than 0.
Kendall’s τ
Type Predictor Cutoff Q1 Mean (Std) Q3 Sign.
pre-retrievalmax-idf100−0.5417−0.1085 (0.4825)0.11833
1000−0.5295−0.0973 (0.4755)0.22631
10,000−0.5699−0.1227 (0.4628)0.15842
mean-idf100−0.4214−0.0449 (0.5272)0.33334
1000−0.4821−0.0617 (0.4898)0.21674
10,000−0.4214−0.0549 (0.4698)0.24733
std-idf100−0.4880−0.1606 (0.4927)0.09154
1000−0.4190−0.0999 (0.5107)0.19385
10,000−0.4064−0.1537 (0.4347)0.15761
sum-scq100−0.2381−0.0102 (0.4021)0.23811
1000−0.29850.0893 (0.4276)0.40004
10,000−0.31260.0150 (0.4558)0.27505
mean-scq100−0.32500.0322 (0.5231)0.39016
1000−0.38980.0135 (0.4838)0.33335
10,000−0.32500.0005 (0.4505)0.29853
max-scq100−0.3541−0.0369 (0.4333)0.27651
1000−0.3341−0.0312 (0.4447)0.25682
10,000−0.3459−0.0484 (0.4441)0.19122
post-retrievalwig100−0.6790−0.1743 (0.5206)0.13764
1000−0.4088−0.0266 (0.5031)0.25196
10,000−0.42140.0171 (0.5185)0.38496
nqc100−0.5611−0.0880 (0.5554)0.23816
1000−0.4405−0.1244 (0.5004)0.15114
10,000−0.4850−0.1539 (0.4991)0.15393
smv100−0.5621−0.1653 (0.4836)0.18491
1000−0.5542−0.1626 (0.4604)0.18590
10,000−0.6243−0.2207 (0.4882)0.09942
Table 4. Kendall’s τ correlation observed for the task of predicting the query formulation recall, using the similarity based approaches. In bolt, best mean score for each cutoff. Note that all the methods considered perform better than traditional predictors(cft Table 3). We also have a higher number of significant rankings compared to the one observed before.
Table 4. Kendall’s τ correlation observed for the task of predicting the query formulation recall, using the similarity based approaches. In bolt, best mean score for each cutoff. Note that all the methods considered perform better than traditional predictors(cft Table 3). We also have a higher number of significant rankings compared to the one observed before.
Kendall’s τ
Predictor Cutoff Q1 Mean (Std) Q3 Sign.
Original Gain [21]1000.00000.3422 (0.4696)0.683115
10000.00000.3783 (0.3527)0.660913
10,0000.00000.1600 (0.2606)0.47652
Mean Gain1000.14560.3822 (0.4936)0.732016
10000.24170.4042 (0.4796)0.732017
10,0000.27090.5111 (0.3984)0.800019
Similarity based Gain1000.14290.3768 (0.4636)0.658113
10000.30830.5069 (0.3505)0.714317
10,0000.25210.5443 (0.3930)0.887624
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Di Nunzio, G.M.; Faggioli, G. A Study of a Gain Based Approach for Query Aspects in Recall Oriented Tasks. Appl. Sci. 2021, 11, 9075. https://0-doi-org.brum.beds.ac.uk/10.3390/app11199075

AMA Style

Di Nunzio GM, Faggioli G. A Study of a Gain Based Approach for Query Aspects in Recall Oriented Tasks. Applied Sciences. 2021; 11(19):9075. https://0-doi-org.brum.beds.ac.uk/10.3390/app11199075

Chicago/Turabian Style

Di Nunzio, Giorgio Maria, and Guglielmo Faggioli. 2021. "A Study of a Gain Based Approach for Query Aspects in Recall Oriented Tasks" Applied Sciences 11, no. 19: 9075. https://0-doi-org.brum.beds.ac.uk/10.3390/app11199075

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop