Next Article in Journal
Metrological Evaluation of Human–Robot Collaborative Environments Based on Optical Motion Capture Systems
Next Article in Special Issue
Influence of Features on Accuracy of Anomaly Detection for an Energy Trading System
Previous Article in Journal
Analysis of Gender Differences in HRV of Patients with Myalgic Encephalomyelitis/Chronic Fatigue Syndrome Using Mobile-Health Technology
Previous Article in Special Issue
Modeling of Recommendation System Based on Emotional Information and Collaborative Filtering
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Criteria Recommendation Systems to Foster Online Grocery

by
Manar Mohamed Hafez
1,*,
Rebeca P. Díaz Redondo
2,
Ana Fernández Vilas
2 and
Héctor Olivera Pazó
2
1
College of Computing and Information Technology, Arab Academy for Science, Technology and Maritime Transport (AASTMT)-Smart Village, Giza P.O. Box 12676, Egypt
2
AtlanTTic, Information & Computing Lab, Universidade de Vigo, 36310 Vigo, Spain
*
Author to whom correspondence should be addressed.
Submission received: 15 April 2021 / Revised: 12 May 2021 / Accepted: 24 May 2021 / Published: 28 May 2021
(This article belongs to the Collection Intelligent Security Sensors in Cloud Computing)

Abstract

:
With the exponential increase in information, it has become imperative to design mechanisms that allow users to access what matters to them as quickly as possible. The recommendation system (RS) with information technology development is the solution, it is an intelligent system. Various types of data can be collected on items of interest to users and presented as recommendations. RS also play a very important role in e-commerce. The purpose of recommending a product is to designate the most appropriate designation for a specific product. The major challenge when recommending products is insufficient information about the products and the categories to which they belong. In this paper, we transform the product data using two methods of document representation: bag-of-words (BOW) and the neural network-based document combination known as vector-based (Doc2Vec). We propose three-criteria recommendation systems (product, package and health) for each document representation method to foster online grocery shopping, which depends on product characteristics such as composition, packaging, nutrition table, allergen, and so forth. For our evaluation, we conducted a user and expert survey. Finally, we compared the performance of these three criteria for each document representation method, discovering that the neural network-based (Doc2Vec) performs better and completely alters the results.

1. Introduction

According to [1], digital transformation facilitates new ways of value creation at all stages of the consumer decision process: pre-purchase (need recognition, information search, consideration or evaluation of alternatives), the purchase (choice, ordering, payment), and the post-purchase (consumption, use, engagement, service requests). This value creation is especially relevant in retailing to ensure competitiveness and gain a larger market share. Digital transformation came hand in hand with the penetration of mobile devices and data science in e-commerce. Although digital transformation [2] has been addressed from several approaches; multi-channel solutions, user modeling, Internet of Things, and so forth; all of them rely to some extent on the availability of information on operations, supply chains and consumer and shopper behaviors. One of the imperatives in this digital transformation is obtaining a view of customer insights.
From the early steps (Amazon, 2003 [3]), the time to select the desired product has been the main issue for customers, especially if the high volume and rhythm of incorporation of products are considered. From more than two decades, Recommender Systems (RS) in e-commerce have tried to provide the most suitable products of services, to mitigate the product overload problem and to narrow down the set of choices [4,5,6]. Success of major products & service providers mainly relies on RS, such as Amazon [3], Netflix [7], and Google [8]. RSs improve customer satisfaction by reducing customer search efforts and as a consequence, they increase product/service sales. RSs provide users with items based on their interests, the preferences of other users and the item attributes. The recommendation can be carried out with several approaches depending on the type of data collected and the ways it is used by the RS: Content-Based (CB) filtering, Collaborative Filtering (CF), and hybrid. Both systems CB and CF are widely used, and specially the item-based collaborative filtering where the similarity between items is calculated using users’ ratings of those items. (developed by Amazon [3]).
Although RSs are used by users regularly in almost all digitalized sectors, its popularization in the grocery market, that is, a retail store that primarily sells food products, has been delayed as a consequence of the low penetration of online grocery shopping, the implementation of e-commerce for grocery goods. Recently, as well as in other sectors, the grocery industry is harnessing digital to innovate through data-drive business models. Online grocery is considered a central element in the new normal. In this respect, grocery recommendation uses customer’s shopping history and product information to address various added value scenarios; predicting customers’ future shopping, selecting best value for money products, offering new products user may like, and so forth. Besides, the availability of data about products and shopping positively affects the retailer by easing a sustainable business; offers & featured products, stock management, customer profiling, and so forth.
To meet the challenges above, in this paper, we use two document representation methods—BOW and Doc2Vec—to manage product data. We also address the three-criteria recommendation systems; Product, Package and Health for each document representation model to the specific problem of, given a source product P, applying RSs to suggest similar alternative products where similarity is defined on the basis of a product taxonomy, as well as product characteristics; composition, packaging, nutrition table, allergens, etc. The solution to this problem supports various regular use cases in the grocery market, such as out of stock products, inventory clearance, best value options, new products, etc. In order to obtain the recommender model and to validate them, we use a real grocery dataset, referred to as MDD-DS, provided by Midiadia, a Spanish company that works on grocery catalogs. MDD-DS was constructed by analyzing the product’s information (product labeling) and by experts’ manual annotation so that products are assigned to a specific variety in a hierarchical structure for products. Therefore, the major contributions of this research work are the following:
  • Definition of an appropriate data structure to manage the different kinds of information linked to commercial products (especially in the food industry).
  • Definition and identification of the appropriate document representation that works with MDD-DS to represent the products.
  • Design and implementation of a RS that automatically provides alternative products when the user’s choice is not available. The RS do not work with user’s profile, it is exclusively based on the product’s characteristics and the available catalogue.
  • Design of three recommendation approaches based on the product’s characteristics; composition, packaging, nutritional table, allergens, etc.
  • Proof of concept and validation to test the RS performance. We have conducted a survey for users and for experts to evaluate the RS approaches.
The rest of this document is organized as follows: In Section 2, we briefly reviewed RS and document representation methods to manage product data in RS. The grocery MDD-DS is describing in Section 3. In Section 4, the recommendation methodology is introduced with three specific approaches to product similarity, based on product composition, packaging, and healthy characteristics. To implement these three approaches to product similarity, we deployed two kinds of document representation techniques: a simple BOW (Bag of Words, in Section 5) and a neural network-based word embedding, Doc2Vec in Section 6. For the two product representation models, experimental evaluation and discussion are described in Section 7. Finally, in Section 8, we conclude the current work with some future research directions.

2. Recommender Systems

RS are a fundamental task for e-commerce, as the personal RS recommends providing items or products that satisfy the interests of different users according to their different interests and also recommends unknown items for the users that satisfy their interests [9]. As mentioned above, the three most commonly used methods in the RS are CB filtering, C F , and hybrid approach.
CB filtering [10,11,12] is one of the standard techniques used by RS. CB identifies items based on an analysis of the item’s content, similar to items known to be of interest to the user. For example, a CB website recommendation service can work by analyzing the user’s favorite web pages to generate a profile of commonly occurring terms. Then use this profile to find other web pages that include some or all of these terms.
CB technique has several issues and limitations [13,14,15]. For example, (i) having no mechanism to assess the quality of an item supported by CB methods. Furthermore, CB methods generally require items to include some type of content that is amenable to feature extraction algorithms. As a result, CB technique tend to be ill-suited for recommending products, movies, music titles, authors, restaurants and other types of items with little or no useful and analyzable content; (ii) CB is also have another problem that they rarely reflect current user community preferences. In a technique that recommends products to users, for example, there is no mechanism to favor items that are currently “hot sellers”. Moreover, existing systems do not provide a mechanism to recognize that the user can search for a particular type or category.
CF [16,17] is another common recommendation technique. In general, the CF recommends the item to the user based on a community of user interests, without any analysis of the item content. CF idea is to build a personal profile of ratings data through each item sold and rate it through the user. Besides the CF technique’s concept to recommend the item to the user, the user’s profile is initially compared with other users’ profiles to identify one or more similar users. These similar users’ highly-rated items are recommended to the user. A significant benefit of CF is that it overcomes the previously mentioned shortcomings of CB filtering.
The main issue in the above is how to measure user similarity. This problem inspires memory-based methods [18], which can be implemented as user-based [19] or item-based [20,21]. User and item-based methods have similar mechanisms, but item-based methods are used more to perform better at scale and with a lower rating density.
A hybrid approach is an approach that combines CB and CF (user-based and item-based) filtration approaches with attempts to eliminate their flaws and provides a more efficient result. It usually performs better than either filtering method alone. Here, the hybrid approach does combine the CB and CF to solve the significant problems that are the cold start [22] and sparsity problems [23]. The cold start problem occurs when there is not enough new user data or ratings for a new item, so it is difficult to make recommendations for that new user or present new items to a user. Regarding sparsity, it occurs when the user has not rated most of the items and the ratings are sparse.
In our work, we have some issues in providing a recommendation service and associated methods for generating personalized items. Science, the recommendation is based on the user’s interests without considering the user profile. This paper focuses solely on the user’s interest and how to recommend suitable items to each user. The benefit of this work is also that recommended items are identified by lists of similar items to the desired item. As mentioned above, in our paper we worked on combining CB filtering and CF (item-to-item), such as Amazon [3]. Amazon invented an algorithm that began looking at items themselves. It analyzes the recommendations through the items purchased or rated by the user and matches them with similar items, using metrics and composing a list of recommendations. That algorithm is called “item-based collaborative filtering”. This approach was also very appropriate and faster, especially for huge data sets. It was developed in 2017 [24], to aggregate data about the user to develop an RS to rely on the data and the user behavior in selecting the items. It is still based only on the analysis of the items. However, it combines the analysis of the items with the user’s data and choices. Regarding the related works, we see that the most widely used in the previous works is collaborative filtering, as shown in the following paragraphs.
In [25], the authors used a collaborative filtering method to create the proposal for various items using accessible ratings and comments on Twitter. The authors have also evaluated the reviews given by blipper (a review website) for four unique products using the CF method. When dealing with video as data to find suitable items for the user, there are also research works that apply collaborative filtering to recommend products through this kind of data. For instance, in [26], the authors introduced an approach that includes item-to-item collaborative filtering to discover exciting and meaningful videos among the large-scale videos. This method runs on Qizmt, which is a.NET MapReduce framework. The RS in [27] also depends on monitoring the video content the user watches, the customer carrier database, and the vector database of products; therefore, the idea is to identify an item related to a part of the video content the user viewed that, and consequently determine the product category associated with the item, then analyze the characteristics of items similar to the item. That has been identified through the video’s visualization, and it compares the customer value vectors and the product characteristics vectors. Moreover, start showing the recommended product to the customer. Other approaches take user interactions into account to recommend the right products. For instance, in [28], the recommender system collaborative filtering uses user interactions and keeps them to benefit the recommendation. It does not stop at the items that have been selected only from the users, but the proposed system is related to the category of items.
Recommendation systems usually require a large amount of user data. Safeguarding the privacy of this information is an important aspect that must be taken into account. For instance, in [29], an arbitrable remote data auditing scheme is proposed. This is based on a non third-party auditor for the network storage-as-a-service paradigm. The authors have designed a network storage service system based on blockchain, in which the user and the network storage service provider will generate the integrity metadata of the corresponding original data block respectively. All of that reach a consensus on the matter by means of the use of the blockchain technique.
Other approaches solve some problems in the recommendation system, such as scalability and the cold start problem. For instance, the authors of [30] implement a user-based collaborative filtering algorithm on a distributed cloud computing platform that is Hadoop to solve the scalability problem of the collaborative filtering method. Besides, the authors of [31] propose a keyword-Aware Service Recommendation method called KASR. They also present a personalized service recommendation list and keywords used to indicate user preferences. A user-based collaborative filtering algorithm is adopted to generate the recommendations. They implemented KASR on Hadoop with real-world data sets to improve its scalability and efficiency in a big data environment. Furthermore, in [32] proposed a novel approach based on item-based C F use of BERT [33] to help understand the items and work to show the connections between the items and solve problems that are related to the traditional recommender system as cold start. This experiment was performed with an actual data set large scale with a whole cold start scenario, and this approach has overtaken the popular Bi-LSTM model. It used the item title as content along with the item token to solve the cold start problem. The approach also further identifies the interests of the user. Other approaches consider recommending products that are in line with the user’s interests without being affected by the problems faced by the recommendation system mentioned above and the problem of data sparsity. For instance in [34], a product recommendation system is proposed where an autoencoder based on a collaborative filtering method is employed. The experiment result shows a very low Root Mean Squared Error (RMSE) value, considering that the users’ recommendations are in line with their interests and are not affected by the data sparsity problem as the datasets are very sparse.
In e-commerce, user data and purchasing behavior play an important role [35,36]. However, in our scenario we are totally agnostic about the customer behavior. The company Midiadia does not provide complete e-commerce solutions, but provides enriched catalogues to e-commerce platforms. Consequently, Midiadia has not information about the customers interactions, habits or any kind of profiling. To the best of our knowledge, no other study provides a solution to this problem (recommending a similar product) taking exclusively into account the product information: ingredients, size, packaging, health messages, allergens, etc. All this consideration without going back to the customer data, depends only on the product description, such as name, brand, ingredients, legal name, and size; likewise, other data helps to know that the product is also healthy, such as sugars, fats, carbohydrates and excluding all the contents that can cause allergies. Our proposition fills an exciting void for many e-commerce dominants.

Representation Models

Regarding document representation models, we provide some representation models regarding the techniques used in this paper. We start with simple techniques such as Bag-Of-Words, TF-IDF. First, Bag-Of-Words (a.k.a. BOW [37,38]) is a basic, popular, and most straightforward approach among all other feature extraction methods. It is used to create document representations in Natural Language Processing (NLP) [39] and Information Retrieval (IR) [40]. The text is represented as a bag that contains many words. It forms a word presence feature set from all the words of an instance. The method does not care how often the word appears or the order of the words; the only thing that matters is whether the word is in the word list. It is generally used to extract features from text data in various ways. A bag of words is the presentation of text data. It specifies the frequency of words in the document. A feature generated by bag-of-words is a vector where n is the number of words in the input documents vocabulary. Second, TF-IDF [41] short for term frequency–inverse document frequency, is a technique that can be used as a weighting factor not only in IR solutions but also in text mining and user modeling. This method, as in the bag-of-words model, counts how many times a word appears in a document. However, words which are repeated so many times like the stopwords (the, of, …) are penalized with this technique because of the inverse documentary frequency weighting. Here, the more documents a word appears in, the less relevant it is. Therefore, a word that is distinctive and frequent will be high-ranked if it appears in the query introduced by the user.
On the other hand, word embedding is a term used for the representation of words for text analysis [42,43,44,45]. It also maps of words in vectors of real numbers using the neural network, the probabilistic model, or the dimension reduction on the word co-occurrence matrix. Word embeddings are also very useful in mitigating the curse of dimensionality, a recurring problem in artificial intelligence [46]. Without word embedding, the unique identifiers representing the words generate scattered data, isolated points in a vast sparse representation [47]. With word embedding, on the other hand, the space becomes much more limited in terms of dimensionality with a widely richer amount of semantic information [48]. With such numerical features, it is easier for a computer to perform different mathematical operations like matrix factorization, dot product, and so forth, which are mandatory to use shallow and deep learning techniques.
Regarding word embedding, unfortunately, the representation of meaning with different symbols cannot orchestrate the same meaning as words. Early attempts solved this problem by clustering words based on the meaning of their endings and representing the words as high-dimensional spaced vectors. A new idea was recently proposed inspired by the neural network language model, and the model proposed is known as Word to Vector (word2vec) [49]. These embeddings are easy to work with since the vectors can be manipulated by many algorithms like dimensionality reduction, clustering, classification, similarity searching, and many more.
Two models generate the representation of word2vec have been presented in order to produce such dense word embeddings: the Continuous Bag of Word (CBOW) model [50] and the Skip-Gram model [51,52]. Each of the two models train a network to predict neighboring words. Suppose that a sequence of tokens ( t 1 , , t n ) is provided. The CBOW model, first randomly initializes the vector of each word and then using a single layer neural network whose outcome is the vector of the predicted word, optimizes the original guesses. One can easily understand that the size of the Neural Network controls the size of the word vector. The Skip-gram model uses the word, in order to predict the context words. After explaining the meaning of Word2Vec, however, the goal of doc2vec is to create a digital representation of the document, regardless of its length. Unlike words, documents don’t come in logical structures like words. In [53] they used Word2Vec template and added below paragraph id to build doc2vec.

3. Dataset

The data set used in this paper was provided by Midiadia, a Spanish company which works to convert textual information in the product package into product category and product attributes by mixing automated natural language processing and manual annotation. The Midiadia DataSet (MDD-DS) is taxonomy where the 3 upper levels are called Category, Subcategory and Variety. Every product in MDD-DS includes; the taxonomy position, that is, values for Category, Subcategory and Variety as well as a set of product attributes. for example, name, ingredients, legal name, brand, product size, and so forth, as shown the extract of real data in Table 1. We have also used these product components before in [54,55] to provide a solution to automatically categorize the constantly changing products in the market, which is the first part of our investigation.
  • ’European Article Number’ (EAN) is an internationally recognized standard that describes the barcode and numbering system used in world trade to identify a specific product that is specifically packaged and has a specific manufacturer in retail.
  • ’Category’, ’Subcategory’, and ’Variety’ are a hierarchy and can be displayed by a company as catalog organization levels in the classification. The companies manufacture the products and each company has an identifying name and is listed as the brand.
  • In addition, there are some properties compatible with the EU regulation [56], for example, name, legal name and ingredients, as indicated in Table 2.
  • ’Servings’ is a number that determined based on the amount of product and is sufficient for how many people.
In addition, Midiadia supported us with two versions of MDD-DS to implement recommendation systems and cover all the company’s requirements. The basic version which was called MDD-DS1, contained all the above information plus some information related to the nutrition table, such as sugar and fat, and some messages on the product packaging such as the sugar-free or the free gluten and other messages on the cover of the product. Of course, these messages are placed according to the components of each product, as shown in Table 3.
The extended version which was named MDD-DS2, contained all the above information besides the characteristics of the Brand Type and Brand attributes, and the price was also added randomly besides more information about the nutrition table such as carbohydrates ( C a r b s ), dietary fiber ( d f ) , and a percentage of saturated fat ( s f ) and good fat ( g f ) , protein ( p n ) and salt ( s a ) . It also contains allergens such as soy, fish, eggs, nuts, and so forth, as characteristics that will be mentioned in detail and how they are used in our research, as shown in Table 4.
  • ’Carbohydrates’ are considered one of the three main food categories and a source of energy, and they are also basically sugars and starches that the body breaks down into glucose (we can say that it is a simple sugar that the body can use to nourish its cells).
  • ’Dietary Fiber’ is part of the food that has been separated from plants and cannot be completely broken down by human digestive enzymes.

4. Methodology Overview

Taking into account our dataset, the proposed recommender system does not have information about a user’s history so that C F should be excluded. An hybrid item-based C F is designed for the specific scenario of finding similar products to a source product P where similarly will be defined according to, first, the  V a r i e t y of the product in the MDD taxonomy and, second, other attributes of the product. The alternative product to P will be a product in the same V a r i e t y which moreover meets other similarity requirements over the product attributes. Three similarity approaches have been defined: (i) Product Composition (PRO-COM), where similarity is scored according to product composition (ingredient, name, legal name, etc.); (ii) Package-based (PK-BD), where similarity is scored according to the size of the product chosen by the user; and (iii) Health-based (HTH-BD), where similarity is scored according to a healthy grade by using the product nutrition table. The recommendation methodology considers allergens apart from these three similarity approaches as follows. In MDD-DS, several product attributes are related to allergens: (Nuts, egg, hazelnuts, fish, sulfates, peanuts, mollusks, lupine, gluten, mustard, soy, crustaceans, milk and its derivatives including lactose, sunflower seeds and sesame). Allergens are considered pre-conditions for suggesting an alternative product, that is, if the user-chose a product which includes sugar, water and nuts), the allergen precondition for the alternative products is possibly containing nuts but not other allergen. So, the alternative product may contain nuts or not, but it should not contain other allergens.
The proposed methodology is shown in Figure 1. In order to obtain the model, a training set is defined in order to obtain the recommender model with the following steps: (A) MDD-DS is preprocessed; (B) for every product P the dataset is filtered by allergen preconditions; (C) the three similarity scores are obtained (PRO-COM, PK-BD, and HTH-BD). Then at the bottom of the model is the automated recommendation when the user selects the product. The recommendation system recommends an alternative based on the three approaches. A survey is conducted to consider the users in the three approaches.
To implement the recommendation, we carried out collaborative filtering as a first model. Then we add more features and a neural network solution to improve our results. The Figure 2, illustrate the strategy of this paper. First model, The dataset (MDD-DS1) is analyzed by preprocessing. Three approaches were then developed, which are (PRO-COM, PK-BD, and HTH-BD) by collaborative filtering. A survey is carried out to take the users’ opinions in the three approaches.
In the second model, the approaches are redeveloped based on user feedback. We added more features and more filters, such as filtering by allergen features. We added a neural network solution to improve our results. Therefore, the company extends the dataset—called (MDD-DS2)—to contain additional features to develop the approaches, so the data are analyzed through preprocessing. A neural network is built on the products. Then, it extracts the product as a vector and compares it to the rest of the products using similarity techniques and then makes the approaches (PRO-COM, HTH-BD, and PK-BD).
All three approaches take allergens’ features into account, which means that, as explained above, if the product is, for example, nut-free, the alternative products are too. Then the approaches are sent to an expert by the company for evaluation. This has indicated that the modification is suitable for the company’s requirements. Hence, a questionnaire was published for users to evaluate the recommendation system after these modifications. Finally, a comparison of the evaluation of the users was carried out.

5. Recommendation System Based on Item-Based Collaborative Filtering (RS-CF)

This paper proposes a methodology to develop RS-CF for the retail sale of products. Three recommendation methods have been developed, each of which recommends alternative products to help the user obtain the product of interest. Our solution implementation takes the data (shown in the Section 3) for each approach as the input control variables. Alternative products are then recommended for each approach and then presented to the user to choose the right product for him and evaluate RS-CF. The modeling methodology consists of 2 main steps as show in Figure 3: (A) data pre-processing; (B) build the RS-CF approach; the RS-CF was done in three ways, namely: (i) Product Composition (PRO-COM) approach, where similarity is scored according to product component (ingredient, name, legal name, etc.); (ii) Package-based (PK-BD) approach, where similarity is scored based on the PRO-COM result besides the size of the product chosen by the user; and (iii) Health-Based (HTH-BD) approach, where the similarity is scored according to the PRO-COM result and taking into account that the allergen information is being considered along with a healthy degree using the product nutrition table. In order to evaluate the RS-CF approaches by the user, we conducted a survey that includes many of the products and similar products.
It begins with PRO-COM and PK-BD approach to performing preprocessing and building the product matrix. The algorithm in both approaches will be explained similarly, with the added feature of the PK-BD approach.

5.1. Preprocessing

The PRO-COM and PK-BD approaches are the first two RS-CF approaches that use the first data release (MDD-DS1). MDD-DS1 comprises 29,236 products, so the data was processed and cleaned by removing the empty rows from the variable name and serving as well. The EAN column is then scanned for duplicates and removed. Next, we ignore that the variety contains fewer than four products; this number is required to implement the algorithm; the main focus of this investigation is when the primary product is not found. Therefore, keep at least three alternatives of the same variety. Therefore, the number of products after cleaning the data is 20,371, we mentioned the last steps in Algorithm 1 as a C l e a n i n g (MDD-DS1) step. The data is then preprocessed by extracting all the words for the attributes name, legal name, and ingredients. Consider a Corpus C of each product p, C(p[Name], p[Legal Name], p[Ingredients]). That means that the three attributes are combined in a single text to describe the product d e s ( p ) . This description was obtained ( d e s ( p ) ) after cleaning the product C l e a n _ p by following these steps: (i) transform the parentheses into space; (ii) the numbers, stopwords, punctuation, and extra spaces are removed; (iii) all letters are converted to lowercase; and (iv) duplicate strings are removed. Algorithm 1 shows all the preprocessing steps for PRO-COM and PK-BD.
Algorithm 1 RS-CF: PRO-COM and PK-BD preprocessing pseudocode.
1:
procedurePreprocess(MDD-DS1)
2:
     C l e a n i n g (MDD-DS1)
3:
     p r o d u c t _ w o r d s [ ] new_list(m)
4:
     a l l _ p r o d u c t s _ w o r d s new_vector (0)
5:
    for i ← 1: m do
6:
         p MDD-DS1 [ i , ]
7:
         d e s ( p ) ← C(p[Name], p[Legal Name], p[Ingredients])
8:
         d e s ( p ) C l e a n _ p ( d e s ( p ) )
9:
         p r o d u c t _ w o r d s [ i ] d e s ( p )
10:
        a l l _ p r o d u c t s _ w o r d s
11:
       ← a l l _ p r o d u c t s _ w o r d s d e s ( p )
12:
    end for
13:
end procedure
Thus, the words are divided and a vector of words is created for p r o d u c t _ w o r d s ( p ) , an example is shown in Table 5.
We obtain a l l _ p r o d u c t s _ w o r d s unique tokens/words extracted from the corpus C(p[Name], p[Legal Name], p[Ingredients]), which is the different meaningful tokens in the dataset after preprocessing. Therefore, a l l _ p r o d u c t s _ w o r d s contains 10,707 unique tokens, an example (We have translated p r o d u c t _ w o r d s ( p ) and a l l _ p r o d u c t s _ w o r d s to make it readable) shows in the Table 6. Let t be the n-dimensional vector obtained from a l l _ p r o d u c t s _ w o r d s such that t = ( t 1 , , t n ) and k [ 1 , , n ] , t k is a string ∈ a l l _ p r o d u c t s _ w o r d s and N = dim ( a l l _ p r o d u c t s _ w o r d s ) . The N tokens will form d e s ( p ) and the count vector size in product matrix X will be given by M x N .
The product matrix X [ m , n ] is a M x N matrix, where each row M represents a product p of the MDD-DS1 so that M is the total number of products; and each column N represents a token t i t . The next step is to use BOW to rate the words on each product. The goal is to convert each free text product into a vector that we can use as an RS model input. Since we know that the vocabulary in a l l _ p r o d u c t s _ w o r d s contains 10,707 words, we can use a fixed-length document-representation of 10,707, with a position in the vector to score each word. The simplest scoring method is to mark the presence of words as a boolean value, 0 for absent, 1 for present. Using the arbitrary order of a l l _ p r o d u c t s _ w o r d s listed above in our vocabulary d e s ( p ) , we can loop through the products and convert them to a binary vector, as shown in Table 7.

5.2. RS-CF: Product Composition Approach

Product composition (PRO-COM) is the main approach upon which the recommendation system is built. PRO-COM is built to obtain the alternative product based on the similarity ratio. A product matrix is used and these steps are followed to build PRO-COM approach. Let Z a variety. Let d be a | Z | -dimensional vector. Here, d = ( d 1 , , d | Z | ) where d i = d ( q i , q j ) denoting the following distance between the products q i and ( q j Z ).
The product p is calculated by getting the absolute value of the difference considering each column of the product matrix ( a l l _ p r o d u c t s _ w o r d s ) and then adding up all the distances as shown in Equation (1). When a product p Z that is not available, the RS-CF (PRO-COM) recommended p a Z alternative product would be obtained. If there is more than one p a value p a = ( p a 1 , , p a t ) , a t-dimensional vector p a is created being t the number of alternatives. The alternative products given ( p a ) will be those that have the lower distance to the product p as show that in Equation (2) and can be seen an example of a distance vector p a = ( p a 1 , , p a t ) in Table 8.
d ( q i , q j ) = k = 1 n | X [ q i , k ] X [ q j , k ] | / q i , q j Z
p a = p a i / d ( p , p a i ) = min i = 1 , , | Z | p q i { d ( p , q i ) } / p , q i Z .

5.3. RS-CF: Package-Based Approach

The PK-BD approach is to offer alternative products with product size in mind, so the PK-BD approach adds more condition using an additional feature called serving (size per person), at the same time taking the PRO-COM distance result into account. They are compared again to an unavailable product p but with regards to the package size. Each product p of the variety Z, a distance between the product p is calculated the absolute value of the package size difference. Then, the alternative products p a given will be those that have the lower distance with respect to the product p. Let s be the vector that contains the package size of the products in Z ( | Z | -dimensional vector). Hence, s = ( s 1 , , s | Z | ) . Let d s ( q i , q j ) the following distance between the products q i and q j according to their package size as shown in Equation (3).
d s ( q i , q j ) = | s [ q i ] s [ q j ] | / q i , q j Z .
Considering the PK-BD approach, if a product p Z is not available there are two steps to follow in order to get the alternative product p a Z ; (i) First of all, the PRO-COM distance is taken into account by applying Equation (2). (ii) Next, the package size distance is additionally applied to the products in vector p a | p = ( p a 1 | p , , p a t | p ) in order to select the alternative product p a to be offered to the user as show that in Equation (4). If there is more than one p a | l value, a u-dimensional vector p a | l = ( p a 1 | l , , p a u | l ) is created, u being the number of alternatives. An example of a matrix, that has two distance vectors taking into account the criterions selected, is shown in Table 9.
p a | l = p a i | l / d s ( p , p a i | l ) = min q i p a p q i { d s ( p , q i ) } / p , q i Z .

5.4. RS-CF: Health-Based Approach

The health-based approach (HTH-BD) is the tricky one to consider, recommending health products to the user based on their choices. The most common nutritional table properties fats, sugar are used to help recommend healthy products. The cleanliness of the data mentioned in Section 5.1 is used in addition to replacing the serving with the additional properties, which are fats, sugars, so that the rows with blank values for the name of the product, sugar and fat are eliminated, so that the sugar values in the remaining products range between 1 and 1087 g, and the fat values in the remaining products also range between 1 and 937 g. Additionally, 13 additional columns named Messages that provide allergen information are being considered. The number of products becomes 20,259 products and 24 features after cleaning the data.
About the Messages columns, after analyzing all the tags indicating the absence of allergens 50 different strings are obtained in the form Table 10 and stored in a vector named w i t h o u t w o r d s . Here, taking into account the law of the European Union (BOE, Regulation (EU) n. 1169/2011 of the European Parliament and of the Council) on the labeling of food products that obliges companies to report certain allergens that may endanger the health of the customer, sensitivity will be taken into account.
A law with similar objectives was previously approved in Spain in 2004 and amended in 2008 (BOE, Royal Decree 1245/2008, of 18 July, which modifies the General Regulations for Labeling, Presentation and Advertising of Food Products, Approved by Royal Decree 1334/1999 of 31 July). Of the 50 different strings obtained previously, 17 are relevant in terms of allergies, as can be seen in Table 11. After performing the necessary analysis and clarification, the information obtained from MDD-DS1 is useful for developing the RS-CF HTH-BD algorithm.
Aside from the data obtained from the Ingredients variable, the Messages columns associated with the respective product are also obtained for each iteration. Here, for each product, the 13 Messages columns are handled in the following way: (1) 13 columns for the current product in the iteration are obtained, with the blank columns removed; (2) To remove additional information unrelated to the allergen, values are also removed from columns that do not begin with the string “without”; (3) The duplicate strings obtained are removed, strings are converted to lowercase; (4) The strings are divided by a point followed by a space, substrings preceded by a comma are removed; (5) Some incorrect parsed characters (overridden characters such as ∖r and ∖n backslashes) are removed, as well as some strings with errors and full stops are removed. The word vector is constructed with the resulting string.
As in the PRO-COM and PK-BD approaches, the  p r o d u c t _ w o r d s list is generated with the difference that here just the Ingredients column is considered. This is, it contains a number of elements equal to the number of different products existing in the MDD-DS1 (in the HTH-BD approach, the MDD-DS1 has 20,259 elements). The vector of words belonging to each product obtained in the text string processing is stored in each element of the list after using the steps of C l e a n _ p . The list is shown in Table 12.
In addition, a list called withoutlist, which stores the vector with the healthy features obtained from the Messages columns for each product, is created. The vector withoutwords stores the different healthy features once, having 50 elements. The list and the vector are shown in Table 13 and Table 14, respectively. The entire preprocessing is shown in Algorithm 2. It is relevant to know that a subset comprising 17 elements of the withoutwords vector is considered in order to check for allergens in a product, whose data about it can be accessed by indexing the withoutlist with the index of the product in the MDD-DS1.
Algorithm 2 RS-CF: HTH-BD: Preprocessing pseudocode.
1:
procedurePreprocess(MDD-DS1)
2:
     C l e a n i n g (MDD-DS1)
3:
     p r o d u c t _ w o r d s [ ] new_list(m)
4:
     w i t h o u t l i s t new_list(m)
5:
     w i t h o u t w o r d s new_vector (0)
6:
    for i ← 1: m do
7:
         p MDD-DS1 [ i , ]
8:
         d e s ( p ) ← (p[Ingredients])▹ Just the Ingredientes column is taken into account
9:
         d e s ( p ) p
10:
       d e s ( p ) C l e a n _ p ( d e s ( p ) )
11:
       p r o d u c t _ w o r d s [ i ] d e s ( p )
12:
     m MDD-DS1 [ i , M e s s a g e s 1 : M e s s a g e s 13 ] The information contained in the 13 Messages columns is loaded
13:
       m remove_empty_strings(m)
14:
       m m[which(m.startsWith(“without ”)]▹Just the values in the m vector which start with the string “without ” are loaded
15:
         m remove_duplicates(m)
16:
         m transform_into_lowercase(m)
17:
         m split(m,“[.] ”)▹ The strings are splitted by a dot followed by a space
18:
         m split(m,“[,].*”)▹ The strings are splitted by a comma, removing what is after it
19:
         m remove_malformed_strings(m)
20:
         m remove_full_stops(m)
21:
         w i t h o u t w o r d s w i t h o u t w o r d s m
22:
         w i t h o u t l i s t [ i ] m
23:
    end for
24:
end procedure
After processing the data to be valid for building the health-based approach, let g be the withoutwords vector (50 elements). Then, g = ( g 1 , , g 50 ) . Let a be the subset of the withoutwords vector considering allergens (17 elements). Hence, a g / a = ( a 1 , , a 17 ) . Let v be a the m-dimensional wordvectors list. Each element contains a vector v i . Hence v = ( v 1 , , v m ) . Likewise, v i = ( v i [ 1 ] , , v i [ d v i ] ) where d v i is the length of the vector contained in the i element of the list v . Note that k [ 1 , , d v i ] , v i [ k ] is a string. Let v s be a | Z | -dimensional subset of the Z elements of the m-dimensional v wordvectors list. Each element contains a vector v s i . Hence v s v and v s = ( v s 1 , , v s | Z | ) . Likewise, v s i = ( v s i [ 1 ] , , v s i [ d v s i ] ) where d v s i is the length of the vector contained in the i element of the list v s . Note that k [ 1 , , d v s i ] , v s i [ k ] is a string. Let n p l be a | Z | -dimensional vector. Here, n p l = ( n p l 1 , , n p l | Z | ) where n p l i = d v s q i . Each element denotes the processing level of a product.
Let w be the m-dimensional withoutlist list. Each element contains a vector w i . Hence w = ( w 1 , , w m ) . Likewise, w i = ( w i [ 1 ] , , w i [ d w i ] ) where d w i is the length of the vector contained in the i element of the list w . Note that k [ 1 , , d w i ] , w i [ k ] is a string. Let w s be a | Z | -dimensional subset of the Z elements of the m-dimensional w withoutlist list. Each element contains a vector w s i . Hence w s w and w s = ( w s 1 , , w s | Z | ) . Likewise, w s i = ( w s i [ 1 ] , , w s i [ d w s i ] ) where d w s i is the length of the vector contained in the i element of the list w s . Note that k [ 1 , , d w s i ] , w s i [ k ] is a string. Let n h be a | Z | -dimensional vector. Here, n h = ( n h 1 , , n h | Z | ) where n h i = d w s q i . Each element denotes the number of healthy features of a product.
Let c be a | Z | -dimensional vector. Here, c = ( f 1 + s 1 , , f | Z | + s | Z | ) where c i = f i + s i . It stores the fat and sugar features about the products. Here, f i and s i denote, respectively, the fat and sugar quantities in grams of the product q i . Let d a i = d a ( q i , q j ) . This denotes the following similarity measure (taking into account allergens) of the product q i with respect to the product q j as shown in Equation (5).
d a ( q i , q j ) = { l 1 , if h / a [ h ] w s [ q i ] a [ h ] w s [ q j ] / q i , q j Z 0 , if h / a [ h ] w s [ q i ] a [ h ] w s [ q j ] / q i , q j Z
where a [ h ] w s [ q i ] k / w s q j [ k ] = a [ h ] , k [ 1 , , d w s q i ] .
The product p Z being unavailable, the alternative product p a Z is obtained by the following the next steps: (1) The first criterion is to consider the similarity about allergens. Thus, the alternative product p a | a is selected according to that measure:
p a | a = p a i | a / d a ( p , p a i | a ) = max i 1 , , | Z | p q i { d a ( p , q i ) } / p , q i Z .
If there is more than one p a | a value, a  u a -dimensional vector p a | a = ( p a 1 | a , , p a u a | a ) is created being u a the number of alternatives. (2) Secondly, the sum of the sugar and fat quantities are considered to select the alternative product p a | c among the ones in vector  p a | a :
p a | c = p a i | c / c [ p a i | c ] = min q i p a | a p q i { c [ q i ] } / p , q i Z .
If there is more than one p a | c value, a  u c -dimensional vector p a | c = ( p a 1 | c , , p a u c | c ) is created being u c the number of alternatives. (3) The next criterion to get the alternative product p a | h (among the ones in vector p a | c ) is the number of healthy features:
p a | h = p a i | h / n h [ p a i | h ] = max q i p a | c p q i { n h [ q i ] } / p , q i Z .
If there is more than one p a | h value, a  u h -dimensional vector p a | h = ( p a 1 | h , , p a u h | h ) is created being u h the number of alternatives. (4) The last step to get the alternative product p a | n p l involves the level of processing of the products selecting from the vector p a | h :
p a | n p l = p a i | n p l / n p [ p a i | n p l ] = min q i p a | h p q i { n p l [ q i ] } / p , q i Z .
If there is more than one p a | n p l value, a  u n p l -dimensional vector p a | n p l = ( p a 1 | n p l , , p a u n p l | n p l ) is created being u n p l the number of alternatives.
In conclusion, Algorithm 3 compares first each q i product in the variety Z to the product p with regards to the similar features about allergens. Similar products are then ranked considering these features in order: the sum of the fat and sugar amounts (in increasing order), the number of healthy features (in decreasing order) and the processing level (in increasing order). An example of a matrix with vectors defining each of the criterions as columns is shown in Table 15.
After building the three approaches, a user survey was conducted. Products and alternatives were presented according to each approach. Subsequently, the analyses of the results were compiled. We developed the approach to improve the results and meet the company’s requirements.
Algorithm 3 HTH-BD approach of RS-CF: Algorithm pseudocode.
1:
procedureAlgorithm(p, a , v s , w s , c )    ▹p is the index of the unavailable product
2:
    for  i 1 : | Z |  do
3:
        if  p = = i  then
4:
           continue
5:
        end if
6:
         a _ i n d e x e s which(a w s [ p ] )
7:
        if a _ i n d e x e s , a [ a _ i n d e x e s ] w s [ i ]  then
8:
            d a [ i ] = 1
9:
        else
10:
            d a [ i ] = 0
11:
        end if
12:
         n h [ i ] = l e n g t h ( w s [ i ] )
13:
         n p l [ i ] = l e n g t h ( v s [ i ] )
14:
    end for
15:
     p a sort(- d a , c , - n h , n p l )    ▹ The products are sorted. The minus sign means the order is decreasing.
16:
end procedure

6. Recommendation System Based on Neural Network-Based (RS-NN)

The idea of improving RS-CF is based on improving the result and considering more conditions and filtering: (1) Adding allergens’ properties as a pre-condition in the recommendation for three approaches (PRO-COM, PK-BD, and HTH-BD). For example, the product includes (flour, eggs, water, nuts, and salt), so the alternative product will include free allergens, or the maximum allergens are eggs and nuts; (2) We also consider more conditions for three approaches based on using more additional features such as brand type, brand attribute and price; (3) Besides considering the more characteristics of the nutritional table, such as carbohydrates, dietary fiber, a percentage of saturated fat, good fat, protein and salt to improve the HTH-BD approach. (4) Rearrange the approaches of PRO-COM, then HTH-BD, then PK-BD. To improve the result, we thought about using a deep neural network like Doc2vec to represent the data set and build a model to help obtain alternative products. That is why we call this model a Recommendation system based on neural networks (RS-NN).
After that, we use many of the similarity techniques such as Cosine, Jaccard, Euclidean and Manhattan to obtain and sort similar products. Subsequently, we conduct a comparative study to determine which technique is best to sort similar products based on the experts’ results.
Figure 4 illustrates that the new model comprises three main steps: (A) Preprocessing the dataset using text mining, filtering, and representing the adaptive dataset with a neural network model. (B) Using neural networks to create a model based on Doc2vec. (C) We apply the three RS-NN approach (PRO-COM, HTH-BD, and PK-BD).
An expert first carries out the evaluation; then we did a user survey based on the result of the expert.

6.1. Preprocessing

In order to build the RS-NN approaches, we will use some more features, such as the brand type and the brand attributes, as well as the addition of 16 characteristics that cause allergies (Nuts, egg, hazelnuts, fish, sulfates, peanuts, mollusks, lupine, gluten, mustard, soy, crustaceans, milk and its derivatives including lactose, sunflower seeds and sesame). So, the Midiadia extend new version of data set (MDD-DS2) which is the number of products 45,198 product and the number of features is 368. To improve the approaches, the data cleansing, preprocessing, and approach building phase was reused, so blank rows were removed from each name, brand,brand type, brand attributes, variety, ingredients and legal name. The elimination of the duplicate rows that have the same name and brand, and finally the empty and duplicate rows, were eliminated from the EAN variable. The main idea of the recommendation system is to have alternatives for the product of the same variety. Therefore, we will remove all variety with less than the first quarter, which equates to 15 products, Table 16 shows the products per variety after eliminate the variety for PRO-COM approach (Section 6.4).
The variety Z, p Z that is not available, the recommended p a Z alternative product. After that, a product p is in a specific variety called “ o t h e r ” ( p Z = o t h e r ”), and is listed at a subcategory. Let SC be a subcategory, where Z S C , so the other subcategories S C will be removed. Let B a a Brands attributes have a product p B a that is not available, the recommended p a B a alternative product. Finally, it filters the products p according to the allergens feature, let a f be allergen features. Eliminating all the products that contain more or different allergens feature a f .
The products p and p a alternative product will be preprocessed by extracting all the words for each of the following attributes: Name, brand, Ingredients, legal Name, and allergens feature for each product p, p a alternative product. Consider a Corpus C of ( N a m e , b r a n d , I n g r e d i e n t s , l e g a l N a m e , a l l e r g e n s f e a t u r e ) to describe the product d e s ( p ) . This description ( d e s ( p ) ) was obtained after the 3-step purge process: (1) Use the tokenize function to make a list to convert everything to lower words and separate each word p r o d u c t _ w o r d s ( p ) ; (2) Use stopwords in Spanish to filter by stopwords such as remove [and, or, etc.], use the number filter to remove all numbers from the list; (3) Use a lemmatization step to takes the tokens and divides each one into a lemma, which is the basic form of the word, it cuts the conjugation and declension method. For example, the word “different” would become “differ”, “running” would become “run” and “trucks” would become “truck”. Lemmatization can even change from “was” to “be” because the lemmatizer (nltk) improves vocabulary rather than relying solely on the algorithm, similar words are then removed, as shown in Table 17 step (3).
After pre-processing, we used the Tagged_Document function for training corpus C( N a m e , b r a n d , I n g r e d i e n t s , l e g a l N a m e , a l l e r g e n s f e a t u r e ), which represents a product along with a tag, input product format for Doc2Vec, a single product, made up of words; a list of unicode string tokens and tags; a list of tokens. Tags may be one or more unicode string tokens, but typical practice (which will also be the most memory-efficient) is for the tags list to include a unique integer i d as the only tag. Let p be the p r o d u c t _ w o r d s [ ] list such that, each element contains a vector p i hence p = ( p 1 , , p m ) . Likewise, p i = ( p i [ 1 ] , , p i [ l ] ) where l = dim ( p r o d u c t _ w o r d s [ i ] ) , length of the vector in the i element of the list p . Note that k [ 1 , , l ] , p i [ k ] is a string. Then beside each p, the Tagged_Document function defines the tag (the product id p i d ), which means simply the zero-based line number as shown in Table 18.

6.2. Product Representation

As explained earlier, the doc2vec in Section 2, it shows the simplest way to convert a token to a fixed-size digital vector, as it proposed a neural network-based word representation method called Word2Vec. Give a sequence of training tokens [ t 1 , t 2 , , t N 1 , t N ] ; the goal of Word2Vec is to maximize the average log probability [57]:
1 N n = s N s log p ( t n | t n s , , t n + s ) ,
where s is the size of the window to preserve contextual information, the token t n can be easily be predicted using a multilabel classifier like SoftMax function:
p ( t n | t n s , , t n + s ) = e j t n i e j t i ,
where each j t i is the ( t i ) output value of a feed-forward neural network calculated with
j = x + h f ( t n s , , t n + s ; R ) ,
where x, h, f, and R are terms for the bias between the hidden and output layers, the weight matrix between the hidden and output layers, the mean or sequence of product tokens, and the word embedding matrix, respectively. Doc2Vec extends from Word2Vec, which tries to define, in this case, a continuous vector fit to a product to preserve the semantic relationship between the different products [58]. Like Word2Vec, each token is represented by a d-dimensional continuous vector ( d < < | v | , which is the size of the vocabulary in the d e s ( p ) ). Furthermore, the same product p is also represented by a continuous vector in the same space as the word vectors. In Doc2Vec, each product p is assigned to a unique vector that is represented by a column in matrix D, while each token in the d e s ( p ) is assigned to a unique vector that is represented by a column in matrix T. Therefore, the only change in the network formulation is to add D to the Equation (12) as follows:
j = x + h f ( t n s , , t n + s ; R , D ) .
When the network is adequately trained, it can obtain a distributed representation of each of the products p. Therefore, the products were trained using three elements of the Doc2Vec model, vector size with 50 dimensions a = ( a 1 , , a 50 ) , and iteration over the training set 40 times. Set the minimum word count to two to ignore words with very few frequencies.
Finally, we have a product matrix X [ m , n ] is a M x N matrix where each column N represents a vector for each product a = ( [ a 1 , , a 50 ] , t a g ) , k [ 1 , , 50 ] , where t a g p i d , and each row M represents a product p of the MDD-DS2 so that M is the total number of products p = ( p 1 , , p m ) , as shown in Table 19.

6.3. Similarity

The RS-NN approaches used the similarity techniques such as (Cosine, Jaccard, Euclidean, and Manhattan) to calculate the distance between the product q i and q j . Let d i = d ( q i , q j ) , this denotes the following similarity measure taking into account variety Z, brand attribute B a , and allergens features a f of the product q i with respect to the product q j as show that in Equation (14).
C o s [ d ( q i , q j ) ] = q i · q j | | q i | | · | | q j | | / q i , q j Z , B a J a c [ d ( q i , q j ) ] = | q i q j | | q i q j | / q i , q j Z , B a E [ d ( q i , q j ) ] = k = 1 n | X [ q i , k ] X [ q j , k ] | 2 / q i , q j Z , B a M [ d ( q i , q j ) ] = k = 1 n | X [ q i , k ] X [ q j , k ] | / q i , q j Z , B a .
Having a product p Z and p B a that is not available, the recommended p a Z and p a B a alternative product would be obtained as follows taking into account the allergen featuresas shown in Equation (15), the first equation is the output from Cosine, Jaccard, the second one for Euclidean, and Manhattan. If there is more than one C J [ p a ] , E M [ p a ] value C J [ p a ] = ( p a 1 , , p a c ) , E M [ p a ] = ( p a 1 , , p a c ) a c-dimensional vector p a is created being c the number of alternatives.
C J [ p a ] = p a i / d ( p , p a i ) = max i = 1 , , | Z , B a | p q i { d ( p , q i ) } E M [ p a ] = p a i / d ( p , p a i ) = min i = 1 , , | Z , B a | p q i { d ( p , q i ) . }

6.4. RS-NN: Product Composition Approach

The product composition (PRO-COM) is where similarity is scored according to product matrix to offer alternative products. In addition, the alternatives taking into account the distance based on d ( q i , q j ) are compared to the unavailable product but with regards to the brand, brand type, and price. For each product of the variety Z, and p B a brand attribute, a distance between the product p is calculated using similarity techniques Equation (14).
Considering the PRO-COM approach, if a product p Z , and p B a is not available in order to get the alternative product, p a Z , and p a B a . Let b be the vector that contains the brand of the products in Z ( | Z | -dimensional vector). Hence, b = ( b 1 , , b | Z | ) . Beside, let b t be the vector that contains the brand type of the products in Z ( | Z | -dimensional vector). Hence, b t = ( b t 1 , , b t | Z | ) . In addition, let P R be the vector that contains the price of the products in Z ( | Z | -dimensional vector). Hence, P R = ( P R 1 , , P R | Z | ) .
Considering verifying the brand b and brand type b t in the product p and that p a alternative product contains the same value for the two variables ( b , b t ) , we found three possibilities:
(1) The alternative product q j has the same attributes value for ( b , b t ) of the product q i .
P o s ( 1 ) = m / b [ m ] = b [ q i ] b t [ m ] = b t [ q i ] b [ m ] = b [ q j ] b t [ m ] = b t [ q j ] / q i , q j Z , B a ;
(2) The alternative product q j has the attribute value of one of ( b , b t ) of the product q i .
P o s ( 2 ) = m / b [ m ] = b [ q i ] b t [ m ] = b t [ q i ] b [ m ] = b [ q j ] b t [ m ] = b t [ q j ] / q i , q j Z , B a ;
(3) The alternative product q j does not have the same value for ( b , b t ) of the product q i .
P o s ( 3 ) = m / b [ m ] = b [ q i ] b t [ m ] = b t [ q i ] b [ m ] b [ q j ] b t [ m ] b t [ q j ] / q i , q j Z , B a
To check the price, there are two options in each possibility of variables ( b , b t ) ; the price P R of the alternative product q j is higher than the product q i or vice versa. Let C J [ d r i ] = C J [ d r ( q i , q j ) ] for cosine and jaccard, let E M [ d r i ] = E M [ d r ( q i , q j ) ] for euclidean and manhattan. This denotes the following similarity measure of the product q i with respect to the product q j as show that in Equation (16).
C J [ d r ( q i , q j ) ] = { l if P R [ q j ]   >   P R [ q i ] , d ( q i , q j ) × ( P R [ q i ] / P R [ q j ] ) if P R [ q j ]     P R [ q i ] , d ( q i , q j ) × ( P R [ q j ] / P R [ q i ] ) E M [ d r ( q i , q j ) ] = { if P R [ q j ]   >   P R [ q i ] , ( P R [ q i ] / P R [ q j ] ) / d ( q i , q j ) if P R [ q j ]     P R [ q i ] , ( P R [ q j ] / P R [ q i ] ) / d ( q i , q j ) .
Then, check the possibilities for p a alternatives product of variables brand and brand type ( b , b t ) , and calculate the distance C J [ d r ( q i , q j ) ] , E M [ d r ( q i , q j ) ] . Let C J [ d m i ] = C J [ d m ( q i , q j ) ] for cosine and jaccard, let E M [ d m i ] = E M [ d m ( q i , q j ) ] for euclidean and manhattan. This denotes the following similarity measure of the product q i with respect to the product q j as shown in Equation (17). Lastly, we will multiply the distance C J [ d r ( q i , q j ) ] , E M [ d r ( q i , q j ) ] with weight (100, 10, 1) to help the p a alternative product’s ordering.
C J [ d m ( q i , q j ) ] = { l d r ( q i , q j ) × 100 ,   if   P o s ( 1 ) d r ( q i , q j ) × 10 ,   if   P o s ( 2 ) d r ( q i , q j ) ,   if   P o s ( 3 ) E M [ d m ( q i , q j ) ] = { l d r ( q i , q j ) ,   if   P o s ( 1 ) d r ( q i , q j ) × 10 ,   if   P o s ( 2 ) d r ( q i , q j ) × 100 ,   if   P o s ( 3 )
The distance is additionally applied to the products in order to select the alternative product to be offered to the user as shown in Equation (18). If there is output from the similarity techniques (Cosine, Jaccard, Euclidean, and Manhattan) more than one alternative product p a | b value, a y-dimensional vector p a | b = ( p a 1 | b , , p a y | b ) is created, y being the number of alternatives.
p a | b = p a i | b / d m ( p , p a i | b ) = max q i p a p q i { d m ( p , q i ) . }
Finally, after its development, PRO-COM works on three main characteristics, which are the brand, brand type and the price. After obtaining a vector p a | b , the alternative products are ordered from closest to furthest.

6.5. RS-NN: Healthy-Based Approach

The health-based (HTH-BD) approach depends on the result of PRO-COM approach and make an equation for nutrition table features. The HTH-BD was based on the most health-based characteristics found in the nutrition table, namely fats  ( f ) , containing a percentage of saturated fat ( s f ) and a percentage of good fats ( g f ) ; ( s f , g f f ) , carbohydrates ( C a r b s ) , and containing dietary fibers ( d f ) and sugars ( s ) ( d f , s C a r b s ) and finally salt ( s a ) and protein ( p n ) .
Eight characteristics play an important role in the product, whether or not it becomes a healthy product. So, the products that do not have values for these characteristics are removed. As we mentioned before, if the product p in the variety Z which means p Z and brand attribute B a include variety Z, Z B a , the HTH-BD approach recommends p a alternative products within a variety Z, then the products p are analyzed within the variety Z, and Z that contain less than four products are removed, which is the first quarter of the value of the products for each variety, so it becomes the minimum variety that contains four products and the maximum number of products per variety is 203, the median is 13 products and the mean contains about 24.3 products for the one variety, this is after analyzing the products p of HTH-BD approach, shown in Table 16.
After that, we check the values of nutrition tables characteristics that have the same unit of measurement such as Grams, %, and so forth. It turns out that the nutrition tables characteristics are measured in grams, except for the percentage of good fats g f and dietary fiber d f , and each of them is measured in percentage. They are converted to grams [59,60] using g f = ( g f / 100 ) f . We converted the dietary fiber variable, d f = ( d f / 100 ) C a r b s .
This approach has used some nutrition books and nutrition experts [61,62,63,64] to arrange the nutritional table features used in this approach. The result of this arrangement was (protein, then good fats, then dietary fiber, then salt, then sugars, then carbohydrates, then saturated fat, and finally fat). In our research, an additional weight value was added to each nutrition table feature to help us arrange the product alternative.
Let h be a sort of nutrition table features list with weight value of each nutrition table feature. Each element contains a vector h i . Likewise, h i = ( p n i 100 + g f i 200 + d f i 300 + s a i 400 + s i 500 + C a r b s i 600 + s f i 700 + f i 800 ) . It stores protein ( p n ) , good fats ( g f ) , dietary fiber ( d f ) , salt ( s a ) , sugars ( s ) , carbohydrates ( C a r b s ) , saturated fat ( s f ) , and fat ( f ) nutrition features about the products.
Let N t be a | Z | -dimensional vector. Here, N t = ( h 1 , , h | Z | ) where N t i = ( h i ) . Let d h ( q i , q j ) , this denotes the following similarity measure according to their nutrition table of the product q i with respect to the product q j Equation (19), the similarity calculated based on the output of PRO-COM Equation (18).
d h ( q i , q j ) = N t q j / d m ( p , q i ) .
The product being unavailable, the alternative product p a | h is selected according to that measure. The less value in the alternative product becomes a healthy product for the user:
p a | h = p a i | h / d h ( p , p a i | h ) = min q i p a | b p q i { d h ( p , q i ) . }
If there is more than one p a | h value, a u h -dimensional vector p a | h = ( p a 1 | h , , p a u h | h ) is created being u h the number of alternatives.

6.6. RS-NN: Package-Based Approach

The package-based (PK-BD) approach is considered to include all the approaches together as it depends on the PRO-COM and HTH-BD approaches. The algorithm was developed based on the result of the HTH-BD approach. First, products that do not contain values for the three variables, which are product size, units of measure and servings, are removed, and these are the variables on which this approach depends. Second, as mentioned above, the product p and alternative products p a must be within a variety Z, and within a brand attribute B a , so the quantity of products within the varieties is analyzed so that the varieties containing less than the first quarter value are removed from the number of products within each Z and its value is 4. Therefore, in the PK-BD approach as shown in Table 16, the minimum product per variety is four products, and maximum of the product per variety is 203 products and the median number of products is 13, and the average becomes 24.3 products. The algorithm is based on arranging alternative products p a based on the servings S g value of the product p. Let S g be the vector that contains the servings of the products in Z ( | Z | -dimensional vector). Hence, S g = ( S g 1 , , S g | Z | ) . The value of servings S g in the product q i is compared to the alternative product q j , and there are two possibilities: namely that the product q i has servings S g value greater than the servings S g value of the alternative product q j , or vice versa. Let d s e i = d s e ( q i , q j ) . This denotes the following similarity measure of the product q i with respect to the product as shown in Equation (21).
d s e ( q i , q j ) = { l ( S g [ q i ] / S g [ q j ] ) / d h ( q i , q j ) , if S g [ q j ] > S g [ q i ] ( S g [ q j ] / S g [ q i ] ) / d h ( q i , q j ) , if S g [ q j ] S g [ q i ] .
The distance is additionally applied to the products in order to select the alternative product p a as shown in Equation (22). When there is output more than one alternative product p a | s value, a j-dimensional vector p a | s = ( p a 1 | s , , p a j | s ) is created being j the number of alternatives.
p a | s = p a i | s / d s e ( p , p a i | s ) = max q i p a | h p q i { d s e ( p , q i ) . }
The arrangement of alternative products p a | s is based on the closest similarity ratio to the product p, taking into account the value of servings S e , which is greater or less in proportion to the value of the servings S e of the product p.

7. Experimental Evaluation

In order to evaluate the effectiveness and performance of our recommender system, which is exclusively based on the product characteristics, we have used the following hardware and software equipment. We have selected the Python language to implement the different recommender system approaches. Our system uses the Window 10 operating system and hardware with the following specifications: Intel(R) Core(TM) i7-5500U, CPU (2.4 GHz), RAM (16 GB) and Storage (1 TB). The response time is very fast, taking approximately 4 s to recommend alternative products of each desired product. In order to perform these tests, we randomly selected the products from MDD-DS for each approach. The alternative products given by the different RS approaches are displayed and stored in a report for the users and experts to check the results. Finally, and in order to deployed the survey, we have decided to conduct a web survey using Python [65] and Django [66,67] on Amazon Elastic Compute Cloud (Amazon EC2), with the following specifications: vCPU (8), Memory (32 GiB), Network Burst Bandwidth (Up to 5 Gbps).
The evaluations are presented through a survey that includes three approaches, which is answered by users and experts. Each survey comprises 30 questions and each group contains ten questions as shown in Figure 5.
Each question includes a product p and three alternative products p a , with alternative products being the first three products with the closest value to the product. The survey depends on the situation in which the person is shopping and the product has not been found, so the user chooses between three alternative products according to each approach.
The experimental evaluation is divided into four subsections, which are the evaluations obtained from the user (Section 7.1), in Section 7.2, the evaluations obtained by the expert, in Section 7.3, the evaluations obtained by the user based on the results of the expert and, finally, evaluation and discussion, which is a comparison between user surveys (Section 7.4).

7.1. RS-CF User Survey

The survey is divided into three blocks. Therefore, the first block is considered, expressing the PRO-COM approach (Section 5.2), the second block is dedicated to the PK-BD approach (Section 5.3), and the last block is also performed to evaluate the HTH-BD approach (Section 5.4).
Survey results were calculated using mean squared error (MSE) after 65 people had responsed. That said, once the products are sorted, there may be more than one substitute. Therefore, a link may occur between the highest-rated products recommended in the order given by the various approaches of the RS, so that the MSE [68] has been calculated for each approach taking into account the following three groups:
  • Group 1: All the questions answered by the users are considered.
  • Group 2: The questions having untied answers and the questions in which just the top-2 choices are tied are considered.
  • Group 3: Only questions with untied answers are considered.
The formula used to calculate the MSE is the following:
M S E = 1 n i = 1 n ( Y i ^ Y i ) 2 ,
where the value Y i ^ is the value of the answer chosen by the user and Y i is the top-1 product, always having a value of 1. The value of Y i ^ would be ( 1 ; 0.5 ; 0 ) if the user chose the first, second or third product of the survey, respectively. The values would be ( 1 ; 1 ; 0.5 ) if there is a tie between the top-2 products and it would be ( 1 ; 1 ; 1 ) if the tie happened between all the products. The results are shown in Table 20, taking into account the three groups.
Accuracy (ACC) was also calculated for the result (only for Group 3) as shown in Equation (24).
A C C = i = 1 n ( ( x i = 1 ) ( x i = 2 ) ) n ,
where the value n is the number of questions in group 3 and x i is the answer chosen by the user of group 3. It includes the answers in which the choice of the first or second product will be declared as positive while the third product will be declared negative. The result is shown in Table 21.

7.2. RS-NN Expert Survey

The company provided experts to evaluate the three approaches in the recommendation system, and expert opinions are important in evaluating the recommendation system for several reasons, the most important of which is that the experts know the products and also know the alternative products, so they can easily give their opinion whether the recommendation system recommends suitable alternative products or not. Four surveys of each approach were sent to the experts; the surveys are the result of the techniques used (Cosine, Jaccard, Euclidean and Manhattan similarity) and those mentioned in Section 6.3. In each survey, the expert must answer three questions, namely: (1) Would you select any of these 3 options (alternative products)? (Yes/no); (2) If yes, select which one? (for example, 3); (3) Elaborate a raking to order the options (from the most similar product to the less similar product). Example: 3, 1, 2.
Figure 6 shows the results of the surveys indicating how many times the alternative product, be it the first, second or third, was chosen for each technique and also for each approach. The results show that the first approach has 80% of the questions that have suitable alternatives. The expert reported that the second approach recommends that 90% of the questions have suitable alternative products. The expert also stated that the third approach also recommends 80% suitable alternative products.
Therefore, a user survey was created based on the opinions of the experts. Cosine similarity was chosen for all three approaches.

7.3. RS-NN User Survey

This survey was built after taking the result from the expert, and this survey was very similar to the first survey we did, but augmented with clear images to make it easier for the user to quickly get to know the product and choose between alternative products, it is easier than. This survey is considered including three blocks. The first block expresses the PRO-COM approach (Section 6.4), the second block is dedicated to the HTH-BD approach (Section 6.5), and the last block is also implemented to evaluate the PK-BD approach (Section 6.6).
As shown in Table 22, the survey results were also calculated using MSE as we did in the first survey after receiving 65 responses from users. The same groups that were used before were used to compare the results between the two investigations.
Accuracy (ACC) for the result of Group 3 was also calculated as shown in Table 23 using Equation (24) as calculated in the first survey.

7.4. Evaluation and Discussion

Performing offline experiments by using a pre-collected data set to let users choose or rate items is the usual way to estimate the performance of recommender systems, such as prediction accuracy [69]. In this case, the dataset is usually divided into (i) a training sample to build the model based on the user rating and (ii) a test sample to calculate the measurement parameters such as accuracy, precision, recall and f-score. Since our recommender system is uniquely based on the characteristics of the products and we are not considering the customer profile, this kind of offline experiment is not provided for our evaluation purposes.
However, we have decided to opt by a most direct evaluation based on the feedback from two important sectors: customers (users) and experts (workers in the food retail sector). For this, we created a large-scale experiment on a prototype through a user survey, that is, an online experiment. The results is the direct feedback and opinion of the performance of the recommender system according to the users’ perspective. Consequently, the feedback obtained would depend on a variety of factors, such as the user’s intent (for example: how specific their information needs are), the user’s context (for example: what items are they already familiar with, in addition, how much they trust the system) and the interface through which the recommendations are presented. This is a more realistic scenario and it will provide strong evidence about the recommender system’s results: that is, if the suggested product is one the user would buy instead of the required one or not, obtaining, therefore, a good value for the accuracy.
Since we do not have information about the customers (profiling, interactions etc.), we have worked on all the data without dividing the data set. In order to save time and optimize operational performance when recommending alternative products, we took the following steps. We filtered and pre-processed the data set with two methods (BOW, Doc2Vec) for each approach based on the desired product characteristics, such as variety, size, allergen, and so forth. In RS-CF, we compared the desired product with the rest of the data and ordered the alternative based on the similarity ratio. In RS-NN, we built the model for the desired product using the neural network and then classified the alternative based on the similarity ratio.
Surveys evaluated the recommendation system, which is the RS-CF user survey and RS-NN user survey, where a comparison was made between them, as shown in Figure 7, which shows the difference between the accuracy results of the two surveys considering group 3. The results showed that the RS-NN user survey performed better in all three approaches. Figure 8 shows the difference between the MSE results, as it showed that the RS-CF user survey results are the best for the PRO-COM approach for the first and second groups, but the RS-NN user survey is the best for the third group. The RS-NN user survey is the best for both approaches: PK-BD and HTH-BD. The comparatives prove that using the neural network completely alters the results, and taking price and brand into account was something that users wanted. Using more nutrition table features gives better results. It also proved that a PK-BD approach based on the HTH-BD approach is far better than relying solely on a PRO-COM approach.
Finally, we evaluated the multi-criteria RS through a user survey using MSE to calculate the average error for the responses of the users of three groups, which is the main evaluation of the users’ responses. We also use accuracy to evaluate the responses of users in group 3 only, because for the other two groups it resulted in approximately 100 % accuracy.

8. Conclusions

The recommendation idea is to implement some approaches that help a user to get the right product. The approaches are made based on the user’s interest. For example, suppose the user is interested in a specific size product or a product that does not contain an allergen, and it is not available in stock. In that case, the RS recommends a similar product with these specifications without referring to the user’s file; recommended depending only on the user’s choice. The recommendation system can recommend an alternative health product to the user. In this paper, to build a recommendation system, we used item-based collaborative filtering (RS-CF) and BOW to represent the dataset as a vector. To build an RS-CF that caters to the largest number of users, we created three approaches, which are product composition (PRO-COM), package-based (PK-BD), and the healthy-based approach (HTH-BD). Essentially, PRO-COM works to obtain a similar product based on the product’s component, whereas the PK-BD approach takes into consideration PRO-COM and adds product size to obtain a similar product. Finally, the HTH-BD approach obtains a similar product by taking PRO-COM and allergen information into account, then an equation is made, consisting of the features of the nutrition table. The user then evaluates these approaches through the survey.
After that, we refine the recommendation to suit the company’s requirements. Optimization of the RS-NN model is done using the neural network as a representation dataset and a model is created using Doc2Vec. RS-NN tries to improve the approaches by adding some considerations (such as allergen features as a pre-condition for all approaches, more features about the nutrition table and brand type, brand attribute and price) and rearranging approaches to PRO-COM, HTH-BD, and then PK-BD. A survey of experts and users was conducted to assess RS-NN. Then, we collected the result, which we compared between the models (RS-CF, RS-NN); the comparatives prove that using the neural network-based model completely alters the results.
For future work, this research will be developed, especially the health approach, so that this approach will be based on the user’s profile and not just on the product’s components, for example, creating a user profile including age, chronic disease, prominent diet, and so forth. This profile will help the recommendation system to recommend a suitable alternative healthy product for the user.

Author Contributions

Conceptualization, M.M.H., R.P.D.R., A.F.V. and H.O.P.; Data curation, M.M.H. and H.O.P.; Formal analysis, M.M.H., R.P.D.R., A.F.V. and H.O.P.; Funding acquisition, R.P.D.R. and A.F.V.; Investigation, Manar Hafez, R.P.D.R., A.F.V. and H.O.P.; Methodology, M.M.H., R.P.D.R., A.F.V. and H.O.P.; Resources, R.P.D.R. and A.F.V.; Software, M.M.H. and H.O.P.; Supervision, R.P.D.R. and A.F.V.; Validation, M.M.H. and H.O.P.; Visualization, M.M.H.; Writing—original draft, M.M.H., R.P.D.R. and A.F.V.; Writing—review & editing, M.M.H., R.P.D.R. and A.F.V. All authors have read and agreed to the published version of the manuscript.

Funding

This work has received financial support from the European Regional Development Fund (ERDF) and the Galician Regional 7 Government, under the agreement for funding the Atlantic Research Center for Information and Communication Technologies (atlanTTIC), and the Spanish Ministry of Economy and Competitiveness, under the National Science Program (TEC2017-84197-C4-2-R).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors would like to thank the European Regional Development Fund (ERDF) and the Galician Regional 7 Government, under the agreement for funding the Atlantic Research Center for Information and Communication Technologies (atlanTTIC), and the Spanish Ministry of Economy and Competitiveness, under the National Science Program (TEC2017-84197-C4-2-R).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Reinartz, W.; Wiegand, N.; Imschloss, M. The impact of digital transformation on the retailing value chain. Int. J. Res. Mark. 2019, 36. [Google Scholar] [CrossRef]
  2. Wessel, L.; Baiyere, A.; Ologeanu-Taddei, R.; Cha, J.; Blegind Jensen, T. Unpacking the Difference between Digital Transformation and IT-enabled Organizational Transformation. J. Assoc. Inf. Syst. 2020, 22, 102–129. [Google Scholar]
  3. Linden, G.; Smith, B.; York, J. Amazon. com recommendations: Item-to-item collaborative filtering. IEEE Internet Comput. 2003, 7, 76–80. [Google Scholar] [CrossRef] [Green Version]
  4. Thorat, P.B.; Goudar, R.; Barve, S. Survey on collaborative filtering, content-based filtering and hybrid recommendation system. Int. J. Comput. Appl. 2015, 110, 31–36. [Google Scholar]
  5. Grbovic, M.; Radosavljevic, V.; Djuric, N.; Bhamidipati, N.; Savla, J.; Bhagwan, V.; Sharp, D. E-commerce in your inbox: Product recommendations at scale. In Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Sydney, Australia, 10–13 August 2015; pp. 1809–1818. [Google Scholar]
  6. Shen, J.; Zhou, T.; Chen, L. Collaborative filtering-based recommendation system for big data. Int. J. Comput. Sci. Eng. 2020, 21, 219–225. [Google Scholar] [CrossRef]
  7. Bennett, J.; Lanning, S. The netflix prize. In Proceedings of the KDD Cup and Workshop, New York, NY, USA, 12 August 2007; Volume 2007, p. 35. [Google Scholar]
  8. Das, A.S.; Datar, M.; Garg, A.; Rajaram, S. Google news personalization: Scalable online collaborative filtering. In Proceedings of the 16th International Conference on World Wide Web, Banff, AB, Canada, 8–12 May 2007; pp. 271–280. [Google Scholar]
  9. Kumar, P.S. Recommendation System for E-Commerce by Memory Based and Model Based Collaborative Filtering. In Proceedings of the 11th International Conference on Soft Computing and Pattern Recognition (SoCPaR 2019); Springer: Berlin/Heidelberg, Germany, 2020; Volume 1182, p. 123. [Google Scholar]
  10. Zhang, Y.; Yang, C.; Niu, Z. A research of job recommendation system based on collaborative filtering. In Proceedings of the 2014 Seventh International Symposium on Computational Intelligence and Design, Hangzhou, China, 13–14 December 2014; Volume 1, pp. 533–538. [Google Scholar]
  11. Pirasteh, P.; Jung, J.J.; Hwang, D. Item-based collaborative filtering with attribute correlation: A case study on movie recommendation. In Proceedings of the Asian Conference on Intelligent Information and Database Systems; Springer: Berlin/Heidelberg, Germany, 2014; pp. 245–252. [Google Scholar]
  12. Bag, S.; Kumar, S.K.; Tiwari, M.K. An efficient recommendation generation using relevant Jaccard similarity. Inf. Sci. 2019, 483, 53–64. [Google Scholar] [CrossRef]
  13. Van Meteren, R.; Van Someren, M. Using content-based filtering for recommendation. In Proceedings of the Machine Learning in the New Information Age: MLnet/ECML2000 Workshop, MLNIA, Barcelona, Spain, 30 May 2000; Volume 30, pp. 47–56. [Google Scholar]
  14. Lops, P.; De Gemmis, M.; Semeraro, G. Content-based recommender systems: State of the art and trends. In Recommender Systems Handbook; Springer: Berlin/Heidelberg, Germany, 2011; pp. 73–105. [Google Scholar]
  15. Saravanan, S. Design of large-scale Content-based recommender system using hadoop MapReduce framework. In Proceedings of the 2015 Eighth International Conference on Contemporary Computing (IC3), Noida, India, 20–22 August 2015; pp. 302–307. [Google Scholar]
  16. Schafer, J.B.; Frankowski, D.; Herlocker, J.; Sen, S. Collaborative filtering recommender systems. In The Adaptive Web; Springer: Berlin/Heidelberg, Germany, 2007; pp. 291–324. [Google Scholar]
  17. Elahi, M.; Ricci, F.; Rubens, N. A survey of active learning in collaborative filtering recommender systems. Comput. Sci. Rev. 2016, 20, 29–50. [Google Scholar] [CrossRef]
  18. Yu, K.; Schwaighofer, A.; Tresp, V.; Xu, X.; Kriegel, H.P. Probabilistic memory-based collaborative filtering. IEEE Trans. Knowl. Data Eng. 2004, 16, 56–69. [Google Scholar]
  19. Koohi, H.; Kiani, K. User based Collaborative Filtering using fuzzy C-means. Measurement 2016, 91, 134–139. [Google Scholar] [CrossRef]
  20. Sarwar, B.; Karypis, G.; Konstan, J.; Riedl, J. Item-based collaborative filtering recommendation algorithms. In Proceedings of the 10th International Conference on World Wide Web, Hong Kong, China, 1–5 May 2001; pp. 285–295. [Google Scholar]
  21. Gao, M.; Wu, Z.; Jiang, F. Userrank for item-based collaborative filtering recommendation. Inf. Process. Lett. 2011, 111, 440–446. [Google Scholar] [CrossRef]
  22. Wei, J.; He, J.; Chen, K.; Zhou, Y.; Tang, Z. Collaborative filtering and deep learning based recommendation system for cold start items. Expert Syst. Appl. 2017, 69, 29–39. [Google Scholar] [CrossRef] [Green Version]
  23. Alaa, R.; Gawich, M.; Fernández-Veiga, M. Personalized Recommendation for Online Retail Applications Based on Ontology Evolution. In Proceedings of the 2020 6th International Conference on Computer and Technology Applications, Antalya, Turkey, 14–16 April 2020; pp. 12–16. [Google Scholar]
  24. Smith, B.; Linden, G. Two decades of recommender systems at Amazon. com. IEEE Internet Comput. 2017, 21, 12–18. [Google Scholar] [CrossRef]
  25. Esparza, S.G.; O’Mahony, M.P.; Smyth, B. Mining the real-time web: A novel approach to product recommendation. Knowl. Based Syst. 2012, 29, 3–11. [Google Scholar] [CrossRef]
  26. Jin, Y.; Hu, M.; Singh, H.; Rule, D.; Berlyant, M.; Xie, Z. MySpace video recommendation with map-reduce on qizmt. In Proceedings of the 2010 IEEE Fourth International Conference on Semantic Computing, Pittsburgh, PA, USA, 22–24 September 2010; pp. 126–133. [Google Scholar]
  27. Wilkinson, B.W.; McHale, B.G.; Mattingly, T.D. Systems and Methods for Providing Content-Based Product Recommendations. U.S. Patent 10,614,504, 7 April 2020. [Google Scholar]
  28. Singh, M.K.; Rishi, O.P. Event driven Recommendation System for E-commerce using Knowledge based Collaborative Filtering Technique. Scalable Comput. Pract. Exp. 2020, 21, 369–378. [Google Scholar] [CrossRef]
  29. Xu, Y.; Ren, J.; Zhang, Y.; Zhang, C.; Shen, B.; Zhang, Y. Blockchain empowered arbitrable data auditing scheme for network storage as a service. IEEE Trans. Serv. Comput. 2019, 13, 289–300. [Google Scholar] [CrossRef]
  30. Zhao, Z.D.; Shang, M.S. User-based collaborative-filtering recommendation algorithms on hadoop. In Proceedings of the 2010 Third International Conference on Knowledge Discovery and Data Mining, Phuket, Thailand, 9–10 January 2010; pp. 478–481. [Google Scholar]
  31. Meng, S.; Dou, W.; Zhang, X.; Chen, J. KASR: A keyword-aware service recommendation method on mapreduce for big data applications. IEEE Trans. Parallel Distrib. Syst. 2014, 25, 3221–3231. [Google Scholar] [CrossRef]
  32. Wang, T.; Fu, Y. Item-based Collaborative Filtering with BERT. In Proceedings of the 3rd Workshop on e-Commerce and NLP, Seattle, WA, USA, 10 July 2020; pp. 54–58. [Google Scholar]
  33. Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv 2018, arXiv:1810.04805. [Google Scholar]
  34. Ferreira, D.; Silva, S.; Abelha, A.; Machado, J. Recommendation system using autoencoders. Appl. Sci. 2020, 10, 5510. [Google Scholar] [CrossRef]
  35. Olbrich, R.; Holsing, C. Modeling consumer purchasing behavior in social shopping communities with clickstream data. Int. J. Electron. Commer. 2011, 16, 15–40. [Google Scholar] [CrossRef]
  36. Qiu, J.; Lin, Z.; Li, Y. Predicting customer purchase behavior in the e-commerce context. Electron. Commer. Res. 2015, 15, 427–452. [Google Scholar] [CrossRef]
  37. Zhang, Y.; Jin, R.; Zhou, Z.H. Understanding bag-of-words model: A statistical framework. Int. J. Mach. Learn. Cybern. 2010, 1, 43–52. [Google Scholar] [CrossRef]
  38. Jiang, H.; Yang, D.; Xiao, Y.; Wang, W. Understanding a bag of words by conceptual labeling with prior weights. World Wide Web 2020, 23, 2429–2447. [Google Scholar] [CrossRef]
  39. Chowdhury, G.G. Natural language processing. Annu. Rev. Inf. Sci. Technol. 2003, 37, 51–89. [Google Scholar] [CrossRef] [Green Version]
  40. Baeza-Yates, R.; Ribeiro-Neto, B. Modern Information Retrieval; ACM Press: New York, NY, USA, 1999. [Google Scholar]
  41. Husain, M.S. Critical Concepts and Techniques for Information Retrieval System. In Natural Language Processing in Artificial Intelligence; Apple Academic Press: New York, NY, USA, 2020; pp. 29–51. [Google Scholar]
  42. Lai, S.; Liu, K.; He, S.; Zhao, J. How to generate a good word embedding. IEEE Intell. Syst. 2016, 31, 5–14. [Google Scholar] [CrossRef]
  43. Li, Y.; Yang, T. Word embedding for understanding natural language: A survey. In Guide to Big Data Applications; Springer: Berlin/Heidelberg, Germany, 2018; pp. 83–104. [Google Scholar]
  44. Hira, Z.M.; Gillies, D.F. A review of feature selection and feature extraction methods applied on microarray data. Adv. Bioinform. 2015, 2015. [Google Scholar] [CrossRef]
  45. Wang, Z.; Ma, L.; Zhang, Y. A hybrid document feature extraction method using latent Dirichlet allocation and word2vec. In Proceedings of the 2016 IEEE first international conference on data science in cyberspace (DSC), Changsha, China, 13–16 June 2016; pp. 98–103. [Google Scholar]
  46. Yin, Z.; Shen, Y. On the dimensionality of word embedding. arXiv 2018, arXiv:1812.04224. [Google Scholar]
  47. Xing, C.; Wang, D.; Liu, C.; Lin, Y. Normalized word embedding and orthogonal transform for bilingual word translation. In Proceedings of the 2015 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Denver, CO, USA, 31 May–5 June 2015; pp. 1006–1011. [Google Scholar]
  48. Ghannay, S.; Favre, B.; Esteve, Y.; Camelin, N. Word embedding evaluation and combination. In Proceedings of the Tenth International Conference on Language Resources and Evaluation (LREC’16), Portoroz, Slovenia, 23–28 May 2016; pp. 300–305. [Google Scholar]
  49. Mikolov, T.; Chen, K.; Corrado, G.; Dean, J. Efficient estimation of word representations in vector space. arXiv 2013, arXiv:1301.3781. [Google Scholar]
  50. Wang, Q.; Xu, J.; Chen, H.; He, B. Two improved continuous bag-of-word models. In Proceedings of the 2017 International Joint Conference on Neural Networks (IJCNN), Anchorage, AK, USA, 14–19 May 2017; pp. 2851–2856. [Google Scholar]
  51. Guthrie, D.; Allison, B.; Liu, W.; Guthrie, L.; Wilks, Y. A closer look at skip-gram modelling. In Proceedings of the LREC, Genoa, Italy, 22–28 May 2006; Volume 6, pp. 1222–1225. [Google Scholar]
  52. Lazaridou, A.; Pham, N.T.; Baroni, M. Combining language and vision with a multimodal skip-gram model. arXiv 2015, arXiv:1501.02598. [Google Scholar]
  53. Le, Q.; Mikolov, T. Distributed representations of sentences and documents. In Proceedings of the International Conference on Machine Learning, Beijing, China, 21–26 June 2014; pp. 1188–1196. [Google Scholar]
  54. Hafez, M.M.; Redondo, R.P.D.; Vilas, A.F. A Comparative Performance Study of Naïve and Ensemble Algorithms for E-commerce. In Proceedings of the 2018 14th International Computer Engineering Conference (ICENCO), Cairo, Egypt, 29–30 December 2018; pp. 26–31. [Google Scholar]
  55. Hafez, M.M.; Fernández Vilas, A.; Redondo, R.P.D.; Pazó, H.O. Classification of Retail Products: From Probabilistic Ranking to Neural Networks. Appl. Sci. 2021, 11, 4117. [Google Scholar] [CrossRef]
  56. European Commission. General Food Law. 2002. Available online: https://ec.europa.eu/food/safety/general_food_law_en (accessed on 20 March 2019).
  57. Karvelis, P.; Gavrilis, D.; Georgoulas, G.; Stylios, C. Topic recommendation using Doc2Vec. In Proceedings of the 2018 International Joint Conference on Neural Networks (IJCNN), Rio de Janeiro, Brazil, 8–13 July 2018; pp. 1–6. [Google Scholar]
  58. Kim, D.; Seo, D.; Cho, S.; Kang, P. Multi-co-training for document classification using various document representations: TF–IDF, LDA, and Doc2Vec. Inf. Sci. 2019, 477, 15–29. [Google Scholar] [CrossRef]
  59. Johnson, L. How to Calculate Percentages Into Grams. Updated on 6 November 2020. Available online: https://sciencing.com/calculate-percentages-grams-6942118.html (accessed on 25 November 2020).
  60. Smith, C.C.; Follmer, D. Food preferences of squirrels. Ecology 1972, 53, 82–91. [Google Scholar] [CrossRef]
  61. Flynn, M.A.; Surprenant, T.; Craig, C.M.; Bergstrom, A. Is it good for me? A content analysis of the healthiness of foods advertised in magazines. Atl. J. Commun. 2020, 1–15. [Google Scholar] [CrossRef]
  62. Egnell, M.; Talati, Z.; Galan, P.; Andreeva, V.; Vandevijvere, S.; Gombaud, M.; Dréano-Trécant, L.; Hercberg, S.; Pettigrew, S.; Julia, C. Objective understanding of the front-of-pack nutrition label Nutri-Score by European consumers. Eur. J. Public Health 2020, 30, ckaa165.902. [Google Scholar] [CrossRef]
  63. Dréano-Trécant, L.; Egnell, M.; Hercberg, S.; Galan, P.; Soudon, J.; Fialon, M.; Touvier, M.; Kesse-Guyot, E.; Julia, C. Performance of the Front-of-Pack Nutrition Label Nutri-Score to Discriminate the Nutritional Quality of Foods Products: A Comparative Study across 8 European Countries. Nutrients 2020, 12, 1303. [Google Scholar] [CrossRef]
  64. Jamieson, J.A.; Neufeld, A. Food sources of energy and nutrients among Canadian adults following a gluten-free diet. PeerJ 2020, 8, e9590. [Google Scholar] [CrossRef]
  65. Swamynathan, M. Mastering Machine Learning with Python in Six Steps: A Practical Implementation Guide to Predictive Data Analytics Using Python; Apress: New York, NY, USA, 2019. [Google Scholar]
  66. Doan, D. A Developer’s Survey on Different Cloud Platforms. Ph.D. Thesis, University of California San Diego, San Diego, CA, USA, 2009. [Google Scholar]
  67. Stigler, S.; Burdack, M. A practical approach of different programming techniques to implement a real-time application using Django. Athens J. Sci. 2020, 7, 43–66. [Google Scholar] [CrossRef]
  68. Torabi, M.; Rao, J.N. Mean squared error estimators of small area means using survey weights. Can. J. Stat. 2010, 38, 598–608. [Google Scholar] [CrossRef]
  69. Shani, G.; Gunawardana, A. Evaluating recommendation systems. In Recommender Systems Handbook; Springer: Berlin/Heidelberg, Germany, 2011; pp. 257–297. [Google Scholar]
Figure 1. Description of the proposed model definition and evaluation.
Figure 1. Description of the proposed model definition and evaluation.
Sensors 21 03747 g001
Figure 2. More details on description of the proposed model definition and evaluation.
Figure 2. More details on description of the proposed model definition and evaluation.
Sensors 21 03747 g002
Figure 3. The modeling methodology for RS-CF.
Figure 3. The modeling methodology for RS-CF.
Sensors 21 03747 g003
Figure 4. The modeling methodology for RS-NN.
Figure 4. The modeling methodology for RS-NN.
Sensors 21 03747 g004
Figure 5. A snapshot of the web survey.
Figure 5. A snapshot of the web survey.
Sensors 21 03747 g005
Figure 6. The number of times the alternative products was chosen.
Figure 6. The number of times the alternative products was chosen.
Sensors 21 03747 g006
Figure 7. Comparative study using accuracy considering group 3.
Figure 7. Comparative study using accuracy considering group 3.
Sensors 21 03747 g007
Figure 8. The MSE of three approaches.
Figure 8. The MSE of three approaches.
Sensors 21 03747 g008
Table 1. Extract of the MDD-DS.
Table 1. Extract of the MDD-DS.
EAN Category Subcategory Variety Brand Name Ingredients Legal   Name Servings Size Unit
10,590FreshFish & ShellfishOtherGeneric MidiadiaCongrioConger congerCongrio1330mL
84,107BeveragesBeersLagerMoritzCerveza moritzBeer. 5.4% vol. alc.Ron3500g
843,654Snacks and nutsNutsSeedsFacundoGian SeedsSunflower seeds and salt (4%)Roasted and salty2250g
Table 2. Product attributes in the dataset.
Table 2. Product attributes in the dataset.
FieldLevelsDescription
EAN Unique Product Number
Category161st Level Category
Subcategory622nd Level Category
Variety1593rd Level Category, referred to as variety
Brand3015Product Brand
Name11,139Product Customary Name
Legal Name8442Official product denomination regarding the European Union provisions
Ingredients List of Ingredients in the product
Table 3. Extract of the MDD-DS1.
Table 3. Extract of the MDD-DS1.
EAN Fat Sugar Message Message
10,5902.80without sugar
84,1074.38 Room temperature
843,6542.1830 Room temperature
Table 4. Extract of the MDD-DS2.
Table 4. Extract of the MDD-DS2.
EAN Brand   Type Brand   Attribute Carbs df sf gf pn sa
10,590manufacturerstandard2.8000102
84,107manufacturerwithout gluten4.38150410
843,654Whitestandard2.183040756
Table 5. Examples of p r o d u c t _ w o r d s for every p.
Table 5. Examples of p r o d u c t _ w o r d s for every p.
p idProduct-Vector
1[’parsley’, ’fresh’, ’leek’, …]
218[’milk’, ’skimmed’, ’leek’, …]
29 , 167 [’oil’, ’parsley’, ’lemon’, …]
Table 6. Extract from a l l _ p r o d u c t s _ w o r d s .
Table 6. Extract from a l l _ p r o d u c t s _ w o r d s .
[’parsley’, ’fresh’, ’leek’, ’raw’, ’cauliflower’, ’raw’, ’thistle’, ’cynara’, ’cardunculu’, ’panettone’, ’flour’, ’wheat’, ’kind’, ’raisin’, ’egg’, ’butter’, ’sugar’, ’orange’,’candied’,’peel’,’lemon’,’syrup’, ’glucose’, ’fructose’, ’regulator’, ’acidity’, ’acid’, ’citric’, ’water’, ’yolk’, ’yeast’, ’bakery’, ’emulsifier’, ’mono’, ’diglyceride’, ’fatty’, ’salt’, ’milk’, ’skimmed’, ’powder’, ’flavoring’, ’preservative’, ’sorbic’, ’plum’, ’conger eel’, ’conger’, ’canon’, ’valerianella’, ’locusta’, ’fruit’, ⋯]
Table 7. Example of a product matrix.
Table 7. Example of a product matrix.
p id’Parsley’’Fresh’’Leek’’Oil’
11110
29,1671001
Table 8. Example of a distance vector p a = ( p a 1 , , p a t ) for PRO-COM.
Table 8. Example of a distance vector p a = ( p a 1 , , p a t ) for PRO-COM.
EAN 50,113,00976,287,83784,240,702
Distance051400
Table 9. PK-BD approach: Example of a matrix with distance vectors as columns after sorting the products.
Table 9. PK-BD approach: Example of a matrix with distance vectors as columns after sorting the products.
EAN Distance (PRO-COM)Distance (PK-BD)
84,862,23823
84,312,15526
54,487,50550
74,410,953754
Table 10. String values obtained from the Messages columns.
Table 10. String values obtained from the Messages columns.
[without colorings, without preservatives, without additives, without transgenics, without gluten, without artificial colors, without trans fats, without artificial flavors, without fat, without molluscs, without lactose, without artificial preservatives, without sugar, without egg, without milk and its derivatives, without cholesterol, without added preservatives, without added sugar, without salt, without palm oil, without soy, without added salt, without nuts, without peanuts, without palm, without sesame, without peanuts, without sulfites, without mustard, without saturated fat, without alcohol, without calories, without caffeine, without sweeteners, without hydrogenated fats, without palm oil and fat, high protein, without added phosphates, without allergen, starch free, celery free, without artificial sweeteners, without fish, without crustaceans, without glutamate, without lupins, low fat, low in energy]
Table 11. Allergen features in the withoutwords vector.
Table 11. Allergen features in the withoutwords vector.
[without allergens, without gluten, without crustaceans, without egg, without fish, without peanuts, without peanuts, without soy, without milk and its derivatives, without lactose, without nuts, without celery, without mustard, without sesame, without sulphites, without lupins, without molluscs]
Table 12. Example of the list which contains the vectors of words belonging to each product considering the Ingredientes column (product_words).
Table 12. Example of the list which contains the vectors of words belonging to each product considering the Ingredientes column (product_words).
p id Des ( p )
1[“oil”, “olive”, …]
2[“oil”, “olive”, …]
20,259[“lettuce”, “green”, …]
Table 13. Example of the list which contains the vectors of the features included in the Messages columns belonging to each product (withoutlist).
Table 13. Example of the list which contains the vectors of the features included in the Messages columns belonging to each product (withoutlist).
p id Messages
1[“without preservatives”, “low fat”]
2[“without gluten”]
20,259[“without peanuts”]
Table 14. Example of the vector which contains all the different features obtained from the Mensajes columns (withoutwords).
Table 14. Example of the vector which contains all the different features obtained from the Mensajes columns (withoutwords).
1250
“without colorants”“without preservatives”“low in energy”
Table 15. HTH-BD: Example of a matrix with the considered criterions as columns after sorting the products.
Table 15. HTH-BD: Example of a matrix with the considered criterions as columns after sorting the products.
EAN Similarity Fat + Sugar Healthy FeaturesProcessing Level
6,431,6491465
7,358,8021430
652,10800.4701
4,452,030010.26415
Table 16. Products per variety after eliminate first quarter.
Table 16. Products per variety after eliminate first quarter.
ApproachesCountMinMedianMeanMax.
PRO-COM342154163.7501
HTH-BD30341426.5233
PK-BD29241324.3203
Table 17. Examples of p r o d u c t _ w o r d s for every p.
Table 17. Examples of p r o d u c t _ w o r d s for every p.
Steps Des ( p )
Original
(p)
[’Wine’, ’White’, ’Sauvignon’, ’Blanc’, ’Type’, ’Grape’, ’Sauvignon’, ’Blanc’, ’13’, ’Vol’, ’Alc’]
(1)[’wine’, ’white’, ’sauvignon’, ’blanc’, ’type’, ’grape’, ’sauvignon’, ’blanc’, ’13’, ’vol’, ’ alc ’]
(2)[’wine’, ’white’, ’sauvignon’, ’blanc’, ’type’, ’grape’, ’sauvignon’, ’blanc’, ’vol’, ’alc’]
(3)[’wine’, ’white’, ’sauvignon’, ’blanc’, ’type’, ’grape’, ’vol’, ’alc’]
Table 18. The first two d e s ( p ) by Tagged_Document.
Table 18. The first two d e s ( p ) by Tagged_Document.
[TaggedDocument(words=[’dessert’, ’dairy’, ’apricot’, ’milk’, ’whole’, ’water’, ’sugar’, ’starch’, ’modified’, ’corn’, ’puree’, ’rice’, ’stabilizer’,’pectin’, ’oil’, ’rapeseed’, ’caseinate’, ’sodium calcium’, ’contains’, ’aroma’, ’natural’, ’corrector’, ’acidity’, ’acid’, ’lactic’, ’food’, ’infant’, ’sugary’, ’derivative’, ’included’, ’lactose’], tags = [’0’]),
TaggedDocument (words = [’dessert’, ’dairy’, ’cocoa’, ’milk’, ’whole’, ’water’, ’sugar’, ’thickener’, ’starch’, ’modified’, ’corn’, ’gum’, ’locust bean’, ’powder’, ’oil’, ’rape’, ’caseinate’, ’soda-calcium’, ’contains’,’ corrector’, ’acidity’, ’hydroxide’, ’sodium’, ’food’, ’infant’, ’sugary’, ’derivative’, ’included’, ’lactose’], tags = [’1’])]
Table 19. The vector for first product a ( p 1 ) .
Table 19. The vector for first product a ( p 1 ) .
array([ 0.00742837, −0.00540146, −0.14862166, −0.00862698, 0.31875622, 0.115518, 0.00795528, −0.06915003, 0.03247217, −0.12760445, −0.20222402, −0.09181757, −0.02992765, −0.09429716, 0.04839283, −0.08727524, −0.08463322, −0.09556159, −0.01945411, −0.0644968, 0.11707045, −0.09715877, −0.24429108, −0.08826657, −0.12004123, −0.17009708, 0.17322347, −0.04258763, 0.03453251, −0.19297938, −0.2081344, 0.23702264, 0.08457132, 0.0120729, 0.03960438, −0.21322013, 0.09752178, −0.03770451, −0.06469689, 0.02615795, 0.20623626, −0.09590556, −0.00720048, −0.12926176, −0.21335329, −0.11945274, 0.06031954, 0.0124997, 0.27832198, −0.10382865], dtype = float32)
Table 20. The MSE considering the three approaches as well as the different groups of products tested.
Table 20. The MSE considering the three approaches as well as the different groups of products tested.
ApproachesGroup 1Group 2Group 3
PRO-COM0.138850.215390.26187
PK-BD0.224230.253040.26695
HTH-BD0.337310.360020.36002
Table 21. The accuracy considering the group 3.
Table 21. The accuracy considering the group 3.
ApproachesGroup 3
PRO-COM81.33%
PK-BD79.28%
HTH-BD70.77%
Table 22. The MSE considering the three approaches and the three groups of products evaluated (Second survey).
Table 22. The MSE considering the three approaches and the three groups of products evaluated (Second survey).
ApproachesGroup 1Group 2Group 3
PRO-COM0.2138460.2293720.235824
HTH-BD0.2107690.2107690.210769
PK-BD0.1807690.1880.188
Table 23. The accuracy of user survey using ML (group 3).
Table 23. The accuracy of user survey using ML (group 3).
ApproachesGroup 3
PRO-COM82.13058%
HTH-BD83.38461%
PK-BD87.68%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Hafez, M.M.; Redondo, R.P.D.; Vilas, A.F.; Pazó, H.O. Multi-Criteria Recommendation Systems to Foster Online Grocery. Sensors 2021, 21, 3747. https://0-doi-org.brum.beds.ac.uk/10.3390/s21113747

AMA Style

Hafez MM, Redondo RPD, Vilas AF, Pazó HO. Multi-Criteria Recommendation Systems to Foster Online Grocery. Sensors. 2021; 21(11):3747. https://0-doi-org.brum.beds.ac.uk/10.3390/s21113747

Chicago/Turabian Style

Hafez, Manar Mohamed, Rebeca P. Díaz Redondo, Ana Fernández Vilas, and Héctor Olivera Pazó. 2021. "Multi-Criteria Recommendation Systems to Foster Online Grocery" Sensors 21, no. 11: 3747. https://0-doi-org.brum.beds.ac.uk/10.3390/s21113747

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop