Next Article in Journal
Rapid Manufacturing of Multilayered Microfluidic Devices for Organ on a Chip Applications
Next Article in Special Issue
INIM: Inertial Images Construction with Applications to Activity Recognition
Previous Article in Journal
An Identification Method for Rotor Direction Based on Charge Induction
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Gaze and Event Tracking for Evaluation of Recommendation-Driven Purchase

by
Piotr Sulikowski
1,*,
Tomasz Zdziebko
2,
Kristof Coussement
3,4,
Krzysztof Dyczkowski
5,
Krzysztof Kluza
6 and
Karina Sachpazidu-Wójcicka
7
1
Faculty of Information Technology and Computer Science, West Pomeranian University of Technology, ul. Żołnierska 49, 71-210 Szczecin, Poland
2
Faculty of Economics, Finance and Management, University of Szczecin, ul. Adama Mickiewicza 64, 71-101 Szczecin, Poland
3
IÉSEG School of Management, 3 Rue de la Digue, F-59000 Lille, France
4
LEM-CNRS 9221, 3 Rue de la Digue, F-59000 Lille, France
5
Faculty of Mathematics and Computer Science, Adam Mickiewicz University, ul. Uniwersytetu Poznańskiego 4, 61-614 Poznań, Poland
6
Faculty of Electrical Engineering, Automatics, Computer Science and Biomedical Engineering, AGH University of Science and Technology, al. Adama Mickiewicza 30, 30-059 Kraków, Poland
7
Faculty of Business and Management, Wrocław University of Economics and Business, ul. Komandorska 118/120, 53-345 Wrocław, Poland
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(4), 1381; https://doi.org/10.3390/s21041381
Submission received: 13 January 2021 / Revised: 4 February 2021 / Accepted: 11 February 2021 / Published: 16 February 2021
(This article belongs to the Special Issue Vision and Sensor-Based Sensing in Human Action Recognition)

Abstract

:
Recommendation systems play an important role in e-commerce turnover by presenting personalized recommendations. Due to the vast amount of marketing content online, users are less susceptible to these suggestions. In addition to the accuracy of a recommendation, its presentation, layout, and other visual aspects can improve its effectiveness. This study evaluates the visual aspects of recommender interfaces. Vertical and horizontal recommendation layouts are tested, along with different visual intensity levels of item presentation, and conclusions obtained with a number of popular machine learning methods are discussed. Results from the implicit feedback study of the effectiveness of recommending interfaces for four major e-commerce websites are presented. Two different methods of observing user behavior were used, i.e., eye-tracking and document object model (DOM) implicit event tracking in the browser, which allowed collecting a large amount of data related to user activity and physical parameters of recommending interfaces. Results have been analyzed in order to compare the reliability and applicability of both methods. Observations made with eye tracking and event tracking led to similar results regarding recommendation interface evaluation. In general, vertical interfaces showed higher effectiveness compared to horizontal ones, with the first and second positions working best, and the worse performance of horizontal interfaces probably being connected with banner blindness. Neural networks provided the best modeling results of the recommendation-driven purchase (RDP) phenomenon.

1. Introduction

The continuous growth of e-commerce, especially nowadays due to the coronavirus disease (COVID-19) pandemic, requires tools, particularly recommendation systems that aim to help discover relevant products for individuals. While online shopping benefits generally exceed disadvantages, lack of personal touch, especially when a customer is overwhelmed by many alternative products is perceived as a major obstacle to shopping online. To alleviate this problem, online stores invest in personalization tools, recommendation systems being a very popular example of them. Recommendation adequacy plays a vital role in overall commercial success and customer satisfaction. However, due to the significant loads of marketing content and increasing browsing speeds, users may often not notice the displayed recommendations. Therefore, the way of presenting a recommendation, and its positioning and usability seem to have an important role in the ultimate recommendation effectiveness.
In a personalized recommendation system, a user model is created based on user demographic data and his/her behavior and transactions [1]. A digital representation of this model is a user profile, which usually reflects the user’s preferences and needs. Online websites can collect a broad spectrum of data [2,3,4,5] concerning users’ demographics and behavior, which are used to construct user profiles and generate personalized recommendations aimed to generate sales.
Most popular recommender engines use well-established, collaborative filtering with improvements, e.g., including user interest [6,7]. Another type of content-based recommender engine relies on presenting products similar to products the user liked in the past. The major advantages and disadvantages of each approach are presented in Table 1. Recently, the use of deep neural networks has grown in interest among recommender system researchers [8]. Shahriar et al. (2020) used eye tracking to learn user behavior by counting how many times a user looked at a particular website element in order to generate a cluster of similar products to constitute a base for recommendations [9].
The recommendation system’s performance and effectiveness depend on the underlying recommendation algorithm itself, but it also goes far beyond this point [10]. There is much less research in selecting the best ways the products should be presented to the customer compared to massive research in the field of recommendation algorithms [11,12]. The most appropriate ways of presenting a recommendation item to users can be learned by observing human–computer interaction, with websites and recommending interfaces in particular. While interacting with websites, human behavior can be observed via eye tracking [13] or by collecting events inside modern web browsers [2], based respectively on gaze sensors and mouse/keystroke sensors. The investigation of visual aspects of recommending interfaces could better integrate those interfaces in online stores [10,14]. The position and layout of a recommendation zone, the number of recommendations, the size of associated images, product titles and prices, etc. can all be evaluated toward interface improvement [15]. In the era of information overload, in particular enormous marketing content, the habituation effect often comes into action, resulting in the banner blindness phenomenon [16]. Even the best recommendations from the algorithm perspective may have little effect unless they are properly presented [16,17,18]—with the right layout and position on a website, at the right moment of the browsing and purchase process, with the right visual intensity (understood as the level of content intrusiveness) [19] and in the right context, etc. [20,21,22,23].
This paper is a substantial extension of a conference paper [24], which presented the initial results of an eye-tracking experiment on visual intensity and layout for a simple-structure recommending interface with regard to reducing habituation. The general aim of the study was to evaluate two layouts and three visual intensity levels of item presentation and varied positions of recommendations in a basic-structure recommending interface, while the levels of attention drawn and inspired product interest were monitored.
This paper focuses on the analysis of the effectiveness of recommendation interfaces based on the eye-tracking research in relation to purchasing based on recommendations using a number of machine learning methods. The eye-tracking study results were compared with the results of an implicit feedback study performed via browser add-ons on a group of participants for major e-commerce websites in Poland. The results were analyzed and models for predicting recommendation effectiveness were built using different machine learning methods to find the best-performing ones. This paper’s original contribution also relates to using two different methods of tracking user behaviors and thus user interest—eye tracking and collecting implicit feedback via a web browser add-on—to compare their applicability for building automated mechanisms of recommending interface personalization.
The remainder of the article is structured as follows: Assumptions for the study and methodology are presented in Section 2. The structure and procedures of the experiment are provided in Section 3. Results are discussed in Section 4, and conclusions are presented in Section 5.

2. Methodology

The study’s main objective is to verify the effectiveness of recommending interfaces of e-commerce websites with their visual aspects in mind using gaze tracking and human–computer interaction events tracking. Different layouts (vertical versus horizontal) different positions, and varying visual intensities of recommending interfaces were compared with regard to attracting customer interest and recommendation-driven purchase (RDP) from the perspective of user experience and business goals.
While users are usually most interested in the main content of a website, positioning and modifying the intensity of the most accurate recommendations so that they would attract the highest possible attention of users allows for the achievement of various website goals. In turn, that could lead to increased customer satisfaction, retention, long-term increased sale, and finally boosting profits.
User interest can be determined by asking the user explicitly or by observing the interaction implicitly in order to infer it. Unfortunately, explicit questioning may disturb natural behavior as it adds an additional burden on the user [25,26,27]. Furthermore, in the case of fast Internet browsing, there could be a lot of unconscious attraction caused by some webpage parts. Therefore, inobtrusive implicit measures are better suited for the purpose of the study, allowing the monitored subjects to focus only on tasks performed, not causing additional cognitive load, and not requiring unique motivation to provide explicit ratings [28,29].
For the purpose of this study, the results of experiments using various observation techniques have been analyzed. Eye-tracking technology was used for implicit user visual behavior observation and analysis. Eye tracking is the most popular technique of observing human visual activity and providing a large amount of implicit feedback and can be performed with standalone or wearable devices. Within the scope of the study, gaze-based data were used and interpreted in a simple e-commerce scenario. Eye movements were used to discover which areas of a shopping website are mostly looked at, attracting user attention as being the most relevant to the user. Raw data were processed and analyzed with the help of popular eye-tracking software and data analytics algorithms to generate metrics regarding eye fixation and gaze.
Although eye movements are often unconscious and chaotic in nature, they are generally tightly connected with cognitive processes [30]. Thus, based on gaze data from eye tracking, it is possible to conclude about user attention and interest. Gaze is an invaluable source of data for providing information on how much attention a user pays to contents on the screen [31]. For the following analyses, total fixation duration is the main gaze-based measure used. This parameter is used as an indicator of attractiveness by a lot of other studies [32,33,34]. It is calculated as the sum of durations of fixations aggregated on a website section, an area of interest (AOI), one with recommendation content (RC), i.e., presentation of the recommended items, and the main section with editorial content (EC), i.e., the main viewed product description, pictures, etc.
An important property of a recommending interface on a website, in addition to the positioning of recommending interface and recommended items within that interface, is the aspect of visual intensity [35], which is also considered in the study. Changing visual intensity is a popular marketing approach to counteract habituation and attract more attention [35]. Three basic levels of intensity were used.
Another possible technique for implicit monitoring of user behavior on websites is based on human–computer interaction monitoring solutions such as scripts or browser extensions that allow registering events resulting from interacting with the website on the client’s side. With this method, it is possible to observe user behavior unobtrusively without extra attention from them [36] or additional equipment.
Our e-commerce customer preference monitor (ECPM) was built in order to collect e-customer behavior data [2]. This monitoring tool was implemented as an extension for the Mozilla Firefox browser, with the core monitoring code being easily detachable and available for external use. ECPM allows for monitoring of human–website interaction by exploiting document object model (DOM) events. HTML document object model events allow JavaScript to register different event handlers on elements in an HTML document. Our ECPM unobtrusively collects data on visited product pages and user interactions, recording numerous parameters describing the page’s physical attributes and interactions. Among many collected parameters connected with the recommending interfaces were times of mouse cursor spent inside recommendation zones and their physical size measured as the number of characters within recommended products section. As other research shows, mouse position may be quite strongly correlated with eye gaze [37,38,39]. Tracking mouse movements can be used as a lower-quality and easier to perform alternative to eye tracking for the purpose of identification of user interest. Measured times with cursor being over recommended products were related to other measures such as total time spent on the product page and page height and length measured by a number of characters, images, etc. In this way, relative measures of cursor times spent inside particular recommendation interfaces were generated.

3. Conceptual System Architecture and Research Experiment Procedure

3.1. Conceptual System Architecture

The conceptual system architecture is presented in Figure 1. The goal of the system is to adapt recommending interfaces to maximize its efficiency. The core of the system is recommending interface efficiency models based on user browsing characteristics. Adaptation is performed on recognized user behavior characteristics. Those models are generated based on data gathered from all users via implicit tracking of their behavior inside browsers. Current use is matched against models’ database, and the most efficient recommending interface is to be selected.

3.2. Research Experiment Procedure

The research experiment procedure is presented in Figure 2. Results presented in this paper are based on two experiments on recommending interfaces in e-commerce websites, described in detail in the next two sections, one performed with eye-tracking equipment and software on a smaller group of participants, and the other performed using our own implicit browsing behavior monitoring tool in the form of a web browser extension—on a larger group of participants. Both experiments used simple e-commerce scenarios where participants were browsing products and expressing interest in them. The main goal was to analyze recommending interface efficiency for two groups of users with different ways of collecting data about users’ activity. Results from both methods have been cross-compared in order to estimate their applicability for automatic recommending interface personalization in e-commerce scenarios, but not limited to them.
Figure 3 shows the unified modeling language (UML) activity diagram depicting implicit data processing with our ECPM behavior monitoring system. Three decision paths exist between the start point and the finish point while the activity is being executed for different types of events.

3.3. Eye-Tracking Experiment Structure and Procedure

This section describes the experiment performed to collect behavior data, which were used to train five different classifiers responsible for evaluating of recommending interfaces. The experimental group of users consisted of 52 people, 14 women and 38 men, all of whom produced valid eye-tracking data. Most of them were attracted by advertisements for the study or invited in person. They ranged in age from 14 years old to 54 years old (mean = 25.2, σ = 8.0). One limitation of the study was the large difference in the ratio of women to men and the big age span, although 90.4% of the participants were 18–39 years old. The group of subjects taking part in the study was not randomly selected and did not constitute a representative subset of the whole population.
The study design was within subjects. All participants were given the same task—to shop online to furnish a studio apartment with six kinds of furniture fully. A subject was asked to go from one product category to another and choose an item from each category. For the purpose of the experiment, a dedicated e-commerce website was developed using Drupal CMS. The website was composed of a recommending interface. Each participant was provided with the same website and the same product sets. The pages were written in Polish and consisted of the title, menu, product images, and descriptive text. The website covered functions such as product list, buying cart, and recommendations.
The editorial content was placed in the central part of the screen, under the main menu, and included product lists that were approximately three-screen pages long and contained 10 products for each product category. Each product had three unique features—name, an image of the product, and price. Products in a category were quite similar visually and similarly priced. There were six product categories PCj— wardrobes, chests of drawers, beds, bedside cabinets, tables, and chairs. Moreover, an “Add to Cart” button was placed under the description of a piece of furniture allowing to store customer selections in a database. From the moment a product is selected, its short description is available in cart preview and the main cart page. It is possible to remove a product from the cart in order to allow a user to make necessary changes to the final selection.
There were two alternative recommendation modes, i.e., horizontal and vertical recommending interface layout, which means that the RC section was anchored in one of two dedicated parts of the screen below the main menu, either on the left side of the page next to the general product list (vertical mode) or at the top of the page, above the general product list (horizontal mode). Only one recommendation layout was available at one time, i.e., when the horizontal mode was on, the vertical one was deactivated and vice versa. Figure 4 and Figure 5 show two alternatives to the general location of the recommended content.
The RC section always consisted of four recommendation items, RC1 to RC4, randomly selected from all products in a given category. The section in each variant did not change its location on the screen throughout browsing products in a product category, regardless of the EC section’s user scroll. Only the general product lists were actually scrollable in order to ensure reliable subject exposure to recommendations.
It was ensured that product features, i.e., name, image, and price did not stand out from other products in a category. It was intended that the distinction of a particular RCi location would be accomplished only by means of visual intensity (VI). Three levels of intensity were used—standard (no highlight) VI1, flickering (slowly disappearing and reappearing every 1–2 s) VI2, and background in red VI3. For each product category, there was a maximum of one RCi at VI2 or VI3. An example of the visual intensity of the last kind (VI3) is presented in Figure 6.
An experimental run proceeded in a way that the examined person was explained what the device for sensing eye movements was, and the eye tracker was calibrated using Gazepoint Control software (Gazepoint, Vancouver, BC, Canada) and a nine-point calibration method. For better accuracy, the calibration was always performed twice the first time to familiarize the subject with the process. There was a double monitor setup, and the operator’s screen was invisible to the participant. Due to the correct calibration, it was possible to determine precise coordinates of the place the user was looking at. We did not employ data loss measures because in these short tasks with a given goal, the data loss rate is very low. To overcome any data loss, calibration was performed twice, as mentioned earlier. All recordings from the eye tracker were also carefully analyzed in order to confirm the proper performance of the experiment. There was no eyewear that could interfere with pupil tracking.
Each participant was informed what their task was, but they were not told about the study’s goal yet. After this introduction, the subject had to furnish the apartment. After choosing one item from a category, the subject clicked “Next” and was automatically moved to the next category. Category after category, the visual intensity of recommendation items changed each time, following the same order for each participant. Furthermore, for the first three categories, the RC layout was vertical. After moving to the fourth category, it changed to horizontal and remained this way for the following categories. On the whole, there were at least six subsequent pages with different recommendation variants presented to each participant.
The eye-tracker device used was Gazepoint GP3 (Gazepoint, Vancouver, BC, Canada) with a 60 Hz update rate (satisfactory taking into account the nature of the research and observing fixation times), the device’s nominal accuracy being 0.5–1° of visual angle, the range of depth movement ±15 cm, calibration 5- and 9- point. Every session was recorded and controlled live using Gazepoint Analysis 4.3.0 software (www.gazept.com (accessed on 16 February 2021)), and it was double-checked continuously on the operator’s monitor that the experimental setup was correct. After completing the task by the participant, basic data such as age were collected, and a question was asked whether a subject felt recommendations influenced them. Finally, all data were saved and stored for further analysis. A typical experimental run lasted around 12 min.
Results from Gazepoint Analysis software have been preprocessed with a simple algorithm whose goal was to calculate total fixation time while the user gazed at particular areas of interest. The fixation point-of-gaze (POG) coordinates data FPOGX and FPOGY, which provide the user’s point-of-gaze as determined by the internal fixation filter, have been used together with information about the location of particular sections (e.g., recommending interface elements) of analyzed websites. Based on those data, the total time while the user was actively gazing at particular elements of the website was calculated. Only valid fixation POG data points (flag FPOGV) have been counted. It needs to be noted that some gaze tracking signals may occur in the absence of any specific state. The nature of gaze itself is variable due to breathing, the gravity of expressed interactions, etc.

3.4. Implicit Event Tracking via ECPM Browser Extension

The ECPM tool was instrumented for five major Polish online stores: Merlin.pl, Agito.pl, Electro.pl, Komputronik.pl, and Morele.net. From the perspective of assortment at the time of the study, Agito.pl and Merlin.pl were horizontal shops, whereas Electro.pl and Morele.net offered mostly electronic goods. Merlin.pl was a major Polish online bookstore.
All 85 participants for the study were non-profit volunteers who were active web users located in Poland, aged between 19 years old and 33 years old, all holding at least a high school degree. The task given to participants was to browse for interesting products and rate them. After leaving the product page, a rating form was displayed, where the user could express his/her level of being pleased by the product. User activity was monitored at the most granular level. Every DOM-fired event connected with the mouse and keyboard was registered together with detailed information about the source element and its position in the structure of a web page. In addition, the HTML code of every visited product page was collected. Those data were used to generate parameters describing user behavior, which were classified into four groups—describing product page attributes, describing user interaction times (in ms), describing user behavior per se, and also relative parameters describing user behavior.
During the study, 1396 products were rated, and all customer interactions with websites were monitored via ECPM. About half of the participants rated below 14 products, while the upper quartile rated more than 20 products. Higher ratings were more popular than lower ones. Among many parameters, ECPM monitored user interaction with recommending interfaces. The main monitored measure was the total time while the mouse pointer was positioned over other recommended products section. This measure was used to reflect user interest in the recommendation section.

3.5. Comparing Visual Aspects of Recommending Interfaces

This section describes the procedure employed to compare visual aspects of the recommending interfaces for observations made with eye tracking and those based on DOM events. The size and resolution of the screen were mostly the same for both studies. The eye-tracking study was conducted in a university laboratory on the same equipment. The implicit tracking with ECPM was conducted in a university laboratory and at personal computers of volunteers taking part in the study from home. The resolution and size of the screen were recorded in the study and were the same or very similar for different participants.
Data collected using the eye-tracking device were used to build a classification model with five different classification modeling methods. Gazepoint Analysis software was used to gather fixations made by participants and process them into lines of data. Overall, 15,922 fixations lines of records have been stored. Those records have been pre-processed to generate features appropriate for analysis and classification modeling. Every row of the resulting dataset contained features shown in Table 2. Only behavior-related parameters were used, without any demographic data, compared to previous studies [24,40].
As a measure of recommendation efficiency add-to-cart actions were used, being the most appropriate due to the selection and purchase task given to participants. The pre-processed data from the eye-tracking study were used for building classification models using five different modeling methods—logistic regression, decision trees, random forests, neural networks, and support vector machines. Those methods were selected as being commonly used and well established [41,42]. Our goal was to compare those methods in terms of classification quality in order to select the most suited one for the task of evaluating recommending interfaces.
The whole data set was split into training and testing sets with a proportion of 70% to 30% with stratified sampling, which resulted in roughly the same proportion of each class cases between both training and test sets. The testing set was used to evaluate models’ metrics. The training was performed with selected values for the hyper-parameters presented in Table 3.
Detailed data collected with ECPM were used to extract parameters concerning user behavior in recommendation zones together with features of the recommending interface and product pages. Every row contained the following features presented in Table 4.
Those parameters were used to compare the effectiveness between different recommendation layouts—horizontal and vertical. As a measure of effectiveness, the time of the mouse being positioned over the recommending interface was used as the most appropriate due to the goal that was given to participants of the study. This measure was related to different physical attributes of the page. Komputronik.pl store was omitted from the analysis, in opposition to previous studies, because it contained only vertical recommending interfaces.
Finally, results from the experiments were compared in order to find similarities or dissimilarities in evaluations of recommending interfaces.

4. Results and Evaluation

4.1. Eye-Tracking Results of Recommending Interface Efficiency Evaluation

Analysis of data collected with eye trackers shows that finishing the task averagely took 2.3 min. Overall, 312 products were chosen for purchase in the preliminary study. The time of fixation on the recommending interface was on average 16.3 s per participant, which constitutes 12% of the mean task finishing time. Subjects devoted an averagely of 8.2 s for observing vertical RC and 8.1 s for horizontal RC. This means that two presented layout variants of the recommending interface (horizontal and vertical) performed equally in fixation times.
The distribution of fixation times for all recommendation item positions is shown more exhaustively in Table 5. The first three RCi positions were most favorable regardless of the layout type, whereas the fourth positions in the list were the least eye-catching ones. Surprisingly, the most popular of all was the RC3 location in a horizontal layout (3.9 s). This might have been affected by the fact that this recommendation item was located right above the main product list. Next in terms of popularity were sections RC2 and RC1, respectively, situated in the vertical layout. The clear popularity of RC2 in this layout was caused by the flickering (VI2) effect shown in one product category. The popularity of RC1 although always shown without any extra visual intensity effects (VI1) may be resulting from that many people tend to perceive the first/top location in a list being the best one. In the case of the vertical layout, the first location performed better than RC3 despite presenting it with very strong intensity VI3 for one product category. Both items RC3 in the vertical layout and RC2 in the horizontal layout performed equally despite that the latter was distinguished with flickering effect (VI2) for one product category.
From the sales view, 12% of products were added to carts directly from the recommending interfaces, while around 13% of products were firstly seen by subjects in RC and added from the main EC. This constitutes roughly 25% of all products being added to the cart directly from RC or indirectly (influenced by RC). Around 75% of products were selected directly from the main EC without any influence from RC. Coincidentally, it is exactly the same proportion as the one of recommending interface fixation time to task completion time, which somehow shows the importance of focusing attention on recommended items. Vertical RC layout accounted for 60% of adding to cart actions, while the rest were thanks to horizontal RC layout—the vertical one turned out to be much more effective than the horizontal. This may connect with the so-called banner blindness phenomenon, where banners have often been located in the same site area as horizontal recommendations in the analyzed experiment.
The vertical layout resulted in a much better rate of adding products to the cart despite a similar average time of gaze for both interfaces. It is presumed that this may be caused by the fact that the horizontal recommending interface was located at the top of the page directly under the main menu. This location of the menu required often forced unintentional gaze at recommending interface. Initial analysis of eye-tracking video recordings confirmed this hypothesis in many cases. After adding products from vertical RC or main EC, users were often stopped at horizontal RC before selecting the next category from the menu.
In the case of the vertical layout, for RC with all RCi at standard intensity level (VI1), the purchases directly or indirectly driven by recommendations (RDPs) were equally distributed between recommended products; for RC with RC2 at a flickering intensity level (VI2), the item attracted 7 out of 18 RDPs in the product category; for RC with RC3 in the red background (VI3), the item attracted surprisingly only 2 out of 15 RDPs in the product category. Overall, RC2 performed best, which means that the second recommendation in the vertical list generally resulted in the biggest sales (43% of RDPs for vertical RC, and 35% of all RDPs). The volumes of recommendation driven purchase per participant are presented more exhaustively in Table 6.
Additional side remark resulting from the visual analysis of fixations paths shows that the flickering effect (VI2) of a recommendation item appeared to have an extended effect on fixation after moving to the next product category, which means that despite the visual intensity reverted to standard this recommendation location continued to attract more attention of users.
After completing the task, a survey was conducted, and 33% of the participants responded that they felt their selections were affected by recommendation content (6% felt strongly about it). In contrast, the rest responded that they felt the opposite, including 52% who claimed definitely that they ignored recommendations. This last group showed strong resistance to recommendations—when presented the RC elements after the test, some were surprised to see they had probably neglected most of them, taking them for unwanted advertisements, which suggests the effect of habituation occurred.

4.2. Modeling Recommending Interface Efficiency Based on the Eye-Tracking Experiment

Five different classification methods were used to model the recommending interface efficiency based on add to basket actions. Models were built separately for three different target variables, namely, add_to_cart_direct, add_to_cart_indirect, and add_to_cart_both. Model validation techniques and hyper-parameters used for training various classifiers have been presented in Table 3.

4.2.1. Classification Accuracy for Target Parameter add_to_cart_both

For target parameter add_to_cart_both, ranking of variables with Gain ratio and Gini is mostly similar showing that most valuable variables are fixation_time_location, share_time_location_category, rc_location_intensity, fixation_time_location_layout, and share_time_layout_category. On the opposite side, variables with the lowest ranks are fixation_time_category, rc_location, and rc_layout.
All models targeting the add_to_cart_both parameter have very good precision measures with accuracy classification score (CA) above 90% for all models and area under the curve (AUC) having high values for all models except Binary Tree (Table 7).
All models obtained for the target parameter being add_to_cart_both with accuracy measured as average over classes have relatively high precision with Neural Network performing best (Table 8).
For the target class being 1, which denotes the action of direct or indirect addition to cart, all models have lower precision compared to models with accuracy measured as average over classes. However, precision and specificity are still very high (Table 9).
For the target class being 0, which denotes that the product was not added to the cart directly or indirectly from the recommending interface but was added from editorial content without registering gaze at this product within the recommendation zone. These models have the highest classification precision of all three types of models meaning that models are performing best in predicting not adding product to basket recommending interface directly or indirectly (Table 10).

4.2.2. Classification Accuracy for Target Parameter add_to_cart_direct

For the target parameter add_to_cart_direct, a ranking of variables with Gain ratio and Gini is mostly similar to where add_to_cart_both was denoted as the target variable. The only difference is the lower rank of rc_location_intensity. General precision metrics, both AUC and CA, for all models are very high showing their outstanding capabilities (Table 11).
With accuracy measured as average over classes, all models have precision above 93% with the Binary Tree, also having the highest specificity (Table 12).
For the target class being 1, which denotes the action of direct addition to cart from the recommending interface, all models have much lower precision compared to models with accuracy measured as average over classes (Table 13). Again, Binary Tree is the best with 69.2% precision.
Precision for not adding product to cart directly or indirectly from recommending interface is again higher for all models (Table 14).

4.2.3. Classification Accuracy for Target Parameter add_to_cart_indirect

For the target parameter add_to_cart_indirect, a ranking of variables with Gain ratio and Gini is mostly similar to other target variables. The only difference is the higher rank of rc_location_intensity.
Models for indirectly adding products to cart have different accuracy classification scores with Random Forests and Neural Network having the highest values (Table 15).
With accuracy measured as average over classes, all models have precision above 86%, with Neural Network having the highest value (Table 16).
For the target class being 0, which means that products have been seen only on the main EC and added to cart from that place, all models have relatively high precision (Table 17).
For the target class being 1, which means that the user gazed at a product before adding it to the cart from EC, all models have relatively low precision from 52% to 64.8% (Table 18).
Classification accuracy metrics (AUC, CA) for all three target variables obtained for different models are presented for easier comparison in Figure 7.

4.3. Results of Recommending Interfaces Efficiency Evaluation Based on the Event Tracking Study

Comparing recommending interfaces efficiency based on data gathered during our ECPM-based experiment shows that average prod_recommend_time was significantly higher for vertical layout than horizontal layout in case of three of four shops (Table 19).
Only Morele.net was in the opposite, which may well be caused by a higher number of horizontal recommending zones than vertical (two versus one) in this shop. The parameter that shows the level of interest in recommendation rel_recommend_time_recommend_length compared to its length proves that vertical zones gained more attention than horizontal ones. Other relative parameters presented in Table 19 confirm this tendency.

5. Conclusions

Recommender systems are commonly used to draw customers’ attention to products being consistent with their preferences. They are expected to result in higher conversion rates, purchase volumes, and client retention. This study showed the influence of the recommending interface on user behavior in simply structured and real-world e-commerce stores. By combining evaluation based on two different methodologies—eye tracking and implicit behavior event tracking inside the browser, which is an original contribution of this paper—a large amount of data related to user activity and physical parameters of recommending interfaces were collected. Results were analyzed in order to compare the reliability and applicability of both methods, and the classification quality of popular modeling methods was verified with regard to recommending interfaces evaluation.
In the eye-tracking experiment, an average of 12% of task completion time was devoted to looking at the recommending interfaces, and the same percentage of goods were bought with direct use of recommendations. Comparing the vertical and horizontal recommending interface layouts, they performed equally in terms of fixation times. Still, from the perspective of purchase commitments, the vertical layout turned out to be almost twice as effective as the horizontal. In the better performing vertical layout, the most attractive were the first and second positions in the list. A slow flickering effect used in the second position influenced its attractiveness. On the other hand, the high visual attractiveness of the first position in the list, despite the lack of any visual highlight, may have resulted from the conception that the first one is always the best (similarly to search results). The worse effect of the horizontal layout may be connected with the so-called banner blindness. The event tracking experiment also showed the higher attractiveness of vertical zones compared to horizontal ones. To the best of our knowledge, this is the first study that verified this result using different methods and data sources.
Various classification models for predicting actions of adding product to cart directly from recommending interface or indirectly (influenced by displayed recommendation – gaze tracked) or any of those actions show overall very good precision. The accuracy measured as average over classes for all models with target variable add_to_cart_both varied in precision from 90.1 to 92.9%, with Neural Network performing best. The accuracy of predicting adding product to cart exclusively directly or indirectly from recommending interface ranged from 55.6 to 92.6% for precision metric, again with Neural Network performing best. Generally, good quality was provided with models based on Neural Network, Logistic Regression, and Random Forests.
The study showed that both approaches, gaze and event-tracking based, showed similar results in terms of recommendation interface presentation evaluation, and as such, they are similarly applicable for building automated mechanisms of recommending interface personalization and improving human–computer interaction. Both techniques confirmed the higher efficiency of vertical recommending interfaces compared to horizontal ones. Users spent relatively more time looking at vertical recommending interfaces, which resulted in more interactions related to adding products to the cart. The recommendation position inside the recommending interface proved to play an important role in a good performance for middle positions—second for vertical and third for horizontal layout. The visual intensity had a varied influence with the flickering effect attracting most users.
The study had some limitations, in particular, the subjects in the study were not randomly selected and were not a representative subset of the whole population. They, however, provided cognitive value in the context of the effectiveness of tracking techniques and machine learning methods for identifying user behavior indicative of purchase intent.
In future research, we intend to extend the study to a larger and statistically significant scale by partnering with a growing number of e-commerce stores and leveraging the usage of implicit activity tracking to receive the ultimate benefit of recommendation optimization. We will use the most promising modeling methods. In our experiments, we have not dealt with eye fatigue and eye–hand coordination, which can affect the results’ reliability [43,44,45,46], nor with the relation of gaze tracking to the cognitive realm, which is a significant part when deciding whether to purchase or not. Therefore, we plan to use more advanced eye tracking for gathering more data, considering the variable nature of gaze and detecting eye fatigue issues. We also intend to include the cognitive emotion information based on the fact that different types of eye motions can represent different emotions [47,48] and integrate a facial expression recognition (FER) module to include data on human sentiment while surfing the e-commerce sites. We will also perform both eye tracking and implicit events tracking at the same time to cross-compare their results and possibly come up with an optimized hybrid approach. Moreover, we want to develop our events-based solutions to recognize users’ activity, including time spent displaying particular AOIs in recommending interfaces on computers and mobile devices, focusing on visual aspects such as layout and intensity, in order to seek best-performing solutions for this multidimensional problem.

Author Contributions

Conceptualization, P.S.; methodology, P.S. and T.Z.; software, P.S. and T.Z.; supervision, P.S.; validation, P.S. and T.Z.; formal analysis, P.S. and T.Z.; investigation, P.S. and T.Z.; resources, P.S., T.Z., K.C., K.D., and K.S.-W.; data curation, P.S. and T.Z.; writing—original draft preparation, P.S. and T.Z.; writing—review and editing, P.S., K.C., K.D., K.K., and K.S.-W.; visualization, P.S. and T.Z.; project administration, P.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research project was supported by the National Science Centre of Poland, Grant No. 2017/27/B/HS4/01216, coordinated by Prof. Jaroslaw Jankowski (West Pomeranian University of Technology); and within the framework of the program of the Minister of Science and Higher Education under the name Regional Excellence Initiative in the years 2019–2022, Project No. 001/RID/2018/19, the amount of financing PLN 10,684,000.

Institutional Review Board Statement

The study was conducted according to the guidelines of the Declaration of Helsinki. Ethical review and approval were not required in accordance with the local legislation and institutional requirements.

Informed Consent Statement

All participants were informed about the purpose and the procedure of the study and provided their informed consent prior to the experiment. The experiment did not involve any risk or discomfort for the participants.

Data Availability Statement

Some data presented in the study are available from the corresponding author, upon request, data sharing being limited due to personal information content, and the code used during the study being proprietary.

Acknowledgments

We would like to thank Chair of Information Systems Engineering Jaroslaw Jankowski (West Pomeranian University of Technology) for grant coordination, and student Lukasz Dobrowolski (West Pomeranian University of Technology) for support, initial gaze data being used in his master’s thesis with the permission granted by the first author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kelly, D. Implicit Feedback: Using Behavior to Infer Relevance. In New Directions in Cognitive Information Retrieval; Spink, A., Cole, C., Eds.; The Information Retrieval Series; Springer: Dordrecht, The Netherlands, 2006; Volume 19, pp. 169–186. [Google Scholar]
  2. Sulikowski, P.; Zdziebko, T.; Turzyński, D.; Kańtoch, E. Human-website interaction monitoring in recommender systems. Procedia Comput. Sci. 2018, 126, 1587–1596. [Google Scholar] [CrossRef]
  3. Sulikowski, P.; Zdziebko, T.; Turzyński, D. Modeling online user product interest for recommender systems and ergonomics studies. Concurr. Comput. Pract. Exp. 2019, 31, e4301. [Google Scholar] [CrossRef]
  4. Wątróbski, J.; Jankowski, J.; Karczmarczyk, A.; Ziemba, P. Integration of Eye-Tracking Based Studies into e-Commerce Websites Evaluation Process with eQual and TOPSIS Methods. In Proceedings of the 10th SIGSAND/PLAIS EuroSymposium 2017, Gdańsk, Poland, 22 September 2017; Springer: Cham, Switzerland, 2017; Volume 300, pp. 56–80. [Google Scholar]
  5. Jankowski, J.; Ziemba, P.; Wątróbski, J.; Kazienko, P. Towards the Tradeoff Between Online Marketing Resources Exploitation and the User Experience with the Use of Eye Tracking. In Proceedings of the 8th Asian Conference, ACIIDS 2016, Da Nang, Vietnam, 14–16 March 2016; Springer: Berlin/Heidelberg, Germany, 2016; Volume 9621, pp. 330–343. [Google Scholar]
  6. Zhang, L.; Qin, T.; Teng, P. An Improved Collaborative Filtering Algorithm Based on User Interest. J. Softw. 2014, 9, 999–1006. [Google Scholar] [CrossRef]
  7. Geuens, S.; Coussement, K.; De Bock, K.W. A Framework for Configuring Collaborative Filtering-based Recommendations Derived from Purchase Data. Eur. J. Oper. Res. 2018, 265, 208–218. [Google Scholar] [CrossRef]
  8. Zhang, S.; Yao, L.; Sun, A.; Tay, Y. Deep Learning Based Recommender System. Acm Comput. Surv. 2019, 52, 1–38. [Google Scholar] [CrossRef] [Green Version]
  9. Shahriar, A.; Moon, M.; Mahmud, H.; Hasan, K. Online Product Recommendation System by Using Eye Gaze Data. In Proceedings of the International Conference on Computing Advancements (ICCA 2020), Dhaka, Bangladesh, 10–12 January 2020; Association for Computing Machinery: New York, NY, USA, 2020. Article 61. pp. 1–7. [Google Scholar]
  10. Chen, L.; Pu, P. Eye-tracking study of user behavior in recommender interfaces. In International Conference on User Modeling, Adaptation, and Personalization; Springer: Berlin/Heidelberg, Germany, 2010; pp. 375–380. [Google Scholar]
  11. Bobadilla, J.; Ortega, F.; Hernando, A.; Gutierrez, A. Recommender systems survey. Knowl. Based Syst. 2013, 46, 109–132. [Google Scholar] [CrossRef]
  12. Lu, J.; Wu, D.; Mao, M.; Wang, W.; Zhang, G. Recommender system application developments: A survey. Decis. Support Syst. 2015, 74, 12–32. [Google Scholar] [CrossRef]
  13. Cazzato, D.; Leo, M.; Distante, C.; Voos, H. When I Look into Your Eyes: A Survey on Computer Vision Contributions for Human Gaze Estimation and Tracking. Sensors 2020, 20, 3739. [Google Scholar] [CrossRef]
  14. Pu, P.; Chen, L.; Kumar, P. Evaluating Product Search and Recommender Systems for E-Commerce Environments. Electron. Commer. Res. 2008, 8, 1–27. [Google Scholar] [CrossRef] [Green Version]
  15. Bortko, K.; Bartków, P.; Jankowski, J.; Kuras, D.; Sulikowski, P. Multi-criteria Evaluation of Recommending Interfaces towards Habituation Reduction and Limited Negative Impact on User Experience. Procedia Comput. Sci. 2019, 159, 2240–2248. [Google Scholar] [CrossRef]
  16. Portnoy, F.; Marchionini, G. Modeling the effect of habituation on banner blindness as a function of repetition and search type: Gap analysis for future work. In CHI’10 Extended Abstracts on Human Factors in Computing Systems; ACM: New York, NY, USA, 2010; pp. 4297–4302. [Google Scholar]
  17. Ha, L.; McCann, K. An integrated model of advertising clutter in offline and online media. Int. J. Advert. 2008, 27, 569–592. [Google Scholar] [CrossRef]
  18. Jin, X.; Zhou, Y.; Mobasher, B. Task-oriented web user modeling for recommendation. In International Conference on User Modeling; Springer: Berlin/Heidelberg, Germany, 2005; pp. 109–118. [Google Scholar]
  19. Jankowski, J.; Hamari, J.; Watróbski, J. A gradual approach for maximising user conversion without compromising experience with high visual intensity website elements. Internet Res. 2019, 29, 194–217. [Google Scholar] [CrossRef]
  20. Rohrer, C.; Boyd, J. The rise of intrusive online advertising and the response of user experience research at yahoo! In CHI’04 Extended Abstracts on Human Factors in Computing Systems; ACM: New York, NY, USA, 2004; pp. 1085–1086. [Google Scholar]
  21. Jankowski, J. Modeling the Structure of Recommending Interfaces with Adjustable Influence on Users, Intelligent Information and Database Systems. Lect. Notes Comput. Sci. 2013, 7803, 429–438. [Google Scholar]
  22. Sałabun, W.; Karczmarczyk, A.; Mejsner, P. Experimental Study of Color Contrast Influence in Internet Advertisements with Eye Tracker Usage. In Neuroeconomic and Behavioral Aspects of Decision Making. Springer Proceedings in Business and Economics; Nermend, K., Łatuszyńska, M., Eds.; Springer: Cham, Switzerland, 2017. [Google Scholar]
  23. Gomes, L.; Almeida, C.; Vale, Z. Recommendation of Workplaces in a Coworking Building: A Cyber-Physical Approach Supported by a Context-Aware Multi-Agent System. Sensors 2020, 20, 3597. [Google Scholar] [CrossRef] [PubMed]
  24. Sulikowski, P. Evaluation of Varying Visual Intensity and Position of a Recommendation in a Recommending Interface Towards Reducing Habituation and Improving Sales. In Advances in E-Business Engineering for Ubiquitous Computing, Proceedings of the 16th International Conference on E-Business Engineering, ICEBE 2019, Shanghai, China, 11–13 October 2019; Lecture Notes on Data Engineering and Communications Technologies; Chao, K.M., Jiang, L., Hussain, O., Ma, S.P., Fei, X., Eds.; Springer Nature: Cham, Switzerland, 2020; Volume 41, pp. 208–218. [Google Scholar]
  25. Jannach, D.; Lerche, L.; Zanker, M. Recommending based on implicit feedback. In Social Information Access; Springer: Cham, Swizerland, 2018; pp. 510–569. [Google Scholar]
  26. Lerche, L. Using Implicit Feedback for Recommender Systems: Characteristics, Applications, and Challenges. Ph.D. Thesis, Technische Universität Dortmund, Dortmund, Germany, December 2016. [Google Scholar]
  27. Zhao, Q.; Harper, F.M.; Adomavicius, G.; Konstan, J.A. Explicit or implicit feedback? Engagement or satisfaction? A field experiment on machine-learning-based recommender systems. In Proceedings of the 33rd Annual ACM Symposium on Applied Computing, Pau, France, 9–13 April 2018; ACM: New York, NY, USA, 2018; pp. 1331–1340. [Google Scholar]
  28. Zhou, M.; Ding, Z.; Tang, J.; Yin, D. Micro behaviors: A new perspective in e-commerce recommender systems. In Proceedings of the Eleventh ACM International Conference on Web Search and Data Mining, Los Angeles, CA, USA, 5–9 February 2018; pp. 727–735. [Google Scholar]
  29. Tian, G.; Wang, J.; He, K.; Sun, C.; Tian, Y. Integrating implicit feedbacks for time-aware web service recommendations. Inf. Syst. Front. 2017, 19, 75–89. [Google Scholar] [CrossRef]
  30. Liversedge, S.P.; Drieghe, D.; Li, X.; Yan, G.; Bai, X.; Hyönä, J. Universality in eye movements and reading: A trilingual investigation. Cognition 2016, 147, 1–20. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  31. Buscher, G.; Dengel, A.; Biedert, R.; Van Elst, L. Attentive documents: Eye tracking as implicit feedback for information retrieval and beyond. ACM Trans. Interact. Intell. Syst. 2012, 1, 1–30. [Google Scholar] [CrossRef]
  32. Steichen, B.; Carenini, G.; Conati, C. Adaptive Information Visualization-Predicting user characteristics and task context from eye gaze. In Proceedings of the International Conference on User Modeling, UMAP Workshops, Montreal, QC, Canada, 16–20 July 2012; Volume 872. [Google Scholar]
  33. Loyola, P.; Brunetti, E.; Martinez, G.; Velásquez, J.D.; Maldonado, P. Leveraging Neurodata to Support Web User Behavior Analysis. In Wisdom Web of Things; Web Information Systems Engineering and Internet Technologies Book Series; Zhong, N., Ma, J., Liu, J., Huang, R., Tao, X., Eds.; Springer: Cham, Switzerland, 2016; pp. 181–207. [Google Scholar]
  34. Sharafi, Z.; Shaffer, T.; Sharif, B.; Guéhéneuc, Y.-G. Eye-Tracking Metrics in Software Engineering. In Proceedings of the 2015 Asia-Pacific Software Engineering Conference, New Delhi, India, 1–4 December 2015; IEEE Press: Piscataway, NJ, USA, 2015; pp. 96–103. [Google Scholar]
  35. Lee, J.; Ahn, J.-H.; Park, B. The effect of repetition in internet banner ads and the moderating role of animation. Comput. Hum. Behav. 2015, 46, 202–209. [Google Scholar] [CrossRef]
  36. Akuma, S.; Iqbal, R.; Jayne, C.; Doctor, F. Comparative analysis of relevance feedback methods based on two user studies. Comput. Hum. Behav. 2016, 60, 138–146. [Google Scholar] [CrossRef] [Green Version]
  37. Boi, P.; Fenu, G.; Spano, L.; Vargiu, V. Reconstructing User’s Attention on the Web through Mouse Movements and Perception-Based Content Identification. Acm Trans. Appl. Percept. 2016, 13, 1–21. [Google Scholar] [CrossRef] [Green Version]
  38. Zhao, Q.; Chang, S.; Harper, M.; Konstan, J.A. Gaze Prediction for Recommender Systems. In Proceedings of the 10th ACM Conference on Recommender Systems (RecSys ’16); Association for Computing Machinery: New York, NY, USA, 2016; pp. 131–138. [Google Scholar]
  39. Sulikowski, P.; Zdziebko, T. Horizontal vs. Vertical Recommendation Zones Evaluation Using Behavior Tracking. Appl. Sci. 2021, 11, 56. [Google Scholar] [CrossRef]
  40. Sulikowski, P.; Zdziebko, T. Deep Learning-Enhanced Framework for Performance Evaluation of a Recommending Interface with Varied Recommendation Position and Intensity Based on Eye-Tracking Equipment Data Processing. Electronics 2020, 9, 266. [Google Scholar] [CrossRef] [Green Version]
  41. Kantardzic, M. Data Mining: Concepts, Models, Methods, and Algorithms; John Wiley & Sons Publishing: Hoboken, NJ, USA, 2003. [Google Scholar]
  42. Al-Nabiand, D.L.A.; Ahmed, S.S. Survey on classification algorithms for data mining: (comparison and evaluation). Comput. Eng. Intell. Syst. 2013, 4, 18–24. [Google Scholar]
  43. Vasiljevas, M.; Gedminas, T.; Ševčenko, A.; Jančiukas, M.; Blažauskas, T.; Damaševičius, R. Modelling eye fatigue in gaze spelling task. In Proceedings of the 2016 IEEE 12th International Conference on Intelligent Computer Communication and Processing (ICCP), Cluj-Napoca, Romania, 8–10 September 2016; pp. 95–102. [Google Scholar]
  44. Kim, T.; Lee, E.C. Experimental Verification of Objective Visual Fatigue Measurement Based on Accurate Pupil Detection of Infrared Eye Image and Multi-Feature Analysis. Sensors 2020, 20, 4814. [Google Scholar] [CrossRef]
  45. Katona, J.; Kovari, A.; Heldal, I.; Costescu, C.; Rosan, A.; Demeter, R.; Thill, S.; Stefanut, T. Using Eye- Tracking to Examine Query Syntax and Method Syntax Comprehension in LINQ. In Proceedings of the 2020 11th IEEE International Conference on Cognitive Infocommunications (CogInfoCom), Stockholm, Sweden, 23–25 September 2020. [Google Scholar]
  46. Ujbanyi, T.; Kovari, A.; Sziládi, G.; Katona, J. Examination of the eye-hand coordination related to computer mouse movement. Infocommun. J. 2020, 12, 26–31. [Google Scholar] [CrossRef]
  47. Maskeliunas, R.; Raudonis, V. Are you ashamed? Can a gaze tracker tell? PeerJ Comput. Sci. 2016, 2, e75. [Google Scholar] [CrossRef]
  48. Costescu, C.; Rosan, A.; Brigitta, N.; Hathazi, A.; Kovari, A.; Katona, J.; Heldal, I.; Helgesen, C.; Thill, S.; Demeter, R. Emotion recognition in typical and atypical development–a technology-based paradigm. In Proceedings of the 10th IEEE International Conference on Infocommunications (CogInfoCom), Naples, Italy, 23–25 October 2019; pp. 349–352. [Google Scholar]
Figure 1. Conceptual system architecture.
Figure 1. Conceptual system architecture.
Sensors 21 01381 g001
Figure 2. Research experiment architecture.
Figure 2. Research experiment architecture.
Sensors 21 01381 g002
Figure 3. Activity diagram depicting implicit data processing with the behavior monitoring tool.
Figure 3. Activity diagram depicting implicit data processing with the behavior monitoring tool.
Sensors 21 01381 g003
Figure 4. Vertical layout of the recommending interface.
Figure 4. Vertical layout of the recommending interface.
Sensors 21 01381 g004
Figure 5. Horizontal layout of the recommending interface.
Figure 5. Horizontal layout of the recommending interface.
Sensors 21 01381 g005
Figure 6. Example of visual intensity VI3 (red background) in the vertical recommending interface.
Figure 6. Example of visual intensity VI3 (red background) in the vertical recommending interface.
Sensors 21 01381 g006
Figure 7. Classification accuracy metrics for all target variables received for different models.
Figure 7. Classification accuracy metrics for all target variables received for different models.
Sensors 21 01381 g007
Table 1. Main recommendation techniques—advantages and disadvantages.
Table 1. Main recommendation techniques—advantages and disadvantages.
Recommendation TechniqueAdvantagesDisadvantages
Collaborative filteringSerendipity—user can discover new items/products
Intuitive and rationale recommendations
Easy scalability
No need for content data about items/products
Cold start problem for both new users and new items
Low recommendation quality for sparse data
Model training can be expensive
Content-basedUser independence—the model needs data only about the current user
Explainability of the results
Ability to generate a recommendation for the user with a unique taste
Lack of recommendation novelty
Over-specialization
Cold start problem for both user and item
Requires description of products/items
Table 2. Features generated based on eye-tracking device.
Table 2. Features generated based on eye-tracking device.
FeatureFeature Description
rc_layoutRC layout (horizontal/vertical)
rc_location_intensityintensity level of a recommendation location (1–3)
rc_locationRCi position (1–4)
fixation_time_locationoverall fixation time for RCi position
fixation_time_layoutoverall fixation time for RC layout
fixation_time_categoryoverall time spent by the participant on a product category page
share_time_location_categoryfraction of time when fixation was registered inside RCi position with respect to the overall time spent on a category page
share_time_layout_categoryfraction of time when fixation was registered inside RC layout to the overall time registered on a product category page
share_time_location_layoutproportion of time when fixation was observed inside RCi position to the overall time registered on RC layout
add_to_cart_directaction of adding the product to cart directly from RC
add_to_cart_indirectaction of adding the product to cart action from EC while product was looked at in RC
add_to_cart_bothadding the product to cart action from RC or EC while product was being observed in RC
Table 3. Values of classification model hyper-parameters.
Table 3. Values of classification model hyper-parameters.
ModelHyper-ParameterHyper-Parameter Value
Neural Networkactivation functionRectified Linear (ReLU)
Neural Networkmaxium number of neurons in hidden layer150
Decision treesstop extending a tree when majority reaches95%
Decision treesdo not split subsets smaller than5
Logistic Regressionregularization functionRidge (L2)
Logistic RegressionstrengthC = 0.18
Random Forestsmaxium number of trees15
Random Forestsdepth limit of individual tree3
Random Forestssplitting stop subset limit5
SVMkernelRBF with Auto Gamma
SVMcost1
SVMregression loss epsilon0.1
Table 4. Features generated based on data gathered via the e-commerce customer preference monitor (ECPM) monitoring tool.
Table 4. Features generated based on data gathered via the e-commerce customer preference monitor (ECPM) monitoring tool.
FeatureFeature Description
rc_layoutRC layout (horizontal/vertical)
recommend_lengthnumber of characters within all recommended products sections
document_lengthnumber of characters within all texts on the page
page_timetime between page load and page unload
tab_activ_timetime while tab containing particular page is active
user_activ_timetime while user is actively interacting with page (generating keyboard or mouse events)
rc_prod_recommend_timetime while mouse pointer was positioned over recommended products
rel_recommend_time_document_lengthrelative time while mouse pointer was positioned over recommending interface in relation to the number of characters within all texts on the page
rel_recommend_time_page_timerelative time, while mouse pointer was positioned over recommending interface, in relation to time between page load and page unload
rel_recommend_time_tab_activrelative time while mouse pointer was positioned over recommending interface in relation to time while tab containing particular page is active
rel_recommend_time_user_activrelative time while mouse pointer was positioned over recommending interface in relation to time while user is actively interacting with page (generating keyboard or mouse events)
Table 5. Average fixation time (s) for each recommendation location.
Table 5. Average fixation time (s) for each recommendation location.
Recommendation LocationTime (s)
Vertical RCHorizontal RC
RC12.41.3
RC23.12.1
RC32.13.9
RC40.60.8
Total8.28.1
Table 6. Recommendation-driven purchase and visual intensity for each recommendation location RCi and product category PCj.
Table 6. Recommendation-driven purchase and visual intensity for each recommendation location RCi and product category PCj.
Recommendation LocationVertical RCHorizontal RC
PC1PC2PC3PC4PC5PC6
RC1 direct0.019 (VI1)0.019 (VI1)0 (VI1)0.038 (VI1)0 (VI1)0.077 (VI1)
RC1 indirect0.019 (VI1)0 (VI1)0 (VI1)0.019 (VI1)0 (VI1)0.058 (VI1)
RC2 direct0.038 (VI1)0.077 (VI2)0.096 (VI1)0 (VI1)0.058 (VI2)0 (VI1)
RC2 indirect0.038 (VI1)0.058 (VI2)0.077 (VI1)0.019 (VI1)0.038 (VI2)0.019 (VI1)
RC3 direct0.019 (VI1)0.077 (VI1)0.019 (VI3)0 (VI1)0.019 (VI1)0.038 (VI3)
RC3 indirect0.058 (VI1)0.096 (VI1)0.019 (VI3)0 (VI1)0.038 (VI1)0.058 (VI3)
RC4 direct0.038 (VI1)0 (VI1)0.038 (VI1)0 (VI1)0.038 (VI1)0 (VI1)
RC4 indirect0.019 (VI1)0.019 (VI1)0.038 (VI1)0 (VI1)0.058 (VI1)0.019 (VI1)
Total direct0.1150.1730.1540.0380.1150.115
Total indirect0.1350.1730.1350.0380.1350.154
Table 7. Models accuracy metrics for add_to_cart_both as target parameter.
Table 7. Models accuracy metrics for add_to_cart_both as target parameter.
MetricNeural NetworkRandom ForestsLogistic RegressionSVMBinary Tree
AUC0.8830.9130.9380.9050.681
CA0.9320.9270.9210.9210.910
Table 8. Models accuracy metrics average over classes with add_to_cart_both as target parameter.
Table 8. Models accuracy metrics average over classes with add_to_cart_both as target parameter.
ModelF1PrecisionRecallSpecificity
Neural Network0.9290.9290.9320.694
Random Forests0.9190.9230.9270.582
Logistic Regression0.9110.9160.9210.544
SVM0.9110.9160.9210.544
Binary Tree0.9010.9010.9100.543
Table 9. Models accuracy metrics for add to cart action (class 1) with add_to_cart_both as target parameter.
Table 9. Models accuracy metrics for add to cart action (class 1) with add_to_cart_both as target parameter.
ModelF1PrecisionRecallSpecificity
Random Forests0.6490.8570.5220.987
Logistic Regression0.6110.8460.4780.987
SVM0.6110.8460.4780.987
Neural Network0.7140.7890.6520.974
Binary Tree0.5790.7330.4780.974
Table 10. Models accuracy metrics for not adding a product from recommending interface (class 0) with add_to_cart_both as target parameter.
Table 10. Models accuracy metrics for not adding a product from recommending interface (class 0) with add_to_cart_both as target parameter.
ModelF1PrecisionRecallSpecificity
Neural Network0.9620.9490.9740.652
Random Forests0.9590.9330.9870.522
Logistic Regression0.9560.9270.9870.478
SVM0.9560.9270.9870.478
Binary Tree0.9490.9260.9740.478
Table 11. Models accuracy metrics for add_to_cart_direct as target parameter.
Table 11. Models accuracy metrics for add_to_cart_direct as target parameter.
MetricBinary TreeSVMRandom ForestsNeural NetworkLogistic Regression
AUC0.8970.9610.9500.9600.971
CA0.9660.9440.9550.9550.944
Table 12. Models accuracy metrics average over classes with add_to_cart_direct as target parameter.
Table 12. Models accuracy metrics average over classes with add_to_cart_direct as target parameter.
ModelF1PrecisionRecallSpecificity
Binary Tree0.9670.9690.9660.828
SVM0.9410.9390.9440.487
Random Forests0.9550.9550.9550.657
Neural Network0.9530.9510.9550.573
Logistic Regression0.9410.9390.9440.487
Table 13. Models accuracy metrics for add to cart action (class 1) with add_to_cart_direct as target parameter.
Table 13. Models accuracy metrics for add to cart action (class 1) with add_to_cart_direct as target parameter.
ModelF1PrecisionRecallSpecificity
Binary Tree0.7500.6920.8180.976
SVM0.5000.5560.4550.976
Random Forests0.6360.6360.6360.976
Neural Network0.6000.6670.5450.982
Logistic Regression0.5000.5560.4550.976
Table 14. Models accuracy metrics for not adding product from recommending interface (class 0) with add_to_cart_direct as target parameter.
Table 14. Models accuracy metrics for not adding product from recommending interface (class 0) with add_to_cart_direct as target parameter.
ModelF1PrecisionRecallSpecificity
Binary Tree0.9820.9880.9760.818
SVM0.9700.9640.9760.455
Random Forests0.9760.9760.9760.636
Neural Network0.9760.9700.9820.545
Logistic Regression0.9700.9640.9760.455
Table 15. Models accuracy metrics for add_to_cart_indirect as target parameter.
Table 15. Models accuracy metrics for add_to_cart_indirect as target parameter.
ModelNeural NetworkBinary TreeLogistic RegressionRandom ForestsSVM
AUC0.8580.5480.7290.9050.750
CA0.9040.8910.9270.9210.921
Table 16. Models accuracy metrics average over classes with add_to_cart_indirect as target parameter.
Table 16. Models accuracy metrics average over classes with add_to_cart_indirect as target parameter.
ModelF1PrecisionRecallSpecificity
Neural Network0.9020.9000.9040.562
Binary Tree0.8840.8860.8810.526
Logistic Regression0.8970.8690.9270.508
Random Forests0.8940.8680.9210.496
SVM0.8940.8680.9210.446
Table 17. Models accuracy metrics for add to cart action (class 1) with add_to_cart_indirect as target parameter.
Table 17. Models accuracy metrics for add to cart action (class 1) with add_to_cart_indirect as target parameter.
ModelF1PrecisionRecallSpecificity
Neural Network0.9490.9260.9520.616
Binary Tree0.9360.9190.9330.562
Logistic Regression0.9620.9120.9940.562
Random Forests0.9590.8910.9880.568
SVM0.9590.8910.9880.538
Table 18. Models accuracy metrics for not adding product from recommending interface (class 0) with add_to_cart_indirect as target parameter.
Table 18. Models accuracy metrics for not adding product from recommending interface (class 0) with add_to_cart_indirect as target parameter.
ModelF1PrecisionRecallSpecificity
Neural Network0.7950.6480.7680.424
Binary Tree0.7860.6460.7510.418
Logistic Regression0.6980.5620.7170.403
Random Forests0.6570.5200.6260.387
SVM0.6570.5200.6260.340
Table 19. Recommending interface quality parameters based on the ECPM study.
Table 19. Recommending interface quality parameters based on the ECPM study.
Store/rc_layoutprod_recommend_time
[ms]
rel_recommend_time_
recommend_length [ms/char]
rel_recommend_time_
document_length
[ms/char]
rel_recommend_time_
page_time
rel_recommend_time_
tab_activ
rel_recommend_time_
user_activ
Agito.pl
vertical
26691.5840.0750.0590.0670.101
Agito.pl
horizontal
19201.3040.0540.0430.0480.073
Electro.pl
vertical
40292.0480.1570.0890.0960.174
Electro.pl
horizontal
18580.9440.0720.0410.0440.080
Merlin.pl
vertical
19770.3510.1080.0490.0840.156
Merlin.pl
horizontal
13120.2850.0720.0320.0560.104
Morele.net
vertical
17770.9050.1040.0540.0520.079
Morele.net
horizontal
19530.7520.1140.0590.0580.087
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Sulikowski, P.; Zdziebko, T.; Coussement, K.; Dyczkowski, K.; Kluza, K.; Sachpazidu-Wójcicka, K. Gaze and Event Tracking for Evaluation of Recommendation-Driven Purchase. Sensors 2021, 21, 1381. https://doi.org/10.3390/s21041381

AMA Style

Sulikowski P, Zdziebko T, Coussement K, Dyczkowski K, Kluza K, Sachpazidu-Wójcicka K. Gaze and Event Tracking for Evaluation of Recommendation-Driven Purchase. Sensors. 2021; 21(4):1381. https://doi.org/10.3390/s21041381

Chicago/Turabian Style

Sulikowski, Piotr, Tomasz Zdziebko, Kristof Coussement, Krzysztof Dyczkowski, Krzysztof Kluza, and Karina Sachpazidu-Wójcicka. 2021. "Gaze and Event Tracking for Evaluation of Recommendation-Driven Purchase" Sensors 21, no. 4: 1381. https://doi.org/10.3390/s21041381

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop