Next Article in Journal
Active Learning Strategies in Computer Science Education: A Systematic Review
Previous Article in Journal
Metaverse & Human Digital Twin: Digital Identity, Biometrics, and Privacy in the Future Virtual Worlds
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

OnMapGaze and GraphGazeD: A Gaze Dataset and a Graph-Based Metric for Modeling Visual Perception Differences in Cartographic Backgrounds Used in Online Map Services

by
Dimitrios Liaskos
and
Vassilios Krassanakis
*
Department of Surveying and Geoinformatics Engineering, School of Engineering, University of West Attica, 28 Agiou Spyridonos Str., 12243 Egaleo, Greece
*
Author to whom correspondence should be addressed.
Multimodal Technol. Interact. 2024, 8(6), 49; https://doi.org/10.3390/mti8060049
Submission received: 16 May 2024 / Revised: 7 June 2024 / Accepted: 9 June 2024 / Published: 13 June 2024

Abstract

:
In the present study, a new eye-tracking dataset (OnMapGaze) and a graph-based metric (GraphGazeD) for modeling visual perception differences are introduced. The dataset includes both experimental and analyzed gaze data collected during the observation of different cartographic backgrounds used in five online map services, including Google Maps, Wikimedia, Bing Maps, ESRI, and OSM, at three different zoom levels (12z, 14z, and 16z). The computation of the new metric is based on the utilization of aggregated gaze behavior data. Our dataset aims to serve as an objective ground truth for feeding artificial intelligence (AI) algorithms and developing computational models for predicting visual behavior during map reading. Both the OnMapGaze dataset and the source code for computing the GraphGazeD metric are freely distributed to the scientific community.

1. Introduction

In our constantly developing world, the map has evolved from a static view of the world to an interactive, mobile, dynamic, and collaborative interface. Such an interface now serves as a bridge between individuals and the ever-changing environment. Nowadays, a map can be considered as a platform that allows users to manipulate spatial information. As reported by Roth et al. [1], web maps are adaptive to the use and user’s context (e.g., [2,3]), interactive to user’s requests (e.g., [4,5]), mobile (e.g., [6,7]), multiscale (e.g., [8,9]), and updated in real time (e.g., [10,11]). Ultimately, the map has become a powerful tool for enhancing communication and decision-making that values a dynamic approach to spatial information. The widespread presence of digital technologies has transformed the way the general audience accesses maps and spatial information [12]. Maps are ingrained in our daily lives, and their role is crucial in different tasks, ranging from navigation to trip planning purposes. However, their effectiveness depends on how well users can identify the information they need, such as points of interest and navigation cues [13]. Modern cartographic products are mainly distributed through online map services and are accessed by millions of web users every day. It is therefore important to understand how users observe and understand maps to improve their design and usability [9].
Eye-tracking and eye movement analysis a valuable research technique for assessing different aspects related to the interaction between a map and its user [14,15]. Although map designers strive to create informative maps, there is a constant need to examine how map users perceive and prioritize the objects depicted on a map. The examination of visual behavior during map reading could be based on both the analysis and visualization of experimental gaze data [16]. Over the last decades, a considerable number of experimental studies have been conducted to investigate both the effectiveness and efficiency of cartographic products such as interactive maps [17,18], animated maps [19,20,21,22,23,24], and web maps [13,25,26,27]. The study of map users’ behaviors requires developing simple or more sophisticated approaches that reveal crucial information related to user perception and cognition. In particular, specific metrics that could model the performed visual strategies (e.g., [28]), as well as gaze activity patterns during the execution of map tasks (e.g., [29]), are of high importance for the development of robust methodological frameworks for cartographic design evaluation processes. Thus, the need to develop cartographic-oriented user response measures is clear and could contribute significantly to a better understanding of map usability issues. This need becomes even more apparent when considering the large amount of experimental data collected with modern eye-tracking equipment (i.e., modern devices can record gaze data at frequencies up to 2000 Hz).
Today, artificial intelligence (AI) has become an active part of state-of-the-art technology, science, and society. GeoAI is the use of AI and machine learning in geography [30] to develop approaches related to spatiotemporal problems [31]. AI could substantially support the production of personalized maps [32,33]. User data collected utilizing eye tracking or other user response methods (e.g., mouse tracking [34]), can be used as input for on-the-fly map design processes. In this direction, the existence of rich and open-access eye movement datasets is valuable, especially considering that they can serve as benchmarks [35]. Undoubtedly, freely distributed benchmark datasets facilitate the sharing of open data, encourage collaboration, empower the scientific community to tackle challenges, and support research related to AI applications (e.g., visual saliency prediction). Currently, there are several efforts to distribute open datasets to the scientific community.
The aim of the present study is twofold. Firstly, we present a new eye-tracking dataset, called OnMapGaze, which includes both experimental and analyzed gaze data collected during the observation of different cartographic backgrounds utilized in modern online map services. Our dataset aims to serve as an objective ground truth for feeding AI algorithms and developing computational models for predicting visual behavior during map reading. Secondly, we introduce a new graph-based metric, called GraphGazeD, suitable for modeling existing visual perception differences based on aggregated gaze behavior data. Both the OnMapGaze dataset and the source code for computing the GraphGazeD metric are freely distributed to the scientific community (see Supplementary Materials).
Section 2 describes the related work that involves the basic quantitative eye-tracking metrics used in eye movement analyses, the processes for aggregated gaze behavior modeling utilizing statistical grayscale heatmaps, and the use of the aforementioned products in cartographic research as well as the existing gaze datasets. In Section 3, the approaches and methods for the design, execution, and analysis of an eye-tracking experiment are presented, including the introduced metric. Section 4 describes the results of the present study, which mainly involve the components of the new dataset. Section 5 discusses the results of the present study, outlines the conclusions, and includes suggestions for future work.

2. Related Work

2.1. Basic Quantitative Eye-Tracking Metrics

Eye tracking is a valuable experimental technique that involves the recording of gaze coordinates spatiotemporally during the observation of visual scenes [36]. Using this technique, researchers can study eye movement events, changes in pupil size (pupillometry), and overall gaze behavior [37]. Fixations and saccades constitute the fundamental metrics utilized to analyze eye movements [38]. Fixations correspond to periods of time during which the eyes remain relatively still and focused on a specific location or object in the observed visual scene. Saccades occur between fixations and involve rapid eye movements from one point of interest to another.
Several derived metrics can be computed from the fundamental metrics, including the scan path corresponding to fixation–saccade–fixation sequences [39]. Scan paths reveal the patterns in which different areas or objects are visited. Additionally, pupil size (i.e., the measurement of pupil diameter) and the blink rate are metrics that provide insight into visual attention and user behavior during the execution of specific tasks [40]. Pupil size measurements are therefore connected to cognitive processes and cognitive load. The blink rate is the frequency at which individuals blink their eyes. It indicates attention levels and changes in blink rate reflect shifts in cognitive states [41].

2.2. Aggregated Gaze Behavior Modeling Using Statistical Grayscale Heatmaps

Eye-tracking technology has advanced over the years, facilitating the collection of eye movement data and increasing the demand for enhanced visual analysis methods [42]. Several techniques have been developed to visualize eye-tracking data, including heatmaps and gaze plots [43]. Understanding the variations in eye movements among the data collected from different participants is a key goal in eye-tracking studies [17,44]. Hence, the visualization of aggregated gaze behavior data is important in eye-tracking studies [16]. Such visualizations highlight the allocation of visual attention during the observation of visual stimuli, thereby aiding in qualitative examination and providing a comprehensive overview of scanning behavior. Among the performed visualization methods, heatmaps provide a clearer representation of eye movement data spatial distribution [45]. This method is applicable to both static and dynamic stimuli and can be implemented either post-experimentally or in real-time applications (e.g., [46]). Statistical grayscale heatmaps indicate the distribution of visual attention during the observation of a visual stimulus, and they can be generated using either raw gaze data or the centers of fixation points [16]. Additionally, the aforementioned products could be used to model the overall spatial distribution of visual attention across multiple observers. The creation of statistical grayscale heatmaps is based on the application of a Gaussian filter that takes into account the visual range of foveal vision. The filtering process produces a grayscale image where pixel values are normalized within a predefined range, reflecting the likelihood of gaze activity occurring within each pixel area. Therefore, they can effectively represent, both qualitatively and quantitatively, the salient locations of the visual stimuli. Moreover, statistical indices can be computed to describe the visual strategies of participants during free-viewing conditions or during the execution of a specific task (see e.g., [29]). Statistical grayscale heatmaps are one of the main products during the generation of an eye-tracking dataset. Such datasets could serve as the main ground truth for the production of saliency models (see e.g., [47,48,49,50,51]). Additionally, it is important to mention that grayscale statistical heatmaps could also be computed for the analysis of mouse tracking behavioral experiments since the nature of the corresponding data is similar (see e.g., [52]). Moreover, Krassanakis [16] presented a new method for aggregated gaze data visualization, which is based on the combined use of grayscale statistical heatmaps and contiguous irregular cartograms.

2.3. Statistical Grayscale Heatmaps for Cartographic Purposes

As mentioned above, grayscale statistical heatmaps could represent the cumulative spatial distribution of visual attention across different visual stimuli, as observed by multiple individuals. In cartography, however, even if different visual stimuli depict the same geographic area, their cartographic design (i.e., symbolization) may differ significantly. For example, cartographic backgrounds from different online map services are characterized by differences in their cartographic symbols [53]. Therefore, the result of the comparison of grayscale statistical heatmaps produced by such visual stimuli could reveal existing differences in the allocation of visual attention [54]. Comparing grayscale statistical heatmaps is a process that enhances comprehension and validation of the saliency of objects within the analyzed visual stimuli. While some grayscale statistical heatmaps may show similar gaze patterns, suggesting uniformity in visualization, closer examination through comparison reveals disparities. The information extracted from this type of product contributes to the description of the differences among cartographic products, enriching the knowledge to create more effective and efficient maps. In addition, statistical grayscale heatmaps can be used to examine gaze activity over the entire area of a map or over different areas of interest (AOIs) within a map. For example, in a recent study by Cybulski and Krassanakis [29], specific statistical indices were developed in order to examine the overall allocation of gaze activity during the visual search of labels on cartographic backgrounds.

2.4. Eye-Tracking Dataset Distribution in Cartography and Related Fields

As mentioned in Section 2.2, eye-tracking datasets are commonly used to predict salient locations on visual stimuli. However, it should be noted that eye-tracking datasets may differ in terms of the number of participants involved, the type and total number of visual stimuli used, the tasks that the participants were asked to perform, and the duration of each visual stimulus. Additionally, the age range and/or the level of expertise of the participants may vary. Furthermore, the existing eye-tracking datasets can be generated using different types of experimental equipment.
A number of eye-tracking datasets have been designed and distributed in the past years. The MIT/Tuebingen Saliency Benchmark [55] constitutes an extensive collection that provides eye-tracking data and saliency maps for different types of (audio)visual stimuli. The majority of models have been designed specifically to predict visual attention when viewing natural images. Nonetheless, a map, being an artificial and abstract depiction of the real world, poses a challenge for typical saliency models in predicting user behavior when interacting with maps [56]. Gaze datasets produced during the observation of cartographic products (and/or related products) could significantly aid in modeling visual attention and broaden the scope of cartographic research to explore emerging topics. EyeTrackUAV [57] and EyeTrackUAV2 [58] were created to study participants’ visual attention while observing unmanned aerial vehicle (UAV) videos under free-viewing and task-based (surveillance) conditions. Tzelepis et al. [59] provided an eye-tracking dataset collected during the observation of hill-shading images. GeoEye [51] contains eye movement data during the observation, under free-viewing conditions, of geospatial images, such as thematic maps, remote sensing, and street view images. CartoGaze [13] consists of eye-tracking data acquired during the execution of a controlled memorability experiment on 2D web maps. As Keskin et al. [35] also point out, sharing open benchmark datasets as well as developing specific guidelines are of major importance in understanding map usability issues.

3. Materials and Methods

3.1. Experimental Design

3.1.1. Visual Stimuli

The experimental visual stimuli consisted of cartographic backgrounds that were excerpted from web maps provided by five different online map services (Google Maps, Wikimedia, Bing Maps, ESRI, OSM) utilizing the open-source Geographic Information Software QGIS (v.3.22.4).and the QuickMapServices plugin (v.0.19.32). The visual stimuli were adjusted to the native resolution of the utilized display screen (1920 × 1080 px). The visual stimuli depict fifteen different cities in Greece at three different zoom levels (12z, 14z, and 16z). Hence, five different cities were selected for each zoom level, where cities with higher populations correspond to higher zoom levels. In total, 75 different visual stimuli were generated. All web map services use the WGS84 (EPSG: 4326) coordinate reference system and the Web Mercator projection, which was introduced by Google Maps and constitutes the new version of the Mercator projection [53]. Figure 1 depicts some indicative examples of the experimental visual stimuli.
The display order of the visual stimuli was randomized to reduce order bias, while each visual stimulus was displayed for three seconds. Between the experimental stimuli, a blank screen of white color (RGB 255, 255, 255) was displayed for one second to initialize the observer’s visual exploration from the starting point, the center of the image. The display time was determined after several trials, taking into account the total duration of the experimental process, since longer experimental durations can lead to physical fatigue in the participants.

3.1.2. Participants

In total, thirty participants, with normal or corrected to normal vision, participated in the experiment (53% female). The right eye was dominant for 73% of the participants. The age of the participants varied with a range from 19 to 61 years old, with the majority of the participants being between the ages of 21 to 23 years old. None of the participants had any previous experience with eye-tracking experiments. All the participants were volunteers. The total number of participants used in this study was in line with the typical number of volunteers used in map user studies [60]. Since the experimental process required the approval of the Research and Ethics Committee of the University of West Attica, each participant was given a consent form to read and complete before participation. This document outlined the objectives of this research study, detailed the experimental process, and explained the utilization of the collected data.

3.1.3. Experimental Equipment and Setup

The EyeLink® 1000 Plus [61] eye tracker is used for the performance of the presented study. The experimental setup consists of two Personal Computers (PCs) and two display monitors. The first PC is used as the display computer, where EyeLink® software is installed along with SR Research’s tools, and the second PC is used as the host computer where all the computations of true gaze and eye-motion event detection and analysis take place. Moreover, the host monitor (primary monitor) is manipulated by the operator of the experiment, and the display monitor (secondary monitor) serves as the projection of the visual stimuli. Both PCs communicate via Ethernet. For the performance of the present study, gaze samples are collected at 500 Hz. The distance between the display monitor and the eye-tracking camera is fixed during the execution of the experimental process, while the distance between the participant and the display monitor is approximately 60 cm (the participant is positioned at approximately 1.75 times the width of the display, ensuring that the entire monitor is visible without exceeding a visual angle of 32° and 25° in the horizontal and in vertical dimensions, respectively).

3.1.4. Experimental Programming

The experiment’s overall structure is designed using the SR Research Experiment Builder (SREB) [62]. The workflow consists of actions followed by triggers, and the same pattern is repeated throughout the experiment. This sequence results in the creation of a hierarchically organized flow diagram. The components of the SRSE are categorized as “actions” and “triggers”, and they contain nodes that are inserted directly in the flow diagram utilizing visual programming. In more detail, visual programming facilitates the experimental design and procedures such as dragging nodes with different properties and functions, nesting different sequence blocks, and connecting them. The main experimental flow consists of three main blocks and a starting node. The first node added is the Camera Setup, which triggers the procedures of calibration, validation, and the setup of the eye-tracking device’s camera for every participant at the beginning of the experimental procedure. These blocks are nested in one another, and each one serves a different purpose. The first block is called the Sequence Block, the second block is called the Trial Block, and the third block is called the Recording Block. The Sequence Block contains the Drift Correction node, the Trial Block contains the Prepare Sequence node and the data source of the experiment, and the Recording Block contains two Display Screen nodes, each followed by a Timer node. Figure 2 depicts a flowchart of the main experiment.
Calibration accuracy must be maintained throughout the experimental procedure. Drift checks between a certain number of presented visual stimuli aid in keeping track of the accuracy. The trial data source is divided into three parts, each with 25 visual stimuli. The drift correction procedure is started after the observation of each part of the visual stimuli. If the observed accuracy does not fall into the desired limits, the calibration procedure is performed again. Therefore, the Sequence Block is iterated as many times as the number of parts of the data source. The Trial Block contains the data that are recorded in real time, and the Sequence Block draws the visual stimuli from the source to the display screen.

3.1.5. Procedure

The experiment is conducted in a controlled environment where variables such as lighting conditions, environmental vibrations, ambient noise, and the presence of others are carefully manipulated to optimize the accuracy of gaze data collection. Only the participant and the experimenter are allowed in the experimental room. Moderate natural ambient lighting is present, and adjustments are made to the lighting, especially when participants wear glasses or mascara. Potential questions expressed by the participants are addressed prior to the experiment during their review and signing of the consent form (see also Section 3.1.2). Prior to the beginning of the main experiment, the participants undergo a demo experiment to familiarize themselves with the process. At this point, it is important to mention that since any changes may reduce the accuracy of the collected data and require recalibration, the participants must maintain a comfortable position throughout the experiment during the experimental process.
After completing the necessary preparations, calibration and validation become essential steps in the process. Calibration involves the use of point targets with known locations, specific colors, sizes, and shapes. The target’s size is determined by the following two parameters: outer size and inner size. The default target is a filled circle with a central hole, where the outer size represents the diameter of the outer circle and the inner size represents the diameter of the inner circle, contributing to calibration accuracy. For peripheral detectability, the outer circle size is set to 17 pixels, and the inner circle size is set to 5 pixels. The chosen colors are red (RGB 255, 0, 0) for the outer circle and white (RGB 255, 255, 255) for the inner circle, against a gray background (RGB 230, 230, 230) to ensure sufficient contrast. The calibration process is based on the use of thirteen point targets.
The main experiment’s total duration is estimated at about six minutes, while the demo experiment lasts around one minute. The aforementioned durations are based on ideal conditions and do not include the required time to set up the participant in an optimal position. Calibration and validation together take one minute. Three drift corrections are incorporated between the three divisions of the data source, and a blank screen follows each stimulus appearance. The experiment is conducted under free-viewing conditions (i.e., participants do not have to answer any questions about the stimuli). This approach is followed in order to examine the salient locations of the investigated cartographic backgrounds, minimizing the cognitive parameters that may influence the procedure under the execution of specific map tasks.

3.2. Eye Movement Analysis

3.2.1. Basic Eye Movement Indices

A maximum acceptable error of one degree of visual angle is set for the collection of raw binocular gaze data. It is important to report that the implementation of intermediate drift corrections was found to be an effective method of maintaining accuracy throughout the process, with the capabilities of the equipment facilitating this approach. Based on the collected raw data, both fixation and saccade events are calculated using the corresponding detection algorithms incorporated in EyeLink® Data Viewer software [63]. The minimum threshold for fixation duration and saccade amplitude is set to 80 ms and one degree of visual angle, respectively.
For each visual stimulus and for each one of the basic derived metrics of fixation duration, the total number of fixations, and saccade amplitude, five indices are calculated in order to quantitatively describe the overall visual behavior of the participants during the observation. Specifically, for the three metrics mentioned above, the minimum, the maximum, the average, the median, and the standard deviation values are computed for each visual scene. The computation of the reported eye movement indices is fully automated using the scripting language of MATLAB (MathWorks®) software (R2022a).
The calculation of such indices is important since it facilitates the analysis of the variables under consideration. By comparing the relationships between these variables, meaningful conclusions can be drawn. The minimum values indicate the smallest observation among the selected categories, helping to identify outliers and understand the lower boundary of the range. Conversely, the maximum values represent the largest observation, helping to identify outliers and the upper bound of the range. Average values are measures of central tendency and represent the typical value. The median values, as the middle values in a sorted dataset, provide a robust measure of central tendency, which is particularly useful in the presence of outliers. Finally, the standard deviation values quantify variability by measuring how spread out the values are around the mean, with a higher standard deviation indicating greater variability.

3.2.2. Statistical Grayscale Heatmaps

Statistical grayscale heatmaps are produced to determine the salient objects observed by participants in the visual stimuli. Their generation is based on the execution of the corresponding function provided by the EyeMMV toolbox [64]. In more detail, the Gaussian parameter is determined by considering factors of the average distance between the participant and the display screen and the dimensions of the display screen. Converting the result from millimeters to pixels yields an approximate sigma value of 38 pixels. The Gaussian kernel is then computed as 6 times the sigma value, while the maximum vertical and horizontal values correspond to the resolution of the display screen. All pixel values are normalized within the range of 0 to 255, resulting in the generation of 8-bit grayscale images, where each intensity represents the probability of occurrence of each point. Hence, higher intensity values (255) indicate a higher probability of gaze.

3.2.3. The GraphGazeD Metric and Python Tool

The produced statistical grayscale heatmaps are utilized to calculate a new metric, called GraphGazeD, that describes quantitatively the differences in gaze patterns among the visual stimuli. The calculation of this metric facilitates the description of the differences in the allocation of visual attention by comparing the generated heatmaps, while a Python (v.3) tool is developed in order to support the computation of the introduced metric. In practice, comparisons are made among heatmaps that refer to cartographic backgrounds depicting the same geographic areas, at the same zoom level but derived from different online map services. However, instead of the result being a new map showing the differences, these differences are represented by a graph indicating the percentage of pixels that differ for the possible number of differences that can occur in terms of pixel intensity. Moreover, to model the produced graphs, curve-fitting techniques are applied. The GraphGazeD Python tool involves two main functions. The first function utilizes the OS, CV2, and Numpy Python libraries, while in order to execute the second function, the CSV, matplotlib.pyplot, scipy.optimize, and sklearn.metrics Python libraries/functions are imported. The scipy.optimize library is used to import the curve_fit function, while the sklearn.metrics library is employed to import the r_2 score function.
The first function refers to the comparison of the generated statistical grayscale heatmaps. The input is the directory of the heatmap folder, and the output is a simple .csv file that contains the generated data that are used to generate the corresponding graphs. There are 75 statistical grayscale heatmaps in total, divided into sets of five that derive from cartographic backgrounds depicting the same area at the same zoom level and distributed from different services. There are ten possible comparison pairs in each set, producing 150 combinations. Consequently, for each pair of heatmaps to be compared, one heatmap is subtracted from the other, generating another grayscale image with its pixels having the values of the absolute difference in the value of the compared heatmaps’ pixels. A threshold is set, ranging from values 0 to 255 as are the intensity values of the pixels of the initial statistical grayscale heatmaps. The values of the differences that are equal to or lower than the threshold are recorded and the number of those values is divided by the total number of pixels, thus calculating the percentage of pixels below or equal to the threshold’s value. At the end of this process, an output file is created that contains the name of each comparison pair of heatmaps, the threshold value set normalized in the range between 0 and 1, and the percentage of pixels below this threshold.
The second function is referred to as graph generation. The input is the .csv file produced from the first function and the output is a folder containing the graphs generated for each comparison pair. In each graph, the 256 threshold values normalized in the range between 0 and 1 are on the x-axis, and the percentage of pixels whose absolute difference is equal to or lower than the corresponding value is on the y-axis. The interval of the values on the graph depends on the data. Figure 3 visualizes the aforementioned process in three successive steps. For this example, an interval of 0.2 is selected. At this point, it is important to mention that the final plot is generated automatically (having a standard format) by the Python tool that is developed within the present study.
Curve-fitting techniques are applied to model the generated graphs. Specifically, three different mathematical models are employed, including (a) a hexic (6th degree) polynomial function, (b) the rectangular hyperbola, and (c) the logistic function (Figure 4). The above-mentioned functions are selected based on the similarity in their curves to the graphs generated from the data derived from the computations of the first function.

4. Results

The results of this research, along with the materials used for their computation, are the components of the OnMapGaze dataset. The 75 cartographic backgrounds that served as visual stimuli in the experiment are included in their native resolution (1920 × 1080 px). The raw gaze data, collected in free-viewing conditions for each participant observing each visual stimulus, are stored in text files. The data were recorded at a sample rate of 500 Hz, resulting in 1500 samples for each cartographic background observed by each participant. This implies that the total number of samples for all participants across all cartographic backgrounds is equal to 3,375,000. Furthermore, the aggregated data for each visual stimulus observed by all participants are included in the dataset as .txt files. The total number of samples for each cartographic background is 45,000. The 75 statistical grayscale heatmaps are included in their native resolution (1920 × 1080 px). The GraphGazeD metric is appended to the dataset, and it is composed of a .csv file that is made up of the data of the differences and the plots of these data along with the curve fitting functions. The components of the dataset are presented in Figure 5 and in more detail in Section 4.1, Section 4.2 and Section 4.3.

4.1. Basic Eye Movement Indices

Basic eye movement metrics generated based on the collected data were first presented by Liaskos and Krassanakis [65]. These metrics are incorporated into the OnMapGaze dataset. However, the previous study was a preliminary work that aimed to investigate the representative eye movement metrics that characterize the observation of this specific type of visual stimuli. In the present study, further analysis is performed on these data in order to create three different lists that rank (from higher to lower value) the experimental stimuli based on the generated values of basic eye movement metrics. In more detail, the indices of median fixation duration, median number of fixations, and median saccade amplitude are used for this analysis. The results of the ranking process are shown in Table 1.

4.2. Statistical Grayscale Heatmaps

The OnMapGaze dataset contains the statistical grayscale heatmaps generated for each visual stimulus of the performed experiment. These heatmaps constitute a cumulative product as they describe the visual behavior of all participants. Figure 6 shows three different examples of the highest-ranking visual stimuli presented in Table 1.

4.3. GraphGazeD Metric

GraphGazeD is calculated for 150 different comparison pairs of statistical grayscale heatmaps. The results are 150 graphs depicting the curve that quantitatively describes the difference between each pair. The graphs depict the curve fitting techniques while testing three different functions. The first test is performed with a hexic polynomial equation that has a good fit, and the coefficient of determination has values ranging from 0.71 to 0.98. The second test is performed with the rectangular hyperbola, which fits in most cases with a coefficient of determination equal to 1.00; however, there are cases without adjustment and a small value of the coefficient of determination. The last method is the sigmoidal function, which fits the data with a coefficient of determination value ranging from 0.91 to 0.99. Figure 7 depicts an example of a higher difference pair (on the left side) and a lower difference pair (on the right side).
Furthermore, Figure 8 presents the fitted curves that correspond to the highest and the lowest values of R2.

5. Discussion and Conclusions

In this study, the examination of aggregated gaze data provides insights into the cumulative behavior of participants while observing cartographic backgrounds from different online map services. The GraphGazeD metric is an attempt to scale down the problem of quantitatively describing the differences between two statistical grayscale heatmaps, and thus, the distinction between the salient objects of two cartographic backgrounds becomes evident. Additionally, the curve fitting functions are selected and applied based on their fit to the data visualized on the graphs. The functions are non-linear, and the coefficient of determination is used to measure their goodness of fit. In the case of nonlinear models, the R2 measure is often misinterpreted [66]. Although R2 has high values, additional data are required to assess the suitability of a regression equation for modeling the generated graphs [67]. In the case of the rectangular hyperbola function specifically, R2 is equal to 1 in the best-fitting case and equal to 0.39 in the worst-fitting case. This means that the coefficient represents the noise rather than the genuine relationship. In the case of the hectic polynomial function, the fitting is good and a further assessment of the suitability of the regression is needed, though the main setback is the computational capacity needed for such an equation. As for the logistic function, the fitting is good according to R2 and the data visualization resembled the s-shaped curve in the first quadrant of the sigmoid function’s curve. The relationship of the data was expected to be sigmoidal because the difference in the heatmaps is actually the difference in the intensity values of each heatmap that represent the probability of occurrence of each point.
The ability to describe the qualitative differences between two heatmaps is a valuable tool in the creation of user-oriented maps. Statistical grayscale heatmaps present the allocation of visual attention to the cartographic backgrounds observed by the participants. Comparing the differences between the observation and gaze behavior based on the salient objects, a further understanding can be obtained concerning the design of the visual stimuli. The examination of the cartographic elements that differ in the same geographic area, but using different online map services, reveals the hierarchical organization of the map elements and how it changes with the symbolization used. In this way, evaluating the visual variables that make the visual stimuli have many differences from each other will enhance the hierarchical organization and will achieve a balanced symbology [53]. Furthermore, the computation of the GraphGazeD metric can be utilized to model gaze behavior. Statistical grayscale heatmaps can serve as the main input for the creation of cartographic free-viewing saliency models [13,29,68] that predict participant gaze behavior, and the specific metric can serve to model the differences and similarities in gaze behavior [69]. If gaze behavior and its differences are known, the investigation of an effective and efficient map design is reinforced by evaluating alternative symbolization techniques on cartographic backgrounds and/or web maps, resulting in the design of user-oriented maps.
The basic eye movement indices contribute to describing the aggregate allocation of visual attention during the participants’ observations of the experiment’s visual stimuli. The computation of such indices refers to the gaze behavior produced and influenced by the salient objects of the cartographic backgrounds because the experiment is conducted under free-viewing conditions. The ranking list of the selected indices (Table 1) can be used as a reference to describe the complexity in the cartographic design of the visual stimuli in relation to the zoom level and the online map service. The highest-ranking cartographic backgrounds regarding the median fixation duration, the median number of fixations, and the median saccade amplitude are “Katerini_16z_ESRI”, “Alexandroupoli_16z_Google” and “Ioannina_14z_Google”, respectively. The qualitative description of the basic eye movement indices allows for the quantitative cross-check with the statistical grayscale heatmaps and the cartographic backgrounds, hence enhancing our comprehension. Observing the “Katerini_16z_ESRI” visual stimulus, it is evident that the allocation of visual attention is found on the complex road network with its contrasting color. As Skopeliti and Stamou [53] mention, color is the most frequently employed method of commenting upon the structural elements of the map [70]. In the case of “Alexandroupoli_16z_Google”, the high number of fixations is due to the large number of pictorial symbols concentrated in the depicted area. As for the case of “Ioannina_14z_Google”, the high ranking in median saccade amplitude can be understood when observing the salient objects disclosed by the statistical grayscale heatmap and the symbology present at these locations. The high saccade amplitude value is the distance between the point furthest to the left on the heatmap and the point furthest to the right on the heatmap. Both points follow the Gestalt principle of the focal point, attracting the participants’ attention, because they are pictorial points with darker hues than their backgrounds. Consequently, the calculated statistics complement the qualitative analysis of the grayscale statistical heatmaps. No patterns were detected for the correlation between the ranking and the online map service or the zoom level of the visual stimuli in this phase of the analysis; therefore, further analysis is required to gain a more comprehensive understanding of this issue.
The source code developed for the computation of the GraphGazeD metric is provided in a GitHub repository (see Supplementary Materials). There are two different versions that can be used for different purposes. The first version is created for the comparison of images within a directory with the same naming form as the files in this study. In this way, the flexibility of the script can aid the user in easily manipulating the data and making comparisons dedicated to the online map service used, the zoom level, or the geographic area depicted on the visual stimulus. The second version is designed for the comparison of images in general. Additionally, both versions can be found in a Jupyter notebook along with some indicative examples.
As mentioned above, the tool created to compute the GraphGazeD metric is provided freely. In this manner, the users are able to manipulate their data. For example, they can decide if their data will be used as input, either from a directory like in the script or as plain files. Additionally, the user can change the type of the output file where the differences are stored. The images to be compared can be any type they want, not only uint8 images like in this study. The threshold used is another variable that can be modified for more targeted results. As for the curve-fitting function, more coefficients can be added to examine the goodness of fit. Finally, the tool can be employed to create a comprehensive toolbox that will enable modeling the differences in aggregate gaze behavior among heatmaps. Hence, such a tool can be integrated with others, such as the PeyeMMV toolbox [71], in order to deliver complete and open-source software solutions to the scientific community.
The OnMapGaze dataset can be employed to model the visual behavior of participants who observed cartographic backgrounds derived from online map services in free-viewing conditions. The statistical grayscale heatmaps are produced by considering many samples collected with high accuracy, making them suitable to serve as objective ground truth in deep learning approaches [13,57,58]. Considering that cartographic backgrounds constitute, in practice, abstract (and not natural) images, modeling approaches can result in cartographic-oriented visual saliency models. The statistical indices provided offer insight into the visual strategies employed by the participants. The open-access nature of the collected dataset allows other researchers to analyze the data further and/or create new indices. However, the designed approach for heatmap comparison can be implemented in any other type of visual stimuli, including natural or more abstract images. Moreover, although the proposed metric cannot be directly compared to typical similarity measures used in image recognition since it is based on a graph-based approach that reveals the aggregated behavior in a range of threshold values, an interesting point for future research involves the comparison of the modeled graphs with such type of measures (e.g., Jaccard index, dice coefficient, and BF score).
The typical eye movement indices, the statistical grayscale heatmaps, and the computation of new metrics (such as GraphGazeD) could direct AI algorithms toward predicting visual behavior during the observation of different types of cartographic products. As a consequence, this will contribute to a better understanding of how map users react to cartographic backgrounds and, at the same time, serve as an objective evaluation approach for alternative choices in the design variables (i.e., visual, dynamic, and sound variables) and/or Graphical User Interfaces (GUIs).
This presented research study examines the visual perception of static cartographic backgrounds. This may be a limitation when modeling the visual behavior of dynamic products is needed. Therefore, in the next step, similar experiments could be performed considering animated maps and/or animated cartographic symbols [22]. Furthermore, the produced gaze dataset refers to different cities in Greece. Considering that the cartographic symbol density could differ from region to region (even in the same map service and/or zoom level), the development of similar datasets based on cartographic backgrounds referring to different geographical regions will help to better model the visual behavior of map users.

Supplementary Materials

The OnMapGaze dataset is available at https://doi.org/10.5281/zenodo.11520659 (accessed on 6 June 2024). The GraphGazeD tool is available at https://github.com/dgliaskos/GraphGazeD (accessed on 6 June 2024).

Author Contributions

Conceptualization, D.L. and V.K.; data curation, D.L.; methodology, D.L. and V.K.; software, D.L.; supervision, V.K.; visualization, D.L.; writing—original draft, D.L.; writing—review and editing, D.L. and V.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

This research received the approval of the Research and Ethics Committee of the University of West Attica (no. 47307, 15 May 2023).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The OnMapGaze dataset is available under Creative Commons Attribution 4.0 International. The GraphGazeD tool is available under the GNU General Public License v3.0.

Conflicts of Interest

The authors declare no conflicts of interest.

Correction Statement

This article has been republished with a minor correction to the order of the two subfigures of Figure 7. This change does not affect the scientific content of the article.

References

  1. Roth, R.E.; Donohue, R.G.; Sack, C.M.; Wallace, T.R.; Buckingham, T.M.A. A Process for Keeping Pace with Evolving Web Mapping Technologies. Cartogr. Perspect. 2014, 2014, 25–52. [Google Scholar] [CrossRef]
  2. Friedmannová, L.; Konečný, M.; Staněk, K. An adaptive cartographic visualization for support of the crisis management. In Proceedings of AutoCarto; Cartography and Geographic Information Society: Vancouver, Washington, USA, 2006; pp. 100–105. [Google Scholar]
  3. Reichenbacher, T. Adaptive methods for mobile cartography. In Proceedings of the 21st International Cartographic Conference, Durban, South Africa, 10–16 August 2003; International Cartographic Association: Durban, South Africa, 2003; pp. 1311–1322. [Google Scholar]
  4. Andrienko, G.L.; Andrienko, N.V. Interactive Maps for Visual Data Exploration. Int. J. Geogr. Inf. Sci. 1999, 13, 355–374. [Google Scholar] [CrossRef]
  5. Roth, R.E. An Empirically-Derived Taxonomy of Interaction Primitives for Interactive Cartography and Geovisualization. IEEE Trans. Vis. Comput. Graph. 2013, 19, 2356–2365. [Google Scholar] [CrossRef] [PubMed]
  6. Clarke, K.C. Mobile Mapping and Geographic Information Systems. Cartogr. Geogr. Inf. Sci. 2004, 31, 131–136. [Google Scholar] [CrossRef]
  7. Meng, L.; Reichenbacher, T. Map-based mobile services. In Map-Based Mobile Services: Theories, Methods and Implementations; Springer: Berlin/Heidelberg, Germany, 2005; pp. 1–10. [Google Scholar]
  8. Brewer, C.A.; Buttenfield, B.P. Framing Guidelines for Multi-Scale Map Design Using Databases at Multiple Resolutions. Cartogr. Geogr. Inf. Sci. 2007, 34, 3–15. [Google Scholar] [CrossRef]
  9. Roth, R.E.; Brewer, C.A.; Stryker, M.S. A Typology of Operators for Maintaining Legible Map Designs at Multiple Scales. Cartogr. Perspect. 2011, 29–64. [Google Scholar] [CrossRef]
  10. Boulos, M.N.K.; Burden, D. Web GIS in Practice V: 3-D Interactive and Real-Time Mapping in Second Life. Int. J. Health Geogr. 2007, 6, 51. [Google Scholar] [CrossRef] [PubMed]
  11. Goldsberry, K.P. Real-Time Traffic Maps; University of California: Santa Barbara, CA, USA, 2007. [Google Scholar]
  12. Liu, B.; Dong, W.; Zhan, Z.; Wang, S.; Meng, L. Differences in the Gaze Behaviours of Pedestrians Navigating between Regular and Irregular Road Patterns. ISPRS Int. J. Geoinf. 2020, 9, 45. [Google Scholar] [CrossRef]
  13. Keskin, M.; Krassanakis, V.; Çöltekin, A. Visual Attention and Recognition Differences Based on Expertise in a Map Reading and Memorability Study. ISPRS Int. J. Geoinf. 2023, 12, 45. [Google Scholar] [CrossRef]
  14. Krassanakis, V.; Cybulski, P. A Review on Eye Movement Analysis in Map Reading Process: The Status of the Last Decade. Geod. Cartogr. 2019, 68, 191–209. [Google Scholar] [CrossRef]
  15. Krassanakis, V.; Cybulski, P. Eye Tracking Research in Cartography: Looking into the Future. ISPRS Int. J. Geoinf. 2021, 10, 411. [Google Scholar] [CrossRef]
  16. Krassanakis, V. Aggregated Gaze Data Visualization Using Contiguous Irregular Cartograms. Digital 2021, 1, 130–144. [Google Scholar] [CrossRef]
  17. Çöltekin, A.; Heil, B.; Garlandini, S.; Fabrikant, S.I. Evaluating the Effectiveness of Interactive Map Interface Designs: A Case Study Integrating Usability Metrics with Eye-Movement Analysis. Cartogr. Geogr. Inf. Sci. 2009, 36, 5–17. [Google Scholar] [CrossRef]
  18. Herman, L.; Popelka, S.; Hejlova, V. Eye-Tracking Analysis of Interactive 3d Geovisualization. J. Eye Mov. Res. 2017, 10. [Google Scholar] [CrossRef] [PubMed]
  19. Opach, T.; Nossum, A. Evaluating the usability of cartographic animations with eye-movement analysis. In Proceedings of the 25th International Cartographic Conference, Paris, France, 3–8 July 2011; p. 11. [Google Scholar]
  20. Dong, W.; Liao, H.; Xu, F.; Liu, Z.; Zhang, S. Using Eye Tracking to Evaluate the Usability of Animated Maps. Sci. China Earth Sci. 2014, 57, 512–522. [Google Scholar] [CrossRef]
  21. Krassanakis, V.; Filippakopoulou, V.; Nakos, B. Detection of Moving Point Symbols on Cartographic Backgrounds. J. Eye Mov. Res. 2016, 9. [Google Scholar] [CrossRef]
  22. Cybulski, P.; Krassanakis, V. Motion Velocity as a Preattentive Feature in Cartographic Symbolization. J. Eye Mov. Res. 2023, 16. [Google Scholar] [CrossRef]
  23. Cybulski, P.; Krassanakis, V. The Role of the Magnitude of Change in Detecting Fixed Enumeration Units on Dynamic Choropleth Maps. Cartogr. J. 2021, 58, 251–267. [Google Scholar] [CrossRef]
  24. Cybulski, P. Animating Cartographic Meaning: Unveiling the Impact of Pictorial Symbol Motion Speed in Preattentive Processing. ISPRS Int. J. Geoinf. 2024, 13, 118. [Google Scholar] [CrossRef]
  25. Alaçam, Ö.; Dalc, M. A usability study of webmaps with eye tracking tool: The effects of iconic representation of information. In Proceedings of the Human-Computer Interaction. New Trends: 13th International Conference, HCI International 2009, Proceedings, Part I 13, San Diego, CA, USA, 19–24 July 2009; pp. 12–21. [Google Scholar]
  26. Manson, S.M.; Kne, L.; Dyke, K.R.; Shannon, J.; Eria, S. Using Eye-Tracking and Mouse Metrics to Test Usability of Web Mapping Navigation. Cartogr. Geogr. Inf. Sci. 2012, 39, 48–60. [Google Scholar] [CrossRef]
  27. Cybulski, P.; Horbiński, T. User Experience in Using Graphical User Interfaces of Web Maps. ISPRS Int. J. Geoinf. 2020, 9, 412. [Google Scholar] [CrossRef]
  28. Dolezalova, J.; Popelka, S. The use of simple graphs and cliques for analysis of cartographic eye-tracking data. In Proceedings of the 16th ITAT Conference Information Technologies—Applications and Theory, Tatranské Matliare, Slovakia, 15–19 September 2016; pp. 206–211. [Google Scholar]
  29. Cybulski, P.; Krassanakis, V. The Effect of Map Label Language on the Visual Search of Cartographic Point Symbols. Cartogr. Geogr. Inf. Sci. 2022, 49, 189–204. [Google Scholar] [CrossRef]
  30. Janowicz, K.; Sieber, R.; Crampton, J. GeoAI, Counter-AI, and Human Geography: A Conversation. Dialogues Hum. Geogr. 2022, 12, 446–458. [Google Scholar] [CrossRef]
  31. Robinson, A.C.; Çöltekin, A.; Griffin, A.L.; Ledermann, F. Cartography in GeoAI: Emerging themes and research challenges. In Proceedings of the 6th ACM SIGSPATIAL International Workshop on AI for Geographic Knowledge Discovery, Hamburg, Germany, 13 November 2023; pp. 1–2. [Google Scholar]
  32. Kang, Y.; Gao, S.; Roth, R. A review and synthesis of recent geoai research for cartography: Methods, applications, and ethics. In Proceedings of the AutoCarto, Redlands, CA, USA, 2–4 November 2022; pp. 2–4. [Google Scholar]
  33. Kang, Y.; Gao, S.; Roth, R.E. Artificial Intelligence Studies in Cartography: A Review and Synthesis of Methods, Applications, and Ethics. Cartogr. Geogr. Inf. Sci. 2024, 1–32. [Google Scholar] [CrossRef]
  34. Krassanakis, V.; Misthos, L.-M. Mouse Tracking as a Method for Examining the Perception and Cognition of Digital Maps. Digital 2023, 3, 127–136. [Google Scholar] [CrossRef]
  35. Keskin, M.; Qin, T.; Liu, B. A framework and practical guidelines for sharing open benchmark datasets in cartographic user research utilizing neuroscientific meth-ods. In Proceedings of the 18th International Conference on Location Based Services, Ghent, Belgium, 20–23 November 2023; pp. 204–213. [Google Scholar]
  36. Gökstorp, S.G.E.; Breckon, T.P. Temporal and Non-Temporal Contextual Saliency Analysis for Generalized Wide-Area Search within Unmanned Aerial Vehicle (UAV) Video. Vis. Comput. 2022, 38, 2033–2040. [Google Scholar] [CrossRef]
  37. Holmqvist, K.; Blignaut, P. Small Eye Movements Cannot Be Reliably Measured by Video-Based P-CR Eye-Trackers. Behav. Res. Methods 2020, 52, 2098–2121. [Google Scholar] [CrossRef]
  38. Duchowski, A. Eye Tracking Methodology: Theory and Practice. Springer: London, UK, 2007; ISBN 978-1-84628-608-7. [Google Scholar]
  39. Poole, A.; Ball, L.J. Eye Tracking in Human-Computer Interaction and Usability Research: Current Status and Future Prospects. Encycl. Hum. Comput. Interact. 2005, 211–219. [Google Scholar]
  40. Kiefer, P.; Giannopoulos, I.; Raubal, M.; Duchowski, A. Eye Tracking for Spatial Research: Cognition, Computation, Challenges. Spat. Cogn. Comput. 2017, 17, 1–19. [Google Scholar] [CrossRef]
  41. Brychtova, A.; Popelka, S.; Dobesova, Z. Eye-tracking methods for investigation of cartographic principles. In Proceedings of the SGEM 2012 Conference Proceedings, Albena, Bulgaria, 17–23 June 2012; Volume 2, pp. 1041–1048. [Google Scholar]
  42. Raschke, M.; Blascheck, T.; Burch, M. Visual analysis of eye tracking data. In Handbook of Human Centric Visualization; Springer: Berlin/Heidelberg, Germany, 2014; pp. 391–409. [Google Scholar]
  43. Blascheck, T.; Kurzhals, K.; Raschke, M.; Burch, M.; Weiskopf, D.; Ertl, T. Visualization of eye tracking data: A taxonomy and survey. Comput. Graph. Forum 2017, 36, 260–284. [Google Scholar] [CrossRef]
  44. Wang, J.; Antonenko, P.; Celepkolu, M.; Jimenez, Y.; Fieldman, E.; Fieldman, A. Exploring Relationships between Eye Tracking and Traditional Usability Testing Data. Int. J. Hum. Comput. Interact. 2019, 35, 483–494. [Google Scholar] [CrossRef]
  45. Špakov, O.; Miniotas, D. Visualization of Eye Gaze Data Using Heat Maps; Visualization of Eye Gaze Data using Heat Maps. Elektron. Ir Elektrotechnika 2007, 74, 55–58. Available online: https://eejournal.ktu.lt/index.php/elt/article/view/10372 (accessed on 6 June 2024).
  46. Pfeiffer, T.; Memili, C. Model-based real-time visualization of realistic three-dimensional heat maps for mobile eye tracking and eye tracking in virtual reality. In Proceedings of the Ninth Biennial ACM Symposium on Eye Tracking Research & Applications, Charleston, SC, USA, 14–17 March 2016; pp. 95–102. [Google Scholar]
  47. Ullah, I.; Jian, M.; Hussain, S.; Guo, J.; Yu, H.; Wang, X.; Yin, Y. A Brief Survey of Visual Saliency Detection. Multimed. Tools Appl. 2020, 79, 34605–34645. [Google Scholar] [CrossRef]
  48. Yan, F.; Chen, C.; Xiao, P.; Qi, S.; Wang, Z.; Xiao, R. Review of Visual Saliency Prediction: Development Process from Neurobiological Basis to Deep Models. Appl. Sci. 2021, 12, 309. [Google Scholar] [CrossRef]
  49. Paramanandam, K.; Kanagavalli, R. A review on deep learning techniques for saliency detection. In Information and Communication Technology for Competitive Strategies (ICTCS 2021) ICT: Applications and Social Interfaces; Springer: Berlin/Heidelberg, Germany, 2022; pp. 279–289. [Google Scholar]
  50. Abraham, S.E.; Kovoor, B.C. Visual Saliency Modeling with Deep Learning: A Comprehensive Review. J. Inf. Knowl. Manag. 2023, 22, 2250066. [Google Scholar] [CrossRef]
  51. He, B.; Dong, W.; Liao, H.; Ying, Q.; Shi, B.; Liu, J.; Wang, Y. A Geospatial Image Based Eye Movement Dataset for Cartography and GIS. Cartogr. Geogr. Inf. Sci. 2023, 50, 96–111. [Google Scholar] [CrossRef]
  52. Pappa, A.; Krassanakis, V. Examining the Preattentive Effect on Cartographic Backgrounds Utilizing Remote Mouse Tracking. Abstr. ICA 2022, 5, 111. [Google Scholar] [CrossRef]
  53. Skopeliti, A.; Stamou, L. Online Map Services: Contemporary Cartography or a New Cartographic Culture? ISPRS Int. J. Geoinf. 2019, 8, 215. [Google Scholar] [CrossRef]
  54. Liaskos, D. Searching for Salient Objects on Cartographic Backgrounds Used in Online Map Services: An Experimental Study Based on Eye Movement Analysis. Master’s Thesis, University of West Attica, Aigaleo, Greece, 2023. [Google Scholar]
  55. Kummerer, M.; Wallis, T.S.A.; Bethge, M. Saliency benchmarking made easy: Separating models, maps and metrics. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 770–787. [Google Scholar]
  56. Krassanakis, V.; Lelli, A.; Lokka, I.-E.; Filippakopoulou, V.; Nakos, B. Searching for salient locations in topographic maps. In Proceedings of the First International Workshop on Solutions for Automatic Gaze-Data Analysis (SAGA 2013), Bielefeld, Germany, 24–25 October 2013; pp. 41–44. [Google Scholar]
  57. Krassanakis, V.; Da Silva, M.P.; Ricordel, V. Monitoring Human Visual Behavior during the Observation of Unmanned Aerial Vehicles (Uavs) Videos. Drones 2018, 2, 36. [Google Scholar] [CrossRef]
  58. Perrin, A.-F.; Krassanakis, V.; Zhang, L.; Ricordel, V.; Perreira Da Silva, M.; Le Meur, O. Eyetrackuav2: A Large-Scale Binocular Eye-Tracking Dataset for Uav Videos. Drones 2020, 4, 2. [Google Scholar] [CrossRef]
  59. Tzelepis, N.; Kaliakouda, A.; Krassanakis, V.; Misthos, L.-M.; Nakos, B. Evaluating the Perceived Visual Complexity of Multidirectional Hill-Shading. Geod. Cartogr. 2020, 69, 161–172. [Google Scholar]
  60. Bergmann Martins, V.; Rosa Amorim, F.; Augusto Reolon Schmidt, M.; Stamato Delazari, L. Study about the Appropriate Number of Participants in Map User Studies. Int. J. Cartogr. 2023, 9, 196–209. [Google Scholar] [CrossRef]
  61. SR Research Ltd. EyeLink® 1000 Plus User Manual Desktop, LCD Arm, Tower, Primate and Long Range Mounts Remote, 2000 Hz and Fiber Optic Camera Upgrades; SR Research Ltd.: Oakville, ON, Canada, 2021. [Google Scholar]
  62. SR Research Ltd. SR Research Experiment Builder 2.3.1; SR Research Ltd.: Oakville, ON, Canada, 2020. [Google Scholar]
  63. SR Research Ltd. EyeLink® Data Viewer 4.2.1; SR Research Ltd.: Oakville, ON, Canada, 2021. [Google Scholar]
  64. Krassanakis, V.; Filippakopoulou, V.; Nakos, B. EyeMMV Toolbox: An Eye Movement Post-Analysis Tool Based on a Two-Step Spatial Dispersion Threshold for Fixation Identification. J. Eye Mov. Res. 2014, 7. [Google Scholar] [CrossRef]
  65. Liaskos, D.; Krassanakis, V. Modeling the visual perception of cartographic backgrounds used in web services through eye tracking techniques (in Greek). In Proceedings of the 16th National Cartographic Conference, Egaleo, Greece, 2–4 November 2023; Hellenic Cartographic Society: Thessaloniki, Greece, 2023; pp. 229–242. [Google Scholar]
  66. Kva Lseth, T.O. Note on the R2 Measure of Goodness of Fit for Nonlinear Models. Bull. Psychon. Soc. 1983, 21, 79–80. [Google Scholar] [CrossRef]
  67. Cornell, J.A.; Berger, R.D. Factors That Influence the Value of the Coefficient of Determination in Simple Linear and Nonlinear Regression Models. Phytopathology 1987, 77, 63–70. [Google Scholar] [CrossRef]
  68. Keskin, M.; Kettunen, P. Potential of Eye-Tracking for Interactive Geovisual Exploration Aided by Machine Learning. Int. J. Cartogr. 2023, 9, 150–172. [Google Scholar] [CrossRef]
  69. Wooding, D.S. Eye Movements of Large Populations: II. Deriving Regions of Interest, Coverage, and Similarity Using Fixation Maps. Behav. Res. MethodsInstrum. Comput. 2002, 34, 518–528. [Google Scholar] [CrossRef] [PubMed]
  70. Robinson, A.H. The Look of Maps: An Examination of Cartographic Design. Am. Cartogr. 1986, 13, 280. [Google Scholar] [CrossRef]
  71. Krassanakis, V. PeyeMMV: Python Implementation of EyeMMV’s Fixation Detection Algorithm. Softw. Impacts 2023, 15, 100475. [Google Scholar] [CrossRef]
Figure 1. Indicative samples of experimental visual stimuli.
Figure 1. Indicative samples of experimental visual stimuli.
Mti 08 00049 g001
Figure 2. A flowchart of the main experiment in the SR Research Experiment Builder environment.
Figure 2. A flowchart of the main experiment in the SR Research Experiment Builder environment.
Mti 08 00049 g002
Figure 3. Graph-based metric computation in three successive steps. For the illustrative example, an interval of 0.2 is selected.
Figure 3. Graph-based metric computation in three successive steps. For the illustrative example, an interval of 0.2 is selected.
Mti 08 00049 g003
Figure 4. Curve-fitting examples for modeling the graph-based metric. Blue lines represent the calculated values before fitting process.
Figure 4. Curve-fitting examples for modeling the graph-based metric. Blue lines represent the calculated values before fitting process.
Mti 08 00049 g004
Figure 5. The components of the OnMapGaze dataset. Blue lines represent the calculated values before fitting process.
Figure 5. The components of the OnMapGaze dataset. Blue lines represent the calculated values before fitting process.
Mti 08 00049 g005
Figure 6. Aggregated statistical grayscale heatmaps produced for the highest-ranking visual stimuli.
Figure 6. Aggregated statistical grayscale heatmaps produced for the highest-ranking visual stimuli.
Mti 08 00049 g006
Figure 7. An example of a higher difference pair (on the left side) and a lower difference pair (on the right side).
Figure 7. An example of a higher difference pair (on the left side) and a lower difference pair (on the right side).
Mti 08 00049 g007
Figure 8. Fitted curves (hexic (sixth degree) polynomial—(left), rectangular hyperbola—(middle), logistic function—(right)) that correspond to the highest (up) and lowest (down) values of R2. Blue lines represent the calculated values before fitting process.
Figure 8. Fitted curves (hexic (sixth degree) polynomial—(left), rectangular hyperbola—(middle), logistic function—(right)) that correspond to the highest (up) and lowest (down) values of R2. Blue lines represent the calculated values before fitting process.
Mti 08 00049 g008
Table 1. Ranking process of all experimental visual stimuli based on basic eye movement metrics.
Table 1. Ranking process of all experimental visual stimuli based on basic eye movement metrics.
Median Fixation Duration (ms)Median Number of FixationsMedian Saccade Amplitude (°)
Katerini_16z_ESRIAlexandroupoli_16z_GoogleIoannina_14z_Google
Thessaloniki_12z_GoogleIoannina_14z_BingAthens_12z_Google
Heraklion_12z_OSMChalcis_14z_GoogleThessaloniki_12z_ESRI
Katerini_16z_BingThessaloniki_12z_WikimediaHeraklion_12z_Google
Katerini_16z_OSMTrikala_14z_GoogleVolos_14z_ESRI
Katerini_16z_WikimediaXanthi_16z_ESRILarissa_12z_Google
Heraklion_12z_ESRIAlexandroupoli_16z_BingXanthi_16z_Wikimedia
Larissa_12z_BingKavala_16z_BingAthens_12z_Bing
Athens_12z_BingSerres_14z_ESRILarissa_12z_ESRI
Volos_14z_GoogleKaterini_16z_GoogleLarissa_12z_OSM
Alexandroupoli_16z_ESRIKavala_16z_GoogleAthens_12z_OSM
Volos_14z_WikimediaChalcis_14z_BingHeraklion_12z_Bing
Serres_14z_BingTrikala_14z_OSMThessaloniki_12z_Wikimedia
Serres_14z_GoogleTrikala_14z_WikimediaVolos_14z_Wikimedia
Larissa_12z_WikimediaChalcis_14z_OSMPatras_12z_Google
Thessaloniki_12z_BingIoannina_14z_OSMKalamata_16z_Wikimedia
Heraklion_12z_GoogleSerres_14z_BingAlexandroupoli_16z_Google
Larissa_12z_OSMVolos_14z_ESRIPatras_12z_ESRI
Athens_12z_ESRIPatras_12z_WikimediaIoannina_14z_Bing
Thessaloniki_12z_ESRITrikala_14z_ESRITrikala_14z_ESRI
Volos_14z_ESRIIoannina_14z_GoogleSerres_14z_Wikimedia
Patras_12z_ESRIXanthi_16z_OSMXanthi_16z_ESRI
Volos_14z_OSMHeraklion_12z_OSMTrikala_14z_OSM
Athens_12z_WikimediaKaterini_16z_OSMLarissa_12z_Wikimedia
Ioannina_14z_BingKaterini_16z_WikimediaVolos_14z_OSM
Kalamata_16z_OSMLarissa_12z_BingKaterini_16z_Wikimedia
Heraklion_12z_WikimediaAlexandroupoli_16z_ESRIKalamata_16z_Google
Larissa_12z_ESRIVolos_14z_WikimediaSerres_14z_ESRI
Xanthi_16z_GoogleLarissa_12z_WikimediaHeraklion_12z_Wikimedia
Serres_14z_WikimediaThessaloniki_12z_BingThessaloniki_12z_Bing
Athens_12z_GoogleHeraklion_12z_GoogleIoannina_14z_OSM
Patras_12z_BingAthens_12z_ESRIChalcis_14z_Bing
Xanthi_16z_BingThessaloniki_12z_ESRIChalcis_14z_OSM
Kalamata_16z_WikimediaPatras_12z_ESRITrikala_14z_Google
Patras_12z_GoogleVolos_14z_OSMIoannina_14z_Wikimedia
Volos_14z_BingAthens_12z_WikimediaPatras_12z_Bing
Chalcis_14z_GoogleKalamata_16z_OSMAlexandroupoli_16z_Bing
Patras_12z_WikimediaLarissa_12z_ESRIKaterini_16z_ESRI
Kalamata_16z_ESRIXanthi_16z_GoogleThessaloniki_12z_OSM
Serres_14z_OSMSerres_14z_WikimediaXanthi_16z_OSM
Alexandroupoli_16z_WikimediaPatras_12z_BingKavala_16z_ESRI
Thessaloniki_12z_WikimediaXanthi_16z_BingAlexandroupoli_16z_OSM
Ioannina_14z_ESRIKalamata_16z_WikimediaXanthi_16z_Google
Trikala_14z_BingVolos_14z_BingLarissa_12z_Bing
Ioannina_14z_WikimediaKalamata_16z_ESRIKaterini_16z_Bing
Trikala_14z_GoogleSerres_14z_OSMPatras_12z_Wikimedia
Athens_12z_OSMAlexandroupoli_16z_WikimediaVolos_14z_Bing
Kavala_16z_OSMIoannina_14z_ESRIAthens_12z_ESRI
Thessaloniki_12z_OSMTrikala_14z_BingKaterini_16z_Google
Xanthi_16z_ESRIIoannina_14z_WikimediaChalcis_14z_ESRI
Kavala_16z_ESRIAthens_12z_OSMTrikala_14z_Bing
Alexandroupoli_16z_OSMKavala_16z_OSMChalcis_14z_Wikimedia
Kavala_16z_WikimediaThessaloniki_12z_OSMChalcis_14z_Google
Alexandroupoli_16z_BingKavala_16z_ESRIIoannina_14z_ESRI
Kalamata_16z_BingAlexandroupoli_16z_OSMHeraklion_12z_OSM
Heraklion_12z_BingKavala_16z_WikimediaKalamata_16z_ESRI
Larissa_12z_GoogleKalamata_16z_BingKavala_16z_Wikimedia
Xanthi_16z_WikimediaHeraklion_12z_BingKavala_16z_Google
Kavala_16z_BingLarissa_12z_GoogleXanthi_16z_Bing
Chalcis_14z_ESRIXanthi_16z_WikimediaVolos_14z_Google
Kalamata_16z_GoogleChalcis_14z_ESRISerres_14z_Google
Trikala_14z_ESRIKalamata_16z_GoogleAlexandroupoli_16z_ESRI
Ioannina_14z_GoogleChalcis_14z_WikimediaTrikala_14z_Wikimedia
Chalcis_14z_WikimediaPatras_12z_OSMPatras_12z_OSM
Xanthi_16z_OSMKaterini_16z_BingKaterini_16z_OSM
Patras_12z_OSMHeraklion_12z_ESRIHeraklion_12z_ESRI
Serres_14z_ESRIAthens_12z_BingKalamata_16z_Bing
Katerini_16z_GoogleKaterini_16z_ESRIAthens_12z_Wikimedia
Kavala_16z_GoogleThessaloniki_12z_GoogleKalamata_16z_OSM
Chalcis_14z_BingVolos_14z_GoogleSerres_14z_OSM
Alexandroupoli_16z_GoogleSerres_14z_GoogleKavala_16z_Bing
Trikala_14z_OSMLarissa_12z_OSMKavala_16z_OSM
Trikala_14z_WikimediaHeraklion_12z_WikimediaThessaloniki_12z_Google
Chalcis_14z_OSMAthens_12z_GoogleSerres_14z_Bing
Ioannina_14z_OSMPatras_12z_GoogleAlexandroupoli_16z_Wikimedia
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liaskos, D.; Krassanakis, V. OnMapGaze and GraphGazeD: A Gaze Dataset and a Graph-Based Metric for Modeling Visual Perception Differences in Cartographic Backgrounds Used in Online Map Services. Multimodal Technol. Interact. 2024, 8, 49. https://doi.org/10.3390/mti8060049

AMA Style

Liaskos D, Krassanakis V. OnMapGaze and GraphGazeD: A Gaze Dataset and a Graph-Based Metric for Modeling Visual Perception Differences in Cartographic Backgrounds Used in Online Map Services. Multimodal Technologies and Interaction. 2024; 8(6):49. https://doi.org/10.3390/mti8060049

Chicago/Turabian Style

Liaskos, Dimitrios, and Vassilios Krassanakis. 2024. "OnMapGaze and GraphGazeD: A Gaze Dataset and a Graph-Based Metric for Modeling Visual Perception Differences in Cartographic Backgrounds Used in Online Map Services" Multimodal Technologies and Interaction 8, no. 6: 49. https://doi.org/10.3390/mti8060049

Article Metrics

Back to TopTop