Predicting mid-air gestural interaction with public displays based on audience behaviour

https://doi.org/10.1016/j.ijhcs.2020.102497Get rights and content

Highlights

  • A 35-days long field study focused on a public display deployment.

  • Results show that audience size and behaviour significantly influence user(s) interactions.

  • Predictor models are built to forecast users’ interaction duration and distance.

  • A visualisation tool is made available to visualise predictions based on audience bahaviour.

  • Both the visualisation tool and the predictor model can be adapted to other deployments.

Abstract

Knowledge about the expected interaction duration and expected distance from which users will interact with public displays can be useful in many ways. For example, knowing upfront that a certain setup will lead to shorter interactions can nudge space owners to alter the setup. If a system can predict that incoming users will interact at a long distance for a short amount of time, it can accordingly show shorter versions of content (e.g., videos/advertisements) and employ at-a-distance interaction modalities (e.g., mid-air gestures). In this work, we propose a method to build models for predicting users’ interaction duration and distance in public display environments, focusing on mid-air gestural interactive displays. First, we report our findings from a field study showing that multiple variables, such as audience size and behaviour, significantly influence interaction duration and distance. We then train predictor models using contextual data, based on the same variables. By applying our method to a mid-air gestural interactive public display deployment, we build a model that predicts interaction duration with an average error of about 8 s, and interaction distance with an average error of about 35 cm. We discuss how researchers and practitioners can use our work to build their own predictor models, and how they can use them to optimise their deployment.

Introduction

Interactive public displays can today be found in airports, universities, shopping malls and more. Although it is generally known that users interact with public displays for very short amounts of time, interaction durations vary widely (Davies et al., 2014; Müller et al., 2012; Parker et al., 2018). The same is true for interaction distances, particularly when using interaction techniques such as mid-air gestures (Müller et al., 2012; Nancel et al., 2015). Knowing the expected interaction duration and distance upfront can bring in a plethora of benefits to the different stakeholders of a public display. For instance, if a system is aware that the current situation will result in the user interacting at a certain distance, it can dynamically determine which interaction modality to employ (Dingler et al., 2015b). Furthermore, depending on the expected interaction duration, the system could, for example, dynamically choose which version of an advertisement video to show according to the video length. In addition to run-time benefits, knowing which factors influence interaction duration and distance can be of great value to stakeholders who would then be able to tweak their setups to achieve the optimal user experience.

To address such issues, designers developed methods that try to keep users interacting longer by, for example, showing autopoiesic content (Memarovic et al., 2011), or presenting new content immediately after the user has finished interacting with current content (Alt et al., 2016). Moreover, previous works reported that users interact at different distances, particularly when using at-a-distance interaction techniques such as mid-air gestures (Müller et al., 2012; Nancel et al., 2015). The surrounding environment also affects how users position themselves: for instance, benches surrounding the display potentially result in larger audiences, which in turn discourage users from positioning themselves close to the display (Gentile et al., 2017a). These facts have led researchers to study how to guide users into positioning themselves in the optimal sweet spot (Alt et al., 2015; Zhang et al., 2014).

In contrast to display-centred approaches discussed in previous work, in this paper we focus on the people near the display, and in particular on estimating how the audience could affect possible interactions with the display itself. In more detail, here we present a method for building models to predict the duration and distance of interactions based on the behaviour of users (i.e., people who actually interact with the display) and audience (i.e., people who sit or stay around the display), and on the relationship between them.

To this end, we first analysed the behaviour of users and audience in a real-world deployment of a public display. We chose a deployment in which mid-air gestures were employed for interaction because this modality often results in varying interaction durations (Ackad et al., 2016) and interaction distances (Müller et al., 2012). Data analysis revealed that the users’ interaction duration and distance from the display are significantly influenced by the number of users, the size of the audience, the user-audience relationship, and the audience’s gaze towards the user(s). We then used the collected data to train predictor models based on our method, able to estimate the probability density functions (PDFs) for interaction duration and distance (see Fig. 1), using audience-related information. Our solution estimates the interaction duration with a mean absolute error (MAE) of about 8 seconds, and the interaction distance with a MAE of about 35 cm. We describe our approach in details and discuss how the outcomes of our study, which include the predictor model and a visualisation tool, can be reused by researchers and designers to build predictor models for their public display setups.

We discuss how this work can help space owners and designers to optimise user engagement and experience, e.g., by (a) building similar models and (b) using them to predict how settings will influence users to (c) ultimately optimise the setup. We also explain how to leverage this knowledge to adapt content and interaction modalities in real time.

The contribution of this work is threefold:

  • we report our findings from a 35-day long observation of our deployment to identify factors that influence users’ behaviour;

  • we propose and evaluate a machine learning approach based on expectation maximisation, aimed at generating predictor models using the collected data;

  • we describe how similar models can be developed for predicting user behaviour on public displays.

To help visualise the predictions, we implemented a visualisation tool that can be adapted to other similar deployments. The tool is freely available and open source1

Section snippets

Related work

Our work builds on previous work on the analysis, modelling, and prediction of the behaviour of users of interactive public displays. In particular, we focused on display applications that use mid-air gestures in order to provide interactivity to the users.

In this section, we provide an overview of the most significant previous works on these main topics.

Audience's impact on interaction

To study the impact of audience presence and behaviour on interaction in our deployment, we conducted a longitudinal study in order to observe users’ behaviour while interacting with a public display. In this section, we provide a brief description of the deployment we analysed, the data collection process and the outcome from our data analysis.

Predicting interaction duration and distance

In the previous section, we described the process of observing users’ behaviour and collecting data related to a set of surrounding information, e.g., the audience’s presence and behaviour, in order to understand the relationship between such audience-related data and the users’ behaviour while interacting with a display. Using this information, we concluded that some of the observed variables have a more significant influence than others.

In this section, our goal is to understand how to use

Discussion and uses of the model

In this section, we discuss the results of the field study and the subsequent statistical analysis, and we describe how display providers and space owners can benefit from the proposed model.

Future work

As discussed in Section 3.2, the data collection process can be very time consuming if not automated. A direction for future work is to automatically detect the situation around the display. Tools such as those proposed by Williamson and Williamson (2014) and by Elhart et al. (2017) already track some of the relevant aspects of audience behaviour (e.g., audience size). These tools can be extended to collect additional information such as the audience’s gaze by, for example, using head pose

Conclusion

In this work, we studied the impact of audience size and behaviour, along with contextual information, to predict duration and distance of interactions via mid-air gesture on public displays. Through a field study, we found that the interaction duration is influenced by the number of users, the audience size and the relationship between the users and the audience. We also found that the interaction distance is influenced by the audience size and whether they gaze at the user(s). We used the

CRediT authorship contribution statement

Vito Gentile: Conceptualization, Methodology, Software, Validation, Formal analysis, Investigation, Data curation, Writing - original draft, Writing - review & editing, Visualization. Mohamed Khamis: Conceptualization, Methodology, Validation, Formal analysis, Investigation, Data curation, Writing - original draft, Writing - review & editing, Visualization. Fabrizio Milazzo: Conceptualization, Methodology, Software. Salvatore Sorce: Conceptualization, Validation, Investigation, Data curation,

Declaration of Competing Interest

We wish to confirm that there are no known conflicts of interest associated with this publication and there has been no significant financial support for this work that could have influenced its outcome. We confirm that the manuscript has been read and approved by all named authors and that there are no other persons who satisfied the criteria for authorship but are not listed. We further confirm that the order of authors listed in the manuscript has been approved by all of us. We confirm that

Acknowledgements

This research was supported by the Bavarian State Ministry of Education, Science and the Arts in the framework of the Center Digitization.Bavaria (ZD.B; Grant no. M7426.6.4) in Germany, and the German Research Foundation (DFG), Grant No. AL 1899/2-1. Moreover, this work was partially funded on two research grants by the Italian Ministry of University and Research (MIUR), namely project BookAlive (Grant no. PAC02L2_00068) and project NEPTIS (Grant no. PON03PE_00214_3). This work was supported,

References (98)

  • C.P. Gerba et al.

    Bacterial contamination of computer touch screens

    Am. J. Infect. Control

    (2016)
  • N. Valkanova et al.

    Public visualization displays of citizen data: Design, impact and implications

    International Journal of Human-Computer Studies

    (2015)
  • C. Ackad et al.

    An in-the-wild study of learning mid-air gestures to browse hierarchical information at a large interactive public display

    Proceedings of the 2015 ACM International Joint Conference on Pervasive and Ubiquitous Computing

    (2015)
  • C. Ackad et al.

    Skeletons and silhouettes: comparing user representations at a gesture-based large display

    Proceedings of the 2016 CHI Conference on Human Factors in Computing Systems

    (2016)
  • D. Ahlström et al.

    Are you comfortable doing that?: acceptance studies of around-device gestures in and for public settings

    Proceedings of the 16th International Conference on Human-computer Interaction with Mobile Devices & #38; Services

    (2014)
  • R. Aigner et al.

    Understanding mid-air hand gestures: a study of human preferences in usage of gesture types for hci

    Microsoft Res. TechReport MSR-TR-2012-111

    (2012)
  • F. Alt et al.

    Gravityspot: guiding users in front of public displays using on-screen visual cues

    Proceedings of the 28th Annual ACM Symposium on User Interface Software &; Technology

    (2015)
  • F. Alt et al.

    Don’t disturb me: understanding secondary tasks on public displays

    Proceedings of the 5th ACM International Symposium on Pervasive Displays

    (2016)
  • M. Argyle et al.

    Eye-contact, distance and affiliation

    Sociometry

    (1965)
  • T. Ballendat et al.

    Proxemic interaction: designing for a proximity and orientation-aware environment

    ACM International Conference on Interactive Tabletops and Surfaces

    (2010)
  • R. de la Barré et al.

    Touchless interaction-novel chances and challenges

  • M. Behrens et al.

    Exploring the effect of spatial layout on mediated urban interactions

    Proceedings of the 2Nd ACM International Symposium on Pervasive Displays

    (2013)
  • G. Beyer et al.

    The puppeteer display: attracting and actively shaping the audience with an interactive public banner display

    Proceedings of the 2014 Conference on Designing Interactive Systems

    (2014)
  • C.M. Bishop

    Machine Learning and Pattern Recognition

    (2006)
  • F. Block et al.

    Fluid grouping: Quantifying group engagement around interactive tabletop exhibits in the wild

    Proceedings of the 33rd Annual ACM Conference on Human Factors in Computing Systems

    (2015)
  • P. Bottoni et al.

    Copuppet : collaborative interaction in virtual puppetry

  • H. Brignull et al.

    Enticing people to interact with large public displays in public spaces

    In Proceedings of the IFIP International Conference on Human-Computer Interaction (INTERACT 2003

    (2003)
  • G. Broll et al.

    Design and evaluation of techniques for mobile interaction with dynamic nfc-displays

    Proceedings of the Fifth International Conference on Tangible, Embedded, and Embodied Interaction

    (2011)
  • D. Buschek et al.

    Improving accuracy, applicability and usability of keystroke biometrics on mobile touchscreen devices

    Proceedings of the 33rd Annual ACM Conference on Human Factors in Computing Systems

    (2015)
  • D. Buschek et al.

    Improving accuracy in back-of-device multitouch typing: A clustering-based approach to keyboard updating

    Proceedings of the 19th International Conference on Intelligent User Interfaces

    (2014)
  • S. Clinch et al.

    A survey of pervasive displays for information presentation

    IEEE Pervasive Comput.

    (2016)
  • J. Coenen et al.

    The concurrent use of touch and mid-air gestures or floor mat interaction on a public display

    Proceedings of the 6th ACM International Symposium on Pervasive Displays

    (2017)
  • P. Dalsgaard et al.

    Designing urban media façades: Cases and challenges

    Proceedings of the SIGCHI Conference on Human Factors in Computing Systems

    (2010)
  • P. Dalsgaard et al.

    Participation gestalt: analysing participatory qualities of interaction in public space

    Proceedings of the 2016 CHI Conference on Human Factors in Computing Systems

    (2016)
  • P. Dalsgaard et al.

    Performing perception—staging aesthetics of interaction

    ACM Trans. Comput.-Hum. Interact.

    (2008)
  • N.S. Dalton et al.

    Display blindness?: Looking again at the visibility of situated displays using eye-tracking

    Proceedings of the 33rd Annual ACM Conference on Human Factors in Computing Systems

    (2015)
  • N. Davies et al.

    Pervasive displays: Understanding the future of digital signage

    (2014)
  • K. Deb et al.

    A fast and elitist multiobjective genetic algorithm: nsga-ii

    IEEE Trans. Evol. Comput.

    (2002)
  • T. Dingler et al.

    Ucanvas: A Web Framework for Spontaneous Smartphone Interaction with Ubiquitous Displays

  • T. Dingler et al.

    Interaction proxemics: combining physical spaces for seamless gesture interaction

    Proceedings of the 4th International Symposium on Pervasive Displays

    (2015)
  • I. Elhart et al.

    Audience monitor: An open source tool for tracking audience mobility in front of pervasive displays

    Proceedings of the 6th ACM International Symposium on Pervasive Displays

    (2017)
  • O. Erazo et al.

    Predicting task execution time on natural user interfaces based on touchless hand gestures

    Proceedings of the 20th International Conference on Intelligent User Interfaces

    (2015)
  • G. Fanelli et al.

    Real time head pose estimation with random regression forests

    CVPR 2011

    (2011)
  • S. Frintrop et al.

    Computational visual attention systems and their cognitive foundations: a survey

    ACM Trans. Appl. Percept.

    (2010)
  • V. Gentile

    Designing Touchless Gestural Interfaces for Public Displays

    (2017)
  • V. Gentile et al.

    They are looking at me!: understanding how audience presence impacts on public display users

    Proceedings of the 6th ACM International Symposium on Pervasive Displays

    (2017)
  • V. Gentile et al.

    Investigating how user avatar in touchless interfaces affects perceived cognitive load and two-handed interactions

    Proceedings of the 6th ACM International Symposium on Pervasive Displays

    (2017)
  • V. Gentile et al.

    Touchless interfaces for public displays: can we deliver interface designers from introducing artificial push button gestures?

    Proceedings of the International Working Conference on Advanced Visual Interfaces

    (2016)
  • E. Hall

    The hidden dimension

    (1992)
  • J.D. Hincapié-Ramos et al.

    Consumed endurance: a metric to quantify arm fatigue of mid-air interactions

    Proceedings of the 32Nd Annual ACM Conference on Human Factors in Computing Systems

    (2014)
  • K. Hinckley et al.

    Pre-touch sensing for mobile interaction

    Proceedings of the 2016 CHI Conference on Human Factors in Computing Systems

    (2016)
  • J. Huang et al.

    User see, user point: gaze and cursor alignment in web search

    Proceedings of the SIGCHI Conference on Human Factors in Computing Systems

    (2012)
  • B. Huber et al.

    Detecting user intention at public displays from foot positions

    Proceedings of the 33rd Annual ACM Conference on Human Factors in Computing Systems

    (2015)
  • J.D. Hunter

    Matplotlib: a 2d graphics environment

    Comput. Sci. Eng.

    (2007)
  • M. Khamis et al.

    A field study on spontaneous gaze-based interaction with a public display using pursuits

    Adjunct Proceedings of the 2015 ACM International Joint Conference on Pervasive and Ubiquitous Computing and Proceedings of the 2015 ACM International Symposium on Wearable Computers

    (2015)
  • M. Khamis et al.

    Which one is me?: Identifying oneself on public displays

    Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems

    (2018)
  • M. Khamis et al.

    Gtmopass: two-factor authentication on public displays using gazetouch passwords and personal mobile devices

    Proceedings of the 6th International Symposium on Pervasive Displays

    (2017)
  • M. Khamis et al.

    Eyescout: active eye tracking for position and movement independent gaze interaction with large public displays

    Proceedings of the 30th Annual ACM Symposium on User Interface Software & Technology

    (2017)
  • M. Khamis et al.

    Textpursuits: Using text for pursuits-based interaction and calibration on public displays

    Proceedings of the 2016 ACM International Joint Conference on Pervasive and Ubiquitous Computing

    (2016)
  • Cited by (2)

    View full text