Next Article in Journal
Secure PD-NOMA with Multi-User Cooperation and User Clustering in Both Uplink and Downlink PD-NOMA
Next Article in Special Issue
A near Real-Time Monitoring System Using Public WI-FI Data to Evaluate COVID-19 Social Distance Measures
Previous Article in Journal
MTM-Inspired Graphene-Based THz MIMO Antenna Configurations Using Characteristic Mode Analysis for 6G/IoT Applications
Previous Article in Special Issue
COVIDSensing: Social Sensing Strategy for the Management of the COVID-19 Crisis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Sound Feedback for Social Distance: The Case for Public Interventions during a Pandemic

by
William Primett
1,*,
Hugo Plácido Da Silva
2 and
Hugo Gamboa
1
1
LIBPhys Faculdade de Ciências e Tecnologia, Universidade Nova de Lisboa, 2829-516 Caparica, Portugal
2
Instituto de Telecomunicações (IT), 1049-001 Lisbon, Portugal
*
Author to whom correspondence should be addressed.
Electronics 2022, 11(14), 2151; https://doi.org/10.3390/electronics11142151
Submission received: 19 May 2022 / Revised: 28 June 2022 / Accepted: 29 June 2022 / Published: 9 July 2022

Abstract

:
Within the field of movement sensing and sound interaction research, multi-user systems have gradually gained interest as a means to facilitate an expressive non-verbal dialogue. When tied with studies grounded in psychology and choreographic theory, we consider the qualities of interaction that foster an elevated sense of social connectedness, non-contingent to occupying one’s personal space. Upon reflection of the newly adopted social distancing concept, we orchestrate a technological intervention, starting with interpersonal distance and sound at the core of interaction. Materialised as a set of sensory face-masks, a novel wearable system was developed and tested in the context of a live public performance from which we obtain the user’s individual perspectives and correlate this with patterns identified in the recorded data. We identify and discuss traits of the user’s behaviour that were accredited to the system’s influence and construct four fundamental design considerations for physically distanced sound interaction. The study concludes with essential technical reflections, accompanied by an adaptation for a pervasive sensory intervention that is finally deployed in an open public space.

Graphical Abstract

1. Introduction

The Coronavirus Disease 2019 (COVID-19) was affirmed as a pandemic by the World Health Organization (WHO) on 11 March 2020 [1] presenting numerous unforeseen novelties given that symptoms would range from severe to unrecognisable over an indeterminate timeline [2]. As the effects of the pandemic intensified on a global scale, central governments enforced new regulations that compromise public space interactions against the preceding normality. While the specific constraints varied between countries and eased accordingly over time, a key component throughout was to prohibit physical contact almost entirely where viable and to avoid close proximity with others [3,4]. Simultaneously, the sudden ubiquity of protective face coverings has introduced a plethora of unanticipated complications regarding verbal and non-verbal expressions that are habitually relied upon to interpret emotional states [5,6,7].
In accordance with these measures, a vast majority of public events were jeopardised, particularly in the case of live performances and artistic installations. As pandemic regulations were gradually alleviated, however, public activities were re-authorised on the grounds that all participants complied with mask-wearing and distancing orders [8]. We foresaw a need to address the tensions surrounding safe social conduct as well as the risks of viral transmission newly associated with on-body sensor technologies [9,10]. The physical space observed between bodies can be linked to various social cues during an interaction [11,12,13]. However, these inferences can deviate massively between individuals and therefore it is unintuitive to commit to a one-size-fits-all model that is wholly representative of all sorts of social contexts and cultural environments [14,15].
We propose an intervention to challenge the conventional dualisms between physical distancing and social connectedness, based on sound and movement. This begins by reconstructing our approach to both sensor-based monitoring and the delivery of live performance, looking beyond the introspective experience, and considering affective processes that occur externally from the body. We then present a wearable system for sound–movement interaction centralised around interpersonal distance and collective movement. Following a series of ideation sessions, over a two-week period, the system was situated in a live performance setting from which we evaluate the performer’s quality of movement in relation to the sound feedback. We describe some of the system’s key influences contextualised in the socio-emotional domain, inducing entrainment, spontaneity and awareness. We formulate the following research questions: RQ1 How can sound feedback improve collective awareness and sensibility to interpersonal distance? RQ2 What movement patterns are influenced by external mediation materials? and RQ3 What technological interventions are appropriate for proxemic interaction in public space?
Our work continues to extend the rich design space for collective sound environments with movement sensing. We outline our perspectives on mediating social interactions with regard to the proposed coupling of distance sensing with sound actuation. Following this, we acknowledge the vital limitations of our initial framework and put forward research opportunities that may arise from extended exploration. Namely, the incentive of transitioning into open public environments where participation can occur pervasively. Additionally, we document the necessary measures taken with respect to the restrictions put into force by the COVID-19 pandemic, the impact of which remains prevalent up until the time of writing. Howbeit, our findings should be interpreted as universally relevant toward orchestrating spatially aware interventions even in non-pandemic circumstances.

2. Background

2.1. Social Distancing and Public Presence

It is evident that regular social engagement plays an important role in community wellbeing. Not only bonding with our close companions in organised situations but inclusive of unanticipated interactions that emerge in public [16,17,18]. The first wave of the COVID-19 pandemic elevated some attention to this. To exist outside of our own home meant weighing out the chances of a life-threatening infection against the consequences of avoiding social connection, endangering one’s health in other ways [19,20,21,22].
As emergency COVID-19 measures were becoming standardised globally, epidemiologists from the WHO firmly recommended a linguistic shift from ’social’ to ’physical’ distancing on the basis that modern technology is capable of keeping us connected in spite of the new regulations [23]. Given that the anticipation of the first wave would in essence reject in-person contact in the way it was known; this narrative was therefore appointed to the normalisation of fully remote communication. However, as society would phase in and out of non-confinement periods, conventional face-to-face interaction could take place again, persisting with the caution of physical distance, embracing Social Proximity as a safe practice for social wellbeing [24]. Irrespective of this major adaptation to safer re-socialisation, we argue that the development of in-person interventions has not matured to the same extent as remote interaction technologies and that there still lies a vast design space yet to be fulfilled.

2.2. Proxemic Behaviour and Re-Socialisation

The attention toward interpersonal distance coincides with the study of proxemics, examining the function of physical space during face-to-face interactions [25,26]. Proxemic theory has been given a lot of attention in a wide range of behavioural studies [27] and thus spurs incentives from several ubiquitous computing projects [28,29]. A majority of these accept the proxemic zones set out by Hall [25], by which interpersonal distances are generally categorised into the following boundaries: Intimate, up to 1.5 feet (0.45 m); Personal, 1.5 to 4 feet (1.2 m); Social, 4 to 12 feet (3.6 m); and Public, more than 12 feet (7.6 m). We believe, however, that the rich contextual nature of interpersonal movement behaviour poses a challenge for conventional computational modelling, calling for expertise to break down and articulate expressive features of human motion and sensation [30,31]. With regard to proxemics, we cannot solely depend upon the measure of distances and angles to define the affective characteristics that occur during any given exchange.
Coping with the current risks of infection in everyday social contact, the new sociable space implies a new urban etiquette that expands upon the standardised dimensions proposed by Hall’s proxemic theory, preserving conversational affordances at double the distance [32]. In a general sense, spontaneous, informal encounters are recognised as an essential part of growing one’s social circle, fostering a sense of belongingness within the community [33]. However, the circumstances in which these are likely to occur are vulnerable to the spatial dynamics and physical distance between bodies [34,35]. These types of relationships fit into the social formation of indirect contact, which can take place seamlessly by way of intergroup behaviour, alleviating the pressures of face-to-face enactment [36] that are increasingly present in the midst of pandemic concerns [37]. While, of course, many will long to reconnect with their peers during moments of close physical bonding, these new sociable spaces provide an enlightening deconstruction of proxemic acceptability where strangers and non-strangers are both welcomed into everyday social encounters. The linguistic connotation of ’distancing’ assumes active separation from others, while its function should instead be geared towards social affirmation, organised in a safe manner. We propose an interactional view on social distancing that is unconditional to definitive measurements, advancing upon the outlooks contained in the following study [38] that insists on expanding Hall’s discrete interpretation of proxemic zones when applied to continuous mappings of movement.
The orchestration presented in this case study is set out to capture the affective outcomes when proxemic behaviours are exaggerated during social exchange, where sensory intervention goes beyond being an assessment tool, but becomes a modality for non-verbal expression.

2.3. Sensor Technology and the Right to the City

When considering proxemic behaviours in public, urban pedestrian areas have long exemplified the effectiveness of street-installed technology to manage the mobility of crowds; however, the vital function is vulnerable to being overthrown in the case of high-intensity overcrowding [39,40], which has been associated with feelings of anxiety, frustration and claustrophobia [41]. Certain demographics are more sensitive, while others are far less cautious of invading the personal space of others [42,43]. With respect to interpersonal distance, initiatives in response to pandemic conditions rely upon a level of altruistic responsibility, by which the public are inclined to follow a protective etiquette [44]. When taking a look at the technologies that have become ubiquitous in response to the pandemic: Contact Tracing Applications, Skin Temperature Scanners, and contactless patient monitoring, for example [45,46], we note that the user is constrained to one interpretation of the data made available to them. When contrasted with alternative interventions that explore emotional bonding in urban environments, e.g., [47,48], we begin to question the limitations of the prior, and subsequently, practices that exist at the intersection of social affirmation and long-lasting public health. Building upon the insights presented by Howell [49], calling for a progressive turn in public space sensing technologies [49], we insist that the data-driven motives pushed onto current smart city development schemes are non-compliant with these sorts of shared emotional experiences [50,51,52].
The long-standing notion known as “the right to the city” has been embraced by Interaction Design researchers, commentating on the industrialised approach that we are seeing with sensor technology that is being increasingly embedded into urban spaces, shifting the perspective from data-driven smart cities to ’social’ or ’playful cities’ [49,53]. Such case studies express a necessity for inclusive participation to establish grounds for progressive social integration, framed as the collective right to be in control of the surroundings through co-creation. Acting upon the current situation, we enquire into a sensory intervention that aids awareness of one’s physical presence and self-affirmed boundaries, without the need to declare any action as right or wrong.

2.4. Appropriation of Social Distance and Interpersonal Touch

The urgent nature of the pandemic demanded disruption to common interactional norms, namely the concern of keeping distance and avoiding touch [24,54]. However, despite a reactionary appeal for spatially-aware behaviours, this call for action has been relevant for a while now. There exist several rationals for designing systems that are considerate of one’s personal space, which in our view, have been neglected by relevant research areas.
For many, the apprehension of close contact in urban space serves as an evaluation criterion for everyday safety [42,55]. A preference for greater interpersonal distances is also evident for those suffering from anxiety disorders, commonly assumed to an avoidance of social interactions entirely [56]. Intersecting this view with the subjective quality of skin-to-skin contact, praised by a substantial volume of research for inciting profound therapeutic sensations [57], presented as an appealing method to relieve anxiety, strengthen social bonds, and even elicit physical health benefits [58,59,60]. However, this is not always the case if we take into account, for instance, those who experience hypersensitivity (or lack of) to introspective stimuli [61,62]. Furthermore, the perceived benefits associated with affective touch are largely subject to the qualities of the pre-existing relationship of the dyad [63].
Drawing parallels between inclusive proxemic interventions and safe re-socialisation [24], we want to understand the liberties that can emerge from contact-less mediation while protecting the right to physical presence. Therefore, it is important to depart from the assumption that everybody is entirely comfortable with close contact, and understand that the degree of comfort largely depends upon who is approaching them and in what context [64,65]. In Section 5, we discuss the importance of flexible boundaries when designing for interpersonal distance, conditional to the social context at hand and the proxemic sensitives of the individual.

2.5. Non-Verbal Contingencies and Face Coverings

Proxemic interactions, given the alliance with mutual gaze, would assume the inclusion of facial expressions, where conventionally, more salient emotional attributes are depicted between the nose and the chin [66,67]. The visibility of facial expressions and clarity of speech is both highly relied upon in everyday communication, but the regulated use of the face mask disregards this modality almost entirely [6]. This becomes even more crucial in the context of physical distancing regimes [68], where proxemic studies report a habitual reduction of eye contact when spaced more than twelve feet (3.6 m) apart, i.e., public distance [69]. Additionally, studies related to pandemic behaviour have demonstrated how face masks influence interpersonal distances [70]. Our facial muscles can expose a great deal of how we are feeling [71,72], sometimes completely unknowingly to the extent that one may exert themselves into forcibly concealing these expressions while under pressure [73]. In comparison, how we move in space is normally a consequence of deliberate coordination during an interaction [74,75]. Hadley and Ward [76] point close attention to the physical gestures used as a substitute for verbal exchange where background noises impair vocal comprehension [76]. The normalisation of face masks worn during conversation has been shown to degrade the acoustic quality of the voice as a result of suppressing higher frequency ranges, commonly depended upon to recognise articulations of consonant sounds [77,78,79]. This poses a further disadvantage to those hard of hearing [80] as well as non-native speakers, often more dependent on reading the face [81,82].
In Section 3.3, we outline our fabrication methods, taking the newly ubiquitous face mask, commonly condemned as a social hindrance and reshaping this as sensory material for non-verbal exchange, isolating communication channels aside from the face.

2.6. Measuring Interpersonal Movement

Aside from interpersonal distance, we are also interested in movement qualities that can characterise aspects of a social situation. Such qualities are recognised by Rudolf Laban’s Movement Analysis (LMA), a well-established notation framework used to depict expressive features of human movement, originating from the perspective of dance and physical therapy [83], and since adapted to all kinds of contexts (e.g., routines of factory workers) [84]. With ongoing advancements toward body-centred applications, Laban’s theory on movement has firmly settled itself into Human–Computer Interaction research [85].
It is apparent here that the intersections of HCI and Laban Movement principles tend to focus on individual accounts. Pluralist qualities, on the other hand, comprised of two or more persons moving together, are observed as the Relationship category. The following article from Roudposhti et al. affirms a scarcity of literature in this domain [86], expressing the unexplored potential for social interactive systems. The authors propose a global feature space that combines Pentland’s analysis model of social signals [87] with LMA qualities, taking upon the following descriptors: Indicator, Empathy, Interest and Emphasis. In Section 5, we borrow two of these qualities in our evaluation, Indicator to describe the exchange between influent and influenced members, presumed by the difference in energy between them, and Interest representing one’s engagement to the situation or outside context, gauged by energetic movements. Similarly to Laban’s Movement Analysis, each quality operates on a continuous scale between two polarities.

2.7. Sound Interaction as Social Mediation

Throughout the extensive literature surrounding proxemic interfaces, we came across a surprising lack of studies related to sound, given this is already an established modality for movement interaction with ties to affective representation [88], supported by a base understanding that physical action and sound perception are mutually responsive [89]. As a tentative presumption, we can point to the inherent limitations of distance detection with typical camera-based tracking technologies [90] as well as the obtrusiveness of on-body sensor devices [91]. An affirming study that fits into this criteria installs a proxemic augmentation into a gallery space [92], supporting the role of proxemic audio interactions in a “post-screen world” [93]. A recent study comparing common actuation modalities by Alfaras et al. demonstrates the usefulness of audio-based biofeedback to foster physiological synchronisation and somatic awareness, noting that the human hearing system is highly sensitive and that sound is a convenient medium to share amongst many users [94]. Furthermore, recent literature suggests that synchronous motor activity can be indicative of prosocial affiliation [76], particularly when contextualised with sound, be it disruptive or complimentary [95].
Within the context of music performance, the concept of a collaborative system is not a novel phenomenon, far from it in fact. Reports date back as far as 1978 [96] with progressions to real-time remote interaction [97], later supporting the establishment and maturity of interactive music systems (IMSs), e.g., [98,99]. A recent review from Aly et al. discusses the pervasive nature of sound interactions in the context of biosignal-driven IMSs when capturing data from many users simultaneously [100]. Giving attention to collective movement, Hege presents an artefact by which the members of the Princeton Laptop Orchestra showcase democratic expertise through intentional yet delicate control of the sound output [101]. IRCAM researchers advocate for a human-centred framework for gestural sound control in a group scenario [102], not only for performance but also justified in clinical use-cases [103].
Moving away from the audience–performer dynamic, related case studies demonstrate how social encounters can be mediated through embodied sensor data [49,104,105]. We would like to further investigate this approach to engage multiple users simultaneously, in this instance, representing an assembly of interpersonal distances through sound.

3. Materials and Methods

3.1. Composition of a Wearable Sensing Medium

We decided to design a wearable sensing medium centred around the face mask. These were a mandatory possession for local citizens that had already become a cultural norm for one’s public appearance [106]. Our pursuit towards an on-body device was supported by Montanari et al.’s investigation into a novel proxemic sensor, welcoming a sacrifice in the high-level attributes that come with camera-based tracking in preference of environmental flexibility, quoting the affordance “to collect data even in areas that cannot be instrumented, like public spaces or during large events” [107]. We consider the worthwhile benefits of a wearable solution that is more versatile in non-laboratory situations [95,108,109]. In essence, this bypasses the challenges articulated by Jürgens et al. directed towards using an unobtrusive markerless motion capture technology in an on-stage environment for contemporary dance performance [110]. The authors highlight digital errors inflicted by particular lighting conditions, clothing contrast, as well as scenarios where performers were in close proximity to each other. For our study, we were incentivised to capture the point of view of the user, aligning the sensing trajectory with the user’s gaze during an interaction, as detailed in the following sections.

3.2. Overview of Components

We have interfaced low-cost HC-SR04 ultrasonic distance sensors with the BITalino R-IoT microcontroller using a modified firmware (included in Appendix A); this allows the acquisition of proxemic data from multiple participants at 10 samples per second (s). When initiated, each module streams data wirelessly over a designated local Wi-Fi network to a host computer, which is then responsible for signal processing and sonification. With this sampling rate, the maximum response time is accepted as 100 milliseconds (ms) plus any wireless latency, averaging at 10–11 ms in such conditions (report provided in Appendix B). This sensing technology is highly prevalent in robotics and IoT educational fields, typically used as part of introductory curricula [111,112] while sustaining relevance in the state of the art (e.g., [113,114]). Fundamentally, the sensor measures the distance from the first physical interference at a given direction by transmitting and receiving ultrasonic frequencies outside of the human hearing range [115]. The range and accuracy benchmarks for the ultrasonic sensors are partially dependent on environmental variables such as ambient temperature and humidity. When working in typical indoor conditions, the ultrasonic sensors are expected to ensure a stable accuracy of up to 13 feet (4 m), within a 15 degree angle [116].
With respect to providing a wearable form factor, and minimising the size and weight of the components, we implemented a discrete voltage divider onto the sensor’s Ground and Echo pins to comply with the power specifications commonly found in smaller microcontrollers, usually rated at 3.3 Volt. With this configuration, there was a minor but notable drop in accuracy compared to our tests using 5-Volt compatible microcontrollers. However, this was mostly resolved with signal processing to the extent that the effects would not hinder the overall sound-based mediation experience. Following technical directions from Kielas-Jensen [117], we were able to incorporate data from the R-IoTs embedded temperature sensor into the distance acquisition function, this improved the reliability and consistency of the readings when transitioning between distinct environmental conditions. For example, a computer lab, open exhibition space, performance theatre holding maximum occupancy, or even installed in an open outdoor area.

3.3. Material Design and Fabrication

In order to publicly distribute the sensory masks in a safe manner, respecting the government guidelines, we assigned the following design principles: (1) The sensory components are modular and detachable, allowing for sanitation while the mask fabric is being replaced; (2) These components must remain at a fixed position and be robust in situations of rapid movement; and (3) The mask should feel sufficiently comfortable for users to wear for prolonged periods of time. Additionally, for the system to accommodate mass participation, we favoured a scalable solution that was low-cost and easily reproducible (4). For each wearable, we modelled two separate housing elements for the respective components, i.e., the microcontroller and the proximity sensor. The 3D models are included in Appendix A.
For the sensing mechanism, the ultrasonic sensor casing was merged with an arched nose clip and secured onto the mask fabric. The microcontroller and battery were encased in a custom mask strap that would be tied around the back of the head. The components are connected via a flexible cable, splitting into four wires each connecting to the corresponding inputs of the sensor and microcontroller, transmitting analogue signals back-and-forth to retrieve proximity data, represented as trigger and echo in Figure 1. These are assembled by positioning the sensor component above the nasal dorsal, resting front-most of the face, in line with the gaze with minimal obstruction. To overcome the precariousness of the sensors dropping down from the nose during movement, we increased the tension of the elastic ear straps and fastened a flexible rod between the mask fabric and sensor attachment. While this modification improved the structural reliability, our capacity to freely substitute sensory modules, along with spontaneous user alterations was limited as a consequence.

3.4. Data Processing and Sound Mapping Strategies

The orchestration does not require participants to conform to a fixed minimum distance. We lend our trust to the users to coordinate themselves in a safe manner without explicit orders, as it is expected in their daily life. Prompted by RQ1, the sound interaction is purposed to strengthen one’s awareness of the other’s presence, unbounded from discrete categorisation, engaging one’s auditory senses while verbal modalities are restricted. In addition to this, the experience should not call for virtuosity or a specialist training process. The system assumes engagement from the moment the user is being sensed and become progressively accustomed to the sound–movement affordances. This aligns with the presumption that the participants are expected to work together to develop their understanding of the system and stay vigilant to each other’s actions.
During these first trials, we used granular synthesis (Grain Scanner by Amazing Noises: https://www.ableton.com/en/packs/grreciprocateain-scanner-amazing-noises/ (acessed on 6 July 2022) to interpolate between 4 field recordings as base sound textures. Inspired by Laban’s qualities of space and relationship, we proposed a transition between “open” to “closed” sound textures, continuously modulated in real-time as the average distance reaches the minimum sensing boundaries of 4 feet (1.2 m), before succeeding social distance.
This process starts with a lot of fluctuations in timbre and then closes in on a limited range of frequencies that form into a staccato rhythmic feel. A one minute sound sample is visualised in Figure 2.

3.5. Performance Structure

The system’s initial deployment exposed a significant level of confusion from new users when given the chance to freely experiment with the device. Aside from the technical ideation, we speculated upon ways of stimulating movement whilst sustaining an appropriate degree of improvisational freedom. This reiterates that we were not interested in forcing a strict sequence of movement, but rather providing geometrically informed cues set out to prompt dyadic gestures. We devised three scenes that each focus on distinct affordances of proxemic mediation, as illustrated in Figure 3 and presented in the following order:
The group was made aware of the performance structure and interactive elements described in Section 3.2, while granted the freedom to explore the space and move intuitively to the sound feedback, whilst maintaining a minimum interpersonal distance of 5 feet (1.5 m). For these actions, a 225 m2 performance area was used accompanied by two concert-grade loudspeakers.

3.5.1. Performers and Objects

These geometric arrangements are predominantly inspired by William Forsythe’s architectural approach to choreographic environments [118], a radical staple in contemporary dance culture, bridging deeply into other artistic mediums [119].
Giving attention to RQ2, the following passage redirects the proxemic attention from the neighbouring bodies, onto the surroundings (Figure 3(ii)), supported by Kinns’s recommendations for mutual engagement with shared representations [120]. In the extended proxemic criteria presented by [38], the relationship to other objects, both digital and non-digital, is embraced. Adopting theories of gaze-based interaction into the format of proxemic awareness, authors claim to enrich the user’s attention to the space by attending to implicit responses from the user’s surroundings.
Autonomous flying drones have been incorporated into a stage performance [121] and movement-centred practices [122] proceeding efforts to enhance kinesthetic awareness through intercorporeal engagement [123]. During Scene ii, the drone represents more generally an unpredictable external influence, comparable to the external nuances that occur in public space environments.
In relation to this, Ballendat’s evaluation of screen-based proxemic interactions considers the attention directed to other people, allowing for conventional social exchange to coexist alongside the technological artefact itself, accepting natural influences that would arise in coherence with an everyday social situation [38].

Scene I, Geometric Boundaries and Interceptions

To begin, the performers position themselves around the boundaries of the performance area, facing the centre (Figure 3(i)), standing 18 feet (5 m) apart. The first sounds are activated as users cross to the opposite side, alternating back and forth until coinciding with the linear path of one another, after which an anticipated diversion is required. One participant offers the following interpretation: “The soundscape intensifies as the collective tightens in space. Before collision, we have to figure out our next steps, to dodge, retrace or simply pause for a moment”. Our improvisational framework here is designed to inspire proxemic exchange, while the primary responsibility of avoiding near-contact is handed over to the performers.

Scene II, Interacting with the Non-Human

A technician is assigned the role of navigating the stage with a flying drone, weaving between the masked performers. When advancing towards the bodies on stage, specifically targeting the sensory components, the drone serves as an external trigger as it reaches close enough to the face. As a result, this synthetic artefact would provoke instantaneous reflex responses from the performers that consequentially would instigate dynamic changes in the soundscape.

Scene III, Participation from External Users

Finally, we invite external users that are not individually equipped with any wearable sensors but are authorised to manipulate the behaviour of the mask-wearing user group. Nine additional participants were instructed to approach the mask-wearing group, directing them to move back until reaching the end of the stage. The two collectives both march between the right and left extremities of the stage, constantly maintaining a mutual gaze and a forward distance of approximately 6 to 10 feet (1.8 to 3.0 m) from the closest person opposite (Figure 3(iii)).

3.6. Study Outline

Our study derives from the experimental framework of Performance-Led Research in the Wild [124]. This compromise relieves some of the social confines imposed in a lab setting while bypassing the highly unpredictable nature that comes with arbitrary participation [125]. In this current research action, we set up an open call, as a result of which 12 local artists were chosen to take part in a 2-week residency that would conclude with a public performance. This was structured to build upon compositional practices that were stimulated by a series of performance and improvisation workshops that took place 3 weeks prior. From the residency group, 4 participants were selected to use the wearable during the final performance, presenting the three scenes described above, which lasted just over 8 minutes in total.
We recorded video and sensor data throughout each scene in order to evaluate the qualities of interaction. From this sub-group, 3 identified as male and 1 female with ages between 24 and 51 years old. All 4 held long-term experience as highly-skilled musicians, performing internationally and retaining regular musical practice. An additional 4 participants, 50:50 split male and female, trialled primitive iterations of the system during the prior rehearsal stages, from which we produced field notes and recorded user feedback. Aside from a handful of local encounters, the user group was considered strangers amongst one another, not closely bonded and only acquainted during the preparation stages.
The first study took place in Lisbon, Portugal between October and December 2020, before the national vaccination rollouts were initiated. Cultural activities, such as live performances, were permitted in accordance with the measures set out by local authorities [8]. In the scope of COVID-19 distancing measures, the user groups were respected as a support bubble throughout the research period, assuring that in-person interaction was permitted under the same conditions as those they share a home [126]. With that said, at no point would the sessions explicitly require a disruption to the standardised safety regimes, particularly those concerning close contact and sanitation of shared materials to minimise viral transmission.

4. Preliminary Results

Here, we summarise the perspectives of the individual participants and indicate moments where the first-person accounts are complemented by the acceleration data. We apply statistical analysis to the collective sensor pool to discern group-level features, later used to substantiate relevant design guidelines for proxemic intervention strategies. We acknowledge the analytic restraints when subjected to a single recording from an individual user group that by no means should be taken as a comprehensive survey of the proxemic orchestration. This provisional study is purposed to demonstrate novel adaptations for proxemic interaction and prescribe design insights for future work.

4.1. Data Collection

Throughout a 2-week prototyping period, we recorded participant observations and first-person experiences, based upon the user-centred design actions proposed by Bernardo et al. [127]. This was comprised of eight day-long rehearsal sessions that each allowed one hour focused on exploring the wearable followed by a brief interview with participants, asking about any limitations that were discovered and features of the interaction that influenced their behaviour. These sessions took place alongside a course of improvisation exercises fitted towards musical performance. After the study, we extract comments from a series of recordings and field notes. Participants were made aware that they were taking part in a study set out to evaluate the general usability of the wearable device and sound feedback mapping.
We recorded video and audio of the final performance and rehearsals using a single-point microphone and camera. The video clips were colour graded to obtain a clearer view, especially when there was a lack of ambient light being projected onto the stage. When probing deeper into a handful of significant events depicted by our written assessment, we extract still frames from the recordings and overlay these with graphical annotations to comment on the positional qualities present. Audio was processed through Sonic Visualiser (Sonic Visualiser: https://www.sonicvisualiser.org/doc/reference/4.4/en/ (accessed on 25 June 2022) for feature extraction. The sensory masks were embedded with an accelerometer sensing component located at the back of the head, recording three axes of directional acceleration from each user. The triaxial sensor data are unified by computing the Signal Vector Magnitude (SVM), representative of the combined acceleration coordinates x, y and z, as validated by Ward et al.’s proposed method for monitoring synchronous motor behaviour within a group of live performers [128]. Each accelerometer was recorded at 10 samples per second, with the data being synchronised and smoothed in post-processing.
When probing deeper into a handful of significant events depicted by our written assessment, we extract still frames from the recordings and overlay these with graphical annotations used to commentate on the positional qualities present. In Figure 4, we present the median average magnitude acceleration measured from the four users to represent the collective movement, indicated by a series of peaks throughout the three scenes. The median acceleration timeline (top) serves as a guide to navigate the general progression of movement qualities presented throughout the performance, while separating the data stream for each individual user (middle), and assessing the alignment of the peaks, provides a numerical evaluation of interpersonal responsiveness or absence of. Finally, the peak acceleration points are partitioned according to a series of successive bursts over time (bottom), detected using a k-means clustering algorithm. The clusters help us to convey the collective movement dynamics and relationship qualities that are observed throughout each scene, partly revealed in the average duration and amplitude of the clusters.
We formulate the following clustering features: Peak Density, the total number of peaks detected relative to the duration of the cluster, and User Effort Dissimilarity, based on an equal alteration of active users.
Let C ( x ) = i = 1 n [ s i = x ]
User Effort Dissimilarity = i = 1 n ( s i C ( x ) 1 n ) · 100 .
The User Effort Dissimilarity is calculated according to the proportion of peaks exerted by each individual user, where C(x) represents the total number of acceleration peaks, si being those exerted by each individual user.

4.2. Interpersonal Synchrony

During Scene iii, we observed participants engaging more confidently before and after the sound mapping is disrupted by the unmasked participants. In these moments, the mask-wearing group march in parallel alignment, maintaining a consistent pace with each other. Additionally, we note the persistent use of eye contact towards the unmasked group, voiding obstruction until reaching the end of the stage. As the procedural soundscape is overridden by the choral chants voiced by the unmasked performers, we see the groups disperse in the opposite direction, subverting the momentum and common alignment. This interplay repeats itself 4 times, where the two sub-groups voluntarily hand over the leading role of the march, alternating every 6–10 s. In Figure 5, four masked performers walk towards and away from a group of those without wearable sensors whilst maintaining a forward distance. For these frames, the annotator was tasked to visually mark the alignment and walking direction of the mask-wearing group.
We compare the synchronous activity of each scene based on the separation of the clustered peaks. From this measure, we determine that the most consistent levels of synchronous movement take place during Scene iii, with far fewer occurrences during Scene ii. Both scenes have a similar number of peaks per cluster (Table 1), with a greater density of peaks in Scene iii; this is shown by the constrained duration of the clusters formulated in more consistent intervals. This perception is reassured by the increased amplitude of acceleration values throughout, perceived as an intentional syncopation of movements that are highly responsive to one another, and agreeable to the feedback given by one of the users.
we had only a short moment to respond, similar to a call-and-response situation you sometimes see in concerts. The motivation was no longer about controlling the timbre (of the sound), but just to assert dramatic impulses before the others take the lead.
Our observations and data taken from this final arrangement indicate vastly different group behaviours from before. We would suggest that the major influence lies behind the “call-and-response” routines as the opposing bodies assert themselves into the space, inducing a series of structured interruptions that we probe deeper into in the Discussion, Section 5.2. We will not neglect here that these behaviours were inherently dramatised under theatrical persuasion, but, nonetheless, insist that the core reactions are authentic to the invasive confrontation as the opposing bodies assert themselves into the space.

4.3. Spontaneous and Sustained Engagement

Given the premise that regular social engagement is an effective precursor to community wellbeing, generating and sustaining interactions are both considered to play a key role in urban design practices [129,130,131]. We question how sensor-based mediation can be used to not only provoke, but actually extend the longevity and quality of a new encounter, described as “what makes the experience comfortable, interesting, and meaningful” [131]. In this case study, we gauge spontaneous engagement from the rotation of users asserting themselves as a leading influence (i.e., a new person exerting an acceleration peak), and how peaks are partitioned between individual users. Sustained interactions are more nuanced, recognised as lingering behaviours that take place amidst the succession of new movements, for which we draw upon the user’s control over the sound output.
Responding to our research question RQ2, we enquire into the user’s relationship with the external mediation materials, and the way these momentarily interfere with the continuity of the sound output in Scenes ii and iii, separating these external influences by their level of predictability. In Figure 6, we recognise a major disparity in the arrangement of the non-structured interruptions incited by the flying drone compared to those anticipated by the unmasked group (i.e., structured interruptions). The unpredictable manoeuvres imposed by the flying drone disrupt the course of sustained dyadic gestures, from which we dissect the rapid, unplanned interchange of users controlling the sound output. Referring back to Figure 4, this parallels with sporadic changes in peak intervals, with individual exertions coming at arbitrary intervals. Scene iii, on the other hand, the course of sonic deviations better emphasises a uniform sequence of interruptions. This repeatable turn in controllability infers an attentive reciprocation with the other users, showing an advancement from sporadic reflex actions to meaningful group gestures, also granted by steadier peak intervals and the overall progression of synchronous movement that is detailed in the subsection above. One participant describes an impulsive reaction while being approached by the flying drone, suggesting its non-human form carries a provocation for spontaneous engagement:
the drone would keep coming closer, like I was being chosen out of the crowd, so I would start following. I was comfortable with running towards the wall or even a flying object, but into somebody else, it’s not the same. Even if we do not get so close, I feel like I am threatening someone or just being a nuisance.
This correspondence is not so apparent in the acceleration data, where we actually notice a drop in energy regarding the lower mean cluster density and substantial User Effort Dissimilarity, supposing an imbalance of individuals dominating the space. However, we may testify that the shared positional influence of the drone on stage engaged all users in the shared space, even when at a standstill.

4.4. Technical Venerability against Proxemic Awareness

In hindsight, the placement of the sensors of the face, constantly shifting their projected angles meant that misreadings were highly probable. When tested in lab conditions, these proximity sensors are expected to perform reliably at differing trajectories [116], but when we inspect the sound recording and video footage, we observed many cases of unexpected readings during the performances in contrast to the testing phase. Participants would move closer together with no perceivable feedback. Moreover, the detectability of the ultrasonic reflections was highly dependent on the material of the occlusion. Table 2 provides a set of sample signals recorded when approaching the sensor using three different surfaces. These were recorded separately in controlled conditions. The ceramic tile provided the greatest detection range, while the clothing material resulted in more noise and vulnerability to drop-outs. Though our signal filtering methods helped remove extreme anomalies, there would still be a great deal of inaccurate measurements being fed into the system. For these reasons and more, this particular sensor technology has been discouraged for measuring interpersonal distance in a review of wearable devices for proxemic interaction, which exposes multiple scenarios by which ultrasonic sensing was proven to be partly inadequate [107].
Fdili Alaou writes about the perceived messiness that inevitably comes with adopting personal tracking devices into performance-based practice (noise, sensor placement, classification failure) [132]. However, instead of seeing these as issues that need to be resolved, actually encourages artists to embrace the technical nuances, turning technology resistance into creative material. It is important to appreciate that, in non-lab conditions, imperfections will constantly prevail and, therefore, we should welcome and validate the experiences that come with each iteration. In spite of such cases where the sound–movement relationship was not sensible to the performer or even the audience, as revealed repeatedly during the user studies, we contend that there remained a genuine influence from the physical artefact alone. Throughout the study, users demonstrated a strong awareness of their surroundings to control the sounds, recognise technical faults and overcome them through persistent trial and error.

5. Discussion: Preparing for Public Instalment

5.1. New Sociable Space

Modern communication technology has been shown to facilitate rich social engagements in a remote setting, by which some degree of face-to-face affairs continue to be viewed redundant, even after confinement measures have subsided [133]. We, therefore, propose that solutions should also exist to support meaningful discourse from an extendable distance that is suitable for the social environment and individuals involved. Taking sound feedback as the core of proxemic mediation, we adopt Mehta’s new sociable space concept into our design considerations, this being the capability to capture and hold one’s attention from a comfortable distance, encouraging spontaneous encounters with non-acquaintances before confrontation into personal space [32]. Inviting flexibility to traditional proxemic theory, we can call upon Roudposhti’s behavioural model that was introduced in Section 2.6, demonstrating mannerisms of Indicator and Interest.
While this desirable social dynamic is provisional to a wide spread of design implications, we put forth the benefit of authorising drop-in and drop-out participation, whereby the formation of a group can be altered at any point to openly accommodate new users. Additionally, this functionality insists upon an agreed focal point of interest that establishes a proxemic cornerstone regardless of the group’s composition, continually subject to change. We find this inclusive control structure to be very much customary in the context of interactive installations, such that the initiation of the sound feedback invites new unsuspecting members to engage before even being aware of the artefact’s existence [134,135]. In our performance-led study, we were enlightened through the inclusion of foreign mediation artefacts, acting as a vital component for exposing new dynamics of the sound, shown less prominently when the mask-wearing group was isolated from the external surroundings in Scene i.
This draws upon the gestural limitations while adhering to a minimum interpersonal distance of 4 feet (1.2 m), absent from any positional cues to prompt collective engagement.

5.2. Proxemic Sensibility and Sensitivity

Our literature review in Section 2 supports the vital influence that routine social interactions have on public spaces and the proceeding benefits that come with this. We also bear in mind the negative view that undesired social isolation can be detrimental to one’s mental condition [21,22,136]. The pandemic forced prolonged periods of isolation that would ultimately cause a rise in self-reported loneliness [137], reported to be particularly harmful for those who already experienced anxiety prior to the pandemic period [19,138]. Such conditions have been shown to suppress one’s tolerance for engagement within intimate proxemic boundaries [139]. We can also include recent findings related to isolation and cognitive function, locating harmful effects on the brain region associated with spatial orientation, learning and memory [20,140], which presumably foreshadows a long-term disassociation with in-person social situations. A critical motivation behind our work is to examine what interventions open up a safe intermediate to re-socialisation for individuals who do not yet feel comfortable exposing themselves in public [141], understanding that each person will hold their own preference for personal space.
Moving away from a standardised proxemic model, we form an empathetic view around personal space, appreciative of boundaries that are unfixed and individualistic in light of one’s past experiences and various other factors that are undisclosed between strangers. Reinforcing Pentland’s descriptors assigned previously, this derives from the sentiments that embody Empathy and Interest. Addressing RQ1, we articulate the call for proactive awareness as part of the following design considerations with regard to sensory intervention, Proxemic Sensibility and Sensitivity. Sensibility is the altruistic responsibility that ensures safe coordination of bodies, mindful of the surroundings and presence of any individual, paired with Sensitivity, for those to stay receptive to the actions projected by others, with the willingness to alter their paths accordingly. We still cannot be certain of the system’s influence without any sound feedback since the study does not include a specific control condition for this. However, given the mixed experiences that arose in the presence of external interactive artefacts on stage, these being the flying drone and the unmasked participant group, we speculate on the suitable conditions for effective social signalling through the participatory engagement with sound. Rather than trying to incorporate all of the users simultaneously, we suggest that individualistic control mechanisms can help to elevate one’s agency to the surroundings, while the anticipation of regular interchange is necessary to preserve attention. Consequenty, we favour the use of structured interruptions by way of turn-taking procedures, whereupon users are compelled to listen to one another, then allocated a sufficient time window in order to react accordingly.

5.3. Constrained Complexity

This work fits into the domain of proxemic interaction strategies, asserting novelty in the non-categorisation of physical distances. The early phases of experimentation lead us to try complex, nonlinear sound-distance mapping strategies without the anticipation of hindered usability. However, we were enlightened early into prototyping that such sophisticated mapping strategies are only intuitive on the presumption that users are sufficiently acquainted with the apparatus and expected results, reiterating discussions around the virtuosity of musical instruments within the NIME research community [142]. We first insisted on more abstract mappings, layering sound elements each influenced by multiple users simultaneously; this approach was designed to provoke collective actions and, consequentially, for the group to act upon the movement patterns that made the most sense to them. From this composite sound feedback strategy, however, a great deal of confusion emerged from new users, leaving them with a feeling of disempowerment. We found this ambitious setup expected far too much knowledge from new users, made clear during testing, where participants would express their frustrations with the system’s behaviour while not understanding how their individual actions influence what they were hearing. At a minimum, they would want to walk closer or farther from someone and hear an instantaneous reaction. We were able to recover the user’s association to the sound feedback when resorting to a linear distance to amplitude mapping according to the smallest distance detected, allocating control to only one individual at a time in favour of being perceived as more receptive. Though, in many cases, the irregular substitutions would cause excessive shifts in volume, disrupting the continuity of the controlling gesture. In particular, when incorporating external objects and additional users onto the stage, we find more instances of abrupt sound bursts, depicted by the intermittent gaps in the audio recording, as shown in Figure 6.
While working in this semi-controlled setting, we benefited from the user’s past experiences using digital musical instruments. However, when gearing toward a public space intervention, the importance of bridging with broader audiences would become most salient, to engage those unfamiliar with the system and each other, as described in Section 6. Such a pursuit calls for an accessible means of individual control that also encourages collaboration amongst strangers. Here, we advocate for a turn-taking framework that operates on a fixed time interval. This conveniently aligns with the step sequencer metaphor proposed in Bengler and Bryan-Kinns’s work with non-musicians, studying how a constrained control paradigm can improve engagement with the general public [143]. To build upon these considerations, we prescribe the function of sequencing for distributed control and attention in group interaction, maintaining the usability of single-user mappings whilst emphasising the quality of observing the other.

5.4. Precautions for Public Inclusion

The change in pandemic circumstances that happened during the research timeline meant that only a maximum of four participants were authorised to use the wearable at any one time, dismissing any substitution of users between daily intervals. This compromised the conditions for open public inclusion while the measures gradually became stricter, bringing serious doubts in justifying any sort of artistic action that persisted in congregating different social bubbles. In these circumstances, we were confronted with the fragility of the sensory components embedded onto the mask. More often than not, assistance was required from the workshop coordinator to secure the wearable around the user’s head; this would unfavourably call for physical contact near the mouth, posing an additional risk of viral transmission. That considered, a misfortune such as this would have been far more problematic if widespread into public hands.
The deferred extremity of the situation was only realised around one month later, as these measures would come forth as a final phase of provisionary measures before the nation was required to fall under a compulsory confinement period (i.e., lockdown). On these terms, the study was committed to a public space intervention with elevated awareness of safety and robustness, particularly in regard to the exchange of wearable components between alternating user groups.
At the time of reorganising the study, this absence of open public participation was deemed somewhat a pitiful solution, albeit one that avoids abandoning the in-person field study indefinitely. Though, in hindsight, we realise these intermediate steps were absolutely necessary before the system could be made freely accessible to the public, ensuring usability and safety to a minimum standard. This mentality of turning restrictions into opportunities for precaution is highly acclaimed in the discussions drawn out from Howell et al. [49]. This work rationalises the function of private space experimentation which would merit design guidelines for public interventions, detailing concerns around safety that would omit the likelihood of welcoming fruitful interactions between strangers. Nonetheless, due to a lack of public exposure, the opportunity arises to grasp deeper insights from the user group. We factor this progression into our design considerations, insisting that new systems intended for public use should first be evaluated in a low-risk environment for a substantial period of time, taking opportunities to carry out data collection and open-ended experimentation. The controllable nature of the performance-led study allowed us to gain insights from a consenting user group, willing to follow instructions, undergo trial & error testing, and contribute to data collection. As a result, we were able to construct benchmarks for usability, later informing design adaptations for pubic inclusion.

6. The Case for Public Interventions during a Pandemic

Taking what was learned during the first phase of experimentation and preliminary results, we introduce an adaptation for a proxemic-based sound intervention designed for open urban spaces, which materialised some months later. The installation was publicly active during October 2021; at this point, the fully vaccinated population was reported at over 80% [144], and the majority of cultural actives could resume, as the most prominent pandemic regulations were already lifted [145]. This demonstration was set out to provoke collaborative engagement through sound, similar to what was observed during the closed user study, but in contrast, situated in the public space format. The system was installed for three full days in a touristic square close to the city centre, completely open to any passing members of the general public. Participants were expected to make use of the system independently, and offered only a set of essential instructions that were made accessible online through a mobile web application. Enlightened by our design considerations, we detail an alternative method for distance-to-sound interaction and assess to what extent, this intervention was capable of preserving the qualities of interaction that were preconceived in the discussion. Here, we evaluate the following strategies: drop-in and out participation, structured interruptions, and sequencing.

6.1. Instrumentation and Physical Arrangement

The public sensing environment brought many challenges to the system’s physical orchestration. The major conditions here may be subject to the installation sustaining itself outdoors, and the considerations in order for it to operate independently without facilitators. We were also intrigued to experiment more deeply into proxemic affordances shaped by the surrounding environment and non-human objects. First off, we reconsidered the mask-worn device to grant user independence, improve safety against viral transmission and refrain from dealing with inaccuracies. As an alternative, four proximity sensors were secured around the lower branches of a tree, pointing slightly downward to establish a path from the tree’s crown to a seating area set up 16 feet (4.8 m) away from the base of the tree.
Along with each sensor hangs a brightly coloured ribbon from the branch, representing the origin of the individual paths, indicating the course for users to walk under, as annotated in Figure 7. The extent of the projected sensing area was bounded by the maximum sensing distance, capped at 12 feet (3.6 m) to maintain reliability. Granted that the sensor’s performance is dependable on the ambient temperature and humidity, we noticed the detectable range varied more throughout the day than when we were working indoors. The air temperatures would fluctuate from 9–20 °C daily, with generally worsening detection rates during the night.

6.2. Spatially Distributed Sound Output

It was in our prime interest that the intervention would preserve the inherent tranquillity of the pedestrian area. To avoid any unnecessary disruption, we insisted that the installation would not exert any sound until participants were willing to engage with the system. We speculated upon a feasible solution by which the public could initiate the sensing mechanism and listen to the installation from their mobile phone. A web app was developed using the Soundworks web framework by Matuszewski et al. [146]. Upon loading the app, participants are prompted to select one of four colours each allocated to one of the sensor placements, determinant of a walking path (Figure 8). The individual sensor measurements are broadcast to each of the mobile phones connected to the current session, activating new notes when someone is detected in the space and continues to do so while the participant moves within the measurable path.
The occlusion distance from each sensor was recorded in consecutive order, cycling through the complete batch every quarter of a second. The distances are converted into MIDI notes on a single octave chromatic scale, allowing 12 possibilities spread out between 30 cm intervals, rising in pitch as the distance measured from the sensor increases. When a user is detected, the app arpeggiates through the incoming notes every cycle. As one note is released, a new note is triggered from the neighbouring user according to their detected distance. Each note is emitted from the mobile device that is assigned to the coloured sensing path, accumulating into short loops that are continuously recorded and echoed in ongoing circulation. A video extract is provided in the Supplementary Material.

6.3. From Lab to City, Transferable Experiential Qualities

Over the course of the installation period, we made notes of third-person observational accounts, similar to the rehearsal sessions and performance described prior to this. In particular, we give close attention to moments of group engagement that resonated with the spatial qualities examined previously, thus disclosing the experiences that were transferable from the wearable device to the public space adaptation. This numerical evaluation here does not intend to go as in-depth as the initial study; instead, this closing segment can be considered as a technical primer to fortify our design considerations for future research set in public space environments.
Advancing upon Proxemic Sensibility and turn-taking qualities, we found the mobile distributed sequencer fostered a certain degree of mutual agency, in that participants would invite their peers to join them and instinctively feel inclined to listen to one another. While exploring different spatial configurations, users would continuously experience new melodic loops with each note coming from separate directions according to the user’s position. The separation of the mobile speakers in itself provided an additional modality to the acoustic quality of the soundscape, formed by the relationship between perceived intensity and distance. For the most part, however, users were staring directly at their mobile phones during the interaction process, with their bodies constantly turned towards the sensing apparatus standing parallel to each other, majorly discouraging prospects for mutual acknowledgement through eye contact.
From an observational standpoint, it was not possible to discern synchronous movement patterns inspired by the installation, or any definitive collective movement traits for that matter. Compared to the first case study, resourced with a large stage, the spatial exploration was far more limited here, restricted only to a linear sensing area, revolving around one focal point. Users would insist on staying idle, waiting for the sound to loop for a while, perhaps experimenting with walking forwards or backwards a few steps to trigger new notes. This coincides with the outcomes that presented themselves in Scene ii (Section 3.5) in the way that, when the directional influence of the sensors is displaced from the body and onto the surroundings, users show more infatuation with their own movements over anyone else’s. With the ambition to cater for a new sociable space, the non-wearable arrangement combined with the drop-in and out functionality of the app was assumed to incentivise a flexible interchange of users, inclined to welcome those non-acquainted into the space. That said, we did not observe any instances of strangers simultaneously engaging at the same time, only those already affiliated, supposedly conditioned to the tensions when asserting oneself into a predetermined social clique.
In this instance, we wish to examine how well the intervention harmonises with the everyday operation of the space, blending with naturally occurring social exchanges, and staying respectful to those not actively participating. We found from a sample of 10 individuals and groups passing by during a weekday afternoon that 7 of 10 would continue walking, three would feel captivated to read the information board with one going as far as entering themselves into the application, and starting to engage with the sound feedback. During the weekend period, the installation would be stationed along the route of a few public walking tours, serving as an amusing artefact to intrigued bystanders, without pulling enough attention for anyone to abandon their personal schedule. In the evening, we observed a spontaneous social gathering take place in the square comprised of 15 or more people mingling beside the installation area. Small groups would approach the installation and briefly engage in a new session, triggering just a few notes before returning back to rejoin the social event. Accepting the severe limitations in retaining interest from new users, these impulsive engagements showcase a strong starting point for public inclusion, by which the artefact successfully captures the attention of broader audiences enough for voluntary initiation. This can partly be owed to our preparation stages, pleading for Constrained Complexity to minimise the learning curve.

6.4. Limitations of Public Space Adaptations

In this supplementary case study, we were granted the opportunity to confront the challenges of proxemic sensing in public space; this exposed a number of external factors that were less problematic in semi-controlled conditions. To confront our final research question RQ3, we discuss design decisions that were influenced by environmental changes, technical durability and usability, contributing to pervasiveness and inclusion. The non-wearable solution was less susceptible to errors, but at the expense of a confined sensing area. The linear note-based interaction combined with the fixed placement of the sensor improved the system’s usability when it was made openly accessible to a public audience. With that said, we believe this approach minimised the user’s resilience to unexpected outcomes, and made apparent the moment that sound would stop playing, user’s would immediately lose interest and move on to proceed with the rest of their day. In lack of a firm recommendation here, these outcomes continue to linger onto the feasibility of engaging unskilled users with novel interactive music systems [147], only to be exaggerated in a pervasive setting where prolonged participation is nonobligatory. In its wearable form, the original orchestration was purposed specifically to capture group dynamics by way of geometric and temporal relationships. However, here, we discern that the users were not highly aware of the movements happening around them. This persuasion may be accredited to the animations that are triggered concurrently within the application window, for which the extra stimulation has an overriding effect on mutual engagement, as is inferred by Bryan-Kinns’s study [120]. Coinciding with the issues we faced when introducing additional visual elements, the authors ultimately caution against an excessive exposure to non-essential information in order to maintain attention in a collaborative interaction setting.

7. Conclusions

This article reflects upon the common attitudes associated with interpersonal distancing and social connection during the most critical moments of pandemia. We speculate upon a spatially-informed intervention to be deployed as part of a performance; this incentivised the design of a sensory face mask, coupled with a system for sound interaction. This wearable orchestration was trailed over a series of workshop sessions, inviting participatory feedback used to refine the physical design and mapping strategies in anticipation to be presented in a live performance setting.
Reflecting upon observational notes and data analysis, we construct design considerations that respond to the pivotal challenges surrounding safe, inclusive re-socialisation in public and in theory, what spatially sensitive systems can offer to overcome such issues. This ultimately calls for an individualistic understanding of proxemic boundaries, giving agency to neighbouring bodies through sequential control, adaptable participation and constrained complexity strategies. We frame our findings in a broader perspective on sensory interventions that are not solely relevant to pandemic measures, generating critical reflections that later inform future developments as we appropriate the system towards urban sensing environments, proving transferable qualities amidst persisting limitations when subjected to the general public.
We outline the standout progressions between the two sensing mechanisms, one situated directly onto the body, the other installed into the surroundings, proving transferable qualities amidst persisting limitations when subjected to the general public. For future work, we foresee the benefit of incorporating a hybrid system comprised of wearable and environmental sensors, suitable for large open spaces in the confidence of robust operation. From here, we also look towards long-term studies with diverse user groups, crucial in forming generalisable conclusions of proxemic behaviour and sound feedback strategies.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/electronics11142151/s1, Figure 4: Acceleration Data Graph; Video S1: Video Extract; Video S2: Performance Video Extract; Video S3: Sound Installation Video.

Author Contributions

H.P.D.S. and H.G. are founders of the company responsible for the design and manufacturing of some of the sensory technology used for the wearable presented in the article, specifically the microcontroller component (PLUX Wireless Biosignals). W.P. was an employed researcher with the company during the time of the study. All authors have read and agreed to the published version of the manuscript.

Funding

This work has been supported by Marie Skłodowska Curie Actions ITN AffecTech (ERC H2020 Project 1059 ID: 722022).

Informed Consent Statement

The research meets all ethical guidelines, including adherence to the legal requirements of the study country. Informed consent was obtained from all subjects involved in the study to publish this paper.

Data Availability Statement

All available data will be furnished upon request to the corresponding author, W.P., upon reasonable request.

Acknowledgments

We thank the performers for partaking in the study and the audience members for their attendance. Additionally, we acknowledge the work of Bruno Reis for conducting the artistic residency and final performance hosted by Carpintarias de São Lázaro, Câmara Municpal de Lisboa for supporting the public installation. In addition, finally, we thank André Pinto for operating the drone.

Conflicts of Interest

H.P.D.S. and H.G. are founders of the company who are responsible for the design and manufacturing of some of the sensory technology that was used for the wearable presented in the article, specifically the microcontroller component (PLUX Wireless Biosignals). W.P. was an employed researcher with the company during the time of the first study.

Appendix A

Appendix B

References

  1. Cucinotta, D.; Vanelli, M. WHO Declares COVID-19 a Pandemic. Acta Bio Medica Atenei Parm. 2020, 91, 157–160. [Google Scholar] [CrossRef]
  2. Woelfel, R.; Corman, V.M.; Guggemos, W.; Seilmaier, M.; Zange, S.; Mueller, M.A.; Niemeyer, D.; Vollmar, P.; Rothe, C.; Hoelscher, M.; et al. Virological assessment of hospitalized patients with COVID-2019. Nature 2020, 581, 465–469. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Toquero, C.M. Challenges and Opportunities for Higher Education amid the COVID-19 Pandemic: The Philippine Context. Pedagog. Res. 2020, 5, em0063. [Google Scholar] [CrossRef] [Green Version]
  4. Work and Well-being during COVID-19: Impact, Inequalities, Resilience, and the Future of Work. Available online: https://www.gov.uk/government/publications/coronavirus-act-report-november-2020 (accessed on 17 December 2021).
  5. Marta, C.; Nunzio, L.; Francesca, F.; Martina, M.; Anna, K.; Vittorio, G.; Alessandra, U.M. The consequences of COVID-19 on social interactions: An online study on face covering. Sci. Rep. 2021, 11, 2601. [Google Scholar] [CrossRef]
  6. Carbon, C.C. Wearing Face Masks Strongly Confuses Counterparts in Reading Emotions. Front. Psychol. 2020, 11, 566886. [Google Scholar] [CrossRef]
  7. Grundmann, F.; Epstude, K.; Scheibe, S. Face masks reduce emotion-recognition accuracy and perceived closeness. PLoS ONE 2020, 16, e0249792. [Google Scholar] [CrossRef]
  8. Covid-19 Mitigation Phase—Recovery Guide 28th May. Available online: https://www.safecommunitiesportugal.com/major-incidents/guides/ (accessed on 22 April 2022).
  9. Seshadri, D.R.; Davies, E.V.; Harlow, E.R.; Hsu, J.J.; Knighton, S.C.; Walker, T.A.; Voos, J.E.; Drummond, C.K. Wearable Sensors for COVID-19: A Call to Action to Harness Our Digital Infrastructure for Remote Patient Monitoring and Virtual Assessments. Front. Digit. Health 2020, 2, 8. [Google Scholar] [CrossRef]
  10. Jeong, H.; Rogers, J.A.; Xu, S. Continuous on-body sensing for the COVID-19 pandemic: Gaps and opportunities. Sci. Adv. 2020, 6, eabd4794. [Google Scholar] [CrossRef]
  11. Kroczek, L.O.H.; Pfaller, M.; Lange, B.; Müller, M.; Mühlberger, A. Interpersonal Distance During Real-Time Social Interaction: Insights From Speripersonal spaceubjective Experience, Behavior, and Physiology. Front. Psychiatry 2020, 11, 561. [Google Scholar] [CrossRef]
  12. Vinciarelli, A.; Mohammadi, G. Towards a technology of nonverbal communication: Vocal behavior in social and affective phenomena. In Affective Computing and Interaction: Psychological, Cognitive and Neuroscientific Perspectives; IGI Global: Hershey, PA, USA, 2011; pp. 133–156. [Google Scholar]
  13. Sundstrom, E.; Altman, I. Interpersonal relationships and personal space: Research review and theoretical model. Hum. Ecol. 1976, 4, 47–67. [Google Scholar] [CrossRef]
  14. Yu, X.; Xiong, W.; Lee, Y.C. An Investigation Into Interpersonal and Peripersonal Spaces of Chinese People for Different Directions and Genders. Front. Psychol. 2020, 11, 981. [Google Scholar] [CrossRef] [PubMed]
  15. Sorokowska, A.; Sorokowski, P.; Hilpert, P.; Cantarero, K.; Frackowiak, T.; Ahmadi, K.; Alghraibeh, A.; Aryeetey, R.; Bertoni, A.; Bettache, K.; et al. Preferred Interpersonal Distances: A Global Comparison. J. Cross-Cult. Psychol. 2017, 48, 002202211769803. [Google Scholar] [CrossRef] [Green Version]
  16. Ang, S. Your friends, my friends, and our family: Informal social participation and mental health through the lens of linked lives. Soc. Sci. Med. 2021, 276, 113848. [Google Scholar] [CrossRef] [PubMed]
  17. Simões Aelbrecht, P. Rethinking urban design for a changing public life. J. Place Manag. Dev. 2010, 3, 113–129. [Google Scholar] [CrossRef]
  18. Sugiyama, T.; Leslie, E.; Giles-Corti, B.; Owen, N. Associations of neighbourhood greenness with physical and mental health: Do walking, social coherence and local social interaction explain the relationships? J. Epidemiol. Community Health 2008, 62, e9. [Google Scholar] [CrossRef] [Green Version]
  19. Brooks, S.K.; Webster, R.K.; Smith, L.E.; Woodland, L.; Wessely, S.; Greenberg, N.; Rubin, G.J. The psychological impact of quarantine and how to reduce it: Rapid review of the evidence. Lancet 2020, 395, 912–920. [Google Scholar] [CrossRef] [Green Version]
  20. Rivera, D.S.; Lindsay, C.B.; Oliva, C.A.; Codocedo, J.F.; Bozinovic, F.; Inestrosa, N.C. Effects of long-lasting social isolation and re-socialization on cognitive performance and brain activity: A longitudinal study in Octodon degus. Sci. Rep. 2020, 10, 18315. [Google Scholar] [CrossRef]
  21. Zorzo, C.; Méndez-López, M.; Méndez, M.; Arias, J.L. Adult social isolation leads to anxiety and spatial memory impairment: Brain activity pattern of COx and c-Fos. Behav. Brain Res. 2019, 365, 170–177. [Google Scholar] [CrossRef] [Green Version]
  22. Evans, I.E.; Martyr, A.; Collins, R.; Brayne, C.; Clare, L. Social Isolation and Cognitive Function in Later Life: A Systematic Review and Meta-Analysis. J. Alzheimer’S Dis. 2019, 70, S119–S144. [Google Scholar] [CrossRef] [Green Version]
  23. Romania, V. Interactional Anomie? Imaging Social Distance after COVID-19: A Goffmanian Perspective. Sociologica 2020, 14, 51–66. [Google Scholar] [CrossRef]
  24. Long, E.; Patterson, S.; Maxwell, K.; Blake, C.; Bosó Pérez, R.; Lewis, R.; McCann, M.; Riddell, J.; Skivington, K.; Wilson-Lowe, R.; et al. COVID-19 pandemic and its impact on social relationships and health. J. Epidemiol. Community Health 2021, 76, 128–132. [Google Scholar] [CrossRef] [PubMed]
  25. Hall, E.T. The Hidden Dimension; Doubleday: Garden City, NY, USA, 1966; Volume 609. [Google Scholar]
  26. Guerrero, L.; Fleuriet, C. Proxemics. In Communication; Oxford University Press: Oxford, UK, 2015. [Google Scholar] [CrossRef]
  27. McCall, C. Mapping Social Interactions: The Science of Proxemics. In Social Behavior from Rodents to Humans: Neural Foundations and Clinical Implications; Current Topics in Behavioral Neurosciences; Wöhr, M., Krach, S., Eds.; Springer International Publishing: Cham, Switzerland, 2017; pp. 295–308. [Google Scholar] [CrossRef]
  28. Pérez, P.; Roose, P.; Cardinale, Y.; Dalmau, M.; Masson, D.; Couture, N. Mobile proxemic application development for smart environments. In Proceedings of the 18th International Conference on Advances in Mobile Computing & Multimedia, Chiang Mai, Thailand, 30 November–2 December 2020; ACM: New York, NY, USA, 2020; pp. 94–103. [Google Scholar] [CrossRef]
  29. Marquardt, N.; Greenberg, S. Proxemic Interactions: From Theory to Practice. Synth. Lect. Hum. Centered Inform. 2015, 8, 1–199. [Google Scholar] [CrossRef]
  30. Fdili Alaoui, S.; Carlson, K.; Cuykendall, S.; Bradley, K.; Studd, K.; Schiphorst, T. How Do Experts Observe Movement? In Proceedings of the 2nd International Workshop on Movement and Computing; ACM: New York, NY, USA, 2015; pp. 84–91. [Google Scholar] [CrossRef] [Green Version]
  31. Schiphorst, T. Self-evidence: Applying somatic connoisseurship to experience design. In Proceedings of the CHI ’11 Extended Abstracts on Human Factors in Computing Systems, 7–12 May 2011; Association for Computing Machinery: Vancouver, BC, Canada, 2011; pp. 145–160. [Google Scholar] [CrossRef]
  32. Mehta, V. The new proxemics: COVID-19, social distancing, and sociable space. J. Urban Des. 2020, 25, 669–674. [Google Scholar] [CrossRef]
  33. Ye, J. The ambivalence of familiarity: Understanding breathable diversity through fleeting encounters in Singapore’s Jurong West: The ambivalence of familiarity. Area 2016, 48, 77–83. [Google Scholar] [CrossRef]
  34. van den Berg, P.; Sharmeen, F.; Weijs-Perrée, M. On the subjective quality of social Interactions: Influence of neighborhood walkability, social cohesion and mobility choices. Transp. Res. Part A Policy Pract. 2017, 106, 309–319. [Google Scholar] [CrossRef]
  35. Fayard, A.L.; Weeks, J. Photocopiers and Water-coolers: The Affordances of Informal Interaction. Organ. Stud. 2007, 28, 605–634. [Google Scholar] [CrossRef]
  36. White, F.A.; Borinca, I.; Vezzali, L.; Reynolds, K.J.; Blomster Lyshol, J.K.; Verrelli, S.; Falomir-Pichastor, J.M. Beyond direct contact: The theoretical and societal relevance of indirect contact for improving intergroup relations. J. Soc. Issues 2021, 77, 132–153. [Google Scholar] [CrossRef]
  37. Durnová, A.; Mohammadi, E. Intimacy, home, and emotions in the era of the pandemic. Sociol. Compass 2021, 15, e12852. [Google Scholar] [CrossRef]
  38. Ballendat, T.; Marquardt, N.; Greenberg, S. Proxemic interaction: Designing for a proximity and orientation-aware environment. In Proceedings of the ACM International Conference on Interactive Tabletops and Surfaces-ITS ’10, Saarbrücken, Germany, 7–10 November 2010; ACM Press: New York, NY, USA, 2010; p. 121. [Google Scholar] [CrossRef]
  39. Gehl, J.; Gemzøe, L. (Eds.) Public Spaces Public Life, 3rd ed.; The Danish Architectural Press: Copenhagen, Denmark, 2004. [Google Scholar]
  40. Machleit, K.A.; Eroglu, S.A.; Mantel, S.P. Perceived retail crowding and shopping satisfaction: What modifies this relationship? J. Consum. Psychol. 2000, 9, 29–42. [Google Scholar] [CrossRef]
  41. Kendrick, V.L.; Haslam, R.A. The User Experience of Crowds—A Human Factors Challenge. Proc. Hum. Factors Ergon. Soc. Annu. Meet. 2010, 54, 2000–2004. [Google Scholar] [CrossRef] [Green Version]
  42. Peimani, N.; Kamalipour, H. Where gender comes to the fore: Mapping gender mix in urban public spaces. Spaces Flows Int. J. Urban Extraurban Stud. 2016, 8, 19–30. [Google Scholar] [CrossRef]
  43. Remland, M.S.; Jones, T.S.; Brinkman, H. Interpersonal distance, body orientation, and touch: Effects of culture, gender, and age. J. Soc. Psychol. 1995, 135, 281–297. [Google Scholar] [CrossRef]
  44. Honey-Rosés, J.; Anguelovski, I.; Chireh, V.K.; Daher, C.; van den Bosch, C.K.; Litt, J.S.; Mawani, V.; McCall, M.K.; Orellana, A.; Oscilowicz, E.; et al. The impact of COVID-19 on public space: An early review of the emerging questions—Design, perceptions and inequities. Cities Health 2020, 5, 1–17. [Google Scholar] [CrossRef]
  45. Mehrdad, S.; Wang, Y.; Atashzar, S.F. Perspective: Wearable Internet of Medical Things for Remote Tracking of Symptoms, Prediction of Health Anomalies, Implementation of Preventative Measures, and Control of Virus Spread During the Era of COVID-19. Front. Robot. AI 2021, 8, 610653. [Google Scholar] [CrossRef] [PubMed]
  46. Taylor, W.; Abbasi, Q.H.; Dashtipour, K.; Ansari, S.; Shah, S.A.; Khalid, A.; Imran, M.A. A Review of the State of the Art in Non-Contact Sensing for COVID-19. Sensors 2020, 20, 5665. [Google Scholar] [CrossRef] [PubMed]
  47. Adhitya, S.; Scott, D. The London Soundmap: Integrating sonic interaction design in the urban realm. In Proceedings of the Audio Mostly 2018 on Sound in Immersion and Emotion, Wrexham, UK, 12–14 September 2018; ACM: New York, NY, USA, 2018; pp. 1–7. [Google Scholar] [CrossRef]
  48. Gatehouse, C. Feral Screens: Queering Urban Networked Publics. In Proceedings of the 2016 ACM Conference Companion Publication on Designing Interactive Systems, Brisbane, QLD, Australia, 3–4 June 2014; ACM: New York, NY, USA, 2014; pp. 99–104. [Google Scholar] [CrossRef]
  49. Howell, N.; Niemeyer, G.; Ryokai, K. Life-Affirming Biosensing in Public: Sounding Heartbeats on a Red Bench. In Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems, Scotland, UK, 4–9 May 2019; Association for Computing Machinery: New York, NY, USA, 2019; pp. 1–16. [Google Scholar] [CrossRef]
  50. Battarbee, K.; Baerten, N.; Hinfelaar, M.; Irvine, P.; Loeber, S.; Munro, A.; Pederson, T. Pools and satellites: Intimacy in the city. In Proceedings of the 4th Conference on Designing Interactive Systems: Processes, Practices, Methods, and Techniques, London, UK, 25–28 June 2002; Association for Computing Machinery: New York, NY, USA, 2002; pp. 237–245. [Google Scholar] [CrossRef]
  51. Lange, M.D.; Waal, M.D. First Monday. In Owning the City: New Media and Citizen Engagement in Urban Design; Taylor & Francis Group: Berlin/Heidelberg, Germany, 2013. [Google Scholar] [CrossRef]
  52. Bueno, A.d.O. From Smart Cities to Social Cities: Technology to Support Community Life. In Proceedings of the 2016 CHI Conference Extended Abstracts on Human Factors in Computing Systems, San Jose, CA, USA, 7–12 May 2016; Association for Computing Machinery: New York, NY, USA, 2016; pp. 198–202. [Google Scholar] [CrossRef]
  53. Castro Seixas, E. Urban (Digital) Play and Right to the City: A Critical Perspective. Front. Psychol. 2021, 12, 636111. [Google Scholar] [CrossRef] [PubMed]
  54. Katila, J.; Gan, Y.; Goodwin, M.H. Interaction rituals and ‘social distancing’: New haptic trajectories and touching from a distance in the time of COVID-19. Discourse Stud. 2020, 22, 418–440. [Google Scholar] [CrossRef]
  55. Farina, L.; Boussauw, K.; Plyushteva, A. Moving safely at night? Women’s nocturnal mobilities in Recife, Brazil and Brussels, Belgium. Gender Place Cult. 2021, 1–22. [Google Scholar] [CrossRef]
  56. Givon-Benjio, N.; Okon-Singer, H. Biased estimations of interpersonal distance in non-clinical social anxiety. J. Anxiety Disord. 2020, 69, 102171. [Google Scholar] [CrossRef]
  57. Crucianelli, L.; Filippetti, M.L. Developmental Perspectives on Interpersonal Affective Touch. Topoi 2020, 39, 575–586. [Google Scholar] [CrossRef] [Green Version]
  58. Field, T. Touch for socioemotional and physical well-being: A review. Dev. Rev. 2010, 30, 367–383. [Google Scholar] [CrossRef]
  59. Peterson, A.M.; Cline, R.J.; Foster, T.S. Parents’ interpersonal distance and touch behavior and child pain and distress during painful pediatric oncology procedures. J. Nonverbal Behav. 2007, 31, 79–97. [Google Scholar] [CrossRef]
  60. Dolin, D.J.; Booth-Butterfield, M. Reach out and touch someone: Analysis of nonverbal comforting responses. Commun. Q. 1993, 41, 383–393. [Google Scholar] [CrossRef]
  61. Bischoff-Grethe, A.; Wierenga, C.E.; Berner, L.A.; Simmons, A.N.; Bailer, U.; Paulus, M.P.; Kaye, W.H. Neural hypersensitivity to pleasant touch in women remitted from anorexia nervosa. Transl. Psychiatry 2018, 8, 161. [Google Scholar] [CrossRef] [PubMed]
  62. Sivik, T. Alexithymia and hypersensitivity to touch and palpation. Integr. Physiol. Behav. Sci. 1993, 28, 130–136. [Google Scholar] [CrossRef] [PubMed]
  63. Gulledge, A.K.; Hill, M.; Lister, Z.; Sallion, C. Non-erotic physical affection: It’s good for you. In Low-Cost Approaches to Promote Physical and Mental Health; Springer: Berlin/Heidelberg, Germany, 2007; pp. 371–384. [Google Scholar]
  64. LaFrance, M.; Vial, A.C. Gender and nonverbal behavior. In APA Handbook of Nonverbal Communication; Matsumoto, D., Hwang, H.C., Frank, M.G., Eds.; American Psychological Association: Washington, DC, USA, 2016; pp. 139–161. [Google Scholar] [CrossRef]
  65. Suvilehto, J.T.; Glerean, E.; Dunbar, R.I.M.; Hari, R.; Nummenmaa, L. Topography of social touching depends on emotional bonds between humans. Proc. Natl. Acad. Sci. USA 2015, 112, 13811–13816. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  66. Fischer, A.H.; Gillebaart, M.; Rotteveel, M.; Becker, D.; Vliek, M. Veiled Emotions: The Effect of Covered Faces on Emotion Perception and Attitudes. Soc. Psychol. Personal. Sci. 2012, 3, 266–273. [Google Scholar] [CrossRef]
  67. Blais, C.; Roy, C.; Fiset, D.; Arguin, M.; Gosselin, F. The eyes are not the window to basic emotions. Neuropsychologia 2012, 50, 2830–2838. [Google Scholar] [CrossRef]
  68. Campagne, D.M. The problem with communication stress from face masks. J. Affect. Disord. Rep. 2021, 3, 100069. [Google Scholar] [CrossRef]
  69. Baldassare, M.; Feller, S. Cultural Variations in Personal Space: Theory, Methods, and Evidence. Ethos 1975, 3, 481–503. [Google Scholar] [CrossRef]
  70. Cartaud, A.; Quesque, F.; Coello, Y. Wearing a face mask against Covid-19 results in a reduction of social distancing. PLoS ONE 2020, 15, e0243023. [Google Scholar] [CrossRef] [PubMed]
  71. Qian, Z.; Linlin, C.; Qun, Y. The effects of facial features on judicial decision-making. Adv. Psychol. Sci. 2018, 26, 698. [Google Scholar] [CrossRef]
  72. Todorov, A. Face Value: The Irresistible Influence of First Impressions; Princeton University Press: Princeton, NJ, USA, 2017. [Google Scholar]
  73. Billings, D.; Davidson, A.; Schaeffer, J.; Szafron, D. The challenge of poker. Artif. Intell. 2002, 134, 201–240. [Google Scholar] [CrossRef]
  74. Poyatos, F. Interactive functions and limitations of verbal and nonverbal behaviors in natural conversation. Semiotica 1980, 30. [Google Scholar] [CrossRef]
  75. Jones, R.G. Communication in the Real World: An Introduction to Communication Studies, 1st ed.; University of Minnesota Libraries Publishing: Minneapolis, MN, USA, 2016. [Google Scholar] [CrossRef]
  76. Hadley, L.V.; Ward, J.A. Synchrony as a measure of conversation difficulty: Movement coherence increases with background noise level and complexity in dyads and triads. PLoS ONE 2021, 16, e0258247. [Google Scholar] [CrossRef] [PubMed]
  77. Saunders, G.H.; Jackson, I.R.; Visram, A.S. Impacts of face coverings on communication: An indirect impact of COVID-19. Int. J. Audiol. 2021, 60, 495–506. [Google Scholar] [CrossRef] [PubMed]
  78. Rahne, T.; Fröhlich, L.; Plontke, S.; Wagner, L. Influence of face surgical and N95 face masks on speech perception and listening effort in noise. PLoS ONE 2021, 16, e0253874. [Google Scholar] [CrossRef]
  79. Nobrega, M.; Opice, R.; Lauletta, M.M.; Nobrega, C.A. How face masks can affect school performance. Int. J. Pediatr. Otorhinolaryngol. 2020, 138, 110328. [Google Scholar] [CrossRef]
  80. McKee, M.; Moran, C.; Zazove, P. Overcoming Additional Barriers to Care for Deaf and Hard of Hearing Patients During COVID-19. JAMA Otolaryngol. Neck Surg. 2020, 146, 781. [Google Scholar] [CrossRef]
  81. Inceoglu, S. Language Experience and Subjective Word Familiarity on the Multimodal Perception of Non-native Vowels. Lang. Speech 2021, 65, 002383092199872. [Google Scholar] [CrossRef]
  82. Jabber, K.W.; Mahmood, A.A. Non-verbal Communication between Two Non-native English Speakers: Iraqi and Chinese. Theory Pract. Lang. Stud. 2020, 10, 189. [Google Scholar] [CrossRef]
  83. Laban, R.V.; Ullmann, L. The Mastery of Movement; Northcote House: Plymouth, UK, 1988. [Google Scholar]
  84. Davies, E. Beyond Dance: Laban’s Legacy of Movement Analysis; Routledge: New York, NY, USA, 2006. [Google Scholar]
  85. Ziegelmaier, R.S.; Correia, W.; Teixeira, J.M.; Simoes, F.P.M. Laban Movement Analysis applied to Human-Computer Interaction. In Proceedings of the 2020 22nd Symposium on Virtual and Augmented Reality (SVR), Porto de Galinhas, Brazil, 7–10 November 2020; pp. 30–34. [Google Scholar] [CrossRef]
  86. Roudposhti, K.K.; Santos, L.; Aliakbarpour, H.; Dias, J. Parameterizing interpersonal behaviour with Laban movement analysis—A Bayesian approach. In Proceedings of the 2012 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, Providence, RI, USA, 16–21 June 2012; pp. 7–13. [Google Scholar] [CrossRef]
  87. Pentland, A.; Heibeck, T. Honest Signals: How They Shape Our World; The MIT Press: Cambridge, MA, USA, 2010. [Google Scholar]
  88. Landry, S.; Jeon, M. Interactive sonification strategies for the motion and emotion of dance performances. J. Multimodal User Interfaces 2020, 14, 167–186. [Google Scholar] [CrossRef] [Green Version]
  89. Krzyzaniak, M.; Frohlich, D.; Jackson, P.J. Six types of audio that DEFY reality!: A taxonomy of audio augmented reality with examples. In Proceedings of the 14th International Audio Mostly Conference: A Journey in Sound, Nottingham, UK, 18–20 September 2019; ACM: New York, NY, USA, 2019; pp. 160–167. [Google Scholar] [CrossRef]
  90. Wu, C.J.; Quigley, A.; Harris-Birtill, D. Out of sight: A toolkit for tracking occluded human joint positions. Pers. Ubiquitous Comput. 2017, 21, 125–135. [Google Scholar] [CrossRef] [Green Version]
  91. Hensel, B.K.; Demiris, G.; Courtney, K.L. Defining Obtrusiveness in Home Telehealth Technologies: A Conceptual Framework. J. Am. Med. Inform. Assoc. 2006, 13, 428–431. [Google Scholar] [CrossRef] [Green Version]
  92. Rector, K.; Salmon, K.; Thornton, D.; Joshi, N.; Morris, M.R. Eyes-Free Art: Exploring Proxemic Audio Interfaces For Blind and Low Vision Art Engagement. Proc. ACM Interactive Mobile Wearable Ubiquitous Technol. 2017, 1, 1–21. [Google Scholar] [CrossRef]
  93. Bryan-Kinns, N. Interaction Design with Audio: Speculating on Sound in Future Design Education. In Proceedings of the 4th Central China International Design Science Seminar, Belfast, UK, 4–6 July 2017; pp. 19–22. [Google Scholar]
  94. Alfaras, M.; Primett, W.; Umair, M.; Windlin, C.; Karpashevich, P.; Chalabianloo, N.; Bowie, D.; Sas, C.; Sanches, P.; Höök, K.; et al. Biosensing and Actuation—Platforms Coupling Body Input-Output Modalities for Affective Technologies. Sensors 2020, 20, 5968. [Google Scholar] [CrossRef]
  95. Solberg, R.T.; Jensenius, A.R. Group behaviour and interpersonal synchronization to electronic dance music. Music. Sci. 2019, 23, 111–134. [Google Scholar] [CrossRef] [Green Version]
  96. Bischoff, J.; Gold, R.; Horton, J. Music for an Interactive Network of Microcomputers. Comput. Music. J. 1978, 2, 24–29. [Google Scholar] [CrossRef]
  97. Barbosa, A. Displaced Soundscapes: A Survey of Network Systems for Music and Sonic Art Creation. Leonardo Music. J. 2003, 13, 53–59. [Google Scholar] [CrossRef]
  98. Jordà, S.; Geiger, G.; Alonso, M.; Kaltenbrunner, M. The reacTable: Exploring the synergy between live music performance and tabletop tangible interfaces. In Proceedings of the 1st International Conference on Tangible and Embedded Interaction-TEI ’07, Baton Rouge, LA, USA, 15–17 February 2007; ACM Press: New York, NY, USA, 2007; p. 139. [Google Scholar] [CrossRef]
  99. Freeman, J.; Varnik, K.; Ramakrishnan, C.; Neuhaus, M.; Burk, P.; Birchfield, D. Auracle: A voice-controlled, networked sound instrument. Organised Sound 2005, 10, 221–231. [Google Scholar] [CrossRef] [Green Version]
  100. Aly, L.; Silva, H.; Bernardes, G.; Penha, R. Appropriating Biosensors as Embodied Control Structures in Interactive Music Systems. Hum. Technol. 2021, 17, 45–80. [Google Scholar] [CrossRef]
  101. Hege, A.K. Spirit in the Flesh: Essays on the Relationship between the Body and Meaning in Performance; Princeton University: Princeton, NJ, USA, 2014. [Google Scholar]
  102. Schnell, N.; Robaszkiewicz, S.; Bevilacqua, F.; Schwarz, D. Collective Sound Checks: Exploring Intertwined Sonic and Social Affordances of Mobile Web Applications. In Proceedings of the Ninth International Conference on Tangible, Embedded, and Embodied Interaction, Stanford, CA, USA, 15–19 January 2015; ACM: New York, NY, USA, 2015; pp. 685–690. [Google Scholar] [CrossRef]
  103. Bevilacqua, F.; Segalen, M.; Marchand-Pauvert, V.; Peyre, I.; Pradat-Diehl, P.; Roby-Brami, A. Exploring different movement sonification strategies for rehabilitation in clinical settings. In Proceedings of the 5th International Conference on Movement and Computing, Genoa, Italy, 28–30 June 2018; Association for Computing Machinery: New York, NY, USA, 2018; pp. 1–6. [Google Scholar] [CrossRef]
  104. Roo, J.S.; Gervais, R.; Lainé, T.; Cinquin, P.A.; Hachet, M.; Frey, J. Physio-Stacks: Supporting Communication with Ourselves and Others via Tangible, Modular Physiological Devices. In Proceedings of the 22nd International Conference on Human-Computer Interaction with Mobile Devices and Services, Oldenburg, Germany, 5–8 October 2020; ACM: New York, NY, USA, 2020; pp. 1–12. [Google Scholar] [CrossRef]
  105. Ashford, R. The EEG visualising pendant for social situations. In Proceedings of the Adjunct Proceedings of the 2019 ACM International Joint Conference on Pervasive and Ubiquitous Computing and Proceedings of the 2019 ACM International Symposium on Wearable Computers, London, UK, 9–13 September 2019; Association for Computing Machinery: New York, NY, USA, 2019; pp. 457–458. [Google Scholar] [CrossRef]
  106. COVID-19 and Your Health. Available online: https://www.cdc.gov/coronavirus/2019-ncov/prevent-getting-sick/prevention.html (accessed on 26 November 2021).
  107. Montanari, A.; Tian, Z.; Francu, E.; Lucas, B.; Jones, B.; Zhou, X.; Mascolo, C. Measuring Interaction Proxemics with Wearable Light Tags. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2018, 2, 25. [Google Scholar] [CrossRef] [Green Version]
  108. Hale, J.; Ward, J.A.; Buccheri, F.; Oliver, D.; Hamilton, A.F.D.C. Are You on My Wavelength? Interpersonal Coordination in Dyadic Conversations. J. Nonverbal Behav. 2020, 44, 63–83. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  109. Hamilton, A.; Pinti, P.; Paoletti, D.; Ward, J. Seeing into the brain of an actor with mocap and fNIRS. In Proceedings of the 2018 ACM International Symposium on Wearable Computers, Singapore, 8–12 October 2018; pp. 216–217. [Google Scholar] [CrossRef]
  110. Jürgens, S.; Correia, N.N.; Masu, R. Designing Glitch Procedures and Visualisation Workflows for Markerless Live Motion Capture of Contemporary Dance. In Proceedings of the 7th International Conference on Movement and Computing, Jersey City, NJ, USA, 15–17 July 2020; Association for Computing Machinery: New York, NY, USA, 2020. [Google Scholar] [CrossRef]
  111. Hernandez-Barrera, A. Teaching introduction to robotics: Using a blend of problem- and project-based learning approaches. In Proceedings of the IEEE SOUTHEASTCON 2014, Lexington, KY, USA, 13–16 March 2014; pp. 1–5. [Google Scholar] [CrossRef]
  112. Garcia-Ruiz, M.A.; Santana-Mancilla, P.C.; Gaytan-Lugo, L.S. Integrating Microcontroller-Based Projects in a Human-Computer Interaction Course. Int. J. Comput. Inf. Eng. 2018, 12, 946–950. [Google Scholar]
  113. Eguchi, S.; Yazaki, Y.; Kato, R.; Arita, Y.; Moriya, T.; Tanaka, H. Proto-Chair: Posture-Sensing Smart Furniture with 3D-Printed Auxetics. In Proceedings of the Extended Abstracts of the 2020 CHI Conference on Human Factors in Computing Systems, Honolulu, HI, USA, 25–30 April 2020; Association for Computing Machinery: New York, NY, USA, 2020; pp. 1–7. [Google Scholar] [CrossRef]
  114. Xu, W.; Yan, C.; Jia, W.; Ji, X.; Liu, J. Analyzing and Enhancing the Security of Ultrasonic Sensors for Autonomous Vehicles. IEEE Internet Things J. 2018, 5, 5015–5029. [Google Scholar] [CrossRef]
  115. Pal, J. Distance Measurement of Object by Ultrasonic Sensor HC-SR 04 Manpreet Kaur. Int. J. Sci. Res. Dev. 2015, 3, 503–505. [Google Scholar]
  116. Abreu, D.; Toledo, J.; Codina, B.; Suárez, A. Low-Cost Ultrasonic Range Improvements for an Assistive Device. Sensors 2021, 21, 4250. [Google Scholar] [CrossRef]
  117. Improve Ultrasonic Range Sensor Accuracy. Available online: https://www.instructables.com/Improve-Ultrasonic-Range-Sensor-Accuracy/ (accessed on 7 May 2022).
  118. Forsythe, W.; Kaiser, P. Dance Geometry. Perform. Res. 1999, 4, 64–71. [Google Scholar] [CrossRef]
  119. Clark, J.O.; Ando, T. Geometry, embodied cognition and choreographic praxis. Int. J. Perform. Arts Digit. Media 2014, 10, 179–192. [Google Scholar] [CrossRef]
  120. Bryan-Kinns, N. Mutual engagement and collocation with shared representations. Int. J. Hum. Comput. Stud. 2013, 71, 76–90. [Google Scholar] [CrossRef]
  121. Eriksson, S.; Unander-Scharin, A.; Trichon, V.; Unander-Scharin, C.; Kjellström, H.; Höök, K. Dancing With Drones: Crafting Novel Artistic Expressions Through Intercorporeality. In Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems, Glasgow, UK, 4–9 May 2019; Association for Computing Machinery: New York, NY, USA, 2019; pp. 1–12. [Google Scholar]
  122. La Delfa, J.; Baytas, M.A.; Patibanda, R.; Ngari, H.; Khot, R.A.; Mueller, F.F. Drone Chi: Somaesthetic Human-Drone Interaction. In Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems, Honolulu, HI, USA, 25–30 April 2020; ACM: New York, NY, USA, 2020; pp. 1–13. [Google Scholar] [CrossRef]
  123. Tezza, D.; Andujar, M. The State-of-the-Art of Human–Drone Interaction: A Survey. IEEE Access 2019, 7, 167438–167454. [Google Scholar] [CrossRef]
  124. Benford, S.; Greenhalgh, C.; Crabtree, A.; Flintham, M.; Walker, B.; Marshall, J.; Koleva, B.; Rennick Egglestone, S.; Giannachi, G.; Adams, M.; et al. Performance-Led Research in the Wild. ACM Trans. Comput. Hum. Interact. 2013, 20, 1–22. [Google Scholar] [CrossRef]
  125. Heron, J.; Kershaw, B. On PAR: A dialogue about performance-as-research. In Performance as Research; Routledge: London, UK, 2017. [Google Scholar]
  126. Trotter, S. Ways of Being Together During the COVID-19 Pandemic: Support Bubbles and the Legal Construction of Relationships. Front. Sociol. 2021, 6, 730216. [Google Scholar] [CrossRef] [PubMed]
  127. Bernardo, F.; Grierson, M.; Fiebrink, R. User-centred design actions for lightweight evaluation of an interactive machine learning toolkit. J. Sci. Technol. Arts 2018, 10, 25–38. [Google Scholar] [CrossRef] [Green Version]
  128. Ward, J.A.; Richardson, D.; Orgs, G.; Hunter, K.; Hamilton, A. Sensing interpersonal synchrony between actors and autistic children in theatre using wrist-worn accelerometers. In Proceedings of the 2018 ACM International Symposium on Wearable Computers-ISWC ’18, Singapore, 8–12 October 2018; ACM Press: New York, NY, USA, 2018; pp. 148–155. [Google Scholar] [CrossRef]
  129. Latham, A.; Layton, J. Social infrastructure and the public life of cities: Studying urban sociality and public spaces. Geogr. Compass 2019, 13, e12444. [Google Scholar] [CrossRef]
  130. Okkels, N.; Kristiansen, C.B.; Munk-Jørgensen, P.; Sartorius, N. Urban mental health: Challenges and perspectives. Curr. Opin. Psychiatry 2018, 31, 258–264. [Google Scholar] [CrossRef]
  131. Mehta, V. Look Closely and You Will See, Listen Carefully and You Will Hear: Urban Design and Social Interaction on Streets. J. Urban Des. 2009, 14, 29–64. [Google Scholar] [CrossRef]
  132. Fdili Alaoui, S. Making an Interactive Dance Piece: Tensions in Integrating Technology in Art. In Proceedings of the 2019 on Designing Interactive Systems Conference, San Diego, CA, USA, 23–28 June 2019; ACM: New York, NY, USA, 2019; pp. 1195–1208. [Google Scholar] [CrossRef]
  133. Cotofan, M. Work and Well-being during COVID-19: Impact, Inequalities, Resilience, and the Future of Work; World Happiness Report; University of Oxford: Oxford, UK, 2021; pp. 153–190. Available online: https://worldhappiness.report/ed/2021/work-and-well-being-during-covid-19-impact-inequalities-resilience-and-the-future-of-work/ (accessed on 17 December 2021).
  134. Goudarzi, V.; Gioti, A.M. Engagement and Interaction in Participatory Sound Art. In Proceedings of the 13th Sound and Music Computing Conference, Hamburg, Germany, 31 August–3 September 2016; pp. 179–184. [Google Scholar] [CrossRef]
  135. Rostami, A.; McMillan, D.; Márquez Segura, E.; Rossito, C.; Barkhuus, L. Bio-Sensed and Embodied Participation in Interactive Performance. In Proceedings of the Eleventh International Conference on Tangible, Embedded, and Embodied Interaction, Yokohama, Japan, 20–23 March 2017; Association for Computing Machinery: New York, NY, USA, 2017; pp. 197–208. [Google Scholar] [CrossRef]
  136. Mumtaz, F.; Khan, M.I.; Zubair, M.; Dehpour, A.R. Neurobiology and consequences of social isolation stress in animal model—A comprehensive review. Biomed. Pharmacother. 2018, 105, 1205–1222. [Google Scholar] [CrossRef]
  137. Forte, G.; Favieri, F.; Tambelli, R.; Casagrande, M. COVID-19 Pandemic in the Italian Population: Validation of a Post-Traumatic Stress Disorder Questionnaire and Prevalence of PTSD Symptomatology. Int. J. Environ. Res. Public Health 2020, 17, 4151. [Google Scholar] [CrossRef]
  138. Liang, L.; Gao, T.; Ren, H.; Cao, R.; Qin, Z.; Hu, Y.; Li, C.; Mei, S. Post-traumatic stress disorder and psychological distress in Chinese youths following the COVID-19 emergency. J. Health Psychol. 2020, 25, 1164–1175. [Google Scholar] [CrossRef]
  139. Layden, E.A.; Cacioppo, J.T.; Cacioppo, S. Loneliness predicts a preference for larger interpersonal distance within intimate space. PLoS ONE 2018, 13, e0203491. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  140. Ritchie, K.; Chan, D.; Watermeyer, T. The cognitive consequences of the COVID-19 epidemic: Collateral damage? Brain Commun. 2020, 2, fcaa069. [Google Scholar] [CrossRef] [PubMed]
  141. Lupton, D.; Lewis, S. Coping with COVID-19: The sociomaterial dimensions of living with pre-existing mental illness during the early stages of the coronavirus crisis. Emot. Space Soc. 2022, 42, 100860. [Google Scholar] [CrossRef] [PubMed]
  142. Wu, Y.; Bryan-Kinns, N. Supporting Non-Musicians? Creative Engagement with Musical Interfaces. In Proceedings of the 2017 ACM SIGCHI Conference on Creativity and Cognition, Singapore, 27–30 June 2017; ACM: New York, NY, USA, 2017; pp. 275–286. [Google Scholar] [CrossRef]
  143. Bengler, B.; Bryan-Kinns, N. Designing collaborative musical experiences for broad audiences. In Proceedings of the 9th ACM Conference on Creativity & Cognition, Sydney, Australia, 17–20 June 2013; ACM: New York, NY, USA, 2013; pp. 234–242. [Google Scholar] [CrossRef]
  144. Ritchie, H.; Mathieu, E.; Rodés-Guirao, L.; Appel, C.; Giattino, C.; Ortiz-Ospina, E.; Hasell, J.; Macdonald, B.; Beltekian, D.; Roser, M. Coronavirus Pandemic (COVID-19). Our World In Data. 2020. Available online: https://ourworldindata.org/covid-vaccinations (accessed on 25 June 2022).
  145. Saúde (DGS), D. COVID-19—Eventos de Grande Dimensão (Desportivos, Culturais, Corporativos e Outros). 2022. Available online: https://www.ordemenfermeiros.pt/media/25065/orienta%C3%A7%C3%A3o-014-2021-covid-19-eventos-de-grande-dimens%C3%A3o-desportivos-culturais-corporativos-e-outros-atualizada-a-09012022.pdf (accessed on 25 June 2022).
  146. Matuszewski, B.; Schnell, N.; Bevilacqua, F. Interaction Topologies in Mobile-Based Situated Networked Music Systems. Wirel. Commun. Mob. Comput. 2019, 2019, 9142490. [Google Scholar] [CrossRef]
  147. McPherson, A.; Morreale, F.; Harrison, J. Musical Instruments for Novices: Comparing NIME, HCI and Crowdfunding Approaches. In New Directions in Music and Human-Computer Interaction; Springer Series on Cultural Computing; Holland, S., Mudd, T., Wilkie-McKenna, K., McPherson, A., Wanderley, M.M., Eds.; Springer International Publishing: Cham, Switzerland, 2019; pp. 179–212. [Google Scholar] [CrossRef]
Figure 1. Proximity sensor enclosure fitted onto the face mask with trigger (output) and echo (receiver) signals.
Figure 1. Proximity sensor enclosure fitted onto the face mask with trigger (output) and echo (receiver) signals.
Electronics 11 02151 g001
Figure 2. Spectrogram sample of interactive soundscape showing “open” to “closed” transition.
Figure 2. Spectrogram sample of interactive soundscape showing “open” to “closed” transition.
Electronics 11 02151 g002
Figure 3. Visual representations for each scene. From left to right: (i) Geometric boundaries and interceptions, (ii) Interacting with the non-human, (iii) Participation from external users.
Figure 3. Visual representations for each scene. From left to right: (i) Geometric boundaries and interceptions, (ii) Interacting with the non-human, (iii) Participation from external users.
Electronics 11 02151 g003
Figure 4. Acceleration data recorded from scenes i–iii: the top row displays group median averages (1), with individual user data shown below (2). The final row aligns the peak cluster periods detected along the x-axis (3). A high-resolution version of the image is available in the Supplementary Material.
Figure 4. Acceleration data recorded from scenes i–iii: the top row displays group median averages (1), with individual user data shown below (2). The final row aligns the peak cluster periods detected along the x-axis (3). A high-resolution version of the image is available in the Supplementary Material.
Electronics 11 02151 g004
Figure 5. Annotations from Scene iii rehearsal video. The arrows show the walking direction with a dashed line to trace the dispersion of mask-wearing group. A video recording is included in the Supplementary Material.
Figure 5. Annotations from Scene iii rehearsal video. The arrows show the walking direction with a dashed line to trace the dispersion of mask-wearing group. A video recording is included in the Supplementary Material.
Electronics 11 02151 g005
Figure 6. Spectrogram recording from two scenes, ii and iii, marked with interruptions of the granular soundscape.
Figure 6. Spectrogram recording from two scenes, ii and iii, marked with interruptions of the granular soundscape.
Electronics 11 02151 g006
Figure 7. Public installation using four proximity sensors placed inside foliage with hanging ribbon. Sensors are physically separated by sensing trajectories, identified by colour.
Figure 7. Public installation using four proximity sensors placed inside foliage with hanging ribbon. Sensors are physically separated by sensing trajectories, identified by colour.
Electronics 11 02151 g007
Figure 8. User instructions for installation: (1) Access QR code; (2) Load the app and select a colour; (3) Increase the device volume, and (4) Walk towards the chosen colour to initiate sound.
Figure 8. User instructions for installation: (1) Access QR code; (2) Load the app and select a colour; (3) Increase the device volume, and (4) Walk towards the chosen colour to initiate sound.
Electronics 11 02151 g008
Table 1. Peak and cluster statistics from each scene, calculated from the individual user acceleration data that correspond to rows (2) and (3) in Figure 4.
Table 1. Peak and cluster statistics from each scene, calculated from the individual user acceleration data that correspond to rows (2) and (3) in Figure 4.
SceneMean Peak Interval (s)SD Peak Interval (s)Median Peak Amplitude (g)
i183.7149.90.31
ii182.1141.40.29
iii155.8138.20.36
User Effort Dissimilarity (%)Alternating User Peaks (%)Mean Cluster Concentration
(Peaks/Cluster)
Mean Cluster Duration (s)
i25.590.24.58.9
ii44.778.73.810.0
iii17.296.64.15.6
Table 2. Comparison of reflected materials from benchmark test recordings. From left to right: ceramic tile, skin from the hand, torso covered by clothing.
Table 2. Comparison of reflected materials from benchmark test recordings. From left to right: ceramic tile, skin from the hand, torso covered by clothing.
MaterialMinimum
Detection Range
Maximum
Detection Range
Dropout
Rate
Ceramic Tile5.1 cm288.7 cm0.0%
Skin10.2 cm246.4 cm3.85%
Clothing Fabric25.4 cm201.8 cm4.55%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Primett, W.; Da Silva, H.P.; Gamboa, H. Sound Feedback for Social Distance: The Case for Public Interventions during a Pandemic. Electronics 2022, 11, 2151. https://doi.org/10.3390/electronics11142151

AMA Style

Primett W, Da Silva HP, Gamboa H. Sound Feedback for Social Distance: The Case for Public Interventions during a Pandemic. Electronics. 2022; 11(14):2151. https://doi.org/10.3390/electronics11142151

Chicago/Turabian Style

Primett, William, Hugo Plácido Da Silva, and Hugo Gamboa. 2022. "Sound Feedback for Social Distance: The Case for Public Interventions during a Pandemic" Electronics 11, no. 14: 2151. https://doi.org/10.3390/electronics11142151

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop