1 Introduction
Until today, immersive mixed reality (MR) systems have been primarily confined to homes or other dedicated spaces, typically offering users the choice between controllers or hand tracking for their individual needs and experiences. Recent changes in the MR domain have made mobile settings increasingly prevalent, facilitating, e.g. remote collaboration in distant locations [
14], on-site urban city planning [
9], or mobile offices on trains [
31], and in future self-driving cars [
41]. This growing emphasis on MR mobility highlights the impracticality of carrying separate controllers alongside head-mounted displays (HMDs), emphasising controller-free MR interactions. However, while an HMD’s integrated hand-tracking feature reduces the hardware burden and allows high-fidelity user input, the absence of a tangible input device limits haptic explorations and immersion. This introduces the challenge of maintaining immersive haptic experiences within controller-free MR.
Previous research has been exploring different approaches to substitute dedicated handheld controllers with more ubiquitous devices for MR interactions by integrating, e.g. everyday objects [8], smartwatches [47] and touch-based gestures [44, 49] or 3D motion gestures with a mobile device [12, 20, 30].Given their widespread accessibility and advanced sensing capabilities via touch screens and internal measurement units (IMUs), mobile devices in particular hold promising potential for facilitating haptic MR interactions. Further, incorporating an HMD’s hand tracking feature can expand the sensing capabilities without adding hardware to the mobile device, enabling the detection of hand and finger movements in contact with the nowadays device’s non-sensing parts (e.g., overall shape, back and sides and immediate space around the device). Utilising mobile devices with extended sensing capabilities as an input device opens up the traditional design space of touch screen gestures more towards novel device-gestures for touching and haptically experiencing virtual objects in MR. Moreover, using mobile devices to provide haptic feedback for MR facilitates accessibility and mobility of future haptic technology. It is, therefore, important to understand how people haptically explore objects in the physical world and incorporate relevant aspects into the gesture design.
Therefore, to reintroduce haptic experiences into controller-free MR systems, we explore the utilisation of mobile devices as a replacement for specialised MR controllers. Our research methodology comprises two user studies. An initial online survey (
n = 50) empirically defines the object candidate set that manifests six common haptic object properties, including size, shape, texture, hardness, temperature, and weight [
24,
25,
26], which we use in a follow-up step when investigating haptic exploration. There, we conducted an in-depth gesture elicitation study (
n = 18) to investigate (1) user-defined hand gestures performed with prospective mobile devices and assess (2) users’ preferences for real-time haptic feedback for exploring haptic object properties in MR.
By synthesising our findings into a design space, we outline
ten main gesture types, including
grip + move,
pinch,
stroke,
press,
grip,
two-handed grip,
unsupported holding,
tap,
hover and
wearing gesture for virtual object exploration. Furthermore, we provide a variety of related haptic sensations voiced by our participants, including e.g.,
force feedback,
surface deformation,
weight shift,
shape change, and many more.
Moreover, variations of gestures within our design space illustrate three distinct modes that the device can adopt including (1) object mode, (2) hand mode and (3) tool mode (see Figure 1). We discuss our results and their implications for using the mobile device as an input tool for haptically exploring virtual objects in MR. In summary, this paper makes the following contributions:
(1)
It contributes a set of user-defined exploration gestures performed with a mobile device and with hands only in MR that extract specific haptic properties of virtual objects.
(2)
It offers a comprehensive design space that links haptic attributes of virtual objects with user-defined gestures and with the desired real-time haptic feedback.
(3)
Finally, our paper provides an in-depth understanding of how users adapt natural exploratory procedures to MR object interaction using mobile-device-input gestures.
This work broadens the scope of haptic interactions possible in MR with prospective mobile devices. By doing so, we hope that our design space fosters a knowledge-driven approach to integrating user-defined mobile device gestures that seamlessly align with user-defined haptic feedback. Moreover, our findings can guide the design of mobile devices, enrich their input and haptic output capabilities and shape future haptic experiences in mobile MR.
3 Study 1: Online Survey
Our online survey aims to understand people’s expectations about a
3D object’s haptic properties based on visual information.
Given the subjectivity of the haptic expectations that are mainly derived from the object’s graphical appearance to compensate for the absence of dexterous exploration, we employed an empirical approach to balance subjective perspectives and to capture a more objective and generalisable haptic understanding. Participants were asked to rate 100
virtual objects according to their expectations of the six haptic properties that we
derived from previous work [
22].
Lederman & Klatzky studied in a series of haptic exploration experiments what haptic knowledge people extract through touch [
24,
25,
26]
and identified six haptic attributes that comprehensively capture an object’s tactility: size, shape, texture, stiffness, temperature, and weight. The sum of all
presented 100
virtual objects covers
a range of these six haptic properties, contributing towards the generalisability of our results.
To present participants with visual representations of the 100 objects, we first collected 3D objects on various 3D modelling platforms, including cgtrader.com , sketchfab.com , and free3d.com . We were searching for publicly available 3D models and for 3D models that exhibit a visual fidelity that is not too abstract (e.g., a pixelated red sphere to represent an apple) and not too photorealistic (e.g., a highly realistic photocopy of an apple) to ensure we can re-use the 3D models later in our MR-based gesture elicitation study. We also aimed for a similar visual fidelity of all objects to keep it consistent across all 100 candidate objects, contributing towards internal validity. We then captured the original preview images as 2D copies of the 3D objects for the online survey, and stored the original 3D objects for later use in our gesture elicitation study.
3.1 Study Design and Participants
After preparing the study material, we deployed a within-subjects online survey to understand people’s perceptions of the six haptic properties of each candidate object. These properties include
size,
shape,
texture,
hardness,
temperature, and
weight, which cover the six most important characteristics when identifying objects [
22]. Each property manifests two characteristics on the level of each property. For example, for the property
size, participants could rate on a continuous 0-to-100 scale from small (0) to large (100) using a slider. We then calculated means to identify two objects each at both ends of each property continuum. For example, if one object received the lowest mean score across all objects for size, we considered this object to be a suitable candidate for representing a “small object” in our follow-up gesture elicitation study. The same was done for the highest mean score, representing a suitable candidate for, e.g. a “large object”.
Fifty participants
n = 50 (age ranging from 20 to 59, M = 28.65, SD = 8.41) were recruited using Prolific, an online crowd-sourcing platform that has found widespread application for empirical HCI research, e.g. [
33,
36]. The study was distributed evenly to male and female participants using Prolific’s internal distribution feature. We added three attention-check questions to ensure data quality and excluded survey takers who clearly did not put in significant effort in filling out the survey. To further contribute towards high data quality, we added a 5-point Likert Scale asking how genuinely the participants filled in the survey. This was done after four initial pilot tests. We emphasised that the answer to this question has no impact on their reward. Overall, the online survey was scheduled for 30 minutes and eventually manifested a median of 34.35 minutes. None of the survey takers had to be excluded for the data analysis based on the attention checks, and the participants’ commitment was M = 4.89 (SD = 0.31). Participants received an average reward of £6,01/hr using Prolific’s reimbursement system.
3.2 Procedure and Ethics
After an initial screening by the ethics board of our University of St. Gallen, our research was exempt from a formal full review. All participants signed up for the study directly on Prolific and were then forwarded to our survey, which was hosted on
SoSci Survey (https://www.soscisurvey.de/), a professional survey platform that allows designing and hosting surveys. Participants then provided informed consent of their participation and went through an example object that introduced them to their task (i.e.,
“In this survey, you will see images of objects and rate their properties. A short example at the beginning will illustrate the task.”). After going through the example, participants were asked to indicate their expectations of the size, shape, texture, temperature, weight, and stiffness of 100 objects using sliders from 0 to 100, as described in Section
3.1.
3.3 Results: Pre-Selection of Object Candidates
To identify objects that cover both ends of the category continuum (e.g., small size and large size), we
normalised the dataset and calculated means and standard deviations
(SDs) of participants’
indicated expectations of all six haptic properties of each object. This process identified the 12 objects that cover all properties (i.e., size, shape, texture, stiffness, temperature, weight) and all characteristics (e.g., small size or large size). For example, for size, the
paper clip object was perceived as the smallest object and the
umbrella object as the largest object out of all 100 objects by Klatzky et al. [
22]. Table
1 shows the identified objects at both ends of the individual continua, including the means and the SDs from participants’ ratings.
Overall, all participants rated their expected quality of a specific haptic attribute consistently in the online survey. In particular, while the visually displayed object sizes in the 2D copies are not in a realistic proportion, all size ratings across objects and participants show a low variability among the individual SDs with an average of 0.09 (SD = 0.06). This indicates that participants did not rate an object’s size as visually displayed in the 2D copy, but the expected actual object size, as intended by the study task.As a next step, we then integrated the 12 objects in our haptic exploration study to investigate how people would elicit haptic sensations when interacting with them in MR. While our objects cover various object categories and properties (e.g., low temperature and high temperate) and are informed by previous user-centred object recognition research [
22], we want to emphasise that our candidates do not represent a non-exhaustive list of possible MR objects; however, we aimed at identifying objects that cover a broad spectrum of characteristics in an initial step to then investigate possible related gestures for exploration, which we study in more depth in our gesture elicitation study.
5 Results: Eliciting Gestures and Haptic Feedback
The video recordings were transcribed and analysed by performing an iterative process of open coding, using the MaxQDA data analysis software, and thematic analysis [
3]. Three researchers performed an initial round of open coding on two data sets, covering 11.76% of the overall data (2 out of 17). With an additional researcher, four researchers in total completed the remaining data sets for the construction of the design space.
Gestures.
We identified over 200 suggested individual gestures across the data sets (17 participants × at least one gesture for each of the twelve objects), from which we derived ten high-level categories of gestures (see Table 2, top). In total, these gesture categories were employed to explore a specific haptic attribute in 31 instances. While Grip + Move was employed for the highest number of min and max attribute variants, Stroke was the most versatile, as it was applied to explore all six haptic attributes. Tap, Hover, and Wearing Gesture are situated at the lower end of the frequency spectrum, each associated with only a single haptic attribute. The idea of the wearing gesture was to push the fingers through the device to achieve isolated finger placement similar to wearing a glove. A detailed overview of the individual gestures in relation to the objects is illustrated in Figure 3, demonstrating the broad and heterogeneous design space of gesture interaction with a mobile phone as a haptic interface. Nonetheless, we identified the mobile phone’s relative positioning to the virtual object as a varying factor among gesture categories. One variation shows the phone often portraying the object itself while participants manually explored areas of interest of the phone (see e.g., Pinch the paper clip for the size in Figure 3). In another variation, participants held the phone inside their palm and moved it alongside their hand around the object (see e.g., Press the balloon down for the weight in Figure 3). Lastly, in some instances, the object was explored by tapping it with the roughly grasped phone (see e.g., Grip+Move of the paperclip or umbrella for the size in Figure 3). Based on these observed positional relationships between the phone, the hand and the object, we derived three modes that the phone can assume as a crucial aspect of the gesture design: (1) object mode, (2) hand mode, and (3) tool mode (see Figure 1). Overall, during the study, participants did not comment on the two rendering layers (virtual, physical) inside MR or their virtual hands and objects occluding the tangible artefact. Participants accurately described most haptic features of the objects in the study, with only two deviations noted. The umbrella was perceived as smaller than real-life, sometimes likened to a children’s umbrella, while the paper clip was seen as larger than its real-life counterpart. However, as these size variations remained within a narrow margin and still depicted realistic versions of the objects, we believe the gesture elicitation was unaffected.
Haptic Feedback. For real-time feedback
to haptically experience the object properties during the gesture interaction, participants envisioned eleven distinct types of haptic feedback (see Table 2, bottom). These haptic outputs were suggested in a total of 24 combinations with the object properties. While no feedback mechanism was consistently applied across all six attributes, force feedback/resistance was envisioned to be the most versatile, being associated with five attributes. This was followed by passive haptics through device features and surface deformation, both of which were linked to four attributes. The remaining concepts were specific to either two or one haptic attribute. Gestures x Haptic Feedback.
An overview of the relationship between both interaction parameters, the gesture types and the corresponding haptic feedback, is presented in Table 3. Here, Grip + Move was combined by far with the most types of haptic feedback and max and min variants, reaching a total of 17 combinations. The second-highest level of versatility is demonstrated by Stroke, which encompasses nine combinations. The remaining connections among gesture category, haptic feedback and attribute variant were at or below five. As an additional visualisation, Figure 4 in the appendix depicts the data flow within this complex design space, constrained by the parameters haptic attribute, attribute variant, gesture category, and haptic feedback. In the following, we discuss in more detail the interactions with the individual haptic attributes.
5.1 Size
Participants’ design of gestures for exploring an object’s size was strongly impacted by an object’s large or small dimensions. Often, the large object was spatially "sampled" through motion gesture, and the small object was mapped onto the 2D surface of the artefact. When exploring the large object, the artefact was often used as a poking tool and moved through space to detect the object’s outline. Envisioned haptic feedback upon collision included force feedback, resisting arm movement, or vibration. A variant incorporating both hands "hugged" the object’s outline and involved force feedback at the second hand, which was not holding the artefact or feeling the umbrella’s round outline along the arms. The relative distance between both arms also provided a proprioceptive sensation for the object’s large dimensions. Another way to sense the size of the umbrella involved grasping the artefact as a substitute for the umbrella’s handle and moving it up and down. Participants then described the size sensation as the vertical resistive force coming from the air resistance. A concrete idea for implementing such vertical force feedback utilised magnetic repelling force between the grasping hand and an envisioned second magnet positioned at the top of the virtual umbrella.
In contrast to gestures heavily relying on spatial movement, participants mainly mapped the small virtual paperclip onto the artefact’s 2D surface, focusing the interaction on the exploration of the flat surface. The paperclip’s 2D outline was then examined through a pinching movement, squeezing the artefact’s front and back, guided by resistive force upon outline collision. For a more in-depth exploration, finger-pads either stroked or were pressed onto the surface to sense the size through envisioned surface deformation or additionally attached wire. Nonetheless, the paperclip’s outline was also experienced through the artefact resting inside the palm, with the size outline pressing into the palm’s skin via surface deformation on the artefact’s back. A motion gesture, similar to the exploration of the large umbrella, was also employed for the paperclip. In this context, the grip was restricted to the artefact’s corner to enable a smaller grip pose appropriate for the object while also providing passive haptic feedback from the physical feature.
5.2 Shape
Haptic knowledge about shape was mainly retrieved through gestures that allowed participants to enclose their hands or fingers around the physical shape of the artefact, serving as a substitute for the virtual object. The simple and straight shape of the toothpick was often mapped onto the edges or flat surface of the artefact. To experience the toothpick’s elongated form, participants commonly pinched either the device’s long edge from top to bottom or rotated the device by 90 degrees to pinch the short edge of the device. To capture the sensation of the toothpick’s sharp ends, a wire was suggested that would come out of both ends of the artefact’s edges. Alternatively, a deformation was described that mimics both cone-shaped ends. The toothpick’s thin diameter was investigated by pinching the round edge, providing passive haptic feedback similar to the object’s original shape. Additional rolling of both fingers then mimicked the natural rolling of the toothpick between the fingers. It was also suggested to incorporate one of the physical buttons on the side instead of the artefact’s whole edge, as the button’s shape is closer to the toothpick’s. In scenarios, in which participants mapped the toothpick onto the flat surface, the index and thumb were in contact with the artefact’s front and back side, feeling the shape through surface deformation recreating the toothpick’s thin diameter. The same surface deformation just on the front side was also explained as a meaningful simulation of the toothpick’s shape, which was then felt through stroking.
For the complex shape of the baseball glove, both hands were often used to grip the artefact from opposing sides. This approach was complemented by a desired shape-changing feature, aiming at enhancing the ability to perceive the object’s volume within both palms. Furthermore, participants applied pressure to the artefact, bending it to gain insights into the shape’s flexibility, conveyed by haptic feedback simulating the level of artefact deformation. Participants were also curious about the glove’s opening for the hand and suggested a gesture for pushing individual fingers through openings in the artefact. "Wearing" the artefact as a glove in this way would then restrict the individual fingers’ movements and convey the arched shape.
5.3 Hardness
Central concepts in the device interaction for exploring the level of hardness involved squeezing and pushing the object, with the primary gestures
Grip, Two-handed grip, Pinch and Press (see Figure
3). Grip gestures, one-handed and two-handed, squeezed the artefact’s width and aimed for haptic feedback related to soft or firm device deformation. Additionally, two-handed grip gestures were suggested to bend the artefact and convey a sense of hardness by responding with a break-resistance sensation. Another variant of the two-handed squeezing compressed the virtual object between the user’s second hand and the hand holding the artefact. Participants described a resistive force sensation that would then restrict both hands from approaching each other further and convey a level of softness. In the context of motion gestures, participants gripped the artefact again as a poking tool to push into the virtual object and sense its hardness upon collision through resistive force feedback. Here, participants exhibited different preferences for hand poses while holding the device.
Participants’ investigations of hardness on an object’s local scale were often conducted through a Pinch gesture. During the interaction, participants focused, e.g. on the concept of compressing the thin and pliable sheet of toilet paper. To convey the sheet’s flexibility and deformation, sensing the pushback of the opposing finger was highlighted as an integral component of the tactile encounter, which was complemented by material deformation and ripping sensation. The expected sensation when pressing the artefact’s flat surface with a single finger mostly targeted the feeling of a change in the device material’s softness. Furthermore, proposed gestures utilising the unique tactile qualities of the artefact involved pressing a physical button on the artefact’s side with envisioned adaptive button resistance. Additional ideas for higher hardness involved stroking the edges of the device or its camera component (considerably sticking out from the used Google Pixel 7) to perceive the tactile hardness of the co-located virtual clamp’s edge.
5.4 Texture
The texture was mainly explored through Pinch and Stroke for which the virtual object was mapped onto the artefact’s surface or other physical features. Notably, these gestures targeted a rather small area of the artefact, highlighting a more local haptic exploration.
Participants frequently engaged in stroke gestures, executing long lateral motions using their index, thumb, or multiple fingers, primarily to explore scratchy surfaces. Their expectations centred around experiencing vibration and surface deformation to convey the granularity of the sandpaper’s texture effectively. Moreover, a unique idea emerged, involving placing an index finger on the back of the device and virtually touching the object. In this scenario, the index finger would sense the device’s scratchiness upon collision. In the case of the sandpaper, Some participants used the device as a tool for scratching the object’s surface, combining grip and move gestures to create desired vibrations and force feedback. Participants also employed pinch gestures, often combined with rolling or stroking motions, to explore the texture. The rice grain was sometimes mapped onto the edge or physical button, to take advantage of their smooth surface as well as of the round shape of the edge and size of the button, which was described to be similar to the rice grain. Pinching the artefact’s flat front and back side was more common for the sandpaper, as participants then expected both sides to respond with different texture sensations, rough on top and smooth on the bottom. Moreover, some participants even suggested substituting parts of the device with small squares of fabric to provide the appropriate sensation.
5.5 Weight
The prevalent gestures for experiencing weight through vertical force feedback were grip and rocking, forming a large portion of participants’ interactions (see Figure
3). Various techniques for holding the artefact were exhibited, including gripping the device and moving it up and down, gripping and rotating the wrist, holding the thumb onto the screen or pushing a physical button to simulate grasping the string attached to a balloon and moving the device downward. Additionally, a touch-based gesture involved stroking downwards, pulling a virtual string attached to a balloon, which was mapped onto the screen. This gesture allowed participants to perceive the sensation of vertical upward force through their thumb, adding to the diversity of interactions explored.
In a contrasting approach, some participants opted for unsupported holding of the device, effectively using it as a scale and moving it up and down to engage with the vertical force feedback. Similarly, one participant rested the device on a table and placed their palm on it, expecting an upward push from the device to simulate the negative weight of the balloons. For interactions involving the hammer, participants primarily grasped the handle, focusing on sensing the imbalanced weight distribution. An interesting concept was also introduced, wherein a second hand was used to create a magnetic repelling force, pushing down on the hand holding the device to simulate a unique haptic experience.
5.6 Temperature
Participants primarily utilised Hover and Tap to sense the temperature of the virtual objects. In the context of Hover, the radiating heat of the burning candle was mapped onto the artefact and explored by hovering one or two hands above the device. We observed different variations in the artefact’s orientation. In one scenario, the device was placed on a table, emanating the heat from its front side, while in another, the device was held upright, with heat radiating from the top section to mimic the flame. An additional airflow coming out of the top complimented the haptic sensation of the hot flame.
In contrast to hovering, some participants interacted with hot and cold temperatures through contact with the entire artefact by gripping, tapping, pressing, or resting the device in the palm. For the cold key, participants projected the key’s shape onto the flat surface as the only cold area. To provide localised feedback, temperature panels were suggested that are incorporated into the artefact. Another contact-based gesture aimed to replicate the sensation of the highest temperature at the artefact’s top by pinching its upper edge. One participant suggested tapping the flashlight of the artefact as the source of the heat. The lower sections of the artefact’s surface represented the colder wax, which participants explored through stroking and press movements. Additionally, surface deformation was mentioned to simulate the varying surface tensions of wax at different temperatures.
6 Discussion
The set of user-defined gestures derived from our user study exhibited an extensive variety (see Table
3 and Figure
3), complemented by a diverse range of haptic sensations (see Table
2). Overall, all gestures can be grouped under
ten high-level type of gestures
Grip, Grip + Move, Two-handed grip, Pinch, Stroke, Tap, Press, Hover,
Unsupported holding and Wearing Gesture.
As a real-time response during the interaction, eleven types of haptic feedback were suggested. In the following, we discuss the impact of haptic attributes on gesture design, the relation between gesture and haptic feedback and explore the idea of mobile multi-haptic feedback devices.
6.1 Gesture Design and Three Interaction Modes for the Mobile Device
A key design aspect across all gestures is the variation in the positioning of the mobile device relative to the virtual object and the user’s hand. Based on our observations, we identified three distinct modes that the device can adopt during the interaction: (1) object mode, (2) hand mode, and (3) tool mode (see Figure 1). In the following, we explore each of these roles and examine the impact of object properties on these gesture variations. Object mode. In the object mode, the device acts as a physical representation of the virtual object. Here, touching the device is interpreted as touching the virtual object. The device remains stationary in front of the user, with the hand holding it. Participants described the object as either co-located with the handheld device or remaining in the augmented environment. In our study, we further observed three variations of the object mode. First, when the virtual object is smaller than the mobile device, the gesture utilises only a part of the device. For instance, when the small paper clip or cold key was mapped onto a partial area of the flat surface, the simple shape of the toothpick was aligned with the device’s edge or the rice’s texture was explored by touching the device’s similarly shaped physical button. The remaining physical components of the device do not serve as interactive elements in the artefact-gesture. Second, when the virtual object is larger than the device, the device depicts only a part of the object that is relevant to the explored haptic attribute. For example, gripping the mobile device in landscape orientation with the entire hand was associated with gripping the middle part of the clamp positioned on the side. Third, the mobile device depicts only the haptic attribute itself and does not portray the object. For example, the tapping of the mobile phone’s flashlight as a heat source of the candle. Here, the phone only conveys the temperature, but not the candle. All three variations of the object mode are not always clearly distinguishable, and sometimes, more than one can be observed. For example, one could argue that the device did not take on the clamp’s shape during the Grip gesture and, therefore, did not accurately assume the part of the object but rather the sensation of the hard material. However, the positioning of the hand mirrors the one that would be used on the virtual clamp. Therefore, we consider the device representing the clamp’s middle part to be more dominant in the gesture design than its representation of the isolated hardness. While the third variation of the object mode of isolating a specific haptic attribute might be influenced by our study task, it provides a valuable approach to simplifying a comprehensive simulation of object properties and conveying only the haptic property most relevant for the MR experience.
Hand mode. In the hand mode, the object takes a physical representation of the user’s hand. As the device is cradled in the open palm with fingers slightly curled to stabilise it, this mode can also be interpreted as extending the inside of the hand and enabling its sensing capabilities towards the virtual object. The envisioned haptic feedback mainly is restricted to the palm. During the object interaction, the device hand moves around the space, approaches the object and establishes physical contact with its outer shape. For example, the softness of the toilet paper was tested through a two-handed grip, with one hand holding the device and pressing against the paper roll. Other examples included using Grip + Move to press down the lightweight balloon or to "sample" the outer boundaries of the large umbrella and feel the collisions inside the palm.
Tool mode. In the tool mode, the device is used as a tool for indirect interaction with the object. The spatial movement of the device during interaction is similar to that in the hand mode and elicits haptic feedback upon colliding with the outer boundaries of the object. However, the device is held more with a harder grip and the haptic sensation focuses strongly on proprioceptive feedback such as force feedback/resistance in mid-air. For example, when poking the clamp with gripping + moving, the collision expressed the hardness of the clamp or the large dimensions of the umbrella.
Form factors of the device in the gesture design. In some instances, participants held the device either in portrait or landscape orientation while performing the same type of gesture, such as during the Pinch gesture to explore the size of the paper clip or to trace the shape of the toothpick along the short or long device edge. However, in these interactions, the device’s orientation appeared to be more of a personal preference rather than a deliberate design decision. Therefore, allowing different device orientations would ensure coverage for users’ individual preferences. The recent reintroduction of foldable mobile phones and tablets in the consumer market, however, will likely impact participants’ choices in gesture design. Unfolded mobile devices offer larger dimensions and an extended flat surface, likely prompting more two-handed types of gestures for larger objects. In contrast, when folded, the devices become compact and small, potentially encouraging gestures that prioritize one-handed gestures and exploration of small and fine-grained haptic properties. To understand the impact of a foldable form factor on gesture design, further research is needed.
6.2 Multi-Haptic Feedback in Mobile Devices
To cover the variety of
haptic features involved in the object’s tactile perception, participants suggested various hardware mechanisms and concepts. Overall, they can be broadly categorised into active feedback, passive haptics (also established through shape-change or appendages), arm poses for proprioceptive sensation and adding a second mobile device. Vibration as the nowadays most commonly used haptic feedback in consumer MR systems was suggested for only two attributes (size and texture). Going beyond vibration, a plethora of previous works have already explored different approaches to equip an artefact with form factors of a mobile device with enhanced haptic rendering capabilities, as reviewed in Section
2.2. As exhibited by recent work from Shultz and Harrison, a shape-changing display can already render bumps and holes [
45], and further advancements might be able to achieve a more fine-grained deformation of the device’s surface to create a local shape outline of a key or to render distinct lines representing the individual wires of a paperclip. An existing method that could create sharper elevations and edges on the small form factor of a mobile device is a pin-array [
19]. By repositioning the pins to the mobile device’s edges, users could experience the sharpness of, e.g. a virtual toothpick during a pinch gesture. Other mechanisms could simulate a change in the device’s global shape [
40] or the temperature gradient of a virtual candle through heat panels [
58]. However, incorporating this variety of hardware components will be highly complex and increase the device’s weight, which goes against the vision of everyday, mobile MR experiences. Considering scalable solutions, a flexible setup that allows users to add or remove elaborated haptic rendering capabilities could be in the form of haptic mobile cases, as done in the work by Maiero et al. [
29]. Furthermore, software-based concepts known from VR inducing pseudo-haptic effects such as the manipulation of the control-display ratio [
43,
48] or hand-redirection [
59] could be explored for haptic MR interaction. Manipulating the movement gain when moving the device up and down could convey a sense of weight, or redirecting arm movements while hugging a large object could induce a proprioceptive sensation of different sizes. However, in contrast to VR, where users only see the manipulated hands, MR displays users’ physical hands, making pseudo-haptic illusions challenging. A solution might be blink-suppressed redirecting [
60] combined with manipulating the video passthrough, to shift users’ displayed arms unnoticed during a blink. Overall, future work should explore the design of mobile devices as a multi-haptic interface to enable haptic experiences in mobile MR. While software-based pseudo-haptic approaches offer more flexibility and scalability, their application and perceptual mechanisms in MR have to be further researched.
6.3 Multi-Haptic Experiences for Mobile MR
As we singled out individual haptic attributes for eliciting exploration gestures in our study, we gained an in-depth understanding of the core hand movements and haptic sensations. However, our derived gestures and haptic renderings could be combined and integrated into a multi-haptic sensation to achieve even more realistic and comprehensive haptic experiences of virtual objects. The wish to not only sense one single haptic property was indeed expressed by a small number of participants in our study. For example, weight was reported as a desired indicator for the size of the umbrella or that simulating the metal texture for the hard surface of the clamp could improve the hardness perception. To understand meaningful and appropriate combinations of attributes, gestures and haptic feedback, we can leverage our identified relations between one type of gesture and multiple haptic attributes, illustrated in Table
3. In the following, we illustrate the use of our design space using three concrete applications.
6.3.1 Experiencing Augmented Buttons Through Physical Device Features.
Our gesture set shows, that a pinching gesture is suitable for exploring a small size, a simple shape, a hot temperature, a rough texture and both levels of hardness. In this example, the multi-feedback rendering could then include force feedback/resistance, passive haptics through device features, surface deformation and contact temperature. A potential use case scenario might focus on expanding the device’s interaction space, similar to previous work utilising the space around a smartwatch for off-device gestural interaction [
23]. For example, the device’s edges and corners could be augmented and extended with virtual 3D buttons or sliders. This would allow the experience of haptic button sensations, varying in resistance or temperature, to communicate the significance of an action associated with the button. This concept also facilitates new ways to engage with digital content, departing from the device’s conventional means of interaction.
6.3.2 Experiencing Virtual Objects Resting on the Palm.
An unsupported holding gesture (see Figure
3) could allow to simulate a small and light object with low temperatures (size-min, weight-max and temperature-min relating to Unsupported holding in Table
3) via force feedback/resistance, surface deformation, and contact temperature (types of haptic feedback relating to Unsupported holding in Table
3). This haptic description is, e.g., applicable for handling smaller fruits or vegetables including tomatoes or apples, which might be encountered in MR grocery shopping experiences [
46].
6.3.3 Grasping Large Objects with Two Hands.
Larger objects whose size is explored through a Two-handed grip could exhibit a simple or complex shape and soft or hard characteristics when applying a combined rendering of force feedback/resistance, shape-change and force feedback/resistance for the second hand. A potential application scenario might involve the haptic simulation of a virtual 3D binder, connected to the mobile device in a MR mobile office. As the virtual binder fills or empties, changes in the device’s shape offer users a tactile perception of the binder’s content volume.
Overall, when designing multi-haptic experiences, an important consideration, that was mentioned by a few participants, addresses the need to maintain a safe grip on the device during the gesture to prevent the device from slipping out of users’ hands. Lastly, a few participants expressed that they would also sometimes prefer visual feedback over haptic exploration. Depending on the use case, a visual inspection of an object’s size can already suffice and be more time-efficient than manual haptic exploration. However, other attributes, such as texture, depend more on the retrieval of haptic information through manual exploration, which might not be possible using visual inspection only.
In summary, our research and design space carries significant practical implications for the development of future mobile phones as haptic interfaces and user-defined gestures for haptic exploration in mixed reality. As new iterations of mixed reality technology continue to emerge, we advocate for future research to explore how ubiquitous devices such as mobile phones can be designed to support the prospective use case of serving as an MR controller and as physical substitutes for virtual objects. In particular, variations of a phone’s outer shape need to be explored to cover a wider spectrum of object features, haptic attributes and gestures. While the current slim design of a mobile phone has already provided meaningful passive haptic feedback in some instances in our study, e.g. edges depicting the long shape of the virtual toothpick, it falls short for the majority of our explored cases. For example, the slim shape can not appropriately convey thicker objects. We believe a shape-changing form factor would allow the device to assume, at least partially, crucial parts of the object for a better haptic experience. For example, a growing edge could portray the thick handle of the umbrella or the clamp. A sharpening of the phone’s corner via a pin array would allow passive haptic feedback for sharper objects, such as the virtual toothpick. These features can provide a physical approximation of the virtual object’s shape and facilitate indirect interaction. However, the exploration of new and evolving form factors should also consider how to maintain the traditional use of the phone and ensure its balance with serving as an MR controller. Elaborated shape-changing abilities might allow users to switch between both use cases. The technical feasibility of this concept is suggested by the recent reintroduction of flip phones and the release of bendable phones on the consumer market which hold a promising outlook for a flexible shape of a future MR phone. Further, the built-in technology for providing real-time haptic feedback needs to be extended beyond the currently available vibrotactile motors and include components that allow the simulation of haptic sensations such as temperature or weight shift. While we recognise that utilising everyday objects, like mobile devices, for substitutional haptic explorations may limit the range of interactions that are possible compared to a dedicated MR controller, our findings demonstrate that everyday life objects can enable users to perceive and explore approximations of the properties of various virtual objects with different user-preferences depending on the object’s characteristics (Figure 3). Recent advancements, such as Apple’s Vision Pro, suggest that specialised MR controllers may become obsolete in future mixed reality experiences. This necessitates a re-evaluation by researchers and practitioners on how users can effectively explore, perceive, and interact with virtual objects, a direction to which our exploration of mobile devices as haptic interfaces provides strong fundamental work. 6.4 Limitations
There are a few limitations that are worth discussing. First, we investigated the use of a mobile phone as an initial important step towards haptic mobile MR interactions. We did not explore other form factors of mobile devices, such as tablets or smartwatches, which might impact the design of user-centred gestures and add to our design space. While we acknowledge that our design space for haptic mobile MR interactions is not exhaustive, we do believe it provides a comprehensive and detailed overview of gestures that are employed by users when using their smartphones for object exploration in MR. Second, we explored individual haptic attributes to construct an initial design space of user-centred gestures in MR. Exploring multiple haptic attributes simultaneously might reveal additional gestures, that can further broaden our design space in Figure
3. Nonetheless, as our findings show overlaps among the gestures and targeted object properties, our findings show a multi-attribute coverage of the gesture types.
Furthermore, with future applications increasingly taking place in shared public spaces, making mid-air gestures less preferred by users [32, 34, 51], social acceptance becomes a critical consideration in designing haptic MR interactions for mobile contexts [13]. While our participants expressed that the presence of others around them in the semi-public study setup did not impact their gesture design, future studies should explore this aspect in a more controlled setting. Finally, all our participants in the gesture elicitation study were right-handed; therefore, gestures performed by users with their left hand being the dominant hand are not covered in our design space. While we believe this does not significantly impact the design space per se, it remains an unexplored area in our research that surely deserves attention in the future, including people with mixed handedness.
7 Conclusion
This paper investigates the use of mobile devices as replacements for specialised mixed reality (MR) controllers when exploring virtual objects in MR. In a 30-minute online survey on Prolific (n = 50) and a 1-hour gesture elicitation study (n = 18), we gained a deeper understanding of the tangible and haptic interactions users employ when exploring the size, shape, texture, hardness, temperature, and weight of virtual objects in MR. Our results revealed three core haptic exploration modes a mobile device takes on when used to understand MR object characteristics and properties: using the device as a substitutional MR object, as a replacement for hands or fingers, and as an additional tool for interaction, similar to how traditional MR controllers are used. Furthermore, we constructed and contributed a design space that maps over 200 gestures (17 participants × at least one gesture for each of the twelve objects) to ten high-level artefact-gestures, facilitating the simulation of sensation, interaction, and expected output haptics. In conclusion, we hope that our design space of haptic interactions and haptic feedback modalities inspires and guides researchers and practitioners when designing future mobile MR experiences that incorporate interactions with, and explorations of, virtual objects in MR.
Acknowledgments
This research benefited from the support of multiple funding bodies. Specifically, the German Research Foundation (DFG) under Germany’s Excellence Strategy (EXC 2077, University of Bremen) and the Swedish Research Council - 2022-03196. This research was also made possible through the International Postdoctoral Fellowships (IPF) of the University of St. Gallen. Furthermore, this work was conducted within the framework of the project ’Illusionary Surface Interfaces’, which is funded by the Deutsche Forschungsgemeinschaft (DFG, German Research Foundation) – 425869442 as part of the Priority Program SPP2199 Scalable Interaction Paradigms for Pervasive Computing Environments. We also thank SIGCHI and the internal funding from the University of St. Gallen for their instrumental role in realising the CHI Together event in St. Gallen. Moreover, we want to extend our appreciation to Luca-Maxim Meinhardt, Thomas Mildner and Ava Scott for their support in the development of the figures. Finally, we thank the reviewers for their thoughtful feedback and our participants for their valuable input.
In this paper, we used Overleaf’s built-in spell checker and the current version of ChatGPT (GPT 3.5). These tools helped us fix spelling mistakes and get suggestions to improve the writing of the paper. If not noted otherwise in a specific section, these tools were not used in other forms.