Academia.eduAcademia.edu

The sound of touch

2013, Proceedings of the 2013 ACM international conference on Interactive tabletops and surfaces

Recent work has shown that the body provides an interesting interaction platform. We propose a novel sensing technique based on transdermal low-frequency ultrasound propagation. This technique enables pressure-aware continuous touch sensing as well as arm-grasping hand gestures on the human body. We describe the phenomena we leverage as well as the system that produces ultrasound signals on one part of the body and measures this signal on another. The measured signal varies according to the measurement location, forming distinctive propagation profiles which are useful to infer on-body touch locations and on-body gestures. We also report on a series of experimental studies with 20 participants that characterize the signal, and show robust touch and gesture classification along the forearm.

The Sound of Touch: On-body Touch and Gesture Sensing Based on Transdermal Ultrasound Propagation Adiyan Mujibiya1,2, Xiang Cao1,4, Desney S. Tan1, Dan Morris1, Shwetak N. Patel1,3, Jun Rekimoto2 1 Microsoft Research Beijing, China; Redmond, USA {desney,dan}@microsoft.com 2 The University of Tokyo Tokyo, Japan {adiyan, rekimoto}@acm.org 3 University of Washington Seattle, USA [email protected] 4 Lenovo Research & Technology Beijing, China [email protected] ABSTRACT Recent work has shown that the body provides an interesting interaction platform. We propose a novel sensing technique based on transdermal low-frequency ultrasound propagation. This technique enables pressure-aware continuous touch sensing as well as arm-grasping hand gestures on the human body. We describe the phenomena we leverage as well as the system that produces ultrasound signals on one part of the body and measures this signal on another. The measured signal varies according to the measurement location, forming distinctive propagation profiles which are useful to infer on-body touch locations and on-body gestures. We also report on a series of experimental studies with 20 participants that characterize the signal, and show robust touch and gesture classification along the forearm. Author Keywords On-body sensing; Gestures; Skin; Ultrasound propagation; ACM Classification Keywords H.5.2. Information interfaces and presentation (e.g., HCI): User Interfaces - Input devices and strategies; INTRODUCTION We propose a novel on-body touch and gesture sensing technology based on active ultrasound signal transmission. It provides rich contextual information of on-body touch and gesture such as: 1) continuous indication of touch or gesture presence, 2) continuous localization of touch (potentially supporting slider-like interaction), 3) pressuresensitive touch (potentially supporting a touch-and-click event), and 4) an arm-grasping hand gesture (Figure 1). This paper highlights an active system with transducers that resonate the skin surface with low-frequency ultrasound and a receiver that measures the signal at some other point on the body. Significant portions of the signal actually transform into a surface wave when placed in perpendicular conPermission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]. ITS '13, October 06 - 09 2013, St Andrews, United Kingdom. Copyright 2013 ACM 978-1-4503-2271-3/13/10…$15.00. http://dx.doi.org/10.1145/2512349.2512821 Figure 1. We proposed a novel on-body touch sensing method based on transdermal ultrasound propagation, and we explore the design space using two exemplary sensor configurations: (a) A wearable transmitter/receiver pair to detect location, duration, and applied pressure. (b) One armband that combines the transmitter and receiver to detect handgrasping gestures. This method can be immediately useful in: (1) information access and user input for the visually impaired; (2, 3) user input for users with limited access to an input device. tact with human skin (Figure 2). Complex body properties such as muscle mass, geometry, and bone structures attenuate the signal in relatively distinct ways and provide reasonably good signal differentiation. We measure the received signal at a multitude of frequencies. Our proposed sensing principle is by nature able to continuously detect touch and gestures, by which we mean that – unlike previous work – we can detect both the onset and offset of a touch event. We can also continuously sense the location of a touch event, by measuring the signal amplitude, which degrades as the function of the distance between the transmitter and receiver. When pressing harder against the skin, the amplitudes of the measured signal dramatically increases across our transmitting frequencies (in contrast to more subtle changes related to location), therefore applying an adaptive threshold is sufficient to sense changes in pressure. To sense more complex on-body touch and gesture events such as discrete touch location sensing and hand gesture sensing, we leverage the feature-distinctive propagation profile of the measured signal, and then use machine learning classification techniques. One crucial factor in instrumenting the body to add touch and gesture is to make the sensing hardware configuration as simple and as minimally intrusive as possible. Our proposed sensing method requires as few as two ultrasound transducers acting as transmitter and receiver to be in perpendicular contact with the skin. Furthermore, both of the transducers can be implemented with low-profile sound cards already existing in popular handheld devices, such as smart phones and music players. The sensing unit is also safe, low-power, and inexpensive in both hardware and software cost. The sensing technology presented is effective for various body parts. Users may attach the designated active signal transmitter and the receiver on their area of interest to add touch and gesture sensitivity. In this work, we focused our sensing experiment on the forearm, considering user convenience and affordance of the supported interaction modalities. However, our signal propagation experiments suggest that the sensing method could be extended to various body parts. Hence the potential input space is more diverse than what could be achieved with e-textiles, finger tap body acoustics [6], or capacitive sensing [14]. Despite past research exploring feasibility of using ultrasound range sensors propagating through air and hitting the forearm [7], we are not aware of previous work exploring on-body propagation of the signal for touch and gesture interaction. The specific contributions of this paper are: 1) Proposal of novel on-body touch and gesture sensing, using ultrasound signal transmission and acquisition. 2) Description of a wearable system to demonstrate the usability and reliability of the proposed method. 3) Presentation of results from a series of experiments aimed at profiling on-body ultrasound wave propagation as well as exploring the capabilities, robustness, and limitations of the system. 4) Description of the interaction design supported using the proposed method. We also note on-body sensormounting options for specific interaction modality. RELATED WORK Recently, multiple researchers have explored using wearable sensors that measure signals passing through the human body to turn it into the core interaction device. For example, researchers have studied approaches leveraging electromagnetic (EM) propagation on the human body to create Personal Area Networks [21] as well as their uses in HCI work [22]. Cohn et al. [2, 3] leverage power-line noise picked up by the human body acting as an antenna to recognize touched locations on uninstrumented walls. Saponas et al. [13] demonstrated the feasibility of classifying tapping and lifting gestures across all five fingers by sensing the signals of muscle activation through electromyography (EMG). This approach typically requires high-cost amplification systems and is limited to finger gestures. Our sensing method utilizes cheap ultrasound transducers and can be applied to various body parts, and senses a wider interaction area with the sensor placed on the forearm. Sato et al. proposed Touché [14], which leveraged swept frequency capacitive sensing to add touch and gesture sensitivity to an object with a single electrode. While this work also notes the feasibility of the sensing method to be deployed on the human body, the supported touch gestures are limited to body parts that meet the requirements of spatial separation between the source and sensor electrodes (such as between fingers in different hands). Therefore, more detailed interaction on the user’s forearm offering more affordable and subtle interactions are missed. Our sensing method further adds the possibility to sense continuous touch location (slider), touch pressure, and arm-grasping gestures. Point Upon Body (PUB) [7] and SonarWatch [8] proposed a method to explore human factors with respect to how users can interact with their forearms, and explored the design space using an ultrasound rangefinder to detect fingertapped positions on an arm based on the time of flight of air-propagated ultrasound. This setup is limited to sensing distances roughly along a single direction, and requires line of sight from the sensor. In contrast, our method is based on transdermal propagation that allowed us to enable touch sensing for positions surrounding the arm, as well as pressure-sensitive touch and hand-grasp gesture sensing. Takemura et al. [18] reported their early work on the usage of active signal injection (bone conduction sound at 800 Hz), but measure only elbow angle. In separate work, Harrison et al. present Skinput, a system that utilizes passive bio-acoustic sensors placed on the body to detect the presence and locations of taps on the body [6]. This system relies on the fact that these taps generate lowfrequency signals that propagate through the body and “sound” quite different depending on the composition of body parts they pass through on the way to the sensor. While it was interesting to conceive of interfaces that utilize the body as a tap surface, our work attempts to extend the capabilities of Skinput to sense temporally and spatially continuous touch points, adding gesture sensing capability, and to sense touch on a larger range of body parts. In early explorations, we found that low-frequency (between 20 and 100 kHz) ultrasound propagated quite well through the human body, but in a non-uniform manner that might allow us to differentiate touches. In fact, medical researchers have shown that low-frequency ultrasound is effective in enhancing transdermal transport of various molecules [1, 10, 11, 16, 17, 19]. Within this work, they have demonstrated that the frequencies around 40 kHz have excellent transdermal propagation with minimal heating [17]. Each of these transducers was set to a power output of 0.2 W, and we used the stereo I/O from a computer’s sound card to drive them. It would be relatively easy to embed the system into a smaller device using smaller form-factor ultrasound sensors (e.g. surface-mount devices) and substituting the signal-processing unit with a general-purpose microcontroller and analog-to-digital converter (ADC). This would allow the entire sensing unit to be embedded on a mobile/handheld device. Figure 2. The properties of ultrasound signal propagation along the skin. A transducer placed perpendicular to the skin results in surface wave propagation. Currently, we are not aware of previous attempts to exploit transdermal ultrasound propagation in designing on-body touch and gesture sensing. TRANSDERMAL ULTRASOUND PROPAGATION When ultrasound energy penetrates the body tissues, biological effects can be expected to occur if the tissues absorb the energy [20]. As illustrated in Figure 2, we leverage the surface mechanical wave of low-frequency ultrasound. Consequent with Surface Acoustic Wave (SAW) phenomenon [9], when a sound wave traveling through a medium encounters a boundary with a dissimilar medium that lies perpendicular to the direction of the wave, a portion of the wave energy will be reflected (with some portions transforming into a surface wave) and a portion will continue straight ahead (forming a shear wave). The percentage of reflection versus transmission is related to the acoustic impedances of the two materials. In the case of a boundary between aluminum (acoustic impedance of 40.6 to 45.7) and human skin (1.53 to 1.68) [12], the calculated reflection coefficient is 92%. Consequently, in our studies, we put ultrasound transducers in direct perpendicular contact with the skin. Exciting the skin with ultrasound produces a surface wave whose signal strength depends on the transmission power; more transmission power allows a larger sensing area. The non-EM nature of the signal also adds to the reliability of the sensing result compared to capacitive-sensing-based approaches such as Touché [14], which are highly affected by whether user’s body is connected to ground, or whether there are noise sources in the surrounding environment. IMPLEMENTATION Sensor Design We chose to work near 40 kHz for a number of reasons. First, it is inaudible. Second, there is very little ambient noise in the environment in this range, as energy is lost relatively rapidly as waves propagate through air. Third, these frequencies are already popular for tasks like range sensing, and many low-powered, cheap commodity transducers exist on the market. In fact, transducers with a center frequency between 25 and 60 kHz cost less than US$1 each. For our prototype, we used 14 mm diameter aluminum housing transducers with a center frequency of 40 kHz. In our approach, preventing the receiver from catching the signal propagated through the air is highly important, because it will affect the robustness of the system. Fortunately, there is less noise in the low-frequency ultrasound range; reflection and/or refraction from the transmitters are highly unlikely to disperse through the air, due to the nature of our mounting design (perpendicular skin contact). We used sealed, waterproof ultrasound transducers to provide an acoustic pathway between the transducer and the skin. This method minimizes any air coupling and adapts the contours of the probe to the skin. While we could have further enhanced this coupling by using sonic conduction gel, we chose not to as we do not believe this is practical in realworld scenarios. Based on early trials, we applied 10 V RMS driving voltage and a sound pressure level (SPL) of 20.4 dB. The transmitter’s SPL was measured at the center part of the transducer. Canada, Japan, Russia, and the International Radiation Protection Agency recommend safe levels of 110 dB SPL ceiling operation for frequencies from 25 to 50 kHz [15]. Therefore, our signal is significantly below this threshold. To sample the signal seen by the receiver, we also used the same conventional computer’s sound card, which has a sampling rate of 192 kHz at 24-bit precision. The Nyquist frequency was 96 kHz, which provided enough headroom for our designated transmitting frequency. This kind of audio codec chip has been widely used in PC motherboards, as well as high-end mobile audio players and mobile phone; therefore, deploying both the transmission and reception parts of our approach should be relatively easy. Sensor Configurations There are many sensor configurations possible using our sensing approach. In this paper, we narrowed our explorations down to two basic sensing configurations based on the usability and the contextual information we hoped to achieve. First, we formed a wearable pair, consisting of armband mounted ultrasound transmitters and a ringmounted receiver (Figure 3). We hypothesized that the spatial mapping between the transmitters and body parts with its complex geometry would exhibit variance in signal propagation, helpful in disambiguating touch location (Figure 8 shows our varied signal reading). In this configuration, user can perform interactions (as described in Figure 1a) on the body part that is attached to the transmitter with the minimally intrusive ring-mounted receiver. Figure 3. Wearable pair configuration: 4 ultrasound transmitters mounted on an armband; and a receiver mounted on a ring. We separate transmitters based on frequency. The second configuration is a combination of transmitters and receivers in a single armband (Figure 4). We aimed to recognize simple arm-grasping gestures (as illustrated in Figure 1b), while alleviating the need for finger-mounted receiver. Since these gestures displace external tissues and form different acoustic paths, we believed we could sense meaningful signals with only a single armband. Figure 4. Combination of transmitters and receivers in an armband. Using this configuration, system complexity is reduced and allows direct interaction between user’s hands. heights and body types was important so that we can appropriately generalize the results. Study 1: Measuring Signal Propagation on Body Parts The first experiment examined the profile of on-body ultrasound propagation, and to gain the profile of which body parts possess consistent signal amplitude deterioration that may further be utilized to detect spatially continuous touch gesture along the skin (e.g. potentially used as slider). Furthermore, we demonstrate effects of the applied pressure on the receiver’s signal. While the separate mounting requirement seems cumbersome, in the future, transmitters may be embedded into various accessories people are already wearing (wristwatches, sport armbands, headsets, or clothes). On the other hand, the receiver is by default embedded in a minimally-intrusive ring form factor for finger touch, but may also be embedded in other everyday devices such as pens or mobile phones for using them as a pointing tool on the body. In this study, we aimed to generally characterize transdermal ultrasound propagation; therefore we transmitted a single 40 kHz of sinusoidal ultrasound signal when the transmitter unit was placed on the following extensive body parts selection (with summaries of why each was selected): EVALUATION • We used aforementioned sensing configurations and conducted a set of controlled experiments to profile the signal propagation through various body parts, as well as to evaluate the performance and robustness of our system. Procedure • • Participants We recruited 20 participants (6 female), divided into two groups. The first group of 10 participants (2 female) completed our first three studies. They were 25 to 35 years old, 165 to 180 cm tall, and weighing between 54 and 85 kg. Participant’s Body Mass Index (BMI) ranged from 18.69 (normal-underweight) to 28.73 (obese as defined in Asia, where the work was conducted), with an average of 21.56 (SD=3.27). A second group of 10 participants (4 female) completed the 4th study of arm grasping gesture recognition. They were 21 to 28 years old, 159 to 178 cm tall, with BMI from 15.64 to 23.82. The data collection was distributed over approximately 5 days span for the first group, and 7 days span for the second group. The time span adds the real-world parameters variation, such as temperature and humidity. The diversity of • • • Forearm: anterior (body part #1) and posterior (#2). Large and convenient interaction surface. Upper arm: anterior (#3) and posterior (#4). Large and convenient (sports armband mounting strategy). Forehead (#5). Boney part of human body with relatively large area (headphone mounting strategy). Back-of-neck (#6). Collar mounting strategy. Foot: anterior (#7). Profiling uniformed boney structure. Foot: posterior (#8). Profiling firm muscular structure. We measured the signal amplitude with the finger-mounted receiver; by subsequently perform skin touch at 5 cm, 10 cm, and 15 cm away from the transmitters, as well as no skin touch (no contact between skin and receiver). For body parts where armband transmitter mounting was difficult (such as forehead and back-of-neck), we used a single transducer acting as a transmitter placed perpendicularly on the skin surface; e.g. position near the ear on backof-neck, and position above the eye on forehead. The participants were seated in a conventional chair while the experimenter measured the received signal at the desig- Figure 5. Experiment scene using wearable pair configuration (study 1, 2, and 3). nated position (transmitter and receiver placement was consistent across participants). The experimenter also controlled touch pressure application. See Figure 5 for sensor configuration and on-body sensor placement example. Figure 6. Signal propagation results for various body parts. Note the deterioration of the signal according to the distance. Error bars show standard deviation. Results We calculated the average signal power across participants for each measurement position. The results are compiled in Figure 6. We measured the propagation along the skin, with attenuation based on distance. We also compared the propagation for different body parts where fleshy parts of the body tend to propagate the signal better than boney parts. Also, body parts with firm muscular volume seem to propagate the signal better, as experiment results show that the posterior side of forearm yield significantly higher amplitudes than the associated anterior areas. Back-of-neck and forehead measurement results were particularly interesting due to the inconsistency of the correlation between distance and amplitude. In the back-of-neck, amplitude counter-intuitively increases with distance from the transmitter, while forehead exhibited slight amplitude decrease for measurement position near the center of the forehead. These phenomenon were actually consistent with the underlying physics on transdermal ultrasound propagation, where a portion of the signal will form surface wave that propagates along the skin, and another portion will form shear wave, which exhibit reflections from the muscular tissue underneath the skin, thus affect signal amplitude on circular measurement positions (i.e. neck and forehead). These results encouraged us to adopt machine learning to sense on-body touch on radial positions of the body, where the signal amplitude often do not decrease with distance from the transmitter. To check the effects of BMI in signal propagation, we divide our participants into two groups, participants with BMI over or below 23 (borderline for normal to overweight/obese). The group with higher BMI exhibited weaker signal, especially for positions such as forearm and upper arm. However, signals were present in all. As an example of the measurement results regarding to the BMI effects, we compiled forearm and upper arm signal propagation results in Figure 7. Figure 7. Signal propagation results on forearm and upper arm for participants with BMI over and below 23. Note that overall the group with higher BMI exhibited weaker signal, except posterior side of upper-arm where the effects of firm muscular volume were considered more significant than BMI effects. Error bars show standard deviation. In explorations on the effects of the applied touch pressure, we found that the signal amplitude increased considerably when pressing the receiver harder against the skin. This could potentially be leveraged as a click signal. This contextual information will further enrich the usability of skin as an interactive surface that also pressure sensitive. Study 2: Space-domain Continuous Touch Position and Touch Pressure Detection To achieve space-domain continuous touch sensing, we leverage the measured signal amplitude that generally degrades as the function of the distance from the transmitter. When pressing harder against the skin, the amplitudes of the measured signal significantly increased across all of our transmitting frequencies (in contrast to more subtle and non-uniform change in spatially continuous touch events). We further apply adaptive threshold to sense these events. Based on the results of our prior study and potential interaction application, we focused on the forearm as it showed consistent deterioration of the measured signal amplitudes, and its convenience for providing interaction affordance [6,13]. Procedure We adopted the same sensing configuration as our previous study, using wearable pair of transmitter and receiver transducers (Figure 3 & 5). However, in this study the transmitters are injecting 4 transmitting frequency (35, 40, 45, and 50 kHz of sinusoidal wave) on the user’s forearm. The usage of multiple transmitting frequencies enriches the feature set usable for inferring touch and gesture, because different frequencies propagate differently [6], thus different frequency response profiles are created at different positions (Figure 8). the system to perform interactive sliding gestures. In this study, we used the same gesture recognition frameworks as the prior study. However, instead of performing post hoc cross-validation, we mapped the following gestures to realtime interactive events: moving away from or closer to the transmitter mapped to sliding events (forward or backward respectively, where sliding velocity depends on moving speed), and pressing gesture (without slide) mapped to selection event. We designed a simple GUI of a sliding menu with 11 selectable items (as shown in Figure 11). Participants were seated in front of a computer displaying the GUI, while instructed to select 5 randomly assigned items, by performing pressing gesture to confirm selection events. In the beginning of the experiment and after each selection event, the menu was designed to slide back to middle (6th) item. We gain an average successful select rate of 92%, with an average selection time of 2.28 seconds (to select one item within 11 selectable menu). This result seems encouraging to design real-time application for on-body interactions. Study 3: Discrete Touch Position Classification Position Set for Classification Figure 8. Signal captured from a participant during (a) nogesture, (b, c, d) touch gesture on positions along the forearm gradually away from transmitter. Our signal visualization highlighted raw signal, spectrum (nearby 40 kHz), and it’s linear average. Transducer’s center frequency of 40 kHz is highlighted in red. Note the variance of the signal measured in different locations, which forms the basis of our classification approach. We selected 7 classification sets from the multitude of possible test position combinations (Figure 9). These represent logical sets of locations on the forearm, and test the limit of our sensing and classification capability. All of classification sets below include no-touch gesture. 1. 3 Positions (palm, and back-of-palm) 2. 4 Positions (3 positions along the anterior) 3. 5 Positions (4 positions surrounds the wrist) 4. 5 Positions (4 positions surrounds middle part) 5. 9 Positions (4 positions along the anterior and posterior) 6. 9 Positions (4 positions along the sides of anterior and posterior) 7. Gesture and no-gesture segmentation across all classification sets Participants were asked to perform four gestures: 1) Continuously sliding the ring-mounted receiver away from the transmitter and 2) towards the transmitter, and 3-4) are same gestures as 1 and 2 but with considerable pressure applied. Each participant performed 5 rounds of the designated 4-gesture set. Each gesture’s signal is automatically detected and segmented, with one-second-gesture window length. In a recent study conducted by Lin et al. [7], users were generally able to distinguish up to 6 points in eyes-free situations. We expect in scenarios where visual attention is allowed users can achieve higher resolution, thus we experimented with above resolutions. Our sensing method also allows radial sensing, i.e. sensing positions surrounding the arm, not limited to a single side along the arm. Results Participants were seated in a chair in front of a desktop computer showing the experiment steps and directions, as well as visualization of the signals (Figure 5). The armband was adjusted to ensure good contact between the sensor and the skin. The armband was placed on participant’s nondominant arm, and measurements were done with participant’s dominant hand. Similar to study 2, transmitter frequencies were adjusted to 35, 40, 45, and 50 kHz in sinusoidal wave, configured as wearable pair (Figure 3). Using this configuration, we were able to aggregate richer feature Averaged across participants, our real-time adaptive threshold correctly detects 98.21% (SD=4.52, chance=25%) of the aforementioned gestures. This provides evidence that by simply observing the behavior of amplitude spikes across the designated frequencies, our system is able to reliably detect sliding and pressing gestures. Interactive study: mapping for continuous gestures We conducted a study on event mapping for our spatially continuous touch sensing. We aim to clarify the usability of Procedure Figure 9. Classification accuracy for the designated 7 position sets. Labels show accuracy value, chance level, and standard deviation. Error bars indicate standard deviations. Position sets 1-4 demonstrate immediate usable accuracy for interaction, while position set 5 and 6 show the classifier’s scalability when increasing the number of positions. Position set 7 shows our system’s robustness to segment gesture and no-gesture across all the previous 6 position sets. set by leveraging the variability of the frequency response profiles created at different positions, as well as spatial separation between transmitters. To calculate the classification accuracy for various conditions, we ran five-fold cross-validations. Each fold consisted of a single round through all measurement positions in the classification set. Each was separated by several minutes to avoid the possibilities of over-fitting. We pick one round as test, and use the rest (four other rounds) as a training set. We repeat this condition for each round (fold), and we aggregate our classification results. Signal Processing and Classification To train the system, we collected 100 samples (a sample consists of the averages for every FFT bin in a 25 ms window) for each position. Samples were separated by 2 ms. Including software overhead, sample collection for each of the locations took about 3 seconds. We calculated a 2048-point FFT, resulting in 1024 spectral samples with 93.75 Hz frequency resolutions, to which we applied a Hamming window for smoothing. We generated 120 features for each sample. Our features are inspired by previous work on bio-acoustics (Skinput [6]), and we use a fairly exhaustive feature set to provide readers with a complete palette for subsequent implementations. In detail, our feature space consists of: 1. Acoustic power (84): The amplitudes of each transmitting frequency and their surrounding bands (±1 kHz). 2. Amplitude ratio between each transmitting frequency (6): Signal amplitude fluctuates when pressure applied, but the ratios between amplitudes were frequently stable. 3. Average amplitude (4): The average of acoustic power from each transmitter frequency with ±1 kHz bandwidth. This will essentially eliminate temporal fluctuations. 4. Average amplitude ratio between each transmitting frequency (6): We also considered ratios between average amplitude described above as feature. 5. Standard deviation (4): The different signal spectrum form (valley and peak) can be quantified in standard deviation between each transmitter frequency’s 21 bands. 6. Linear average (10): We calculated linear average grouped for 32 bands (Figure 4, bottom). This has essentially wider bandwidth and potentially useful for capturing harmonic signals as well as outlier removal. We included frequency range from 28.5 to 55.5 kHz 7. Log average (5): We also included log average for 28 to 72 kHz. This represents the broader band, as well as considering combined transmitting frequency amplitude and their surrounding bands as feature. 8. Zero crossings (1): We included total amount of zero crossed value of raw signal stream. We adopted the Sequential Minimal Optimization (SMO) implementation of Support Vector Machine (SVM) within the WEKA [5] for our classification. This is a supervised machine learning technique that constructs a classification hyperplane in our high dimensional feature space. Results Classification results are shown in Figure 9. Classification sets 1 to 4 demonstrate useable accuracy readily applicable for real-time interaction systems. Classification sets 5 and 6 pushed the limit for our sensing method to a usable 79.96% and 81.87% in average for all participants. Classification set 7 shows the robustness of our sensing method to segment gesture and no-gesture signal across all 6 classification sets. The inclusion of the no-touch condition in position sets 1 to 6 reflects our system’s target scenarios, where differentiating touch from no-touch is as important as position classification. To get more insights on our system’s capability to classify on-body touch gestures, we also run our classification without including no-touch gesture (i.e. only classify the actual touch locations). Averaged across all 10 participants, we aggregated slightly lower accuracy (e.g. for set 1 at 93.81% (SD=10.1%), set 2 at 98.29% (SD=2.57%), set 3 at 84.20% (SD=10.72%), set 4 at 89.8% (SD=11.54%), and set 5 at 88.55% (SD=7.37%)). Overall accuracy decrease are insignificant, therefore show promising real-world applicability of our classifier. We also examine the effects of participant’s BMI on our classification results (compiled in Figure 10). Overall, there system to classify simple arm-grasping gestures as an example of what our sensing method is able to support. Based on early trials, we found that applying significant pressure on the forearm exhibits significant fluctuations on the signals sensed by receivers that are collocated with transmitters in a single armband. Performing these gestures displace external and internal tissues, thus form different acoustic paths for our signal. To evaluate our system’s capability on sensing armgrasping gestures, we defined a gesture set consisted of five gestures: no touch, one, two, three, and four finger grasping (See Figure 1b for gesture illustration). Procedure Figure 10. Classification accuracy for the designated 7 position sets divided between participants with BMI over and below 23. Note that there is no clear indication of accuracy degradation relative to participant’s BMI. Error bars show standard deviation. is no indication of accuracy degradation based on participant’s BMI, which is encouraging. In PUB [7] the accuracies for discrete touch sensing were 84%, 66.6% and 65.7% for 5, 6 and 7 points, respectively. With our approach, we aggregate 98.75% (set 2 with 4 points) and 81.87% (set 6 with 9 points). This clearly shows the advantage of our transdermal ultrasound approach compared to previous work. Classification set 1, 3, 4, and 6 also showed the advantage of our sensing method, which enables touch sensing for positions surrounding the arm (no limitation in sensing along a single axis on the forearm). To explore our feature space, we ranked features by the square of the weight assigned by the SVM. Among the consistently best features we observed across all the classification sets were standard deviations, transmitter frequency amplitudes, and their ratios. Note that standard deviations indicate strong correlation between measurement position and the dispersion of the received signal amplitudes. We repeated our classification test for the whole sets using highly ranked features only (standard deviations (4), transmitter’s frequency amplitudes (4) and their ratios (6), average amplitude for 2 kHz bandwidth (4) and their ratios (6)); shrinking the total number of features to 24. The classification results only decreased within an average of 0.63% while the classification time were noticeably faster (1.89 times faster for training and 4.36 times faster for classification). On average, we obtain 30 classified samples per second, which is sufficient for real-time recognition. Study 4: On-body Gesture Classification The ability to sense on-body touch gestures is also important in supporting eyes-free and subtle interaction modalities. In this study, we demonstrate the capability of our The participants placed the armband of ultrasound transducers on their upper-forearm, and performed arm-grasping gestures on the forearm. The sensing configuration used was the combined one-armband transmitter and receiver (Figure 5). Participants were instructed to apply maximal pressure while maintaining comfortable feeling and avoid pain, when performing arm-grasping gestures. We inherit the procedures and techniques of our discrete touch position classification (Study 3). Results The average accuracy for 10 participants was 86.24% (SD=6.72%, chance=20%). The result seems promising considering that the classification was conducted with baseline machine learning techniques with plenty of headroom for fine-tuning the parameters. INTERACTION DESIGN & ENVISIONED APPLICATIONS Based on the studies described in previous section, the following design implications are derived: 1. Per-user training is most desirable. There are significant individual differences of how user performs touch gestures. Fortunately, the training phase of our method only took considerably short time (±3 seconds per- gesture). 2. Appropriate sensor configuration for specific interaction is crucial. We used different sensor configuration for study 1-3 and study 4. In our current setup, fingermounted receiver is required to robustly infer touch locations. However, as we discussed in study 4, different arm-grasping gestures exhibits significant variety in signal propagation profiles, therefore the requirement of finger-mounted receiver can be mitigated. Similar approach can be taken when designing interaction based on this sensing method on other body parts. 3. Leverage the sensing capability to perform legato touch sensing. Participants have the tendency to locally shift their finger to adjust the final position. Our approach’s ability to continuously sense touch can help determining the time threshold for a touch event to be finalized. 4. Sensing on fleshy and firm muscular volume body parts. Based on our study, those body parts gain better propagation, good for higher power-to-sensing area ratio. We intend to perform further studies on amplifying propagation of the low-frequency ultrasound mechanical wave for skin-to-skin contact (specially for wearable pair sensing configuration), so that users can interact with their fingers rather than with the sensing ring. Exploring the application of this sensing method for capturing rhythmic pattern of user’s gestures as input method is also feasible [4]. CONCLUSION Figure 11. We developed a proof of concept sliding menu application, which also served as our experiment platform in our interactive study (as a part of study 2). In this study, we investigated event mapping for continuous and pressureaware touch gestures. Note that in the future, the sensor unit (transmitter and receiver transducers) and the display contents can be embedded into wearable devices enabling truly mobile experience. We addressed on-body sliding menu user input in Study 2 (also presented in Figure 11). Furthermore, we envisioned two exemplary application domains in which our approach can be immediately useful: 1) supporting interaction modalities for visually impaired users (who can accurately touch their own body parts using proprioception, as illustrated by Figure 1.1), and 2) for users with limited access to an input device (illustrated by Figure 1.2 and 1.3, e.g. when driving, biking, running, etc.). Other interactive application examples uniquely enabled by this technique includes: a) Wrist mounted mobile phone will embed the sensing unit and appropriates the forearm to be an input surface, b) Headphones that simultaneously plays music and transmits ultrasound signals through the skull, and s/he may touch different parts of the head/neck/face to control the music, c) Multiple sensors placed in various body parts can be used to sense wide-body gestures. Exploring and developing these applications remains future work. DISCUSSION AND FUTURE WORK Our proposed approach has demonstrated that signal amplitude on the receiver deteriorates according to the distance from the transmitter on certain body parts such as forearm, upper arm, and foot. However, during our sample collection we noticed random fluctuation in the signal reading when sliding the receiver on the skin. This produces potential error-prone samples, greatly reducing the robustness of our classification. To address this limitation, we suggest smoothing over longer windows, so as to negate the temporal spikes. Also, in the current design of our prototype, we do not measure phase changes or effects of Doppler shift in response to user interaction. These aspects remain future work. We have demonstrated the feasibility, discussed interaction design as well as present proof-of-concept applications of a novel on-body touch and gesture sensing approach by leveraging low-frequency ultrasound wave propagation along the skin. This sensing method enables a breadth of viable applications using the human body as an input surface. We discussed the signal propagation across different body parts representing boney, fleshy, and firm-muscular tissues. By examining the variations of the signal propagation across different positions on the skin, we have shown that we can robustly classify a series of touch gestures performed at different locations on the forearm. We also demonstrated the feasibility of spatially continuous touch sensing along the skin, pressure-sensitive touch sensing, and arm-grasping gesture sensing. The sensor unit we used in our study consists of off-the-shelf components which are inexpensive in both bill of materials and computing power. Furthermore, it can be replaced with smaller form factor, allowing the whole setup including the signal-processing unit to be embedded into a mobile device. ACKNOWLEDGEMENTS We are thankful to anonymous reviewers for their thoughtful comments and subjects who participated in the experiments. REFERENCES 1. Cui, J., Wei, Y., & Wang, H. The Study of Lowfrequency Ultrasound to Enhance Transdermal Drug Delivery. Complex Medical Engineering, 2007, 1221– 1224. 2. Cohn, G., Morris, D., Patel, S. N., & Tan, D. S. Humantenna: Using the Body as an Antenna for Real-Time Whole Body Interaction. CHI ‘12, 1901–10. 3. Cohn, G., Morris, D., Patel, S. N., & Tan, D. S. Your Noise is My Command: Sensing Gestures Using the Body as an Antenna. CHI ‘11, 791–800. 4. Ghomi, E., Faure, G., Huot, S., Chapuis, O., and Beaudouin-Lafon, M. Using Rhythmic Patterns as an Input Method. CHI ‘12, 1253–1262. 5. Hall, M., Frank, E., Holmes, G., Pfahringer, B., Reutemann, P., and Witten, I. H. The Weka Data Mining Software: an Update. SIGKDD Explor. Newsl. 11, 1 (Nov. 2009), 10–18. 6. Harrison, C., Tan, D. S., & Morris, D. Skinput: Appropriating the Body as an Input Surface. CHI ‘10, 453– 462. 7. Lin, S.-Y., Su, C.-H., Cheng, K.-Y., Liang, R.-H., Kuo, T.-H., and Chen, B.-Y. PUB - Point Upon Body: Explor- ing Eyes-Free Interaction and Methods on an Arm. UIST ‘11, 481–488. 8. Liang, R.-H., Lin, S.-Y., Su, C.-H., Cheng, K.-Y., Chen, B.-Y., and Yang, D.-N. 2011. SonarWatch: appropriating the forearm as a slider bar. In SIGGRAPH Asia 2011 Emerging Technologies (SA ‘11). Article 5, 1 page. 9. Lord Rayleigh. On Waves Propagated along the Plane Surface of an Elastic Solid. Proc. London Math. Soc. s117 (1)(1885), 4–11. 10. Mitragotri, S., Blankschtein, D., and Langer, R. Transdermal Drug Delivery Using Low-frequency Sonophoresis. Pharmaceutical Research 13 (1996), 411–420. 11. Mitragotri, S., Farrell, J., Tang, H., Terahara, T., Kost, J., and Langer, R. Determination of Threshold Energy Dose for Ultrasound-induced Transdermal Drug Transport. Controlled Release 63, 1-2 (2000), 41–52. 12. Ogura, I., Kidikoro, T., Iinuma, K., Takehara, Y., Tanaka, K., and Matsuda, A. Measurement of Acoustic Impedance of Skin. Ultrasound in Medicine 4, RC 78.7, U4 A 5a (1978), 535. 13. Saponas, T. S., Tan, D. S., Morris, D., Balakrishnan, R., Turner, J., and Landay, J. A. Enabling Always-available Input with Muscle-Computer Interfaces. UIST ‘09, 167– 176. 14. Sato, M., Poupyrev, I., Harrison, C. Touché: Enhancing Touch Interaction on Humans, Screens, Liquids, and Everyday Objects. CHI ‘12, 483–492. 15. Non-Ionizing Radiation Section Bureau of Radiation and Medical Devices Department of National Health and Welfare. Guidelines for the safe use of ultrasound: Part II - Industrial and Commercial applications. Available online at: http://www.hc-sc.gc.ca/ewhsemt/pubs/radiation/safety-code_24securite/index_e.html. Accessed at April 5th, 2013. 16. Shung, K. K., Thieme, G. A., and Dunn, F. Ultrasonic Scattering in Biological Tissues. Journal of the Acoustical Society of America 94 (1993), 3033. 17. Suchkova, V., Siddiqi, F. N., Carstensen, E. L., Dalecki, D., Child, S., & Francis, C. W. Enhancement of Fibrinolysis with 40-khz Ultrasound. Circulation 98, 10, 1030–1035. 18. Takemura, K., Ito, A., Takamatsu, J., and Ogasawara, T. Active bone-conducted sound sensing for wearable interfaces. UIST ‘11 Posters. 53–54. 19. Terahara, T., Mitragotri, S., Kost, J., and Langer, R. Dependence of Low-frequency Sonophoresis on Ultrasound Parameters; Distance of the Horn and Intensity. International Journal of Pharmaceutics 235, 1-2 (2002), 35–42. 20. Wells, P. Physics of ultrasound. Ultrasonic Exposimetry. Ziskin, M. and Lewin, P., eds). 35. 21. Zimmerman, T. G. Personal Area Networks: Near-Field Intrabody Communication. IBM Systems Journal 35, 3.4, 609–617. 22. Zimmerman, T. G., Smith, J. R., Paradiso, J. A., Allport, D., and Gershenfeld, N. Applying Electric Field Sensing to Human-Computer Interfaces. CHI ‘95, 280–287.