1
SonarWatch: Appropriating the Forearm as a Slider Bar Rong-Hao Liang *† Shu-Yang Lin * Chao-Huai Su * Kai-Yin Cheng * Bing-Yu Chen * De-Nian Yang * National Taiwan University Academia Sinica 1 introduction Human bodies become an emerging type of human-computer in- terfaces recently. Not only because our skin is a surface that is al- ways available and highly accessible, but also the sense of how our body is configured in space allows us to accurately interact with our bodies in an eye-free manner. Hence, this input method is suitable to be applied on extending the interaction space of mobile devices [Harrison et al. 2010] or providing more degrees-of-freedom for en- hancing gaming experiences such as Kinect 1 . Nevertheless, since the additional gesture detector may be obtrusive or not so portable for users, this approach can hardly be applied in everyday life. This work presents SonarWatch, a less-obtrusive, wearable gesture detector. Unlike the device introduced by Nakatsuma et al. [2011], we equipped an ultrasonic rangefinder and a capacitive touch sensor on a wristwatch (Figure 1), thus a user’s forearm can be appropri- ated as a readily available input surface for interacting with com- puters. Several click and slide gestures are allowed to be performed by the user. To demonstrate the possible applications and the ef- fectiveness of this implicit design, three applications are provided: photo browsing, music playlist controlling and augmenting Kinect for gaming. Capacitive Touch Sensor Ultrasonic Rangefinder Micro-processor on Arduino Mini Figure 1: SonarWatch, a wristwatch equipped with an ultrasonic rangefinder and a capacitive touch sensor. Sensor outputs are re- trieved by the Arduino Mini platform 2 SonarWatch Our device consists of a low-power consumption Devantech SRF10 ultrasonic rangefinder, which provides distance readings from 6cm to 6m, and a capacitive touch sensor mounted for detecting user events. The user can touch the sensor to initiate the detection (Figure 2(a)(b)). Once releasing the finger from the sensor, the rangefinder starts detecting the gesture between the following 0.3s to 1s at 30Hz sampling rate. Noisy outputs would be classified as outliers. All sensor data is retrieved by the microprocessor and transferred to the server running on a PC for analysis. Two types of gestures are developed for users to interact with SonarWatch: Click and Slide.A Click gesture (Figure 2(a)) is rec- ognized if there is no finger detected within 20cm in the first three samplings. Once detecting a click, the sampling session would be immediately terminated for listening a Double-Click. A Slide ges- ture is recognized if fingers are detected within the range such like 1 http://www.xbox.com/kinect/ (Figure 2(b)). In this case, the sampling session would be contin- ued, and the maximum distance M and the final position P would be obtained. If M 20cm, Slide-to-Select (Figure 2(c)) is de- termined with regarding selection at P ; if M> 20cm, it is deter- mined as either Slide-Up (P> 20cm, Figure 2(d)) or Slide-Down (P 20cm, Figure 2(e)). (a) (b) (c) (e) (d) Figure 2: Available gestures. (a) Posture for a click. (b) Posture for a slide. (c) Slide-to-Select. (d) Slide-Up. (e) Slide-Down. 3 Applications We have implemented three different applications for demonstrat- ing the effectiveness of our approaches: photo browsing, playlist controlling, and augmenting Kinect for gaming. Photo browsing: Gestures can help users browsing photos more efficiently. In the photo browser, the users can Slide-Up/Down for turning pages, Slide-to-Select thumbnails for browsing photos, Single-click to select an album, and Double-click to return. Playlist controlling: While working with listening to the mu- sic, users may want to control the playlist in an eyes-free manner. Hence, the users can Single-Click to play/pause, Slide-to-Select to adjust volume, and Slide-Up/Down to switch to another song. Augmenting Kinect for gaming: Utilizing body as controller may somehow lacking haptic sensation, but we can use our propriocep- tion as a complement. We illustrate this idea by the Angry Birds 2 game. Since a user’s arm position can be captured by Kinect, she can perform a shot by moving her arm up and down to adjust the an- gle of slingshot, sliding the fingers on the arm to select the desired intensity, and removing the fingers to launch the Angry Birds. References HARRISON, C., TAN, D., AND MORRIS, D. 2010. Skinput: ap- propriating the body as an input surface. In Proc. ACM CHI ’10, 453–462. NAKATSUMA, K., SHINODA, H., MAKINO, Y., SATO, K., AND MAENO, T. 2011. Touch interface on back of the hand. In ACM SIGGRAPH 2011 E-Tech., 39. 2 http://www.angrybirds.com/

SonarWatch: Appropriating the Forearm as a Slider Barhowieliang/... · H ARRISON,C.,TAN,D.,AND M ORRIS, D. Skinput: appropriat-ing the body as an input surface. In Proc. ACM CHI Õ10,

  • Upload
    others

  • View
    1

  • Download
    0

Embed Size (px)

Citation preview

Page 1: SonarWatch: Appropriating the Forearm as a Slider Barhowieliang/... · H ARRISON,C.,TAN,D.,AND M ORRIS, D. Skinput: appropriat-ing the body as an input surface. In Proc. ACM CHI Õ10,

SonarWatch: Appropriating the Forearm as a Slider Bar

Rong-Hao Liang∗† Shu-Yang Lin∗ Chao-Huai Su∗ Kai-Yin Cheng∗ Bing-Yu Chen∗ De-Nian Yang†∗National Taiwan University †Academia Sinica

1 introduction

Human bodies become an emerging type of human-computer in-terfaces recently. Not only because our skin is a surface that is al-ways available and highly accessible, but also the sense of how ourbody is configured in space allows us to accurately interact with ourbodies in an eye-free manner. Hence, this input method is suitableto be applied on extending the interaction space of mobile devices[Harrison et al. 2010] or providing more degrees-of-freedom for en-hancing gaming experiences such as Kinect1. Nevertheless, sincethe additional gesture detector may be obtrusive or not so portablefor users, this approach can hardly be applied in everyday life.

This work presents SonarWatch, a less-obtrusive, wearable gesturedetector. Unlike the device introduced by Nakatsuma et al. [2011],we equipped an ultrasonic rangefinder and a capacitive touch sensoron a wristwatch (Figure 1), thus a user’s forearm can be appropri-ated as a readily available input surface for interacting with com-puters. Several click and slide gestures are allowed to be performedby the user. To demonstrate the possible applications and the ef-fectiveness of this implicit design, three applications are provided:photo browsing, music playlist controlling and augmenting Kinectfor gaming.

SonarWatch: Appropriating the Forearm as a Slider Bar

Rong-Hao Liang1,2 Shu-Yang Lin1 Zhao-Huai Su1 Kai-Yin Cheng1 Bing-Yu Chen1 De-Nian Yang2

1Graduate Institute of Networking and Multimedia, National Taiwan University2Institute of Information Science, Academia Sinica

1 introduction

Human bodies become an emerging type of human-computer in-terfaces recently. Not only because our skin is a surface that isalways available and highly accessible, the sense of how our bodyis configured in space also allows us to accurately interact with ourbodies in an eye-free manner. Hence, this method is suitable to beapplied on extending the interaction space of mobile devices [Har-rison et al. ] or providing more degrees-of-freedom for enhancinggaming experiences1. Nevertheless, the additional gesture detectormay be obtrusive or hard to be portable for users in daily life.

Based on this insight, we’re motivated to design SonarWatch, awristwatch equipped with an ultrasonic rangefinder and a capacitivetouch sensor, to appropriate an user’s forearm as a readily availableinput interface. By wearing the SonarWatch, a user can slide on anarm or tap on the watch for performing corresponding operationswith passive haptic sensation on the skin. Three applications, photobrowsing, music playlist controlling and gaming, are then devel-oped to demonstrate the effectiveness of using SonarWatch in ourdaily life. We believe this more implicit design can facilitate body-sensing based interaction techniques.

2 System Overview

Our system consists of two components: SonarWatch and its server.Following we discuss the implementation details and gesture recog-nition of this two parts.

2.1 Implementation

SonarWatch consists of a low-power consumption DevantechSRF10 ultrasonic rangefinder, which provides distance readingsfrom 6cm to 6m, and a capacitive touch sensor are mounted fordetecting user events. Sensor outputs are retrieved by connectingto an ATmega328 microprocessor employed by the Arduino Miniplatform, which can be also attached on the watchband. Raw datais transfered to the server part through serial connection.

The server is implemented by Java. Once a raw data string is re-ceived, it would be analyzed by the gesture recognition procedure.After a gesture event is recognized, it would be transferred to theclient application by XML sockets for further processing.

1http://www.xbox.com/kinect/

Capacitive Touch Sensor

Ultrasonic Rangefinder

Micro-processoron Arduino Mini

Figure 1: SonarWatch, a wristwatch equipped with an ultrasonicrangefinder and a capacitive touch sensor. Sensor outputs are re-trieved by the Arduino Mini platform

2.2 Gesture recognition

Our gesture recognition procedure is illustrated as Figure 2. Whenuser touch or pinch the capacitive touch sensor, the sensor wouldoutput in high, which can be regarded as the “ready” signal.Once the capacitive touch sensor is released and the ultrasonicrangefinder reveals that no finger is within the 20cm range of de-tection, a Click gesture (Figure 3(a)(b)) can be immediately de-termined. Otherwise, the ultrasonic rangefinder would start sam-pling in 15Hz in 1 second, and user is allowed to perform theirgesture in this period. In the end of sampling, a Slide gesturecan be determined. Advanced Click gestures such as Single-clickand Double-click; Slide gestures such as Slide-Up, Slide-Down,and Slide-to-Select can also be determined by simple classification(Figure 3(c)(d)(e)) on sensor values.

Figure 2: The gesture recognition procedure for processing rawdata to gestural events in the gesture recognition server.

Figure 1: SonarWatch, a wristwatch equipped with an ultrasonicrangefinder and a capacitive touch sensor. Sensor outputs are re-trieved by the Arduino Mini platform

2 SonarWatch

Our device consists of a low-power consumption Devantech SRF10ultrasonic rangefinder, which provides distance readings from 6cmto 6m, and a capacitive touch sensor mounted for detecting userevents. The user can touch the sensor to initiate the detection(Figure 2(a)(b)). Once releasing the finger from the sensor, therangefinder starts detecting the gesture between the following 0.3sto 1s at 30Hz sampling rate. Noisy outputs would be classifiedas outliers. All sensor data is retrieved by the microprocessor andtransferred to the server running on a PC for analysis.

Two types of gestures are developed for users to interact withSonarWatch: Click and Slide. A Click gesture (Figure 2(a)) is rec-ognized if there is no finger detected within 20cm in the first threesamplings. Once detecting a click, the sampling session would beimmediately terminated for listening a Double-Click. A Slide ges-ture is recognized if fingers are detected within the range such like

1http://www.xbox.com/kinect/

(Figure 2(b)). In this case, the sampling session would be contin-ued, and the maximum distance M and the final position P wouldbe obtained. If M ≤ 20cm, Slide-to-Select (Figure 2(c)) is de-termined with regarding selection at P ; if M > 20cm, it is deter-mined as either Slide-Up (P > 20cm, Figure 2(d)) or Slide-Down(P ≤ 20cm, Figure 2(e)).

Figure 3: Gestures supported by SonarWatch. (a) and (b) are Clickgesture. (c) and (d) are Slide-to-Select gesture. (c)and (e) are Slide-Up gesture.

3 Application

We implement three different applications for demonstrating the ef-fectiveness of our approaches: photo browsing, playlist controlling,and augmenting Kinect for gaming. The first two applications areimplemented by ActionScript 3.0 and Flash, and the third applica-tion is implemented by OpenCV and OpenNI2.

3.1 Photo browsing

Rich gestures can help user browse photo more efficiently. Inour photo browsing application, user can Slide-Up/Down for turn-ing pages up/down, Slide-to-Select thumbnail for browsing photos,Single-click to enter selected album, and Double-click to return.

3.2 Playlist controlling

User may want to control background application such as theirmusic player in an eyes-free manner. In playlist controlling ap-plication, user can Single-click to play/pause the music, Slide-to-Select to adjust volume, and Slide-Up/Down for switching to thenext/previous song.

3.3 Augmenting Kinect for gaming

Utilizing body as controller may somehow lack-of-haptic sensation.Hence, our proprioception can be utilized as complement. We usethe Angry Birds3 game for illustration. By using Kinect to captureuser’s arm position, user can adjust the angle of bow. For launch-ing, user can Slide-to-Select the desired intensity, then remove theirfinger to launch the Angry Birds.

2http://www.openni.org3http://www.angrybirds.com/

4 Conclusion

We present SonarWatch, a less-obtrusive, wearable gesture detec-tor. By equipping an ultrasonic rangefinder and a capacitive touchsensor on the wristwatch, user’s forearm can be appropriated as areadily available input surface for interacting with computers. Sev-eral click and slide gestures are allowed to be performed by users.To demonstrate the possible applications and the effectiveness ofthis approach, three applications: photo browsing, music playlistcontrolling and augmenting Kinect for gaming, are developed forthe SonarWatch. We believe that this work can facilitate the skin-sensing technologies in our daily life.

References

HARRISON, C., TAN, D., AND MORRIS, D. Skinput: appropriat-ing the body as an input surface. In Proc. ACM CHI ’10, 453–462.

Figure 3: Gestures supported by SonarWatch. (a) and (b) are Clickgesture. (c) and (d) are Slide-to-Select gesture. (c)and (e) are Slide-Up gesture.

3 Application

We implement three different applications for demonstrating the ef-fectiveness of our approaches: photo browsing, playlist controlling,and augmenting Kinect for gaming. The first two applications areimplemented by ActionScript 3.0 and Flash, and the third applica-tion is implemented by OpenCV and OpenNI2.

3.1 Photo browsing

Rich gestures can help user browse photo more efficiently. Inour photo browsing application, user can Slide-Up/Down for turn-ing pages up/down, Slide-to-Select thumbnail for browsing photos,Single-click to enter selected album, and Double-click to return.

3.2 Playlist controlling

User may want to control background application such as theirmusic player in an eyes-free manner. In playlist controlling ap-plication, user can Single-click to play/pause the music, Slide-to-Select to adjust volume, and Slide-Up/Down for switching to thenext/previous song.

3.3 Augmenting Kinect for gaming

Utilizing body as controller may somehow lack-of-haptic sensation.Hence, our proprioception can be utilized as complement. We usethe Angry Birds3 game for illustration. By using Kinect to captureuser’s arm position, user can adjust the angle of bow. For launch-ing, user can Slide-to-Select the desired intensity, then remove theirfinger to launch the Angry Birds.

2http://www.openni.org3http://www.angrybirds.com/

4 Conclusion

We present SonarWatch, a less-obtrusive, wearable gesture detec-tor. By equipping an ultrasonic rangefinder and a capacitive touchsensor on the wristwatch, user’s forearm can be appropriated as areadily available input surface for interacting with computers. Sev-eral click and slide gestures are allowed to be performed by users.To demonstrate the possible applications and the effectiveness ofthis approach, three applications: photo browsing, music playlistcontrolling and augmenting Kinect for gaming, are developed forthe SonarWatch. We believe that this work can facilitate the skin-sensing technologies in our daily life.

References

HARRISON, C., TAN, D., AND MORRIS, D. Skinput: appropriat-ing the body as an input surface. In Proc. ACM CHI ’10, 453–462.

Figure 3: Gestures supported by SonarWatch. (a) and (b) are Clickgesture. (c) and (d) are Slide-to-Select gesture. (c)and (e) are Slide-Up gesture.

3 Application

We implement three different applications for demonstrating the ef-fectiveness of our approaches: photo browsing, playlist controlling,and augmenting Kinect for gaming. The first two applications areimplemented by ActionScript 3.0 and Flash, and the third applica-tion is implemented by OpenCV and OpenNI2.

3.1 Photo browsing

Rich gestures can help user browse photo more efficiently. Inour photo browsing application, user can Slide-Up/Down for turn-ing pages up/down, Slide-to-Select thumbnail for browsing photos,Single-click to enter selected album, and Double-click to return.

3.2 Playlist controlling

User may want to control background application such as theirmusic player in an eyes-free manner. In playlist controlling ap-plication, user can Single-click to play/pause the music, Slide-to-Select to adjust volume, and Slide-Up/Down for switching to thenext/previous song.

3.3 Augmenting Kinect for gaming

Utilizing body as controller may somehow lack-of-haptic sensation.Hence, our proprioception can be utilized as complement. We usethe Angry Birds3 game for illustration. By using Kinect to captureuser’s arm position, user can adjust the angle of bow. For launch-ing, user can Slide-to-Select the desired intensity, then remove theirfinger to launch the Angry Birds.

2http://www.openni.org3http://www.angrybirds.com/

4 Conclusion

We present SonarWatch, a less-obtrusive, wearable gesture detec-tor. By equipping an ultrasonic rangefinder and a capacitive touchsensor on the wristwatch, user’s forearm can be appropriated as areadily available input surface for interacting with computers. Sev-eral click and slide gestures are allowed to be performed by users.To demonstrate the possible applications and the effectiveness ofthis approach, three applications: photo browsing, music playlistcontrolling and augmenting Kinect for gaming, are developed forthe SonarWatch. We believe that this work can facilitate the skin-sensing technologies in our daily life.

References

HARRISON, C., TAN, D., AND MORRIS, D. Skinput: appropriat-ing the body as an input surface. In Proc. ACM CHI ’10, 453–462.

Figure 3: Gestures supported by SonarWatch. (a) and (b) are Clickgesture. (c) and (d) are Slide-to-Select gesture. (c)and (e) are Slide-Up gesture.

3 Application

We implement three different applications for demonstrating the ef-fectiveness of our approaches: photo browsing, playlist controlling,and augmenting Kinect for gaming. The first two applications areimplemented by ActionScript 3.0 and Flash, and the third applica-tion is implemented by OpenCV and OpenNI2.

3.1 Photo browsing

Rich gestures can help user browse photo more efficiently. Inour photo browsing application, user can Slide-Up/Down for turn-ing pages up/down, Slide-to-Select thumbnail for browsing photos,Single-click to enter selected album, and Double-click to return.

3.2 Playlist controlling

User may want to control background application such as theirmusic player in an eyes-free manner. In playlist controlling ap-plication, user can Single-click to play/pause the music, Slide-to-Select to adjust volume, and Slide-Up/Down for switching to thenext/previous song.

3.3 Augmenting Kinect for gaming

Utilizing body as controller may somehow lack-of-haptic sensation.Hence, our proprioception can be utilized as complement. We usethe Angry Birds3 game for illustration. By using Kinect to captureuser’s arm position, user can adjust the angle of bow. For launch-ing, user can Slide-to-Select the desired intensity, then remove theirfinger to launch the Angry Birds.

2http://www.openni.org3http://www.angrybirds.com/

4 Conclusion

We present SonarWatch, a less-obtrusive, wearable gesture detec-tor. By equipping an ultrasonic rangefinder and a capacitive touchsensor on the wristwatch, user’s forearm can be appropriated as areadily available input surface for interacting with computers. Sev-eral click and slide gestures are allowed to be performed by users.To demonstrate the possible applications and the effectiveness ofthis approach, three applications: photo browsing, music playlistcontrolling and augmenting Kinect for gaming, are developed forthe SonarWatch. We believe that this work can facilitate the skin-sensing technologies in our daily life.

References

HARRISON, C., TAN, D., AND MORRIS, D. Skinput: appropriat-ing the body as an input surface. In Proc. ACM CHI ’10, 453–462.

Figure 3: Gestures supported by SonarWatch. (a) and (b) are Clickgesture. (c) and (d) are Slide-to-Select gesture. (c)and (e) are Slide-Up gesture.

3 Application

We implement three different applications for demonstrating the ef-fectiveness of our approaches: photo browsing, playlist controlling,and augmenting Kinect for gaming. The first two applications areimplemented by ActionScript 3.0 and Flash, and the third applica-tion is implemented by OpenCV and OpenNI2.

3.1 Photo browsing

Rich gestures can help user browse photo more efficiently. Inour photo browsing application, user can Slide-Up/Down for turn-ing pages up/down, Slide-to-Select thumbnail for browsing photos,Single-click to enter selected album, and Double-click to return.

3.2 Playlist controlling

User may want to control background application such as theirmusic player in an eyes-free manner. In playlist controlling ap-plication, user can Single-click to play/pause the music, Slide-to-Select to adjust volume, and Slide-Up/Down for switching to thenext/previous song.

3.3 Augmenting Kinect for gaming

Utilizing body as controller may somehow lack-of-haptic sensation.Hence, our proprioception can be utilized as complement. We usethe Angry Birds3 game for illustration. By using Kinect to captureuser’s arm position, user can adjust the angle of bow. For launch-ing, user can Slide-to-Select the desired intensity, then remove theirfinger to launch the Angry Birds.

2http://www.openni.org3http://www.angrybirds.com/

4 Conclusion

We present SonarWatch, a less-obtrusive, wearable gesture detec-tor. By equipping an ultrasonic rangefinder and a capacitive touchsensor on the wristwatch, user’s forearm can be appropriated as areadily available input surface for interacting with computers. Sev-eral click and slide gestures are allowed to be performed by users.To demonstrate the possible applications and the effectiveness ofthis approach, three applications: photo browsing, music playlistcontrolling and augmenting Kinect for gaming, are developed forthe SonarWatch. We believe that this work can facilitate the skin-sensing technologies in our daily life.

References

HARRISON, C., TAN, D., AND MORRIS, D. Skinput: appropriat-ing the body as an input surface. In Proc. ACM CHI ’10, 453–462.

Figure 3: Gestures supported by SonarWatch. (a) and (b) are Clickgesture. (c) and (d) are Slide-to-Select gesture. (c)and (e) are Slide-Up gesture.

3 Application

We implement three different applications for demonstrating the ef-fectiveness of our approaches: photo browsing, playlist controlling,and augmenting Kinect for gaming. The first two applications areimplemented by ActionScript 3.0 and Flash, and the third applica-tion is implemented by OpenCV and OpenNI2.

3.1 Photo browsing

Rich gestures can help user browse photo more efficiently. Inour photo browsing application, user can Slide-Up/Down for turn-ing pages up/down, Slide-to-Select thumbnail for browsing photos,Single-click to enter selected album, and Double-click to return.

3.2 Playlist controlling

User may want to control background application such as theirmusic player in an eyes-free manner. In playlist controlling ap-plication, user can Single-click to play/pause the music, Slide-to-Select to adjust volume, and Slide-Up/Down for switching to thenext/previous song.

3.3 Augmenting Kinect for gaming

Utilizing body as controller may somehow lack-of-haptic sensation.Hence, our proprioception can be utilized as complement. We usethe Angry Birds3 game for illustration. By using Kinect to captureuser’s arm position, user can adjust the angle of bow. For launch-ing, user can Slide-to-Select the desired intensity, then remove theirfinger to launch the Angry Birds.

2http://www.openni.org3http://www.angrybirds.com/

4 Conclusion

We present SonarWatch, a less-obtrusive, wearable gesture detec-tor. By equipping an ultrasonic rangefinder and a capacitive touchsensor on the wristwatch, user’s forearm can be appropriated as areadily available input surface for interacting with computers. Sev-eral click and slide gestures are allowed to be performed by users.To demonstrate the possible applications and the effectiveness ofthis approach, three applications: photo browsing, music playlistcontrolling and augmenting Kinect for gaming, are developed forthe SonarWatch. We believe that this work can facilitate the skin-sensing technologies in our daily life.

References

HARRISON, C., TAN, D., AND MORRIS, D. Skinput: appropriat-ing the body as an input surface. In Proc. ACM CHI ’10, 453–462.

(a)

(b) (c)

(e)(d)

11年10月14日星期五

Figure 2: Available gestures. (a) Posture for a click. (b) Posturefor a slide. (c) Slide-to-Select. (d) Slide-Up. (e) Slide-Down.

3 Applications

We have implemented three different applications for demonstrat-ing the effectiveness of our approaches: photo browsing, playlistcontrolling, and augmenting Kinect for gaming.

Photo browsing: Gestures can help users browsing photos moreefficiently. In the photo browser, the users can Slide-Up/Downfor turning pages, Slide-to-Select thumbnails for browsing photos,Single-click to select an album, and Double-click to return.

Playlist controlling: While working with listening to the mu-sic, users may want to control the playlist in an eyes-free manner.Hence, the users can Single-Click to play/pause, Slide-to-Select toadjust volume, and Slide-Up/Down to switch to another song.

Augmenting Kinect for gaming: Utilizing body as controller maysomehow lacking haptic sensation, but we can use our propriocep-tion as a complement. We illustrate this idea by the Angry Birds2

game. Since a user’s arm position can be captured by Kinect, shecan perform a shot by moving her arm up and down to adjust the an-gle of slingshot, sliding the fingers on the arm to select the desiredintensity, and removing the fingers to launch the Angry Birds.

References

HARRISON, C., TAN, D., AND MORRIS, D. 2010. Skinput: ap-propriating the body as an input surface. In Proc. ACM CHI ’10,453–462.

NAKATSUMA, K., SHINODA, H., MAKINO, Y., SATO, K., ANDMAENO, T. 2011. Touch interface on back of the hand. In ACMSIGGRAPH 2011 E-Tech., 39.

2http://www.angrybirds.com/