The $-family of recognizers isn’t just about how to build better recognition algorithms, it’s also about understanding patterns and inconsistencies in how people make gestures, too. This kind of knowledge will help inform gesture interaction both in terms of developing better recognizers, and designing appropriate gesture sets. In this vein, I have had a paper accepted, along with my collaborators, Jacob O. Wobbrock and Radu-Daniel Vatavu, to the Graphics Interface 2013 conference, on characterizing patterns in people’s execution of surface gestures from existing datasets. The paper is titled “Understanding the Consistency of Users’ Pen and Finger Stroke Gesture Articulation,” and here is the abstract:
Little work has been done on understanding the articulation patterns of users’ touch and surface gestures, despite the importance of such knowledge to inform the design of gesture recognizers and gesture sets for different applications. We report a methodology to analyze user consistency in gesture production, both between-users and within-user, by employing articulation features such as stroke type, stroke direction, and stroke ordering, and by measuring variations in execution with geometric and kinematic gesture descriptors. We report results on four gesture datasets (40,305 samples of 63 gesture types by 113 users). We find a high degree of consistency within-users (.91), lower consistency between-users (.55), higher consistency for certain gestures (e.g., less geometrically complex shapes are more consistent than complex ones), and a loglinear relationship between number of strokes and consistency. We highlight implications of our results to help designers create better surface gesture interfaces informed by user behavior.
As usual, you may download the camera-ready version of our paper if you are interested. See you in Regina!
My NSF-funded MTAGIC project will appear at two workshops at the upcoming CHI 2013 conference in Paris, France! The first is the RepliCHI workshop, which is focusing on what role replication studies can play in the HCI literature. In our paper “Challenges of Replicating Empirical Studies with Children in HCI,” we are presenting a series of empirical studies that we have run with different age groups over the past 18 months, essentially replicating a similar methodology. We will specifically describe how our methodology has had to be adapted to work with very young children. This will be a two-day workshop. For more information, check out the camera-ready version of our paper; here is the abstract for a quick overview:
In this paper, we discuss the challenges of conducting a direct replication of a series of mobile device usability studies that were originally conducted with adults and older children (ages 7 to 17). The original studies were designed to investigate differences in how adults and children use mobile devices to touch targets and create surface gestures. In this paper, we report on a replication we conducted with young children (ages 5 to 7). We discuss several methodological changes that were needed to elicit the same quality of data from the replication with young children as had been obtained from the older children and adults. The insights we present are relevant to the extension of empirical studies in HCI in general to younger children.
The second workshop is the Mobile Accessibility workshop, which is focusing on how to improve the accessibility of mobile devices to users with different abilities and to users in different contexts. In our paper “Towards Designing Adaptive Touch-Based Interfaces,” we are presenting our vision of how the work we’ve been doing on MTAGIC will lead to universally accessible mobile touchscreen interaction, by highlighting some of the technical extensions we believe our work points to. Again, for more information, check out our camera-ready paper, and here is the abstract:
As the use of mobile devices by non-typical users increases, so does the need for platforms that can support the unique ways in which these special users engage with them. We posit that, by developing an understanding of patterns in input behaviors for different user groups, we can design and develop interactions that support such non-typical users. We prove this technique with children: we present findings from two empirical studies showing how interaction patterns differ among younger children, older children, and adults. These findings point to a model of how to develop touch-based interactive technologies that can adapt to users of different ages or abilities. Such adaptations will serve to better support natural interactions by user populations with distinctive needs.
If you work in the area of kids and touch + gesture interaction, or mobile device interaction in general, find a MTAGIC project member at CHI and say hi!
February was a great month over here with lots of good news coming in about conference and journal paper acceptances! The MTAGIC project will be well-represented at the upcoming CHI 2013 conference, with two workshop papers accepted on different aspects of the project (the workshops are RepliCHI and Mobile Accessibility). We’ve also heard great news that two papers about our work with kids and mobile touchscreen devices will appear at IDC 2013 and in an upcoming special issue of the Journal of Personal and Ubiquitous Computing!
In other news, my project with Jacob O. Wobbrock and Radu-Daniel Vatavu on using patterns in how people make surface gestures to inform the design of better gesture sets and gesture recognizers (e.g., the $-family of recognizers) will appear at GI 2013. And, last but not least, my side project with Leah Findlater on understanding how people with physical impairments, including children, are using mainstream mobile touchscreen devices in their daily lives will receive a ‘Best Paper Award’ at CHI 2013! This award is an honor only the top 1% of submissions receive, and we are very honored our work was selected to be among such great company.
Look for more details on each of these upcoming papers in blog posts throughout March and April, and you can already see them listed in my current CV if you are interested.