There’s one thing I’ve noticed in the 15+ years I’ve been practicing user-centered design and leading User Experience (UX) teams: one of the best ways to judge the experience of a User Experience practitioner is to assess the number and variety of the design solutions, or “tools,” they have available in their personal UX “tool belt.” Usability problems come in many shapes and sizes, and the solutions need to be equally varied – seasoned UX professionals don’t often fall into the trap of thinking just because they’ve mastered a standard set of design “hammers,” that every usability problem they see is a “nail.”
This being said, most of us User Interface (UI) designers have spent our whole careers using visual tools to solve usability problems. Some of us have branched out to audio tools too, but the other human senses are rarely considered part of our practice. However, recent research involving haptic, or touch, -based interfaces have given the UX community new insights into how tactile “tools” can be used to solve a number of emerging usability issues in situations where visual and audio interfaces may fall short. Now is the time for well-rounded UI designers to consider adding haptic “tools” to their UX “tool belts.”
Haptic interfaces are nothing new. In fact, one of the first haptic interfaces was introduced in 1829 by Louis Braille. Braille devised an alphabet of characters consisting of physically raised dots that the blind, who lack access to a visual UI, could use to read. Video games have used vibrating controls to simulate explosions and crashes for decades and, more recently, most smartphones and tablets come with tiny motors that allow applications to “vibrate” the device when a visual or auditory cue is not appropriate. These uses of haptic feedback have not been optimal – they have either required extensive training, as is the case with Braille, or were used as a simply binary (either “on” or “off” state) cues, requiring the user to reference another interface, usually visual, to get the meaning of the vibration.
In the past few years, however, cutting edge UI designers have begun to see a different, more expansive role for haptic feedback. Recent research seems to show that using haptic feedback may not always require extensive training or complementary visual and audio feedback. In a 2013 study at the University of Tampere, Finland, researchers found that haptic feedback can be used in more than a binary fashion, using a variety of vibration patterns, called “vibrotactile icons,” to convey richer information to the user and, in some cases, completely substitute for visual and auditory feedback.
In this research 24 participants were split up, with half receiving training on the meaning of three different vibration patterns; one for “speeding up,” one for “stay at their pace,” and one for “slowing down.” The other half were not trained at all. All the participants were fitted with vibration devices on both their chest and wrists. Then both the groups of participants were asked to do exercises and use the haptic stimuli to guide their pace.
The trained participants interpreted the haptic stimuli correctly 88-100% of the time, whereas the untrained participants were correct 71-83%, and the researchers concluded: “As the unprimed participants performed similarly to the primed participants, the results suggest that vibrotactile stimulation can be intuitively understood.” In other words, it’s possible “vibrotactile icons” could convey more than just an “on” or “off” state and that haptic feedback can be understood by most users “out of the box” – without training.
UI designers must now consider haptic “icons” when faced with design challenges on UIs for devices that either have or could provide vibrotactile feedback. This additional design tool is already showing up in cutting-edge product and application designs. One excellent example comes from automaker General Motors, who has created a haptic interface for their latest version of their premium Cadillac line. Dubbed the “Safety Seat Alert,” the driver’s seat generates vibrating pulse patterns on the left and/or right side of the lower bolster to alert the driver of potential dangers, such as drifting from a traffic lane or toward nearby objects while parking. Threats from the front and rear trigger pulses on both sides of the seat. “It’s akin to someone tapping on your shoulder in a crowd to get your attention,” said General Motors Active Safety Technical Fellow Raymond Kiefer. “Using the tactile sense to communicate crash threat direction provides an effective and intuitive way to cut through the clutter of visual and auditory sensory information that drivers routinely experience.”
When should designers consider using haptic icons? Here are few user challenges that may indicate their use:
What design considerations go with haptic icons? Here are some best practices to consider:
Experimenting with tactile icons is fairly easy.
For Android smartphone users, there is a “Haptic Effect Preview” app available for free from a company called Immersion, which also sells a developer’s library of code to enable haptic icons within Android apps.
For iPhone users, the latest iOS allows you to customize the haptic icon associated with each of the entries in your contact list – try setting separate vibration patterns for your boss and your spouse or significant other and see if you can use the haptic icon you created to select which call to pick up during your next business meeting!
Haptic Effect Preview App for Android
Lylykangas, J., Surakka, V., Rantala, J., and Raisamo, R. 2013. Intuitiveness of vibrotactile speed regulation cues. ACM Trans. Appl. Percept. 10, 4, Article 24 (October 2013), 15 pages
Immersion Corporation Apps from Google Play
Download Haptic Effect Preview app from Google Play
Mactrust.com: “How to Create Custom Vibration Alerts in iOS 7”; December 30, 2013
Sign up to get our Newsletter delivered straight to your inbox