LONDON – As a student at the University of Bristol, Tom Carter became obsessed with a seemingly impossible notion: letting people feel and manipulate virtual objects. A professor turned him on to an esoteric concept — using ultrasound to simulate tactile sensations — that was floated and abandoned back in the 1970s.
Today, Carter, 29, is co-founder and chief technology officer of Ultrahaptics, which uses clever algorithms and an array of ultrasound emitters to simulate a range of feelings: tiny bubbles bursting on your fingertips, a stream of liquid passing over your hand, the outlines of three-dimensional shapes. Carter says dozens of companies making everything from computer games to cars and appliances are testing the technology, an example of what computer scientists call haptic feedback (haptic in Ancient Greek basically meant “coming into contact with something”).
While other tech companies are working on their own versions of haptic feedback, Ultrahaptics says it is the only one that lets people feel and manipulate virtual objects in the air. There are many potential applications — using an invisible slider to pump up the bass on a home stereo, adjusting the car air conditioning with the twist of a virtual dial — but Carter said the technology’s greatest promise may lie in making virtual reality “feel” more real.
“Touch is a really essential sense to make compelling virtual reality,” Carter said. “If you go and get the best possible virtual-reality goggles and the best possible surround-sound headphones, it is going to be very cool, but it is going to be very difficult if you don’t have the sense of touch to interact with the things that are there.”
Owing to nondisclosure agreements, Carter can’t say which companies are testing the technology or what they are working on. But there are tantalizing hints. Last month, high-end stereo maker Harman demonstrated a prototype audio speaker system featuring a gesture-control system powered by Ultrahaptics technology at the Consumer Electronics Show. Jaguar Land Rover has announced plans to integrate Ultrahaptics’ technology into a gesture-control system for its cars. Steve Cliffe, Ultrahaptics’ chief executive officer, says the first computer game using Ultrahaptics will be launched this year but declined to provide more details.
YouTube: Ultrahaptics in action
The technology has its limitations. While ultrasound can simulate the sensation of touching the outline of an object, it cannot create the illusion of solidity. You will always be able to push your fingers or hands through the area of vibration. Users also can hear a distracting buzz when the ultrasound waves reverberate off the skin. The company says it will find a way to minimize the noise. And while humans can’t hear the ultrasound waves themselves, dogs and cats can. Ultrahaptics says the sound hasn’t triggered a canine reaction in tests, but pledges that pets won’t be affected by whatever products come to market.
Patrick Baudisch, a professor specializing in human-computer interaction at the University of Pottsdam’s Hasso Plattner Institute in Germany, admires the research team behind Ultrahaptics. “They are one of the most innovative groups working on this in Europe,” he said. However, the inability to create the illusion of grasping a solid object or the resistance one encounters when bumping up against things in the real world might limit the technology’s use in virtual reality, he said. Instead, Baudisch sees greater applications in so-called augmented reality — when a user needs to interact with both real world and virtual objects simultaneously.
Ultrahaptics’ offices are in The Engine Shed — a sleek, postmodern incubator housed inside what was once a train shed for the Bristol railway station built by Isambard Kingdom Brunel, a 19th-century disrupter known for his pioneering work on trains, steamships and bridges. Sitting in a conference room, Carter frankly explains that the hardware — an array of ultrasound emitters hooked up to an off-the-shelf gesture-control platform — is nothing special.
“The clever bit is in our software,” said Carter, who is lanky and boyish, with a mop of brown hair that wouldn’t look out of place on George Harrison circa 1964. “It is actually in the algorithms of how you drive the emitters to create the sensations.”
The software can be programmed to find your hand and direct sound to it, projecting sensations onto a spot as small as a fingertip from as much as six feet away. It can also create multiple sensations simultaneously — allowing different hands, for instance, to “touch” different things.
Founded in 2013, Ultrahaptics grew out of research Carter began as a student at the University of Bristol. There he worked under the supervision of computer-science professor Sriram Subramanian, who ran a lab devoted to improving human-computer interaction. Subramanian, who has since moved to the University of Sussex, said the idea of using ultrasound to simulate touch was floated in the 1970s, but no one could figure out the programming. “I knew this could be an interesting idea,” he said. “But before Tom came along, I didn’t really imagine anyone would want to take this on.”
Five years of experimentation — and “a lot of really hard maths” later — Carter put his Ph.D. on hold to start Ultrahaptics. Subramaniam and another research assistant from the same lab, Benjamin Long, joined as co-founders. (Both still serve as scientific advisers to the company.) Going from the lab to a commercially viable product required overcoming big technical challenges: An early prototype took almost 20 minutes to render a single image of a hand gesture and produce a corresponding sensation. Now Ultrahaptics software can render 100,000 frames per second with a fraction of the original computing power, making it viable for home appliances and cars.
In 2014, Carter and his partners secured £600,000 ($865,000) in seed money from IP Group Plc, a publicly listed British venture capital firm, and went looking for a chief executive. They quickly hired Cliffe, who had held a variety of executive roles in the semiconductor industry. Cliffe said he immediately grasped the applications of Ultrahaptics’ product. And he liked Carter. “Tom and I hit it off straight away,” Cliffe said. He was impressed that Carter seemed to know what he didn’t know — and was willing to listen to a more seasoned hand. Ultrahaptics hired Cliffe in December 2014.
Last year, Ultrahaptics secured a £10.1 million Series A investment round, led by Woodford Investment Management, a large U.K. money management firm run by Neil Woodford, a former top Invesco portfolio manager. IP Group also pumped in more money. Since its founding, the company has sold more than 50 kits, which cost $20,000 apiece including technical support, that companies can use to evaluate the technology. It has more than 20 employees and plans to double in size this year.
With funding lined up and a U.S. office opening later this year, the firm is racing to help customers get the first products using Ultrahaptics to market. The company has received a €1.5 million ($1.64 million) grant from the European Union to put the technology on a single chip for small businesses keen to use the technology — for instance, bringing gesture control to household lamps.
There are many more applications besides virtual reality and home appliances. For example, Ultrahaptics technology could enable the blind and deaf to navigate more safely.
And then there is the tractor beam.
Last year, Long and Subramanian were part of a team, along with other researchers from the University of Bristol and Spain’s University of Navarre, that used Ultrahaptics gear to levitate small polystyrene beads, grabbing them and manipulating them through the air.
At the Bristol Interaction and Graphics lab, the same university lab that gave birth to Ultrahaptics, a Ph.D. student is already applying the tractor beam technology to create three-dimensional data visualizations. The levitating beads making up a chart rearrange themselves in midair as the data changes.
And Bruce Drinkwater, a professor who specializes in ultrasound at Bristol and participated in the tractor beam project, said there are potential medical uses for the technology, including using the beam to grab and help remove kidney stone fragments.