Queen's Gazette | Queen's University

Search form

In the palm of your hand

Queen’s researchers unveil holographic smartphone.

Researchers at the Human Media Lab at Queen’s University have developed the world’s first holographic flexible smartphone. The device, dubbed HoloFlex, is capable of rendering 3D images with motion parallax and stereoscopy to multiple simultaneous users without head tracking or glasses.

“HoloFlex offers a completely new way of interacting with your smartphone. It allows for glasses-free interactions with 3D video and images in a way that does not encumber the user,” says Roel Vertegaal (School of Computing), Director of the Human Media Lab.

A HoloFlex phone in use for videoconferencing. Researchers at the Human Media Lab are calling the HoloFlex the world's first holographic flexible smartphone. (Human Media Lab)

HoloFlex features a 1920x1080 full high-definition Flexible Organic Light Emitting Diode (FOLED) touchscreen display. Images are rendered into 12-pixel wide circular blocks, which create the full view of the 3D object from a particular viewpoint. These pixel blocks project through a 3D printed flexible microlens array consisting of more than 16,000 fisheye lenses, creating a 3D image in 160 x 104 resolution that which can be viewed from any angle simply by rotating the phone.

Building on the success of the ReFlex flexible smartphone, HoloFlex is also equipped with a bend sensor, which allows the user to bend the phone as a means of moving objects along the z-axis of the display. HoloFlex is powered by a 1.5 GHz Qualcomm Snapdragon 810 processor and 2 GB of memory. The board runs Android 5.1 and includes an Adreno 430 GPU supporting OpenGL 3.1.

Dr. Vertegaal envisions a number of applications for the new functionality of the HoloFlex technology. A first application is the use of bend gestures for Z-Input to facilitate the editing of 3D models, for example, when 3D printing. Using the touchscreen, a user can swipe to manipulate objects in the x and y axes, while squeezing the display to move objects along the z-axis. Due to the wide view angle, multiple users can examine a 3D model simultaneously from different points of view.

“By employing a depth camera, users can also perform holographic video conferences with one another,” Dr. Vertegaal says. “When bending the display users literally pop out of the screen and can even look around each other, with their faces rendered correctly from any angle to any onlooker.”

HoloFlex also can be used for holographic gaming. In a game such as Angry Birds, for example, users would be able to bend the side of the display to pull the elastic rubber band that propels the bird. When the bird flies across the screen, the holographic display makes the bird literally pop out of the screen in the third dimension.

Queen’s researchers will unveil HoloFlex in San Jose, Calif., at the top conference in human-computer interaction, ACM CHI 2016, on May 9. This research was support by Immersion Canada Inc. and the Natural Sciences and Engineering Research Council of Canada (NSERC).

For more information on the HoloFlex, please visit the Human Media Lab website