Science News
from research organizations

Device allows users to manipulate 3-D virtual objects more quickly

Date:
May 1, 2017
Source:
North Carolina State University
Summary:
Researchers have developed a user-friendly, inexpensive controller for manipulating virtual objects in a computer program in three dimensions. The device allows users to manipulate objects more quickly -- with less lag time -- than existing technologies.
Share:
FULL STORY

Researchers at North Carolina State University have developed a user-friendly, inexpensive controller for manipulating virtual objects in a computer program in three dimensions. Called CAPTIVE, the device allows users to manipulate objects more quickly -- with less lag time -- than existing technologies.
Credit: Zeyuan Chen

Researchers at North Carolina State University have developed a user-friendly, inexpensive controller for manipulating virtual objects in a computer program in three dimensions. The device allows users to manipulate objects more quickly -- with less lag time -- than existing technologies.

The device, called CAPTIVE, offers six degrees of freedom (6DoF) for users -- with applications ranging from video gaming to medical diagnostics to design tools. And CAPTIVE makes use of only three components: a simple cube, the webcam already found on most smartphones and laptops, and custom software.

The cube is plastic, with differently colored balls at each corner. It resembles a Tinkertoy, but is made using a 3-D printer. When users manipulate the cube, the image is captured by the webcam. Video recognition software tracks the movement of the cube in three dimensions by tracking how each of the colored balls moves in relation to the others.

"The primary advantage of CAPTIVE is that it is efficient," says Zeyuan Chen, lead author of a paper on the work and a Ph.D. student in NC State's Department of Computer Science. "There are a number of tools on the market that can be used to manipulate 3-D virtual objects, but CAPTIVE allows users to perform these tasks much more quickly."

To test CAPTIVE's efficiency, researchers performed a suite of standard experiments designed to determine how quickly users can complete a series of tasks.

The researchers found, for example, that CAPTIVE allowed users to rotate objects in three dimensions almost twice as fast as what is possible with competing technologies.

"Basically, there's no latency; no detectable lag time between what the user is doing and what they see on screen," Chen says.

CAPTIVE is also inexpensive compared to other 6DoF input devices.

"There are no electronic components in the system that aren't already on your smartphone, tablet or laptop, and 3-D printing the cube is not costly," Chen says. "That really leaves only the cost of our software."

The paper, "Performance Characteristics of a Camera-Based Tangible Input Device for Manipulation of 3D Information," will be presented at the Graphics Interface conference being held in Edmonton, Alberta, May 16-19. The paper was co-authored by Christopher Healey, a professor of computer science at NC State and in the university's Institute for Advanced Analytics; and Robert St. Amant, an associate professor of computer science at NC State. The work was done with support from the National Science Foundation under grant number 1420159.

Video demonstrating CAPTIVE can be seen here: https://youtu.be/gRN5bYtYe3M.


Story Source:

Materials provided by North Carolina State University. Note: Content may be edited for style and length.


Cite This Page:

North Carolina State University. "Device allows users to manipulate 3-D virtual objects more quickly." ScienceDaily. ScienceDaily, 1 May 2017. <www.sciencedaily.com/releases/2017/05/170501094331.htm>.
North Carolina State University. (2017, May 1). Device allows users to manipulate 3-D virtual objects more quickly. ScienceDaily. Retrieved May 28, 2017 from www.sciencedaily.com/releases/2017/05/170501094331.htm
North Carolina State University. "Device allows users to manipulate 3-D virtual objects more quickly." ScienceDaily. www.sciencedaily.com/releases/2017/05/170501094331.htm (accessed May 28, 2017).

RELATED STORIES