A wearable, picture-taking health monitor created by University of Pittsburgh researchers has received a recent facelift. Now, in addition to documenting what a person eats, the eButton prototype can accurately match those images against a geometric-shape library, providing a much easier method for counting calories.
Published in Measurement Science and Technology, the Pitt study demonstrates a new computational tool that has been added to the eButton -- a device that fastens to the shirt like a pin. Using its newly built comprehensive food-shape library, the eButton can now extract food from 2D and 3D images and, using a camera coordinate system, evaluate that food based on shape, color, and size.
"Human memory of past eating is imperfect," said Mingui Sun, lead investigator and Pitt professor of neurological surgery and bioengineering and professor of electrical and computer engineering. "Visually gauging the size of a food based on an imaginary measurement unit is very subjective, and some individuals don't want to track what they consume. We're trying to remove the guess work from the dieting process."
eButton -- which is built with a low-power central processing unit, a random-access memory communication interface, and an Android operating system -- now includes a library of foods with nine common shapes: cuboid, wedge, cylinder, sphere, top and bottom half spheres, ellipse, half ellipse, and tunnel. The device snaps a series of photos while a person is eating, and its new formula goes to work: removing the background image, zeroing in on the food, and measuring its volume by projecting and fitting the selected 3D shape to the 2D photograph using a series of mathematical equations.
The Pitt team tested their new design on 17 popular favorites like jelly, broccoli, hamburgers, and peanut butter. Using a Logitech webcam, they captured five high-resolution images at different locations on diners' plates. Likewise, they applied the eButton to real-world dining scenarios in which diners were asked to wear the eButton on their chests, recording their eating. For each image, the eButton's new configuration method was implemented to automatically estimate the food portion size after the background was removed. To account for eaters leaving food behind, the Pitt team analyzed the last photograph taken during a meal. This leftover food was estimated and subtracted from the original portion size, as documented by earlier photographs.
"For food items with reasonable shapes, we found that this new method had an average error of only 3.69 percent," said Sun. "This error is much lower than that made by visual estimations, which results in an average error of about 20 percent."
While Sun and his colleagues were pleased with the results, there were three common foods that presented problems: ketchup, haddock, and ice cream. Because the properties of these provisions can change, results varied. Also, varying cooking techniques presented mixed consistencies, Sun said.
Even though the eButton is still not available commercially, Sun hopes to get it on the market soon. He and his team are now fine-tuning the device, working toward improving the accuracy of detecting portion sizes for irregularly shaped foods.
- Hsin-Chen Chen, Wenyan Jia, Yaofeng Yue, Zhaoxin Li, Yung-Nien Sun, John D Fernstrom, Mingui Sun. Model-based measurement of food portion size for image-based dietary assessment using 3D/2D registration. Measurement Science and Technology, 2013; 24 (10): 105701 DOI: 10.1088/0957-0233/24/10/105701
Cite This Page: