New! Sign up for our free email newsletter.
Science News
from research organizations

Image Processing Methods For Computer Vision-Based Fuel Gauge Developed

Date:
April 6, 2000
Source:
Penn State
Summary:
Taking a "look" at how much fuel is left in the tank could become literally possible now that Penn State computer engineers have developed image processing methods necessary for a computer vision-based fuel gauge.
Share:
FULL STORY

University Park, Pa. -- Taking a "look" at how much fuel is left in the tank could become literally possible now that Penn State computer engineers have developed image processing methods necessary for a computer vision-based fuel gauge.

The developers say a computer vision-based gauge would be much safer than current gauges that rely on a sensor with low voltage electrical leads that come in contact with the fuel. Electrical contact fuel measuring systems are often suspected of having contributed to explosions in aircraft disasters.

The image processing methods were developed by Srivatsan Chakravarthy, who earned his master's degree in August at Penn State, Dr. Rangachar Kasturi, professor of computer science and engineering, and Dr. Rajeev Sharma, assistant professor of computer science and engineering. The work was part of Chakravarthy's master's thesis.

In the method the Penn State team developed, nothing need be in the fuel tank except the fuel, Kasturi says. Two transparent glass portholes in the top of the fuel tank allow all the access needed to determine the depth of the liquid inside. One port enables a light source to flash the image of two crossed lines on the surface of the liquid while the other port allows an off-the-shelf video camera attached to a computer to record the position of the lines.

The computer's image processing software, developed by the researchers, is trained in a calibration process in which the tank is filled gradually and the position of the crossed lines, as recorded by the camera, are noted at the various depths. The correspondence between the depths in the actual 3-D scene and the 2-D image recorded by the camera is established in a process called digital mapping. The computer can then calculate the change in depth of the liquid in the tank when any one of the units, or pixels, in the camera's image of the crossed lines changes.

The researchers note that the mathematical problems involved in determining depth in a three-dimensional space from a two-dimensional camera image forms a part of almost all computer vision applications. A similar problem has to be solved whether one is equipping a robot with a 2-D camera "eye" to navigate in the real 3-D world or determining the depth of a liquid from its 2-D image. The researchers used two standard mathematical approaches to solving the problem for the fuel tank and found that one, triangulation, yielded simple, direct solutions that came at low computational cost. Although their computations were performed with a stationary system, they think their approach can be adapted to systems undergoing vibration, turbulence or other displacement.

The researchers conclude that computer vision offers an attractive alternative to the currently available options. They note that their experimental results are available now for studying the feasibility of implementing an actual system that could be incorporated in aircraft, automobiles or other applications.


Story Source:

Materials provided by Penn State. Note: Content may be edited for style and length.


Cite This Page:

Penn State. "Image Processing Methods For Computer Vision-Based Fuel Gauge Developed." ScienceDaily. ScienceDaily, 6 April 2000. <www.sciencedaily.com/releases/2000/04/000404204059.htm>.
Penn State. (2000, April 6). Image Processing Methods For Computer Vision-Based Fuel Gauge Developed. ScienceDaily. Retrieved March 18, 2024 from www.sciencedaily.com/releases/2000/04/000404204059.htm
Penn State. "Image Processing Methods For Computer Vision-Based Fuel Gauge Developed." ScienceDaily. www.sciencedaily.com/releases/2000/04/000404204059.htm (accessed March 18, 2024).

Explore More

from ScienceDaily

RELATED STORIES