- Surfing a Wake of Light
- Atherosclerosis: Standing Beliefs Overturned
- Tundra Study: Climate Warming in Arctic
- Driverless Taxis to Roam City Streets?
- Extra DNA: 'Spare Tire' for Our Genomes
- Astronomers See Pebbles Poised to Make Planets
- Hidden Supermassive Black Holes Revealed
- Micro-Organisms Explain Features On Comets?
- Elusive Structure of HIV Protein Unraveled
- The Very Hungry Sea Anemone

Science News

from research organizations

- Date:
- March 2, 2013
- Source:
- Massachusetts Institute of Technology
- Summary:
- A new technique for solving 'graph Laplacians' is drastically simpler than its predecessors, with implications for a huge range of practical problems.
- Share:

FULL STORY

In the last decade, theoretical computer science has seen remarkable progress on the problem of solving graph Laplacians -- the esoteric name for a calculation with hordes of familiar applications in scheduling, image processing, online product recommendation, network analysis, and scientific computing, to name just a few. Only in 2004 did researchers first propose an algorithm that solved graph Laplacians in "nearly linear time," meaning that the algorithm's running time didn't increase exponentially with the size of the problem.

At this year's ACM Symposium on the Theory of Computing, MIT researchers will present a new algorithm for solving graph Laplacians that is not only faster than its predecessors, but also drastically simpler. "The 2004 paper required fundamental innovations in multiple branches of mathematics and computer science, but it ended up being split into three papers that I think were 130 pages in aggregate," says Jonathan Kelner, an associate professor of applied mathematics at MIT who led the new research. "We were able to replace it with something that would fit on a blackboard."

The MIT researchers -- Kelner; Lorenzo Orecchia, an instructor in applied mathematics; and Kelner's students Aaron Sidford and Zeyuan Zhu -- believe that the simplicity of their algorithm should make it both faster and easier to implement in software than its predecessors. But just as important is the simplicity of their conceptual analysis, which, they argue, should make their result much easier to generalize to other contexts.

**Overcoming resistance**

A graph Laplacian is a matrix -- a big grid of numbers -- that describes a graph, a mathematical abstraction common in computer science. A graph is any collection of nodes, usually depicted as circles, and edges, depicted as lines that connect the nodes. In a logistics problem, the nodes might represent tasks to be performed, while in an online recommendation engine, they might represent titles of movies.

In many graphs, the edges are "weighted," meaning that they have different numbers associated with them. Those numbers could represent the cost -- in time, money or energy -- of moving from one step to another in a complex logistical operation, or they could represent the strength of the correlations between the movie preferences of customers of an online video service.

The Laplacian of a graph describes the weights between all the edges, but it can also be interpreted as a series of linear equations. Solving those equations is crucial to many techniques for analyzing graphs.

One intuitive way to think about graph Laplacians is to imagine the graph as a big electrical circuit and the edges as resistors. The weights of the edges describe the resistance of the resistors; solving the Laplacian tells you how much current would flow between any two points in the graph.

Earlier approaches to solving graph Laplacians considered a series of ever-simpler approximations of the graph of interest. Solving the simplest provided a good approximation of the next simplest, which provided a good approximation of the next simplest, and so on. But the rules for constructing the sequence of graphs could get very complex, and proving that the solution of the simplest was a good approximation of the most complex required considerable mathematical ingenuity.

**Looping back**

The MIT researchers' approach is much more straightforward. The first thing they do is find a "spanning tree" for the graph. A tree is a particular kind of graph that has no closed loops. A family tree is a familiar example; there, a loop might mean that someone was both parent and sibling to the same person. A spanning tree of a graph is a tree that touches all of the graph's nodes but dispenses with the edges that create loops. Efficient algorithms for constructing spanning trees are well established.

The spanning tree in hand, the MIT algorithm then adds back just one of the missing edges, creating a loop. A loop means that two nodes are connected by two different paths; on the circuit analogy, the voltage would have to be the same across both paths. So the algorithm sticks in values for current flow that balance the loop. Then it adds back another missing edge and rebalances.

In even a simple graph, values that balance one loop could imbalance another one. But the MIT researchers showed that, remarkably, this simple, repetitive process of adding edges and rebalancing will converge on the solution of the graph Laplacian. Nor did the demonstration of that convergence require sophisticated mathematics: "Once you find the right way of thinking about the problem, everything just falls into place," Kelner explains.

**Paradigm shift**

Daniel Spielman, a professor of applied mathematics and computer science at Yale University, was Kelner's thesis advisor and one of two co-authors of the 2004 paper. According to Spielman, his algorithm solved Laplacians in nearly linear time "on problems of astronomical size that you will never ever encounter unless it's a much bigger universe than we know. Jon and colleagues' algorithm is actually a practical one."

Spielman points out that in 2010, researchers at Carnegie Mellon University also presented a practical algorithm for solving Laplacians. Theoretical analysis shows that the MIT algorithm should be somewhat faster, but "the strange reality of all these things is, you do a lot of analysis to make sure that everything works, but you sometimes get unusually lucky, or unusually unlucky, when you implement them. So we'll have to wait to see which really is the case."

The real value of the MIT paper, Spielman says, is in its innovative theoretical approach. "My work and the work of the folks at Carnegie Mellon, we're solving a problem in numeric linear algebra using techniques from the field of numerical linear algebra," he says. "Jon's paper is completely ignoring all of those techniques and really solving this problem using ideas from data structures and algorithm design. It's substituting one whole set of ideas for another set of ideas, and I think that's going to be a bit of a game-changer for the field. Because people will see there's this set of ideas out there that might have application no one had ever imagined."

**Story Source:**

The above post is reprinted from materials provided by **Massachusetts Institute of Technology**. The original item was written by Larry Hardesty. *Note: Materials may be edited for content and length.*

**Journal Reference**:

- Jonathan A. Kelner,
Lorenzo Orecchia,
Aaron Sidford,
Zeyuan Allen Zhu.
**A Simple, Combinatorial Algorithm for Solving SDD Systems in Nearly-Linear Time**.*Submitted to ArXiv*, 2013 [link]

**Cite This Page**:

Massachusetts Institute of Technology. "Short algorithm, long-range consequences." ScienceDaily. ScienceDaily, 2 March 2013. <www.sciencedaily.com/releases/2013/03/130302125400.htm>.

Massachusetts Institute of Technology. (2013, March 2). Short algorithm, long-range consequences. *ScienceDaily*. Retrieved July 6, 2015 from www.sciencedaily.com/releases/2013/03/130302125400.htm

Massachusetts Institute of Technology. "Short algorithm, long-range consequences." ScienceDaily. www.sciencedaily.com/releases/2013/03/130302125400.htm (accessed July 6, 2015).