+ All Categories
Home > Documents > Science DOI:10.1145/1592761.1592768 Gary Anthes Deep Data ...

Science DOI:10.1145/1592761.1592768 Gary Anthes Deep Data ...

Date post: 25-Jan-2022
Category:
Upload: others
View: 5 times
Download: 0 times
Share this document with a friend
2
N news NOVEMBER 2009 | VOL. 52 | NO. 11 | COMMUNICATIONS OF THE ACM 13 PHOTOGRAPH BY JONATHAN HILLER, CORNELL UNIVERSITY M INING SCIENTIFIC DATA for patterns and rela- tionships has been a common practice for decades, and the use of self-mutating genetic algorithms is nothing new, either. But now a pair of computer scientists at Cornell Univer- sity have pushed these techniques into an entirely new realm, one that could fundamentally transform the methods of science at the frontiers of research. Writing in a recent issue of the journal Science, Hod Lipson and Mi- chael Schmidt describe how they programmed a computer to take un- structured and imperfect lab measure- ments from swinging pendulums and mechanical oscillators and, with just the slightest initial direction and no knowledge of physics, mechanics, or geometry, derive equations represent- ing fundamental laws of nature. Conventional machine learning sys- tems usually aim to generate predictive models that might, for example, calcu- late the future position of a pendulum given its current position. However, the equations unearthed by Lipson and Schmidt represented basic invariant relationships—such as the conserva- tion of energy—of the kind that govern the behavior of the universe. The technique may come just in time as scientists are increasingly confronted with floods of data from the Internet, sensors, particle accel- erators, and the like in quantities that defy conventional attempts at analysis. “The technology to collect all that data has far, far surpassed the technology to analyze it and understand it,” says Schmidt, a doctoral candidate and member of the Cornell Computational Synthesis Lab. “This is the first time a computer has been used to go directly from data to a free-form law.” The Lipson/Schmidt work features two key advancements. The first is their look for invariants, or “conserva- tions,” rather than for predictive mod- els. “All laws of nature are essentially laws of conservation and symmetry,” says Lipson, a professor of mechanical engineering. “So looking for invariants is fundamental.” Given crude initial conditions and some indication of what variables to consider, the genetic program churned through a large number of possible equations, keeping and building on the most promising ones at each itera- tion and eliminating the others. The project’s second key advance was find- ing a way to identify the large number of trivial equations that, while true and invariant, are coincidental and not di- rectly related to the behavior of the sys- tem being studied. Lipson and Schmidt found that trivial equations could be weeded out by looking at ratios of rates of change in the variables under consideration. The program was written to favor those Science | DOI:10.1145/1592761.1592768 Gary Anthes Deep Data Dives Discover Natural Laws Computer scientists have found a way to bootstrap science, using evolutionary computation to find fundamental meaning in massive amounts of raw data. Cornell University’s Michael Schmidt, left, and Hod Lipson with one of the double pendulums used in their experiments.
Transcript
Page 1: Science DOI:10.1145/1592761.1592768 Gary Anthes Deep Data ...

nnews

november 2009 | vol. 52 | no. 11 | communications of the acm 13

Ph

ot

og

ra

Ph

by

Jo

na

th

an

hi

LL

er

, C

or

ne

LL

Un

iV

er

Si

ty

MiN iN g sCieNTiFiC daTa for patterns and rela-tionships has been a common practice for decades, and the use of

self-mutating genetic algorithms is nothing new, either. But now a pair of computer scientists at Cornell Univer-sity have pushed these techniques into an entirely new realm, one that could fundamentally transform the methods of science at the frontiers of research.

Writing in a recent issue of the journal Science, Hod Lipson and Mi-chael Schmidt describe how they programmed a computer to take un-structured and imperfect lab measure-ments from swinging pendulums and mechanical oscillators and, with just the slightest initial direction and no knowledge of physics, mechanics, or geometry, derive equations represent-ing fundamental laws of nature.

Conventional machine learning sys-tems usually aim to generate predictive models that might, for example, calcu-late the future position of a pendulum given its current position. However, the equations unearthed by Lipson and Schmidt represented basic invariant relationships—such as the conserva-tion of energy—of the kind that govern

the behavior of the universe. The technique may come just in

time as scientists are increasingly confronted with floods of data from the Internet, sensors, particle accel-erators, and the like in quantities that defy conventional attempts at analysis. “The technology to collect all that data has far, far surpassed the technology to analyze it and understand it,” says

Schmidt, a doctoral candidate and member of the Cornell Computational Synthesis Lab. “This is the first time a computer has been used to go directly from data to a free-form law.”

The Lipson/Schmidt work features two key advancements. The first is their look for invariants, or “conserva-tions,” rather than for predictive mod-els. “All laws of nature are essentially laws of conservation and symmetry,” says Lipson, a professor of mechanical engineering. “So looking for invariants is fundamental.”

Given crude initial conditions and some indication of what variables to consider, the genetic program churned through a large number of possible equations, keeping and building on the most promising ones at each itera-tion and eliminating the others. The project’s second key advance was find-ing a way to identify the large number of trivial equations that, while true and invariant, are coincidental and not di-rectly related to the behavior of the sys-tem being studied.

Lipson and Schmidt found that trivial equations could be weeded out by looking at ratios of rates of change in the variables under consideration. The program was written to favor those

Science | DOI:10.1145/1592761.1592768 Gary Anthes

Deep Data Dives Discover natural Laws Computer scientists have found a way to bootstrap science, using evolutionary computation to find fundamental meaning in massive amounts of raw data.

cornell university’s michael schmidt, left, and hod Lipson with one of the double pendulums used in their experiments.

Page 2: Science DOI:10.1145/1592761.1592768 Gary Anthes Deep Data ...

14 communications of the acm | november 2009 | vol. 52 | no. 11

news

equations that were able to use these ratios to predict connections between variables over time. “This was one of the biggest challenges we were able to overcome,” Lipson says. “There are infinite trivial equations and just a few interesting ones.”

Like human scientists, the software favors equations with the fewest terms. “We want to find the simplest equation powerful enough to predict the dynam-ics of the system,” Schmidt says.

applying artificial intelligenceScientific data has become so volumi-nous and complex in many disciplines today that scientists often don’t know what to look for or even how to start an-alyzing it. So they are applying artificial intelligence, via machine learning, to giant data sets without precisely speci-fying in advance a desired outcome. Unlike AI systems of the past, which were usually driven by hard-coded ex-pert rules, the idea now is to have the software evolve its own rules primed with an arbitrary starting point and a few simple objectives.

Automating the discovery of natu-ral laws marks a major step into a realm that was previously inhabited solely by humans.

Eric Horvitz, an AI specialist at Mi-crosoft Research, says it’s only the be-ginning. “Computers will grow to be-come scientists in their own right, with intuitions and computational variants of fascination and curiosity,” says Hor-vitz. “They will have the ability to build and test competing explanatory mod-els of phenomena, and to consider the likelihoods that each of the proposed models is correct. And they will under-stand how to progress through a space of inquiry, where they consider the best evidence to gather next and the best new experiments to perform.”

A major challenge facing the Eu-ropean Organization for Nuclear Re-search (CERN) is how to use the 40 terabytes of data that its Large Hadron Collider is expected to produce every day. Processing that amount of data would be a challenge if scientists knew exactly what to look for, but in fact they can hardly imagine what truths might be revealed if only the right tests are performed. CERN researchers have turned to Lipson and Schmidt for help in finding a way to search for scientific

laws hidden in the data. “It could be a killer app for them,” Schmidt says.

Indeed, Lipson and Schmidt have received so many requests to apply their techniques to other scientists’ data that they plan to turn their meth-odology and software into a freely dis-tributed tool.

Josh Bongard, a computer scien-tist and roboticist at the University of Vermont, says the Lipson/Schmidt ap-proach is noteworthy for its ability to find equations with very little assumed in advance about their form. “That gives the algorithm more free rein to derive relationships that we might not know about,” he says.

Bongard says earlier applications of machine learning to discovery have not scaled well, often working for simple systems, such as a single pendulum, but breaking down when applied to a chaotic system like a double pendu-lum. Further evidence of the scalabil-ity of the Lipson/Schmidt algorithm is its apparent ability to span different domains, from mechanical systems to very complex biological ones, he says.

The Lipson/Schmidt work takes search beyond “mining”—where a spe-cific thing is sought—to “discovery,” where “you are not sure what you are looking for, but you’ll know it when you find it,” Bongard says. A key to making that possible with large stores of com-plex data is having algorithms that are able to evolve building blocks from simple systems into successively more complex models.

Such methods aim to complement the efforts of scientists but not replace them, as some critics have suggested. “These algorithms help to bootstrap science, to help us better investigate the data and the models by acting like an intelligent filter,” says Bongard.

Scientific research for decades has followed a well-known path from data collection (observation) to model formu-lation and prediction, to laws (expressed as equations), and finally to a higher-level theoretical framework or interpre-tation of the laws. “We have shown we can go directly from data to laws,” says Schmidt, “so we are wondering if we can go from laws to the higher theory.”

He and Lipson are now trying to au-tomate that giant last step, but admit they have little idea how to do it. Their tentative first step uses a process of

analogy; a newly discovered but poorly understood equation is compared with similar equations in areas that are un-derstood.

For example, they recently mined a large quantity of metabolic data pro-vided by Gurol Suel, assistant professor at the University of Texas Southwestern Medical Center. The algorithm came up with two “very simple, very elegant” invariants—so far unpublished—that are able to accurately predict new data. But neither they nor Suel has any idea what the invariants mean, Lipson says. “So what we are doing now is trying to automate the interpretation stage, by saying, ‘Here’s what we know about the system, here’s the textbook biology; can you explain the new equations in terms of the old equations?’ ”

Lipson says the ultimate challenge may lie in dealing with laws so compli-cated they defy human understanding. Then, automation of the interpretation phase would be extremely difficult. “What if it’s like trying to explain Shakespeare to a dog?” he asks.

“The exciting thing to me is that we might be able to find the laws at all,” says Lipson. “Then we may have to write a program that can take a very complicated concept and break it down so humans can understand it. That’s a new challenge for AI.”

Further Reading

Lipson, H. and Schmidt, M. Distilling free-form laws from experimental data. Science 234 (Apr. 3, 2009), 81–85.

Waltz, D. and Buchanan, B. Automating science. Science 234 (Apr. 3, 2009), 43–44.

King, R., Whelan, K., Jones, F., Reiser, P., Bryant, C., Muggleton, S., Kell, D., Oliver, S. Functional genomic hypothesis generation and experimentation by a robot scientist. Nature 427, 6971 (Jan. 15, 2004), 247–252.

Bongard, J. and Lipson, H. Automated reverse engineering of nonlinear dynamical systems. In Proceedings of the National Academy of Sciences 104, 24 (June 6, 2007), 9943–9948.

Koza, J. Genetic Programming: On the Programming of Computers by Means of Natural Selection. MIT Press, Cambridge, MA, 1992.

Gary Anthes is a technology writer and editor based in arlington, Va.

© 2009 aCM 0001-0782/09/1100 $10.00


Recommended