Smarter, faster algorithm cuts number of steps to solve problems
What if a large class of algorithms used today — from the algorithms that help us avoid traffic to the algorithms that identify new drug molecules — worked exponentially faster?
Computer scientists at the Harvard John A. Paulson School of Engineering and Applied Sciences (SEAS) have developed a completely new kind of algorithm, one that exponentially speeds up computation by dramatically reducing the number of parallel steps required to reach a solution.
The researchers will present their novel approach at two upcoming conferences: the ACM Symposium on Theory of Computing (STOC), June 25-29 and International Conference on Machine Learning (ICML), July 10 -15.
A lot of so-called optimization problems, problems that find the best solution from all possible solutions, such as mapping the fastest route from point A to point B, rely on sequential algorithms that haven’t changed since they were first described in the 1970s. These algorithms solve a problem by following a sequential step-by-step process. The number of steps is proportional to the size of the data. But this has led to a computational bottleneck, resulting in lines of questions and areas of research that are just too computationally expensive to explore.
“These optimization problems have a diminishing returns property,” said Yaron Singer, Assistant Professor of Computer Science at SEAS and senior author of the research. “As an algorithm progresses, its relative gain from each step becomes smaller and smaller.”
Singer and his colleague asked: what if, instead of taking hundreds or thousands of small steps to reach a solution, an algorithm could take just a few leaps?
“This algorithm and general approach allows us to dramatically speed up computation for an enormously large class of problems across many different fields, including computer vision, information retrieval, network analysis, computational biology, auction design, and many others,” said Singer. “We can now perform computations in just a few seconds that would have previously taken weeks or months.”
“This new algorithmic work, and the corresponding analysis, opens the doors to new large-scale parallelization strategies that have much larger speedups than what has ever been possible before,” said Jeff Bilmes, Professor in the Department of Electrical Engineering at the University of Washington, who was not involved in the research. “These abilities will, for example, enable real-world summarization processes to be developed at unprecedented scale.”
Traditionally, algorithms for optimization problems narrow down the search space for the best solution one step at a time. In contrast, this new algorithm samples a variety of directions in parallel. Based on that sample, the algorithm discards low-value directions from its search space and chooses the most valuable directions to progress towards a solution.
Take this toy example:
You’re in the mood to watch a movie similar to The Avengers. A traditional recommendation algorithm would sequentially add a single movie in every step which has similar attributes to those of The Avengers. In contrast, the new algorithm samples a group of movies at random, discarding those that are too dissimilar to The Avengers. What’s left is a batch of movies that are diverse (after all, you don’t want ten Batman movies) but similar to The Avengers. The algorithm continues to add batches in every step until it has enough movies to recommend.
This process of adaptive sampling is key to the algorithm’s ability to make the right decision at each step.
“Traditional algorithms for this class of problem greedily add data to the solution while considering the entire dataset at every step,” said Eric Balkanski, graduate student at SEAS and co-author of the research. “The strength of our algorithm is that in addition to adding data, it also selectively prunes data that will be ignored in future steps.”
In experiments, Singer and Balkanski demonstrated that their algorithm could sift through a data set which contained 1 million ratings from 6,000 users on 4,000 movies and recommend a personalized and diverse collection of movies for an individual user 20 times faster than the state-of-the-art.
The researchers also tested the algorithm on a taxi dispatch problem, where there are a certain number of taxis and the goal is to pick the best locations to cover the maximum number of potential customers. Using a data set of two million taxi trips from the New York City taxi and limousine commission, the adaptive-sampling algorithm found solutions 6 times faster.
“This gap would increase even more significantly on larger scale applications, such as clustering biological data, sponsored search auctions, or social media analytics,” said Balkanski.
Of course, the algorithm’s potential extends far beyond movie recommendations and taxi dispatch optimizations. It could be applied to:
- designing clinical trials for drugs to treat Alzheimer’s, multiple sclerosis, obesity, diabetes, hepatitis C, HIV and more
- evolutionary biology to find good representative subsets of different collections of genes from large datasets of genes from different species
- designing sensor arrays for medical imaging
- identifying drug-drug interaction detection from online health forums
This process of active learning is key to the algorithm’s ability to make the right decision at each step and solves the problem of diminishing returns.
“This research is a real breakthrough for large-scale discrete optimization,” said Andreas Krause, professor of Computer Science at ETH Zurich, who was not involved in the research. “One of the biggest challenges in machine learning is finding good, representative subsets of data from large collections of images or videos to train machine learning models. This research could identify those subsets quickly and have substantial practical impact on these large-scale data summarization problems.”
Singer-Balkanski model and variants of the algorithm developed in the paper could also be used to more quickly assess the accuracy of a machine learning model, said Vahab Mirrokni, a principal scientist at Google Research, who was not involved in the research.
“In some cases, we have a black-box access to the model accuracy function which is time-consuming to compute,” said Mirrokni. “At the same time, computing model accuracy for many feature settings can be done in parallel. This adaptive optimization framework is a great model for these important settings and the insights from the algorithmic techniques developed in this framework can have deep impact in this important area of machine learning research.”
Learn more: ‘Breakthrough’ algorithm exponentially faster than any previous one
The Latest on: Machine learning
via Google News
The Latest on: Machine learning
- Researchers explore machine learning to automate sorting of microcapsules in real-time on April 17, 2019 at 6:04 am
Micro-encapsulated CO 2 sorbents (MECS)—tiny, reusable capsules full of a sodium carbonate solution that can absorb carbon dioxide from the air—are a promising technology for capturing carbon ... […]
- Crystallizing knowledge with a learning machine on April 17, 2019 at 4:54 am
Transforming a new drug from a set of liquid ingredients in a lab to a pill in a box can be an exercise in complex chemistry. To better understand how drug ingredients crystallize, UConn ... […]
- Machine Learning Identifies Weapons in the Christchurch Attack Video. We Know, We Tried It on April 17, 2019 at 4:34 am
This piece is part of an ongoing Motherboard series on Facebook's content moderation strategies. You can read the rest of the coverage here. 29 minutes after the Christchurch terrorist started ... […]
- Latest Innovation in Machine Learning in Healthcare Industry with Top Players by 2024- Philips, Google, Intel, Oracle, HP and IBM on April 17, 2019 at 3:43 am
Apr 17, 2019 (The Expresswire via COMTEX) -- Machine Learning in Healthcare Industry Report presents analysis by size, growth rate, key players, regions, types & applications. Global Machine Learning ... […]
- NCI inks partnership with machine-learning company on April 17, 2019 at 3:00 am
NCI Information Systems Inc. notched another alliance with an innovator in the emerging technology space Tuesday, partnering with North Carolina-based machine learning company Tanjo Inc. The ... […]
- Robust data governance is key for machine learning success on April 17, 2019 at 2:48 am
The terms artificial intelligence and machine learning are often projected as two sides of the same coin. Principally though, whilst the ML algorithms enhance AI capabilities, and enables them to do ... […]
- Using Semi-Supervised Machine Learning in Laser Powder-bed Fusion Fault Detection on April 17, 2019 at 1:13 am
Researchers from the University of Liverpool outline their findings regarding the automatic detection of faults in additive manufacturing products in a recently published paper, ‘Automatic fault ... […]
- Machine Learning Vs. Artificial Intelligence: How Are They Different? on April 16, 2019 at 10:17 am
If you’re hoping to use one or the other in your business, it’s important to know which one to focus on. ML and AI are related, but they aren’t the same, and they aren’t necessarily suited to the same ... […]
- Why You Should Consider Google AI Platform For Your Machine Learning Projects on April 16, 2019 at 7:28 am
At Cloud Next 2019, Google announced the launch of AI Platform, a comprehensive machine learning service for developers and data scientists. Google has many investments in the space of machine ... […]
via Bing News