Algorithms the chunks of code that allow programs to sort, filter and combine data, among other things are the standard tools of modern computing. Like tiny gears inside a watch, algorithms execute well-defined tasks within more complicated programs.
Theyre ubiquitous, and in part because of this, theyve been painstakingly optimized over time. When a programmer needs to sort a list, for example, theyll reach for a standard sort algorithm thats been used for decades.
Now researchers are taking a fresh look at traditional algorithms, using the branch of artificial intelligence known as machine learning. Their approach, called algorithms with predictions, takes advantage of the insights machine learning tools can provide into the data that traditional algorithms handle. These tools have, in a real way, rejuvenated research into basic algorithms.
Machine learning and traditional algorithms are two substantially different ways of computing, and algorithms with predictions is a way to bridge the two, said Piotr Indyk, a computer scientist at the Massachusetts Institute of Technology. Its a way to combine these two quite different threads.
The recent explosion of interest in this approach began in 2018 with a paper by Tim Kraska, a computer scientist at MIT, and a team of Google researchers. In it, the authors suggested that machine learning could improve a well-studied traditional algorithm called a Bloom filter, which solves a straightforward but daunting problem.
Imagine you run your companys IT department and you need to check if your employees are going to websites that pose a security risk. Naively, you might think youll need to check every site they visit against a blacklist of known sites. If the list is huge (as is likely the case for undesirable sites on the internet), the problem becomes unwieldly you cant check every site against a huge list in the tiny amount of time before a webpage loads.
The Bloom filter provides a solution, allowing you to quickly and accurately check whether any particular sites address, or URL, is on the blacklist. It does this by essentially compressing the huge list into a smaller list that offers some specific guarantees.
Bloom filters never produce false negatives if they say the site is bad, its bad. However, they can produce false positives, so perhaps your employees wont be able to visit some sites they should have access to. Thats because they trade some accuracy for an enormous amount of data compression a trick called lossy compression. The more that Bloom filters compress the original data, the less accurate they are, but the more space they save.
To a simple Bloom filter, every website is equally suspicious until its confirmed to not be on the list. But not all websites are created equal: Some are more likely than others to wind up on a blacklist, simply because of details like their domain or the words in their URL. People understand this intuitively, which is why you likely read URLs to make sure theyre safe before you click on them.
Kraskas team developed an algorithm that can also apply this kind of logic. They called it a learned Bloom filter, and it combines a small Bloom filter with a recurrent neural network (RNN) a machine learning model that learns what malicious URLs look like after being exposed to hundreds of thousands of safe and unsafe websites.
When the learned Bloom filter checks a website, the RNN acts first and uses its training to determine if the site is on the blacklist. If the RNN says its on the list, the learned Bloom filter rejects it. But if the RNN says the site isnt on the list, then the small Bloom filter gets a turn, accurately but unthinkingly searching its compressed websites.
By putting the Bloom filter at the end of the process and giving it the final say, the researchers made sure that learned Bloom filters can still guarantee no false negatives. But because the RNN pre-filters true positives using what its learned, the small Bloom filter acts more as a backup, keeping its false positives to a minimum as well. A benign website that could have been blocked by a larger Bloom filter can now get past the more accurate learned Bloom filter. Effectively, Kraska and his team found a way to take advantage of two proven but traditionally separate ways of approaching the same problem to achieve faster, more accurate results.
Kraskas team showed that the new approach worked, but they didnt formalize why. That task fell to Michael Mitzenmacher, an expert on Bloom filters at Harvard University, who found Kraskas paper innovative and exciting, but also fundamentally unsatisfying. They run experiments saying their algorithms work better. But what exactly does that mean? he asked. How do we know?
In 2019, Mitzenmacher put forward a formal definition of a learned Bloom filter and analyzed its mathematical properties, providing a theory that explained exactly how it worked. And whereas Kraska and his team showed that it could work in one case, Mitzenmacher proved it could always work.
Mitzenmacher also improved the learned Bloom filters. He showed that adding another standard Bloom filter to the process, this time before the RNN, can pre-filter negative cases and make the classifiers job easier. He then proved it was an improvement using the theory he developed.
The early days of algorithms with predictions have proceeded along this cyclical track innovative ideas, like the learned Bloom filters, inspire rigorous mathematical results and understanding, which in turn lead to more new ideas. In the past few years, researchers have shown how to incorporate algorithms with predictions into scheduling algorithms, chip design and DNA-sequence searches.
In addition to performance gains, the field also advances an approach to computer science thats growing in popularity: making algorithms more efficient by designing them for typical uses.
Currently, computer scientists often design their algorithms to succeed under the most difficult scenario one designed by an adversary trying to stump them. For example, imagine trying to check the safety of a website about computer viruses. The website may be benign, but it includes computer virus in the URL and page title. Its confusing enough to trip up even sophisticated algorithms.
Indyk calls this a paranoid approach. In real life, he said, inputs are not generally generated by adversaries. Most of the websites employees visit, for example, arent as tricky as our hypothetical virus page, so theyll be easier for an algorithm to classify. By ignoring the worst-case scenarios, researchers can design algorithms tailored to the situations theyll likely encounter. For example, while databases currently treat all data equally, algorithms with predictions could lead to databases that structure their data storage based on their contents and uses.
And this is still only the beginning, as programs that use machine learning to augment their algorithms typically only do so in a limited way. Like the learned Bloom filter, most of these new structures only incorporate a single machine learning element. Kraska imagines an entire system built up from several separate pieces, each of which relies on algorithms with predictions and whose interactions are regulated by prediction-enhanced components.
Taking advantage of that will impact a lot of different areas, Kraska said.
Here is the original post:
Machine Learning Reimagines the Building Blocks of Computing - Quanta Magazine