Adversarial Machine Learning
Adversarial machine learning happens when model outputs are manipulated by carefully constructed inputs. This phenomenon opens up potential security risks via adversarial example attacks but, if leveraged correctly, also provides an interesting and powerful way to train machine learning models. In some sense, adversarial attacks and adversarial training are the two sides of the same coin.
As a security risk, whenever a machine learning system takes input from users or the physical world – such as machine-learning-as-a-service (MLaaS) or computer vision recognition in autonomous vehicles – an adversary can feed the model carefully perturbed data that appear to be legitimate but trigger unexpected behaviour from the machine learning system.
However, if the induced error is leveraged correctly, adversarial training, and – in particular, generative adversarial nets (GANs) – provide a new approach for learning . We typically train ML models by minimizing an objective with respect to their model parameters. This objective function is usually hand-designed to capture the way a model’s outputs deviate from the correct outputs. Coming up with a good objective function is a challenging research problem in itself and has traditionally required knowledge both about the data and the problem to be solved. GANs remove or reduce the need for humans to engineer the objective function; instead, the system learns the objective function from the data. This has great potential impact on the field as it automates another part of the data-processing pipeline.
At Borealis AI we are actively contributing to both sub-areas of adversarial machine learning. We believe that adversarial perturbations are a unique probe for better fundamental understanding of ML models. For example, we don’t yet understand what causes these adversarial examples and why adversarial examples can often be usefully transferred between models. With regards to adversarial training, our research interests are twofold: i) deepening theoretical understanding of GANs with the hope to improve their stability and solution; and ii) applying GANs onto novel data domains and exploiting adversarial training for better learning and inference in other ML models.
With applications ranging from the travelling salesman problem to clustering and scheduling, the field of combinatorial optimization involves discrete variable problems that have exponential complexity (NP hard) and are pervasive in our everyday lives. At Borealis AI, we’re interested in developing new and more efficient approaches to obtain better polynomial time-approximate solutions to these problems.
Specifically, we’re looking into developing methods that are rooted in two areas of machine learning: graphical models and reinforcement learning.
In the first area, casting a combinatorial problem as inference on a graphical model allows us to break the problem down into a set of smaller, partially-overlapping, local problems for computational speed up. Then, we alternate between solving these local subsets and exchanging their solutions where the subsets overlap to more quickly gain an approximate answer to the original problem.
From an RL standpoint, the goal in designing reward functions for combinatorial problems is twofold: i) avoiding handcrafted branching heuristics, which are often instance-specific and do not transfer over well from one combinatorial problem to another; and ii) finding solutions to particularly hard instances of these problems where the solution space is sparse and disjointed.
Reinforcement learning is a powerful framework for solving sequential decision tasks, including Atari games, robot control, and data center optimization. At Borealis AI, we’re working on fundamental improvements to the framework by incorporating human knowledge — although our goal is often super-human performance, leveraging a user’s abilities can help us learn faster and better across a wide range of tasks.
An additional, related goal is to learn directly from human feedback. Video games have a score to directly maximize. But how should your banking app work to help you? What should the program try to optimize? Our goal is to learn from both explicit and implicit user feedback to our decisions. Once successful, programs will be able to better understand what a given user finds useful and autonomously maximize user satisfaction.
Reinforcement Learning (Applied)
Reinforcement learning is one of the most promising sets of technologies for advancing artificial intelligence. Our team uses cutting edge methods in deep and asynchronous reinforcement learning to automate and amplify complex processes that do not lend themselves to traditional modelling as prediction or classification problems. In addition to applying state of the art methods, we look for opportunities to push the state of the art forward, developing new reinforcement learning techniques specifically designed to address the real world challenges that we face.
Natural Language Processing
Natural language processing, a catchall term for the intersection between computer systems and natural language, has existed in some form for the better part of a century. It’s only recently, however, that advances in deep learning have revitalized the field, enabling NLP researchers to tackle previously intractable problems. With open source tools such as Tensorflow’s SLING and Spacy, it’s now possible to quickly create end-to-end models that achieve state-of-the-art performance in natural language understanding tasks, and to apply those models to real-world data.
Our NLP team uses these recent advances to help bring structure to and extract logic from text. These tools allow the team to develop knowledge graphs constructed from a rich dataset of entities and events based on news articles, social media and financial reports. Our models allow us to discover latent information underlying entities and their relations, and to surface the right information at the right time.
Unsupervised learning is a fundamental pillar of machine learning, yet also one of the most challenging. The core challenge lies in how we can gain insight into unlabeled, and possibly unstructured, data in order to make actionable decisions. Businesses use unsupervised learning daily in the form of clustering, for example, but this clustering is only as good as the space in which the data is represented. The Borealis AI unsupervised learning team explores ways to leverage the vast amounts of detailed but unlabeled data that exists in the world. Our primary objective is to find new and creative ways to learn the true latent features and natural distributions of the data. Moreover, while we leverage these methods to find lookalike entities, we also strive to understand the evolution of entities through time and identify new phenomena that are important to the business and to society.