## Projects## List of Projects## Sorting by ReversalsAndrew Childs, Aniruddha Bapat, Alexey Gorshkov, Eddie Schoute In the Sorting By Reversals (SBR) problem, we wish to implement a given permutation using a minimum number of reversals. A reversal σ(i,j) inverts a subsequence as The SBR problem asks us to find the minimum number of reversals , for , such that . Caprara showed that SBR is NP-hard. (Other work even showed that it is APX-hard, which means hard to even approximate.) We now consider a variant of this problem where we are allowed to do reversals on disjoint segments (i.e., and with ) in parallel and we wish to minimize the number of such parallel reversal steps. This seems like a harder problem than SBR but its complexity is not known. In this project we will study the notion of “parallel sorting by reversals” and try to find and implement efficient algorithms for solving it. Furthermore, it would be interesting to prove hardness, and possibly the inapproximability, of this problem. ## Diversity in Matching Markets
How should a firm allocate its limited interviewing resources to select the optimal cohort of new employees from a large set of job applicants? How should that firm allocate cheap but noisy resume screenings and expensive but in-depth in-person interviews? We view this problem through the lens of combinatorial pure exploration (CPE) in the multi-armed bandit setting, where a central learning agent performs costly exploration of a set of arms before selecting a final subset with some combinatorial structure. In our case, that combinatorial structure will incorporate elements of diversity, where including multiple elements of the same type in the final set results in diminishing marginal gain. We will apply our algorithms to real data from university admissions and the hiring processes from major tech firms. ## Using Generative Adversarial Networks to Guarantee Fairness
Learning a probability model from data is a fundamental problem in machine learning and statistics. Building on the success of deep learning, there has been a modern approach to this problem using Generative Adversarial Networks (GANs). GANs view to this problem as a game between two sets of functions: a generator whose goal is to generate synthetic samples that are close to the real data training samples and a discriminator whose goal is to distinguish between the real and fake samples. In this project, we aim to leverage some of the state-of-the-art GAN architectures to produce distributions that are unbiased with respect to some protected features. ## Machine Learning for Ramsey Games (not fun) and NIM with Cash (maybe fun)William Gasarch and Joshua Twitty
Consider the following game: RED and BLUE take turns coloring numbers RED and BLUE, with RED going first. They are coloring . The first one to get an arithmetic sequence of length 4 in their color wins! (A variant: loses!). Does either player have a winning strategy? Can Machine Learning techniques help us discern such a strategy? This is an example of a Ramsey Game. The origin of such games is in a deep field of math called Ramsey Theory. Throughout this project we will learn a variety of AI techniques for playing games: mini-max, alpha-beta, Deep Q-learning, monte carlo methods, and others. You will also learn Ramsey theory. We may later apply AI techniques to variants of NIM, including NIM with Cash. ## Generating Information From Neural Network Models
This project will explore ways to extract information from trained neural networks.
We will explore this issue from two angles. First, we'll consider
## Learning Fair Representations
Machine learning algorithms are very powerful and they are being used everywhere today. Unfortunately, machine learning algorithms can learn bias as they learn. This could have drastic consequences. Many concerns have been voiced about the susceptibility of data-intensive AI algorithms to error and misuse, and the possible ramifications for gender, age, racial, or economic classes. The proper collection and use of data for AI systems, in this regard, represent an important challenge. Beyond purely data-related issues, however, larger questions arise about the design of AI to be inherently just, fair, transparent, and accountable. Researchers must learn how to design these systems so that their actions and decision-making are transparent and easily interpretable by humans, and thus can be examined for any bias they may contain, rather than just learning and repeating these biases. There are serious intellectual issues about how to represent and ‘‘encode’’ value and belief systems. Scientists must also study to what extent justice and fairness considerations can be designed into the system, and how to accomplish this within the bounds of current engineering techniques. The overall goal is to obtain a representation of our original data that has the same dimension as the original data, and preserves as much of information in the original data as possible while de-correlating the sensitive attribute. ## Cryptanalysis for Lattice-Based Cryptography
Next-generation cryptosystems may be based on newer assumptions related to the hardness of solving certain problems on lattices. This project will explore implementations of known algorithms from the literature for solving these problems in an effort to understand their concrete hardness. ## Algorithms for Group Decision-making and NegotiationThe overarching objective in project Decidio is deep insight on what cyber-supported practices in structured discussion bring about the best outcomes from a group that is charged with advising a “decider” (hence the project name.) Examples? A committee might be tasked with vetting research proposals to advise on funding decisions; experts may be tasked with evaluating consequences of one or another policy decision under consideration; and engineers may debate ways forward to solve a tough problem brought to them. In all cases the quality of a decision made with group input depends on that group's dynamics. Technology is an enabler of larger-scale interaction but also increases the risk of cacophony. Our hypothesis is that “rules matter" in structured decisions, which is to say that some conventions for interaction more so than others will demonstrably lead to better outcomes over time. This is especially true as the size of groups increases. The rules – the algorithms that choreograph interactions between participants – can be captured, and their efficacy measured. Ours is an experimental project. We're building a system that supports interaction within large, distributed groups. Decidio enables expression of rule-sets to define various kinds of interaction scenarios, which then can proceed using our communication framework. We anticipate that REU students will join our development team for the summer. Applicants should have Python programming skills, a basic understanding of full stack development, willingness to rapidly get up to speed on relevant human factors and an interest in seeing their ideas become published experiments. ## Tough Graphs. Mentor: Wisely Wong.This project looks into the toughness of a graph, which is a parameter on graph connectivity. The value is the minimum of the ratio of |S|/c(GS), where S runs through all vertex cuts of the graph, and c(GS) is the number of components from disconnecting the graph. It is NP-hard to determine the toughness value of a graph. We determined the toughness for the Kneser graphs K(n,k) for sufficiently large n, and all values of n for K(n,3), K(n,4). Moreover, we classified the exact vertex cuts that obtain the value. In the project, a variety of mathematical tools are used, including extremal set theory and spectral bounds in vertex cuts and independence number. |