The Corso group’s main research thrust is high-level computer vision and its relationship to human language, robotics and data science. They primarily focus on problems in video understanding such as video segmentation, activity recognition, and video-to-text; methodology, models leveraging cross-model cues to learn structured embeddings from large-scale data sources as well as graphical models emphasizing structured prediction over large-scale data sources are their emphasis. From biomedicine to recreational video, imaging data is ubiquitous. Yet, imaging scientists and intelligence analysts are without an adequate language and set of tools to fully tap the information-rich image and video. His group works to provide such a language. His long-term goal is a comprehensive and robust methodology of automatically mining, quantifying, and generalizing information in large sets of projective and volumetric images and video to facilitate intelligent computational and robotic agents that can natural interact with humans and within the natural world.
My lab creates systems that use a combination of both human and machine computation to solve problems quickly and reliably. We have introduced the idea of continuous real-time crowdsourcing, as well as the ‘crowd agent’ model, which uses computer-mediated groups of people submitting input simultaneously to create a collective intelligence capable of completing tasks better than any constituent member.
Professor Subramanian is interested in a variety of stochastic modeling, decision and control theoretic, and applied probability questions concerned with networks. Examples include analysis of random graphs, analysis of processes like cascades on random graphs, network economics, analysis of e-commerce systems, mean-field games, network games, telecommunication networks, load-balancing in large server farms, and information assimilation, aggregation and flow in networks especially with strategic users.
I study patterns in large, complex data sets, and make quantitative predictions and inferences about those patterns. Problems I’ve worked on include classification, anomaly detection, active and semi-supervised learning, transfer learning, and density estimation. I am primarily interested in developing new algorithms and proving performance guarantees for new and existing algorithms.
Raj Nadakuditi, PhD, is Associate Professor of Electrical Engineering and Computer Science, College of Engineering, at the University of Michigan, Ann Arbor.
Prof. Nadakuditi received his Masters and PhD in Electrical Engineering and Computer Science at MIT as part of the MIT/WHOI Joint Program in Ocean Science and Engineering. His work is at the interface of statistical signal processing and random matrix theory with applications such as sonar, radar, wireless communications and machine learning in mind.
Prof. Nadakuditi particularly enjoys using random matrix theory to address problems that arise in statistical signal processing. An important component of his work is applying it in real-world settings to tease out low-level signals from sensor, oceanographic, financial and econometric time/frequency measurements/time series. In addition to the satisfaction derived from transforming the theory into practice, real-world settings give us insight into how the underlying techniques can be refined and/or made more robust.
Professor Balzano and her students investigate problems in statistical signal processing and optimization, particularly dealing with large and messy data. Her applications typically have missing, corrupted, and uncalibrated data as well as heterogeneous data in terms of sensors, sensor quality, and scale in both time and space. Her theoretical interests involve classes of non-convex problems that include Principal Components Analysis (or the Singular Value Decomposition) and many interesting variants such as PCA with sparse or structured principal components, orthogonality and non-negativity constraints, nonlinear variants such as low-dimensional algebraic variety models, and even categorical data or human preference data. She concentrates on fast gradient methods and related optimization methods that are scalable to real-time operation and massive data. Her work provides algorithmic and statistical guarantees for these algorithms on the aforementioned non-convex problems, and she focuses carefully on assumptions that are realistic for the relevant applications. She has worked in the areas of online algorithms, real-time computer vision, compressed sensing and matrix completion, network inference, and sensor networks.
Jenna Wiens, PhD, is Assistant Professor of Computer Science and Engineering (CSE) in the College of Engineering at the University of Michigan, Ann Arbor.
Prof. Wiens currently heads the MLD3 research group. Her primary research interests lie at the intersection of machine learning, data mining, and healthcare. Within machine learning, she is particularly interested in time-series analysis, transfer/multitask learning, causal inference, and learning intelligible models. The overarching goal of her research is to develop the computational methods needed to help organize, process, and transform patient data into actionable knowledge. Her work has applications in modeling disease progression and predicting adverse patient outcomes. For several years now, Prof. Wiens has been focused on developing accurate patient risk stratification approaches that leverage spatiotemporal data, with the ultimate goal of reducing the rate of healthcare-associated infections among patients admitted to hospitals in the US. In addition to her research in the healthcare domain, she also spends a portion of my time developing new data mining techniques for analyzing player tracking data from the NBA.
Matthew Johnson-Roberson, PhD, is Assistant Professor of Naval Architecture and Marine Engineering and Assistant Professor of Electrical Engineering and Computer Science, College of Engineering, the University of Michigan, Ann Arbor.
The increasing economic and environmental pressures facing the planet require cost-effective technological solutions to monitor and predict the health of the earth. Increasing volumes of data and the geographic dispersion of researchers and data gathering sites has created new challenges for computer science. Remote collaboration and data abstraction offer the promise of aiding science for great social benefit. Prof. Johnson-Roberson’s research in this field has been focused on developing novel methods for the visualization and interpretation of massive environments from multiple sensing modalities and creating abstractions and reconstructions that allow natural scientists to predict and monitor the earth through remote collaboration. Through the promotion of these economically efficient solutions, his work aims to increase access to hundreds of scientific sites instantly without traveling. In undertaking this challenge he is constantly aiming to engage in research that will benefit society.
Kevyn Collins-Thompson, PhD, is Associate Professor of Information, School of Information and Associate Professor of Electrical Engineering and Computer Science, College of Engineering, at the University of Michigan, Ann Arbor.
My lab explores algorithms and interfaces for intelligent information systems that can infer when and how to help people learn and discover. Examples include search engines that can deliver the right kind of personalized information at the right time, and intelligent tutoring systems that learn when and how to be most helpful in teaching a particular student. Toward these goals, I employ data-centric methods that include machine learning from interaction traces and large-scale text mining and retrieval. My current research is centered on education, but I’m also interested in mobile and health-related applications.
My main research interest is in the old-fashioned goal of Artificial Intelligence (AI), that of building autonomous agents that can learn to be broadly competent in complex, dynamic, and uncertain environments. The field of reinforcement learning (RL) has focused on this goal and accordingly my deepest contributions are in RL.
A very recent effort combines Deep Learning and Reinforcement Learning.
From time to time, I take seriously the challenge of building agents that can interact with other agents and even humans in both artificial and natural environments. This has led to research in:
Over the past few years, I have begun to focus on Healthcare as an application area.