Skip to main content
  • Description

    Data Science Institute Postdoctoral Scholars Program

    Application

    The Data Science Institute (DSI) at the University of Chicago invites applications for Postdoctoral Scholars who wish to advance cutting-edge data science approaches, methods, and applications in research. The Data Science Institute (DSI) executes the University of Chicago’s bold, innovative vision of Data Science as a new discipline by advancing foundational and interdisciplinary research, partnerships with industry, government, and social impact organizations, and holistic approach to data science education.

    The DSI is part of a vibrant and growing data science community which includes departments across campus as well as the Polsky Center for Entrepreneurship and Innovation, Argonne National Laboratory, Fermi National Accelerator Laboratory, and Toyota Technological Institute at Chicago (TTIC). These on-campus partnerships enable and encourage unique interdisciplinary research partnerships, reducing early-stage friction to hasten meaningful relationships and high-impact research.

    This unique program provides postdocs with the opportunity to pursue original research on significant questions in data science. Drawing on the University of Chicago’s top-ranked programs, world-renowned faculty, as well as a vibrant and quickly expanding data science ecosystem, this program will allow postdoctoral scholars to engage in field-defining data science and artificial intelligence research. Our positions carry a competitive salary, generous research funding allowances, and benefits. It is equally important that our postdocs work and grow in an environment that is supportive and focused on community and career development; as such, this program includes lectures focused on communication and workplace culture, and aims to develop a diverse cohort of well-rounded scholars pursuing careers across data science disciplines.

    Postdocs may (but do not need to) choose to express interest in working on one of the DSI’s three research initiatives, AI+Science, Data & Democracy, and Internet Equity.

    Equal Employment Opportunity Statement:

    We seek a diverse pool of applicants who wish to join an academic community that places the highest value on rigorous inquiry and encourages diverse perspectives, experiences, groups of individuals, and ideas to inform and stimulate intellectual challenge, engagement, and exchange. The University’s Statements on Diversity are at https://provost.uchicago.edu/statements-diversity.

    The University of Chicago is an Affirmative Action/Equal Opportunity/Disabled/Veterans Employer and does not discriminate on the basis of race, color, religion, sex, sexual orientation, gender identity, national or ethnic origin, age, status as an individual with a disability, protected veteran status, genetic information, or other protected classes under the law. For additional information please see the University’s Notice of Nondiscrimination.

    Job seekers in need of a reasonable accommodation to complete the application process should call 773-834-3988 or email equalopportunity@uchicago.edu with their request

    Contact: 

    For questions about this application, please contact: data-science@uchicago.edu.

    The ZhengTong Fellowship Fund, which supports the Data Science Institute Postdoctoral Scholars program, was established with generous support from ZhengTong Group.

  • Application

    Apply Here

    PDF Version

    Timeline:

    • Applications are open now and will remain open until all positions are filled.
    • Review of applications will begin on May 15th, 2023; applications received after this date will still be considered but we encourage interested applicants to submit by early May.
    • Estimated Start Date: September 1, 2023; earlier starts will be considered on a case-by-case basis

    Eligibility:

    If you have any questions about your eligibility, please feel free to send an email to data-science@uchicago.edu.

    • Completion of all requirements for a Ph.D. required at the time of appointment.
    • Applicants may only submit one application.
    • We welcome applications from researchers who are using data science to advance the state of the art in their respective field (e.g., humanities, social sciences, natural and physical sciences).
    • Postdoctoral scholars will be expected to be active participants in the DSI at UChicago, and may be requested to take on leadership roles in one or more of the DSI initiatives (e.g. supporting the AI+Science Summer School program, or engaging colleagues for speaker series events).

    Application Materials:

    • Curriculum vitae;
    • Summary of the candidate’s current research (250 words);
    • Research statement that outlines research goals and significance, research plan, and motivation for seeking a postdoctoral appointment at UChicago (maximum of 3 pages);
    • If you have indicated interest in one of the research initiatives, please include information about how your research and vision matches with the stated goals and foci of the selected initiative.
    • 1-2 representative publications or manuscripts;
    • Names and contact information for at least two and up to five references (the application is complete only when two letters of reference have been submitted, so please contact referees early in the application process). Referees will be provided with a link to the submission portal;
    • Names of potential UChicago faculty mentors;
    • (Optional) Indicate interest in any of the DSI Research Initiatives (AI+Science, Data & Democracy, and Internet Equity);
    • (Optional) Applicants may include a letter of collaboration from a UChicago faculty mentor who has agreed to mentor the applicant if the scholar is accepted into the program. Please use the following template for the letter:
      • “If Dr. [insert full name of applicant] is accepted as a Data Science Institute Postdoctoral Scholar at the University of Chicago, it is my intent to act as a mentor on a project of mutual interest.”

    Evaluation Criteria:

    Proposals will be reviewed by the DSI Postdoctoral Committee using the following factors:

    • Research Potential: Applicant displays overall potential for research excellence, demonstrated by research statement, academic progress, publications to date, endorsements from faculty recommenders, and long-term career goals.
    • Data Science Background: Experience or coursework in computer science, statistics, data science, AI, or a related field.
    • Impact: Applicant’s research evinces approaches, methods or theories that advance research innovation in interdisciplinary or foundational approaches in data science, or real world challenges.
    • Research Alignment: Relevance of research plan to DSI’s research initiatives and projects.

    Contact:

    For questions about this application, please contact data-science@uchicago.edu.

  • Program Structure & Benefits

    What you’ll do:  

    • Independent Research: Scholars will have the freedom to pursue their own research interests with a majority of their time spent working on scholar-driven research projects and no teaching responsibilities.
    • Joint Research with Mentor: Scholars will help lead and execute collaborative work on cutting-edge research projects with mentors in their academic field and area of interest.
    • Professional Development: Scholars will gain training and experience with: required mentoring and outreach through our Summer Lab and Clinic programs (required three quarters annually); communicating your research to a broad audience; engaging with the media and external stakeholders; and applying for and securing funding.
    • Program Benefits
    • Mentorship

      Mentors will provide ongoing research and career guidance through regular meetings, as well as opportunities to promote the scholar’s accomplishments in public forums. Scholars have the option to receive mentorship from a single faculty member or joint mentorship from a data science researcher and a domain expert.

    • Unique Datasets

      Scholars will have privileged, unique access to large-scale datasets from a variety of sectors.

    • Cohort Program

      The program will host activities where scholars can connect with members of their cohort, share knowledge, and gain insight through guest lectures, industry speakers, and other activities. Scholars will have autonomy and resources to select, host, and invite speakers, with support from DSI administrative staff.

    • Outreach and Impact

      Scholars will have considerable opportunities to establish new relationships and translate their research into real-world impact by leveraging our network of academic, civic, government, and industry connections.

    • Academia/Industry Ready

      Experience gained during the program will help scholars prepare for diverse career paths from tenure-track academic positions to leadership opportunities within innovative companies.

  • Mentors & Research Focus Areas

    As part of the Data Science Institute Postdoctoral Scholars program, postdocs will have the opportunity to work on collaborative projects in cutting-edge research areas. Learn more about potential faculty mentors  and sample research focus areas below.

    If you are interested in working with a particular mentor (listed below or elsewhere at UChicago), please indicate the area(s) and mentor name within your application. Please note that the list of example mentors and research focus areas below is not exhaustive.

    Applicants will also have the opportunity to indicate interest in the following Research Initiatives at the Data Science Institute:

    • Internet Equity and AccessDeveloping new data science tools to measure broadband access and working with community partners to make high-speed internet accessible to all.
    • AI & Science: Advancing the new frontier of artificial intelligence and scientific discovery.
    • Data & Democracy: Transformational research initiative focused on democracy in the digital age.

    AI & Machine Learning

    I am an Assistant Professor of Computer Science at the University of Chicago. I founded and direct 3DL (threedle! ), a group of enthusiastic researchers passionate about 3D, machine learning, and visual computing. I obtained my Ph.D. in 2021 from Tel Aviv University under the supervision of Daniel Cohen-Or and Raja Giryes.

    My research is focused on building artificial intelligence for 3D data, spanning the fields of computer graphics, machine learning, and computer vision. Deep learning, the most popular form of artificial intelligence, has unlocked remarkable success on structured data (such as text, images, and video), and I am interested in harnessing the potential of these techniques to enable effective operation on unstructured 3D geometric data.

    We have developed a convolutional neural network designed specifically for meshes, and also explored how to learn from the internal data within a single shape (for surface reconstructiongeometric texture synthesis, and point cloud consolidation) – and I am interested in broader applications related to these areas. Additional research directions that I am aiming to explore include: intertwining human and machine-based creativity to advance our capabilities in 3D shape modeling and animation; learning with less supervision, for example to extract patterns and relationships from large shape collections; and making 3D neural networks more “interpretable/explainable”.

    Chenhao Tan is an assistant professor at the Department of Computer Science and the UChicago Data Science Institute. His main research interests include language and social dynamics, human-centered machine learning, and multi-community engagement. He is also broadly interested in computational social science, natural language processing, and artificial intelligence.

    Homepage.

    Project: Causality for Credible AI. The backbone of modern ML is evaluation on randomly held out data. However, models with strong held out performance frequently still exhibit disturbing failures. For example, they can degrade when deployed out of domain, learn to depend on apparently irrelevant input features, or change substantively in response to tiny variations in training procedure (e.g., changing the random seed). The aim of this project is to develop methods to mitigate this kind of failure by baking in causal knowledge to model design and evaluation. For example, we may know changing certain input features shouldn’t change predictions (e.g., your medical diagnosis shouldn’t change if we edit your zipcode), or we may know something about common causal structure between domains (rain should affect a self-driving car the same in Chicago and New York). The challenge here is to translate domain knowledge and desiderata into formal requirements, and then determine how to enforce and measure these using available data.

    Bio: I am an assistant professor of Statistics and Data Science at the University of Chicago and a research scientist at Google Cambridge. My recent work revolves around the intersection of machine learning and causal inference, as well as the design and evaluation of safe and credible AI systems. Other noteable areas of interests include network data, and the foundations of learning and statistical inference.

    I was previously a Distinguished Postdoctoral Researcher in the department of statistics at Columbia University, where I worked with the groups of David Blei and Peter Orbanz. I completed my Ph.D. in statistics at the University of Toronto, where I was advised by Daniel Roy. In a previous life, I worked on quantum computing at the University of Waterloo. I won a number of awards, including the Pierre Robillard award for best statistics thesis in Canada.

    Energy & Environment

    Project: Designing next generation batteries with AI

    Description: A sustainable energy economy requires an immediate shift to renewable
    technologies, and batteries are vital to counterbalance the intermittency of solar and wind.
    Batteries consist of an anode, cathode, and an electrolyte; while the anode/cathode selection
    determines the theoretical energy density, the electrolyte choice often controls battery lifetime.
    Furthermore, electrolyte selection can determine how fast a battery charges, battery operation
    temperature, and safety. The conflicting requirements for electrolytes often mean that
    electrolytes are discovered using trial-and-error approaches; slowing down research progress and
    time to commercialization. Here, data science and the ability of AI to extract meaning from
    complex data is of vital interest. We have a multitude of data science + battery projects. One
    example project involves the development of natural language processing (NLP) and computer
    vision algorithms to extract battery data from the literature to build machine learning models.
    Although the scientific literature contains decades of relevant battery information, utilization of
    this data has been stifled by the lack of NLP tools capable of digesting the information. The
    second example project is focused on developing ML tools and applying these tools to discover
    novel electrolyte compounds for novel battery chemistries. Combining your data science
    expertise with the battery (and materials) domain knowledge that you will obtain from our
    laboratory will provide you with a skillset to exploit the battery sector that is expected to grow to
    over 100 billion dollars by 2025.

    Mentor: Chibueze Amanchukwu is a Neubauer Family Assistant Professor in the Pritzker School of
    Molecular Engineering at the University of Chicago, and holds a joint appointment in the
    Chemical Sciences and Engineering Division at Argonne National Laboratory. His research
    involves the design, synthesis, and understanding of ion transport in electrolytes for batteries and
    electrocatalytic applications. His team is especially interested in correlating bulk solvation
    properties to electrochemical interfacial phenomena, and they use data science tools coupled
    with synthesis and novel characterization approaches to answer these questions. He obtained his
    B.S. in chemical engineering at Texas A&M University, PhD in chemical engineering as a
    NDSEG Fellow at MIT and was a TomKat Center Postdoctoral Fellow at Stanford University.

    Projects: Data-driven Environmental Enforcement

    The Energy and Environment Lab invites a postdoc to collaborate on a suite of projects that leverage advances in monitoring technology and machine learning approaches to inform environmental policy, under the mentorship of Michael Greenstone, the Milton Friedman Distinguished Service Professor in Economics, the College, and the Harris School; Director of the Energy and Environment Lab, the Becker Friedman Institute, and the Energy Policy Institute at Chicago.

    Congestion and Traffic Safety

    Many cities across the United States have adopted Vision Zero, the policy goal of eliminating all traffic-related deaths and serious injuries. But what are the costs of achieving Vision Zero and what are the most efficient policy instruments to get there? Monitoring technologies offer the potential to revolutionize urban policy by providing governments with big data to inform policymaking. As part of NYC’s Vision Zero, the Department of Transportation is more than quintupling the number of speed cameras in the city. Leveraging our access to unique data of taxi, for-hire vehicle, and city fleet trips to model the impacts of new traffic cameras on vehicle crashes, slowdowns, and congestion spillovers. The post-doc would utilize large administrative datasets on camera enforcement, vehicle crashes, segment-level traffic speeds, and high-resolution driver behavior, to help measure the costs and benefits of enforcement strategies for fatality/injury reduction to inform optimal policy for urban traffic safety.

    Leveraging Satellite Data to Reduce Oil & Gas Methane Emissions

    The meteoritic rise of shale oil and gas (O&G) drilling in the United States poses significant challenges for reducing greenhouse gas emissions. The methane emitted has around 30 times greater short-term global warming potential than CO2, contributing aggressively to climate change. Reliable estimates of emitted methane are essential to fully understand and mitigate the environmental threat presented by shale drilling. While some estimates suggest that approximately 2.3% of gross natural gas production is leaked per year, accurate monitoring of emissions remains extremely challenging. Currently, regulators visit individual facilities to measure emissions; but due to budgetary constraints and a fast-growing industry inspector can visit only a fraction of the facilities each year. This project will leverage a wealth of administrative data and novel remote sensing data from recently-launched satellites to estimate facility-level methane emissions. Leveraging these unique data and state-of-the-art machine learning techniques, the project will help regulators re-design their monitoring and enforcement strategy to realize improvements in regulatory efficiency and reductions in greenhouse gases.

    Beyond Inspection Targeting: Deterrence through Machine Learning

    Building on a three-year partnership with the Environmental Protection Agency (EPA), this project aims to scale a machine learning-driven framework across inspection targeting programs at EPA. The Clean Water Act (CWA) is one program where data-driven inspection targeting can directly influence environmental policy. Using state-of-the-art machine learning models, we can generate risk scores for the likelihood individual firms will violate CWA standards, and use these model-generated risk scores to study facility compliance behavior and identify the most effective approaches to deterrence in a randomized field trial.

    Mentor: Michael Greenstone, Milton Friedman Distinguished Service Professor in Economics, the College, and the Harris School, University of Chicago; Director, Becker Friedman Institute for Research in Economics; Director, Energy Policy Institute at the University of Chicago (EPIC); Director, Tata Center for Development at the University of Chicago

    Michael Greenstone is the Milton Friedman Distinguished Service Professor in Economics, the College, and the Harris School, as well as the Director of the Becker Friedman Institute and the interdisciplinary Energy Policy Institute at the University of Chicago. He previously served as the Chief Economist for President Obama’s Council of Economic Advisers, where he co-led the development of the United States Government’s social cost of carbon. Greenstone also directed The Hamilton Project, which studies policies to promote economic growth, and has since joined its Advisory Council. He is an elected member of the American Academy of Arts and Sciences, a fellow of the Econometric Society, and a former editor of the Journal of Political Economy. Before coming to the University of Chicago, Greenstone was the 3M Professor of Environmental Economics at MIT.

    Greenstone’s research, which has influenced policy globally, is largely focused on uncovering the benefits and costs of environmental quality and society’s energy choices. His current work is particularly focused on testing innovative ways to increase energy access and improve the efficiency of environmental regulations around the world. Additionally, he is producing empirically grounded estimates of the local and global impacts of climate change as a co-director of the Climate Impact Lab. He also created the Air Quality Life Index™ that provides a measure of the gain in life expectancy communities would experience if their particulates air pollution concentrations are brought into compliance with global or national standards.

    Greenstone received a Ph.D. in economics from Princeton University and a BA in economics with High Honors from Swarthmore College.

    Foundations of Data Science

    Raul Castro Fernandez is an Assistant Professor of Computer Science at the University of Chicago. In his research he builds systems for discovering, preparing, and processing data. The goal of his research is to understand and exploit the value of data. He often uses techniques from data management, statistics, and machine learning. His main effort these days is on building platforms to support markets of data. This is part of a larger research effort on understanding the Economics of Data. He’s part of ChiData, the data systems research group at The University of Chicago.

    Homepage.

    Language learning has come to be a central theme in both cognitive science and artificial intelligence. The nature of language learning has long been a topic of interest for cognitive scientists, and machine learning has begun to dominate natural language processing (NLP) in modern AI. NLP systems have benefited tremendously from machine learning. However, the learning systems developed using these procedures often don’t achieve the efficiency and robustness of human language acquisition. Insights from language acquisition have the potential to help address this problem. But there are two critical challenges in exploring this possibility: (1) identifying the innate learning biases that enable fast, robust language learning in humans, and (2) determining how to translate theoretical insights about these biases into effective implementation for learning in NLP systems. This project will tackle both of these issues, making use of insights from special linguistic populations.

    Our first challenge –– identifying innate language learning predispositions –– is driven by the fact that most children are exposed to linguistic input from birth, making it difficult to disentangle innate characteristics versus characteristics that are rapidly learned from input. The rare cases in which children do not have usable linguistic input can help here by allowing us make important headway in identifying these predispositions. Congenitally deaf children who cannot learn the spoken language that surrounds them, and who have not been exposed to sign language by their hearing families, are in the unique situation of being without language input early in life. These children use their hands to communicate –– they gesture –– and those gestures (called “homesigns”) take on many, but not all, of the forms and functions of languages that have been handed down from generation to generation. The properties of these naturally-arising gestures provide evidence for the nature of linguistic predispositions independent of input.

    Drawing candidate biases from homesign, we will then tackle the second challenge –– incorporating biases into machine learning systems –– by systematic testing of models against real-world child language acquisition data. The goal of this phase will be to identify effective means of instantiating proposed human biases, and to test whether models incorporating these biases will successfully simulate the learning trajectories exhibited by children. Models with the proposed biases will be compared against minimally-different baseline models lacking the biases; stronger fit to human data will be taken as support that the biases are actual human predispositions. An important priority of this phase will be to balance scientific and engineering needs –– to maintain transparency of the models’ cognitive implications and to simulate human learning patterns as closely as possible, but also to use models that will interface smoothly with modern NLP systems, with promise to scale to larger datasets and broader domains.

    Mentor: Allyson Ettinger, Assistant Professor, Department of Linguistics

    Dr. Allyson Ettinger’s research is focused on language processing in humans and in artificial intelligence systems, motivated by a combination of scientific and engineering goals. For studying humans, her research uses computational methods to model and test hypotheses about mechanisms underlying the brain’s processing of language in real time. In the engineering domain, her research uses insights and methods from cognitive science, linguistics, and neuroscience in order to analyze, evaluate, and improve natural language understanding capacities in artificial intelligence systems. In both of these threads of research, the primary focus is on the processing and representation of linguistic meaning.

    Project: Operational Analytics for Communications Systems

    Many applications of machine learning to computer communications systems such as the Internet rely on models that are trained offline, on snapshots of data. Yet, in many operational systems, data arrives as a continuous stream—often as a timeseries, and decisions must be made on short timescales (e.g., milliseconds). In operational systems, designers must face difficult challenges and design tradeoffs concerning encoding of timeseries data, efficiently labeling large quantities of data, distinguishing anomalous activity from model drift, and trading off model accuracy versus model or feature complexity. In this research project, we will explore these challenges in the context of networked systems. Possible avenues include device identification and anomaly detection in industrial control systems and consumer “IoT” smart homes; streaming video quality estimation; content moderation on social media platforms (e.g., Facebook); and security vulnerability detection in networked systems.

    Mentor: Nick Feamster, Neubauer Professor of Computer Science; Director, Center for Data and Computing

    Nick Feamster is Neubauer Professor of Computer Science and the Director of Center for Data and Computing (CDAC) at the University of Chicago. Previously, he was a full professor in the Computer Science Department at Princeton University, where he directed the Center for Information Technology Policy (CITP); prior to Princeton, he was a full professor in the School of Computer Science at Georgia Tech.

    His research focuses on many aspects of computer networking and networked systems, with a focus on network operations, network security, and censorship-resistant communication systems. He received his Ph.D. in Computer science from MIT in 2005, and his S.B. and M.Eng. degrees in Electrical Engineering and Computer Science from MIT in 2000 and 2001, respectively. He was an early-stage employee at Looksmart (acquired by AltaVista), where he wrote the company’s first web crawler; and at Damballa, where he helped design the company’s first botnet-detection algorithm.

    Nick is an ACM Fellow. He received the Presidential Early Career Award for Scientists and Engineers (PECASE) for his contributions to cybersecurity, notably spam filtering. His other honors include the Technology Review 35 “Top Young Innovators Under 35” award, the ACM SIGCOMM Rising Star Award, a Sloan Research Fellowship, the NSF CAREER award, the IBM Faculty Fellowship, the IRTF Applied Networking Research Prize, and award papers at ACM SIGCOMM (network-level behavior of spammers), the SIGCOMM Internet Measurement Conference (measuring Web performance bottlenecks), and award papers at USENIX Security (circumventing web censorship using Infranet, web cookie analysis) and USENIX Networked Systems Design and Implementation (fault detection in router configuration, software-defined networking). His seminal work on the Routing Control Platform won the USENIX Test of Time Award for its influence on Software Defined Networking.

    Language learning has come to be a central theme in both cognitive science and artificial intelligence. The nature of language learning has long been a topic of interest for cognitive scientists, and machine learning has begun to dominate natural language processing (NLP) in modern AI. NLP systems have benefited tremendously from machine learning. However, the learning systems developed using these procedures often don’t achieve the efficiency and robustness of human language acquisition. Insights from language acquisition have the potential to help address this problem. But there are two critical challenges in exploring this possibility: (1) identifying the innate learning biases that enable fast, robust language learning in humans, and (2) determining how to translate theoretical insights about these biases into effective implementation for learning in NLP systems. This project will tackle both of these issues, making use of insights from special linguistic populations.

    Our first challenge –– identifying innate language learning predispositions –– is driven by the fact that most children are exposed to linguistic input from birth, making it difficult to disentangle innate characteristics versus characteristics that are rapidly learned from input. The rare cases in which children do not have usable linguistic input can help here by allowing us make important headway in identifying these predispositions. Congenitally deaf children who cannot learn the spoken language that surrounds them, and who have not been exposed to sign language by their hearing families, are in the unique situation of being without language input early in life. These children use their hands to communicate –– they gesture –– and those gestures (called “homesigns”) take on many, but not all, of the forms and functions of languages that have been handed down from generation to generation. The properties of these naturally-arising gestures provide evidence for the nature of linguistic predispositions independent of input.

    Drawing candidate biases from homesign, we will then tackle the second challenge –– incorporating biases into machine learning systems –– by systematic testing of models against real-world child language acquisition data. The goal of this phase will be to identify effective means of instantiating proposed human biases, and to test whether models incorporating these biases will successfully simulate the learning trajectories exhibited by children. Models with the proposed biases will be compared against minimally-different baseline models lacking the biases; stronger fit to human data will be taken as support that the biases are actual human predispositions. An important priority of this phase will be to balance scientific and engineering needs –– to maintain transparency of the models’ cognitive implications and to simulate human learning patterns as closely as possible, but also to use models that will interface smoothly with modern NLP systems, with promise to scale to larger datasets and broader domains.

    Mentor: Susan Goldin-Meadow, Beardsley Ruml Distinguished Service Professor in the Department of Psychology and Committee on Human Development

    Susan Goldin-Meadow is the Beardsley Ruml Distinguished Service Professor in the Department of Psychology and Committee on Human Development at the University of Chicago. A year spent at the Piagetian Institute in Geneva while an undergraduate at Smith College piqued her interest in the relationship between language and thought, interests she continued to pursue in her doctoral work at the University of Pennsylvania (Ph.D. 1975). At Penn and in collaboration with Lila Gleitman and Heidi Feldman, she began her studies exploring whether children who lack a (usable) model for language can nevertheless create a language with their hands. She has found that deaf children whose profound hearing losses prevent them from learning the speech than surrounds them, and whose hearing parents have not exposed them to sign, invent gesture systems which are structured in language-like ways. This interest in how the manual modality can serve the needs of communication and thinking led to her current work on the gestures that accompany speech in hearing individuals. She has found that gesture can convey substantive information – information that is often not expressed in the speech it accompanies. Gesture can thus reveal secrets of the mind to those who pay attention.

    Professor Goldin-Meadow’s research has been funded by the National Science Foundation, the Spencer Foundation, the March of Dimes, the National Institute of Child Health and Human Development, and the National Institute of Neurological and Communicative Disorders and Stroke. She has served as a member of the language review panel for NIH, has been a Member-at-Large to the Section on Linguistics and Language Science in AAAS, and was part of the Committee on Integrating the Science of Early Childhood Development sponsored by the National Research Council and the Institute of Medicine and leading to the book Neurons to Neighborhoods. She is a Fellow of AAAS, APS, and APA (Divisions 3 and 7). In 2001, she was awarded a Guggenheim Fellowship and a James McKeen Cattell Fellowship which led to her two recently published books, Resilience of Language and Hearing Gesture. In addition, she edited Language in Mind: Advances in the Study of Language and Thought in collaboration with Dedre Gentner. She has received the Burlington Northern Faculty Achievement Award for Graduate Teaching and the Llewellyn John and Harriet Manchester Quantrell Award for Excellence in Undergraduate Teaching at the University of Chicago. She is currently the President of the Cognitive Development Society and the editor of the new journal sponsored by the Society for Language Development, Language Learning and Development. Professor Goldin-Meadow also serves as chair of the developmental area program.

    Chenhao Tan is an assistant professor at the Department of Computer Science and the UChicago Data Science Institute. His main research interests include language and social dynamics, human-centered machine learning, and multi-community engagement. He is also broadly interested in computational social science, natural language processing, and artificial intelligence.

    Homepage.

    Project: Machine Learning for Physical Systems
    Much of machine learning is focused on recognizing patterns and making predictions based on training data. However, in many physical science settings, the complexity of the task is too high for effective learning giving the amount of available data. In these settings, it is essential to incorporate knowledge of the underlying physical system to mitigate the effect of limited data. Examples include using a combination of training data and models of a CT scanners operation to develop better medical image reconstruction methods, leveraging both observational and simulated data to develop better climate predictions, and building deep learning-based surrogate models for computationally demanding PDE-based simulators of physical systems. While there are isolated examples of successes in these regimes, little is known on a fundamental level. What are optimal machine learning methods that leverage both training data and physical models? How does sample complexity scale with the type of physical system and the accuracy of our models? Which kinds of PDE models are most amenable to deep surrogate models? This project will focus on developing new methodology and theory for machine learning for physical system that will address these and other open problems.

    Mentor: Rebecca Willett, Professor, Statistics, Computer Science, and the College

    Rebecca Willett is a Professor of Statistics and Computer Science at the University of Chicago. Her research is focused on machine learning, signal processing, and large-scale data science. She completed her PhD in Electrical and Computer Engineering at Rice University in 2005 and was an Assistant then tenured Associate Professor of Electrical and Computer Engineering at Duke University from 2005 to 2013. She was an Associate Professor of Electrical and Computer Engineering, Harvey D. Spangler Faculty Scholar, and Fellow of the Wisconsin Institutes for Discovery at the University of Wisconsin-Madison from 2013 to 2018. Willett received the National Science Foundation CAREER Award in 2007, is a member of the DARPA Computer Science Study Group, and received an Air Force Office of Scientific Research Young Investigator Program award in 2010. Willett has also held visiting researcher or faculty positions at the University of Nice in 2015, the Institute for Pure and Applied Mathematics at UCLA in 2004, the University of Wisconsin-Madison 2003-2005, the French National Institute for Research in Computer Science and Control (INRIA) in 2003, and the Applied Science Research and Development Laboratory at GE Healthcare in 2002.

    Project: Causality for Credible AI. The backbone of modern ML is evaluation on randomly held out data. However, models with strong held out performance frequently still exhibit disturbing failures. For example, they can degrade when deployed out of domain, learn to depend on apparently irrelevant input features, or change substantively in response to tiny variations in training procedure (e.g., changing the random seed). The aim of this project is to develop methods to mitigate this kind of failure by baking in causal knowledge to model design and evaluation. For example, we may know changing certain input features shouldn’t change predictions (e.g., your medical diagnosis shouldn’t change if we edit your zipcode), or we may know something about common causal structure between domains (rain should affect a self-driving car the same in Chicago and New York). The challenge here is to translate domain knowledge and desiderata into formal requirements, and then determine how to enforce and measure these using available data.

    Bio: I am an assistant professor of Statistics and Data Science at the University of Chicago and a research scientist at Google Cambridge. My recent work revolves around the intersection of machine learning and causal inference, as well as the design and evaluation of safe and credible AI systems. Other noteable areas of interests include network data, and the foundations of learning and statistical inference.

    I was previously a Distinguished Postdoctoral Researcher in the department of statistics at Columbia University, where I worked with the groups of David Blei and Peter Orbanz. I completed my Ph.D. in statistics at the University of Toronto, where I was advised by Daniel Roy. In a previous life, I worked on quantum computing at the University of Waterloo. I won a number of awards, including the Pierre Robillard award for best statistics thesis in Canada.

    Haifeng Xu is an assistant professor in the Department of Computer Science and the Data Science Institute at UChicago. He directs the Strategic IntelliGence for Machine Agents (SIGMA) research lab which focuses on designing algorithms/systems that can effectively elicit, process and exploit information, particularly in strategic environments. Haifeng has published more than 55 publications at leading venues on computational economics, machine learning and theoretical computer science, such as EC, ICML, NeurIPS, STOC and SODA. His research has been recognized by multiple awards, including the Google Faculty Research Award, ACM SIGecom Dissertation Award (honorable mention), IFAAMAS Victor Lesser Distinguished Dissertation Award (runner-up), Google PhD fellowship, and multiple best paper awards.

    The following research themes are the recent focus of our research lab. Please refer to our lab’s website for more details.

    • The economics of data/information, including selling, acquiring, and exploiting information
    • Machine learning in multi-agent setups under information asymmetry, incentive conflicts, and deception
    • Resource allocation in adversarial domains, with applications to security and privacy protection

    Medicine & Health

    Project: Creating Personalized Incentives to Drive Diabetes Patients’ Behavior

    Physical exercise can product a significant health benefit for diabetics, but not all patients have the same natural inclinations for exercise. We are working to develop a heterogenous treatment model that would create individualized incentives to help diabetes patients succeed at an exercise regimen. Come have an outsized impact on this project in its early stages as we formulate initial data needs and begin sourcing additional measures with which to create our model.

    Mentor: Rebecca Dizon-Ross, Associate Professor of Economics and Charles E. Merrill Faculty Scholar, Booth School of Business

    Rebecca Dizon-Ross is a development economist with an interest in human capital. Much of her current work is on the demand-side, aiming to understand the determinants of households’ investments in health and education.

    Before joining Booth, Dizon-Ross was a Prize Fellow in Economics, History, and Politics at Harvard University and a Postdoctoral Fellow in the Abdul Latif Jameel Poverty Action Lab at the Massachusetts Institute of Technology. She received a Ph.D. in Economics from Stanford University and a B.A. (summa cum laude) from Harvard University. Prior to graduate school, she worked as an analyst at McKinsey & Co.

    Project: Nightingale Project

    Machine learning, we are told, will transform medical diagnosis and patient care: by integrating ‘big data’ on patients’ history and physiology, algorithms can dramatically improve the quality of doctors’ decisions, with the potential both to reduce waste, avoid misdiagnosis, and produce breakthrough discoveries. For example, if massive datasets of ECG waveforms could be linked to national mortality registries, we could supercharge the current research, and find better, more consistent ways to allocate life-saving defibrillators. But most clinical data like this is siloed by different institutions and unavailable to researchers. Further complicating things, in order to protect patient privacy, public medical datasets are almost universally limited to a single, easily de-identified stream of information, like a set of X-rays.

    The goal of the Nightingale Project in the Booth Center for Applied AI is to gather and share just the sort of rich, multidimensional data needed to feed AI-enabled discovery. Work with a team of engineers, data analysts, and medical experts to build a secure platform that can warehouse curated de-identified clinical datasets linked to ground truth outcomes. Using this initiative as a proof of concept to develop other privacy tools, own and work through the de-identification, sharing, and privacy components of large-scale datasets — work that will include writing models and creating security-related challenges — with the goal of making the data available to researchers securely.

    Mentor: Sendhil Mullainathan, Faculty Director, Center for Applied AI, Roman Family University Professor of Computation and Behavioral Science, Chicago Booth

    Sendhil Mullainathan is the Roman Family University Professor of Computation and Behavioral Science at Chicago Booth. His current research uses machine learning to understand complex problems in human behavior, social policy, and especially medicine, where computational techniques have the potential to uncover biomedical insights from large-scale health data. He currently teaches a course on Artificial Intelligence.

    In past work he has combined insights from economics and behavioral science with causal inference tools—lab, field, and natural experiments—to study social problems such as discrimination and poverty. Papers include: the impact of poverty on mental bandwidth; how algorithms can improve on judicial decision-making; whether CEO pay is excessive; using fictitious resumes to measure discrimination; showing that higher cigarette taxes makes smokers happier; and modeling how competition affects media bias.

    Mullainathan enjoys writing. He recently co-authored Scarcity: Why Having too Little Means so Much and writes regularly for the New York Times. Additionally, his research has appeared in a variety of publications including the Quarterly Journal of Economics, Science, American Economic Review, Psychological Science, the British Medical Journal, and Management Science.

    Mullainathan helped co-found a non-profit to apply behavioral science (ideas42), co-founded a center to promote the use of randomized control trials in development (the Abdul Latif Jameel Poverty Action Lab), serves on the board of the MacArthur Foundation, has worked in government in various roles, is affiliated with the NBER and BREAD, and is a member of the American Academy of Arts and Sciences.

    Prior to joining Booth, Mullainathan was the Robert C. Waggoner Professor of Economics in the Faculty of Arts and Sciences at Harvard University, where he taught courses about machine learning and big data. He began his academic career at the Massachusetts Institute of Technology.

    Mullainathan is a recipient of the MacArthur “Genius Grant,” has been designated a “Young Global Leader” by the World Economic Forum, was labeled a “Top 100 Thinker” by Foreign Policy Magazine, and was named to the “Smart List: 50 people who will change the world” by Wired Magazine (UK).

    Project: Machine Learning to Improve Targeted Cancer Therapy

    Advances in genomics have created many new cancer therapies that target specific genetic or molecular features, raising the potential for personalized treatment that improves effectiveness and decreases side effects. However, the majority of patients treated with targeted therapies do not respond as predicted, and detailed patient genomic information is expensive to acquire.

    This collaboration proposes to improve targeting of cancer therapies by developing new AI approaches that recommend the best treatment based on a combination of genetic and pathology data. Researchers will build new methods drawing upon computer vision and machine learning to gather essential contextual information about individual cancers from tumor samples, utilizing both genomic and image-based features. This work will advance both artificial intelligence and cancer-focused data science by developing innovative solutions to emerging machine learning problems in cancer research, ultimately benefiting patients through more targeted, effective treatments.

    Mentors: Samantha Riesenfeld and Alex Pearson

    Medical oncologist Alexander T. Pearson, MD, PhD, cares for adults with head and neck cancers, especially squamous cell carcinoma, a common form of cancer that develops in the squamous cells that make up the middle and outer layer lining the mouth and throat. He also specializes in salivary gland cancers.

    His research combines laboratory experiments and mathematical models to more fully understand how head and neck cancers form and how to better design treatments for these cancers. He is currently the principal investigator on an NIH-funded study on the development of combination therapies for head and neck cancer.

    Dr. Pearson has more than a dozen peer-reviewed publications in the primary literature, has had several abstracts selected for oral presentation and is editor of the Oncology Boards Flash Review, 2nd edition.

    Project: Machine Learning to Improve Targeted Cancer Therapy

    Advances in genomics have created many new cancer therapies that target specific genetic or molecular features, raising the potential for personalized treatment that improves effectiveness and decreases side effects. However, the majority of patients treated with targeted therapies do not respond as predicted, and detailed patient genomic information is expensive to acquire.

    This collaboration proposes to improve targeting of cancer therapies by developing new AI approaches that recommend the best treatment based on a combination of genetic and pathology data. Researchers will build new methods drawing upon computer vision and machine learning to gather essential contextual information about individual cancers from tumor samples, utilizing both genomic and image-based features. This work will advance both artificial intelligence and cancer-focused data science by developing innovative solutions to emerging machine learning problems in cancer research, ultimately benefiting patients through more targeted, effective treatments.

    Mentors: Samantha Riesenfeld and Alex Pearson

    Samantha Riesenfeld is an Assistant Professor of Molecular Engineering and of Genetic Medicine, a member of the Committee on Immunology, an Associate Member of the Comprehensive Cancer Center, and co-director of the new Computational and Systems Immunology PhD track in Immunology and Molecular Engineering. She leads an interdisciplinary research program focused on developing and applying genomics-based machine learning approaches to investigate the cellular components, transcriptional circuitry, and dynamics underlying complex biological systems, with a special interest in inflammatory immune responses and solid tumor cancer.

    Project: Early Childhood Metric Initiative

    Early childhood suffers from a lack of quantifiable data that is easy to collect at scale. Work with a team of engineers, computer scientists, and early childhood experts to build a non-intrusive, wearable technology that leverages machine-learning to collect real-time, real-world data to measure young children’s early language environments (i.e. the quantity and quality of language interactions they are exposed to). As part of the project, develop machine learning algorithms and models to analyze the large-scale adult-child interaction data collected from this wearable technology. This large dataset will enable researchers and practitioners to better understand the relationships between family demographics, parental inputs, and child outcomes and identify effective approaches, as well as enable policymakers to hone in on the most effectives programs and policies to enhance children’s early language environments. This audio dataset will also allow experts in natural language processing to develop and refine speech processing algorithms.

    Mentor: Dana Suskind, MD, Professor of Surgery and Pediatrics, Director, Pediatric Cochlear Implantation Program, Co-Director, Thirty Million Words (TMW) Center for Early Learning + Public Health

    Dana Suskind, MD, is a pediatric otolaryngologist who specializes in hearing loss and cochlea implantation. She directs the University of Chicago Medicine’s Pediatric Hearing Loss and Cochlear Implant program.

    Recognized as a national thought leader in early language development, Dr. Suskind has dedicated her research and clinical life to optimizing foundational brain development and preventing early cognitive disparities and their lifelong impact. She is founder and co-director of the TMW Center for Early Learning + Public Health, which aims to create a population-level shift in the knowledge and behavior of parents and caregivers to optimize the foundational brain development in children from birth to five years of age, particularly those born into poverty.

    Her book “Thirty Million Words: Building a Child’s Brain” was published in 2015.

    Dr. Suskind has received several awards for her work, including the Weizmann Women for Science Vision and Impact award, the SENTAC Gray Humanitarian Award, the LENA Research Foundation Making a Difference Award, the Chairman’s Award from the Alexander Graham Bell Association for the Deaf and Hard of Hearing in 2018, and the John D. Arnold, MD Mentor Award for Sustained Excellence from the Pritzker School of Medicine.

    Project: Pediatric Cancer Data Commons

    Collecting, aggregating, harmonizing, and sharing data from children with cancer is essential to making new discoveries and developing new cures. Too often, data are siloed and disconnected, drastically reducing the usefulness of these valuable resources. The Pediatric Cancer Data Commons (PCDC) at UChicago brings together researchers from around the world with the goal of building data dictionaries for all types of pediatric cancer. Consensus data models are balloted with experts from around the world, including clinicians, ontologists/taxonomists, statisticians, and data scientists. The resulting dictionary is used for harmonizing data from completed clinical trials and is subsequently leveraged as a framework for collecting data on new studies. The data are made available to the worldwide research community through a public-facing cohort discovery tool. Data are further connected to other sources through common identifiers, allowing novel new data sets to be developed for research and discovery.

    Potential areas of research include: ontology development and data dictionary creation, data harmonization, automated methods of metadata extraction and data ingestion, development and deployment of novel data visualization tools and analytics, data governance and provenance methods and tools, developing novel methods of combining disparate data sets, and developing analytic methods for new modes of risk stratification. Experience with clinical data is preferred but not required.

    Mentor: Samuel L. Volchenboum, Associate Professor of Pediatrics & Associate Chief Research Informatics Officer, UChicago Medicine

    Samuel L. Volchenboum, MD, PhD, MS, is an expert in pediatric cancers and blood disorders. He has a special interest in treating children with neuroblastoma, a tumor of the sympathetic nervous system.

    In addition to caring for patients, Dr. Volchenboum studies ways to harness computers to enable research and foster innovation using large data sets. He directs the development of the International Neuroblastoma Risk Group Database project, which connects international patient data with external information such as genomic data and tissue availability. The Center he runs provides computational support for the Biological Sciences Division at the University of Chicago, including high-performance computing, applications development, bioinformatics, and access to the clinical research data warehouse.

    Public Policy & Society

    Project: Corporate influence on the rulemaking process within the United States

    The rulemaking process in the United States includes an opportunity for public comment in between a new regulation and implementation of a rule change. During this comment period, not only the public at large, but corporations as well are able to exert influence over rule changes. Using textual analysis techniques, we aim to understand how this process of influence works – to what extent corporate influence affects rule changes and what kinds of changes ultimately result.

    Mentor: Marianne Bertrand, Chris P. Dialynas Distinguished Service Professor of Economics and Willard Graham Faculty Scholar, Booth School of Business

    Marianne Bertrand is the Chris P. Dialynas Distinguished Service Professor of Economics at the University of Chicago Booth School of Business. She is a Research Fellow at the National Bureau of Economic Research, the Center for Economic Policy Research, and the Institute for the Study of Labor.

    Professor Bertrand is an applied micro-economist whose research covers the fields of labor economics, corporate finance, and development economics. Her research in these areas has been published widely, including numerous research articles in the Quarterly Journal of Economics, the Journal of Political Economy, the American Economic Review, and the Journal of Finance.

    Professor Bertrand is Faculty Director of Chicago Booth’s Rustandy Center for Social Sector Innovation and the Faculty Director of the Poverty Lab at the University of Chicago Urban Labs. Professor Bertrand also serves as co-editor of the American Economic Review.

    She has received several awards and honors, including the 2004 Elaine Bennett Research Prize, awarded by the American Economic Association to recognize and honor outstanding research in any field of economics by a woman at the beginning of her career, and the 2012 Society of Labor Economists’ Rosen Prize for Outstanding Contributions to Labor Economics. She is a Fellow of the American Academy of Arts and Sciences.

    Born in Belgium, Professor Bertrand received a Bachelor’s Degree in economics from Belgium’s Universite Libre de Bruxelles in 1991, followed by a Master’s Degree in econometrics from the same institution the next year. She moved to the United States in 1993 and earned a Ph.D. in economics from Harvard University in 1998. She was a faculty member in the Department of Economics at Princeton University for two years before joining Chicago Booth in 2000.

    Project: Data science relies on people at every part of the data science pipeline from deciding what data to collect, how to clean it, what algorithms to use, to eventually determining how to interpret, visualize, and explain the results in a human-friendly manner. Throughout this process, there may also be privacy and security considerations to make at various stages. I am interested in the human aspects of data science and how to help translate data into meaningful results, avoid bias, and manage privacy and security concerns and potential projects that touch on these user-centered issues.

    Bio: Marshini Chetty is an assistant professor in the Department of Computer Science at the University of Chicago where she directs the Amyoli Internet Research Laboratory (AIR lab). She specializes in human-computer interaction, usable privacy and security, and ubiquitous computing. Her work has won best paper awards at SOUPS, CHI, and CSCW, and she was a co-recipient of the Annual Privacy Papers for Policymakers award. Her research has been featured in the NYTimes, CNN, Washington Journal, BBC, Chicago Tribune, The Guardian, WIRED, and Slashdot. She has received generous funding from the National Science Foundation, through grants and a CAREER award, as well as the National Security Agency, Facebook, and multiple Google Faculty Research Awards.

     

    Prior to this position, Marshini was research faculty in the Department of Computer Science at Princeton University where she founded and directed the Princeton Human Computer Interaction Laboratory. Before working at Princeton, Marshini was an assistant professor at the College of Information Studies at the University of Maryland, College Park where she directed the NetCHI laboratory. In the past, Marshini also completed two post-doctoral research fellowships at ResearchICTAfrica in Cape Town, South Africa and with Prof. W. Keith Edwards at the College of Computing at  Georgia Institute of Technology. Marshini received her Ph.D. in Human-Centered Computing from Georgia Institute of Technology where she was advised by Prof. Rebecca E. Grinter. She started her journey in the USA after she completed her MSc., BSc.(Hons), and BSc. in Computer Science at the University of Cape Town, South Africa (her beautiful home country).

    Project: Combining human and machine intelligence for policy impact

    The success of artificial intelligence (AI) for engineering and commercial applications has led to growing interest in using these tools to help solve important social problems like inequality in income, education, health, or criminal justice system involvement. But any realistic assessment of how AI will be used in these areas suggests it will be a complement to, not substitute for, human judgment. That is, AI will be used as decision aids, not decision makers. In previous work (Kleinberg, Lakkaraju, Leskovec, Ludwig and Mullainathan, 2018 Quarterly Journal of Economics) we have found in the context of criminal justice system decision-making that humans on net add negative value to the machine’s predictions of defendant risk, although in principle the private information humans can access that algorithms cannot (such as courtroom discussion about the details of the case) could help the human add positive value in at least some cases. Similar issues arise in numerous other policy domains such as medical diagnosis, hiring, credit, and education admissions. The goal of this project is to better understand the potential sources of human and machine comparative advantage by measuring the private information humans have access to in different decision-making domains, trying to understand what are useful sources of signal versus sources of noise for human decisions about when to follow versus over-ride the algorithm’s recommendations, and then try to build decision-making systems that lead to the human plus machine together to outperform the decisions implied by the machine’s predictions alone.

    Mentor: Jens Ludwig, Edwin A. and Betty L. Bergman Distinguished Service Professor, Harris School of Public Policy, Director of University of Chicago Crime Lab, Co-director Education Lab

    Jens Ludwig is the Edwin A. and Betty L. Bergman Distinguished Service Professor, director of the University of Chicago’s Crime Lab, codirector of the Education Lab, and codirector of the National Bureau of Economic Research’s working group on the economics of crime.

    In the area of urban poverty, Ludwig has participated since 1995 on the evaluation of a HUD-funded randomized residential-mobility experiment known as Moving to Opportunity (MTO), which provides low-income public housing families the opportunity to relocate to private-market housing in less disadvantaged neighborhoods. In the area of education he has written extensively about early childhood interventions, and about the role of social conditions in affecting children’s schooling outcomes. In the area of crime, Ludwig has written extensively about gun-violence prevention. Through the Crime Lab he is also involved in partnering with policymakers in Chicago, New York City, and across the country to use tools from social science, behavioral science, and computer science to identify effective (and cost-effective) ways to help prevent crime and violence. This includes studies of various social programs, helping the Chicago Police Department use data to reduce gun violence and strengthen police-community relations, and work underway to use data science to help New York City build and implement a new pretrial risk tool as part of the city’s goal to close Riker’s Island. Crime Lab projects have helped redirect millions of dollars of public-sector resources to evidence-based strategies and have been featured in national news outlets such as the New York Times, Washington Post, Wall Street Journal, PBS News Hour and National Public Radio. In 2014 the Crime Lab was the recipient of a $1 million MacArthur Award for Creative and Effective Institutions, the organizational equivalent of the foundation’s “genius prize.”

    His research has been published in leading scientific journals across a range of disciplines including Science, New England Journal of Medicine, Journal of the American Medical Association, American Economic Review, Quarterly Journal of Economics, the Economic Journal, and the American Journal of Sociology. His coauthored article on race, peer norms, and education with Philip Cook was awarded the Vernon Prize for the best article in the Journal of Policy Analysis and Management. He is also coauthor with Cook of Gun Violence: The Real Costs (Oxford University Press, 2000), coeditor with Cook of Evaluating Gun Policy (Brookings Institution Press, 2003), and coeditor with Cook and Justin McCray of Controlling Crime: Strategies and Tradeoffs (University of Chicago Press, 2012).

    Prior to coming to Harris, Ludwig was a professor of public policy at Georgetown University. He is currently on the editorial board of the American Economic Review and was formerly coeditor of the Journal of Human Resources, and currently serves on the National Academy of Sciences Committee on the Neurobiological and Socio-behavioral Science of Adolescent Development and Its Applications. In 2012 he was elected vice president of the Association for Public Policy Analysis and Management (APPAM), the professional society for public policy schools. Ludwig received his BA in economics from Rutgers College and his MA and PhD in economics from Duke University. In 2006 he was awarded APPAM’s David N. Kershaw Prize for Contributions to Public Policy by Age 40. In 2012 he was elected to the Institute of Medicine of the National Academies of Science.

    I am an Assistant Professor in the Department of Political Science at The University of Chicago.

    I work on quantitative methodology for social science research, with a focus on causal inference, machine learning, and experimental design–particularly for adaptive experiments. My PhD is from Yale, joint in Political Science and Statistics & Data Science.

    Previously, I was a post-doctoral fellow in Susan Athey’s Golub Capital Social Impact Lab at the Stanford Graduate School of Business.

    In addition to the PhD, I hold a Masters in Statistics, also from Yale, and a Masters in Public Affairs, from the Princeton School of Public and International Affairs. My undergraduate degree was in cultural anthropology from Grinnell College; after college, I spent a year in Lesotho, teaching high school students, and two years in Madagascar, as a Peace Corps volunteer.

    Website

    Project: Preventing violent encounters with first responders

    People who live with serious mental illness or related challenges face heightened risks of violent encounters with first responders. Administrative and qualitative data from police, fire, and other first responders allow us to identify individuals, places, and events associated with such violent encounters. This project will use predictive analytics to improve preventive services and emergency responses for individuals and families who face these risks.

    Project: Predicting mortality among high-users of safety-net services in Illinois

    Individuals who pass through jails, homeless services, and other safety-net institutions face severe risks of premature mortality from opioid overdose, homicide, and other causes. This machine learning project uses integrated administrative data from diverse city, county, and state data sources in Illinois to identify key risk-factors for premature mortality.

    Mentor: Harold Pollack, Helen Ross Professor of Social Service Administration, Co-director, University of Chicago Health Lab

    Harold Pollack is the Helen Ross Professor at the School of Social Service Administration. He is also an Affiliate Professor in the Biological Sciences Collegiate Division and the Department of Public Health Sciences.

    Co-founder of the University of Chicago Crime Lab, he is co-director of the University of Chicago Health Lab. He is a committee member of the Center for Health Administration Studies (CHAS) at the University of Chicago. His current NIH-funded research concerns improved services for individuals at the boundaries of the behavioral health and criminal justice systems, disabilities, and two major new efforts to address the opioid epidemic in Illinois and across the nation.

    Past President of the Health Politics and Policy section of the American Political Science Association, Professor Pollack has been appointed to three committees of the National Academy of Sciences. He received his undergraduate degree, magna cum laude, in Electrical Engineering and Computer Science from Princeton University. He holds master’s and doctorate degrees in Public Policy from the Kennedy School of Government, Harvard University. Before coming to SSA, Professor Pollack was a Robert Wood Johnson Foundation Scholar in Health Policy Research at Yale University and taught Health Management and Policy at the University of Michigan School of Public Health.

    He has published widely at the interface between poverty policy and public health. His research appears in such journals as Addiction, Journal of the American Medical Association, American Journal of Public Health, Health Services Research, Pediatrics, and Social Service Review. His journalism regularly appears in such outlets as Washington Post, the Nation, the New York TimesNew Republic, and other popular publications. His American Prospect essay, “Lessons from an Emergency Room Nightmare” was selected for the collection Best American Medical Writing, 2009.

    Project: Understanding the Effects of Gender on Policy Through Textual Analysis of Congressional Data

    Congress generates a substantial amount of textual data from its hearings, meetings, speeches, etc. By conducting sentiment analysis on this data, we are trying to understand how gender influences the likelihood of member participation and the resulting policy decisions. Currently, this project has preliminary results and is entering a second phase of analysis, which will involve additional scraping of data and the generation of new methodologies for textual analysis.

    Mentor: Heather Sarsons, Assistant Professor of Economics and Diane Swonk Faculty Fellow, Booth School of Business

    Heather Sarsons is an economist with research interests in labor, personnel, and behavioral economics. Much of her work focuses on understanding how norms, stereotypes, and biases influence labor market outcomes and inequality.

    Prior to joining Booth, Sarsons was a post-doctoral fellow at the University of Toronto’s GATE Institute and the U of T Economics Department.

    Sarsons received a PhD in economics from Harvard, and a BA in economics from The University of British Columbia. While pursuing her PhD, she was also a visiting student at the London School of Economics.

    Aaron is an Assistant Professor in the Statistics Department and Data Science Institute at UChicago. His research develops methodology in Bayesian statistics, causal inference, and machine learning for applied problems in political science, economics, and genetics, among other fields. Prior to joining UChicago, Aaron was a postdoctoral fellow in the Data Science Institute at Columbia University. He received his PhD in Computer Science from UMass Amherst, as well as an MA in Linguistics and a BA in Political Science. He is on Twitter @AaronSchein.

    Chenhao Tan is an assistant professor at the Department of Computer Science and the UChicago Data Science Institute. His main research interests include language and social dynamics, human-centered machine learning, and multi-community engagement. He is also broadly interested in computational social science, natural language processing, and artificial intelligence.

    Homepage.

    Project: News-based Sentiment Analysis to Understand Market Movement

    Most financial analysis is quantitative, but there is a wealth of data contained in textual artifacts as well. In this project, we are using natural language processing methodologies to conduct sentiment analysis on global news reports in order to better understand how news-based sentiment affects the movement of markets. Other possible avenues of investigation using new NLP methodologies include understanding the macro effects of global economic sentiment and attempting to detect the existence of fake news.

    Mentor: Dacheng Xiu, Professor of Econometrics and Statistics, Booth School of Business

    Dacheng Xiu’s research interests include developing statistical methodologies and applying them to financial data, while exploring their economic implications. His earlier research involved risk measurement and portfolio management with high-frequency data and econometric modeling of derivatives. His current work focuses on developing machine learning solutions to big-data problems in empirical asset pricing.

    Xiu’s work has appeared in Econometrica, the Journal of Econometrics, the Journal of the American Statistical Association, the Annals of Statistics, and the Journal of Finance. He is a Co-Editor for the Journal of Financial Econometrics, an Associate Editor for the Journal of Econometrics, the Journal of Business & Economic Statistics, the Journal of Empirical Finance, and Statistica Sinica, and also referees for several journals in the fields of econometrics, statistics, and finance. He has received several recognitions for his research, including the Fellow of the Society for Financial Econometrics, the Fellow of the Journal of Econometrics, the 2018 Swiss Finance Institute Outstanding Paper Award, the 2018 AQR Insight Award, and the Best Conference Paper Prize at the 2017 Annual Meeting of the European Finance Association.

    In 2017, Xiu launched a website that provides up-to-date realized volatilities of individual stocks, as well as equity, currency, and commodity futures. These daily volatilities are calculated from the intraday transactions and the methodologies are based on his research of high-frequency data.

    Xiu earned his PhD and MA in applied mathematics from Princeton University, where he was also a student at the Bendheim Center for Finance. Prior to his graduate studies, he obtained a BS in mathematics from the University of Science and Technology of China.

    Aloni Cohen is an Assistant Professor of Computer Science and Data Science at the University of Chicago. Previously, Cohen was a Postdoctoral Associate at Boston University, with a joint appointment at the Hariri Institute for Computing and the School of Law. His research explores the interplay between theoretical cryptography, privacy, law, and policy. Aloni earned his PhD in electrical engineering and computer science at MIT where he was advised by Shafi Goldwasser and supported by a Facebook Fellowship and an NSF Graduate Student Fellowship. Aloni is a former affiliate at the Berkman Klein Center for Internet & Society and a Fellow at the Aspen Tech Policy Hub.

    My research explores the interplay between theoretical cryptography, privacy, law, and policy. Specifically, I aim to understand and resolve the tensions between the theory of cryptography and the privacy and surveillance law that governs its eventual real-world context. Right now, I’m thinking about differential privacy, GDPR, the Fifth Amendment, encryption, multiparty computation, and the Census.

    Visualization and Communication

    Alex Kale is an Assistant Professor in Computer Science and the Data Science Institute at the University of Chicago. Previously, he earned his PhD at the University of Washington where he worked with Jessica Hullman. Alex leads the Data Cognition Lab, focused on creating data visualization and analysis software that explicitly represents the user’s cognitive processes.

    Alex creates and evaluates tools for helping people think with data, specializing in data visualization and reasoning with uncertainty. He publishes in top human-computer interaction and data visualization venues such as ACM CHI and IEEE VIS, where his work has won Best Paper and Honorable Mention Awards. Alex’s work addresses gaps in dominant theories and models of what makes visualization effective for inferences and decision making.

     

  • Committee

    DSI Postdoctoral Program Committee

    Chibueze Amanchukwu is a Neubauer Family Assistant Professor at the Pritzker School of Molecular Engineering at the University of Chicago. He received his bachelor’s degree in chemical engineering from Texas A&M University as the department’s Outstanding Graduating Student, and his PhD in chemical engineering from the Massachusetts Institute of Technology.

    As a graduate student with Paula Hammond, he elucidated polymer degradation mechanisms and tuned polymer electrolyte behavior in lithium-air batteries. His graduate work was supported by the National Defense Science and Engineering Graduate (NDSEG) Fellowship, GEM Fellowship, and the Alfred P. Sloan Minority Fellowship. As a postdoctoral fellow with Zhenan Bao at Stanford University, he developed new small molecule electrolytes that decoupled ionic conductivity from electrochemical instability for lithium metal batteries. His postdoctoral work was supported by the TomKat Center Postdoctoral Fellowship in Sustainable Energy at Stanford. His research has been recognized with awards from the American Chemical Society (Excellence in Graduate Polymer Research) and the American Institute of Chemical Engineers (Session’s Best Paper).

    Raul Castro Fernandez is an Assistant Professor of Computer Science at the University of Chicago. In his research he builds systems for discovering, preparing, and processing data. The goal of his research is to understand and exploit the value of data. He often uses techniques from data management, statistics, and machine learning. His main effort these days is on building platforms to support markets of data. This is part of a larger research effort on understanding the Economics of Data. He’s part of ChiData, the data systems research group at The University of Chicago.

    Homepage.

    I am an Assistant Professor of Computer Science at the University of Chicago. I founded and direct 3DL (threedle! ), a group of enthusiastic researchers passionate about 3D, machine learning, and visual computing. I obtained my Ph.D. in 2021 from Tel Aviv University under the supervision of Daniel Cohen-Or and Raja Giryes.

    My research is focused on building artificial intelligence for 3D data, spanning the fields of computer graphics, machine learning, and computer vision. Deep learning, the most popular form of artificial intelligence, has unlocked remarkable success on structured data (such as text, images, and video), and I am interested in harnessing the potential of these techniques to enable effective operation on unstructured 3D geometric data.

    We have developed a convolutional neural network designed specifically for meshes, and also explored how to learn from the internal data within a single shape (for surface reconstructiongeometric texture synthesis, and point cloud consolidation) – and I am interested in broader applications related to these areas. Additional research directions that I am aiming to explore include: intertwining human and machine-based creativity to advance our capabilities in 3D shape modeling and animation; learning with less supervision, for example to extract patterns and relationships from large shape collections; and making 3D neural networks more “interpretable/explainable”.

    Sanjay Krishnan is an Assistant Professor of Computer Science. His research group studies the theory and practice of building decision systems that are robust to corrupted, missing, or otherwise uncertain data. His research brings together ideas from statistics/machine learning and database systems. His research group is currently studying systems that can analyze large amounts of video, certifiable accuracy guarantees in partially complete databases, and theoretical lower-bounds for lossy compression in relational databases.

    Homepage.

    Aaron is an Assistant Professor in the Statistics Department and Data Science Institute at UChicago. His research develops methodology in Bayesian statistics, causal inference, and machine learning for applied problems in political science, economics, and genetics, among other fields. Prior to joining UChicago, Aaron was a postdoctoral fellow in the Data Science Institute at Columbia University. He received his PhD in Computer Science from UMass Amherst, as well as an MA in Linguistics and a BA in Political Science. He is on Twitter @AaronSchein.

    Chenhao Tan is an assistant professor at the Department of Computer Science and the UChicago Data Science Institute. His main research interests include language and social dynamics, human-centered machine learning, and multi-community engagement. He is also broadly interested in computational social science, natural language processing, and artificial intelligence.

    Website

    Samuel L. Volchenboum, MD, PhD, MS, is an expert in pediatric cancers and blood disorders. He has a special interest in treating children with neuroblastoma, a tumor of the sympathetic nervous system.

    In addition to caring for patients, Dr. Volchenboum studies ways to harness computers to enable research and foster innovation using large data sets. He directs the development of the International Neuroblastoma Risk Group Database project, which connects international patient data with external information such as genomic data and tissue availability. The Center he runs provides computational support for the Biological Sciences Division at the University of Chicago, including high-performance computing, applications development, bioinformatics, and access to the clinical research data warehouse.

    Homepage

  • FAQ
    • Frequently Asked Questions