Views Navigation

Event Views Navigation

Today

IDSS Distinguished Seminar – Essential Concepts of Causal Inference: A Remarkable History

MIT Building 32, Room 141 The Stata Center (32-141), 32 Vassar Street, Cambridge, MA, United States

  Abstract I believe that a deep understanding of cause and effect, and how to estimate causal effects from data, complete with the associated mathematical notation and expressions, only evolved in the twentieth century. The crucial idea of randomized experiments was apparently first proposed in 1925 in the context of agricultural field trails but quickly…

Connections between structured estimation and weak submodularity

MIT Building E18, Room 304 Ford Building (E18), 50 Ames Street, Cambridge, MA, United States

Abstract:  Many modern statistical estimation problems rely on imposing additional structure in order to reduce the statistical complexity and provide interpretability. Unfortunately, these structures often are combinatorial in nature and result in computationally challenging problems. In parallel, the combinatorial optimization community has placed significant effort in developing algorithms that can approximately solve such optimization problems…

Data Science and Big Data Analytics: Making Data-Driven Decisions

online

The seven-week course launches February 5, 2018. This course was developed by over ten MIT faculty members at IDSS. It is specially designed for data scientists, business analysts, engineers, and technical managers looking to learn the latest theories and strategies to harness data.

Machine Learning and Causal Inference

MIT Building 32, Room 141 The Stata Center (32-141), 32 Vassar Street, Cambridge, MA, United States

Abstract: This talk will review a series of recent papers that develop new methods based on machine learning methods to approach problems of causal inference, including estimation of conditional average treatment effects and personalized treatment assignment policies. Approaches for randomized experiments, environments with unconfoundedness, instrumental variables, and panel data will be considered. Bio: Susan Athey…

Variable selection using presence-only data with applications to biochemistry

MIT Building E18, Room 304 Ford Building (E18), 50 Ames Street, Cambridge, MA, United States

Abstract: In a number of problems, we are presented with positive and unlabelled data, referred to as presence-only responses. The application I present today involves studying the relationship between protein sequence and function and presence-only data arises since for many experiments it is impossible to obtain a large set of negative (non-functional) sequences. Furthermore, if…

User-friendly guarantees for the Langevin Monte Carlo

MIT Building E18, Room 304 Ford Building (E18), 50 Ames Street, Cambridge, MA, United States

Abstract: In this talk, I will revisit the recently established theoretical guarantees for the convergence of the Langevin Monte Carlo algorithm of sampling from a smooth and (strongly) log-concave density. I will discuss the existing results when the accuracy of sampling is measured in the Wasserstein distance and provide further insights on relations between, on the one…

Submodular Optimization: From Discrete to Continuous and Back

34-101

Abstract Many procedures in statistics and artificial intelligence require solving non-convex problems. Historically, the focus has been to convexify the non-convex objectives. In recent years, however, there has been significant progress to optimize non-convex functions directly. This direct approach has led to provably good guarantees for specific problem instances such as latent variable models, non-negative…

Optimization’s Implicit Gift to Learning: Understanding Optimization Bias as a Key to Generalization

MIT Building E18, Room 304 Ford Building (E18), 50 Ames Street, Cambridge, MA, United States

Abstract: It is becoming increasingly clear that implicit regularization afforded by the optimization algorithms play a central role in machine learning, and especially so when using large, deep, neural networks. We have a good understanding of the implicit regularization afforded by stochastic approximation algorithms, such as SGD, and as I will review, we understand and…

Provably Secure Machine Learning

32-G449 (Kiva/Patel)

Abstract:  The widespread use of machine learning systems creates a new class of computer security vulnerabilities where, rather than attacking the integrity of the software itself, malicious actors exploit the statistical nature of the learning algorithms. For instance, attackers can add fake data (e.g. by creating fake user accounts), or strategically manipulate inputs to the system…


© MIT Institute for Data, Systems, and Society | 77 Massachusetts Avenue | Cambridge, MA 02139-4307 | 617-253-1764 |