NIPS 2016 - Reliable machine learning in the wild Workshop

9 December 2016
Jacob Steinhardt, Dylan Hadfield-Menell, Adrian Weller, David Duvenaud, Percy Liang

A workshop at NIPS Conference, Barcelona

Speakers include:

  • Jacob Steinhardt
  • Dylan Hadfield-Menell
  • Adrian Weller
  • David Duvenaud
  • Percy Liang

More information

When will a system that has performed well in the past continue to do so in the future? How do we design such systems in the presence of novel and potentially adversarial input distributions? What techniques will let us safely build and deploy autonomous systems on a scale where human monitoring becomes difficult or infeasible? Answering these questions is critical to guaranteeing the safety of emerging high stakes applications of AI, such as self-driving cars and automated surgical assistants. This workshop will bring together researchers in areas such as human-robot interaction, security, causal inference, and multi-agent systems in order to strengthen the field of reliability engineering for machine learning systems. We are interested in approaches that have the potential to provide assurances of reliability, especially as systems scale in autonomy and complexity. We will focus on four aspects — robustness (to adversaries, distributional shift, model mis-specification, corrupted data); awareness (of when a change has occurred, when the model might be mis-calibrated, etc.); adaptation (to new situations or objectives); and monitoring (allowing humans to meaningfully track the state of the system). Together, these will aid us in designing and deploying reliable machine learning systems.

Sponsors: LFCI, CSER and Clifford Chance