Natasha Jaques
Natasha Jaques
Awards
Press
Featured
Publications
Topics
Talks
Communities
Light
Dark
Automatic
Communication and Language
Moral Foundations of Large Language Models
Moral Foundations theory decomposes human moral reasoning into five factors, which vary reliably across different human populations and political affiliations. We use moral foundations to analyze large language models like GPT-3 to determine what, if any, consistent moral values it brings to conversations, whether these can be deliberately manipulated, and whether holding a particular moral stance affects downstream tasks.
M. Abdulhai
,
C. Crepy
,
D. Valter
,
J. Canny
,
S. Levine
,
Natasha Jaques
2022
In
Preprint
Cite
Less is More: Generating Grounded Navigation Instructions from Landmarks
We study the automatic generation of natural language navigation instructions in visually realistic indoor environments. Existing generators suffer from poor visual grounding, skip steps, and hallucinate objects. We address this using a large language model which incorporates visual landmark detection.. The model dramatically increases the quality of generated instructions, such that humans can follow them with a 71% success rate (SR); just shy of the 75% SR of real human instructions.
S. Wang
,
C. Montgomery
,
J. Orbay
,
V. Birodkar
,
A. Faust
,
I. Gur
,
Natasha Jaques
,
A. Waters
,
J. Baldridge
,
P. Anderson
2022
In
Computer Vision and Pattern Recognition (CVPR)
PDF
Cite
Dataset
Human-Centric Dialog Training via Offline Reinforcement Learning
We train dialog models with interactive data from conversations with real humans, using a novel Offline RL technique based on KL-control. Rather than rely on manual ratings, we learn from implicit signals like sentiment, and show that this results in better performance.
Natasha Jaques
*
,
J. H. Shen
*
,
A. Ghandeharioun
,
C. Ferguson
,
A. Lapedriza
,
N. Jones
,
S. Gu
,
R. Picard
2020
In
Empirical Methods in Natural Language Processing (EMNLP)
PDF
Cite
Code
Dataset
Slides
EMNLP Talk
Approximating Interactive Human Evaluation with Self-Play for Open-Domain Dialog Systems
Existing metrics for automatically evaluating dialog models correlate poorly with human judgements, and are evaluated on static conversation snippets. Instead, we deploy bots to interact live with humans, then approximate human ratings with state-of-the-art accuracy using conversations generated with self-play.
A. Ghandeharioun
*
,
J. H. Shen
*
,
Natasha Jaques
*
,
C. Ferguson
,
N. Jones
,
A. Lapedriza
,
R. Picard
2019
In
Neural Information Processing Systems (NeurIPS)
PDF
Cite
Code
Dataset
Poster
Hierarchical Reinforcement Learning for Open-Domain Dialog
For the first time, we use hierarchical reinforcement learning to train open-domain dialog models, enabling the optimization of long-term, conversational, rewards, including reducing the toxicity of generated language. Our approach provides significant improvements over state-of-the-art dialog models.
A. Saleh
*
,
Natasha Jaques
*
,
A. Ghandeharioun
,
J. H. Shen
,
R. Picard
2019
In
Association for the Advancement of Artificial Intelligence (AAAI)
Oral (top 7.8% of submissions)
PDF
Cite
Code
Dataset
Talk
Social and Affective Machine Learning
My PhD Thesis spans both Social Reinforcement Learning and Affective Computing, investigating how affective and social intelligence can enhance machine learning algorithms, and how machine learning can enhance our ability to predict and understand human affective and social phenomena.
Natasha Jaques
2019
In
Massachusetts Institute of Technology
PDF
Cite
Thesis Defense
CV News write-up
Social Influence as Intrinsic Motivation for Multi-Agent Deep Reinforcement Learning
Social influence is a unified mechanism for achieving coordination and communication in Multi-Agent Reinforcement Learning, through rewarding agents for having causal influence over other agents’ actions, thus increasing mutual information between agents’ actions. Optimizing for influence leads to agents learning emergent communication protocols. Unlike prior work, influence can be computed in a fully decentralized manner.
Natasha Jaques
,
A. Lazaridou
,
E. Hughes
,
C. Gulcehre
,
P. A. Ortega
,
D. J. Strouse
,
J.Z. Leibo
,
N. de Freitas
2019
In
International Conference on Machine Learning (ICML)
Best Paper Honourable Mention (top 0.26% of submissions)
PDF
Cite
Poster
Slides
Videos
ICML talk
IEEE Spectrum article
ICML 2019 Best Papers
Sequence Tutor: Conservative Fine-Tuning of Sequence Generation Models with KL-control
To combine supervised learning on data with reinforcement learning, we pre-train a supervised data prior, and penalize KL-divergence from this model using RL training. This enables effective learning of complex sequence-modeling problems for which we wish to match the data while optimizing external metrics like drug effectiveness. The approach produces compelling results in the disparate domains of music generation and drug discovery.
Natasha Jaques
,
S. Gu
,
D. Bahdanau
,
J. M. Hernandez-Lobato
,
R. E. Turner
,
D. Eck
2017
In
International Conference on Machine Learning (ICML)
PDF
Cite
Code
ICML talk
Generated music
Magenta blog
MIT Tech Review article
Personality, Attitudes, and Bonding in Conversations
We collect observational data from real human conversations, and develop a measure of how much participants experienced bonding or chemistry. We analyze the effects of personality and attitudes on bonding, and find that attentiveness and excitement are more effective at promoting bonding than traits like attractiveness and humour.
Natasha Jaques
,
Y. K. Kim
,
\& Picard R. Picard R
2016
In
Intelligent Virtual Agents (IVA)
PDF
Cite
Understanding and Predicting Bonding in Conversations Using Thin Slices of Facial Expressions and Body Language
Given only one-minute slices of facial expressions and body language, we use machine learning to accurately predict whether two humans having a conversation will bond with each other. We analyze factors which lead to bonding and discover that synchrony in body language and appropriate, empathetic facial expressions lead to higher bonding.
Natasha Jaques
,
D. McDuff
,
Y. K. Kim
,
\& Picard R. Picard R
2016
In
Intelligent Virtual Agents (IVA)
PDF
Cite
Cite
×