Dan Gutfreund

Principal Research Scientist and Senior Manager, AI Models Science

Dan Gutfreund is a principal investigator at the MIT-IBM Watson AI Lab and a manager of the AI Models Science department. Before he moved to the Cambridge research lab, Gutfreund was at the Haifa Research Lab, where he held several managerial and technical leadership positions. In his last role, Gutfreund was the manager in charge of IBM Project Debater. In 2005, he received a PhD in computer science from the Hebrew University in Jerusalem, Israel. Before joining IBM in 2009, he was a postdoctoral fellow and a lecturer at Harvard University and MIT.

Gutfreund’s research interests are in machine learning with applications to natural language processing and computer vision. Previously, he worked on problems in theoretical computer science in the areas of computational complexity and foundations of cryptography.

Top Work

Moments in Time Dataset: one million videos for event understanding

Moments in Time Dataset: one million videos for event understanding

Computer Vision

ObjectNet: A bias-controlled dataset object recognition

ObjectNet: A bias-controlled dataset object recognition

Computer Vision

Publications with the MIT-IBM Watson AI Lab

Zero-shot linear combinations of grounded social interactions with Linear Social MDPs
Zero-shot linear combinations of grounded social interactions with Linear Social MDPs
 
How hard are computer vision datasets? Calibrating dataset difficulty to viewing time
How hard are computer vision datasets? Calibrating dataset difficulty to viewing time
 
Finding Fallen Objects Via Asynchronous Audio-Visual Integration
Finding Fallen Objects Via Asynchronous Audio-Visual Integration
 
A Bayesian-Symbolic Approach to Reasoning and Learning in Intuitive Physics
A Bayesian-Symbolic Approach to Reasoning and Learning in Intuitive Physics
 
ThreeDWorld: A Platform for Interactive Multi-Modal Physical Simulation
ThreeDWorld: A Platform for Interactive Multi-Modal Physical Simulation
 
3DP3: 3D Scene Perception via Probabilistic Programming
3DP3: 3D Scene Perception via Probabilistic Programming
 
Multi-Moments in Time: Learning and Interpreting Models for Multi-Action Video Understanding
Multi-Moments in Time: Learning and Interpreting Models for Multi-Action Video Understanding
 
AGENT: A Benchmark for Core Psychological Reasoning
AGENT: A Benchmark for Core Psychological Reasoning
 
ObjectNet: A bias-controlled dataset object recognition
ObjectNet: A bias-controlled dataset object recognition
 
ObjectNet: A large-scale bias-controlled dataset for pushing the limits of object recognition models
ObjectNet: A large-scale bias-controlled dataset for pushing the limits of object recognition models
 
SimVAE: Simulator-Assisted Training for Interpretable Generative Models
SimVAE: Simulator-Assisted Training for Interpretable Generative Models
 
Reasoning about Human-Object Interactions through Dual Attention Networks
Reasoning about Human-Object Interactions through Dual Attention Networks
 
Identifying Interpretable Action Concepts in Deep Networks
Identifying Interpretable Action Concepts in Deep Networks
 
Grounding Spoken Words in Unlabeled Video
Grounding Spoken Words in Unlabeled Video
 
Moments in Time Dataset: one million videos for event understanding
Moments in Time Dataset: one million videos for event understanding