Mahdi Haghifam

alt text  [Google Scholar] [Github] [Twitter] [Linkedin]
Email(preferred): haghifam.mahdi@gmail.com
Email: mhaghifam@ttic.edu


About Me

I am a ‪research assistant professor at the Toyota Technological Institute at Chicago (TTIC)‬. I was previously a Distinguished Postdoctoral Researcher at Khoury College of Computer Sciences at Northeastern University, fortunate to be working Jonathan Ullman‬ and Adam Smith‬. I completed my PhD at University of Toronto and Vector Institute‬ where I was fortunate to be advised by ‪Daniel M. Roy‬. I also received my B.Sc. and M.Sc. degrees in Electrical Engineering from Sharif University of Technology.

Recognitions of my work include a Best Paper Award at ICML 2024, Simons Institute-UC Berkeley Research Fellowship, as well as several honors for graduate research excellence from University of Toronto, including the Henderson and Bassett Research Fellowship and the Viola Carless Smith Research Fellowship. Additionally, I was recognized as a top reviewer at NeurIPS in 2021 and 2023.

Outside my research activities, I enjoy playing and watching soccer, reading classic literature, and baking.

Industry Internship Experience

Google DeepMind| Research Intern
Mountain View, CA | September 2022 – December 2022
Mentors: Thomas Steinke and Abhradeep Guha Thakurta
- Developed novel second-order privacy-preserving training algorithm achieving 10–40× improvement over baselines and Studied the impact of public data on privacy-preserving training algorithms
- Resulted in publications at NeurIPS 2023 (link) (code)
and ICML 2023 (link)

ServiceNow Research | Research Intern
Toronto, ON | November 2020 – March 2021
Mentor: Gintare Karolina Dziugaite
- Studied the connections between different generalization approaches in ML
- Resulted in publication in NeurIPS 2021 (Spotlight) (link)

ServiceNow Research | Research Intern
Toronto, ON | February 2019 – May 2019
Mentor: Gintare Karolina Dziugaite
- Proposed a new analytical technique that measures algorithmic stability on random subsets of data, creating a tighter and more empirically accurate connection between the training process and generalization performance, NeurIPS 2019 (link)
- Built generalization-prediction tools in TensorFlow for CNNs and MLP on image classification tasks. Our method achieves for the first time non-vacuous generalization bound (code)

Research Overview and Selected Papers

My research focuses on the foundations and principled algorithm design for ML. More broadly, I am interested in statistical learning theory, statistics, and information theory. The central goal of my research is to address practical challenges in ML by developing tools and algorithms with rigorous theoretical guarantees that assess and ensure validity. This work is crucial for building trustworthy ML systems in high-stakes applications, where balancing responsible deployment with strong empirical performance is essential. Some of the questions I have been thinking about: When and how can models generalize beyond their training data? Under what conditions do they memorize sensitive information? And how can we preserve privacy while still learning effectively from sensitive data?

Generalization in Machine Learning:

Memorization and Privacy Attacks:

Differential Privacy:

Contact Me!

Feel free to reach out if you'd like to discuss research ideas. Also, I'm happy to offer guidance and support to those applying to graduate programs, especially individuals who might not typically have access to such assistance