Talks and Presentations

Differentially Private Federated Learning with Time-Adaptive Privacy Spending

March 28, 2025

Seminar talk, ResearchTrend Connect session ~~~~(https://researchtrend.ai/social-events/researchtrend-connect-fedml),

In this talk, I presented my two Ph.D. projects that explore privacy-preserving machine learning mechanisms to promote more equitable and private collaborative learning. In the first project, we introduced a framework for learning with group identities, allowing individuals to share data within specific groups, such as business data within a company group or personal genomic data within their family group. We modeled and controlled privacy leakage propagation across potentially overlapping group structures. In the second project, we proposed a novel time-adaptive privacy spending mechanism, enabling participants to preserve more privacy during certain training rounds. Together, these works offer new perspectives on how trust and privacy can be formalized and quantified in federated learning systems.

Controlled Privacy Leakage Propagation Throughout Overlapping Grouped Learning

July 07, 2024

Conference talk, IEEE International Symposium on Information Theory (ISIT), Athens, Greece

In this talk, I described our framework for learning with group identities, where individuals may share data selectively within specific groups, such as contributing business data in their company group or personal genomic data in their family group. We designed the modeling and control of privacy leakage propagation across the potentially overlapping group structures.

Successive approximated coded matrix multiplication

June 28, 2022

Conference talk, IEEE International Symposium on Information Theory (ISIT), Aalto University, Espoo, Finaland

In this talk I describe some of our work on coded matrix multiplication when an approximate result will suffice. We are motivated by potential application in optimization and learning where an exact matric product is not required and one would prefer to get a lower-fidelity result faster. We are also motivated by developing rate-distorion analogs for coded computing and particularly by a recent JSAIT paper by Jeong et al. epsilon-coded-computing wherein the authors show that they can recover an intermediate, approximate, result half-way to exact recovery. In this talk I build on that prior work to show how to realize schemes in which there are multiple stages of recovery (more than one) en-route to exact recovery. In analog to successive refinement in rate distortion we terms this successive approximation coding. [Slides]

Random-Sampling Based Techniques for Approximated Matrix Multiplication

February 07, 2021

Seminar Talk, IT@UOFT seminars, Department of Electrical and Computer Engineering, University of Toronto, Toronto, ON

One of the most common, but at the same time expensive operations in linear algebra, is multiplying two matrices. With the rapid development of machine learning and increases in data volume, performing fast matrix intensive multiplications has become a major challenge. Two different approaches to overcome this issue are 1) to approximate the product and 2) to perform the multiplication distributively. In this talk, I focuse on the first approach and summarize some random-sampling based techniques for the approximation of the matrix-matrix multiplication, such as the work by Mahoney et al., in SIAM J. Comput’ 2006. [Link] [Slide]

Hierarchical Coded Matrix Multiplication

June 28, 2019

Poster presentation, NSERC COHESA (Computing Hardware for Emerging Intelligent Sensory Applications), Annual General Meeting, University of Toronto, Toronto, ON