cv

This is a description of the page. You can modify it in '_pages/cv.md'. You can also change or remove the top pdf download button.

Basics

Name Eduardo Fernandes Montesuma
Label AI Researcher
Email edumontesuma@gmail.com
Url https://github.com/eddardd
Summary AI researcher specializing in optimal transport, domain adaptation, and transfer learning with expertise in fine-tuning and adapting large foundation models.

Work

  • 2025.08 - Present
    AI Researcher
    Sigma Nova
    Conducting research primarily on the fine-tuning and adaptation of large pre-trained foundation models.
    • Foundation Models
    • Model Fine-tuning
    • Model Adaptation
  • 2024.10 - 2025.07
    Applied AI Researcher
    GetVocal
    R&D of conversational agents, focusing on automatic speech recognition, large language models and speech synthesis.
    • Conversational AI
    • Speech Recognition
    • Large Language Models
    • Speech Synthesis
  • 2021.10 - 2024.10
    PhD Student
    Commissariat pour l'Énergie Atomique
    Studied the contributions of Optimal Transport for Machine Learning, Transfer Learning and Domain Adaptation.
    • Optimal Transport
    • Transfer Learning
    • Domain Adaptation
    • Machine Learning
  • 2021.04 - 2021.09
    Scientific Researcher in Data Mining
    Dell Lead
    Mining textual data, building visualizations and exploring strategies for machine reading comprehension.
    • Data Mining
    • Text Analysis
    • Machine Reading Comprehension
    • Data Visualization
  • 2020.03 - 2020.08
    Research Intern in Artificial Intelligence
    Commissariat pour l'Énergie Atomique
    Implementing Optimal Transport for Domain Adaptation methods for transfer learning on bacteria cell stain classification.
    • Optimal Transport
    • Domain Adaptation
    • Transfer Learning
    • Computer Vision
  • 2019.05 - 2019.09
    Research Intern in Deep Learning for Image Restoration
    Institut d'Électronique et Télécommunications de Rennes
    Implementing software for comparing image restoration methods, with focus on deep neural network methods. Open-source solution hosted on Github.
    • Deep Learning
    • Image Restoration
    • Open Source Software
    • Neural Networks
  • 2016.08 - 2017.08
    Research Intern in Pattern Recognition
    Universidade Federal do Ceará
    Working on Computer Vision for Detection and Identification of Cervix Diseases (SISCOLO) project, studying and implementing classifiers for identifying cancer on cervix cells.
    • Computer Vision
    • Pattern Recognition
    • Medical Imaging
    • Cancer Detection

Education

  • 2021.10 - 2024.10

    Paris, Île-de-France, France

    PhD
    Université Paris-Saclay
    Data Science
    • Multi-Source Domain Adaptation through Wasserstein Barycenters
  • 2018.09 - 2020.08

    Rennes, Bretagne, France

    Master's Degree
    INSA de Rennes
    Electronic Engineering and Industrial Informatics
    • Optimal Transport for Domain Adaptation with Applications to Bacteria Stain Classification
  • 2015.01 - 2021.04

    Fortaleza, Ceará, Brazil

    Bachelor's Degree
    Universidade Federal do Ceará
    Computer Engineering
    • Cross-Domain Fault Diagnosis through Optimal Transport

Awards

  • BRAFITEC/France program
    Coordenação de Aperfeiçoamento de Pessoal de Nível Superior (CAPES)
    Scholarship for international exchange program between Brazil and France.

Publications

  • 2025.01.01
    Optimal Transport for Domain Adaptation through Gaussian Mixture Models
    Transactions on Machine Learning Research
    Machine learning systems operate under the assumption that training and test data are sampled from a fixed probability distribution. However, this assumptions is rarely verified in practice, as the conditions upon which data was acquired are likely to change. In this context, the adaptation of the unsupervised domain requires minimal access to the data of the new conditions for learning models robust to changes in the data distribution. Optimal transport is a theoretically grounded tool for analyzing changes in distribution, especially as it allows the mapping between domains. However, these methods are usually computationally expensive as their complexity scales cubically with the number of samples. In this work, we explore optimal transport between Gaussian Mixture Models (GMMs), which is conveniently written in terms of the components of source and target GMMs. We experiment with 9 benchmarks, with a total of 85 adaptation tasks, showing that our methods are more efficient than previous shallow domain adaptation methods, and they scale well with number of samples n and dimensions d.
  • 2024.01.01
    Lighter, better, faster multi-source domain adaptation with gaussian mixture models and optimal transport
    Joint European Conference on Machine Learning and Knowledge Discovery in Databases
    In this paper, we tackle Multi-Source Domain Adaptation (MSDA), a task in transfer learning where one adapts multiple heterogeneous, labeled source probability measures towards a different, unlabeled target measure. We propose a novel framework for MSDA, based on Optimal Transport (OT) and Gaussian Mixture Models (GMMs). Our framework has two key advantages. First, OT between GMMs can be solved efficiently via linear programming. Second, it provides a convenient model for supervised learning, especially classification, as components in the GMM can be associated with existing classes. Based on the GMM-OT problem, we propose a novel technique for calculating barycenters of GMMs. Based on this novel algorithm, we propose two new strategies for MSDA: GMM-Wasserstein Barycenter Transport (WBT) and GMM-Dataset Dictionary Learning (DaDiL). We empirically evaluate our proposed methods on four benchmarks in image classification and fault diagnosis, showing that we improve over the prior art while being faster and involving fewer parameters. Our code is publicly available at https://github.com/eddardd/gmm_msda
  • 2024.01.01
    Recent advances in optimal transport for machine learning
    IEEE Transactions on Pattern Analysis and Machine Intelligence
    Recently, Optimal Transport has been proposed as a probabilistic framework in Machine Learning for comparing and manipulating probability distributions. This is rooted in its rich history and theory, and has offered new solutions to different problems in machine learning, such as generative modeling and transfer learning. In this survey we explore contributions of Optimal Transport for Machine Learning over the period 2012 -- 2023, focusing on four sub-fields of Machine Learning: supervised, unsupervised, transfer and reinforcement learning. We further highlight the recent development in computational Optimal Transport and its extensions, such as partial, unbalanced, Gromov and Neural Optimal Transport, and its interplay with Machine Learning practice.
  • 2023.01.01
    Multi-Source Domain Adaptation Through Dataset Dictionary Learning in Wasserstein Space
    26th European Conference on Artificial Intelligence
    This paper seeks to solve Multi-Source Domain Adaptation (MSDA), which aims to mitigate data distribution shifts when transferring knowledge from multiple labeled source domains to an unlabeled target domain. We propose a novel MSDA framework based on dictionary learning and optimal transport. We interpret each domain in MSDA as an empirical distribution. As such, we express each domain as a Wasserstein barycenter of dictionary atoms, which are empirical distributions. We propose a novel algorithm, DaDiL, for learning via mini-batches: (i) atom distributions; (ii) a matrix of barycentric coordinates. Based on our dictionary, we propose two novel methods for MSDA: DaDil-R, based on the reconstruction of labeled samples in the target domain, and DaDiL-E, based on the ensembling of classifiers learned on atom distributions. We evaluate our methods in 3 benchmarks: Caltech-Office, Office 31, and CRWU, where we improved previous state-of-the-art by 3.15%, 2.29%, and 7.71% in classification performance. Finally, we show that interpolations in the Wasserstein hull of learned atoms provide data that can generalize to the target domain.
  • 2021.01.01
    Wasserstein Barycenter for Multi-Source Domain Adaptation
    Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition
    Multi-source domain adaptation is a key technique that allows a model to be trained on data coming from various probability distribution. To overcome the challenges posed by this learning scenario, we propose a method for constructing an intermediate domain between sources and target domain, the Wasserstein Barycenter Transport (WBT). This method relies on the barycenter on Wasserstein spaces for aggregating the source probability distributions. Once the sources have been aggregated, they are transported to the target domain using standard Optimal Transport for Domain Adaptation framework. Additionally, we revisit previous single-source domain adaptation tasks in the context of multi-source scenario. In particular, we apply our algorithm to object and face recognition datasets. Moreover, to diversify the range of applications, we also examine the tasks of music genre recognition and music-speech discrimination. The experiments show that our method has similar performance with the existing state-of-the-art.

Skills

Machine Learning
Optimal Transport
Domain Adaptation
Transfer Learning
Foundation Models
Deep Learning
Programming
Python
TensorFlow/Keras
PyTorch
Scikit-Learn
PythonOT
Matlab

Languages

Portuguese
Native speaker
French
B2
English
B2
Spanish
A2

Interests

Artificial Intelligence
Optimal Transport
Domain Adaptation
Transfer Learning
Foundation Models
Machine Learning

Projects

  • 2019.05 - 2020.01
    OpenDenoising
    An extensible benchmark for building comparative studies of image denoisers with focus on deep neural network methods.
    • Open Source
    • Image Restoration
    • Deep Learning
    • Benchmarking