Vaidotas (Vaidas) Šimkus

I'm a ML scientist who values simple yet effective solutions to challenging problems that hinder the application of machine learning in critical real-world domains.

I did my PhD in Machine Learning at the University of Edinburgh, where I was advised by Michael Gutmann. My research primarily focused on unsupervised machine learning in the presence of missing data—a challenging problem that affects many domains and often hinders the use of modern machine learning methods. This work has broader implications to deep statistical model estimation, probabilistic inference, and tabular machine learning.

I also hold a MSc in Artificial Intelligence from the University of Edinburgh and a BEng in Software Engineering from the University of Southampton.

Email  /  GitHub  /  Twitter  /  Mastodon  /  LinkedIn

profile photo

Publications

project image

Improving Variational Autoencoder Estimation from Incomplete Data with Mixture Variational Families


Vaidotas Šimkus, Michael Gutmann
Transactions on Machine Learning Research (TMLR), 2024
url / bib / arxiv / code / poster / dmlr-at-iclr2024

We show that missing data increases the complexity of the posterior distribution of the latent variables in VAEs. To mitigate the increased posterior complexity we introduce two strategies based on (i) finite and (ii) imputation-based variational-mixtures.

project image

Conditional Sampling of Variational Autoencoders via Iterated Approximate Ancestral Sampling


Vaidotas Šimkus, Michael Gutmann
Transactions on Machine Learning Research (TMLR), 2023
url / bib / arxiv / code

We link a structured latent space in VAEs, a commonly desired property, to poor conditional sampling performance of Metropolis-within-Gibbs (MWG). To mitigate the issues of MWG we introduce two original methods for conditional sampling of VAEs: AC-MWG and LAIR.

project image

Variational Gibbs Inference for Statistical Model Estimation from Incomplete Data


Vaidotas Šimkus, Ben Rhodes, Michael Gutmann
Journal of Machine Learning Research (JMLR), 2023
url / bib / arxiv / code / poster / slides / jmlr-to-neurips2023 / demo

We propose a new method for statistical model estimation from incomplete data, called variational Gibbs inference (VGI). Whilst being general-pupose, the proposed method outperforms existing VAE and normalising flow specific methods.

project image

Learning Job Titles Similarity from Noisy Skill Labels


Rabih Zbib, Lucas Alvarez Lacasa, Federico Retyk, Rus Poves, Juan Aizpuru, Hermenegildo Fabregat, Vaidotas Šimkus, Emilia García-Casademont
FEAST workshop at ECML-PKDD, 2022
url / bib / arxiv / dataset

We propose an unsupervised representation learning method for a job title similarity model using noisy skill labels. We show that it is highly effective for tasks such as text ranking and job normalization.