Tomas Geffner
(After an intense hike)
About
I am a final year PhD student in the College of Information and Computer Sciences at UMass Amherst. I am fortunate to be advised by Justin Domke.
I finished my PhD in 2023. I am now a research scientist at NVIDIA, working on generative modeling within the Fundamental Generative AI Research Group.
Before joining NVIDIA, I am grateful to have interned at VantAI (New York, summer 2023) with Luca Naef and Michael Bronstein; DeepMind (London, summer 2022) with Andriy Mnih and George Papamakarios; MSR (Cambridge UK, summer 2021) with Cheng Zhang, Emre Kiciman and Miltos Allamanis; and Amazon AWS (New York, summer 2018) with ‪Bing Xiang and Ramesh Nallapati.
I work on probabilistic Machine Learning, focusing on generative modeling, sampling, approximate inference, and causality, with the goal of developing methods able to solve increasingly challenging problems. I am interested in fundamental methodological developments as well as applications in different scientific domains.
Before my PhD I also did some work on planning.
This is a link to my google scholar.
Publications (& Preprints)
- Compositional Score modeling for Simulation-based Inference. By Tomas Geffner, George Papamakarios and Andriy Mnih.
ICML 2023.
We introduce a novel method to do simulation-based inference (aka likelihood-free inference) using conditional score modeling. Our method learns the score of (diffused versions) of the target posterior distribution, uses annealed Langevin dynamics to get samples, and can naturally aggregate multiple observations at inference time by exploiting different factorizations of the target distribution.
- Langevin Diffusion Variational Inference. By Tomas Geffner and Justin Domke.
AISTATS 2023.
We present a new method to combine VI and MCMC based on the discretization of Langevin diffusion processes and their time reversals. We show that our approach generalizes previously existing methods, and outperforms them on a range of inference tasks.
- Variational Inference with Locally Enhanced Bounds for Hierarchical Models. By Tomas Geffner and Justin Domke.
ICML 2022 (spotlight).
We present a new family of variational objectives for hierarchical models, based on the application of importance weighting (and other tightening methods) at the local variables level. Our approach yields tight lower bounds and powerful posterior approximations while being compatible with subsampling, thus addressing one of the main limitations of tightening methods.
- Deep End-to-end Causal Inference. By Tomas Geffner*, Javier Antoran*, Adam Foster*, Wenbo Gong, Chao Ma, Emre Kiciman, Amit Sharma, Angus Lamb, Martin Kukla, Nick Pawlowski, Miltadis Allamanis and Cheng Zhang.
Workshop on Causality for Real World Impact (NeurIPS 2022).
We present a new end-to-end simulation-based method for causal inference. Our approach performs causal discovery by fitting autoregressive flows, and uses the learned model to generate samples from interventional distributions and estimate treatment effects.
- MCMC Variational Inference via Uncorrected Hamiltonian Annealing. By Tomas Geffner and Justin Domke.
NeurIPS 2021.
We present a new way of combining VI and MCMC using Annealed Importance Sampling with uncorrected HMC transitions. The method is simple to implement, and all of its parameters (even the ones from the HMC transition kernel!) can be tuned with unbiased reparameterization gradients.
- On the Difficulty of Unbiased Alpha Divergence Minimization. By Tomas Geffner and Justin Domke.
ICML 2021.
Want to minimize an alpha divergence using unbiased gradients? It will be hard. In high dimensions your gradient estimator will likely suffer from an exponentially small SNR.
- Empirical Evaluation of Biased Methods for Alpha Divergence Minimization. By Tomas Geffner and Justin Domke.
AABI 2020 (contributed talk).
Want to minimize an alpha divergence using biased gradients? It will be hard. In high dimensions you may end up minimizing the typical exclusive KL divergence.
- Approximation Based Variance Reduction for Reparameterization Gradients. By Tomas Geffner and Justin Domke.
NeurIPS 2020.
We propose a new control variate to reduce the variance of gradient estimators used by VI. This leads to faster convergence. Our method works better than previously proposed methods based on Taylor expansions. We propose an explanation for this.
- A Rule for Gradient Estimator Selection, with an Application to Variational Inference. By Tomas Geffner and Justin Domke.
AISTATS 2020.
Some gradient estimates are cheap but have a high variance. Some are expensive but have lower variance. Others are somewhere in between. Which one should you use? We study this using known SGD convergence rates and propose a simple gradient estimator selection algorithm that aims to minimize optimization wall-clock time.
- Using Large Ensembles of Control Variates for Variational Inference. By Tomas Geffner and Justin Domke.
NeurIPS 2018.
Using a control variate may lead to a reduction in the variance of gradient estimators, improving optimization convergence. For VI, many control variates are available. We propose a method to use them jointly and show that this may work better than any of the control variates individually.
- Compact Policies for Fully-Observable Non-Deterministic Planning as SAT. By Tomas Geffner and Hector Geffner.
ICAPS 2018.
We introduce a new SAT-based algorithm for FOND planning. The proposed method works well on several probabilistic interesting domains.
- Width-based Planning for General Video-Game Playing. By Tomas Geffner and Hector Geffner.
AAIDE 2015.
Developed agent for GVG-AI planning competition (real time autonomous game playing) based on a combination of planning and learning algorithms. The method got the 1st place in CEEC leg of competition (2015).
Service
Organizing Committee:
- Publication chair (AISTATS 2023)
Conference reviewing:
- NeurIPS (2019, 2020, 2021, 2022)
- ICML (2020)
- ICLR (2021, 2022, 2025)
- AISTATS (2021)
Journal reviewing:
Teaching experience
I was a TA for several courses:
- Probabilistic graphical models.
- Machine learning.
- Introduction to data structures.
- Introduction to Natural language processing.
- Human Computer Interaction.
Other
I try to play tennis.
Contact
Email: t[lastname]@nvidia.com |
Email: [firstname][lastname]@gmail.com |
Disclaimer
The template for this website was obtained from here.