Training Implicit Generative Models via an Invariant statistical loss (ISL)
-
Updated
Mar 6, 2024 - Julia
Training Implicit Generative Models via an Invariant statistical loss (ISL)
Code for our AISTATS '22 paper: Improving Attribution Methods by Learning Submodular Functions.
DPE code - Code used in "Optimal Algorithms for Multiplayer Multi-Armed Bandits" (AISTATS 2020)
PyTorch implementation for " Differentiable Antithetic Sampling for Variance Reduction in Stochastic Variational Inference" (https://arxiv.org/abs/1810.02555).
AISTATS 2019: Lovász Convolutional Networks
Spectral Tensor Train Parameterization of Deep Learning Layers
Know Your Boundaries: Constraining Gaussian Processes by Variational Harmonic Features
A near-optimal exact sampler for discrete probability distributions
AISTATS 2019: Confidence-based Graph Convolutional Networks for Semi-Supervised Learning
For deep RL and the future of AI.
Add a description, image, and links to the aistats topic page so that developers can more easily learn about it.
To associate your repository with the aistats topic, visit your repo's landing page and select "manage topics."