Machine Learning Retrospective, 2021

With 2021 drawing to a close, we would like to take a moment to recognize the wealth of machine learning research produced by the UMass Manning College of Information and Computer Sciences (CICS). This retrospective provides a brief summary of many (not all) of the machine learning papers published by students and/or faculty in CICS. You can browse papers by their name in the index below, or can just scroll through to get a sense for all of the work that we are doing!


  • Paper: RealMVP: A Change of Variables Method For Rectangular Matrix-Vector Products

    Paper: RealMVP: A Change of Variables Method For Rectangular Matrix-Vector Products

    Rectangular matrix-vector products are used extensively throughout machine learning and are fundamental to neural networks such as multi-layer perceptrons, but are notably absent as normalizing flow layers. This paper identifies this methodological gap and plugs it with a tall and wide MVP change of variables formula. Our theory builds up to a practical algorithm that…

  • Paper: High Confidence Generalization for Reinforcement Learning

    Paper: High Confidence Generalization for Reinforcement Learning

    We present several classes of reinforcement learning algorithms that safely generalize to Markov decision processes (MDPs) not seen during training. Specifically, we study the setting in which some set of MDPs is accessible for training. For various definitions of safety, our algorithms give probabilistic guarantees that agents can safely generalize to MDPs that are sampled…

  • Paper: On the Difficulty of Unbiased Alpha Divergence Minimization

    Paper: On the Difficulty of Unbiased Alpha Divergence Minimization

    Short description: Variational inference approximates a target distribution with a simpler one. While traditional inference minimizes the “inclusive” KL-divergence, several algorithms have recently been proposed to minimize other divergences. Experimentally, however, these algorithms often seem to fail to converge. In this paper we analyze the variance of the underlying estimators for these papers. Our results…

  • Paper: Posterior Value Functions: Hindsight Baselines for Policy Gradient Methods

    Paper: Posterior Value Functions: Hindsight Baselines for Policy Gradient Methods

    Hindsight allows reinforcement learning agents to leverage new observations to make inferences about earlier states and transitions. In this paper, we exploit the idea of hindsight and introduce posterior value functions. Posterior value functions are computed by inferring the posterior distribution over hidden components of the state in previous timesteps and can be used to…

  • Paper: Towards Practical Mean Bounds for Small Samples

    Paper: Towards Practical Mean Bounds for Small Samples

    Historically, to bound the mean for small sample sizes, practitioners have had to choose between using methods with unrealistic assumptions about the unknown distribution (e.g., Gaussianity) and methods like Hoeffding’s inequality that use weaker assumptions but produce much looser (wider) intervals. In 1969, Anderson proposed a mean confidence interval strictly better than or equal to…

  • Paper: How and Why to Use Experimental Data to Evaluate Methods for Observational Causal Inference

    Paper: How and Why to Use Experimental Data to Evaluate Methods for Observational Causal Inference

    Rectangular matrix-vector products are used extensively throughout machine learning and are fundamental to neural networks such as multi-layer perceptrons, but are notably absent as normalizing flow layers. This Methods that infer causal dependence from observational data are central to many areas of science, including medicine, economics, and the social sciences. We describe and analyze observational…

  • Paper: DeepWalking Backwards: From Node Embeddings Back to Graphs

    Paper: DeepWalking Backwards: From Node Embeddings Back to Graphs

    We investigate whether node embeddings, which are vector representations of graph nodes, can be inverted to approximately recover the graph used to generate them. We present algorithms that invert embeddings from the popular DeepWalk method. In experiments on real-world networks, we find that significant information about the original graph, such as specific edges, is often…

  • Paper: Faster Kernel Matrix Algebra via Density Estimation

    Consider an n x n Gaussian kernel matrix corresponding to n input points in d dimensions. We show that one can compute a relative error approximation to the sum of entries in this matrix in just O(dn^{2/3}) time. This is significantly sublinear in the number of entries in the matrix – which is n^2. Our…

  • Paper: Structural Credit Assignment in Neural Networks using Reinforcement Learning

    Paper:  Structural Credit Assignment in Neural Networks using Reinforcement Learning

    Consider an n x n Gaussian kernel matrix corresponding to n input points in d dimensions. We show that one In this work, we revisit REINFORCE and investigate if we can leverage other reinforcement learning approaches to improve learning. We formalize training a neural network as a finite-horizon reinforcement learning problem and discuss how this…

  • Paper: Universal Off-Policy Evaluation

    Paper: Universal Off-Policy Evaluation

    When faced with sequential decision-making problems, it is often useful to be able to predict what would happen if decisions were made using a new policy. Those predictions must often be based on data collected under some previously used decision-making rule. Many previous methods enable such off-policy (or counterfactual) estimation of the expected value of…

  • Paper: MCMC Variational Inference via Uncorrected Hamiltonian Annealing

    Paper: MCMC Variational Inference via Uncorrected Hamiltonian Annealing

    When faced with sequential decision-making problems, it is often useful to be able to predict what would Annealed Importance Sampling (AIS) with Hamiltonian MCMC can be used to get tight lower bounds on a distribution’s (log) normalization constant. Its main drawback is that it uses non-differentiable transition kernels, which makes tuning its many parameters hard.…