Search Results for author: Amirmohammad Rooshenas

Found 7 papers, 2 papers with code

Score-Based Multimodal Autoencoders

no code implementations25 May 2023 Daniel Wesego, Amirmohammad Rooshenas

Multimodal Variational Autoencoders (VAEs) represent a promising group of generative models that facilitate the construction of a tractable posterior within the latent space, given multiple modalities.

Energy-Based Reranking: Improving Neural Machine Translation Using Energy-Based Models

1 code implementation ACL 2021 Sumanta Bhattacharyya, Amirmohammad Rooshenas, Subhajit Naskar, Simeng Sun, Mohit Iyyer, Andrew McCallum

To benefit from this observation, we train an energy-based model to mimic the behavior of the task measure (i. e., the energy-based model assigns lower energy to samples with higher BLEU score), which is resulted in a re-ranking algorithm based on the samples drawn from NMT: energy-based re-ranking (EBR).

Computational Efficiency Machine Translation +4

Differential Equation Units: Learning Functional Forms of Activation Functions from Data

1 code implementation6 Sep 2019 MohamadAli Torkamani, Shiv Shankar, Amirmohammad Rooshenas, Phillip Wallis

Most deep neural networks use simple, fixed activation functions, such as sigmoids or rectified linear units, regardless of domain or network structure.

Learning Compact Neural Networks Using Ordinary Differential Equations as Activation Functions

no code implementations19 May 2019 MohamadAli Torkamani, Phillip Wallis, Shiv Shankar, Amirmohammad Rooshenas

Most deep neural networks use simple, fixed activation functions, such as sigmoids or rectified linear units, regardless of domain or network structure.

Search-Guided, Lightly-supervised Training of Structured Prediction Energy Networks

no code implementations22 Dec 2018 Amirmohammad Rooshenas, Dongxu Zhang, Gopal Sharma, Andrew McCallum

In this paper, we instead use efficient truncated randomized search in this reward function to train structured prediction energy networks (SPENs), which provide efficient test-time inference using gradient-based search on a smooth, learned representation of the score landscape, and have previously yielded state-of-the-art results in structured prediction.

Structured Prediction

The Libra Toolkit for Probabilistic Models

no code implementations1 Apr 2015 Daniel Lowd, Amirmohammad Rooshenas

The Libra Toolkit is a collection of algorithms for learning and inference with discrete probabilistic models, including Bayesian networks, Markov networks, dependency networks, and sum-product networks.

Cannot find the paper you are looking for? You can Submit a new open access paper.