Search Results for author: Narsimha Chilkuri

Found 3 papers, 1 papers with code

Debugging using Orthogonal Gradient Descent

no code implementations17 Jun 2022 Narsimha Chilkuri, Chris Eliasmith

In this report we consider the following problem: Given a trained model that is partially faulty, can we correct its behaviour without having to train the model from scratch?

Continual Learning

Language Modeling using LMUs: 10x Better Data Efficiency or Improved Scaling Compared to Transformers

no code implementations5 Oct 2021 Narsimha Chilkuri, Eric Hunsberger, Aaron Voelker, Gurshaant Malik, Chris Eliasmith

Over three orders of magnitude, we show that our new architecture attains the same accuracy as transformers with 10x fewer tokens.

Language Modelling

Parallelizing Legendre Memory Unit Training

2 code implementations22 Feb 2021 Narsimha Chilkuri, Chris Eliasmith

For instance, our LMU sets a new state-of-the-art result on psMNIST, and uses half the parameters while outperforming DistilBERT and LSTM models on IMDB sentiment analysis.

Machine Translation Sentiment Analysis +2

Cannot find the paper you are looking for? You can Submit a new open access paper.