Search Results for author: Leslie N. Smith

Found 15 papers, 8 papers with code

General Cyclical Training of Neural Networks

1 code implementation17 Feb 2022 Leslie N. Smith

This paper describes the principle of "General Cyclical Training" in machine learning, where training starts and ends with "easy training" and the "hard training" happens during the middle epochs.

Data Augmentation Knowledge Distillation

Cyclical Focal Loss

1 code implementation16 Feb 2022 Leslie N. Smith

In this paper, we introduce a novel cyclical focal loss and demonstrate that it is a more universal loss function than cross-entropy softmax loss or focal loss.

FROST: Faster and more Robust One-shot Semi-supervised Training

no code implementations18 Nov 2020 Helena E. Liu, Leslie N. Smith

Specifically, we show that by combining semi-supervised learning with a one-stage, single network version of self-training, our FROST methodology trains faster and is more robust to choices for the labeled samples and changes in hyper-parameters.

Building One-Shot Semi-supervised (BOSS) Learning up to Fully Supervised Performance

1 code implementation16 Jun 2020 Leslie N. Smith, Adam Conovaloff

Reaching the performance of fully supervised learning with unlabeled data and only labeling one sample per class might be ideal for deep learning applications.

Semi-Supervised Image Classification

Empirical Perspectives on One-Shot Semi-supervised Learning

no code implementations8 Apr 2020 Leslie N. Smith, Adam Conovaloff

One of the greatest obstacles in the adoption of deep neural networks for new applications is that training the network typically requires a large number of manually labeled training samples.

Image Classification

A Useful Taxonomy for Adversarial Robustness of Neural Networks

no code implementations23 Oct 2019 Leslie N. Smith

In addition, there are several papers in the literature of adversarial defenses that claim there is a cost for adversarial robustness, or a trade-off between robustness and accuracy but, under this proposed taxonomy, we hypothesis that this is not universal.

Adversarial Robustness

A disciplined approach to neural network hyper-parameters: Part 1 -- learning rate, batch size, momentum, and weight decay

28 code implementations26 Mar 2018 Leslie N. Smith

Although deep learning has produced dazzling successes for applications of image, speech, and video processing in the past few years, most trainings are with suboptimal hyper-parameters, requiring unnecessarily long training times.

Super-Convergence: Very Fast Training of Residual Networks Using Large Learning Rates

no code implementations ICLR 2018 Leslie N. Smith, Nicholay Topin

In this paper, we show a phenomenon, which we named ``super-convergence'', where residual networks can be trained using an order of magnitude fewer iterations than is used with standard training methods.

Super-Convergence: Very Fast Training of Neural Networks Using Large Learning Rates

10 code implementations23 Aug 2017 Leslie N. Smith, Nicholay Topin

One of the key elements of super-convergence is training with one learning rate cycle and a large maximum learning rate.

Best Practices for Applying Deep Learning to Novel Applications

no code implementations5 Apr 2017 Leslie N. Smith

This report is targeted to groups who are subject matter experts in their application but deep learning novices.

Exploring loss function topology with cyclical learning rates

2 code implementations14 Feb 2017 Leslie N. Smith, Nicholay Topin

We present observations and discussion of previously unreported phenomena discovered while training residual networks.

Deep Convolutional Neural Network Design Patterns

1 code implementation2 Nov 2016 Leslie N. Smith, Nicholay Topin

Recent research in the deep learning field has produced a plethora of new architectures.

Gradual DropIn of Layers to Train Very Deep Neural Networks

no code implementations CVPR 2016 Leslie N. Smith, Emily M. Hand, Timothy Doster

In particular, an untrainable deep network starts as a trainable shallow network and newly added layers are slowly, organically added during training, thereby increasing the network's depth.

Cyclical Learning Rates for Training Neural Networks

50 code implementations3 Jun 2015 Leslie N. Smith

This paper describes a new method for setting the learning rate, named cyclical learning rates, which practically eliminates the need to experimentally find the best values and schedule for the global learning rates.

How to find real-world applications for compressive sensing

no code implementations6 May 2013 Leslie N. Smith

The potential of compressive sensing (CS) has spurred great interest in the research community and is a fast growing area of research.

Compressive Sensing

Cannot find the paper you are looking for? You can Submit a new open access paper.