Achieving More with Less: Learning Generalizable Neural Networks With Less Labeled Data and Computational Overheads

TR Number

Date

2023-03-15

Authors

Journal Title

Journal ISSN

Volume Title

Publisher

Virginia Tech

Abstract

Recent advancements in deep learning have demonstrated its incredible ability to learn generalizable patterns and relationships automatically from data in a number of mainstream applications. However, the generalization power of deep learning methods largely comes at the costs of working with very large datasets and using highly compute-intensive models. Many applications cannot afford these costs needed to ensure generalizability of deep learning models. For instance, obtaining labeled data can be costly in scientific applications, and using large models may not be feasible in resource-constrained environments involving portable devices. This dissertation aims to improve efficiency in machine learning by exploring different ways to learn generalizable neural networks that require less labeled data and computational resources. We demonstrate that using physics supervision in scientific problems can reduce the need for labeled data, thereby improving data efficiency without compromising model generalizability. Additionally, we investigate the potential of transfer learning powered by transformers in scientific applications as a promising direction for further improving data efficiency. On the computational efficiency side, we present two efforts for increasing parameter efficiency of neural networks through novel architectures and structured network pruning.

Description

Keywords

Machine Learning Efficiency, Physics-guided Machine Learning, Efficient Neural Architecture, Neural Network Pruning, Transfer Learning

Citation