Revisiting dropout: Escaping pressure for training neural networks with multiple costs

Sangmin Woo, Kangil Kim, Junhyug Noh, Jong Hun Shin, Seung Hoon Na

Research output: Contribution to journalArticlepeer-review

Abstract

A common approach to jointly learn multiple tasks with a shared structure is to optimize the model with a combined landscape of multiple sub-costs. However, gradients derived from each sub-cost often conflicts in cost plateaus, resulting in a subpar optimum. In this work, we shed light on such gradient conflict challenges and suggest a solution named Cost-Out, which randomly drops the sub-costs for each iteration. We provide the theoretical and empirical evidence of the existence of escaping pressure induced by the Cost-Out mechanism. While simple, the empirical results indicate that the proposed method can enhance the performance of multi-task learning problems, including two-digit image classification sampled from MNIST dataset and machine translation tasks for English from and to French, Spanish, and German WMT14 datasets.

Original languageEnglish
Article number989
JournalElectronics (Switzerland)
Volume10
Issue number9
DOIs
StatePublished - 1 May 2021

Bibliographical note

Publisher Copyright:
© Licensee MDPI, Basel 2021 by the authors. Switzerland.

Keywords

  • Cost-Out
  • Dropout
  • Escaping pressure
  • Gradient conflict
  • Multitask learning

Fingerprint

Dive into the research topics of 'Revisiting dropout: Escaping pressure for training neural networks with multiple costs'. Together they form a unique fingerprint.

Cite this