@katharinafuchs

Training two-layer relu networks with gradient descent is inconsistent

, and . arXiv preprint arXiv:2002.04861, (2020)

Links and resources

Tags

community

  • @unibiblio
  • @simtech
  • @davidholzmller
  • @simtechpuma
  • @katharinafuchs
  • @mathematik
@katharinafuchs's tags highlighted