We show that a neural network whose output is obtained as the difference of the outputs of two feedforward networks with exponential activation function in the hidden layer and logarithmic activation function in the output node, referred to as log-sum-exp (LSE) network, is a smooth universal approximator of continuous functions over convex, compact sets. By using a logarithmic transform, this class of network maps to a family of subtraction-free ratios of generalized posynomials (GPOS), which we also show to be universal approximators of positive functions over log-convex, compact subsets of the positive orthant. The main advantage of difference-LSE networks with respect to classical feedforward neural networks is that, after a standard training phase, they provide surrogate models for a design that possesses a specific difference-of-convex-functions form, which makes them optimizable via relatively efficient numerical methods. In particular, by adapting an existing difference-of-convex algorithm to these models, we obtain an algorithm for performing an effective optimization-based design. We illustrate the proposed approach by applying it to the data-driven design of a diet for a patient with type-2 diabetes and to a nonconvex optimization problem.

Calafiore, G.c., Gaubert, S., Possieri, C. (2020). A Universal Approximation Result for Difference of Log-Sum-Exp Neural Networks. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 31(12), 5603-5612 [10.1109/TNNLS.2020.2975051].

A Universal Approximation Result for Difference of Log-Sum-Exp Neural Networks

Possieri C.
2020-01-01

Abstract

We show that a neural network whose output is obtained as the difference of the outputs of two feedforward networks with exponential activation function in the hidden layer and logarithmic activation function in the output node, referred to as log-sum-exp (LSE) network, is a smooth universal approximator of continuous functions over convex, compact sets. By using a logarithmic transform, this class of network maps to a family of subtraction-free ratios of generalized posynomials (GPOS), which we also show to be universal approximators of positive functions over log-convex, compact subsets of the positive orthant. The main advantage of difference-LSE networks with respect to classical feedforward neural networks is that, after a standard training phase, they provide surrogate models for a design that possesses a specific difference-of-convex-functions form, which makes them optimizable via relatively efficient numerical methods. In particular, by adapting an existing difference-of-convex algorithm to these models, we obtain an algorithm for performing an effective optimization-based design. We illustrate the proposed approach by applying it to the data-driven design of a diet for a patient with type-2 diabetes and to a nonconvex optimization problem.
2020
Pubblicato
Rilevanza internazionale
Articolo
Esperti anonimi
Settore ING-INF/04 - AUTOMATICA
English
Data-driven optimization
difference of convex (DC) programming
feedforward neural networks (FFNs)
log-sum-exp (LSE) networks
subtraction-free expressions
surrogate models
universal approximation
Algorithms
Artificial Intelligence
Diabetes Mellitus, Type 2
Diet
Feedback
Humans
Machine Learning
Meals
Neural Networks, Computer
Calafiore, G.c., Gaubert, S., Possieri, C. (2020). A Universal Approximation Result for Difference of Log-Sum-Exp Neural Networks. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 31(12), 5603-5612 [10.1109/TNNLS.2020.2975051].
Calafiore, Gc; Gaubert, S; Possieri, C
Articolo su rivista
File in questo prodotto:
Non ci sono file associati a questo prodotto.

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/2108/294462
Citazioni
  • ???jsp.display-item.citation.pmc??? 0
  • Scopus 26
  • ???jsp.display-item.citation.isi??? 24
social impact