Structured output prediction problems are ubiquitous in machine learning. The prominent approach leverages neural networks as powerful feature extractors, otherwise assuming the independence of the outputs. These outputs, however, jointly encode an object, e.g. a path in a graph, and are therefore related through the structure underlying the output space. We discuss the semantic loss, which injects knowledge about such structure, defined symbolically, into training by minimizing the network's violation of such dependencies, steering the network towards predicting distributions satisfying the underlying structure. At the same time, it is agnostic to the arrangement of the symbols, and depends only on the semantics expressed thereby, while also enabling efficient end-to-end training and inference. We also discuss key improvements and applications of the semantic loss. One limitations of the semantic loss is that it does not exploit the association of every data point with certain features certifying its membership in a target class. We should therefore prefer minimum-entropy distributions over valid structures, which we obtain by additionally minimizing the neuro-symbolic entropy. We empirically demonstrate the benefits of this more refined formulation. Moreover, the semantic loss is designed to be modular and can be combined with both discriminative and generative neural models. We illustrate this point by integrating the semantic loss into generative adversarial networks, yielding constrained adversarial networks, a novel class of deep generative models able to efficiently synthesize complex objects obeying the structure of the underlying domain.

Chapter 22. Semantic Loss Functions for Neuro-Symbolic Structured Prediction / Ahmed, Kareem; Teso, Stefano; Morettin, Paolo; Di Liello, Luca; Ardino, Pierfrancesco; Gobbi, Jacopo; Liang, Yitao; Wang, Eric; Chang, Kai-Wei; Passerini, Andrea; Van den Broeck, Guy. - 369:(2023), pp. 485-505. [10.3233/FAIA230154]

Chapter 22. Semantic Loss Functions for Neuro-Symbolic Structured Prediction

Teso, Stefano;Morettin, Paolo;Di Liello, Luca;Ardino, Pierfrancesco;Passerini, Andrea;
2023-01-01

Abstract

Structured output prediction problems are ubiquitous in machine learning. The prominent approach leverages neural networks as powerful feature extractors, otherwise assuming the independence of the outputs. These outputs, however, jointly encode an object, e.g. a path in a graph, and are therefore related through the structure underlying the output space. We discuss the semantic loss, which injects knowledge about such structure, defined symbolically, into training by minimizing the network's violation of such dependencies, steering the network towards predicting distributions satisfying the underlying structure. At the same time, it is agnostic to the arrangement of the symbols, and depends only on the semantics expressed thereby, while also enabling efficient end-to-end training and inference. We also discuss key improvements and applications of the semantic loss. One limitations of the semantic loss is that it does not exploit the association of every data point with certain features certifying its membership in a target class. We should therefore prefer minimum-entropy distributions over valid structures, which we obtain by additionally minimizing the neuro-symbolic entropy. We empirically demonstrate the benefits of this more refined formulation. Moreover, the semantic loss is designed to be modular and can be combined with both discriminative and generative neural models. We illustrate this point by integrating the semantic loss into generative adversarial networks, yielding constrained adversarial networks, a novel class of deep generative models able to efficiently synthesize complex objects obeying the structure of the underlying domain.
2023
Compendium of Neurosymbolic Artificial Intelligence
NIEUWE HEMWEG 6B, 1013 BG AMSTERDAM, NETHERLANDS
IOS PRESS
9781643684062
9781643684079
Ahmed, Kareem; Teso, Stefano; Morettin, Paolo; Di Liello, Luca; Ardino, Pierfrancesco; Gobbi, Jacopo; Liang, Yitao; Wang, Eric; Chang, Kai-Wei; Passerini, Andrea; Van den Broeck, Guy
Chapter 22. Semantic Loss Functions for Neuro-Symbolic Structured Prediction / Ahmed, Kareem; Teso, Stefano; Morettin, Paolo; Di Liello, Luca; Ardino, Pierfrancesco; Gobbi, Jacopo; Liang, Yitao; Wang, Eric; Chang, Kai-Wei; Passerini, Andrea; Van den Broeck, Guy. - 369:(2023), pp. 485-505. [10.3233/FAIA230154]
File in questo prodotto:
Non ci sono file associati a questo prodotto.

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11572/401437
 Attenzione

Attenzione! I dati visualizzati non sono stati sottoposti a validazione da parte dell'ateneo

Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus ND
  • ???jsp.display-item.citation.isi??? 0
social impact