{"title":"Structured Prediction Networks through Latent Cost Learning","authors":"R. Milidiú, R. Rocha","doi":"10.1109/SSCI.2018.8628625","DOIUrl":null,"url":null,"abstract":"Structured prediction provides a flexible modeling framework to deal with several relevant problems. Sequences, Trees, Disjoint Intervals and Matching are some useful examples of the type of structures we would like to predict. An elegant learning scheme for this prediction setting is the Structured Perceptron algorithm, which is sure to converge under some linear separability conditions. The framework integrates a very simple Structured layer on top of a latent costs network. Our key contribution is a novel loss function that incorporates structural information and simplifies learning. The effectiveness of this framework is illustrated with sequence prediction problems. We explore LSTM neural network architectures to model the latent costs layer, since our experiments concern NLP tasks. We perform basic experiments with Chunking in English. The SPN predictor outperforms its CRF equivalent. Our initial findings strongly indicate that SPN is a versatile framework with a powerful learning strategy.","PeriodicalId":235735,"journal":{"name":"2018 IEEE Symposium Series on Computational Intelligence (SSCI)","volume":"38 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2018-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2018 IEEE Symposium Series on Computational Intelligence (SSCI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/SSCI.2018.8628625","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1
Abstract
Structured prediction provides a flexible modeling framework to deal with several relevant problems. Sequences, Trees, Disjoint Intervals and Matching are some useful examples of the type of structures we would like to predict. An elegant learning scheme for this prediction setting is the Structured Perceptron algorithm, which is sure to converge under some linear separability conditions. The framework integrates a very simple Structured layer on top of a latent costs network. Our key contribution is a novel loss function that incorporates structural information and simplifies learning. The effectiveness of this framework is illustrated with sequence prediction problems. We explore LSTM neural network architectures to model the latent costs layer, since our experiments concern NLP tasks. We perform basic experiments with Chunking in English. The SPN predictor outperforms its CRF equivalent. Our initial findings strongly indicate that SPN is a versatile framework with a powerful learning strategy.