Benjamin Rodatz, Ian Fan, Tuomas Laakkonen, Neil John Ortega, Thomas Hoffman, Vincent Wang-Mascianica
{"title":"A Pattern Language for Machine Learning Tasks","authors":"Benjamin Rodatz, Ian Fan, Tuomas Laakkonen, Neil John Ortega, Thomas Hoffman, Vincent Wang-Mascianica","doi":"arxiv-2407.02424","DOIUrl":null,"url":null,"abstract":"Idealised as universal approximators, learners such as neural networks can be\nviewed as \"variable functions\" that may become one of a range of concrete\nfunctions after training. In the same way that equations constrain the possible\nvalues of variables in algebra, we may view objective functions as constraints\non the behaviour of learners. We extract the equivalences perfectly optimised\nobjective functions impose, calling them \"tasks\". For these tasks, we develop a\nformal graphical language that allows us to: (1) separate the core tasks of a\nbehaviour from its implementation details; (2) reason about and design\nbehaviours model-agnostically; and (3) simply describe and unify approaches in\nmachine learning across domains. As proof-of-concept, we design a novel task that enables converting\nclassifiers into generative models we call \"manipulators\", which we implement\nby directly translating task specifications into code. The resulting models\nexhibit capabilities such as style transfer and interpretable latent-space\nediting, without the need for custom architectures, adversarial training or\nrandom sampling. We formally relate the behaviour of manipulators to GANs, and\nempirically demonstrate their competitive performance with VAEs. We report on\nexperiments across vision and language domains aiming to characterise\nmanipulators as approximate Bayesian inversions of discriminative classifiers.","PeriodicalId":501135,"journal":{"name":"arXiv - MATH - Category Theory","volume":"72 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-07-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - MATH - Category Theory","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2407.02424","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Idealised as universal approximators, learners such as neural networks can be
viewed as "variable functions" that may become one of a range of concrete
functions after training. In the same way that equations constrain the possible
values of variables in algebra, we may view objective functions as constraints
on the behaviour of learners. We extract the equivalences perfectly optimised
objective functions impose, calling them "tasks". For these tasks, we develop a
formal graphical language that allows us to: (1) separate the core tasks of a
behaviour from its implementation details; (2) reason about and design
behaviours model-agnostically; and (3) simply describe and unify approaches in
machine learning across domains. As proof-of-concept, we design a novel task that enables converting
classifiers into generative models we call "manipulators", which we implement
by directly translating task specifications into code. The resulting models
exhibit capabilities such as style transfer and interpretable latent-space
editing, without the need for custom architectures, adversarial training or
random sampling. We formally relate the behaviour of manipulators to GANs, and
empirically demonstrate their competitive performance with VAEs. We report on
experiments across vision and language domains aiming to characterise
manipulators as approximate Bayesian inversions of discriminative classifiers.