Ari Blau, Evan S Schaffer, Neeli Mishra, Nathaniel J Miska, The International Brain Laboratory, Liam Paninski, Matthew R Whiteway
{"title":"跨越监督、无监督和半监督学习范式的动物动作分割算法研究","authors":"Ari Blau, Evan S Schaffer, Neeli Mishra, Nathaniel J Miska, The International Brain Laboratory, Liam Paninski, Matthew R Whiteway","doi":"arxiv-2407.16727","DOIUrl":null,"url":null,"abstract":"Action segmentation of behavioral videos is the process of labeling each\nframe as belonging to one or more discrete classes, and is a crucial component\nof many studies that investigate animal behavior. A wide range of algorithms\nexist to automatically parse discrete animal behavior, encompassing supervised,\nunsupervised, and semi-supervised learning paradigms. These algorithms -- which\ninclude tree-based models, deep neural networks, and graphical models -- differ\nwidely in their structure and assumptions on the data. Using four datasets\nspanning multiple species -- fly, mouse, and human -- we systematically study\nhow the outputs of these various algorithms align with manually annotated\nbehaviors of interest. Along the way, we introduce a semi-supervised action\nsegmentation model that bridges the gap between supervised deep neural networks\nand unsupervised graphical models. We find that fully supervised temporal\nconvolutional networks with the addition of temporal information in the\nobservations perform the best on our supervised metrics across all datasets.","PeriodicalId":501266,"journal":{"name":"arXiv - QuanBio - Quantitative Methods","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-07-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"A study of animal action segmentation algorithms across supervised, unsupervised, and semi-supervised learning paradigms\",\"authors\":\"Ari Blau, Evan S Schaffer, Neeli Mishra, Nathaniel J Miska, The International Brain Laboratory, Liam Paninski, Matthew R Whiteway\",\"doi\":\"arxiv-2407.16727\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Action segmentation of behavioral videos is the process of labeling each\\nframe as belonging to one or more discrete classes, and is a crucial component\\nof many studies that investigate animal behavior. A wide range of algorithms\\nexist to automatically parse discrete animal behavior, encompassing supervised,\\nunsupervised, and semi-supervised learning paradigms. These algorithms -- which\\ninclude tree-based models, deep neural networks, and graphical models -- differ\\nwidely in their structure and assumptions on the data. Using four datasets\\nspanning multiple species -- fly, mouse, and human -- we systematically study\\nhow the outputs of these various algorithms align with manually annotated\\nbehaviors of interest. Along the way, we introduce a semi-supervised action\\nsegmentation model that bridges the gap between supervised deep neural networks\\nand unsupervised graphical models. We find that fully supervised temporal\\nconvolutional networks with the addition of temporal information in the\\nobservations perform the best on our supervised metrics across all datasets.\",\"PeriodicalId\":501266,\"journal\":{\"name\":\"arXiv - QuanBio - Quantitative Methods\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-07-23\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - QuanBio - Quantitative Methods\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2407.16727\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - QuanBio - Quantitative Methods","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2407.16727","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
A study of animal action segmentation algorithms across supervised, unsupervised, and semi-supervised learning paradigms
Action segmentation of behavioral videos is the process of labeling each
frame as belonging to one or more discrete classes, and is a crucial component
of many studies that investigate animal behavior. A wide range of algorithms
exist to automatically parse discrete animal behavior, encompassing supervised,
unsupervised, and semi-supervised learning paradigms. These algorithms -- which
include tree-based models, deep neural networks, and graphical models -- differ
widely in their structure and assumptions on the data. Using four datasets
spanning multiple species -- fly, mouse, and human -- we systematically study
how the outputs of these various algorithms align with manually annotated
behaviors of interest. Along the way, we introduce a semi-supervised action
segmentation model that bridges the gap between supervised deep neural networks
and unsupervised graphical models. We find that fully supervised temporal
convolutional networks with the addition of temporal information in the
observations perform the best on our supervised metrics across all datasets.