Zhixing Hou, Maoxu Gao, Hang Yu, Mengyu Yang, Chio-In Ieong
{"title":"SDP:利用可学习的通道膜阈值实现机器人操纵的尖峰扩散策略","authors":"Zhixing Hou, Maoxu Gao, Hang Yu, Mengyu Yang, Chio-In Ieong","doi":"arxiv-2409.11195","DOIUrl":null,"url":null,"abstract":"This paper introduces a Spiking Diffusion Policy (SDP) learning method for\nrobotic manipulation by integrating Spiking Neurons and Learnable Channel-wise\nMembrane Thresholds (LCMT) into the diffusion policy model, thereby enhancing\ncomputational efficiency and achieving high performance in evaluated tasks.\nSpecifically, the proposed SDP model employs the U-Net architecture as the\nbackbone for diffusion learning within the Spiking Neural Network (SNN). It\nstrategically places residual connections between the spike convolution\noperations and the Leaky Integrate-and-Fire (LIF) nodes, thereby preventing\ndisruptions to the spiking states. Additionally, we introduce a temporal\nencoding block and a temporal decoding block to transform static and dynamic\ndata with timestep $T_S$ into each other, enabling the transmission of data\nwithin the SNN in spike format. Furthermore, we propose LCMT to enable the\nadaptive acquisition of membrane potential thresholds, thereby matching the\nconditions of varying membrane potentials and firing rates across channels and\navoiding the cumbersome process of manually setting and tuning hyperparameters.\nEvaluating the SDP model on seven distinct tasks with SNN timestep $T_S=4$, we\nachieve results comparable to those of the ANN counterparts, along with faster\nconvergence speeds than the baseline SNN method. This improvement is\naccompanied by a reduction of 94.3\\% in dynamic energy consumption estimated on\n45nm hardware.","PeriodicalId":501031,"journal":{"name":"arXiv - CS - Robotics","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-09-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"SDP: Spiking Diffusion Policy for Robotic Manipulation with Learnable Channel-Wise Membrane Thresholds\",\"authors\":\"Zhixing Hou, Maoxu Gao, Hang Yu, Mengyu Yang, Chio-In Ieong\",\"doi\":\"arxiv-2409.11195\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"This paper introduces a Spiking Diffusion Policy (SDP) learning method for\\nrobotic manipulation by integrating Spiking Neurons and Learnable Channel-wise\\nMembrane Thresholds (LCMT) into the diffusion policy model, thereby enhancing\\ncomputational efficiency and achieving high performance in evaluated tasks.\\nSpecifically, the proposed SDP model employs the U-Net architecture as the\\nbackbone for diffusion learning within the Spiking Neural Network (SNN). It\\nstrategically places residual connections between the spike convolution\\noperations and the Leaky Integrate-and-Fire (LIF) nodes, thereby preventing\\ndisruptions to the spiking states. Additionally, we introduce a temporal\\nencoding block and a temporal decoding block to transform static and dynamic\\ndata with timestep $T_S$ into each other, enabling the transmission of data\\nwithin the SNN in spike format. Furthermore, we propose LCMT to enable the\\nadaptive acquisition of membrane potential thresholds, thereby matching the\\nconditions of varying membrane potentials and firing rates across channels and\\navoiding the cumbersome process of manually setting and tuning hyperparameters.\\nEvaluating the SDP model on seven distinct tasks with SNN timestep $T_S=4$, we\\nachieve results comparable to those of the ANN counterparts, along with faster\\nconvergence speeds than the baseline SNN method. This improvement is\\naccompanied by a reduction of 94.3\\\\% in dynamic energy consumption estimated on\\n45nm hardware.\",\"PeriodicalId\":501031,\"journal\":{\"name\":\"arXiv - CS - Robotics\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-09-17\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - CS - Robotics\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2409.11195\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Robotics","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.11195","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
SDP: Spiking Diffusion Policy for Robotic Manipulation with Learnable Channel-Wise Membrane Thresholds
This paper introduces a Spiking Diffusion Policy (SDP) learning method for
robotic manipulation by integrating Spiking Neurons and Learnable Channel-wise
Membrane Thresholds (LCMT) into the diffusion policy model, thereby enhancing
computational efficiency and achieving high performance in evaluated tasks.
Specifically, the proposed SDP model employs the U-Net architecture as the
backbone for diffusion learning within the Spiking Neural Network (SNN). It
strategically places residual connections between the spike convolution
operations and the Leaky Integrate-and-Fire (LIF) nodes, thereby preventing
disruptions to the spiking states. Additionally, we introduce a temporal
encoding block and a temporal decoding block to transform static and dynamic
data with timestep $T_S$ into each other, enabling the transmission of data
within the SNN in spike format. Furthermore, we propose LCMT to enable the
adaptive acquisition of membrane potential thresholds, thereby matching the
conditions of varying membrane potentials and firing rates across channels and
avoiding the cumbersome process of manually setting and tuning hyperparameters.
Evaluating the SDP model on seven distinct tasks with SNN timestep $T_S=4$, we
achieve results comparable to those of the ANN counterparts, along with faster
convergence speeds than the baseline SNN method. This improvement is
accompanied by a reduction of 94.3\% in dynamic energy consumption estimated on
45nm hardware.