{"title":"面向多任务实时系统的快速切换双处理体系结构","authors":"Tein-Hsiang Lin, Jui-ping Liao","doi":"10.1109/ICPADS.1994.589904","DOIUrl":null,"url":null,"abstract":"A new fast switching double processing architecture for pipelined cache-based real-time computer systems is proposed to reduce the CPU stalls due to increased cache misses resulting from frequent task switching in multi-tasking real-time applications. In this architecture, two sets of registers are provided so that two tasks can be executed alternatively on a cycle-by-cycle basis. This architecture helps alleviate the problem of unpredictable cache performance due to frequent context switches in multi-tasking systems. The performance of the double processing is evaluated first through trace driven simulation for various cache configurations. An analytical performance model is then derived to further explain the performance advantage.","PeriodicalId":154429,"journal":{"name":"Proceedings of 1994 International Conference on Parallel and Distributed Systems","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"1994-12-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"A fast switching double processing architecture for multi-tasking real-time systems\",\"authors\":\"Tein-Hsiang Lin, Jui-ping Liao\",\"doi\":\"10.1109/ICPADS.1994.589904\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"A new fast switching double processing architecture for pipelined cache-based real-time computer systems is proposed to reduce the CPU stalls due to increased cache misses resulting from frequent task switching in multi-tasking real-time applications. In this architecture, two sets of registers are provided so that two tasks can be executed alternatively on a cycle-by-cycle basis. This architecture helps alleviate the problem of unpredictable cache performance due to frequent context switches in multi-tasking systems. The performance of the double processing is evaluated first through trace driven simulation for various cache configurations. An analytical performance model is then derived to further explain the performance advantage.\",\"PeriodicalId\":154429,\"journal\":{\"name\":\"Proceedings of 1994 International Conference on Parallel and Distributed Systems\",\"volume\":\"1 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"1994-12-19\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Proceedings of 1994 International Conference on Parallel and Distributed Systems\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICPADS.1994.589904\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of 1994 International Conference on Parallel and Distributed Systems","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICPADS.1994.589904","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
A fast switching double processing architecture for multi-tasking real-time systems
A new fast switching double processing architecture for pipelined cache-based real-time computer systems is proposed to reduce the CPU stalls due to increased cache misses resulting from frequent task switching in multi-tasking real-time applications. In this architecture, two sets of registers are provided so that two tasks can be executed alternatively on a cycle-by-cycle basis. This architecture helps alleviate the problem of unpredictable cache performance due to frequent context switches in multi-tasking systems. The performance of the double processing is evaluated first through trace driven simulation for various cache configurations. An analytical performance model is then derived to further explain the performance advantage.