{"title":"A Hardware Approach For Accelerating Inductive Learning In Description Logic","authors":"Eyad Algahtani","doi":"10.1145/3665277","DOIUrl":null,"url":null,"abstract":"\n The employment of machine learning (ML) techniques in embedded systems, has seen constant growth in recent years, especially for black-box ML techniques (such as artificial neural networks, ANNs). However, despite the successful employment of ML techniques in embedded environments, yet, their performance potential is constrained by the limited computing resources of their embedded computers. Several hardware based approaches were developed (e.g. using FPGAs and ASICs), to address the constraints of limited computing resources. The scope of this work, focuses on improving the performance for Inductive Logic Programming (ILP) on embedded environments. ILP is a powerful logic-based ML technique that uses logic programming, to construct human-interpretable ML models; where those logic-based ML models, are capable of describing complex and multi-relational concepts. In this work, we present a hardware-based approach that accelerate the hypothesis evaluation task for ILPs in embedded environments, that uses Description Logic (DL) languages as their logic-based representation; In particular, we target the\n \n \\(\\mathcal {ALCQ}^{\\mathcal {(D)}} \\)\n \n language. According to experimental results (through an FPGA implementation), our presented approach has achieved speedups up to 48.7 folds for a disjunction of 32 concepts on 100M individuals; where the baseline performance is the sequential CPU performance of the Raspberry Pi 4. For role and concrete role restrictions, the FPGA implementation achieved speedups up to 2.4 folds (for MIN cardinality role restriction on 1M role assertions); all FPGA implemented role and concrete role restrictions, have achieved similar speedups. In the worst case scenario, the FPGA implementation achieved either a similar or slightly better performance to the baseline (for all DL operations); where worst case scenario results from using a small dataset such as: using conjunction & disjunction on < 100 individuals, and using role & concrete (float/string) role restrictions on < 100, 000 assertions.\n","PeriodicalId":50914,"journal":{"name":"ACM Transactions on Embedded Computing Systems","volume":null,"pages":null},"PeriodicalIF":2.8000,"publicationDate":"2024-05-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"ACM Transactions on Embedded Computing Systems","FirstCategoryId":"94","ListUrlMain":"https://doi.org/10.1145/3665277","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"COMPUTER SCIENCE, HARDWARE & ARCHITECTURE","Score":null,"Total":0}
引用次数: 0
Abstract
The employment of machine learning (ML) techniques in embedded systems, has seen constant growth in recent years, especially for black-box ML techniques (such as artificial neural networks, ANNs). However, despite the successful employment of ML techniques in embedded environments, yet, their performance potential is constrained by the limited computing resources of their embedded computers. Several hardware based approaches were developed (e.g. using FPGAs and ASICs), to address the constraints of limited computing resources. The scope of this work, focuses on improving the performance for Inductive Logic Programming (ILP) on embedded environments. ILP is a powerful logic-based ML technique that uses logic programming, to construct human-interpretable ML models; where those logic-based ML models, are capable of describing complex and multi-relational concepts. In this work, we present a hardware-based approach that accelerate the hypothesis evaluation task for ILPs in embedded environments, that uses Description Logic (DL) languages as their logic-based representation; In particular, we target the
\(\mathcal {ALCQ}^{\mathcal {(D)}} \)
language. According to experimental results (through an FPGA implementation), our presented approach has achieved speedups up to 48.7 folds for a disjunction of 32 concepts on 100M individuals; where the baseline performance is the sequential CPU performance of the Raspberry Pi 4. For role and concrete role restrictions, the FPGA implementation achieved speedups up to 2.4 folds (for MIN cardinality role restriction on 1M role assertions); all FPGA implemented role and concrete role restrictions, have achieved similar speedups. In the worst case scenario, the FPGA implementation achieved either a similar or slightly better performance to the baseline (for all DL operations); where worst case scenario results from using a small dataset such as: using conjunction & disjunction on < 100 individuals, and using role & concrete (float/string) role restrictions on < 100, 000 assertions.
期刊介绍:
The design of embedded computing systems, both the software and hardware, increasingly relies on sophisticated algorithms, analytical models, and methodologies. ACM Transactions on Embedded Computing Systems (TECS) aims to present the leading work relating to the analysis, design, behavior, and experience with embedded computing systems.