TY - GEN
T1 - FaultLines - Evaluating the Efficacy of Open-Source Large Language Models for Fault Detection in Cyber-Physical Systems
AU - Muhlburger, Herbert
AU - Wotawa, Franz
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024/9/25
Y1 - 2024/9/25
N2 - Cyber-physical systems are integral to the infrastructure of global communication and transportation networks, which makes it crucial to detect faults, prevent cyber attacks, and ensure operational safety. Although machine learning techniques, including large language models (LLMs), have been explored for fault detection, the efficacy of open-source LLMs remains underexplored. In this work, we assess the capabilities of eight open-source LLMs in identifying faults in cyber-physical systems using a simulation dataset from monitoring an electrified vehicle's battery management system. By applying pretrained LLMs without fine-tuning and incorporating retrieval augmented generation (RAG) techniques alongside textual encoding methods, our study aims to explore the potential of open LLMs in fault detection. Our results show that open LLMs can effectively identify faults, with Mistral out-performing alternative models such as Mixtral, codellama, and Gemma in precision, recall, and Fl-score metrics. Furthermore, our results highlight the importance of textual encoding strategies in enhancing the fault detection capabilities of LLMs, which possess a degree of explanatory power with respect to the detected anomalies. This work demonstrates the feasibility of using open LLMs for fault detection in cyber-physical systems and opens avenues for future research to enhance fault detection and fault localization.
AB - Cyber-physical systems are integral to the infrastructure of global communication and transportation networks, which makes it crucial to detect faults, prevent cyber attacks, and ensure operational safety. Although machine learning techniques, including large language models (LLMs), have been explored for fault detection, the efficacy of open-source LLMs remains underexplored. In this work, we assess the capabilities of eight open-source LLMs in identifying faults in cyber-physical systems using a simulation dataset from monitoring an electrified vehicle's battery management system. By applying pretrained LLMs without fine-tuning and incorporating retrieval augmented generation (RAG) techniques alongside textual encoding methods, our study aims to explore the potential of open LLMs in fault detection. Our results show that open LLMs can effectively identify faults, with Mistral out-performing alternative models such as Mixtral, codellama, and Gemma in precision, recall, and Fl-score metrics. Furthermore, our results highlight the importance of textual encoding strategies in enhancing the fault detection capabilities of LLMs, which possess a degree of explanatory power with respect to the detected anomalies. This work demonstrates the feasibility of using open LLMs for fault detection in cyber-physical systems and opens avenues for future research to enhance fault detection and fault localization.
KW - Anomaly detection
KW - Cyber-physical systems fault detection
KW - Open-source large language models (LLMs)
KW - Retrieval Augmented Generation (RAG) techniques
KW - Textual encoding strategies for LLMs
UR - http://www.scopus.com/inward/record.url?scp=85206479922&partnerID=8YFLogxK
U2 - 10.1109/AITest62860.2024.00014
DO - 10.1109/AITest62860.2024.00014
M3 - Conference paper
AN - SCOPUS:85206479922
T3 - Proceedings - 6th IEEE International Conference on Artificial Intelligence Testing, AITest 2024
SP - 47
EP - 54
BT - Proceedings - 6th IEEE International Conference on Artificial Intelligence Testing, AITest 2024
PB - IEEE Institute of Electrical and Electronics Engineers
T2 - 6th IEEE International Conference on Artificial Intelligence Testing, AITest 2024
Y2 - 15 July 2024 through 18 July 2024
ER -