Welcome your new ai teammate: On safety analysis by leashing large language models

16Citations
Citations of this article
29Readers
Mendeley users who have this article in their library.
Get full text

Abstract

DevOps is a necessity in many industries, including the development of Autonomous Vehicles. In those settings, there are iterative activities that reduce the speed of SafetyOps cycles. One of these activities is "Hazard Analysis Risk Assessment"(HARA), which is an essential step to start the safety requirements specification. As a potential approach to increase the speed of this step in SafetyOps, we have delved into the capabilities of Large Language Models (LLMs). Our objective is to systematically assess their potential for application in the field of safety engineering. To that end, we propose a framework to support a higher degree of automation of HARA with LLMs. Despite our endeavors to automate as much of the process as possible, expert review remains crucial to ensure the validity and correctness of the analysis results, with necessary modifications made accordingly.

Cite

CITATION STYLE

APA

Nouri, A., Cabrero-Daniel, B., Torner, F., Sivencrona, H., & Berger, C. (2024). Welcome your new ai teammate: On safety analysis by leashing large language models. In Proceedings - 2024 IEEE/ACM 3rd International Conference on AI Engineering - Software Engineering for AI, CAIN 2024 (pp. 172–177). Association for Computing Machinery, Inc. https://doi.org/10.1145/3644815.3644953

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free