Title | A Context-Based Decision-Making Trust Scheme for Malicious Detection in Connected and Autonomous Vehicles |
Publication Type | Conference Paper |
Year of Publication | 2022 |
Authors | Eze, Emmanuel O., Keates, Simeon, Pedram, Kamran, Esfahani, Alireza, Odih, Uchenna |
Conference Name | 2022 International Conference on Computing, Electronics & Communications Engineering (iCCECE) |
Date Published | aug |
Keywords | Collaboration, connected vehicles, context-based, decision making, decision-making, false trust, features selection and classifications, machine learning, policy-based governance, Predictive models, psychology, pubcrawl, Real-time Systems, resilience, Resiliency, Roads, Safety, Scalability, Trust |
Abstract | The fast-evolving Intelligent Transportation Systems (ITS) are crucial in the 21st century, promising answers to congestion and accidents that bother people worldwide. ITS applications such as Connected and Autonomous Vehicle (CAVs) update and broadcasts road incident event messages, and this requires significant data to be transmitted between vehicles for a decision to be made in real-time. However, broadcasting trusted incident messages such as accident alerts between vehicles pose a challenge for CAVs. Most of the existing-trust solutions are based on the vehicle's direct interaction base reputation and the psychological approaches to evaluate the trustworthiness of the received messages. This paper provides a scheme for improving trust in the received incident alert messages for real-time decision-making to detect malicious alerts between CAVs using direct and indirect interactions. This paper applies artificial intelligence and statistical data classification for decision-making on the received messages. The model is trained based on the US Department of Technology Safety Pilot Deployment Model (SPMD). An Autonomous Decision-making Trust Scheme (ADmTS) that incorporates a machine learning algorithm and a local trust manager for decision-making has been developed. The experiment showed that the trained model could make correct predictions such as 98% and 0.55% standard deviation accuracy in predicting false alerts on the 25% malicious data |
DOI | 10.1109/iCCECE55162.2022.9875087 |
Citation Key | eze_context-based_2022 |