Next Article in Journal
Sensitivity Enhancement of 2D Material-Based Surface Plasmon Resonance Sensor with an Al–Ni Bimetallic Structure
Next Article in Special Issue
Resource Management for Collaborative 5G-NR-V2X RSUs to Enhance V2I/N Link Reliability
Previous Article in Journal
Map Space Modeling Method Reflecting Safety Margin in Coastal Water Based on Electronic Chart for Path Planning
Previous Article in Special Issue
A Contention-Free Cooperative MAC Protocol for Eliminating Heterogenous Collisions in Vehicular Ad Hoc Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Modular In-Vehicle C-ITS Architecture for Sensor Data Collection, Vehicular Communications and Cloud Connectivity

1
Instituto de Telecomunicações, Universidade de Aveiro, Campus Universitário de Santiago, 3810-193 Aveiro, Portugal
2
Departamento de Eletrónica, Telecomunicações e Informática, Instituto de Telecomunicações, Universidade de Aveiro, Campus Universitário de Santiago, 3810-193 Aveiro, Portugal
3
Instituto de Telecomunicações, Escola Superior de Tecnologia e Gestão de Águeda, Universidade de Aveiro, Campus Universitário de Santiago, 3810-193 Aveiro, Portugal
*
Author to whom correspondence should be addressed.
Submission received: 5 January 2023 / Revised: 25 January 2023 / Accepted: 31 January 2023 / Published: 3 February 2023
(This article belongs to the Special Issue Sensor Networks for Vehicular Communications)

Abstract

:
The growth of the automobile industry in recent decades and the overuse of personal vehicles have amplified problems directly related to road safety, such as the increase in traffic congestion and number of accidents, as well as the degradation of the quality of roads. At the same time, and with the contribution of climate change effects, dangerous weather events have become more common on road infrastructure. In this context, Cooperative Intelligent Transport Systems (C-ITS) and Internet of Things (IoT) solutions emerge to overcome the limitations of human and local sensory systems, through the collection and distribution of relevant data to Connected and Automated Vehicles (CAVs). In this paper, an intra- and inter-vehicle sensory data collection system is presented, starting with the acquisition of relevant data present on the Controller Area Network (CAN) bus, collected through the vehicle’s On-Board-Diagnostics II (OBD-II) port, as well as on an on-board smartphone device and possibly other additional sensors. Short-range communication technologies, such as Bluetooth Low Energy (BLE), Wi-Fi, and ITS-G5, are employed in conjunction with long-range cellular networks for data dissemination and remote cloud monitoring. The results of the experimental tests allow the analysis of the road environment, as well as the notification in near real-time of adverse road conditions to drivers. The developed data collection system reveals itself as a potentially valuable tool for improving road safety and to iterate on the current Road Weather Models (RWMs).

1. Introduction

Road traffic has experienced constant growth in recent decades around the world. As a result of the increase in the number of vehicles in circulation, concerns have arisen about road safety. Despite the efforts of manufacturers to offer more advanced and effective safety triggers in their vehicles, more than 1 million fatalities happen every year globally, according to the World Health Organization [1]. Traffic congestion is also a problem [2] and drivers spend a significant amount of time in traffic queues every day in big cities all over the world. This has consequences on public health, quality of life, and economic costs. Air pollution is also now a major concern for automotive manufacturers.
Combining all negative aspects of the road environment, a need for the development of new technologies not only in vehicles but also in infrastructure arises in order to mitigate driver’s limitations such as slow reaction times and reduced line of sight, providing them and their cars the tools to become more aware of their surroundings and more capable of preventing dangerous situations. By using both short-range and long-range communications technologies among vehicles and between vehicles and the infrastructure, a connected environment can be achieved, creating a large network of vehicular nodes and providing life to the term C-ITS. Solutions already exist where not only vehicles with communication capabilities are integrated, but also legacy vehicles that cannot communicate with other parties, creating a solid base for these technologies emerge in real traffic scenarios. Autonomous vehicles also extend the conventional vehicle’s sensor data to include other sources [3]. This can be used to improve road safety by integrating the sensors with a data collection mechanism that uses Vehicle-to-Everything (V2X) to broadcast dangerous road conditions to neighboring road users.
When sensor data from vehicle-integrated sensor systems and external sensors become available, an improved perception of the surrounding environment can be provided. In addition to that, access to in-vehicle data is increasingly desired by stakeholders in the automotive ecosystem and related sectors, with great potential to be monetized [4,5]. This work directly contributes to this extended perception of the traffic environment by proposing a modular and scalable C-ITS architecture to be deployed in vehicular systems. It integrates the collection of in-vehicular sensor data from the CAN bus, the dashboard smartphone, and other added car sensors with localized vehicular communications, long-range cellular networks, and cloud connectivity. As a result, data dissemination regarding the road environment and more detailed vehicle statuses can be implemented, enabling real-time warning notifications to vehicles and cloud monitoring of the road traffic system.
The rest of the paper is organized as follows. Section 2 focuses on the background and related work of vehicles’ sensor collection and data dissemination through vehicular networks. Then, Section 3 describes the equipment and methods used for identifying vehicular sensors’ parameters in the CAN bus. After that, Section 4 presents the proposed architecture of the in-vehicle C-ITS system, which includes the integration of the vehicle’s sensors data with the ITS-G5 protocol stack and the cloud communications framework. Section 5 describes the implementation details in the On-Board Unit (OBU), as well as on the other components such as the OBD-II reader and the smartphone device. The performed tests and evaluation results used to validate the proposed system are presented in Section 6. Finally, Section 7 concludes the paper with a summary of its contributions and main conclusions.

2. Related Work

Two important topics were analyzed for the state-of-the-art review regarding the vehicular dissemination of in-vehicle sensor data. Firstly, the relevant sensor data that can be extracted from in-vehicle systems were identified, as well as the data collection methods and their respective integration in RWMs. Then, innovative road awareness and perception mechanisms were studied in the context of C-ITS architectures based on V2X communications.

2.1. Vehicular Sensor Data

The success of C-ITS largely depends on the platform used to access, collect, and process accurate data from the environment. Guerrero-Ibaze et al. [6] state that integrating sensors with the current transportation systems allows the mitigation of some of the problems that past and current transportation systems have been facing, such as high levels of CO 2 emissions and high levels of traffic congestion. Additionally, the author discusses how user applications can benefit from sensor technologies embedded in Intelligent Transport Systems (ITS) components, which enhances the experience of various functionalities, namely driver assistance, collision alerts, and traffic control.
Inferring events or statistics from the road environment from vehicle-integrated sensor systems has a been topic of study by various authors. The comparison of each wheel speed, the position of the pedals, and the speed of the vehicle can be used to detect slippery road conditions in real-time [7,8], as well as pavement damage [9]. Adding external sensors such as the ones present in smartphones can also be combined with the vehicle’s own sensors to add redundancy and improve the accuracy of these detection mechanisms. Yunfei et al. [10] use the smartphone’s inertial sensors as well as OBD-II data, without tapping into the proprietary data, to detect slippery road conditions. Other uses of these data include the evaluation of road weather conditions [11,12,13], using the status of sensors such as windshield wipers and headlights, and remote vehicle diagnostic reporting [14] to relay malfunctions. As an effort to improve safety, vehicular data allow for the assessment of adequate driving behaviors [15] in specific road–weather conditions, causing it to be possible to send recommendations to the user [16] based either on dangerous driving or irresponsible fuel consumption.
Recent years have seen an increased focus on tire–road friction research, as it constitutes a vital aspect for determining the effectiveness of active safety systems and minimizing the likelihood and impact of traffic accidents [17,18]. Pomoni [19] explores the benefits and limitations of smart tires that are able to measure the friction of the road surface, enabling the tracking of pavement skid resistance in real time. The author mentions connectivity as a key factor when developing systems that are embedded in tires, as it is expected to be a dominant aspect of future mobility, enabling other stakeholders such as road managers or other vehicles to potentially access the data. The same work concludes that incorporating adhesive sensors within the tires is a cost-efficient method for implementing smart tires and that this approach is likely to aid drivers in adjusting their behavior in response to indications of reduced skid resistance, thereby enhancing road safety.
Bogaerts et al. [20] propose a model to enhance the current RWM by using vehicular sensor data and to enable real-time road weather warnings for local weather phenomena and dangerous road conditions. Vehicles are equipped with a CAN reader and external sensors and the data collected by the fleet are then sent to a cloud back-end using a data distribution framework. Although CAN data were collected, within this study, only data from external sensors were used to investigate the relationship between measurements and road conditions, since access to CAN parameters is not always available and can be erroneous due to the use of reverse engineering processes. Table 1 overviews the external sensors used in this study.
Every 3 s, when the Global Positioning System (GPS) module presented a new position value, the sensor measurements were collected in the current timestamp, as well as all the CAN messages within this time interval, and then sent to a data distribution platform. With the data from around 2000 h of driving collected from 15 vehicles, the authors’ findings indicated the potential value of vehicle sensors in relation to road weather. With these sensor measurements, the patterns in the temperature values were correlated with the instants when the rain started or stopped. The results showed a median Root-Mean-Square Error (RMSE) close to 4 °C between the temperature forecasts performed at a Road Weather Station (RWS) and the ones performed at close observed road segments.
Following that and using the same equipment as in [20], Bogaerts et al. [21] demonstrated two examples of how artificial intelligence and machine learning can be leveraged to process data from a fleet of vehicles and obtain relevant weather information with an improved data quality compared to the raw measurements, using convolution neural networks to extract visibility and precipitation data from raw camera images. With this, a new RWM was presented to enrich the modeling of road weather with car data at a high spatiotemporal resolution. The results showed a median RMSE below 2 °C between the temperature forecasts performed at RWS and those performed at the closely observed road segments close.

2.2. Road Perception

While data collection from the road traffic environment has been technologically evolving, the current trend in Road Weather Information Systems (RWIS) is also in the direction of greater integration with all road agents, namely connected vehicles and Vulnerable Road Users (VRUs). Low latency V2X communications between vehicles, the roadside infrastructure, and cloud systems can be used for the quick local dissemination of weather warnings [22]. For instance, a variable speed limit system can be implemented depending on weather conditions, in which the maximum speed values are disseminated through vehicular networks [23]. The 5G-SAFE-PLUS project is currently integrating 5G communications in the road infrastructure, in order to enable the fast exchange of data with vehicles and the transmission of real-time weather warnings [24].
Similarly, ensuring road users are mindful of their surroundings is essential in enabling safer conditions for them, while much significant work is performed with drivers in mind, VRU entities such as pedestrians also represent a significant portion of the road environment and may be the most neglected, with very few ways of protecting themselves during accidents.
Willecke et al. [25] mention the importance of raising the awareness of VRU objects to both drivers and these users, which can be achieved through the Collective Perception Service (CPS) defined by European Telecommunications Standards Institute (ETSI). To evaluate the impact of the CPS service on VRUs, the authors employ a simulation using Monaco SUMO Traffic (MoST) as a scenario. Here, V2X-enabled vehicles are equipped with two radar sensors to mimic the sensing capabilities of modern vehicles. The first sensor is attached to the front bumper while the second is attached to the rear bumper. Therefore, equipped vehicles have omnidirectional perception with a range, similar to the perception capabilities used for the Tesla Autopilot. The evaluation investigated the influence of VRUs on various metrics. Although the channel load induced by their inclusion was little, the number of perceived objects increased significantly, where almost half of the objects included in Collective Perception Message (CPM) were persons. The safety benefit was measured by counting the amount of VRU objects perceived in the direct vicinity of vehicles. The inclusion of CPS improved vehicle awareness and detection delays. The study concludes that CPS will play an important role in the adoption of V2X and, while much work is being performed on this topic, most of it is focused on vehicles, despite the inclusion of VRUs in the key use-cases of collective perception.
In addition to the CPS, which enables the passive perception of VRUs, VRU Awareness Service (VAS) can also be actively used to leverage their perception by means of V2X communication with the transmission of VRU Awareness Message (VAM)s. Lobo et al. [26] compare the usage of these services both in separate and combined form through the representative roundabout in the Ingolstadt traffic scenario (InTAS). When combining both services, the best perception rate was achieved, in which an average of 6% more VRUs were detected in comparison to the second best (only CPS). This also resulted in the shortest detection time. With this method, the maximum channel busy ratio was increased by 18%. The study concludes that, to improve VRUs safety, both services should be used, providing a better perception environment to each user, while not overloading the communication channels.
The work described in detail in the following sections uses the knowledge shared by these authors to build a reliable mechanism capable of extracting vehicular sensor data and disseminating it to other ITS entities, while also using a mobile application that leverages its user’s road perception.

3. Materials and Methods

In this work, a set of specific hardware and reverse engineering methods were used to extract data from the vehicle CAN bus and correctly identify the intended sensor parameters. This section presents the equipment installed inside the probe vehicle for sensor data collection together with algorithms applied to the CAN messages wiretapped from the OBD-II port.
Concerning the hardware equipment, a Carloop with a RedBear Duo combo is used as a CAN transceiver and controller, respectively. The microcontroller includes both WiFi and BLE interfaces, serving as an OBD-II reader that wirelessly connects to a vehicle’s OBU. This OBU is based on a board from PC Engines with a set of extensions, being the main hardware specifications detailed in Table 2. Figure 1 and Figure 2 show the OBD-II reader kit and the OBU used.
With respect to the reverse engineering methods, the algorithms used in this work for CAN bus parameter identification are divided into binary and non-binary methods.

3.1. Binary Parameters

For binary parameters, such as the headlight status, windshield wiper state, or pedal status, the identification can be performed through manual interaction with each car function [27]. The following phases are defined:
  • Reference—In this phase, the user needs to not perform any action inside the vehicle for a while. The algorithm searches for bits that did not show any transition, saving the reference for the next phase;
  • Monitoring—Here, the user will have the order to perform a specific number of transitions in the parameter he wants to identify (e.g., switching on and off the minimum headlights). The algorithm then filters the bits with a bit-flip count of twice the number of triggers and deletes those that present different results. These filtered parameters are the only ones that the algorithm will be aware of in the next phase;
  • Validation—This phase serves as confirmation to ensure that the identification is correct.

3.2. Non Binary Parameters

These are based on an offline algorithm that uses previously acquiredCAN traces and calculates the Pearson’s correlation coefficient between different CAN messages and other sources, such as the OBD-II. Given a pair of variables ( x , y ) , the correlation ρ is provided by:
ρ = i = 1 n ( x i x ¯ ) ( y i y ¯ ) i = 1 n ( x i x ¯ ) 2 ( y i y ¯ ) 2
Different methods exist for correlating data between different data sources, but the one described takes into account factors such as different offsets and scales.
However, this algorithm assumes that x and y are arrays of the same size, but because different CAN messages do not have the same sampling frequency, the data must be interpolated to achieve the same number of samples in both variables. Given two points ( x 1 , y 1 ) and ( x 2 , y 2 ) , an interpolated value ( x , y ) can be obtained using the linear interpolation method:
y = y 1 + ( x x 1 ) ( y 2 y 1 ) ( x 2 x 1 )
For demonstration purposes, a trip can be carried out while always creating different OBD-II requests through the CAN controller, sending every captured CAN message to an application that is listening and logging all the messages. After acquiring the trace, the following steps are carried out to identify the CAN parameters (Figure 3 illustrates this process):
  • Pre-Processing—In this step, every byte from different CAN IDs is grouped in a data structure. This results in various arrays containing the evolution of different bytes throughout the trace. Every pair ( I D , B y t e i n d e x ) becomes assigned to one of these arrays;
  • Processing—In this step, the algorithm will extract a reference array from an input and calculate its correlation to every other ( I D , B y t e i n d e x ) pair;
  • Validation—Finally, the algorithm outputs the pair ( I D , B y t e i n d e x ) that most resembles the input assigned, with its correlation coefficient.

4. Architecture

Figure 4 shows the devices installed inside the vehicle. A CAN/OBD-II reader is used to extract relevant data from the CAN bus. The smartphone is placed on the dashboard, which collects data from sensors, such as luminosity, barometric pressure, and others. Lastly, the ITS-G5/Long Term Evolution (LTE) station has the responsibility to communicate with other vehicles, the roadside infrastructure, and the cloud (either directly using short-range technologies or through cellular networks). On the intra-vehicle network, every device exchanges data via wireless technologies (BLE or WiFi). The ITS-G5 station also has the capacity to obtain data from an external Global Navigation Satellite System (GNSS) module, located in the OBU.
The collected data are sent to a cloud platform, in which a Message Queuing Telemetry Transport (MQTT) broker is used for real-time message distribution, providing an endpoint for vehicles to publish and subscribe to different topics in order to share live data to other vehicles in motion. Other cloud resources include logging services, databases for storage, and a Web Application Programming Interface (API) that offers users access to historic data.

4.1. Vehicle Sensor Data

Figure 5 summarizes the data collection process. Each vehicle contains many control units that communicate through messages on the CAN bus. These messages are read through a micro-controller attached to the vehicle OBD-II port, which periodically disseminates the relevant data to the ITS-G5 OBU using BLE. In detail, three components are defined here:
  • Vehicle (1): Integrates an internal network based on the CAN protocol, which serves as a medium for monitoring and controlling various parts of vehicles through each Electronic Control Unit (ECU). An OBD-II standardized port is present in every vehicle that allows for tracing these messages;
  • CAN/OBD-II Reader (2): A previous identification of relevant messages is needed to know what to listen to. Many of these messages obey proprietary protocols, which causes it to be hard to acquire every intended parameter. The chosen device for this was a Carloop with a RedBear Duo microcontroller, which contains Universal Serial Bus (USB), BLE, and WiFi communication interfaces;
  • ITS-G5/LTE Platform (3): Responsible for vehicular communications with roadside infrastructure, other vehicles in its vicinity and the cloud. This unit is used as an OBU, based on a platform from PC Engines that includes a WiFi module with an Atheros driver adapted for ITS-G5, a LTE module, and an external GPS receiver. Communication between this platform and the CAN reader is performed via the BLE interface.

4.2. Smartphone Sensor Data

In addition to the collection of data from the CAN bus, a smartphone is also included in this architecture, being installed in the dashboard of the vehicle in order to capture additional environmental data. The communication between the smartphone and the OBU is performed via WiFi. To accomplish this, an access point is created at the OBU by using an USB WiFi dongle that manages the local wireless network. Furthermore, a mobile application is also present with the goal of providing a user interface to the driver and passengers of the vehicle. The app improves the perception of the current road environment by displaying road and meteorological warnings to the user. Figure 6 represents a block diagram of the application architecture, showing the external components to which it is connected and the internal sensors from which it collects data at a configurable rate.
In order to allow, with a singular code base, the generation of binaries for both Android and iOS, the application was developed using React Native. For interaction with the smartphone’s sensors, Expo is used, which is an API/framework written in JavaScript, that includes implementations of many APIs common to both platforms and exposes them to React Native, allowing the code base to be written only in Typescript/Javascript. Expo-Sensors were used for sensor data collection to which the authors directly contributed by providing a Light Sensor implementation for Android ([28]). Finally, React native maps were chosen for the visualization and reporting of events, presented as map markers and/or in-app notifications, through various messages defined by ETSI, such as Decentralized Environmental Notification Messages (DENMs), Cooperative Awareness Messages (CAMs), Infrastructure to Vehicle Information Messages (IVIMs), VAMs, CPMs, and other proprietary messages such as Heartbeats, which are used to monitor the deployed road infrastructure.

4.3. Platform Internal Architecture

Figure 7 represents the organization of the various services installed in the OBU, how they interact with the wireless communication endpoints, and the data exchanges that exist. Three new types of messages are defined:
  • OVSM: OBD-II’s Vehicular Sensor Message—formed with an ETSI ITS standard header and a message body, including the reference position and the sensor data that are available from the OBD-II reader.
  • SPVSM: SmartPhone’s Vehicular Sensor Message—includes an ETSI ITS standard header and a message body with both the reference position and the sensor data available from the smartphone device;
  • VSM: Vehicular Sensor Message—combines the content of the previous messages (whenever available) with an ETSI ITS standard header.
An MQTT broker acts as the central element that behaves as a message distributor, diffusing messages between the smartphone, the OBU, and other clients. OVSMs and SPVSMs are published here. Different services are then defined to accomplish a scalable and flexible data collection mechanism:
  • it2s-obd: Manages the BLE communications between the OBU and the CAN/OBD-II reader, then uses these data to compose the OVSM messages and send them to the MQTT broker.
  • it2s-peripherals: Acts as an MQTT client, filters the received messages, such as the SPVSMs and OVSMs, and ensures their data are available to other services via shared memory files. A shared library is included to interact with them.
  • it2s-gps: Manages the external GPS module, collects its data, and ensures it is available to the services via a shared memory file. A shared library is included to interact with it.
  • it2s-data-collection: Responsible for collecting data from every data endpoint, creating VSM messages, and finally disseminating them to the cloud via cellular networks.
Finally, ETSI ITS-S Stack is the implementation of the ETSI-defined ITS Station (ITS-S) reference architecture [29]. The facilities layer is adapted to use the it2s-peripherals shared library, using the sensor data to include it in various C-ITS messages such as CAMs or DENMs.

5. Implementation

A list of the parameters to collect from each data source (OBD-II, CAN bus, and smartphone) is shown in Table 3.
On the one hand, ideally, the OBD-II parameters would also be collected directly through the CAN bus, because this eliminates the need to send request messages and wait for the responses and also improves the granularity of the collected data. On the other hand, identifying parameters in the non-standardized messages present in the bus is a difficult task, because different vehicle manufacturers develop distinct ways of representing content in CAN messages (e.g., data may be in signed or unsigned values, little-endian or big-endian, and conversion to integers may have different scales and offsets). Because of this, interpreting the data in a generic way is a difficult task. Furthermore, the CAN Database (CAN DBC) files are not easily accessible and the ones that can be found on online forums such as opengarages are often incomplete and not available for all car models, causing it to be impossible to have a ground truth set of CAN IDs.
With this in mind, a hybrid solution is implemented. Individual parameter identification was carried out through state-of-the-art reserve engineering methods in order to identify which CAN IDs were relevant, but the results of their application widely vary between different vehicles. This problem is somewhat diminished by the usage of the standardized OBD-II request/response messages, but this is also limited by the parameter IDs defined by the standard.

5.1. OBD-II Reader

This entity is implemented through the Carloop development kit that interacts with the CAN bus via the OBD-II port in conjunction with a RedBear Duo development kit that acts as a CAN controller and contains the BLE interface to connect to the ITS-G5 platform.
The data exchanges are defined by the Generic Atribute Protocol (GATT) protocol, which allows communication between devices through the transmission of attributes, consisting of concepts such as services or characteristics. A GATT server is implemented with a custom profile on the BLE peripheral (Carloop), while the BLE central (ITS-G5 platform) initiates the connection and uses the profile to receive data. This profile is configured with a rate of 10 Hz for the device to send a burst of notifications that contain every new reading that was captured from the CAN bus. The payloads of these notifications are composed of the notification ID (NID), which is used to identify which parameter it contains, and four bytes of data, encoded as floating point numbers.

5.2. Smartphone Application

The main screen of the mobile application shows a map with the current location of the user, in which other road agents (represented by CAMs, VAMs, or CPMs) and traffic events or information (disseminated via DENM, High-Definition Map (HD-Map), or IVIM messages) are displayed. Three operating modes are implemented: VRU, Traffic Control Center (TCC), and OBU modes.
The VRU mode is intended for pedestrians/cyclists; it connects to the MQTT broker in the cloud through cellular networks and subscribes to the nearby tiles, defined from the quadtree in which the user is currently located. The zoom level used in this tile system is configurable in the settings menu. In this mode, the application sends, at 1 Hz, VAM messages to the central broker in order to disseminate the presence of a vulnerable road user to other ITS stations. The TCC mode is meant for roadside infrastructure operators. In this scenario, the smartphone assumes a stationId value of 1 by default, since this is the identifier of the C-ITS central station. In this mode, every message is received. Finally, the OBU mode is for vehicle drivers to acquire an improved perception of the road environment, connecting to the OBU’s MQTT broker and subscribing to all topics. In this mode, the application connects to the OBU’s broker and sends frequent SmartPhone’s Vehicular Sensor Message (SPVSM) messages to it, in order to be later transmitted by the OBU to the cloud.
The application user interface can be visualized in Figure 8, Figure 9, Figure 10, Figure 11, Figure 12 and Figure 13. Figure 8 shows the representation of CPM messages sent from a nearby Roadside Unit (RSU) equipped with a traffic radar (vehicles in blue), as well as CAM messages sent from other vehicles (in green). In Figure 9, a representation of an IVIM is shown, signaling a speed limit of 80 km/h, which is displayed when the user enters the influence area of that traffic information sign. The current distance to the relevant info (in this case, the traffic sign) is also presented. Figure 10 displays the representation of an HD-Map message, including the temporary traffic signs and the identified obstacle on the road. The application also allows for the generation of DENM messages considering many scenarios. Figure 11 shows the menu for the selection of the detected warning event. Finally, Figure 12 presents a tab with the current sensor data of the vehicle, parsed from OBD-II’s Vehicular Sensor Messages (OVSMs). Figure 13 shows another tab with the data extracted from the smartphone sensors, also sent to the local OBU via SPVSMs. This mobile application is available on the Google Play Store (available at https://play.google.com/store/apps/details?id=com.it2s.it2smobileapp, accessed on 1 February 2023) for testing and general validation of its functionality.

5.3. Smartphone Sensor Data

While the mobile application’s primary objective is to implement an alert interface for drivers, since vehicles also roam through vast geographical areas, it is beneficial to collect data from its sensors, allowing one to infer the current status of the road environment. As the vehicle automatically collects sensor data, the same can be applied to the mobile application.
One of the biggest causes of road accidents is the lack of visibility, which happens often due to the lack of light on the roads. So, since each smartphone has a luminosity sensor, this can be used to automatically monitor the quantity of light available on the road at every location that the user passes by. This and other smartphone’s sensor data are published to the OBU’s MQTT broker, in the content of a SPVSM, and then to a cloud server in Vehicular Sensor Messages (VSMs). The application also benefits from the smartphone’s gyroscope and accelerometer, regularly collecting data regarding the phone’s rotation and acceleration in each geographical location. Sudden perturbations of these values in the same place by different OBU’s may infer the presence of road deformations or holes. Drivers should pay maximum attention to the driving task, so it is essential that these processes execute in a fully automated way, sending the data to the OBU’s broker, which are then redirected to the cloud.

5.4. ITS-G5/LTE Platform

Figure 14 shows the various data flows that exist in the OBU during execution, which are defined by the services that create and process OVSMs, SPVSMs, and VSMs. In detail:
  • OVSMs: Periodically, the Carloop device sends OBD-II requests while listening to every message that is transmitted in the CAN bus and saving the most relevant ones in local registers. A list of relevant CAN IDs and OBD-II PIDs is configured in order to indicate which messages should be sent and listened to. Additionally, at 10 Hz, a burst of notifications is sent to the OBU via BLE, which includes the already decoded values of the sensors where a change in state is observed, followed by a notification signaling the end of the burst, with the payload bits all equal to 1. On the OBU side, the service it2s-obd manages the communications with the OBD-II reader, writing the content of the sensor data in local memory and then sending OVSM messages at 1 Hz to the local broker. These messages will then be parsed by the service it2s-peripherals and written to a shared memory file;
  • SPVSMs: These messages are sent from the smartphone via WiFi to the OBU’s MQTT broker at 1 Hz. They are then received by the service it2s-peripherals, which then parses the message, analyses its content, and writes it to a shared memory file;
  • VSMs: At 1 Hz, the data from every shared memory file (GPS, OBD-II, and smartphone) is read and included in different VSM body containers, by the it2s-data-reader service. Then, the message is sent through the cellular interface LTE/5G to the cloud MQTT broker for remote monitoring of the road environment.
The generation frequencies previously mentioned are the default ones, but each value is independently configurable. On each endpoint the transmissions are asynchronous and its data flow is properly isolated through the shared memory files. These messages are defined by ASN.1 schemas, allowing data structures to be serialized and deserialized in a cross-platform way. The sensor data are located in different containers, which are included in the message bodies of OVSMs and SPVSMs. VSMs then optionally includes each of those bodies, allowing for execution independence between the services; for example, a vehicle may only be equipped with a smartphone and not the OBD-II reader. In this case, the VSMs message only includes the message body of the SPVSMs in addition to the header. The ASN.1 message formats of OVSMs, SPVSMs, and VSMs are presented in Appendix A.

6. Tests and Results

In order to validate the full end-to-end vehicular sensor dataflow, several tests were carried out on the streets of Aveiro (Portugal), where ITS-S stations were deployed on the roadside by Instituto de Telecomunicações (IT) [30], as well as during interurban trips in the Aveiro region. These tests were carried out in uncontrolled real traffic scenarios, with varying road geometry features, traffic conditions, and speed limits.

6.1. Vehicle Sensor Data

The reverse engineering methods used are mostly based on the monitoring of bitflips for binary parameters and on the correlation with other inputs for non-binary parameters, both online and offline, depending on the need to accommodate the memory limitations of the Carloop device.
As an example, the engine speed parameter corresponds to the PID 0x0C in the OBD-II standard, with its value provided by ( 256 A + B ) 4 , where A is the first data byte and B the second. These OBD-II requested values were used to find the best candidates within all of the CAN messages that are read in the bus. The correlation algorithm identified the second and third bytes of the 0xC9 CAN ID as the best candidates, with a very similar evolution compared to the OBD-II data. Figure 15 shows a comparison between both of them.
The relative position of the throttle pedal is also specified as an OBD-II standard parameter. Its value is provided by ( 100 / 255 ) A , where A is the first data byte. The algorithm identified the fifth byte of the 0xC9 CAN ID as the best candidate, even though the scale of the data is inverted. Figure 16 shows a comparison between both of them.
The speed of each wheel can also be easily identified, through the correlation with the vehicle speed when driving in a straight line. To distinguish between the four of them, performing sharp turns while driving results in bigger position circumference lines on the outer wheels than on the inner wheels, which impacts their speed. Figure 17 shows a representation of this. The data about wheel speed values are valuable because they can be used to report aquaplaning events in real-time to other drivers in the vicinity, causing them to be aware of dangerous road conditions as soon as they are detected.
As is clear in the above pictures, the sampling rate of the CAN messages is much higher than what is possible through only OBD-II, since there is no need to wait for the standard request–reply process. However, the identification of every desired parameter is also very difficult to obtain for each distinct vehicle, given the fact that manufacturers might use complimentary buses such as the Local Interconnect Network (LIN) for some sensors or more complex encoding for sensitive CAN messages. Furthermore, a 100% guarantee that the parameters identified are correct is not possible to attain, due to the intrinsic nature of reverse engineering methods. This ultimately removes the option to rely solely on the CAN bus, but its usage as a complement to an already functioning sensor data collection mechanism is valuable.
These results were acquired using an Opel Adam from 2016 and the different offset and scale values were derived to match the corresponding OBD-II data. An outline of the identified parameters and their features is presented in Table 4.

6.2. Smartphone Sensor Data

Illuminance is a parameter that varies a lot depending on the scenario. During the day, the sun can radiate directly to the sensor, which leads to a very high illuminance value, or be completely covered by a shadow from trees or buildings. Meteorological events also affect this, such as heavy fog or rain. During the night, most of the light comes from the road infrastructure or the vehicle interior, which enables this sensor to be a good indicator of whether roads are well-illuminated or for identifying zones where ice is probable to be formed because of the lack of direct sunlight. Figure 18 shows a comparison between the illuminance values recorded from day and night scenarios.
The results show that, during the night, small variations occur with consistently low values. During the day, the opposite happens; the values vary between intermediate and high illuminance, causing it to be possible to clearly distinguish between the two scenarios.
It was also studied whether road hazards such as holes and humps could be detected by variations in the measures recorded by the inertial sensors of the smartphone. The setup involved placing the smartphone in a neutral position, with the screen turned in the direction of the driver. Figure 19 shows its orientation in the three-dimensional reference space.
In order to test this use case, a trial drive was performed mostly on a highway road where large speed humps were located at the beginning and at the end of the trip. Figure 20 shows the evolution of the acceleration of the smartphone in each axis during this trial.
With this setup, the Y axis corresponds to the vertical orientation, which is why it averages 1G (9.8 m/s2). With this orientation, the X corresponds to the horizontal orientation, increasing its values according to the sharpness of the turns performed, while the Z axis represents the depth, increasing its values with the actual acceleration of the vehicle.
While the acceleration values can be useful to report the status of the vehicle, or if the driver has adequate driving, they are not enough to indicate where street humps might exist. For this case, the gyroscope might be a good indicator, because the vehicle, and consequently the smartphone, rotates around the horizontal axis when passing over a hump. Figure 21 shows the evolution of smartphone rotation on each axis during this trial.
With the setup previously described, the values in the Y axis increase while the vehicle is turning, the values on the Z axis increase when one side (both left or right wheels) of the vehicle goes lower in altitude than the other side, and, finally, the values on the X axis increase when passing over a speed bump. Figure 22 shows only the rotation around this axis.
As expected, the module of the rotation measurements from the gyroscope presented bigger values around the beginning and the end of the trip, in accordance with the location of the big speed bumps. On the other hand, recordings of the pedometer evolution did not seem to have a correlation with any specific parameter on the road trip.
This approach can work with any smartphone because the Expo-Sensors process the raw measurements in a consistent way, with configurable sampling frequencies and using the same units. However, more research needs to be performed to understand if this is a good solution, as, in these tests, the smartphone was correctly positioned and standing throughout the whole journey might not be the case during normal use, where the phone might frequently change positions/orientations, causing the measured variations in each axis to be meaningless. A solution that covers this case could rely on the analysis of the absolute value/distance from the zero of the vector (x, y, and z). In this case, it is possible to work only with a single variable, causing it to be easy to identify variations but sacrificing the ability to distinguish between each direction. In the future, this feature for detecting risky road conditions could be enhanced or complemented with other state-of-the-art solutions based on similar sensors. For instance, the work of Gnap et al. The authors of [31] identify road locations with a higher risk of cargo damage, when the accelerometer from deployed inertial sensors reaches certain thresholds (0.5 g) for at least a minimum time interval, while Jagelčák et al. [32] use GNSS coordinates with acceleration data as an input in a statistical model in order to infer the turning radius and lateral acceleration forces.
Data redundancy is also useful if one wants to increase system availability (e.g., in case of failure when reading from one of the sensors) or improve the accuracy obtained in certain sensors’ measurements. Figure 23 shows a comparison between the same parameter (barometric pressure) between different sources, OBD-II and the smartphone, where it is clear that a higher data resolution is present in the smartphone data.

6.3. Communication and Data Layers Integration

Regarding the integration of the sensors with the other layers, the collected data are presented in a web interface to ease the interpretation by humans after being transmitted by the vehicles to a cloud platform. Figure 24 shows part of this interface. These data are sent via cellular networks to a cloud service that stores it in a database and provides an endpoint for historical data access. This endpoint is available on the PASMO API, in the Vehicles section (available at https://pasmo.es.av.it.pt/docs/, accessed on 1 February 2023; Example of query from 8th March 19:00 h–20:00 h (Aveiro to Oliveira de Azeméis): https://pasmo.es.av.it.pt/api/vehicles/69?initialDate=2022-03-08T19%3A00%3A00.000Z&finalDate=2022-03-08T20%3A00%3A00.000Z&limit=100, accessed on 1 February 2023).

7. Conclusions

The presented work proposes a solution to enrich the current RWMs and the C-ITS environment, taking advantage of the inherent mobility of each vehicle and their distribution through large geographical areas. The deployment of this solution is currently limited by the non-standardized nature of the content of CAN messages, as there is a need to configure each vehicle individually. This issue is somewhat mitigated by the usage of the OBD-II protocol, which standardizes the content of its messages, as well as with the usage of external sensors such as the ones available on smartphones. Another limitation is the fact that only light and medium duty vehicles (either conventional, electric, or autonomous) are required to possess an OBD-II connector. So, this solution cannot be deployed reliably in heavy vehicles as they are not required to have this interface, while some manufacturers optionally provide it and others have alternative diagnostic systems such as J1939, J1708, and J1587.
The extracted vehicular sensor data have various potential uses, for example, meteorological stats such as temperature and barometric pressure could be used to improve the current RWM based on stationary RWS by taking advantage of the inherent mobility of vehicles, enabling the identification of areas where ice is more probable to be formed due to low-temperature values, as well as inferring forest fire likelihood, as atmospheric pressure is proportional to ambient temperature and inversely proportional to humidity. It can also be used in extreme cases to detect forest fires, as barometric pressure suddenly drops in the vicinity of forest fires due to the pressure deficit created by oxygen consumption [33,34].
The analysis of the collected data could also be valuable to other drivers and connected cars, because it allows for the detection of anomalies in the traffic environment such as congestion, dangerous driving behavior, hydroplaning, predictive vehicle maintenance, or inadequate driving that causes high engine load values, directly influencing fuel consumption. By providing real-time notifications about important and dangerous events on the road, traffic safety and efficiency can be significantly improved. Furthermore, these data could also be used by authorities to identify zones more sensitive to traffic congestion, low visibility, accidents, and where drivers do not respect speed limits, indicating the need to reinforce surveillance mechanisms in these areas. Furthermore, statistical data on traffic flows are very valuable to marketing companies and can be used for selling points to consumers.
The mobile application described in the paper works as a proof of concept that takes advantage of the ETSI-defined services (e.g., CPS and VAS) in addition to the vehicular sensor data, which were disseminated using messages specifically created for the purpose. This provides an enhanced perception of the road environment, both for drivers and for VRU entities.
Since the architecture described is designed to be modular, extending this system to other data sources should be easy regardless of their radio access technology, as all that is needed is a handler to redirect the data to the MQTT broker, if it does not already support MQTT communication. Future iterations on this work include expanding the mobile application to use the smartphone’s camera, acquiring a CAN DBC for a specific vehicle in order to work with a ground truth set of parameters, performing more tests in various car models in specific and controlled conditions, and, finally, processing and analyzing the data, which can be fed into a machine learning model with the goal of profiling various geographical locations.

Author Contributions

Conceptualization, D.R., J.A. and J.F.; methodology, D.R., G.T. and J.A.; software, D.R., G.T. and E.V.; validation, D.R., G.T. and J.A.; formal analysis, E.V. and J.F.; investigation, D.R. and J.A.; resources, E.V. and J.A.; data curation, D.R. and G.T.; writing—original draft preparation, D.R., G.T. and J.A.; writing—review and editing, E.V., J.A. and J.F.; visualization, D.R. and G.T.; supervision, E.V., J.A. and J.F.; project administration, J.A. and J.F.; funding acquisition, J.F. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the European Union/Next Generation EU, through Programa de Recuperação e Resiliência (PRR) [Project Route 25 with Nr. C645463824-00000063].

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. In any case, some of the data are already available using a public API (available at https://pasmo.es.av.it.pt/docs/, accessed on 1 February 2023).

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

Abbreviations

The following abbreviations are used in this manuscript:
APIApplication Programming Interface
BLEBluetooth Low Energy
CAMCooperative Awareness Message
CANController Area Network
CAN DBCCAN Database
CAVConnected and Automated Vehicle
C-ITSCooperative Intelligent Transport Systems
CPMCollective Perception Message
CPSCollective Perception Service
DENMDecentralized Environmental Notification Message
ECUElectronic Control Unit
ETSIEuropean Telecommunications Standards Institute
GPSGlobal Positioning System
GATTGeneric Atribute Protocol
GNSSGlobal Navigation Satellite System
HD-MapHigh-Definition Map
IoTInternet of Things
ITInstituto de Telecomunicações
ITSIntelligent Transport Systems
ITS-SITS Station
IVIMInfrastructure to Vehicle Information Message
LINLocal Interconnect Network
LTELong Term Evolution
MQTTMessage Queuing Telemetry Transport
OBUOn-Board Unit
OBD-IIOn-Board-Diagnostics II
OVSMOBD-II’s Vehicular Sensor Message
RMSERoot-Mean-Square Error
RWISRoad Weather Information Systems
RWMRoad Weather Model
RWSRoad Weather Station
RSURoadside Unit
SPVSMSmartPhone’s Vehicular Sensor Message
TCCTraffic Control Center
USBUniversal Serial Bus
V2XVehicle-to-Everything
VAMVRU Awareness Message
VASVRU Awareness Service
VSMVehicular Sensor Message
VRUVulnerable Road User

Appendix A

Figure A1. OVSM ASN.1 message format.
Figure A1. OVSM ASN.1 message format.
Sensors 23 01724 g0a1
Figure A2. SPVSM ASN.1 message format.
Figure A2. SPVSM ASN.1 message format.
Sensors 23 01724 g0a2
Figure A3. VSM ASN.1 message format.
Figure A3. VSM ASN.1 message format.
Sensors 23 01724 g0a3

References

  1. World Health Organization. Road Traffic Injuries. Available online: https://www.who.int/news-room/fact-sheets/detail/road-traffic-injuries (accessed on 24 January 2023).
  2. Afrin, T.; Yodo, N. A Survey of Road Traffic Congestion Measures towards a Sustainable and Resilient Transportation System. Sustainability 2020, 12, 4660. [Google Scholar] [CrossRef]
  3. Campbell, S.; O’Mahony, N.; Krpalcova, L.; Riordan, D.; Walsh, J.; Murphy, A.; Ryan, C. Sensor Technology in Autonomous Vehicles: A review. In Proceedings of the 2018 29th Irish Signals and Systems Conference (ISSC), Belfast, UK, 21–22 June 2018; pp. 1–4. [Google Scholar] [CrossRef]
  4. Kountche, D.; Raissi, F.; Rakotondravelona, M.; Bonetto, E.; Brevi, D.; Martin, A.; Otaegui, O.; Velez, G. Monetisation of and Access to in-Vehicle data and resources: The 5GMETA approach. arXiv 2022, arXiv:2208.11335. [Google Scholar] [CrossRef]
  5. Samuel, M.; Gupta, S. Data Monetization Model for Sustaining Smart City Initiatives. In E-Service Digital Innovation; Kang, D.K., Namisango, M.F., Eds.; IntechOpen: Rijeka, Croatia, 2022; Chapter 4. [Google Scholar] [CrossRef]
  6. Guerrero-Ibáñez, J.; Zeadally, S.; Contreras-Castillo, J. Sensor Technologies for Intelligent Transportation Systems. Sensors 2018, 18, 1212. [Google Scholar] [CrossRef] [PubMed]
  7. Enriquez, D.; Bautista, A.; Field, P.; Kim, S.i.; Jensen, S.; Ali, M.; Miller, J. CANOPNR: CAN-OBD programmable-expandable network-enabled reader for real-time tracking of slippery road conditions using vehicular parameters. In Proceedings of the 2012 15th International IEEE Conference on Intelligent Transportation Systems, Anchorage, AK, USA, 16–19 September 2012; pp. 260–264. [Google Scholar] [CrossRef]
  8. Jang, J. Pavement slipperiness detection using wheel speed and acceleration sensor data. Transp. Res. Interdiscip. Perspect. 2021, 11, 100431. [Google Scholar] [CrossRef]
  9. Van Geem, C.; Bellen, M.; Bogaerts, B.; Beusen, B.; Berlémont, B.; Denys, T.; De Meulenaere, P.; Mertens, L.; Hellinckx, P. Sensors on Vehicles (SENSOVO) – Proof-of-concept for Road Surface Distress Detection with Wheel Accelerations and ToF Camera Data Collected by a Fleet of Ordinary Vehicles. Transp. Res. Procedia 2016, 14, 2966–2975. [Google Scholar] [CrossRef]
  10. Hou, Y.; Gupta, A.; Guan, T.; Hu, S.; Su, L.; Qiao, C. VehSense: Slippery Road Detection Using Smartphones. In Proceedings of the 2017 IEEE 85th Vehicular Technology Conference (VTC Spring), Sydney, Australia, 4–7 June 2017; pp. 1–5. [Google Scholar] [CrossRef]
  11. Chapman, M.; Drobot, S.; Jensen, T.; Johansen, C.; Mahoney, W., III; Pisano, P.; McKeever, B. Using vehicle probe data to diagnose road weather conditions—Results from the Detroit IntelliDrive (SM) Field Study. Transp. Res. Rec. 2010, 2169, 116–127. [Google Scholar] [CrossRef]
  12. Mahoney, W.P., III; O’Sullivan, J.M. Realizing the potential of vehicle-based observations. Bull. Am. Meteorol. Soc. 2013, 94, 1007–1018. [Google Scholar] [CrossRef]
  13. Bartos, M.; Park, H.; Zhou, T.; Kerkez, B.; Vasudevan, R. Windshield wipers on connected vehicles produce high-accuracy rainfall maps. Sci. Rep. 2019, 9, 170. [Google Scholar] [CrossRef] [PubMed]
  14. Iqbal, M.N.; Xin, L.Y.; Rehman, W.U.; Rakhio, A.; Siddique, S.; Zahid, D.; Yasin, W.; Waqar, A.B. Diagnostic tool and remote online diagnostic system for Euro standard vehicles. In Proceedings of the 2017 IEEE 3rd Information Technology and Mechatronics Engineering Conference (ITOEC), Chongqing, China, 3–5 October 2017; pp. 415–419. [Google Scholar] [CrossRef]
  15. AbuAli, N. Advanced vehicular sensing of road artifacts and driver behavior. In Proceedings of the 2015 IEEE Symposium on Computers and Communication (ISCC), Washington, DC, USA, 6–9 July 2015; pp. 45–49. [Google Scholar] [CrossRef]
  16. Galanis, I.; Anagnostopoulos, I.; Gurunathan, P.; Burkard, D. Environmental-Based Speed Recommendation for Future Smart Cars. Future Internet 2019, 11, 78. [Google Scholar] [CrossRef]
  17. Cabrera, J.A.; Castillo, J.J.; Pérez, J.; Velasco, J.M.; Guerra, A.J.; Hernández, P. A Procedure for Determining Tire-Road Friction Characteristics Using a Modification of the Magic Formula Based on Experimental Results. Sensors 2018, 18, 896. [Google Scholar] [CrossRef] [PubMed]
  18. Hwang, Y.; Choi, S.B. Adaptive Collision Avoidance Using Road Friction Information. IEEE Trans. Intell. Transp. Syst. 2019, 20, 348–361. [Google Scholar] [CrossRef]
  19. Pomoni, M. Exploring Smart Tires as a Tool to Assist Safe Driving and Monitor Tire-Road Friction. Vehicles 2022, 4, 744–765. [Google Scholar] [CrossRef]
  20. Bogaerts, T.; Watelet, S.; Thoen, C.; Coopman, T.; Van den Bergh, J.; Reyniers, M.; Seynaeve, D.; Casteels, W.; Latré, S.; Hellinckx, P. Enhancement of road weather services using vehicle sensor data. In Proceedings of the 2022 IEEE 19th Annual Consumer Communications & Networking Conference (CCNC), Las Vegas, NV, USA, 8–11 January 2022; pp. 1–6. [Google Scholar] [CrossRef]
  21. Bogaerts, T.; Watelet, S.; De Bruyne, N.; Thoen, C.; Coopman, T.; Van den Bergh, J.; Reyniers, M.; Seynaeve, D.; Casteels, W.; Latré, S.; et al. Leveraging Artificial Intelligence and Fleet Sensor Data towards a Higher Resolution Road Weather Model. Sensors 2022, 22, 2732. [Google Scholar] [CrossRef] [PubMed]
  22. Jaeger, A. Weather Hazard Warning Application in Car-to-X Communication; Springer: Berlin/Heidelberg, Germany, 2016. [Google Scholar]
  23. Khondaker, B.; Kattan, L. Variable speed limit: A microscopic analysis in a connected vehicle environment. Transp. Res. Part Emerg. Technol. 2015, 58, 146–159. [Google Scholar] [CrossRef]
  24. 5G-SAFE-PLUS—5G Enabled Road Safety Services. Available online: https://5gsafeplus.fmi.fi/ (accessed on 1 February 2023).
  25. Willecke, A.; Garlichs, K.; Schulze, F.; Wolf, L.C. Vulnerable Road Users Are Important As Well: Persons in the Collective Perception Service. In Proceedings of the 2021 IEEE Vehicular Networking Conference (VNC), Ulm, Germany, 10–12 November 2021; pp. 24–31. [Google Scholar] [CrossRef]
  26. Lobo, S.; Festag, A.; Facchi, C. Enhancing the Safety of Vulnerable Road Users: Messaging Protocols for V2X Communication. In Proceedings of the 2022 IEEE 96th Vehicular Technology Conference (VTC2022-Fall), London, UK, 26–29 September 2022. [Google Scholar] [CrossRef]
  27. Correia, R. Cooperative in-Vehicle Sensing of Adverse Road-Weather Conditions. Master’s Thesis, Universidade de Aveiro, Aveiro, Portugal, 2021. [Google Scholar]
  28. Light Sensor Pull Request. Available online: http://github.com/expo/expo/pull/18225 (accessed on 28 December 2022).
  29. Almeida, J.; Vieira, E.; Ferreira, J.; Ribeiro, J.; Dias, T.; Moura, L. Roadside and Cloud Architecture for C-ITS Services. In Proceedings of the 27th ITS World Congress, Hamburg, Germany, 11–15 October 2021. [Google Scholar]
  30. Almeida, J.; Rufino, J.; Cardoso, F.; Gomes, M.; Ferreira, J. TRUST: Transportation and Road Monitoring System for Ubiquitous Real-Time Information Services. In Proceedings of the 2020 IEEE 91st Vehicular Technology Conference (VTC2020-Spring), Antwerp, Belgium, 25–28 May 2020; pp. 1–7. [Google Scholar] [CrossRef]
  31. Gnap, J.; Jagelčák, J.; Marienka, P.; Frančák, M.; Kostrzewski, M. Application of MEMS Sensors for Evaluation of the Dynamics for Cargo Securing on Road Vehicles. Sensors 2021, 21, 2881. [Google Scholar] [CrossRef] [PubMed]
  32. Jagelčák, J.; Gnap, J.; Kuba, O.; Frnda, J.; Kostrzewski, M. Determination of Turning Radius and Lateral Acceleration of Vehicle by GNSS/INS Sensor. Sensors 2022, 22, 2298. [Google Scholar] [CrossRef] [PubMed]
  33. Zhang, J.; Li, W.; Yin, Z.; Liu, S.; Guo, X. Forest fire detection system based on wireless sensor network. In Proceedings of the 2009 4th IEEE Conference on Industrial Electronics and Applications, Xi’an, China, 25–27 May 2009; pp. 520–523. [Google Scholar] [CrossRef]
  34. Loboda, E.; Kasymov, D.; Agafontsev, M.; Reyno, V.; Gordeev, Y.; Tarakanova, V.; Martynov, P.; Loboda, Y.; Orlov, K.; Savin, K.; et al. Effect of Small-Scale Wildfires on the Air Parameters near the Burning Centers. Atmosphere 2021, 12, 75. [Google Scholar] [CrossRef]
Figure 1. CAN transceiver and controller.
Figure 1. CAN transceiver and controller.
Sensors 23 01724 g001
Figure 2. OBU with GPS receiver.
Figure 2. OBU with GPS receiver.
Sensors 23 01724 g002
Figure 3. Representation of the proposed algorithm for CAN parameters’ identification.
Figure 3. Representation of the proposed algorithm for CAN parameters’ identification.
Sensors 23 01724 g003
Figure 4. Representation of the devices installed in the vehicle.
Figure 4. Representation of the devices installed in the vehicle.
Sensors 23 01724 g004
Figure 5. Representation of the vehicular data collection mechanism.
Figure 5. Representation of the vehicular data collection mechanism.
Sensors 23 01724 g005
Figure 6. Diagram of the mobile application architecture.
Figure 6. Diagram of the mobile application architecture.
Sensors 23 01724 g006
Figure 7. OBU’s internal architecture and interfaces.
Figure 7. OBU’s internal architecture and interfaces.
Sensors 23 01724 g007
Figure 8. CAM and CPM message representation in the mobile app.
Figure 8. CAM and CPM message representation in the mobile app.
Sensors 23 01724 g008
Figure 9. IVIM message representation in the mobile app.
Figure 9. IVIM message representation in the mobile app.
Sensors 23 01724 g009
Figure 10. HD-Map message representation in the mobile app.
Figure 10. HD-Map message representation in the mobile app.
Sensors 23 01724 g010
Figure 11. DENM message generation in the mobile app.
Figure 11. DENM message generation in the mobile app.
Sensors 23 01724 g011
Figure 12. CAN/OBD-II’s sensor data representation in the mobile app.
Figure 12. CAN/OBD-II’s sensor data representation in the mobile app.
Sensors 23 01724 g012
Figure 13. Smartphone’s sensor data representation in the mobile app.
Figure 13. Smartphone’s sensor data representation in the mobile app.
Sensors 23 01724 g013
Figure 14. Dataflows between OBU’s services and peripheral devices (OBD-II reader and smartphone).
Figure 14. Dataflows between OBU’s services and peripheral devices (OBD-II reader and smartphone).
Sensors 23 01724 g014
Figure 15. Second byte of OBD-II data for engine speed (top) and the best CAN candidate byte (bottom).
Figure 15. Second byte of OBD-II data for engine speed (top) and the best CAN candidate byte (bottom).
Sensors 23 01724 g015
Figure 16. OBD-II data for throttle position (top) and the best CAN candidate byte (bottom).
Figure 16. OBD-II data for throttle position (top) and the best CAN candidate byte (bottom).
Sensors 23 01724 g016
Figure 17. Variation of each wheel speed through a sharp turn.
Figure 17. Variation of each wheel speed through a sharp turn.
Sensors 23 01724 g017
Figure 18. Illuminance values recorded during day hours (orange), and during the night (blue).
Figure 18. Illuminance values recorded during day hours (orange), and during the night (blue).
Sensors 23 01724 g018
Figure 19. Orientation of the smartphone in the 3D reference space.
Figure 19. Orientation of the smartphone in the 3D reference space.
Sensors 23 01724 g019
Figure 20. Acceleration of the smartphone in each axis (x, y, and z).
Figure 20. Acceleration of the smartphone in each axis (x, y, and z).
Sensors 23 01724 g020
Figure 21. Rotation of the smartphone in each axis (x, y, and z).
Figure 21. Rotation of the smartphone in each axis (x, y, and z).
Sensors 23 01724 g021
Figure 22. Rotation of the smartphone in the x-axis.
Figure 22. Rotation of the smartphone in the x-axis.
Sensors 23 01724 g022
Figure 23. Barometric pressure measured from the OBD-II reader (blue) and the smartphone (orange).
Figure 23. Barometric pressure measured from the OBD-II reader (blue) and the smartphone (orange).
Sensors 23 01724 g023
Figure 24. Web interface displaying the collected data.
Figure 24. Web interface displaying the collected data.
Sensors 23 01724 g024
Table 1. Overview of sensors available on test vehicles [20].
Table 1. Overview of sensors available on test vehicles [20].
In CarSensor Box
GPS ModuleGyroscope
CameraAccelerometer
Thermal Imaging SensorTemperature Sensor
Humidity Sensor
Table 2. OBU platform specifications.
Table 2. OBU platform specifications.
Motherboard ModelPC Engines apu3d4 = AMD GX-412TC CPU/4 GB
WiFi + Bluetooth Combo AdapterVRK USB Wifi 5 and Bluetooth 5.0
Cellular ModuleHuawei ME909s-120
ITS-G5 ModuleAtheros Compex WLE200NX
GPS ReceiverG-Star IV
Table 3. Overview of sensors and parameters defined for data collection.
Table 3. Overview of sensors and parameters defined for data collection.
OBD-IICANSmartphone
Vehicle SpeedSteering Wheel AngleGyroscope
Engine SpeedWheels SpeedAccelerometer
Engine LoadHeadlights StatusMagnetometer
Ambient TemperatureWipers StatusAtmospheric Pressure
Throttle Pedal PositionBrake Pedal PositionAltitude
Atmospheric Pressure Illuminance
Pedometer
Table 4. Identified parameters on Opel Adam (2016) with its respective features.
Table 4. Identified parameters on Opel Adam (2016) with its respective features.
Opel Adam (2016)CAN ID (HEX)Byte(s)UnitOffsetScale
Engine Speed0c91rpm01
Engine Load3d11%−204/5
Vehicle Speed3e90–1km/h01/64
Throttle Pedal0c94%100−1
Wheels Speed34a0–1km/h01/32
34a2–3km/h01/32
3480–1km/h01/32
3482–3km/h01/32
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Rocha, D.; Teixeira, G.; Vieira, E.; Almeida, J.; Ferreira, J. A Modular In-Vehicle C-ITS Architecture for Sensor Data Collection, Vehicular Communications and Cloud Connectivity. Sensors 2023, 23, 1724. https://0-doi-org.brum.beds.ac.uk/10.3390/s23031724

AMA Style

Rocha D, Teixeira G, Vieira E, Almeida J, Ferreira J. A Modular In-Vehicle C-ITS Architecture for Sensor Data Collection, Vehicular Communications and Cloud Connectivity. Sensors. 2023; 23(3):1724. https://0-doi-org.brum.beds.ac.uk/10.3390/s23031724

Chicago/Turabian Style

Rocha, David, Gil Teixeira, Emanuel Vieira, João Almeida, and Joaquim Ferreira. 2023. "A Modular In-Vehicle C-ITS Architecture for Sensor Data Collection, Vehicular Communications and Cloud Connectivity" Sensors 23, no. 3: 1724. https://0-doi-org.brum.beds.ac.uk/10.3390/s23031724

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop