Next Article in Journal
Online Judging Platform Utilizing Dynamic Plagiarism Detection Facilities
Previous Article in Journal
Modeling Networked Telemetry
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Improving Response Time for Crowd Management in Hajj

1
Computer Engineering Department, Umm Al-Qura University, Makkah 24231, Saudi Arabia
2
Science and Technology Unit, Umm Al-Qura University, Makkah 24231, Saudi Arabia
*
Authors to whom correspondence should be addressed.
Submission received: 2 March 2021 / Revised: 28 March 2021 / Accepted: 30 March 2021 / Published: 6 April 2021

Abstract

:
Flying Adhoc Network (FANET) is a particular type of Mobile Adhoc Network (MANET) that consists of flying drones or unmanned aerial vehicles (UAVs). MANETs are especially useful in rural and remote areas, where the lack of public networks necessitates data delivery through mobile nodes. Additionally, FANETs provide better coverage where there is a lack of roads. Generally, the goal of FANETs is to provide multimedia data to applications such as search and rescue operations, forest fire detection, surveillance and patrol, environmental monitoring, and traffic and urban monitoring. The above applications’ performance and efficiency depend on the quality and timely delivery of these essential data from an area of interest to control centers. This paper presents a Priority-based Routing Framework for Flying Adhoc Networks (PRoFFAN) for the expedited delivery of essential multimedia data to control centers. PRoFFAN reduces the FANET application’s response time by prioritizing the sending and forwarding of critical image data from the UAV to the control center. Our motivation application is crowd management; we believe that having important image features as early as possible will save lives and enhance the crowd’s safety and flow. We integrated PRoFFAN over the RPL routing layer of Contiki-NG’s IPv6 network stack. We used simulations in Cooja to demonstrate the benefit of PRoFFAN over conventional ZigBee.

1. Introduction

Hajj is an annual religious event where millions of pilgrims from all over the world gather to perform certain ritual, which include mobility from one site to another. With this massive gathering of people, crowd management is a significant challenge to ensure safety and security for pilgrims during their movement.
During Hajj, the crowd might walk for long distances between different sites, extending the risk zones and increasing the challenge of providing good crowd management. Figure 1a shows a crowded street full of pilgrims in one of the Hajj sites called Mina. Figure 1b shows the whole Mina site and most of its interconnected streets and pathways full of crowds.
Hajj organizing authorities have deployed more than 4000 fixed surveillance cameras in different sites to monitor crowd safety. Figure 1c shows a control room receiving feeds from various crowd-monitoring cameras. Despite its massive and wide deployment, fixed cameras have a limited view and limited alleviation. They do not cover large geographical areas without the use of complicated image merging and stitching technology. Moreover, fixed cameras lack a bird’s-eye view, which is required for most crowd flow detection. Therefore, given the vast area that required coverage in the case of Hajj, image data from fixed cameras are insufficient for crowd-flow detection applications.
The utilization of Unmanned Ariel Vehicles (UAV) or drones is becoming part of our daily life, especially in critical missions like search and rescue, surveillance, reconnaissance, and emergency response. In crowded situations, drone utilization is becoming increasingly interesting in the discovery of earlier signs of a stampede, congestion, and other crowd problems. Real-time images and videos taken by drones will have a wider field of view, cover huge areas, have zero obstruction in open areas, and have an adjustable alleviation view. In [1], the author provided a management framework for crowd events. They suggested using technologies to collect the pilgrims’ data in Hajj, and later enhance security measures and crowd management.
Most civilian applications that use UAVs and drones are offline or not time-critical. With mission-critical applications that require real-time imagery feed, capable communication modules should be provided with UAV to transmit real-time feed images.
Finally, if the images captured by the UAVs are transmitted to the base station using typical FANET communication protocols, the FANET applications must wait for the entire image to be received before it can process the image. Image-processing applications require very large amounts of data to be exchanged between the UAV and base-station. In traditional wireless sensor networks that sense light, humidity, pressure, etc., the traffic generated by a sensing node is limited to the scalar data. In most cases, the memory size required to store and send is 16-bits per reading. On the other hand, image data are in kilobytes or even megabytes. For example, a raw Red–Green–Blue (RGB) image of 128 × 128 pixels with 24-bits per pixel (8 bits per color) will be of 128 × 128 × 24 = 393,216 bits (approximately 48 kilobytes). These are magnitudes larger than traditional sensor data. Although image compression reduces the size of data to be transmitted, it is still not enough. Mukherjee et al. [2] presented a comprehensive survey about the FANET. The author discussed the design issues of FANET. These issues are: Node Mobility, Mobility Models, Node Density, Topology Changes, Radio Propagation Model, Power Consumption and Refueling, and Localization. The networking protocols for FANET are also discussed in detail, such as Static, Proactive, Reactive, Hybrid, and Geographical and Position-based protocols. The author also highlighted some of the research areas of FANET. One of the applications is disaster and crowd management.
This paper proposes an image transmission mechanism called a Priority-Based Routing Framework for Flying Area Networks (PRoFFAN), which will enhance the delivery speed of essential image features from a UAV to the base stations. Our motivation application is crowd management, and we believe that obtaining important image features as early as possible will save lives and enhance the safety and flow of the crowd. In this work, we have simulated the transmission from UAVs to the base-station in Cooja. We implemented our framework in Contiki-NG OS. The simulations were carried out using standard aerial images used in image processing. We used image peak signal-to-noise ratio (PSNR) [3] and structural similarity index (SSIM) [4] to show the benefit of using PRoFFAN as compared to typical Zigbee based communication.
As discussed before, the image size can be reduced using image compression. PRoFFAN leverages the ability of image compression algorithms that generate multiple layers of compressed data. The first layer contains the most prominent features of the image, for example, the edges of objects or coarse image data. The subsequent layers contain the details that when merged with the first layer, and restore the original image. Using such image compression algorithms, the response time of FANET applications can be reduced. If the base-station receives the image data required for first pass sooner than data required for subsequent passes, it can start processing the first pass and take action accordingly while the data of subsequent layers arrive.
The rest of the paper is organized as follows. The following section discusses related works. Section 3 discusses the PRoFFAN methodology. Section 4 contains details of the experiment setup. Section 5 elaborates the results, whereas the paper is concluded in Section 6.

2. Related Work

FANET has become very popular after drones, commercial intervention and usage. Now, in many applications such as monitoring infrastructure, civil works, agriculture monitoring, mass crowd monitoring and many others, drones are used to collect multimedia data. FANETs can be described as a group of Unmanned Aerial Vehicles, which communicate without any access point. In contrast, at least one among the group acts as the server or satellite. These are automated systems, which don’t involve any human input for their basic operations [5]. All the network activity, including the discovery of its topology or the sending of messages, is carried out by the nodes that integrate it. This type of network has an arbitrary topology. The main consequence of this mobility is that links can be formed and broken frequently, which implies that the network must be self-organized. The path between an origin and a specific destination that crosses several intermediate nodes can be modified. When this happens, the network must re-organize itself to establish a new path in the least possible time. This functionality requires that each member of the network will forward data on behalf of other members who act as a router. The nodes of a FANET network have components or associated factors which are considered when carrying out an efficient routing. These factors include auto-configuration, bandwidth, distance, energy, routing path, partitions/unions, speed, and variation in the routing path. FANET is based on the peer-to-peer (P2P) concept, where peers are mostly UAVs.
Flammini et al. [6] presented the usage of drones in railways to monitor the situation, issues early warnings, and some decision-support applications. Their study and framework help in automatically checking the railway infrastructure parameters. This monitoring helps provide early warnings, any threats related to faults and security, or any natural hazards or intruder attacks. In railway applications, drones help with three major applications, namely structural monitoring, environmental security monitoring, and physical security monitoring.
Another exciting application area is agriculture monitoring using drones. The drone’s technology helps farmers optimize the resources required for agriculture, like the amount of water required, and to monitor pests and other diseases. It helps the farmer to target the precision in harvesting and fertilizing. The drones help with collecting the data from the fields, its processing analysis, and management. The drones also play an active role in agriculture applications like interventions in the irrigation and fertilizers of crops [7]. Farmers utilize the drones in many ways, like counting the number of fruits on every tree, the leaf area index measurement, the computation of Normalized Difference Vegetation Index (NDVI), identifying plant diseases, and remote sensing for water management and irrigation control applications, so that the farmer can apply the chemical, and pesticides on time [8].
Drones are also used in monitoring disasters. FANETs are deployed to the area to encounter the situation as, most of the time, humans cannot reach the remote location due to obstacles [9]. Drones are also beneficial in search and rescue operations where all the conventional mobile networks have failed or are temporarily unavailable. FANETs can search for hostages and help humans reach those areas which are physically impossible to reach themselves [7].
With the use of FANETs, the monitoring and analysis of the construction area are easy. Humans can verify the quality, progress, and work condition remotely using FANETs [9].
Azevedo et al. [10] discussed the challenges of FANET. The first challenge belongs to signal transmission because most of the directional antennas are omnidirectional, and using them in drones may not be very efficient in terms of quality and energy consumption. The new antenna developed with beamforming technology [11] is more effective and efficient in drones. The routing protocols can be classified into five major categories, including static routing, position-based routing, and Hierarchical routing, position-based routing, swarm-based routing, hierarchical routing, and topology-based routing.
Noor et al. [12] provide a review of the FANET communication perspective on wireless technologies, applications, challenges, and open research topics. Some of the applications they cited include monitoring, search and rescue. These applications require multimedia capture and transmission from drones. In addition to the challenges faced by FANETs, they also provide a list of available options for communication between the drones and base stations.
The authors of [13] designed a FANET with a fleet of drones for monitoring rural areas that do not have Internet access. They argue that monitorung rural areas can be time-consuming, labor-intensive, and expensive due to the limited access and large distances to cover. They use drones to overcome these challenges. In their paper, they presented a FANET video monitoring platform by describing the architecture, the service chains used to realize the video delivery service, and an analytical model used to evaluate the computational load of the platform nodes to allow the network orchestrator to decide the backbone drones where the virtual functions will be run, and the relative resources to be allocated.
Zhang et al. [14] also provide a similar solution. They argue that public ground networks do not provide sufficient coverage, which has resulted in the problematic backhaul of monitoring data from remote areas of interest. Because of the urgent demand for low-cost data collection in such areas, they propose a drone-enabled IoT relay system to provide high-speed data collection to support remote environmental monitoring. Their solution is based on 5 GHz and LoRa technology. They carried out numerous real-world experiments to validate the designed drone-enabled IoT relay the system’s effectiveness and show its capability of high-speed data collection.
Singh et al. [15] suggest using drones for surveillance of violent individuals using ScatterNet Hybrid Deep Learning Network. Their system detects the violent individuals in real-time by processing the drone images in the cloud.
Wang et al. [16] talk about the collaboration of UAVs and unmanned ground vehicles (UVGs) in intelligence gathering, border surveillance, and crowd control. They propose a comprehensive planning and control framework based on a dynamic-data-driven, adaptive multi-scale simulation (DDDAMS). They discussed their solution and shared the results of their framework, algorithms, and testbed.
Almagbile et al. [17] discuss the estimation of crowd density from images captured from UAVs using corner detection and cluster analysis. They propose a testing procedure based on a feature from accelerated segment test (FAST) algorithms with a high level of correctness. Previously, [18], they proposed and tested a crowd monitoring system for pedestrians. They deployed a crowd monitoring technique using real-time images taken by UAVs. Captured images were analysed to calculate crowd density using image segmentation procedures.
Motlagh et al. [19] presented the UAV potential and IoT devices in crowd surveillance. They presented an IoT-based platform and developed a testbed for crowd surveillance. In crowd surveillance, the UAV can help provide safety to people, detect criminal activities, use facial recognition, and many other applications. In their crowd surveillance experimentation, they used a UAV equipped with a video camera, sensing devices and modems, and used an LTE network to connect with the base station. The primary purpose was to evaluate the energy consumption of UAVs during onboard processing and ground station processing. The UAV onboard video and image processing took a lot of time and energy.
Choi et al. [20] presented a geolocation-based routing protocol for FANETs. Their protocol supports multihop routing. The geolocation-based protocols are dynamic and select the next hop based on demand. Their geolocation-based algorithm operated over the traditional IP layer protocol. The geolocation is added to the sender node’s header, and the node selects the closest possible neighbor based on the location in its header. This mechanism appears proactive and efficient.
Chriki et al. [21] monitored a crowded area using multiple UAVs and presented a centralized data-oriented communication architecture for crowd surveillance. The data-oriented approach prioritizes the message before sending it. The ground station of the UAV network served as a central coordinator to manage the bandwidth of UAVs. The bandwidth was optimized by prioritizing the data as critical and vital. The central coordinator manages the UAVs according to the priority of information that UAVs want to send. This approach saves the energy and bandwidth of the UAV network.

3. Methodology

Some FANET applications depend on the process of capturing an image at the camera node (through a camera mounted on UAV) and its transmission to the base station (sink node). The information may be passed in sequential order from the camera node to the sink node through intermediate nodes. Typically, packets are not prioritized based on their type, and hence, the critical information may reach the sink node with a delay. In this paper, we propose a methodology to transmit critical image information in FANET with priority. Figure 2 shows the methodology used in this paper. In FANET, the images are captured by UAVs and need to be transmitted. The captured image is passed to the Encoder for encoding the image.
The first step is to convert the image into grayscale to obtain 8 bits per pixel. We used the bit-plane encoding [22] to encode the image into two layers: high-priority and low-priority. The bit-plane slicing is a technique used in image processing to slice the images into different planes. The layer formed by these planes contains only binary information. The high-priority layers contain the bit-planes of the bit location 8 to 5 of each pixel, whereas the low-priority layer contains bit planes of bit locations of 4 to 1 of that pixel. One pixel contributes to half a byte of each pixel of the resultant layers. Figure 3 shows the encoding mechanism of two consecutive pixels. The four most significant bits (MSBs) of Pixel-1 are stored as MSBs of the first pixel of the high-priority layer, and the four MSBs of Pixel-2 are stored as the four least significant bits (LSBs) of the first pixel of the high-priority layer. Similarly, the four LSBs of Pixel-1 become the MSBs of the first pixel of the low-priority layer, followed by four LSBs of Pixel-2. The process continues for the whole image, and at the end of the encoding mechanism, PRoFFAN converts one grayscale image into two layers.
All the packets are arranged as low-priority and high-priority based on MSBs and LSBs. All the high-priority packets contain the MSBs of the image’s two pixels, while the low-priority packets contain the LSBs of two pixels of the image. Typically, to send an image over a network, the packets are sent in a sequence without priority. In PRoFFAN, after encoding, the layers are prioritized before sending them. The prioritization module enqueues all the high-priority layers first. Hence, the transmission of all packets of the high-priority layer occurs first, before the low-priority layer.
The bit-plane encoding mechanism used does not cause much degradation in the image quality, as the high-priority layer contains the MSBs of pixels representing most information of the pixel as it has high intensity compared to LSBs. The primary purpose of using PRoFFAN is to transmit critical information with priority to handle emergencies efficiently. Hence, the image is received at the sink node with the high-priority layers first. The image received at the sink node is low in quality as compared to the original image. Based on the pixel values in the high-priority layers, the application can easily extract low-quality image information. However, the sink node does not yet receive all the pixels. The significant difference lies in the decision-making time. If the image is sent without PRoFFAN, the application cannot decide until it receives the full image. While using the PRoFFAN, where all the high-priority layers arrive first, the application can quickly obtain enough information for decision-making, with a significant time decrease.
PRoFFAN uses the algorithm outlined in the flow chart of Figure 4 once the bit-planes are ready to be transmitted. The packet generation stage consists of enclosing the maximum number of bytes from a bit-plane into a packet. Depending on the bit-plane, a priority is assigned to each packet. The packet en-queuing stage checks the priority of the packet received and en-queues it either to the head of the transmission queue, if it is a high-priority packet, or at the tail of the transmission queue if the received packet is of low-priority. Finally, the packet transmission stage transmits the packet from the head of the transmission queue. If a collision is detected, the next transmission is delayed based on contention window. The contention window for the high-priority packet is half the size of the contention window for the low-priority packet. In this way, in case of collision, a high-priority packet would be scheduled sooner than a low-priority packet.

4. Experimentation

The scenario shown in Figure 5 was selected to demonstrate the effectiveness of PRoFFAN in FANETs. In this scenario, the base station (sink node) is located near the field of interest. It has a range R of 500 m. Hence, the base station has a one km radio range. Seventeen UAVs pass by the base station, one by one, but at different speeds.
The first UAV passes at a speed of 1 m/s (3.6 km/h). Therefore, the first UAV will be in the range of the base station for 1000 s. In this time, the UAV has to establish a connection with the sink and transmit one image.
The second UAV passes at a speed of 2 m/s (7.2 km/h). As a result, the second UAV will have 500 s to establish a connection with the sink and transmit one image. The exact process continues for all seventeen UAVs. Hence, the last UAV passes at 17 m/s (61.2 km/h) and has about 1 min to establish a connection and transmit one image to the sink.
All UAVs send the same image of 512 × 512 resolution. As discussed in the Research Methodology section, the image is broken down into two layers. The high-priority layer is sent before the low-priority layer. The packet rate is set at 32 packets per second. Each packet is 100 bytes. Each byte contains either high-priority information of two pixels or low-priority information of two pixels. Hence, one packet contains information of 200 pixels.
In our scenario, the UAVs try to send as much image information. In these experiments, we do not optimize the connection establishment between UAV and base station. All experiments are set with default communication parameters of Contiki-NG OS for RPL routing. Once a connection is established between the base station and the UAV, a UAV at 1 m/s should be able to transmit the entire image data, i.e., 2622 packets, in less than 90 s. Note that the 1 m/s UAV is in range of the base station for 1000 s, so it should be able to transmit the entire image. A UAV at 2 m/s should be able to transmit the entire image as well. However, UAVs at a higher speed would have less time to transmit the entire image. Additionally, packet loss may occur due to other reasons, such as multi-path fading, degradation of transmission, channel characteristics, etc.
As soon as a packet of image information is received at the base station, PRoFFAN starts to decode the information and construct a received image. At the base station, PRoFFAN calculates image Peak Signal-to-Noise Ratio (PSNR) and Structural Similarity Index (SSIM) while receiving packets on the go. When the PSNR and SSIM reach the threshold required by the FANET application for processing images, PRoFFAN passed the image to the FANET application. Note that the goal of these experiments is to see if the introduction of PRoFFAN reduces the FANET application’s response time.
This paper focuses on single-hop image transmission between the UAV (Camera Node) and the base station (Sink Node). Therefore, only one UAV is in range for the base station. Each UAV travels a distance of 2 km, of which 1 km is in the sink range. Table 1 summarizes the simulation parameters.
When one UAV reaches its end position, the next UAV starts its journey. In this way, the sink node collects packets from all UAVs, constructs received images, and calculates PSNR and SSIM. Simulations were carried out with PRoFFAN and without PRoFFAN to see the difference. Experiment results are explained in the next section.

5. Results

The following (Figure 6) shows the packet delivery ratios for high-priority and low-priority packets sent using PRoFFAN and all packets sent with PRoFFAN disabled.
At UAV speeds less than 5 m/s (18 km/h), more than 90 percent of packets were received for all three packet types. The reason is that the UAV speed is less enough than the sink received most packets.
As UAV speed increased from 5 m/s (18 km/h) to 11 m/s (40 km/h), approximately 60 percent of packets made it to the sink when PRoFFAN was disabled. When PRoFFAN was enabled, the delivery ratio of high-priority packets was above 90 percent. The delivery ratio of low priority packets was reduced to 30 percent to compensate for this.
At UAV speeds higher than 11 m/s (40 km/h), it can be seen that the base station receives twice the amount of high-priority packet compared to packets sent with PRoFFAN disabled. Whatever the speed, at any time, the information that the base station receives is of more use to the FANET application than information received with PRoFFAN disabled.
The benefit of using PRoFFAN is more evident when we look at the PSNR and SSIM of received images. Figure 7a,b show the PSNR and SSIM of received images. Both figures show that as the speed of the UAVs increase, the PSNR and SSIM decrease. However, with PRoFFAN enabled, the quality of images received at the base station is much higher. The curve of PSNR and SSIM for PRoFFAN is analogous with the number of high-priority packets received at the base station. With a higher delivery ratio of high-priority packets, the base station’s image information is more beneficial for the FANET application because its PSNR and SSIM are better. Hence, PRoFFAN assists FANET applications in improving their response time by providing them with better quality images than networks without PRoFFAN under the same conditions.
Figure 8a,b show the PSNR and SSIM calculated after every one second at the base station for packets received from a UAV traveling at 10 km/h. Without PRoFFAN, the UAV’s entire journey in the base station’s range could transmit packets that contributed to PSNR, reaching a max value of 6 and SSIM of 0.28, respectively, in 120 s (represented by marker 4). However, the same quality of PSNR = 6 reached the base station in almost one-third of the time using PRoFFAN at 34 s (represented by marker 3). Assuming that the FANET application required a PSNR = 6 for processing the incoming image, without PRoFFAN, the image of PSNR = 6 was received in 120 s, whereas, with PRoFFAN enabled, the same quality of image reached the base station in 34 s. Hence, PRoFFAN saved 86 s.
The vertical line at the 60 s mark in Figure 8a depicts the deadline of 60 s. Without PRoFFAN, the image of PSNR = 4 reached the base station (represented by marker 2). On the other hand, with PRoFFAN enabled, an image of PSNR = 14 reached the base station in 60 s (represented by marker 1).
In fact, at 65 s (represented by mark 5), an image with a PSNR of 17.5 had reached the base station. This improvement in the quality of the image was achieved by enabling PRoFFAN. Hence, the application can use images of much better quality for processing. This is how PRoFFAN assists in improving response time and response efficiency in FANET applications with UAVs traveling at low speeds.
Figure 9 shows the image received at Marker 4 in Figure 8a. The image was received in 120 s with a PSNR of 6 without PRoFFAN. Figure 9 also shows the image received at Marker 3 in Figure 8a. This image was received in 34 s with PRoFFAN enabled with the same PSNR of 6. Hence, PRoFFAN saved 86 s of the FANET application.
Figure 10a,b show the PSNR and SSIM calculated after every one second at the base station for packets received from a UAV traveling at 30 km/h. Without PRoFFAN, the UAV’s entire journey in the base station’s range could transmit packets that contributed to PSNR, reaching a max value of 5 and SSIM of 0.19, respectively, in 86 s (represented by marker 4). However, the same quality of PSNR = 5 reached the base station in almost one-third of the time using PRoFFAN at 29 s (represented by marker 3). Assuming that the FANET application required a PSNR = 5 for processing the incoming image, without PRoFFAN, the image of PSNR = 5 was received in 86 s, whereas, with PRoFFAN enabled, the same quality of image reached the base station in 29 s. Hence, PRoFFAN saved 57 s.
The vertical line at 60 s mark in Figure 10a depicts the deadline of 60 s. Without PRoFFAN, the image of PSNR = 4 reached the base station (represented by marker 2). On the other hand, with PRoFFAN enabled, an image of PSNR = 13.4 had reached the base station in 60 s (represented by marker 1).
In fact, at the 63 s (represented by mark 5), an image with a PSNR of 16.4 reached the base station. This improvement in quality of the image was achieved by enabling PRoFFAN. Hence, the application can use images of much better quality for processing. This is how PRoFFAN assists in improving response time and response efficiency in FANET applications with UAVs traveling at medium speeds.
Figure 11a,b show the PSNR and SSIM calculated at after every one second at the base station for packets received from a UAV traveling at 60 km/h. Without PRoFFAN, the UAV entire journey in the base station’s range could transmit packets that contributed to PSNR reaching a max value of 3.2 and SSIM of 0.16, respectively, in 25 s (represented by marker 4). However, the same quality image reached the base station in 13 s PRoFFAN (represented by marker 3). Assuming that the FANET application required a PSNR = 3.2 for processing the incoming image, without PRoFFAN, the image of PSNR = 3.2 was received in 25 s, whereas, with PRoFFAN enabled, the same quality of image reached the base station in 13 s. Hence, PRoFFAN saved 12 s.
The vertical line at 30 s mark in Figure 11a depicts the deadline of 30 s. Without PRoFFAN, the image of PSNR = 3.2 reached the base station (represented by marker 2). On the other hand, with PRoFFAN enabled, an image of PSNR = 5.2 reached the base station in 30 s (represented by marker 1).
In fact, at 49 s (represented by mark 5), an image with a PSNR of 9.7 had reached the base station. This improvement in the quality of the image was achieved by enabling PRoFFAN. Hence, the application can use images of much better quality for processing. This is how PRoFFAN improves response time and response efficiency in FANET applications with UAVs traveling at medium speeds and higher speeds.

6. Conclusions

FANET applications require certain the PSNR and/or SSIM of received images at the base-station in a lossy network. Using simulation experimentation, this paper demonstrates the usefulness of PRoFFAN in improving the response time of FANET applications by prioritizing packets that contribute more to higher PSNR and SSIM. In this paper, we demonstrated that PRoFFAN delivers images of higher quality to base-stations for FANET applications in the same amount of time and network conditions as compared to a typical Zigbee-based network. It was also seen that deadline-based FANET applications receive an image of higher quality in the same time as the typical Zigbee based network.
Some areas of PRoFFAN require improvements in future, especially Encoding and Transmission stages, outlined in Figure 2. More efficient image encoding techniques have to be explored in future. Morevoer, the current implementation of PRoFFAN was tested and compared with Zigbee-based networks. In future, other low-powered transmission techniques have to be considered.
Finally, this paper demonstrated PRoFFAN’s usefulness using simulations. In the future, we plan to implement a real-time FANET test-bed.

Author Contributions

Methodology, A.N.; Literature, A.N. and E.F.; Experimentation and Results, A.A.S.; writing—original draft preparation, E.F., A.N. and A.A.S.; writing—review and editing, E.F.; funding acquisition, E.F. All authors have read and agreed to the published version of the manuscript.

Funding

The authors extend their appreciation to the Deputyship for Research Innovation, Ministry of Education in Saudi Arabia for funding this research work through project number 0909.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Authors can confirm that all relevant data are included in the article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Yamin, M. Managing crowds with technology: Cases of Hajj and Kumbh Mela. Int. J. Inf. Technol. 2019, 11, 229–237. [Google Scholar] [CrossRef] [PubMed]
  2. Mukherjee, A.; Keshary, V.; Pandya, K.; Dey, N.; Satapathy, S.C. Flying Ad hoc Networks: A Comprehensive Survey. In Information and Decision Sciences; Satapathy, S.C., Tavares, J.M.R., Bhateja, V., Mohanty, J.R., Eds.; Springer: Singapore, 2018; pp. 569–580. [Google Scholar]
  3. Horé, A.; Ziou, D. Image Quality Metrics: PSNR vs. SSIM. In Proceedings of the 2010 20th International Conference on Pattern Recognition, Istanbul, Turkey, 23–26 August 2010; pp. 2366–2369. [Google Scholar] [CrossRef]
  4. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Sahingoz, O.K. Networking Models in Flying Ad-Hoc Networks (FANETs): Concepts and Challenges. J. Intell. Robot. Syst. 2014, 74, 513–527. [Google Scholar] [CrossRef]
  6. Flammini, F.; Pragliola, C.; Smarra, G. Railway infrastructure monitoring by drones. In Proceedings of the 2016 International Conference on Electrical Systems for Aircraft, Railway, Ship Propulsion and Road Vehicles International Transportation Electrification Conference (ESARS-ITEC), Toulouse, France, 2–4 November 2016; pp. 1–6. [Google Scholar] [CrossRef]
  7. Sinha, J.P.; Kushwaha, H.; Kushwaha, D.; Singh, N.; Purushottam, M. Prospect of Unmanned Aerial Vehicle (UAV) Technology for Agricultural Production Management. In Proceedings of the International Conference on Emerging Technologies in Agricultural and Food Engineering 27—30th Agricultural and Food Engineering Department, IIT Kharagpur, Kharagpur, India, 27–30 December 2016. [Google Scholar]
  8. Kakamoukas, G.A.; Sarigiannidis, P.G.; Economides, A.A. FANETs in Agriculture—A routing protocol survey. Internet Things 2020, 100183. [Google Scholar] [CrossRef]
  9. da Cruz, E.P.F. A Comprehensive Survey in Towards to Future FANETs. IEEE Lat. Am. Trans. 2018, 16, 876–884. [Google Scholar] [CrossRef]
  10. Azevedo, M.I.B.; Coutinho, C.; Toda, E.M.; Carvalho, T.C.; Jailton, J. Wireless Communications Challenges to Flying Ad Hoc Networks (FANET). In Mobile Computing; Ortiz, J.H., Ed.; IntechOpen: Rijeka, Croatia, 2020; Chapter 1. [Google Scholar] [CrossRef] [Green Version]
  11. Me Biomo, J.M.; Kunz, T.; St-Hilaire, M. Directional antennas in FANETs: A performance analysis of routing protocols. In Proceedings of the 2017 International Conference on Selected Topics in Mobile and Wireless Networking (MoWNeT), Avignon, France, 17–19 May 2017; pp. 1–8. [Google Scholar] [CrossRef]
  12. Noor, F.; Khan, M.A.; Al-Zahrani, A.; Ullah, I.; Al-Dhlan, K.A. A Review on Communications Perspective of Flying Ad-Hoc Networks: Key Enabling Wireless Technologies, Applications, Challenges and Open Research Topics. Drones 2020, 4, 65. [Google Scholar] [CrossRef]
  13. Rametta, C.; Schembra, G. Designing a Softwarized Network Deployed on a Fleet of Drones for Rural Zone Monitoring. Future Internet 2017, 9, 8. [Google Scholar] [CrossRef] [Green Version]
  14. Zhang, M.; Li, X. Drone-Enabled Internet-of-Things Relay for Environmental Monitoring in Remote Areas Without Public Networks. IEEE Internet Things J. 2020, 7, 7648–7662. [Google Scholar] [CrossRef]
  15. Singh, A.; Patil, D.; Omkar, S.N. Eye in the Sky: Real-Time Drone Surveillance System (DSS) for Violent Individuals Identification Using ScatterNet Hybrid Deep Learning Network. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Salt Lake City, UT, USA, 18–22 June 2018; pp. 1710–17108. [Google Scholar] [CrossRef] [Green Version]
  16. Wang, Z.; Li, M.; Khaleghi, A.M.; Xu, D.; Lobos, A.; Vo, C.; Lien, J.M.; Liu, J.; Son, Y.J. DDDAMS-based Crowd Control via UAVs and UGVs. Procedia Comput. Sci. 2013, 18, 2028–2035. [Google Scholar] [CrossRef] [Green Version]
  17. Almagbile, A. Estimation of crowd density from UAVs images based on corner detection procedures and clustering analysis. Geo Spat. Inf. Sci. 2019, 22, 23–34. [Google Scholar] [CrossRef]
  18. Al-Sheary, A.; Almagbile, A. Crowd Monitoring System Using Unmanned Aerial Vehicle (UAV). J. Civ. Eng. Archit. 2017, 11. [Google Scholar] [CrossRef] [Green Version]
  19. Motlagh, N.H.; Bagaa, M.; Taleb, T. UAV-Based IoT Platform: A Crowd Surveillance Use Case. IEEE Commun. Mag. 2017, 55, 128–134. [Google Scholar] [CrossRef] [Green Version]
  20. Choi, S.; Hussen, H.R.; Park, J.; Kim, J. Geolocation-Based Routing Protocol for Flying Ad Hoc Networks (FANETs). In Proceedings of the 2018 Tenth International Conference on Ubiquitous and Future Networks (ICUFN), Prague, Czech, 3–6 July 2018; pp. 50–52. [Google Scholar] [CrossRef]
  21. Chriki, A.; Touati, H.; Snoussi, H.; Kamoun, F. UAV-GCS Centralized Data-Oriented Communication Architecture for Crowd Surveillance Applications. In Proceedings of the 2019 15th International Wireless Communications Mobile Computing Conference (IWCMC), Tangier, Morocco, 24–28 June 2019; pp. 2064–2069. [Google Scholar] [CrossRef]
  22. Kikuchi, H.; Funahashi, K.; Muramatsu, S. Simple bit-plane coding for lossless image compression and extended functionalities. In Proceedings of the 2009 Picture Coding Symposium, Chicago, IL, USA, 6–8 May 2009; pp. 1–4. [Google Scholar]
Figure 1. Crowd Movement in Hajj (a) A Crowded Street During Hajj, (b) Mina Valley During Hajj, (c) A Control Room for Crowd Management.
Figure 1. Crowd Movement in Hajj (a) A Crowded Street During Hajj, (b) Mina Valley During Hajj, (c) A Control Room for Crowd Management.
Computers 10 00046 g001
Figure 2. PRoFFAN’s Steps to Transmit Information.
Figure 2. PRoFFAN’s Steps to Transmit Information.
Computers 10 00046 g002
Figure 3. Bit-Plane Encoding.
Figure 3. Bit-Plane Encoding.
Computers 10 00046 g003
Figure 4. PRoFFAN Packet Transmission Algorithm.
Figure 4. PRoFFAN Packet Transmission Algorithm.
Computers 10 00046 g004
Figure 5. Experiment Scenario.
Figure 5. Experiment Scenario.
Computers 10 00046 g005
Figure 6. Packet Delivery Ratio for PRoFFAN compared to Zigbee.
Figure 6. Packet Delivery Ratio for PRoFFAN compared to Zigbee.
Computers 10 00046 g006
Figure 7. Comparison of PRoFFAN and Zigbee based on PSNR/SSIM vs. Speed. (a) PSNR vs. Speed Comparison, (b) SSIM vs. Speed Comparison.
Figure 7. Comparison of PRoFFAN and Zigbee based on PSNR/SSIM vs. Speed. (a) PSNR vs. Speed Comparison, (b) SSIM vs. Speed Comparison.
Computers 10 00046 g007
Figure 8. Comparison of PRoFFAN and Zigbee based on PSNR/SSIM vs. Deadline for UAV at 10 km/h. (a) Deadline based PSNR at 10 km/h, (b) Deadline based SSIM at 10 km/h.
Figure 8. Comparison of PRoFFAN and Zigbee based on PSNR/SSIM vs. Deadline for UAV at 10 km/h. (a) Deadline based PSNR at 10 km/h, (b) Deadline based SSIM at 10 km/h.
Computers 10 00046 g008
Figure 9. Received images from UAV with speed 10 km/h.
Figure 9. Received images from UAV with speed 10 km/h.
Computers 10 00046 g009
Figure 10. Comparison of PRoFFAN and Zigbee based on PSNR/SSIM vs. Deadline for UAV at 30 km/h. (a) Deadline based PSNR at 30 km/h, (b) Deadline based SSIM at 30 km/h.
Figure 10. Comparison of PRoFFAN and Zigbee based on PSNR/SSIM vs. Deadline for UAV at 30 km/h. (a) Deadline based PSNR at 30 km/h, (b) Deadline based SSIM at 30 km/h.
Computers 10 00046 g010
Figure 11. Comparison of PRoFFAN and Zigbee based on PSNR/SSIM vs. Deadline for UAV at 60 km/h. (a) Deadline based PSNR at 60 km/h, (b) Deadline based SSIM at 60 km/h.
Figure 11. Comparison of PRoFFAN and Zigbee based on PSNR/SSIM vs. Deadline for UAV at 60 km/h. (a) Deadline based PSNR at 60 km/h, (b) Deadline based SSIM at 60 km/h.
Computers 10 00046 g011
Table 1. Simulation Parameters.
Table 1. Simulation Parameters.
ParameterValue
MAC ProtocolCSMA
Routing ProtocolRPL Lite
TransmissionIEEE 802.15.4 (Channel 26)
Band2400–2483.5 DSSS
BitrateUp to 250 kbps
ModulationO-QPSK
Number of UAVs17
UAV Speeds (m/s)1 to 17 (3.6 to 61.2 km/h)
Base Station Range Radius, R500 m
Base Station Coverage on Road1000 m
Travel Distance per node2000 m
Image Resolution512 × 512
Image Type8-bit grayscale
Total Pixels262,144
PRoFFAN Packet Payload Size100 bytes
Packets per image2622 packets
Number of Image Layers2
Bit Planes (High Priority)1–4 (4 MSBs)
Bit Planes (Low Priority)5–8 (4 LSBs)
Packet Rate32 packets per second
PRoFFAN Queue Length32
Number of Simulations30
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Felemban, E.; Sheikh, A.A.; Naseer, A. Improving Response Time for Crowd Management in Hajj. Computers 2021, 10, 46. https://0-doi-org.brum.beds.ac.uk/10.3390/computers10040046

AMA Style

Felemban E, Sheikh AA, Naseer A. Improving Response Time for Crowd Management in Hajj. Computers. 2021; 10(4):46. https://0-doi-org.brum.beds.ac.uk/10.3390/computers10040046

Chicago/Turabian Style

Felemban, Emad, Adil A. Sheikh, and Atif Naseer. 2021. "Improving Response Time for Crowd Management in Hajj" Computers 10, no. 4: 46. https://0-doi-org.brum.beds.ac.uk/10.3390/computers10040046

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop