Next Article in Journal
Special Issue “Security Threats and Countermeasures in Cyber-Physical Systems”
Previous Article in Journal
Architectures and Protocols for Wireless Sensor and Actuator Networks
Previous Article in Special Issue
Industry 4.0 Applications for Medical/Healthcare Services
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Study on Sensor System Latency in VR Motion Sickness

1
Faculty of Computer Science & Electrical Engineering, University of Rostock, Albert Einstein Str. 2, 18059 Rostock, Germany
2
School of Industrial Automation, Engineering Institute of Technology, Melbourne, VIC 3000, Australia
3
School of Electrical and Computer Engineering, Georgia Institute of Technology, Atlanta, GA 30332, USA
*
Author to whom correspondence should be addressed.
J. Sens. Actuator Netw. 2021, 10(3), 53; https://0-doi-org.brum.beds.ac.uk/10.3390/jsan10030053
Submission received: 10 April 2021 / Revised: 8 July 2021 / Accepted: 13 July 2021 / Published: 6 August 2021
(This article belongs to the Special Issue Recent Trends in Innovation for Industry 4.0 Sensor Networks)

Abstract

:
One of the most frequent technical factors affecting Virtual Reality (VR) performance and causing motion sickness is system latency. In this paper, we adopted predictive algorithms (i.e., Dead Reckoning, Kalman Filtering, and Deep Learning algorithms) to reduce the system latency. Cubic, quadratic, and linear functions are used to predict and curve fitting for the Dead Reckoning and Kalman Filtering algorithms. We propose a time series-based LSTM (long short-term memory), Bidirectional LSTM, and Convolutional LSTM to predict the head and body motion and reduce the motion to photon latency in VR devices. The error between the predicted data and the actual data is compared for statistical methods and deep learning techniques. The Kalman Filtering method is suitable for predicting since it is quicker to predict; however, the error is relatively high. However, the error property is good for the Dead Reckoning algorithm, even though the curve fitting is not satisfactory compared to Kalman Filtering. To overcome this poor performance, we adopted deep-learning-based LSTM for prediction. The LSTM showed improved performance when compared to the Dead Reckoning and Kalman Filtering algorithm. The simulation results suggest that the deep learning techniques outperformed the statistical methods in terms of error comparison. Overall, Convolutional LSTM outperformed the other deep learning techniques (much better than LSTM and Bidirectional LSTM) in terms of error.

1. Introduction

In an ever-changing world seeking regular technological advancement, Virtual Reality has been proven to play an imperative role in the further advancement of scientific progress in many sectors of our lives. Virtual Reality (VR) refers to a computer-simulated environment that produces a sensible and intuitive 3D environment that clients can understand through a head-mounted display (HMD). However, the assessment of VR has brought up a well-known issue of many users suffering from motion sickness (MS), which is also known as simulation sickness or visually induced motion sickness (VIMS). In addition, motion sickness can be referred to as a mal-adaptation syndrome when exposed to real and/or apparent motion [1]. The signs or symptoms of motion sickness include nausea, excessive salivation, cold sweating, and drowsiness [2]. MS is categorized by actual motion on land, sea, and air, [1]. The cause of motion sickness is described in multiple theories, i.e., the sensory conflict theory [3], evolutionary/poison theory, postural instability theory, eye movement theory, and rest-frame hypothesis theory [4]. These theories argue for different causal factors describing the effects of mal-adaptation while sensing real or perceived motion.
The most common causal theory of MS, the Sensory Conflict Theory, was first proposed in 1975 [1]. According to the Sensory Conflict Theory, the visual systems generate the sensory inputs to the brain. The vestibular systems are different sensations experienced at different times, which cause MS. The perception of visual change (such as time delays found in the Head-Mounted Displays (HMDs) or Virtual environments (VEs)) causes a situation where the observing visual images are different from the user learned and mapped during the development phase. These are possible sources of sensory conflict [4]. These effects of sensory conflict frequently result in MS. Additionally, the optical and temporal distortions are also considered as sources of MS. These distortions occur due to the change of the desired inputs of the visual and vestibular systems. In [4], the term “simulator sickness” (ss) method was also used to describe MS. Researchers measured simulator sickness estimates and reports about MS symptoms during simulator and VE exposure [5]. Some technical factors cause MS. For example, a large field of view (FOV) can increase MS.
In addition to the FOV, other factors include latency, tracking accuracy, calibration, tracking precision, lack of positional tracking, judder, refresh rate, display response time, and persistence, all of which can be considered reasons for MS. In addition, other individual factors cause MS. For instance, lateral movement, expectations, experience with VR, thinking about sickness, and mental models can cause MS. Further factors include: instruction for how to put on the HMD correctly, sense of balance, Read–Write (RW) experience with the task (the less, the better) [3].
Concurrent to VR, application of machine learning and deep learning methods have also become increasingly common in different sectors, including power grid security [6], energy management [7], weather impact analysis on critical infrastructures [8], game theory analysis [9,10], robotics [11], medical diagnosis [12,13], and agriculture advisory [14], etc. Deep learning methods for human body prediction have made significant progress.
In [15,16] different deep recurrent neural networks (RNN) and Long short-term memory (LSTM) models were proposed to predict the human future trajectories. However, their models attempted to learn general human movement from a large collection of images/videos. The LSTM model, which can learn generic human movement and predict future trajectories, was used in [15]. The authors proposed a Social-LSTM that predicted pedestrian trajectories more accurately than existing algorithms, which was applied on two publicly available datasets.
In [16], the authors proposed a single Gated Recurrent Unit (GRU) LSTM instead of usual multi-layer LSTM architectures for short-term motion prediction using human skeletal data to learn human kinematics. These models are complicated since they are designed to learn patterns from a set of skeletal data and predict up to 80 milliseconds ahead of time. However, head motion prediction in six degrees of freedom (6DoF) is more difficult than in 360-degree video (3DoF) since both the location and viewing direction can change at any time, and users have a much wider virtual space to explore [17,18].
Thus, the resulting precision of the predicted location is much lower than what is needed for pre-rendering in the VR scenarios. The most difficult aspect of this particular method is to meet the ultra-low latency requirements since streaming 360-degree video can consume quite a lot of bandwidth. Good VR user experiences essentially require ultra-low latency (less than 20 ms) [19].
Motivated by the aforementioned latency-related challenges, in this paper, we study VR devices to predict head and body motions to meet the ultra-low latency requirements [20]. As we know, latency issues disrupt the VR device performance when a user portraits the captured motion on display. One approach is to perform predictive pre-rendering using the edge device/computer and then to stream the expected view to HMD (as long as we can predict the users’ head and body movement/motion for the immediate future) [21].
This paper’s contributions include investigating the performance of the position prediction using statistical and deep-learning-based prediction techniques (i.e., Dead Reckoning, Kalman Filtering, deep-learning-based long short-term memory (LSTM), Bidirectional LSTM, and Convolutional LSTM). Initially, we review the theories related to motion sickness (MS) in Section 2. Then, the technical factors causing the MS in VR and Augmented Reality (AR) devices are discussed (see Section 2). In Section 3, we discuss the common predictive algorithms for predictive tracking.
Later, the most critical factor, called latency, is considered for the simulation. In order to reduce the system delay or latency, we investigate the performance of Dead Reckoning and Kalman Filtering at first. Following that, we deploy deep-learning-based approaches LSTM, Bidirectional LSTM, and Convolutional LSTM (see Section 4). Finally, we conclude this study by analyzing and comparing the error performances between these statistical learning and deep-learning-based methods in Section 5.

2. Technical Factors Related to the Motion Sickness in Virtual Environment

In this section, a few technical factors involved in motion sickness in a virtual environment are discussed, such as field of view, latency, sensors, rendering, and display

2.1. Motion Sickness (MS) Incidence, Symptoms, and Theories

The word “Cybersickness” does not refer to a disease or a pathological state, rather to a typical physiological response that comes from an abnormal stimulus [22]. Cybersickness incidence depends on the stimulus, for instance the frequency, duration, and the user measurement criteria. The sickness symptoms were previously grouped into three common categories of effects, including nausea, oculomotor, and disorientation [23].
Miller and Graybiel determined that 90 % to 96 % of participants would undergo stomach symptoms when the participants reached the maximum number of head movements during rotation protocol [1]. Static observers who are healthy might feel significant discomfort caused by motion stimuli and a moving visual field, which is impossible in individuals without vestibular function.
According to the Oxford Dictionary of Psychology, passive movement can cause a mismatch between the information related to the orientation and the actual movement, which are supplied via the visual and vestibular systems. As per the explanation of sensory conflict theory, it is this mismatch that induces feelings of nausea [24]. In 2005, “Johnson” explained that there was a higher chance of experiencing simulator sickness (SS) when the perceived motion was not correlated with the forces transmitted by the users’ vestibular system [1].
Otherwise, if the real-life movements agree with the visual perception, the risk of experiencing SS reduces [25]. According to the Evolutionary/Poison theory, if conflicting information is received from our senses, it implies that something is irregular with our perceptual and motor systems/frameworks. Human bodies have advanced features that help to ensure an individual by limiting unsettling physiological influences created by consumed poisons [3].
Once humans predict what is or is not stationary, they will combine the information received by the visual and inertial system to support their following perceptions or actions [25]. As per Rest Frame (RF) theory, the most effective way to solve SS is by helping people find or create clear and non-conflicting rest frames to reconstruct their spatial perception [1].

2.2. Field of View

In optical devices and sensors, field of view (FOV) describes the specific angle based on which the devices can catch up with electromagnetic radiation. Rather than a single focusing point, FOV allows for coverage of an area. For VR, it is better to consider a larger FOV to obtain an immersive, life-like experience. Similarly, wider FOV provides better sensor coverage or accessibility for other optical devices. Field of view is one of the critical factors contributing to cybersickness, which can be categorized into two issues [26].

2.3. Latency

Latency or movement to photon latency is one of the most significant and critical properties in VR and AR systems [27]. Latency describes the time length between the user performing motion and the display showing the perfect content for the capture motion. More precisely, the duration it obtains for the real-world event to be sensed, processed, and displayed to the user is the system latency. The approximate range of latency is tens to hundreds of milliseconds (ms) [28]. In VR systems, latency has been determined to confound pointing and object motion tasks, catching tasks, and ball bouncing tasks.
Latency is exceptionally perceptional for humans. In VR, the user only sees the virtual content, which is a perception of the misalignment that originates from the muscle motion and the vestibular system. The user is not able to sense up to 20 ms of lag in the system [28]. With that in mind, the commonly used system for overlaying the virtual content onto the real world is via the optical see-through (OST) in the augmented reality space [29].
Since there is no lag in the real world, the latency for AR systems such as these has become visually apparent, as seen from the misregistration between the virtual content and real-world [28]. It is difficult to obtain minimum latency, as there is very little research evidence about these topics. “Moss and Muth” differentiated latency from other system variables. According to them, there was no increase in motion sickness symptoms when the amount of extra HMD system latency was varied [27]. Obtaining low latency is far more complex.

2.4. Sensors

Most mobile AR and VR devices combine cameras and inertial measurement units (IMUs) for their use for motion estimation. For tracking the position and orientation and the low end-to-end latency, IMU plays an important role. Primarily AR and VR systems run the frequency of tracking cameras at 30 Hz. This suggests that only 33 ms is needed for the image to read out and then passed on to be processed. They are assuming that exposure is settled at 20 ms. Thus, the first step is to assign a timestamp for each sensor measurement so that the processing can happen later on [28]. The cameras are operating at 30–60 Hz, so IMUs also run at a higher rate (from 100 Hz to even 1000 Hz) [28].

2.5. Tracking

In tracking the HMD track, the user’s head’s movement updates the rendered scene according to the orientation and location. Three rotational movements: pitch, yaw, and roll, are tracked by rotational tracking. Rotational tracking is performed by IMUs (such as accelerometers, gyroscopes, and magnetometers). In addition, there are three translational movements: forward/back, up/down, and left/right are tracked by positional tracking known as six degrees of freedom (6DOF) [30].
Usually, positional tracking is much more complicated than rotational tracking [31]. Depending on how the AR or VR device has moved and rotated, the tracking absorbs the sensor data to calculate a 6DOF motion estimation (i.e., the pose). Here, 6DOF means that the body is free to move forward or backward, up or down, and left to right.

2.6. Rendering

A 2D image is generated at the rendering stage. Then, the frame buffer is sent to the display. The renderer requires several inputs to generate that image, such as the 3D content. Due to this, it is usual for rendering to happen asynchronously from the point of tracking. When generating a new frame, it is assumed that the latest camera pose estimated by the tracker is used [28].

2.7. Display

The display’s process is critical for AR and VR since it contributes to a significant amount of additional latency, which is also highly visible. During the buffer phase, the data is to the display in pixel-by-pixel and line-by-line mode [30]. Each pixel of the display system stores the three primary colors: red, green, and blue. Thus, the data of the frames look like RGBRGBRGBRGB... [28].
That means the frame is arranged as scanlines so that each scanline has a pixel, and each pixel has RGB color components. The organic light-emitting diode (OLED) and liquid crystal display (LCD) do not usually store received pixel data in an intermediate/temporary buffer but do some extra processing, such as re-scaling [28]. In [32], predictive displays found success in overcoming the AR device system latency issue by providing the user with immediate visual feedback.

3. Common Prediction Algorithm for Predictive Tracking

A particular device latency tester can measure the motion-to-photon latency in a VR device [28]. The desire to perform predictive tracking comes from the latency, as mentioned earlier. The latency increases with the growing delay. By utilizing this predictive tracking, it is possible to reduce the latency [33]. Latency can come from multiple sources, such as sensing delays, processing delays, transmission delays, data smoothing, rendering delays, and frame rate delays [31]. All of the AR/VR devices have minimal delays. To counter this, “predictive tracking” alongside different methods (e.g., time-warping) helps to reduce the apparent latency [31].
One of the most prominent uses of predictive tracking is by decreasing the evident “motion-to-photon” latency, which indicates the time between movement and the instance the movement is portrayed/visualized on the actual display [28]. Although there is a delay between the movements and the instant the movement information is presented on display, the perceived latency can be reduced by an estimated future orientation and the position as information used in refreshing the display.
The reason for choosing predictive tracking in AR devices includes the viewer’s natural world’s rapid movement to look at against the augmented reality overlay [33]. A classic example would be when the user displays a graphical overlay over a physical object that the user watches with an AR headset. The overlay must stay “locked” to the object even when the user pivots his head. This is needed to ensure that it feels like part of the real world. Even though the object might be perceived with a camera, time is needed so that the camera can capture the frame.
A graphics chip renders the processor figures out the positioning of the object in the frame and then the overlay’s new area. The user can potentially decrease the overlay movement contrasted with the natural world by using predictive tracking [31]. For example, while doing head-tracking, the user can see how quickly the human head can turn, and the typical rotation speeds, which can help improve the tracking model [31]. Next are some regularly used prediction algorithms.

3.1. Statistical Methods of Prediction

In this section statistical methods of prediction, the Dead Reckoning algorithm, Alpha-beta-gamma algorithm, and Kalman Filtering are discussed.

3.1.1. Dead Reckoning

Suppose the position, as well as the velocity, are known for a given instance. In that case, the predicted new position presumes that the last known velocity and position are correct and that the velocity is continuous as before. With this algorithm, one crucial issue is that it requires the velocity to be constant regularly. However, as the velocity does not stay constant in most of the cases, that is why it makes the next set of suspicions wrong [31].
Dead Reckoning-based prediction is applied here for the prediction of the future value. From the Dead Reckoning, we applied the polynomial function for the prediction. The reason being that the data were polynomial. The basic equation of a cubic function or, more generally, a third-degree polynomial function is as follows:
y = a x 3 + b x 2 + c x + d
where y is the dependent variable and a , b , c , d are regression coefficients. For the prediction, the values for the four points from the dataset are taken. Then, the next point’s value is predicted using the polynomial cubic function. Next, the quadratic function and the linear function are used for the prediction. The equation of the quadratic and linear function can be expressed as follows:
y = a x 2 + b x + c
y = a x + b

3.1.2. Alpha-Beta-Gamma

The Alpha-beta-gamma (ABG) predictor has close relations to the Kalman predictor, even though it has less complicated mathematical analysis. ABG endeavors to continuously estimate the velocity and acceleration so that it can be used for prediction. Since the estimate considers the actual data, they reduce the noise [31]. This is done by configuring three parameters (e.g., alpha, beta, and gamma) that give the ability to underscore responsiveness rather than noise reduction [30].
A commonly used helpful technique to reduce apparent latency is predictive tracking. This offers sophisticated or straightforward implementations and requires some idea as well as investigation. However, for today’s VR and AR systems, predictive tracking is essential to achieve low latency tracking [33].

3.1.3. Kalman Filtering

As with Dead Reckoning, the Kalman Filtering algorithm estimates some unknown variables based on measurements taken over time. It is used to decrease the sensor noise for systems in which a mathematical model exists for the operation of the system [34]. It is an optimal estimator algorithm that combines data from a sensor and a motion model computationally. When the predicted values contain random error, uncertainty, or variation, it is a continuous cycle of predict–update [34].
The Kalman filter estimates the value more quickly than conventional standard prediction algorithms (e.g., Dead Reckoning and Alpha-beta-gamma predictor). Prediction with Kalman Filtering starts with the initial state estimation and in the estimated state, there is a certain amount of error. During the iterative process, the Kalman filter narrows down the predicted values somewhere close to the actual values very quickly. To predict efficiently and accurately, sufficient data is preferred. With enough data, the uncertainties are small, and the predicted value from the Kalman filter will be close to the actual value [34].
Let us consider X as the state matrix ( X i , X j ) , P as the process co-variance matrix, μ as the control variable matrix, W as the predicted state noise matrix, Q as the process noise covariance matrix, I as the identity matrix, K as the Kalman Gain, R as the noise covariance matrix (measurement error), and Y as the measurement of the state. A and B are the simple matrices multiplied with the state matrix to obtain the object’s current position. H is also a matrix that allows the format of one matrix to fall into the other matrix’s format.
We consider the initial state as X 0 and P 0 , the previous state as X k 1 , and the predicted new state as K p . Thus, the predicted new state X k p and P k p can be written as follows:
X k p = A X k 1 + B μ k + W k
P k p = A P k 1 A T + Q k
Next, the measured position of the object, which needs to be tracked, is converted into the proper format and returned as a vector. A measurement noise or a measurement uncertainty might need to be added since the measurement at face value might not always be taken. There may be some noise that can be defined in a matrix and add that to the measurement, which results in an updated measurement.
Then, the measurement is folded into the predicted state, and, from that, we come up with the Kalman gain K. The Kalman gain decides how much of the estimate needs to be done. Y k is the measured input and is used to calculate the noise in the measurement and variation in the estimate.
Y k = C X k m + Z K
Next, the Kalman gain is calculated, and the new state X k of the object that needs to be tracked is predicted.
K = P k p H H p k p H T + R
X k = X k p + K [ Y k H X k p ]
Next, the process error or the predicted error, P k in the Kalman gain is calculated as follows:
P k = ( I K H ) P k p
In this stage, the calculated current state becomes the previous state and goes through the iteration process for the next prediction. In the Kalman filter, we update the output, the new position X k , and the new predicted error P k .

3.2. Machine Learning-Based Approach

Machine learning has proven to be extremely powerful in the fields of digital image processing, enhancement, and super-resolution. Moreover, machine learning models have been commonly used in production systems for computational photography and video comprehension. However, integrating and implementing machine learning models on mobile graphics pipelines is difficult due to limited computational resources, tight power, and latency constraints in mobile Virtual Reality (VR) systems. For example, suppose we can predict the head and body motion of the user in the immediate/near future.
In that case, it is possible to perform the predictive pre-rendering on the edge device/computer, thus, streaming the expected view to HMD [35]. To construct our sequential learning model, we use the recurrent neural network (RNN) architecture. In this segment, we use an LSTM-based approach of a sequence-to-sequence predictive model that performs well in sequence-to-sequence prediction problems [36]. The predictive model creates a series from the user’s previous viewpoint positions and predicts future viewpoint positions in the same way [37].

3.2.1. Recurrent Neural Networks (RNN)

Recurrent Neural Networks (RNNs) are used to learn sequences in data. RNNs can map sequences into a single output or a series of outputs with ease. RNNs can handle sequences of varying lengths (e.g., VR data) [38]. The VR data is capable of performing the same operation as the time series data in the RNN model [39]. RNN adds a looping mechanism to a Feed Forward Neural Network (FFNN) that allows information to flow from one step to the next. The secret state (memory) representing the previous inputs carries the knowledge from the input data [40].
RNNs can be used for sequential knowledge modeling [41]. To generate output h 0 , the network takes input x 0 , which is fed along with the next input x 1 to generate the second output h 1 . Here, the h 1 output depends on the current x 1 input and its previous h 0 output. The structure of the RNN helps discover the dependencies of the input samples and recalls the meaning behind the sequences during training [16].

3.2.2. Long Short-Term Memory (LSTM)

Long short-term memory (LSTM) is able to manage long-term dependencies between samples of inputs. Unlike RNNs, it can selectively forget or recall data and add fresh information without fully changing the existing information [42]. An LSTM network consists of cells called memory blocks. Within every cell, there are four neural network layers. To go to the next cell, each cell moves two states: the cell state C t 1 and the hidden layer state h t 1 at the previous time step. x t represents the input at the current time step. Both cell state C and hidden layer state h (also the output state) change after passing through LSTM cells to form a new cell state C t and a new hidden layer state h t .
These cells are responsible for remembering essential information. Through gating mechanisms, this information can be manipulated. There are three gates to LSTM: the forget gate f t , input gate i t , and output gate o t . We calculate the forget gate f t , input gate i t , output gate o t , current cell memory s t , and current cell output h t to follow the LSTM structure from [41,43]:
f t = σ ( W f [ h t 1 , x t ] + b f ) , i t = σ ( W i [ h t 1 , x t ] + b i ) , o t = σ ( W o [ h t 1 , x t ] + b o ) , s t = f t s t 1 + i t ( t a n h ( W s [ h t 1 , x t ] + b s ) ) , h t = o t t a n h ( s t ) ,
Here s t 1 is the previous cell memory, s t is the current cell memory, and h t is the current cell output. The activation functions are sigma ( σ ) and tanh. W and b are the weight vectors for the all gates.
The update mechanism of the cell state represents the core of LSTM. This means that cell state C t 1 used the forget gate to discard part of the data/information at the previous time step ( t 1 ) and obtain a new state C t by inserting part of the information through the input gate. The output gate controls and updates a new hidden layer state h t [15]. For motion prediction by using LSTM, the mean square error (MSE) is considered as the loss function:
L o s s = 1 | N t r a i n | y S t r a i n t = 1 L ( y t y ^ t ) 2 ,
where N t r a i n is the total time steps of all trajectories on training set S t r a i n , and L represents the total length for each of the corresponding trajectories. y t represents the predicted output, and y ^ t represents the actual output.

3.2.3. Bidirectional Long Short-Term Memory (BLSTM)

Unidirectional LSTM preserves information from the past only because it can access only the past input. A Bidirectional LSTM will run on the inputs in two ways, one from the past to the future and the other from the future to the past. What distinguishes the bidirectional technique from unidirectional is that, in the LSTM that runs backwards, the system/user saves information from the future. By combining the two hidden states, the system/user may save information from both the past and the future at any point in time.
A bidirectional recurrent neural network (BRNN) was first proposed by M Schuster [44]. In several areas, such as phoneme classification [45], speech recognition [46], and emotion classification [47], bidirectional networks outperform unidirectional networks [48]. While applying to the time-series data, it also passes information backward in time and passes in normal temporal sequences. The BRNN has two hidden layers, each of which is connected to the input and the output. The first layer has recurrent connections from previous time steps, while the second is flipped.
That is how these two layers are differentiated and transfer activation backward on the series [49]. After unfolding over time, a BRNN can be trained using normal back-propagation. However, motion is a temporal-dynamic process influenced by various factors, such as acceleration, velocity, and direction. To learn these types of dependencies within single/multiple windows, time- and context-sensitive neural networks are proposed here. Since LSTM or BLSTMs learn time-independent dependencies, they capture the relationship between measurements within a window and the rest of the measurements in the same window [49,50].

3.2.4. Convolutional LSTM

The Convolutional LSTM architecture combines Convolutional Neural Network (CNN) layers for feature extraction on input data with LSTM to support sequence prediction [51]. CNN’s are a type of feed-forward neural network that performs well in the image and natural language processing [52]. The Convolutional LSTMs were developed for visual time series prediction problems as well as the application of creating textual descriptions from image sequences (e.g., videos). Convolutional LSTM architecture is both spatially and temporally deep and has the flexibility to be applied to a variety of vision tasks involving sequential inputs and outputs.
The network’s learning efficiency increases by local perceptron and weight sharing, which eventually reduces the number of parameters [53]. The convolution layer and the pooling layer are the two key components of CNN [52]. LSTM is commonly used in time series as it expands based on the time sequences [54]. The Convolutional LSTM process of training and prediction is shown in Figure 1.
The main steps of a Convolutional LSTM are described as follows:
  • Input data: Input the data needed for Convolutional LSTM training.
  • Data standardization: To better train the neural network model, the MinMax scaler-based data standardization technique is used to normalize the data. This technique is to re-scale features with a distribution value between 0 and 1.
  • Convolutional LSTM layer calculation: A Convolutional LSTM defined by adding CNN layers on the front end followed by LSTM layers with a Dense layer on the output. Where the input data are subsequently transferred through the convolution layer and pooling layer in the CNN layer, the feature extraction of the input data is carried out, and the output value is obtained. Finally, the CNN layer output data are calculated, and an output value is obtained through the LSTM layer.
  • Error calculation: To determine the corresponding error, the output value calculated by the output layer is compared to the real value of this group of data.
  • Error back-propagation: Proceed to step 3 to continue training the network by propagating the estimated error in the opposite way, updating the weight and bias of each layer and so on.
  • Train and test the model and make a prediction. Save the trained model and make a prediction of that model with the testing data.
  • Data standardization (inverse transform): The output value obtained through the Convolutional LSTM is the standardized value, and the standardized value is restored to the original value by inverse transform.
  • Output result of the predicted values: Output the restored results to complete the prediction process.

3.2.5. Evaluation Metrics for LSTM, Bidirectional LSTM, Convolutional LSTM

To evaluate the performance of the model for prediction we utilized RMSE and MAE. RMSE and MAE can be formulated as follows:
  • Root Mean Square Error (RMSE):
    R M S E = 1 | N t e s t | y S t e s t t = 1 L ( y t y ^ t ) 2 ,
  • Mean Absolute Error (MAE):
    M A E = 1 | N t e s t | y S t e s t t = 1 L ( y t y ^ t ) ,
Here, y t , y ^ t , and N t e s t represent the original value at period t, the predicted value at period t, and the number of total time steps of all trajectories on the test set S t e s t , respectively. The predicted value is compared to the observed value using these measures. The smaller the value of these metrics, the better the prediction efficiency.

4. Experimental Analysis

In this section, the experimental setup is explained and the simulation results are presented and discussed. The experimental setup explains the building blocks of this study, and the results analysis describes the results from the statistical and machine-learning-based methods and compares them.

4.1. Experimental Setup

Figure 2 The experimental setup of the proposed study started with the data generation/collection. Two types of data were collected from the VR devices, i.e., IMU data and camera sensor data. The IMU data was collected in high frequency, and the camera sensor data was collected in low frequency. Since the camera sensor data lead to the latency problem, these were considered to illustrate the effectiveness of the proposed method in the experiment.
The dataset contains the time (ns) and four Quaternions ( w , x , y , z ) of the sensor. The camera data was in the frequency range of 20 Hz, and the sensor data had a frequency of 200 Hz. After collecting the data, both the statistical and machine learning-based approaches were applied for prediction. After the prediction, the error properties of both approaches were measured. Next, the deployment scenario and the considered baseline schemes are described. Finally, the performance evaluation of the proposed approach was evaluated, and some insights from the results are discussed in Section 4.2.

4.2. Result Analysis

The experimental results from the statistical and machine learning-based approaches are discussed. The results from the statistical methods includes Dead Reckoning and Kalman Filtering. The results from the ML-based approaches include LSTM, Bidirectional LSTM, and Convolutional LSTM.

4.2.1. Experimental Results for Statistical Methods-Based Prediction

Cross-correlation: Cross-correlation measures the similarity of one signal and the time-delayed version of another signal as a function of a time-lag applied to them [55], which can be formulated as follows:
R ( τ ) = + x ( t ) y ( t + τ ) d t
Here, x ( t ) and y ( t ) are the functions of time and the time delay. The time delay can be negative, zero or positive. Cross-correlation reaches its maximum when the two signals considered become the most similar to each other. Since the frequency of the camera sensor data is higher than for the IMU sensor data, the data is re-sampled followed by the cross-correlation. First, the cross-correlation of the two data was computed and identified where the cross-correlation was the maximum. The highest lag was found in 0, which indicates that the camera sensor data and IMU sensor data were mostly the same at that point.
Dead Reckoning algorithm: In the Dead Reckoning algorithm, the cubic, quadratic and linear functions are considered to predict the next value and fit the curve. The algorithm first takes the first four data points from the dataset to predict the next value at the corresponding next position using the cubic, quadratic, and linear functions. The algorithm predicts the next value after every four points.
From Figure 3a, we can see the predicted subsequent data at position 5 for the cubic, quadratic, and linear functions. The quadratic was the best fitted compared to the other functions. The predicted following linear function data were close to the actual data, and the predicted following quadratic function data was also close to the actual data. However, the predicted following cubic function data was not too close to the actual data compared to the quadratic function’s predicted data points.
Next, the first 7 data points from the camera dataset were considered to observe the performance. We can see that the algorithm predicted the next value and fitted the curve from Figure 3b. After every 4 points, the algorithm predicted the next value. From the figure, it can be concluded that quadratic and cubic functions were suitable for curve fitting.
However, if all of the data from the camera dataset was considered, the quadratic function outperformed the cubic function for curve fitting. Furthermore, the linear was not the best fitted curve for the actual data. The linear function performed poorly while fitting the curve for the actual data compared to the other two functions. Figure 4 represents the predicted values of the whole dataset for all the three functions mentioned earlier.
Next, the total length for the camera dataset for the cubic, quadratic, and linear functions is compared. Figure 5a shows the actual value far away from the predicted values by quadratic and cubic functions. From the figure, we can see that the quadratic function outperformed the other two functions while predicting the data.
When the previous data with respect to the actual data and the predicted data is compared, how much the actual data needs to be refined can be analyzed. In this case, it can be seen that the quadratic property/function was the preferred property/function. At the same time, the cubic property also performed close to the quadratic property compared to the linear function.
The differences of the predicted cubic, quadratic, and linear function data from the original data are plotted in Figure 5b.
Kalman filter: The Kalman filter is applied to the camera dataset to predict the future value. Figure 6a shows that the algorithm predicted the value close to the actual value and smoothly fitted it. We compare the cubic, quadratic, and linear function methods with the Kalman Filtering technique where the Kalman filter quickly predicted the next position, and the predicted value was close to the actual value.
In Figure 6b, the prediction was made for the whole dataset, and the curve was fitted correctly compared to the cubic, quadratic, and linear function methods. Thus, from the results above, it can be concluded that the Kalman filter’s prediction was more accurate for the future position prediction.

4.2.2. Experimental Results for Machine Learning-Based Prediction

To prove the effectiveness of LSTM, Bidirectional LSTM, and Convolutional LSTM, the same dataset and operating environment were used. The system’s configuration used to conduct the experiments was an Intel i5-4700H 2.6 GHz, 16 GBs of RAM, 500 GB hard disk, and Windows 10 operating system. The dataset was split into 70 % and 30 % for training and testing. The training was executed for 50 epochs. Two hidden dense layers in the LSTM, one dense layer in bidirectional, and one hidden dense layer Convolutional LSTM were used. The optimizer was Adam, and the loss function of the root mean square error (RMSE) was used to train the model.
Prediction using LSTM, Bidirectional LSTM, and Convolutional LSTM:Figure 7, Figure 8 and Figure 9 illustrates the prediction results for LSTM, Bidirectional LSTM, and Convolutional LSTM. These methods are based on the loss function RMSE.
We found that LSTM, Bidirectional LSTM, and Convolutional LSTM-based prediction outperformed the Kalman Filtering and Dead Reckoning-based prediction. The details of the comparative analysis are provided in Section 4.2.3.

4.2.3. Comparison of Error

Table 1 shows the error comparison of the cubic, quadratic, and linear function methods with no prediction data (the difference in the original data from the current position to the previous position).
The no prediction data points were compared to the predicted cubic, quadratic, and linear data points. Table 1 indicates that the prediction using the quadratic function had a lower error than the no prediction error. However, for the standard deviation (SD), the quadratic and linear function had a similar error. The quadratic function had less errors than cubic and linear.
In Table 2, the predicted output of the Kalman filter with our actual data points is compared.
Table 2 shows that the actual data points had a higher rate of errors than the Kalman filter prediction errors. Thus, the Kalman filter is suitable for prediction, which is illustrated in Figure 5. Additionally, it had more minor errors than the actual value of data points. The predicted data points were close to the actual data points. The results from Table 2 indicate that it had significantly fewer errors than the actual data points.
In Table 3, the training and testing RMSE and MAE value from LSTM, Bidirectional LSTM, and Convolutional LSTM model are compared.
The accuracy of the VR sensor prediction presented in Table 3 demonstrates that the proposed Convolutional LSTM and Bidirectional LSTM models incurred the smallest RMSE and Convolutional LSTM incurred the smallest MAE in most of the sessions. The effectiveness of the proposed deep-learning-based techniques to predict VR motion positions is presented by comparing error properties. We found that LSTM performed as superior in every session of the motion prediction with a small RMSE and MAE. The results depict that, among the three approaches, CNN-LSTM outperformed the others. The CNN-LSTM had a MAE of 0.00 and RMSE of 0.00 , which were the smallest among the three prediction models, and it had high prediction accuracy.

5. Conclusions and Future Work

This paper investigated the latency causing MS in AR and VR systems. We adopted a common prediction algorithm to predict the future values to reduce the system delay resulting in reduced latency. The Dead Reckoning and Kalman Filtering techniques were studied to predict future data as a common prediction algorithm. We also adopted deep-learning-based methods with three types of LSTM models that could learn general head motion patterns of VR sensors and predict future viewing directions and locations based on previous traces.
On a real motion trace dataset with low MAE and RMSE, the system performed well. The error was compared with no prediction values. We found that the predicted values had a minor error when compared with no prediction. While using the Dead Reckoning algorithm, the quadratic function had less errors than the cubic and linear functions. However, the fitted curve was not right for the quadratic value.
The error property was relatively more minor in the Dead Reckoning algorithm than with the Kalman Filtering predicted error results. While deploying the deep-learning-based techniques, the Convolutional LSTM, and Bidirectional LSTM were more effective at learning temporal features from the data set. The prediction output can suffer if the number of time lags in historical data is insufficient.
However, the model still has certain shortcomings. For instance, we only considered a small amount of data for training and testing since the experiment only had a limited amount of real-world data from VR sensors. Potential future research would focus on applying our method with a large, real-world dataset from VR devices.
Since the proposed method showed minimal errors with a high prediction rate, based on the experimental results, we conclude that our prediction model can reduce significantly system latency by predicting future values, which could ultimately help to reduce MS in AR and VR environments.

Author Contributions

Conceptualization, R.K.K. and S.P.; Formal analysis, R.K.K., A.R. and S.P.; Investigation, R.K.K. and S.P.; Methodology, R.K.K. and S.P.; Resources, R.K.K. and S.P.; Supervision, R.K.K., A.R. and S.P.; Writing—original draft, R.K.K. and S.P.; Writing—review & editing, R.K.K., A.R. and S.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Lawson, B.D. Motion Sickness Symptomatology and Origins. 2014. Available online: https://www.taylorfrancis.com/chapters/mono/10.1201/b17360-33/motion-sickness-symptomatology-origins-kelly-hale-kay-stanney (accessed on 1 September 2020).
  2. Wood, C.D.; Kennedy, R.E.; Graybiel, A.; Trumbull, R.; Wherry, R.J. Clinical Effectiveness of Anti-Motion-Sickness Drugs: Computer Review of the Literature. JAMA 1966, 198, 1155–1158. [Google Scholar] [CrossRef]
  3. Azad Balabanian, P.L. Motion Sickness in VR: Adverse Health Problems in VR Part I. 2016. Available online: https://researchvr.podigee.io/5-researchvr-005 (accessed on 5 August 2020).
  4. Reason, J.T. Motion sickness adaptation: A neural mismatch model. J. R. Soc. Med. 1978, 71, 819–829. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Wiker, S.; Kennedy, R.; McCauley, M.; Pepper, R. Susceptibility to seasickness: Influence of hull design and steaming direction. Aviat. Space Environ. Med. 1979, 50, 1046–1051. [Google Scholar]
  6. Paul, S.; Ni, Z.; Ding, F. An Analysis of Post Attack Impacts and Effects of Learning Parameters on Vulnerability Assessment of Power Grid. In Proceedings of the 2020 IEEE Power Energy Society Innovative Smart Grid Technologies Conference (ISGT), Washington, DC, USA, 17–20 February 2020; pp. 1–5. [Google Scholar] [CrossRef]
  7. Sunny, M.R.; Kabir, M.A.; Naheen, I.T.; Ahad, M.T. Residential Energy Management: A Machine Learning Perspective. In Proceedings of the 2020 IEEE Green Technologies Conference(GreenTech), Oklahoma City, OK, USA, 1–3 April 2020; pp. 229–234. [Google Scholar] [CrossRef]
  8. Paul, S.; Ding, F.; Kumar, U.; Liu, W.; Ni, Z. Q-Learning-Based Impact Assessment of Propagating Extreme Weather on Distribution Grids. In Proceedings of the 2020 IEEE Power Energy Society General Meeting (PESGM), Montreal, QC, Canada, 3–6 August 2020; pp. 1–5. [Google Scholar] [CrossRef]
  9. Ni, Z.; Paul, S. A Multistage Game in Smart Grid Security: A Reinforcement Learning Solution. IEEE Trans. Neural Netw. Learn. Syst. 2019, 30, 2684–2695. [Google Scholar] [CrossRef]
  10. Paul, S.; Ni, Z.; Mu, C. A Learning-Based Solution for an Adversarial Repeated Game in Cyber–Physical Power Systems. IEEE Trans. Neural Netw. Learn. Syst. 2020, 31, 4512–4523. [Google Scholar] [CrossRef]
  11. Alsamhi, S.; Ma, O.; Ansari, S. Convergence of Machine Learning and Robotics Communication in Collaborative Assembly: Mobility, Connectivity and Future Perspectives. J. Intell. Robot. Syst. 2020, 98. [Google Scholar] [CrossRef]
  12. Maity, N.G.; Das, S. Machine learning for improved diagnosis and prognosis in healthcare. In Proceedings of the 2017 IEEE Aerospace Conference, Big Sky, MT, USA, 4–11 March 2017; pp. 1–9. [Google Scholar] [CrossRef]
  13. Ahsan, M.M.; Ahad, M.T.; Soma, F.A.; Paul, S.; Chowdhury, A.; Luna, S.A.; Yazdan, M.M.S.; Rahman, A.; Siddique, Z.; Huebner, P. Detecting SARS-CoV-2 From Chest X-Ray Using Artificial Intelligence. IEEE Access 2021, 9, 35501–35513. [Google Scholar] [CrossRef]
  14. Ünal, Z. Smart Farming Becomes Even Smarter With Deep Learning—A Bibliographical Analysis. IEEE Access 2020, 8, 105587–105609. [Google Scholar] [CrossRef]
  15. Alahi, A.; Goel, K.; Ramanathan, V.; Robicquet, A.; Fei-Fei, L.; Savarese, S. Social lstm: Human trajectory prediction in crowded spaces. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 961–971. [Google Scholar]
  16. Martinez, J.; Black, M.J.; Romero, J. On human motion prediction using recurrent neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 2891–2900. [Google Scholar]
  17. Hou, X.; Dey, S.; Zhang, J.; Budagavi, M. Predictive view generation to enable mobile 360-degree and VR experiences. In Proceedings of the 2018 Morning Workshop on Virtual Reality and Augmented Reality Network, Budapest, Hungary, 24 August 2018; pp. 20–26. [Google Scholar]
  18. Fan, C.L.; Lee, J.; Lo, W.C.; Huang, C.Y.; Chen, K.T.; Hsu, C.H. Fixation prediction for 360 video streaming in head-mounted Virtual Reality. In Proceedings of the 27th Workshop on Network and Operating Systems Support for Digital Audio and Video, Taipei, Taiwan, 20–23 June 2017; pp. 67–72. [Google Scholar]
  19. Hou, X.; Lu, Y.; Dey, S. Wireless VR/AR with edge/cloud computing. In Proceedings of the 2017 26th International Conference on Computer Communication and Networks (ICCCN), Vancouver, BC, Canada, 31 July–3 August 2017; pp. 1–8. [Google Scholar]
  20. Perfecto, C.; Elbamby, M.S.; Ser, J.D.; Bennis, M. Taming the Latency in Multi-User VR 360°: A QoE-Aware Deep Learning-Aided Multicast Framework. IEEE Trans. Commun. 2020, 68, 2491–2508. [Google Scholar] [CrossRef] [Green Version]
  21. Elbamby, M.S.; Perfecto, C.; Bennis, M.; Doppler, K. Toward low-latency and ultra-reliable Virtual Reality. IEEE Netw. 2018, 32, 78–84. [Google Scholar] [CrossRef] [Green Version]
  22. Tiiro, A. Effect of Visual Realism on Cybersickness in Virtual Reality. Univ. Oulu 2018, 350. Available online: http://urn.fi/URN:NBN:fi:oulu-201802091218 (accessed on 5 August 2020).
  23. Kennedy, R.S.; Lane, N.E.; Berbaum, K.S.; Lilienthal, M.G. Simulator sickness questionnaire: An enhanced method for quantifying simulator sickness. Int. J. Aviat. Psychol. 1993, 3, 203–220. [Google Scholar] [CrossRef]
  24. Sensory Conflict Theory. Available online: https://0-www-oxfordreference-com.brum.beds.ac.uk/view/10.1093/oi/authority.20110803100454911 (accessed on 24 November 2020).
  25. Lu, D. Virtual Reality Sickness during Immersion: An Investigation Ofpotential Obstacles towards General Accessibility of VR Technology. 2016. Available online: https://www.diva-portal.org/smash/get/diva2:1129675/FULLTEXT01.pdf (accessed on 12 January 2021).
  26. Rouse, M. Field of View (FOV). 2017. Available online: https://whatis.techtarget.com/definition/field-of-view-FOV (accessed on 5 August 2020).
  27. Wilson, M.L. The Effect of Varying Latency in a Head-Mounted Display on Task Performance and Motion Sickness. 2016. Available online: https://tigerprints.clemson.edu/cgi/viewcontent.cgi?article=2689&context=all_dissertations (accessed on 20 September 2020).
  28. Wagner, D. Motion to Photon Latency in Mobile AR and VR. 2018. Available online: https://medium.com/@DAQRI/motion-to-photon-latency-in-mobile-ar-and-vr-99f82c480926 (accessed on 5 August 2020).
  29. Hunt, C.L.; Sharma, A.; Osborn, L.E.; Kaliki, R.R.; Thakor, N.V. Predictive trajectory estimation during rehabilitative tasks in augmented reality using inertial sensors. In Proceedings of the 2018 IEEE Biomedical Circuits and Systems Conference (BioCAS), Cleveland, OH, USA, 17–19 October 2018; pp. 1–4. [Google Scholar]
  30. Zheng, F.; Whitted, T.; Lastra, A.; Lincoln, P.; State, A.; Maimone, A.; Fuchs, H. Minimizing latency for augmented reality displays: Frames considered harmful. In Proceedings of the 2014 IEEE International Symposium on Mixed and Augmented Reality (ISMAR), Munich, Germany, 10–12 September 2014; pp. 195–200. [Google Scholar]
  31. Boger, Y. Understanding Predictive Tracking and Why It’s Important for AR/VR Headsets. 2017. Available online: https://www.roadtovr.com/understanding-predictive-tracking-important-arvr-headsets/ (accessed on 7 August 2020).
  32. Richter, F.; Zhang, Y.; Zhi, Y.; Orosco, R.K.; Yip, M.C. Augmented Reality Predictive Displays to Help Mitigate the Effects of Delayed Telesurgery. In Proceedings of the 2019 International Conference on Robotics and Automation (ICRA), Montreal Convention Center, Montreal, QC, Canada, 20–24 May 2019; pp. 444–450. [Google Scholar]
  33. Azuma, R.T. Predictive Tracking for Augmented Reality. Ph.D. Thesis, University of North Carolina, Chapel Hill, NC, USA, 1995. [Google Scholar]
  34. Akatsuka, Y.; Bekey, G.A. Compensation for end to end delays in a VR system. In Proceedings of the IEEE 1998 Virtual Reality Annual International Symposium (Cat. No.98CB36180), Atlanta, GA, USA, 14–18 March 1998; pp. 156–159. [Google Scholar]
  35. Butepage, J.; Black, M.J.; Kragic, D.; Kjellstrom, H. Deep representation learning for human motion prediction and classification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 6158–6166. [Google Scholar]
  36. Greff, K.; Srivastava, R.K.; Koutník, J.; Steunebrink, B.R.; Schmidhuber, J. LSTM: A search space odyssey. IEEE Trans. Neural Netw. Learn. Syst. 2016, 28, 2222–2232. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  37. Liu, J.; Shahroudy, A.; Xu, D.; Wang, G. Spatio-temporal lstm with trust gates for 3d human action recognition. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 8–16 October 2016; pp. 816–833. [Google Scholar]
  38. Graves, A.; Mohamed, A.r.; Hinton, G. Speech recognition with deep recurrent neural networks. In Proceedings of the 2013 IEEE International Conference on Acoustics, Speech and Signal Processing, Vancouver, BC, Canada, 26–31 May 2013; pp. 6645–6649. [Google Scholar]
  39. Lipton, Z.C.; Berkowitz, J.; Elkan, C. A critical review of recurrent neural networks for sequence learning. arXiv 2015, arXiv:1506.00019. [Google Scholar]
  40. Jain, A.; Zamir, A.R.; Savarese, S.; Saxena, A. Structural-rnn: Deep learning on spatio-temporal graphs. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 5308–5317. [Google Scholar]
  41. Zaremba, W.; Sutskever, I.; Vinyals, O. Recurrent neural network regularization. arXiv 2014, arXiv:1409.2329. [Google Scholar]
  42. Duan, Y.; Yisheng, L.; Wang, F.Y. Travel time prediction with LSTM neural network. In Proceedings of the 2016 IEEE 19th international conference on intelligent transportation systems (ITSC), Rio de Janeiro, Brazil, 1–4 November 2016; pp. 1053–1058. [Google Scholar]
  43. Cho, K.; Van Merriënboer, B.; Gulcehre, C.; Bahdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning phrase representations using RNN encoder-decoder for statistical machine translation. arXiv 2014, arXiv:1406.1078. [Google Scholar]
  44. Schuster, M.; Paliwal, K.K. Bidirectional recurrent neural networks. IEEE Trans. Signal Process. 1997, 45, 2673–2681. [Google Scholar] [CrossRef] [Green Version]
  45. Kim, Y.; Sa, J.; Chung, Y.; Park, D.; Lee, S. Resource-efficient pet dog sound events classification using LSTM-FCN based on time-series data. Sensors 2018, 18, 4019. [Google Scholar] [CrossRef] [Green Version]
  46. Hashida, S.; Tamura, K. Multi-channel mhlf: Lstm-fcn using macd-histogram with multi-channel input for time series classification. In Proceedings of the 2019 IEEE 11th International Workshop on Computational Intelligence and Applications (IWCIA), Hiroshima, Japan, 9–10 November 2019; pp. 67–72. [Google Scholar]
  47. Zhou, Q.; Wu, H. NLP at IEST 2018: BiLSTM-attention and LSTM-attention via soft voting in emotion classification. In Proceedings of the 9th Workshop on Computational Approaches to Subjectivity, Sentiment and Social Media Analysis, Brusssels, Belgium, 31 October–1 November 2018; pp. 189–194. [Google Scholar]
  48. Graves, A.; Schmidhuber, J. Framewise phoneme classification with bidirectional LSTM and other neural network architectures. Neural Netw. 2005, 18, 602–610. [Google Scholar] [CrossRef]
  49. Zhao, Y.; Yang, R.; Chevalier, G.; Shah, R.C.; Romijnders, R. Applying deep bidirectional LSTM and mixture density network for basketball trajectory prediction. Optik 2018, 158, 266–272. [Google Scholar] [CrossRef] [Green Version]
  50. Chen, T.; Xu, R.; He, Y.; Wang, X. Improving sentiment analysis via sentence type classification using BiLSTM-CRF and CNN. Expert Syst. Appl. 2017, 72, 221–230. [Google Scholar] [CrossRef] [Green Version]
  51. Qin, L.; Yu, N.; Zhao, D. Applying the convolutional neural network deep learning technology to behavioural recognition in intelligent video. Tehnički Vjesn. 2018, 25, 528–535. [Google Scholar]
  52. Hu, Y. Stock market timing model based on convolutional neural network—A case study of Shanghai composite index. Financ. Econ. 2018, 4, 71–74. [Google Scholar]
  53. Rachinger, C.; Huber, J.B.; Müller, R.R. Comparison of convolutional and block codes for low structural delay. IEEE Trans. Commun. 2015, 63, 4629–4638. [Google Scholar] [CrossRef]
  54. Zhao, Z.; Chen, W.; Wu, X.; Chen, P.C.; Liu, J. LSTM network: A deep learning approach for short-term traffic forecast. IET Intell. Transp. Syst. 2017, 11, 68–75. [Google Scholar] [CrossRef] [Green Version]
  55. Zhang, L.; Wu, X. On cross correlation based-discrete time delay estimation. In Proceedings of the Proceedings, (ICASSP ’05), IEEE International Conference on Acoustics, Speech, and Signal Processing, Philadelphia, PA, USA, 23–25 March 2005; pp. 981–984. [Google Scholar]
Figure 1. Activity diagram of a CNN-LSTM-based prediction method.
Figure 1. Activity diagram of a CNN-LSTM-based prediction method.
Jsan 10 00053 g001
Figure 2. Experimental procedure for the prediction and error calculation.
Figure 2. Experimental procedure for the prediction and error calculation.
Jsan 10 00053 g002
Figure 3. The fitted cubic, quadratic, and linear curves and the predicted next data from the (a) first four data points and (b) first seven camera data.
Figure 3. The fitted cubic, quadratic, and linear curves and the predicted next data from the (a) first four data points and (b) first seven camera data.
Jsan 10 00053 g003
Figure 4. The fitted cubic, quadratic, and linear curves and the predicted next data for all camera data.
Figure 4. The fitted cubic, quadratic, and linear curves and the predicted next data for all camera data.
Jsan 10 00053 g004
Figure 5. (a) Comparison of the total length of the cubic, quadratic, and linear and actual/no prediction length and (b) The differences of the cubic, quadratic, and linear function data from the original data.
Figure 5. (a) Comparison of the total length of the cubic, quadratic, and linear and actual/no prediction length and (b) The differences of the cubic, quadratic, and linear function data from the original data.
Jsan 10 00053 g005
Figure 6. (a) The Kalman filter output for the 30 data points from the data set and (b) the Kalman filter output for the all data points from the data set.
Figure 6. (a) The Kalman filter output for the 30 data points from the data set and (b) the Kalman filter output for the all data points from the data set.
Jsan 10 00053 g006
Figure 7. LSTM prediction output for the (a) 30 data points and (b) all data points from the data set.
Figure 7. LSTM prediction output for the (a) 30 data points and (b) all data points from the data set.
Jsan 10 00053 g007
Figure 8. Bidirectional LSTM prediction output for the (a) 30 data points and (b) all data points from the data set.
Figure 8. Bidirectional LSTM prediction output for the (a) 30 data points and (b) all data points from the data set.
Jsan 10 00053 g008
Figure 9. Convolutional LSTM prediction output for the (a) 30 data points and (b) all data points from the data set.
Figure 9. Convolutional LSTM prediction output for the (a) 30 data points and (b) all data points from the data set.
Jsan 10 00053 g009
Table 1. Error comparison of cubic, quadratic, and linear function methods with no prediction data.
Table 1. Error comparison of cubic, quadratic, and linear function methods with no prediction data.
Predicted
Function Name
Maximum
Error
Minimum
Error
Average
Error
Standard
Deviation Error
Cubic function 0.0096 3.1696 × 10 7 0.0016 0.0018
Quadratic function 0.0051 5.4207 × 10 7 9.8735 × 10 4 0.0011
Linear function 0.0080 5.8936 × 10 7 0.0010 0.0011
No prediction 0.0132 5.8524 × 10 7 0.0032 0.0032
Table 2. Error comparison of Kalman filter predicted data and the actual data.
Table 2. Error comparison of Kalman filter predicted data and the actual data.
Predicted
Function Name
Maximum
Error
Minimum
Error
Average
Error
Standard
Deviation Error
Kalman filter0.0193 2.09 × 10 6 0.00490.0050
Actual data0.99990.84360.96860.0331
Table 3. Error comparison of LSTM, Bidirectional LSTM, and Convolutional LSTM-based prediction RMSE and MAE.
Table 3. Error comparison of LSTM, Bidirectional LSTM, and Convolutional LSTM-based prediction RMSE and MAE.
Predicted Deep
Learning Model Name
Train
RMSE
Train
MAE
Test
RMSE
Test
MAE
LSTM0.020.010.010.001
Bidirectional LSTM0.010.010.0110.001
Convolutional LSTM0.010.000.0120.00
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Kundu, R.K.; Rahman, A.; Paul, S. A Study on Sensor System Latency in VR Motion Sickness. J. Sens. Actuator Netw. 2021, 10, 53. https://0-doi-org.brum.beds.ac.uk/10.3390/jsan10030053

AMA Style

Kundu RK, Rahman A, Paul S. A Study on Sensor System Latency in VR Motion Sickness. Journal of Sensor and Actuator Networks. 2021; 10(3):53. https://0-doi-org.brum.beds.ac.uk/10.3390/jsan10030053

Chicago/Turabian Style

Kundu, Ripan Kumar, Akhlaqur Rahman, and Shuva Paul. 2021. "A Study on Sensor System Latency in VR Motion Sickness" Journal of Sensor and Actuator Networks 10, no. 3: 53. https://0-doi-org.brum.beds.ac.uk/10.3390/jsan10030053

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop