Next Article in Journal
Method to Solve Underwater Laser Weak Waves and Superimposed Waves
Previous Article in Journal
Determination of Optimal Predictors and Sampling Frequency to Develop Nutrient Soft Sensors Using Random Forest
Previous Article in Special Issue
Usability Evaluation of the SmartWheeler through Qualitative and Quantitative Studies
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Efficient Self-Attention Model for Speech Recognition-Based Assistive Robots Control

by
Samuel Poirier
1,2,*,
Ulysse Côté-Allard
3,
François Routhier
1,2 and
Alexandre Campeau-Lecours
1,2
1
Université Laval, Quebec City, QC G1V 0A6, Canada
2
Centre for Interdisciplinary Research in Rehabilitation and Social Integration, CIUSSS de la Capitale-Nationale, Quebec City, QC G1M 2S8, Canada
3
Department of Technology Systems, University of Oslo, 0313 Oslo, Norway
*
Author to whom correspondence should be addressed.
Submission received: 5 May 2023 / Revised: 15 June 2023 / Accepted: 27 June 2023 / Published: 30 June 2023
(This article belongs to the Special Issue Integration of Advanced Sensors in Assistive Robotic Technology)

Abstract

:
Assistive robots are tools that people living with upper body disabilities can leverage to autonomously perform Activities of Daily Living (ADL). Unfortunately, conventional control methods still rely on low-dimensional, easy-to-implement interfaces such as joysticks that tend to be unintuitive and cumbersome to use. In contrast, vocal commands may represent a viable and intuitive alternative. This work represents an important step toward providing a viable vocal interface for people living with upper limb disabilities by proposing a novel lightweight vocal command recognition system. The proposed model leverages the MobileNet2 architecture, augmenting it with a novel approach to the self-attention mechanism, achieving a new state-of-the-art performance for Keyword Spotting (KWS) on the Google Speech Commands Dataset (GSCD). Moreover, this work presents a new dataset, referred to as the French Speech Commands Dataset (FSCD), comprising 4963 vocal command utterances. Using the GSCD as the source, we used Transfer Learning (TL) to adapt the model to this cross-language task. TL has been shown to significantly improve the model performance on the FSCD. The viability of the proposed approach is further demonstrated through real-life control of a robotic arm by four healthy participants using both the proposed vocal interface and a joystick.

1. Introduction

People living with upper limb disabilities face important challenges to accomplish their Activities of Daily Living (ADL). Over the past years, assistive devices have been developed to improve the autonomy of their users [1,2,3]. Robotics, in conjunction with artificial intelligence, has been seen a key solution to improve the quality of life of people living with physical disabilities. Notably, it has been shown that robotic devices can help the users regain a part of their autonomy and facilitate the accomplishment of their daily tasks [4,5].
One of the principal issues that arises with the control of robotic arms comes from the large number of Degrees of Freedom (DoFs) that the control interface allows the user to manage with a limited number of physical inputs. To lessen the difficulty for the users, the different movements are often regrouped in sub-groups. To perform a task, the user has to switch between these different sub-control groups to select the desired command, resulting in a lot of time and effort spent [6,7].
Interface intuitiveness however remains the Achilles’ heel within the field of assistive robotics. From [8], a definition of an intuitive control for human–machine interaction is “a technical system is, in the context of a certain task, intuitively usable while the particular user is able to interact effectively, not consciously using previous knowledge.” In recent years, multiple control interfaces have been proposed such as virtual joysticks [9,10]. However, these interfaces fail to provide an intuitive experience, increasing the difficulty and the cognitive load of the users [11,12]. Additionally, their use is still cumbersome and requires constant calibrations; this is why joysticks and sip-and-puff devices are still the predominantly used interfaces to this day [13].
In search of a simpler and more straightforward solution, speech command has been seen as one of the most promising control modes [14,15]. Nevertheless, speech recognition is a vast and complex field. As a matter of fact, the tone of the voice, speech rate and pronunciation of words are all characteristics that vary from one person to another. Historically, speech command interfaces have faced important performance issues due to the limited computational power and the type of algorithms used [16]. The advent of deep learning paired with high computing capabilities substantially improved the performance of speech control interfaces to the point of now being ubiquitous. Primary examples of said devices are smartphones, Internet of Things (IoT) and home automation [15,17]. These applications, however, use an internet connection to process the speech and produce word recognition. This requirement greatly reduces the applicability of speech recognition devices for assistive technologies as they should operate without any exterior dependency due to safety and autonomy concerns [18,19]. Also, the use of the internet for processing the data can pose some privacy issues for the users.
The general aim of this paper is to enable people living with upper limb disabilities to control assistive robotic arms intuitively in order to increase their autonomy. The objectives are (1) to develop a lightweight speaker independent Keyword Spotting (KWS) machine learning-based model for the control of assistive robotic arms with English words and compare it to state-of-the-art algorithms, (2) to leverage a Transfer Learning (TL) algorithm to enable robotic arm control with French voice commands (French being the native tongue of the target users) and to quantify the performance of the TL and (3) to assess the intuitiveness of the control interface for a robotic assistive arm.
This work’s main contribution is a novel and lightweight offline speaker-independent KWS model which, to the best of the authors knowledge, performs better than any known model (with fewer parameters). This increase in performance is achieved through the introduction of a novel segmented 2D self-attention module. Another contribution of this work is the introduction of a novel KWS dataset titled French Speech Commands Dataset (FSCD) comprising 4963 vocal command utterances from 33 participants (18 women and 15 men). Finally, another contribution of this work is to demonstrate the successful application of TL from the Google Speech Commands Dataset (GSCD) to the FSCD to achieves a vocal interface in French which allows four participants to control a robotic arm in real-time more intuitively than when using a joystick.
This paper is organized as follows. An overview of the related work in speech command recognition through machine learning algorithms and TL is given in Section 2. Section 3 presents two datasets: the Google Speech Commands Dataset (GSCD) and the FSCD. A presentation of the methods followed for the development of the proposed model as well as the audio features, the training setup and the comparison with the state-of-the-art KWS models is given in Section 4. Section 5 describes the implementation consideration of the algorithm for assistive robotic arms and the adaptation to the French vocal commands via TL. Section 6 presents the evaluation of the control interface for a robotic assistive arm with two experiments. Section 7 presents the results of the comparison between our model and the state-of-the-art KWS models and the result of the TL. Finally, the results are discussed in Section 8.

2. Related Work

Studies on voice-controlled interfaces using short commands for people living with disabilities have been carried out for over 30 years [16]. Different solutions leveraging the user-voice-based interaction with computers and robotic devices have been proposed. For instance, an interface designed for human–computer interaction used a vocal joystick based on discrete vocal commands paired with continuous vocal attributes (e.g., pitch and loudness) that are mapped to continuous control parameters [20]. Despite the improvements that this “vocal joystick” implemented compared to previous works, it still produces too many false positives. It is sensitive to noise and needs a specific training on the voice of the user to increase its performance [21]. This solution was then further developed to allow people with motor impairments to draw freely on a computer with only the use of their voice [22], but suffered from the same shortcomings as the vocal joystick. Other works extended the vocal joystick, such as [23], which used the non-verbal vocalization as a control mode for the Lynx 6 arm. In [24], a speaker-dependent algorithm based on Dynamic Time Warping (DTW) and K-Nearest Neighbor (KNN) grants the user the ability to control a custom-designed robot. Given the simplicity of this algorithm, its accuracy is highly dependent on the intonation and speech rate of the user as well as the surrounding noise. Furthermore, like all the previous algorithms, the user must record and train the system on his own voice to be able to use it (i.e., speaker-dependent). In [25], a robotic arm mounted on a wheelchair uses a vision system and a speech command interface where the user can give low-level commands (e.g., hand forward, gripper open) to control the end effector directly, and also high-level commands such as “pour in a drink”. This interface does not focus on the speech recognition algorithm and opts for the use of a now outdated generic speech recognition system, which brings performance limitations. A speech interface has been implemented for rehabilitation purposes as well in [26] where the users can control various movements of an exoskeleton robot with their voice. This interface has an efficient de-noising step based on wavelet for white noise, but it also faces some drawbacks, mainly due to the speaker-dependent system and the DTW algorithm, which requires words to be enunciated with the same intonation and speech rate as the training utterances. Other researchers [27,28] developed systems to control wheelchairs with voice commands. In each case, the speech recognition algorithm requires two models to be defined: the acoustic model and the language model. The use of two different models makes the training more complex and fragments the information that could be leveraged by a unified model to reach higher performance. Also, for these systems to reach a good performance, the users are required to update the acoustic model by producing plenty of dictation, correction and training data.
KWS is one of the most efficient methods used to implement a voice-based command interface that performs well with robotic arms [16]. KWS consists in the recognition of specific words in a context where multiple noises and words can overlap in an audio stream. The first methods [29,30,31] that allowed a high word recognition rate were based on the Hidden Markov Model (HMM). In this generative approach, a distinct HMM was trained for each word of the dictionary, and another HMM was added to handle the audio segments outside of the keywords dictionary. While performing inference, these HMMs use the Viterbi decoding algorithm, which make the real time predictions computationally expensive.
The paradigm then shifted from generative models to discriminative deep learning models, which yield higher performance and have lower computation cost. Several studies leveraged the highly performing deep learning models from the image classification field and transposed them to the speech recognition task. A highly accurate model used feed forward neural network [32] combined with a Gaussian mixture model (GMM), and largely surpassed the previously state-of-the-art HMMs. Another research [33] successfully demonstrated the high prediction accuracy and the possibility to maintain this accuracy while reducing the number of parameters of all the major baseline models (e.g., Deep Neural Network, Convolutional Neural Network, Long Short-Term Memory, Gated Recurrent Unit) of the deep learning universe. Recent research effort [34,35,36] have focused on the development of highly performing networks with an optimized amount of parameters.
The development of high performance KWS models led to the use of state-of-the-art models (e.g., Residual Neural Network) and the integration of attention mechanisms [37,38,39]. In [34], the application of residual layers to KWS outperformed previous state-of-the-art models and also allowed for a more compact architecture. Two of the models from this article are used for comparison: the res15, which is the most performant model and its lightweight counterpart res8-narrow. More recently, an architecture leveraging both residual connection and graph convolutional network has been proposed in [36]. From this article, the one with the best performance (the CENet-GCN-40) and the one with the smallest number of parameters (CENet-6) are also compared to our proposed model.
Since the publication of [40], self-attention mechanisms have been widely used in conjunction with every existing models to increase prediction accuracy and take advantage of its feed forward structure, which offers a low computational time. In the speech recognition sphere, large vocabulary models have leveraged self-attention for end-to-end training [41,42,43]. However, these models are not optimized for KWS tasks due to the high number of parameters that need to be trained to perform well. In [44], self-attention has been added to the extensively used Time Delay Neural Network (TDNN) and achieved the best performance for the least number of parameters found in the scientific literature. For the comparison process, the model with the best performance (tdnn-swsa) and the smallest one (swsa) were selected.
An important drawback of deep learning is its need for a large volume of data even when the model is designed to have a low number of trainable parameters. Since no large dataset is available for the required French voice commands, and creating a large dataset from scratch is expensive in terms of both time and money, TL is seen as a good alternative. TL is a useful method to enhance a model’s performance in a target task where data are scarce by leveraging a trained model in a data-rich related source domain. From the multiple ways to apply TL effectively for the speech processing [45], the performant TL method for deep neural networks described in [46,47] is used for the transfer from a large English speech command dataset to the smaller French dataset. With this technique, the KWS model is first trained on the source language (English). Then, the entire pre-trained model (or subset thereof) is retrained on the target language (French). This fine-tuning approach tends to yield better models than when simply using random initialization [46].

3. Speech Command Datasets

Within this work, two datasets are considered: the GSCD [48] and the novel FSCD which was created by the authors. The GSCD which is available at http://download.tensorflow.org/data/speech_commands_v0.01.tar.gz (accessed on 10 March 2020) served two purposes, namely: (1) to validate the proposed model and compare it to the current state-of-the-art models for the KWS task, (2) to pre-train the proposed model on prior to applying the TL algorithm toward the FSCD. Note that in this work, the first version of the GSCD is employed for comparison purposes with the different state-of-the-art models, and because it is the closest to FCSD. The GSDC consists of a vocabulary of 30 words for a total of 64,727 utterances (i.e., recorded examples of a voice command). The dataset is unbalanced in terms of number of utterances per class (e.g., the “stop” command has 2380 utterances whereas “bed” has 1713 utterances). Every utterance has a length of 1 s or less, recorded with a 16 kHz sample rate and is encoded as linear 16-bit single-channel PCM values. There is a total of 1881 different speakers, with a different number of utterances per speaker. Each speaker is identified via a unique identifier. In doing so, all utterances from a given speaker are contained within the same set (i.e., training, validation or test set). Ten of the 30 words are the actual commands that are used to control the assistive robotic arm in the experiments, the rest being filler words to help the model learn a wide range of phonemes. The words composing this dataset are presented in Table 1.
The FSCD is made readily available at the following link: https://arweave.net/rIRaVzZqR7aRa0U3gg2Hqxgpat0-wL1XFMZVUd0pq9M, accessed on 4 May 2023. This dataset consists of 4963 different vocal command samples from 33 participants (15 males, 18 females) aged between 20 and 45 years old. The FSCD is balanced: the standard deviation of the number of utterances per class being less than 4%. Every utterance has a length of 1 s, recorded with a 16 kHz sample rate and is encoded as linear 16-bit single-channel PCM values. The vocabulary of the dataset is made of 12 commands with a total of 29 words which are presented in Table 1.
A challenge of speech recognition is robustness to background noises and silence. Therefore, when training the model, several minute-long utterances of such signals were also used in the training of the model. From the given background, noise files of the GSCD, 3500 one-second-long silences and noise utterances were randomly sampled. These data were added to both datasets for the training.

4. Methods

4.1. Model Development

Inspired from the highly performant MobileNetV2 [49], the architecture of the proposed model (which can be seen in Figure 1) was designed to be compact and to ensure its efficiency on embedded devices. Five types of blocks compose the model: (1) the initial block, (2) the inverted residual block, (3) the transition block, (4) the Segmented 2D Self-Attention block and (5) the prediction block. The channel count and depth of the network are kept low to avoid overfitting and to reach our goal of a lightweight model. We proposed two models: the Segmented-SA-11k, which displays the best performance, and its smaller version, the Segmented-SA-7k.

4.1.1. Initial Block

The purpose of this block is to increase the channel dimension from 1 to 8 before a first inverted residual block is applied. The use of this block (as opposed to directly using an inverted residual block) increased the performance of the network at an average of 0.3% for the validation set from ten different trainings. The initial block is composed of a bias-free 2D convolution layer with a 3 × 3 kernel and a stride of 1 followed by a Batch Normalization layer and a ReLU6 used as an activation function.

4.1.2. Inverted Residual Block

The inverted residual block proposed in [49] is used as the main feature extractor of the network. This block allows an effective back-propagation of the gradient to the lower layers due to its residual connection between layers with a low number of channels (known as bottleneck layers). The network is composed of three inverted residual blocks. We added dilation with a rate of 2 since it was experimentally found to increase network performance without increasing model complexity.

4.1.3. Transition Block

The first two inverted residual blocks are followed by a transition block. This block’s function is twofold: (1) to halve the frequency and time dimensions, reducing slowly the dimensionality of the input and (2) to increase the channel’s dimension to capture a greater share of meaningful features. Max pooling followed by a convolutional layer could have been used in place of this block to perform the down-sample and channel increase, but it was found experimentally that a large portion of the information would then be lost. This block is formed from a bias-free 2D convolution layer with a 3 × 3 kernel and a stride of 2, a Batch Normalization layer and a ReLU6 used as an activation function.

4.1.4. Segmented 2D Self-Attention

Segmented 2D Self- Attention is a novel idea that extends the regular self-attention mechanism to the KWS task. The reasoning behind this approach is that non-adjacent phonemes do not have any relation with one another. Thus, the idea of the Segmented 2D Self-Attention is to restrict the self-attention to a 2D local space of the feature maps extracted by the network. In such a way, the network only has access to the local features around a phoneme and therefore, it must learn to extract important features directly in the vicinity of a phoneme.
The mechanism of the Segmented 2D Self-Attention is added after the third inverted residual block. This block was specifically placed at this position to leverage all the information extracted by the previous layers and to reduce the number of multiplications that the mechanism requires. A schematic representation of the Segmented 2D Self-Attention is presented in Figure 2. The projection of the input to the query Q , key K and value V spaces is performed with 2D convolution with a kernel size of 1 × 1. The 2D space is described by c different feature maps composed of the frequency f and the time t axes. The projection reduces the number of channels in a bottleneck fashion [50] to c n trainable filters, where n is fixed to obtain an integer multiple of the channel size. The three projections are obtained by the convolution ⊗ of the output tensor X R c × f × t from the preceding layer with the tensors { W Q , W K , W V } R c n × 1 × 1 resulting in { Q , K , V } R c n × f × t .
Q = W Q X ,
K = W K X ,
V = W V X .
By rearranging the tensors { Q , K , V } to R c n × f t , the softmax σ of the matrix multiplication of Q T and K gives a spatial self-attention map of size R f t × f t , and the matrix multiplication of Q and K T results in a channel self-attention map of size R c n × c n . The output of the global self-attention mechanism for the spatial self-attention is presented in Equation (4) whereas the output for the channel self-attention is given by Equation (5):
S = V ( σ ( Q T K f t ) ) ,
C = σ ( Q K T c n ) V .
As noted in [40], the matrix multiplication Q T K and Q K T tends to grow large in magnitude, which causes the gradient to vanish due to the softmax operation. To counter this effect, the matrix multiplications need to be scaled down by a factor of 1 f t and 1 c n respectively.
With the reasoning that global self-attention carried out on a complete utterance does not allow the model to fully capture the local and more subtle variations in the spectrogram, the feature space R c n × f × t is segmented in smaller local grids. For each element ( : , i , j ) of Q R c n × f × t , a local tensor Q ^ : , i , j is extracted with a size of w + 1 in the frequency axis and h + 1 in the time axis, all the channels being included. Each element q i , j of Q ^ : , i , j is a vector composed of the c n feature maps. Supposing that the width w and height h of the local map are odd numbers, each local tensor can be extracted with:
Q ^ : , i , j = { q i ( w 1 ) 2 , j ( h 1 ) 2 , , q i , j , , q i + ( w 1 ) 2 , j + ( h 1 ) 2 } ,
K ^ : , i , j = { k i ( w 1 ) 2 , j ( h 1 ) 2 , , k i , j , , k i + ( w 1 ) 2 , j + ( h 1 ) 2 } ,
V ^ : , i , j = { v i ( w 1 ) 2 , j ( h 1 ) 2 , , v i , j , , v i + ( w 1 ) 2 , j + ( h 1 ) 2 } .
The width w and height h are parameters that need to be explored as well as the overlap between each two consecutive maps. This overlap can be modified by adjusting the stride s f on the frequency axis and s t on the time axis. With s f , s t > 1 , a part of the spectrogram elements are skipped, reducing the number of extracted local tensors. The padding p is adapted so the whole spatial input map ( f × t ) is included in the segmentation process. The maximum number of extracted local tensors for the frequency and time axes is given by the following equations:
u = f + 2 p ( w 1 ) 1 s f + 1 ,
y = t + 2 p ( h 1 ) 1 s t + 1 .
There is a total of u y segmented grids of size R c n × w × h obtained from this operation for Q ^ , K ^ , V ^ where i { 0 , , u 1 } , j { 0 , , y 1 } .
After rearranging each local tensor Q ^ , K ^ , V ^ to R c n × w h , the process of 2D local self-attention is applied to each grid m extracted. Equation (4) and Equation (5) can then be rewritten for the spatial local self-attention to Equation (11) and and the channel local self-attention Equation (12) respectively:
S : , m = V ^ : , m ( σ ( Q ^ : , m T K ^ : , m w h ) ) ,
C : , m = σ ( Q ^ : , m K ^ : , m T c n ) V ^ : , m .
The local output tensors S : , m and C : , m are also of shape R c n × w h . Both local tensors are then summed:
G : , m = S : , m + C : , m .
The concatenation of all the local output tensors G : , m is of shape G R u y × c n × w h . By applying a summation over the overlapping section of all the extracted local tensors and then by reshaping it, we obtain the original shape before the segmentation occurred G R c n × f × t .
The final output is generated by a convolution with a kernel W O of shape R c × 1 × 1 , expanding the channel’s dimension to its initial shape while leaving the other dimensions intact and allowing the residual to be added:
O = W O G + X .

4.1.5. Prediction Block

The network ends with the prediction block. This block is composed of a global average pooling [51] to collapse the output of the Segmented 2D Self-Attention to a vector composed of the average of each channel. This vector is then fed to a fully connected layer with a number of neurons corresponding to the number of classes to predict. A softmax is used as the activation function to output the posteriors of each vocal command.

4.2. Audio Features

As stated in [52], the amount of data needed grows exponentially with the variables (i.e., dimensionality). Since a waveform signal sampled at 16 kHz contains 16,000 dimensions for a one-second utterance, this causes a large data sparsity that a dataset like the GSCD cannot satisfy. To lessen the impact of said sparsity, other audio features have been investigated. To this end, some of the most popular characteristics for speech recognition with a low number of input dimensions—the Mel-frequency cepstral coefficients (MFCCs)—have been considered, as well as the spectrogram.
The MFCCs are widely used throughout speech recognition algorithms. A time window of 20 ms with a stride of 10 ms was used to extract a total of 40 MFCCs. To reduce the edge effects caused by the STFT (also known as spectral leakage), we used a Hamming window function. For a signal of one second padded at its extremities, this pre-processing resulted in 101 contiguous frames of 40 dimensions each.
The second audio feature used was the spectrogram. Since it has been shown that MFCCs are not robust against noises [53], a reasonable choice was to let the neural network learn relevant characteristics from a signal with less data pre-processing. To obtain the spectrogram features, the Short-time Fourier transform (STFT) is first computed with a time window of 20 ms and a stride of 10 ms to which a Hamming window function was applied. These parameters were determined empirically. Then, only the magnitude is kept since overlapping windows allow the signal to be reconstructed from the magnitude alone, the phase only having a minor impact on sound quality. For a one-second signal padded at its extremities, the dimensions of the resulting spectrogram are 161 frequency bins by 101 time frames.

4.3. Training Setup

The GSCD has been used to assess and compare the performance of the proposed approach. The objective of the model is to discriminate between the ten different commands (see Table 1), all the other words of the dataset (20) being integrated in a “filling words” class. Silence utterances (i.e., no word being spoken) are grouped in a “silence” class, making a total of 12 classes. The output with the highest value is designated as the predicted class.
The training dataset comprises 1503 speakers (corresponding to 51,094 utterances), while the validation and test sets both comprise 189 speakers (6798 and 6835 utterances, respectively). The training and validation sets are used in the training phase whereas the test set is used only once at the end to validate the final performances of the trained models.
The model architecture was implemented, trained and tested using Pytorch [54]. Two types of models have been compared: models from the literature and models from the authors. For the models from the literature, they were implemented following the training procedure described in their respective articles. For the models developed by the authors, they have been trained with the following procedure.
The training was first performed by minimizing the categorical cross entropy. The Adam algorithm [55] is employed for the network optimization (lr = 3 × 10 3 ) with a batch size of 32. To reduce overfitting, a weight decay of 1 × 10 2 was added, and early stopping (patience of 20 epochs) was performed. Additionally, a learning rate annealing was applied with a factor of 2 and a patience of 10. The convolutional layers were initialized with the Kaiming initialization [56], the linear layers were initialized with the Xavier initialization [57], and the BN layers were initialized with a constant value of one for the weights and zero for the bias.

4.4. Comparison with Other Works

The performance of the proposed approach has been compared to the state-of-the-art, lightweight models that reached the highest accuracy rate for the lowest amount of parameters found in the scientific literature. The following models were compared to the models proposed in this paper: the res15 and its lightweight counterpart res8-narrow from [34], the CENet-GCN-40 and CENet-6 from [36] and the tdnn-swsa and swsa from [44]. These models have been implemented following their respective articles. A baseline MobileNetV2 model (i.e., the Segmented-SA-11k without the Segmented 2D Self-Attention block) is also compared to the proposed approach.

4.5. Transfer Learning

The FSCD that the authors created corresponds to only 7.7% of the size of the GSCD. To overcome the data scarcity of our relatively small dataset, the transfer of the representation learned on the GSCD has been tested with different TL methods. We chose the Segmented-SA-11k model since it is a good trade-off between performance and footprint. A smaller model also limits the risk of overfitting on a small dataset.
Fine-tuning [58] is arguably the most prevalent transfer learning technique in deep learning [46,59,60] due to both its implementation simplicity and good performance. Consequently, this work employs fine-tuning to leverage the GSCD in order to improve the performance of the model on the FSCD. Fine-tuning consists in initializing a determined number of layers of the model that will be trained on the target dataset with the corresponding layers of the pre-trained model. Then, it is possible to retrain on the target dataset the pre-trained model completely, or only partially, by freezing a portion of the layers (i.e., the weights of these layers are not modified by the backpropagation).
Two TL methods and a baseline are to be compared. The baseline corresponds to a random initialization of the model trained on the target dataset (i.e., no TL applied). The first TL method is a complete copy of the pre-trained model without any layers being frozen during the retraining. The second method also uses a complete copy of the pre-trained model, but all the layers are frozen, except the classification layer (i.e., the fully connected layer). For both methods and the baseline, the number of output neurons of the fully connected layer is increased from 12 to 14 (since there are 12 commands to predict in French, plus silence and unknowns). Thus, the weights of the fully connected layer have to be reinitialized for all the tested methods.
As suggested in [61], a two-step procedure is employed to compare the different transfer learning approaches (including the baseline without TL). First, Friedman’s test ranks the approaches amongst each other. Then, Holm’s post hoc test is applied (n = 11) using the best ranked method, being the entire model retraining, as a comparison basis.

5. Implementation Consideration for Practical Use of the Algorithm

Since the targeted application of the KWS algorithm is the control of an assistive robotic arm, which is for the most part mounted on a wheelchair, an embedded device is necessary for practical use. The memory footprint and the number of operations should then be taken into account in the design and choice of that model.
The memory footprint and the number of operations are both critical when designing a model to be implemented on embedded systems. To reduce the size of a model further, the quantization of the network can be carried out and this has shown to minimally affect the performance [62]. Following [33], a small model for microcontrollers can be defined by a maximum memory size of 80 KB and performs 6 MOPS maximum whereas a medium model has a size between 80 and 200 KB and performs between 6 and 20 MOPS. With this classification, considering the quantization to 8 bits of our models, the Segmented-SA-7k takes 16 KB and performs 5.5 MOPS, putting it in the small category. For the Segmented-SA-11k with a size of 25 KB and performing 6.3 MOPS, this model is in the small category in regard to memory, and in the medium category in regard to the number of operations (although being very close to the limit of the small category). Thus, the memory aspect is not a concern since, memory-wise, both models could easily fit in a microcontroller. Considering that the number of operations required is only 15% greater for the Segmented-SA-11k than the Segmented-SA-7k for a performance gain of 3.3% (using the spectrogram for input features), the Segmented-SA-11k has been chosen as the model for the target task of KWS for French voice commands.

6. Evaluation of the Control Interface for a Robotic Assistive Arm

To assess the performance of the proposed vocal command interface, four able-bodied participants aged 22 to 35 were asked to perform two tasks. All participants gave written informed consent prior to experiment onset; this study was approved by the local ethics committee (CIUSSS-CN; project #2019-1603, RIS_2018-616). Note that, despite the small sample size considered, previous works have shown that small sample sizes are sufficient to document the usability of assistive devices [63,64]. Each task was performed using both a standard Jaco joystick and the vocal command interface. At the start of each task, participants were randomly assigned to the joystick or the vocal command interface. Note that the participants did not need to provide any training data for their voices (i.e., the model did not require retraining), making the proposed method fully user independent.
The experiments were performed with a Jaco assistive robotic arm and the experiment setup presented in Figure 3. Jaco is an assistive robotic arm produced by Kinova Robotics [2] designed to be installed on a motorized wheelchair and used by people living with upper-body mobility limitations. Controlled through the wheelchair’s drive control, the Jaco arm allows the user to manipulate objects in their surroundings. This enhanced autonomy allows the user to perform ADLs such as drinking, eating and picking up items [5]. The experiments are based on TEMPA [65] (Test d’Evaluation des Membres Supérieurs des Personnes Âgées), a performance-based evaluation developed to help occupational therapists efficiently and accurately assess the performance of people living with upper limb disabilities.
The joystick control interface, presented in Figure 4, is composed of a stick—which outputs forward, backward, left and right commands—and a button. The button is used to toggle between modes. When “Mode 1” is active, the four stick signals are used to control the arm’s horizontal translations (forward, backward, left, and right); in “Mode 2”, the same signals control the vertical translation and rotation about the effector axis; in “Mode 3”, the stick controls the two remaining rotations; and in “Mode 4”, the stick commands the fingers’ opening and closing.
The vocal command interface is composed of a microphone (Microsoft LifeChat LX-3000) and a press button. Two main operating modes make up the vocal command interface, “Translation” and “Rotation”, which can be activated at any time by voicing the desired mode. Different action commands, chosen with the help of two occupational therapists for their clarity and simplicity, are then added to each mode. In the “Translation” mode, the available vocabulary includes up, down, forward, backward, left, right, open and close; these commands allow the user to perform the three available translations, as well as opening/closing the fingers. In the “Rotation” mode, the available vocabulary is up, down, forward, backward, left, right, plus, minus, open and close; these commands perform the three available rotations of the arm, as well as the rotation of the effector. Commands to open and close the fingers are also accessible in this mode. Users can change between operating modes at any time by voicing the desired operating mode. The user selects a motion with a vocal command (e.g., left) and by pressing the button, they can then make the robot perform the selected motion. The user can release the button to stop the arm’s motion and then press it again to continue the selected motion. The vocal command interface continuously listens to the surrounding sounds picked up by the microphone. The trained model ignores noise and words outside of its dictionary, classifying them as unknown words. The last recognized command stays active as long as no other command is recognized. Note that the words were voiced in French by the participants but are presented here in English for the sake of clarity. A flowchart for both control interfaces logic is presented in Figure 5.

6.1. Task 1

The first experiment was the first task proposed in the TEMPA (i.e.,“Pick-up and move a jar”) and consisted of controlling the Jaco arm to reach, grasp, and lift a coffee jar placed on the top shelf and move it onto the middle of the bottom shelf. Each participant performed this task with both the joystick and the vocal command interface. At the start of the task, the Jaco arm was in its default home position, and the jar was positioned on the right side of the top shelf. Each participant repeated the experiment three times using each control interface.

6.2. Task 2

The second experiment required participants to perform the third task proposed by the TEMPA (i.e., “Pour water from a pitcher into a glass”). The participants controlled the Jaco arm in order to reach, grasp, and lift a bottle filled with water placed on the top shelf and pour water into a glass placed in the middle of the bottom shelf. The task was performed with both the vocal command interface and the joystick. At the start of the task, the Jaco arm was in its default home position and the bottle was positioned on the right side of the top shelf. This experiment was also repeated three times by each participant and with each control interface.

6.3. Subjective Evaluation

In order to compare the two interfaces, each participant completed a questionnaire consisting of seven questions. The questionnaire, based on [66], was designed to evaluate subjective measures related to the use of the interfaces and gain insights for future development of the vocal command. The questionnaire used a five-point Likert scale.

7. Results

7.1. Comparison with Other Works

The accuracy results of each model trained and tested on the GSCD are presented in Table 2. They were obtained from a test run on the standard test set provided by the GSCD of ten trainings from different initialization seeds. The prediction accuracy is used for the performance comparison of the models. The accuracy is the fraction of correct predictions over all the predictions made. The 95% confidence interval is computed from a t-distribution with nine degrees of freedom.

7.2. Result of Transfer Learning on the French Dataset

To assess the performance of the transfer learning from the GSCD (source) to the FSCD (target), the test dataset of the FSCD is used to compare the different TL methods at each epoch. Of the 33 participants who contributed vocal command utterances to the FSCD, 22 were randomly selected to be used as the training and validation sets while the remaining 11 formed the test set. The test set is only used to assess the performance of the TL. It does not serve the purpose of guiding the training of the model; the training and validation sets are used for this motive. In the same way as the GSCD, all utterances produced by the same speaker are only part of one set. The spectrogram was used for the input features. The results of the two TL methods and the baseline are presented in Figure 6. The score reported at each epoch is the average of ten trainings from different initialization seeds on the test dataset of the target.

7.3. Result of the Task Completion Time

The completion time was recorded for each test in both experiments. The results of the first experiment are shown in Figure 7. The average completion time using the joystick was 55.2 s with a standard deviation of 6.6 s, while the average completion time using the vocal control interface was 86.2 s with a standard deviation of 17.8 s.
The results of the second experiment are shown in Figure 7. The average completion time using the joystick was 91.5 s with a standard deviation of 20.1 s, while the average completion time using the vocal control interface was 109.2 s with a standard deviation of 18.1 s.

7.4. Control Interface Intuitiveness

The subjective evaluation conducted via the questionnaire was aimed at comparing the two interfaces and developing a better understanding of the strengths and weaknesses of the vocal command interface for future improvements. The Likert scale was scored from 1 (Totally Disagree) to 5 (Totally Agree), with 3 being the neutral option (Neither Agree Nor Disagree). The results of the questionnaire are presented in Figure 8.
All participants found the vocal control interface useful (Q2, average: 4.75) and easy to learn (Q7, average: 4.75). However, participants expressed that the vocal command interface did not improve their performance (Q3, average: 1.75), an observation that was consistent with the task completion time.
While it is a less striking result, participants agreed that the vocal command interface was easier to use than the joystick (Q4, average: 3.5); this is consistent with their observation that the joystick required more effort to use than the vocal command (Q1, average: 3.75). More notably, participants found that using the vocal control interface required less concentration than the joystick (Q5, average: 4); participants also found it to be more intuitive to use than the joystick (Q6, average: 4.25). One participant observed that she “didn’t have to think much” when using the vocal command interface. In fact, participants indicated that the vocal command interface design made it easy to navigate within the control modes simply by voicing the mode and the movement to execute.

8. Discussion

8.1. Model Performance

The initial step for the comparison with other works was to assess if our implementation of each model from their respective articles achieved an accuracy close to the reported results. Each implementation obtained a difference in accuracy of less than 0.2% from their original implementation reported in their respective article; hence, they were judged satisfactory.
For the MFCCs features, we can see from Table 2 that the proposed Segmented-SA-11k model achieves a superior classification accuracy to all the other models. More precisely, it shows a gain of 0.6% in accuracy (97.1% vs. 96.5%) from the second best model (i.e., CENet-GCN-40). According to non-parametric Wilcoxon signed-rank tests with paired data [61] where each person of the test dataset is considered as a single dataset (n = 188), the median of the performance of the Segmented-SA-11k is significantly different from the median of the CENet-GCN-40 (p-value < 0.000001). This performance is achieved with a reduction of more than a factor of six in model size (11.3 k vs. 72.3 k) between these two models. Compared to the model with the closest number of parameters (i.e., tdnn-swsa), it shows a performance boost of 1.4% with 0.7 k less parameters (11.3 k vs. 12 k). The smallest proposed model, the Segmented-SA-7k, which is also the one with the smallest number of parameters of all the tested models, yields a performance of 95.4%, which is close to the res15 (95.8%) even though it has more than 33 times fewer parameters. Compared to the model with the nearest number of parameters (SWSA), it shows a performance improvement of 5.1%.
For the spectrogram features for which the results are presented in the last column of Table 2, we note a constant gain in performance for the majority of the models. This result shows that a larger and less pre-processed input can improve the performance of a model with a high representation power. The only models presenting a lower performance with the spectrogram are also the ones with a low number of parameters (i.e SWSA and Segmented-SA-7k). Considering that the spectrogram is four times the size of the MFCCs in terms of frequency and that its features are not as uncorrelated as the MFCCs, this can prevent models with a limited representation from fully capturing all of the features of a larger input. Thus, a smaller and more pre-processed input such as the MFCCs can facilitate the learning of such models. The Segmented-SA-11k model still has the highest performance (97.4%), which is 0.4% more than the second-best model (CENet-GCN-40). The median accuracy of the Segmented-SA-11k is also significantly different from the median of the CENet-GCN-40 according to non-parametric Wilcoxon signed-rank tests with paired data where each person of the test dataset is considered as a single dataset (p-value < 0.000001).
Furthermore, the results showcased in Table 2 highlight the improvement obtained from utilizing the Segmented 2D Self-attention in the proposed models (Segmented-SA-7K and Segmented-SA-11K) compared to the baseline MobileNetV2 for both MFCCs and spectrogram features.

8.2. Transfer Learning Performance

From Figure 6, it can be observed that the impact of the transfer learning on the performance of the model on the FSCD is highly notable. Indeed, the performance of the model trained directly on the FSCD from a random initialization of all its layers yielded a poor performance (green dotted curve), with a maximum of 63.97% reached at the 58th epoch (average of 10 runs from different initialization seeds). This result demonstrated that the French Dataset does not contain enough data for the model to discriminate efficiently between the different speech commands.
Both methods where transfer learning was applied reached a good performance. The retraining of the entire model (no frozen weights) initialized from the pre-trained model on the source dataset yielded the best performance (94.78% as peak performance of the average of 10 runs at the 43rd epoch). The fine-tuning of the fully connected layer, while the rest of the network was frozen, led to a lower performance than the entire model retraining. Nevertheless, this method of transfer learning still reached a performance peak of 90.15% on the average of 10 runs at the 60th epoch. Since the backpropagation was performed only on the last layer of the network, the retraining took a little less than five minutes for 70 epochs. It is also to be noted that, with TL, the model was trained on a dataset 13 times smaller and still reached a classification score relatively close to the one obtained on the GSCD (97.4%).
The TL approach, which retrained the entire model, was shown to achieve better performance than the other two approaches, and this difference was judged significant following the two-step procedure described in section IV.E (p < 0.000001 in both cases).

8.3. Participants’ Evaluation

Both experiments showed that use of the vocal command interface resulted in slower task completion times compared to the standard Jaco joystick (average of 55.2 s vs. 86.2 s for task 1, and 91.5 s vs. 109.2 s for task 2). These longer completion times can be partially explained by speech processing delays. Although vocal command processing creates a short delay (75 milliseconds), the overall delay observed was mainly induced by the algorithm, which waits for the sound signal to be under the noise threshold for half a second before proceeding to the speech recognition.
However, although healthy participants were able to easily operate the joystick and quickly alternate between modes with the help of the push button, people living with upper-body incapacities may have difficulty operating the joystick (e.g., due to a lack of dexterity or the presence of spasms) or switching between the joystick control and the push button. Other users simply cannot operate a joystick (e.g., people with quadriplegia). Since the vocal command interface uses only a push button, which can be placed anywhere (e.g., a head push button for people with quadriplegia), the hypothesis is that for the target users, task completion times with the vocal control interface could be comparable or better than task completion times with the joystick control interface.

8.4. Experiment Limitations

Although these experiments served the purpose of testing the vocal command interface in real-life conditions, we remain cautious in relation to drawing conclusions about these results. While tests involving small samples (n = 4–5) have been shown to be sufficient to document the usability of assistive devices [63,64], this is still a very limited sample. An even more critical factor is that these experiments were not conducted with end users, but rather with healthy participants who had no physical limitations operating the standard Jaco joystick. Both experiments, however, provided good performance assessments of the vocal command interface in real conditions, as well as valuable insights concerning possible improvements. Another important factor to consider is noise level, which was relatively controlled in the test environment where the experiments were conducted. Vocal command interface performance is in fact limited in low signal-to-noise ratio environments.

9. Conclusions

This paper presents a lightweight speaker-independent speech recognition model for the KWS task that was shown to outperform the current state-of-the-art models. Moreover, this work presents the Segmented 2D Self-Attention mechanism, a self-attention mechanism designed to improve the performance of the KWS task. On the test set of the GSCD, the proposed model (Segmented-SA-11k) achieved 97.1% with the MFCCs as input features, which is 0.6% higher than the second best model, the CENet-GCN-40. With the spectrogram features, this model achieved 97.4%, a gain of 0.4% from the CENet-GCN-40. These performance gains were achieved despite the fact that the Segmented-SA-11k has a model size six times smaller compared to the CENet-GCN-40 (11.3 k vs. 72.3 k). The performance differences were shown to be significant for both input features and made the proposed model convenient to use in practice.
In addition, this work aimed at enabling robotic arm control with French voice commands (as French is the first language of the target users). TL methods have been used to overcome the data scarcity problem of the authors custom French dataset by transferring the learned representation of the network from the GSCD toward the FSCD. Both transfer learning methods that have been tested yielded a high performance: 94.78% when retraining the entire model (no frozen weights) and 90.15% for the fine-tuning of the fully connected layer, both results being the peak performance on the average of 10 runs. The random initialization and retraining of the entire network on the FSCD, which is equivalent to applying no TL, yielded a peak result of only 63.97% on the average of 10 runs. This result emphasized the necessity of using a TL approach when implementing a vocal interface from a small dataset. Moreover, the difference between the entire model retraining and the two other methods (fully connected layer fine tuning and baseline without TL) were shown to be significant.
Our vocal command interface was evaluated on two tasks from the TEMPA. Four able-bodied participants performed both tasks using the standard Jaco joystick and the vocal command interface. Importantly, the four participants did not provide any training data to the algorithm, and thus the proposed method was evaluated in a fully user-independent setting. Although participants required more time to complete both experimental tasks using the vocal command interface, they nonetheless expressed that the vocal command interface was more intuitive to use and that performing the tasks required less concentration than with the joystick. Participant perception of the vocal interface was also positive, as they reported that it was both useful and easy to learn.
Future works will focus on three main developments. Firstly, the robustness of the model to noisy environments and out of dictionary words will be increased by adding more examples to the training set. These examples could be gathered from different large publicly available datasets such as the Common Voice [67]. Secondly, the proposed model will be implemented on an embedded system for utility and mobility reasons. Finally, the algorithm will be tested with people living with upper limb disabilities to assess its real-use-case performance.

Author Contributions

Conceptualization, S.P.; Methodology, S.P., U.C.-A., F.R. and A.C.-L.; Software, S.P.; Validation, S.P. and U.C.-A.; Formal analysis, S.P.; Investigation, S.P.; Resources, A.C.-L.; Data curation, S.P.; Writing—original draft, S.P.; Writing—review & editing, U.C.-A., F.R. and A.C.-L.; Visualization, S.P.; Supervision, A.C.-L.; Project administration, A.C.-L.; Funding acquisition, A.C.-L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Centre interdisciplinaire de recherche en réadaptation et intégration sociale (CIRRIS) under Grant #02613 and the Fonds de recherche du Québec–Santé (FRQS) Research Scholar under Grant #34699 and #296761.

Institutional Review Board Statement

The project was approved by the local ethics committee 369 (Ethics # 2019-1603, RIS_2018-616, CIUSSS-CN) and all participants gave informed consent.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Publicly available datasets were analyzed in this study. This data can be found here for the GSCD: http://download.tensorflow.org/data/speech_commands_v0.01.tar.gz (accessed on 10 March 2020) and here for the FSCD: https://arweave.net/rIRaVzZqR7aRa0U3gg2Hqxgpat0-wL1XFMZVUd0pq9M (accessed on 10 March 2020).

Acknowledgments

I wish to express my gratitude to my research supervisor for his infinite patience and support throughout the research and writing process that led to the creation of this article.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Dunning, A.; Herder, J.L. A review of assistive devices for arm balancing. In Proceedings of the 2013 IEEE 13th International Conference on Rehabilitation Robotics (ICORR), Seattle, WA, USA, 24–26 June 2013; pp. 1–6. [Google Scholar]
  2. Campeau-Lecours, A.; Lamontagne, H.; Latour, S.; Fauteux, P.; Maheu, V.; Boucher, F.; Deguire, C.; L’Ecuyer, L.J.C. Kinova Modular Robot Arms for Service Robotics Applications. Int. J. Robot. Appl. Technol. 2017, 5, 49–71. [Google Scholar] [CrossRef]
  3. Turgeon, P.; Dubé, M.; Laliberté, T.; Archambault, P.S.; Flamand, V.H.; Routhier, F.; Campeau-Lecours, A. Mechanical design of a new device to assist eating in people with movement disorders. Assist. Technol. 2020, 34, 170–177. [Google Scholar] [CrossRef] [PubMed]
  4. Beaudoin, M.; Lettre, J.; Routhier, F.; Archambault, P.S.; Lemay, M.; Gélinas, I. Long-term use of the JACO robotic arm: A case series. Disabil. Rehabil. Assist. Technol. 2019, 14, 267–275. [Google Scholar] [CrossRef] [PubMed]
  5. Maheu, V.; Archambault, P.S.; Frappier, J.; Routhier, F. Evaluation of the JACO robotic arm: Clinico-economic study for powered wheelchair users with upper-extremity disabilities. In Proceedings of the 2011 IEEE International Conference on Rehabilitation Robotics (ICORR), Zurich, Switzerland, 29 June–1 July 2011; pp. 1–5. [Google Scholar]
  6. Ka, H.; Ding, D.; Cooper, R.A. Three Dimensional Computer Vision-Based Alternative Control Method For Assistive Robotic Manipulator. Int. J. Adv. Robot. Autom. 2016; 1, in press. [Google Scholar]
  7. Farina, D.; Jiang, N.; Rehbaum, H.; Holobar, A.; Graimann, B.; Dietl, H.; Aszmann, O.C. The extraction of neural information from the surface EMG for the control of upper-limb prostheses: Emerging avenues and challenges. IEEE Trans. Neural Syst. Rehabil. Eng. 2014, 22, 797–809. [Google Scholar] [CrossRef]
  8. Naumann, A.; Hurtienne, J.; Israel, J.H.; Mohs, C.; Kindsmüller, M.C.; Meyer, H.A.; Hußlein, S. Intuitive use of user interfaces: Defining a vague concept. In Proceedings of the 7th International Conference on Engineering Psychology and Cognitive Ergonomics, Beijing, China, 22–27 July 2007; Springer: Berlin/Heidelberg, 2007; pp. 128–136. [Google Scholar]
  9. Palacios, R.H. Robotic Arm Manipulation Laboratory with a Six Degree of Freedom JACO Arm; Naval Postgraduate School: Monterey, CA, USA, 2015. [Google Scholar]
  10. Fall, C.L.; Turgeon, P.; Campeau-Lecours, A.; Maheu, V.; Boukadoum, M.; Roy, S.; Massicotte, D.; Gosselin, C.; Gosselin, B. Intuitive wireless control of a robotic arm for people living with an upper body disability. In Proceedings of the 2015 37th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC 2015), Milan, Italy, 25–29 August 2015; pp. 4399–4402. [Google Scholar] [CrossRef]
  11. Abascal, J. Users with Disabilities: Maximum Control with Minimum Effort. In Articulated Motion and Deformable Objects, Proceedings of the AMDO 2008: Articulated Motion and Deformable Objects, Port d’Andratx, Mallorca, Spain, 9–11 July 2008; Springer: Berlin/Heidelberg, Germany, 2008; pp. 449–456. [Google Scholar] [CrossRef]
  12. Cook, A.M.; Polgar, J.M. Assistive Technologies- E-Book: Principles and Practice; Google-Books-ID: ODWaBQAAQBAJ; Elsevier Health Sciences: Amsterdam, The Netherlands, 2014. [Google Scholar]
  13. Lebrasseur, A.; Lettre, J.; Routhier, F.; Archambault, P.; Campeau-Lecours, A. Assistive robotic device: Evaluation of intelligent algorithms. arXiv 2018, arXiv:1812.07342. [Google Scholar]
  14. Noyes, J.; Frankish, C. Speech recognition technology for individuals with disabilities. Augment. Altern. Commun. 1992, 8, 297–303. [Google Scholar] [CrossRef]
  15. Portet, F.; Vacher, M.; Golanski, C.; Roux, C.; Meillon, B. Design and evaluation of a smart home voice interface for the elderly: Acceptability and objection aspects. Pers. Ubiquitous Comput. 2013, 17, 127–144. [Google Scholar] [CrossRef] [Green Version]
  16. Noyes, J.M.; Haigh, R.; Starr, A.F. Automatic speech recognition for disabled people. Appl. Ergon. 1989, 20, 293–298. [Google Scholar] [CrossRef]
  17. Han, Y.; Hyun, J.; Jeong, T.; Yoo, J.H.; Hong, J.W.K. A smart home control system based on context and human speech. In Proceedings of the 2016 18th International Conference on Advanced Communication Technology (ICACT), PyeongChang, Republic of Korea, 31 January–3 February 2016; pp. 165–169. [Google Scholar] [CrossRef]
  18. Hollier, S.; Abou-Zahra, S. Internet of Things (IoT) as Assistive Technology: Potential Applications in Tertiary Education. In W4A ’18, Proceedings of the 15th International Web for All Conference, Lyon, France, 23–25 April 2018; Association for Computing Machinery: New York, NY, USA, 2018; pp. 1–4. [Google Scholar] [CrossRef]
  19. Kolkowska, E.; Kristofferson, A. Privacy by Design Principles in Design of New Generation Cognitive Assistive Technologies. In ICT Systems Security and Privacy Protection; Hoepman, J.H., Katzenbeisser, S., Eds.; Springer International Publishing: Cham, Switzerland, 2016; Volume 471, pp. 384–397. [Google Scholar] [CrossRef] [Green Version]
  20. Bilmes, J.A.; Malkin, J.; Li, X.; Harada, S.; Kilanski, K.; Kirchhoff, K.; Wright, R.; Subramanya, A.; Landay, J.A.; Dowden, P.; et al. The vocal joystick. In Proceedings of the 2006 IEEE International Conference on Acoustics Speech and Signal Processing Proceedings, Toulouse, France, 14–19 May 2006; Volume 1, p. I-I. [Google Scholar]
  21. Harada, S.; Landay, J.A.; Malkin, J.; Li, X.; Bilmes, J.A. The Vocal Joystick: Evaluation of voice-based cursor control techniques for assistive technology. Disabil. Rehabil. Assist. Technol. 2008, 3, 22–34. [Google Scholar] [CrossRef]
  22. Harada, S.; Wobbrock, J.O.; Landay, J.A. Voicedraw: A hands-free voice-driven drawing application for people with motor impairments. In Proceedings of the the 9th International ACM SIGACCESS Conference on Computers and Accessibility, Tempe, AZ, USA, 15–17 October 2007; pp. 27–34. [Google Scholar]
  23. House, B.; Malkin, J.; Bilmes, J. The VoiceBot: A voice controlled robot arm. In Proceedings of the the SIGCHI Conference on Human Factors in Computing Systems, CHI ’09; Association for Computing Machinery, Boston, MA, USA, 4–9 April 2009; pp. 183–192. [Google Scholar] [CrossRef]
  24. Akçınar, D.; Arıtürk, M.K.; Yıldırım, T. Speaker Dependent Voice Controlled Robotic Arm. In Proceedings of the 2018 Innovations in Intelligent Systems and Applications (INISTA), Thessaloniki, Greece, 3–5 July 2018; pp. 1–5. [Google Scholar] [CrossRef]
  25. Volosyak, I.; Ivlev, O.; Graser, A. Rehabilitation robot FRIEND II-the general concept and current implementation. In Proceedings of the 9th International Conference on Rehabilitation Robotics, ICORR, Chicago, IL, USA, 28 June–1 July 2005; pp. 540–544. [Google Scholar] [CrossRef]
  26. Guo, S.; Wang, Z.; Guo, J.; Fu, Q.; Li, N. Design of the Speech Control System for a Upper Limb Rehabilitation Robot Based on Wavelet De-noising. In Proceedings of the 2018 IEEE International Conference on Mechatronics and Automation (ICMA), Changchun, China, 5–8 August 2018; pp. 2300–2305. [Google Scholar] [CrossRef]
  27. Nishimori, M.; Saitoh, T.; Konishi, R. Voice controlled intelligent wheelchair. In Proceedings of the SICE Annual Conference 2007, Takamatsu City, Japan, 17–20 September 2007; pp. 336–340. [Google Scholar]
  28. Pires, G.; Nunes, U. A wheelchair steered through voice commands and assisted by a reactive fuzzy-logic controller. J. Intell. Robot. Syst. 2002, 34, 301–314. [Google Scholar] [CrossRef]
  29. Rohlicek, J.R.; Russell, W.; Roukos, S.; Gish, H. Continuous hidden Markov modeling for speaker-independent word spotting. In Proceedings of the International Conference on Acoustics, Speech, and Signal Processing, Glasgow, UK, 23–26 May 1989; Volume 1, pp. 627–630. [Google Scholar] [CrossRef]
  30. Rose, R.; Paul, D. A hidden Markov model based keyword recognition system. In Proceedings of the International Conference on Acoustics, Speech, and Signal Processing, Albuquerque, NM, USA, 3–6 April 1990; Volume 1, pp. 129–132. [Google Scholar] [CrossRef]
  31. Wilpon, J.G.; Miller, L.G.; Modi, P. Improvements and applications for key word recognition using hidden Markov modeling techniques. In Proceedings of the ICASSP 91: 1991 International Conference on Acoustics, Speech, and Signal Processing, Toronto, ON, Canada, 14–17 May 1991; Volume 1, pp. 309–312. [Google Scholar] [CrossRef]
  32. Chen, G.; Parada, C.; Heigold, G. Small-footprint keyword spotting using deep neural networks. In Proceedings of the 2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Florence, Italy, 4–9 May 2014; pp. 4087–4091. [Google Scholar] [CrossRef] [Green Version]
  33. Zhang, Y.; Suda, N.; Lai, L.; Chandra, V. Hello Edge: Keyword Spotting on Microcontrollers. arXiv 2018, arXiv:1711.07128. [Google Scholar]
  34. Tang, R.; Lin, J. Deep Residual Learning for Small-Footprint Keyword Spotting. In Proceedings of the 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; pp. 5484–5488. [Google Scholar] [CrossRef] [Green Version]
  35. Zeng, M.; Xiao, N. Effective Combination of DenseNet and BiLSTM for Keyword Spotting. IEEE Access 2019, 7, 10767–10775. [Google Scholar] [CrossRef]
  36. Chen, X.; Yin, S.; Song, D.; Ouyang, P.; Liu, L.; Wei, S. Small-footprint Keyword Spotting with Graph Convolutional Network. arXiv 2019, arXiv:1912.05124. [Google Scholar]
  37. Shan, C.; Zhang, J.; Wang, Y.; Xie, L. Attention-based End-to-End Models for Small-Footprint Keyword Spotting. arXiv 2018, arXiv:1803.10916. [Google Scholar]
  38. Wang, X.; Sun, S.; Shan, C.; Hou, J.; Xie, L.; Li, S.; Lei, X. Adversarial Examples for Improving End-to-end Attention-based Small-footprint Keyword Spotting. In Proceedings of the ICASSP 2019—2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019; pp. 6366–6370. [Google Scholar] [CrossRef]
  39. De Andrade, D.C.; Leo, S.; Viana, M.L.D.S.; Bernkopf, C. A neural attention model for speech command recognition. arXiv 2018, arXiv:1808.08929. [Google Scholar]
  40. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention Is All You Need. arXiv 2017, arXiv:1706.03762. [Google Scholar]
  41. Salazar, J.; Kirchhoff, K.; Huang, Z. Self-attention Networks for Connectionist Temporal Classification in Speech Recognition. In Proceedings of the ICASSP 2019—2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019; pp. 7115–7119. [Google Scholar] [CrossRef] [Green Version]
  42. Povey, D.; Hadian, H.; Ghahremani, P.; Li, K.; Khudanpur, S. A Time-Restricted Self-Attention Layer for ASR. In Proceedings of the 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; pp. 5874–5878. [Google Scholar] [CrossRef]
  43. Pham, N.Q.; Nguyen, T.S.; Niehues, J.; Müller, M.; Stüker, S.; Waibel, A. Very Deep Self-Attention Networks for End-to-End Speech Recognition. arXiv 2019, arXiv:1904.13377. [Google Scholar]
  44. Bai, Y.; Yi, J.; Tao, J.; Wen, Z.; Tian, Z.; Zhao, C.; Fan, C. A Time Delay Neural Network with Shared Weight Self-Attention for Small-Footprint Keyword Spotting. In Proceedings of the Interspeech 2019, ISCA, Graz, Austria, 15–19 September 2019; pp. 2190–2194. [Google Scholar] [CrossRef] [Green Version]
  45. Wang, D.; Zheng, T.F. Transfer learning for speech and language processing. In Proceedings of the 2015 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA), Hong Kong, China, 16–19 December 2015; pp. 1225–1237. [Google Scholar] [CrossRef] [Green Version]
  46. Yosinski, J.; Clune, J.; Bengio, Y.; Lipson, H. How transferable are features in deep neural networks? In Proceedings of the 27th International Conference on Neural Information Processing Systems, Montreal, QC, Canada, 8–13 December 2014; pp. 3320–3328. [Google Scholar]
  47. Kunze, J.; Kirsch, L.; Kurenkov, I.; Krug, A.; Johannsmeier, J.; Stober, S. Transfer Learning for Speech Recognition on a Budget. arXiv 2017, arXiv:1706.00290. [Google Scholar]
  48. Warden, P. Speech Commands: A Dataset for Limited-Vocabulary Speech Recognition. arXiv 2018, arXiv:1804.03209. [Google Scholar]
  49. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.C. MobileNetV2: Inverted Residuals and Linear Bottlenecks. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 4510–4520. [Google Scholar]
  50. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  51. Lin, M.; Chen, Q.; Yan, S. Network in network. arXiv 2013, arXiv:1312.4400. [Google Scholar]
  52. Bellman, R. Dynamic Programming. Science 1966, 153, 34–37. [Google Scholar] [CrossRef] [PubMed]
  53. Bhattacharjee, U.; Gogoi, S.; Sharma, R. A statistical analysis on the impact of noise on MFCC features for speech recognition. In Proceedings of the 2016 International Conference on Recent Advances and Innovations in Engineering (ICRAIE), Jaipur, India, 23–25 December 2016; pp. 1–5. [Google Scholar] [CrossRef]
  54. Paszke, A.; Gross, S.; Chintala, S.; Chanan, G.; Yang, E.; DeVito, Z.; Lin, Z.; Desmaison, A.; Antiga, L.; Lerer, A. Automatic differentiation in pytorch. In Proceedings of the 31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  55. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  56. He, K.; Zhang, X.; Ren, S.; Sun, J. Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1026–1034. [Google Scholar]
  57. Glorot, X.; Bengio, Y. Understanding the difficulty of training deep feedforward neural networks. In Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics, Sardinia, Italy, 13–15 May 2010; pp. 249–256. [Google Scholar]
  58. Hinton, G.E.; Salakhutdinov, R.R. Reducing the dimensionality of data with neural networks. Science 2006, 313, 504–507. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  59. Bengio, Y. Deep learning of representations for unsupervised and transfer learning. In Proceedings of the ICML Workshop on Unsupervised and Transfer Learning, Edinburgh, UK, 26 June–1 July 2012; pp. 17–36. [Google Scholar]
  60. Huh, M.; Agrawal, P.; Efros, A.A. What makes ImageNet good for transfer learning? arXiv 2016, arXiv:1608.08614. [Google Scholar]
  61. Demšar, J. Statistical comparisons of classifiers over multiple data sets. J. Mach. Learn. Res. 2006, 7, 1–30. [Google Scholar]
  62. Wu, J.; Leng, C.; Wang, Y.; Hu, Q.; Cheng, J. Quantized Convolutional Neural Networks for Mobile Devices. arXiv 2016, arXiv:1512.06473. [Google Scholar]
  63. Resnik, L. Development and testing of new upper-limb prosthetic devices: Research designs for usability testing. J. Rehabil. Res. Dev. 2011, 48, 697–706. [Google Scholar] [CrossRef]
  64. Virzi, R.A. Refining the test phase of usability evaluation: How many subjects is enough? Hum. Factors 1992, 34, 457–468. [Google Scholar] [CrossRef]
  65. Desrosiers, J.; Hébert, R.; Dutil, E.; Bravo, G. Development and reliability of an upper extremity function test for the elderly: The TEMPA. Can. J. Occup. Ther. 1993, 60, 9–16. [Google Scholar] [CrossRef]
  66. Schmidtler, J.; Bengler, K.; Dimeas, F.; Campeau-Lecours, A. A questionnaire for the evaluation of physical assistive devices (quead): Testing usability and acceptance in physical human-robot interaction. In Proceedings of the 2017 IEEE International Conference on Systems, Man, and Cybernetics (SMC), Banff, AB, Canada, 5–8 October 2017; pp. 876–881. [Google Scholar]
  67. Ardila, R.; Branson, M.; Davis, K.; Henretty, M.; Kohler, M.; Meyer, J.; Morais, R.; Saunders, L.; Tyers, F.M.; Weber, G. Common voice: A massively-multilingual speech corpus. arXiv 2019, arXiv:1912.06670. [Google Scholar]
Figure 1. The proposed Segmented-SA-11k composed of an initial convolution block followed by three inverted residual blocks interspersed by transition block. The Segmented 2D Self-Attention block is added after the last inverted residual block. The number of output neurons in the prediction block corresponds to the 12 different vocal commands to predict from the Google Speech Commands Dataset (GSCD), including silence and unknown words.
Figure 1. The proposed Segmented-SA-11k composed of an initial convolution block followed by three inverted residual blocks interspersed by transition block. The Segmented 2D Self-Attention block is added after the last inverted residual block. The number of output neurons in the prediction block corresponds to the 12 different vocal commands to predict from the Google Speech Commands Dataset (GSCD), including silence and unknown words.
Sensors 23 06056 g001
Figure 2. A schematic representation of the proposed Segmented 2D Self-Attention block where the upper part is the spatial local self-attention and the lower part describes the channel local self-attention. Both parts are then combined and the residual of the input is added.
Figure 2. A schematic representation of the proposed Segmented 2D Self-Attention block where the upper part is the spatial local self-attention and the lower part describes the channel local self-attention. Both parts are then combined and the residual of the input is added.
Sensors 23 06056 g002
Figure 3. The experiment setup where two tasks proposed in the TEMPA were performed. The Jaco as well as the button used to switch mode when using the joystick or activate the movement of the Jaco when using the vocal command are presented. The starting and ending position of both tasks are also illustrated. Note that the glass used to receive the water in the second task is not in the Figure for the sake of clarity.
Figure 3. The experiment setup where two tasks proposed in the TEMPA were performed. The Jaco as well as the button used to switch mode when using the joystick or activate the movement of the Jaco when using the vocal command are presented. The starting and ending position of both tasks are also illustrated. Note that the glass used to receive the water in the second task is not in the Figure for the sake of clarity.
Sensors 23 06056 g003
Figure 4. The standard Jaco joystick used as a reference point for performance comparison in assessing the qualitative and quantitative aspects of the proposed vocal command interface.
Figure 4. The standard Jaco joystick used as a reference point for performance comparison in assessing the qualitative and quantitative aspects of the proposed vocal command interface.
Sensors 23 06056 g004
Figure 5. The flowchart detailing the process of both the vocal command interface and the joystick in order to control the Jaco movements.
Figure 5. The flowchart detailing the process of both the vocal command interface and the joystick in order to control the Jaco movements.
Sensors 23 06056 g005
Figure 6. Results at each epoch on the test dataset of the French Speech Commands Dataset (FSCD) with the two TL methods and the baseline (without any TL).
Figure 6. Results at each epoch on the test dataset of the French Speech Commands Dataset (FSCD) with the two TL methods and the baseline (without any TL).
Sensors 23 06056 g006
Figure 7. Probability density estimation of the experiments’ completion time using the standard Jaco joystick and the vocal command interface to control the Jaco robot arm for the first and second task.
Figure 7. Probability density estimation of the experiments’ completion time using the standard Jaco joystick and the vocal command interface to control the Jaco robot arm for the first and second task.
Sensors 23 06056 g007
Figure 8. Results of the questionnaire completed by each participant after having performed both tasks to compare the standard Jaco joystick with the vocal command interface. The dots represent the score given by each participant.
Figure 8. Results of the questionnaire completed by each participant after having performed both tasks to compare the standard Jaco joystick with the vocal command interface. The dots represent the score given by each participant.
Sensors 23 06056 g008
Table 1. Commands and filling words of both datasets used for the training, validation and testing of the developed models.
Table 1. Commands and filling words of both datasets used for the training, validation and testing of the developed models.
DatasetsCommandsFilling Words
Google Speech
Commands Dataset
“yes”, “no”, “up”, “down”,
“left”, “right”, “on”, “off”, “stop”, “go”
“bed”, “bird”, “cat”, “dog”,
“happy”, “house”, “marvin”, “sheila”,
“tree”, “wow”, and the digits zero to nine.
French Speech
Commands Dataset
“arrière”, “avant”, “descendre”,
“droite”, “fermer”, “gauche”, “moins”, “monter”,
“ouvrir”, “plus”, “rotation”, “translation”
“assiette”, “bouche”, “combiner”, “doigts”,
“lentement”, “mode”, “position”, “précédent”, “rapide”,
“sauver”, “suivant”, “tourner”, and the digits one to five.
Table 2. Test accuracy with MFCCs and spectrogram for input with 95% confidence intervals (on 10 different initialized trainings) and parameters count of each implemented models.
Table 2. Test accuracy with MFCCs and spectrogram for input with 95% confidence intervals (on 10 different initialized trainings) and parameters count of each implemented models.
ModelRank# Param.Accuracy on MFCCAccuracy on Spectrogram
CENet-6616.2 k93.9% ± 0.2494.7% ± 0.47
CENet-GCN-40272.3 k96.5% ± 0.1997.0% ± 0.22
swsa98 k90.3% ± 0.1888.9% ± 0.36
tdnn-swsa412 k95.7% ± 0.1396.2% ± 0.29
res153238 k95.8% ± 0.2296.4% ± 0.11
res8-narrow819.9 k90.2% ± 0.8490.5% ± 0.45
Baseline MobileNetV279 k92.1% ± 0.2093.6% ± 0.33
Segmented-SA-7k57.1 k95.4% ± 0.2395.0% ± 0.26
Segmented-SA-11k111.3 k97.1% ± 0.1697.4% ± 0.35
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Poirier, S.; Côté-Allard, U.; Routhier, F.; Campeau-Lecours, A. Efficient Self-Attention Model for Speech Recognition-Based Assistive Robots Control. Sensors 2023, 23, 6056. https://0-doi-org.brum.beds.ac.uk/10.3390/s23136056

AMA Style

Poirier S, Côté-Allard U, Routhier F, Campeau-Lecours A. Efficient Self-Attention Model for Speech Recognition-Based Assistive Robots Control. Sensors. 2023; 23(13):6056. https://0-doi-org.brum.beds.ac.uk/10.3390/s23136056

Chicago/Turabian Style

Poirier, Samuel, Ulysse Côté-Allard, François Routhier, and Alexandre Campeau-Lecours. 2023. "Efficient Self-Attention Model for Speech Recognition-Based Assistive Robots Control" Sensors 23, no. 13: 6056. https://0-doi-org.brum.beds.ac.uk/10.3390/s23136056

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop