Next Article in Journal
The Mediating Effect of Life Satisfaction on Relation between Perceived Physical Attractiveness and Health-Promoting Lifestyle in Korean Adults
Next Article in Special Issue
Cluster Analysis of Urban Acoustic Environments on Barcelona Sensor Network Data
Previous Article in Journal
Health-Related Fitness Levels among Title I Elementary School Students
Previous Article in Special Issue
Quieted City Sounds during the COVID-19 Pandemic in Montreal
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Smartphone-Based Crowd-Sourced Database for Environmental Noise Assessment

1
Centre for Studies on Risks, The Environment, Mobility and Urban Planning (CEREMA), Research Unit in Environmental Acoustics (UMRAE), French Institute of Science and Technology for Transport, Development and Networks (IFSTTAR), University Gustave Eiffel, F-44344 Bouguenais, France
2
Lab-STICC CNRS UMR 6285, IUT de Vannes, 8 Rue Montaigne, BP 561, CEDEX, F-56017 Vannes, France
*
Author to whom correspondence should be addressed.
Int. J. Environ. Res. Public Health 2021, 18(15), 7777; https://0-doi-org.brum.beds.ac.uk/10.3390/ijerph18157777
Submission received: 30 April 2021 / Revised: 6 July 2021 / Accepted: 9 July 2021 / Published: 22 July 2021
(This article belongs to the Special Issue Environmental Acoustics Assessments in Urban and Suburban Contexts)

Abstract

:
Noise is a major source of pollution with a strong impact on health. Noise assessment is therefore a very important issue to reduce its impact on humans. To overcome the limitations of the classical method of noise assessment (such as simulation tools or noise observatories), alternative approaches have been developed, among which is collaborative noise measurement via a smartphone. Following this approach, the NoiseCapture application was proposed, in an open science framework, providing free access to a considerable amount of information and offering interesting perspectives of spatial and temporal noise analysis for the scientific community. After more than 3 years of operation, the amount of collected data is considerable. Its exploitation for a sound environment analysis, however, requires one to consider the intrinsic limits of each collected information, defined, for example, by the very nature of the data, the measurement protocol, the technical performance of the smartphone, the absence of calibration, the presence of anomalies in the collected data, etc. The purpose of this article is thus to provide enough information, in terms of quality, consistency, and completeness of the data, so that everyone can exploit the database, in full control.

1. Introduction

Noise is a very significant source of pollution, particularly in urban areas, with significant effects on health. The fight against noise is a fundamental societal and health issue, to which the public authorities are trying to respond by putting regulations in place. In Europe, for example, the directive 2002/49/EC aims to establish an inventory of noise nuisance, to propose actions to reduce nuisance and to communicate to citizens about their exposure to noise [1]. In this regulatory context, the main tool for decision-makers is the production of strategic noise maps.
These maps are generally produced using specific software, integrating noise emission and acoustic propagation models, coupled with geospatial data and traffic information. Although these maps are limited by the calculation assumptions and the quality of the input data, they make it possible to assess the broad outlines of a noise distribution in a city and to evaluate the effect of action plans to reduce noise. However, they generally lack realism, particularly from the point of view of the temporal dynamics of noise. Conversely, noise observatories, consisting of a large number of acoustic sensors, offer a more realistic description of noise environments. However, the limitation of the number of sensors, for technical and cost reasons, does not allow carrying out noise mapping with a sufficient spatial step.
Faced with these observations, alternatives have been proposed. In particular, the use of more affordable sensor networks has been investigated, allowing to densify the observation points [2]. Another way consists in the involvement of citizens as data collectors, in a crowd-sourcing approach. For example, the Smart Citizen System project proposes a low-cost sensor specifically dedicated to collect noise data by citizen action [3]. Considering a soundscape approach, data produced by people on location-based social networks can also be analyzed to produce maps of the sound environment, like with the Chatty maps experiment [4] or more recently by Gasco et al. [5]. The Sound Around You project is another example, by proposing a web interface to collect soundscape recording and opinions [6]. Nowadays, among all the citizen science-oriented approaches, the one based on the use of smartphones is undoubtedly the most developed in the literature. In particular, Santini et al. have demonstrated the capabilities of a smartphone to perform environmental acoustic measurements [7]. It was followed by several works that have given rise to specific noise and soundscape crowd-sourcing type applications and platforms (see Ear-Phone, NoiseSPY, and NoiseTube applications, respectively, in [8,9,10]). What is interesting in these first works is that, despite the technical limitations of the time (i.e., smartphones with limited technical capabilities and resources), almost all the topics related to this issue had already been discussed: smartphone calibration, data quality, noise maps reconstruction, contextualized data collection (perceptual data), the need for a complementary web interface, the need to know the context of the measurement, the implementation of specific events to organize data collection, contributors privacy, motivation of contributors, etc. Subsequently, other contributions have appeared on this subject; the reader may refer to recent literature reviews [11,12,13,14,15] for more details.
The evolution of Information and Communication Technologies (ICTs) and the experience obtained from the past researches allowed the implementation of very advanced solutions, in the last few years, among them the Sense2Health platform (which led to the Ambiciti plateform) [16], integrating a data assimilation model to produce more realistic noise maps; the Hush City platform [13], for collecting data in quiet areas; the City Soundscape platform [17], with the objective to evaluate action plans for road noise reduction; and the GRCSensing platform [14] with an interesting feature for distributing tasks to users in order to capture noise in specific urban areas and times.
Proposed more recently, the NoiseCapture project is completely in line with the last platforms [18], but extends the concept of participatory science to that of open science. Thus, all source codes, whether for the smartphone application, the spatial data infrastructure, or the web interface, are released as open source. In the same way, the data are available in open data in many ways, and, as far as possible, the scientific productions, in open access. Attention was also paid to the long-term sustainability of the system, the NoiseCapture project being part of an operational framework and not in the form of a short-term experimentation. The objective is to ensure a collection of data over several years, in order to constitute a reference database for the study of sound environments over the long term. The respect for privacy and use of personal data is also a founding element of the NoiseCapture project; in order to respect the national regulation, in particular in Europe, no sound or video recordings are made, nor is any personal information collected; the use of the application does not require the creation of an account. Finally, the developers of the application have paid great attention to the quality of the acoustic data collected, by integrating proven signal processing algorithms, and by proposing several methods for smartphone calibration. After more than 3 years of existence, the amount of data collected worldwide thanks to the application is thus considerable (more than 100,000 downloads, 74,000 contributors, 260,000 tracks that represents around 60 million of one second measurement points), showing the interest of the citizens for this participatory approach and offering very promising operational and research perspectives.
Nevertheless, the exploitation of the database, whether in an operational or research context, requires a good knowledge of the inherent limitations of the methodology, such as the lack of control of the measurement protocol, poor acoustic calibration of the application, measurements tainted by uncertainties, the misuse of the application, the metrological limitations of smartphones, the context of the measurement, etc. In order to ensure that any user of the database has a perfect knowledge and control of the information contained in the database, a full description and an analysis of the database is performed in this article, in order to highlight the various uncertainties, irregularities, or inconsistencies that need to be considered before any exploitation of the data. This analysis also highlights future evolution that it would be interesting to consider in order to improve the application and to increase the quality of the collected data but also to collect additional information in order to better take into account the context of the measure in its exploitation. This article does not therefore constitute an acoustic study of sound environments, but provides a framework for understanding the NoiseCapture database for its future exploitation. The study of the noise environments using this database will be the subject of further works.
The NoiseCapture platform is first presented in Section 2. The collected data are then described and analyzed in the Section 3, providing sufficient information to a future user, for an exploitation of the database in total control of the nature of the data and their possible limitations. In Section 4, a discussion is provided for improving the application and the methodology to increase the data quality and analysis. Last, Section 5 concludes this work.

2. NoiseCapture Application and Database Description

2.1. NoiseCapture History

The development of the NoiseCapture application was initiated by the french National Center for Scientific Research (CNRS) and the Université Gustave Eiffel (formerly Ifsttar) within the framework of the European ENERGIC-OD project [19], which aimed at producing and redistributing geospatial information in open data to user communities. The development continued thereafter, as a part of the Noise-Planet project [20], with the objective to combine geomatic and acoustic sciences for the evaluation of outdoor sound environments. In line with the general goal of the Noise-Planet project, it was decided to develop the NoiseCapture application in the framework of an Open Science approach, with the dissemination of source codes in Open Source, data in Open Data, and as far as possible, scientific dissemination through publications in Open Access journals.
The initial objective of the NoiseCapture application was to propose a smartphone application to a community of specialists (technical staff within a local authority for example), in order to assess the outdoor sound environments in their territory, by using a collaborative mapping tool. The target audience was therefore initially people with technical and, possibly, acoustic knowledge, allowing them to understand a rather professional smartphone application.
The NoiseCapture application was designed in order to carry out acoustic measurements over a shorter period of time, if possible while walking, in order to collect data on a large spatial area. The user was expected to keep the smartphone in hand throughout the measurement, especially to control the measurement. Thus, the measurements are user-initiated and not background. Each user can then decide to upload data to a remote server that collects all the data in a database, performs further analysis, and represents the results collected by a set of users in the form of a noise map.
The choice of the development environment was oriented towards the most widespread platform, namely, Android, whose market share has been above 80% for many years (around 15% for iOS (a mobile operating system created and developed by Apple Inc.)) [21]. The porting of the application to iOS has not been achieved, although a gain in terms of metrological quality may be possible due to a lower variability in devices [22]. In order to promote the diffusion of the application worldwide, the application has been translated thanks to volunteers, in several languages (English (en), Chinese (China, zh_CN), French (fr), Greek (el), Polish (pl), Portuguese (Brazil, pt_BR), Spanish (es)). In the rest of this article, the terms and features of the application refer to the 1.2.15 version (release 51) of NoiseCapture with the default language (i.e., “en” for English). The last public NoiseCapture release is available on Google Play (“Google Play” brand is property of Google LLC) [23].

2.2. NoiseCapture Description

2.2.1. NoiseCapture Android App

From a functional point of view, the NoiseCapture application uses the principles of a “pocket” sound level meter. The main screen (Figure 1, “Measurement”) presents the results of an acoustic measurement through several classical acoustic indicators: an instantaneous sound level (calculated on a sliding window), as well as the minimum (Min), maximum (Max), and average (Mean) instantaneous sound levels over the duration of a measurement. The instantaneous spectrum by third octave band between 100 Hz and 16 kHz is also proposed on a specific tab, as well as a spectrogram. The duration of the measurement is also indicated: the user can start, pause/resume, and stop the measurement at their convenience, and the measurement duration can also be automatically be fixed in the application settings (the user starts the measurement, but it stops by itself after a certain duration).
In accordance with the initial objective of the application, i.e., the production of a noise map, each measurement is geolocalized with the last known GPS location. The measurement screen also indicates the position of each measurement, every second (i.e., the “Measurement Point”), and more globally the trace of a measurement (i.e., the “Measurement Track”) according to the user displacement. The accuracy of the location is also indicated both numerically and graphically on the map. At this point, note that the acoustic indicators that are calculated and displayed on this screen, as well as those that are presented on the “Results” screen, do not result in any audio recording; these indicators are calculated on the fly.
After a measurement has been performed, the user accesses a second screen (Figure 1, “Description”), which allows the user to give additional information to the measurement. Filling this form is entirely optional. Some information, such as “Description” and “Pictures”, are only stored on the smartphone, while other data may be collected and transmitted to the NoiseCapture remote data server. The choice of whether or not to transmit the measurements and information can be configured by the user. On this screen, 3 types of information can be provided: (1) information on the perceived quality of the sound environment (“Pleasantness”); (2) information on the measurement conditions using 4 tags (“Test”, “Indoor”, “Rain”, and “Wind”); (3) information on the nature of the sound sources perceived during the measurement, using 14 tags (“Footsteps”, “Voice”, “Natural”, “Mechanical”, “Human”, “Works”, “Air t.”). (i.e., “Air Traffic”), “Entertainment”, “Children”, “Music”, “Road”, “Rail”, “Marine”, “Alarms”, “Industrial”, “Water”, “Animals”, “Vegetation”).
Once this optional information has been validated, the user has access to a summary of the measurement in the “Results” screen (Figure 1). Acoustic indicators are specific to the evaluation of outdoor sound environments, based on 1 s average sound level [24], such as noise levels in percentiles ( L A 10 , L A 50 and L A 90 ), maximum (Max) and minimum (Min) values as well as the average sound level in dB(A) and the average spectrum over the measurement time. In addition, a graphical representation, noted RNE, shows the distribution of the 1-second noise levels.
A “Map” can also be displayed on a specific screen (Figure 1) in order to locate the measurement points and to represent the average values shared by the user community and aggregated by the NoiseCapture remote server. Other functionalities are also offered by the application, such as smartphone calibration and data archiving, but are not detailed in the present paper. More details are given in the following reference [18].

2.2.2. NoiseCapture Web Interface

While the NoiseCapture application can be used to meet the need of a user (i.e., to assess a noise level in his own environment), the overall interest of the approach lies in the sharing of data within a community, which requires the data to be centralized on a remote server. To this end, a Spatial Data Infrastructure (SDI), called OnoMap, has been specifically implemented to propose 3 functionalities: to (1) collect, (2) display, and (3) share all the data produced by the contributors [20,25,26].
The second functionality is the most visible part of this SDI, as it allows to display the collected data to any visitor of the website, in an aggregated and understandable form. Figure 2 illustrates an example of a graphical representation, centered on the city of Lyon in France. Depending on the zoom scale of the map, the main window presents the collected data either in a numerical form, in terms of number of points per geographical area (represented by hexagons of different sizes depending on the zoom level), or in the form of a ‘classical’ noise map. In the latter representation, only certain acoustic indicators are presented, by aggregating all the values collected over a fixed spatial extent (i.e., average of an acoustic indicator in a hexagon). The left-hand side of the web page gives access to additional contents, such as the history of the last 30 series of measurements (almost in real-time), general statistics on all the data collected (most contributing countries, number of measurements, most used tags, etc.). By clicking on a hexagon on the noise map (at the highest zoom levels), it is also possible to access to more detailed information, such as the number of points and the total duration of measurements in the corresponding hexagon, the average equivalent sound level ( L A , eq and L A 50 ), the tags used (in the form of a tag cloud), as well as the hourly distribution of sound levels on different days of the week. All the information presented in this web page results from a direct exploitation of the NoiseCapture database, and illustrates some relatively simple analysis. Downloading the collected data (the third functionality of the SDI) offers many more perspectives of analysis and representation of the data. The upper screenshot of Figure 2, which displays the position of the measurement points, underlines again the interest of the method and the very rich perspectives of analysis of the sound environments, with regard to the quantity of data that can be collected on a given spatial extent. The purpose of this article is precisely to propose a first analysis of these raw data, in Section 3, so that they can be exploited, in a second step, to perform a relevant sound environment analysis.

2.3. NoiseCapture Raw Database

The analysis that is carried out thereafter covers the data collected since the official publication of the application on 29 August 2017 until 28 August 2020 (3 years). These collected data cover several releases of the application (from 28 to 51, see Table 1), some of which make changes to the nature of the data collected and the features. Previous release (before release 28) and intermediate pre-releases correspond to beta versions, published on Google Play to a specific panel of testers. The database available for download may include data from beta and pre-release versions; it may be useful to filter these data both on a period (from the launch of the application) and on the version (since release 28), for a relevant analysis.

2.4. NoiseCapture Installs and Uninstalls

As mentioned above, the initial audience targeted during the development of the application was primarily technical staff, with sufficient expertise to be able to use the application in satisfactory conditions (compliance with a measurement protocol, acoustic calibration of the smartphone, critical analysis of the measurements, etc.). The production of data was therefore initially part of a supervised activity with a professional purpose. In practice, the publication of the application on Google Play, combined with an institutional communication, was relayed by the national and then European media, generating the interest of a wider public than initially foreseen. Very quickly, the application was then downloaded in other countries, notably the United States, by a large audience. This confirms once again the interest of citizens and communities in the issue of noise environments and reaffirms the major societal challenge of research on this subject.
Figure 3a illustrates the number of installs of the application for the two countries (US and FR) that contribute the most to the data collection today; these data are obtained from the application dashboard on Google Play. This figure clearly shows the impact of the launch of the application in France, with a high number of installs in the first few weeks, followed by a decrease to an average level of about 60 installs per week; conversely, there is a gradual increase in the number of installs in the US, to an average level of 800 installs per week. From a global point of view, Figure 3b shows a certain stability around the 1000 weekly installs worldwide, over most of the period concerned. Unsurprisingly, the uninstalls rate follows the rate of installations, but the number of uninstalls tends to exceed the number of installations since the end of 2019, which leads to a decrease in the number of active devices (i.e., devices having installed the application and being turned on over a 30-day period), which has gone from about 17,000 at the end of 2019 to 13,000 at the end of 2020.
Even if it is difficult to make a direct link between the number of installations and the number of different contributors, we can see that on average, about 50% of new installations give rise to at least one contribution on the NoiseCapture server over the period from 2017 to mid 2019 (Figure 4). The ‘break’ that is visible on this figure in mid-2019, due to a very sharp drop in the number of contributors (Figure 3b), is at this stage undetermined. Conversely, Figure 3b shows the interaction between certain events quite well, in particular the impact of the publication of a new release on the number of new installs (also visible on the number of active devices). For example, the decrease in number of installs and contributors observed from the beginning of 2020 coincides with the beginning of the COVID-19 pandemic, particularly visible in the US community (Figure 3a); this is not visible with the French data, but we can quite imagine that there is a link between these two events. A detailed analysis of COVID-19 lockdown and user behavior in each country would undoubtedly lead to some hypotheses. This shows again the interest of such alternative way for collecting data for the study of the noise environment.

3. Analysis of the Collected Data

3.1. Collected Data

As mentioned above, the statistical analysis on NoiseCapture data presented in this section involves data collected from 29 August 2017 to 28 August 2020 (3 years of data). During this period, the NoiseCapture application has proven to be successful to perform and gather acoustics measurements. NoiseCapture has been downloaded more than 160,000 times on Google Play [23], with 76,229 contributors to the database all over the world. Approximately 91.7% of the users present in the database (69,898 of the 76,229 contributors in the present database) have contributed within this period. Table 2 shows that 260,422 tracks (59,685,328 points) have been collected, with an average of 229.2 points (i.e., seconds) per track (a median value of 28 points per track).
All the collected data during the corresponding period of analysis have been integrated into a spatial relational PostGIS database [29] (i.e., a spatial database extender for PostgreSQL object-relational database [30], adding a support for geographic objects). The database is fully available for download [31] and can be used according to the OdBL license [32]. It is important to specify that all the data integrated in this database fully respects the privacy of users as no personal data is collected.
The data collected from smartphones are organized into several tables (Figure 5):
  • For each measurement ‘Point’ (i.e., a measurement performed every second during a ‘Track’), the global ‘noise_level’ value measured at the measurement date ‘time’ is given in the ‘noisecapture_point’ table. In addition, the ‘speed’ at the measurement point, the geolocalization (‘the_geom’), the date of the localization (‘time_location’), the ‘accuracy’ of the geolocalization as well as the smartphone ‘orientation’, all obtained by the smartphone GPS, are given. In this table, the measurement point is defined by a primary key ‘pk_point’ (generated by the database) allowing to make the relation with two other tables ‘noisecapture_freq’ and ‘noise_capture_track’ (via the primary key ‘pk_track’);
  • The ‘noisecapture_freq’ table contains for the measurement point defined by the primary key ‘pk_point’, the ‘noise_level’ spectrum by third octave band ‘frequency’ between 100 Hz and 16 kHz;
  • The ‘noisecapture_track’ table contains all the information associated with a measurement corresponding to a set of measurement points. Each measure is defined by a primary key ‘pk_track’ (generated by the database) and a unique identifier ‘track_uuid’ (generated by the application). Each measurement contains the following information: the user primary key ‘pk_user’, the release number of the application ‘version_number’, the characteristics of the smartphone (the reference ‘device_product’, the model ‘device_model’ and the manufacturer ‘device_ manufacturer’), the date of the start of the measurement ‘record_utc’, the duration ‘time_length’ of the measurement, the average sound level over the duration of the measurement ‘noise_level’ and the perception of the sound environment ‘pleasantness’. Information on the acoustic calibration of the smartphone is also associated with the measurement: the choice of the calibration method ‘calibration method’ and the corresponding calibration value ‘gain calibration’. Finally, if the measurement was performed during a NoiseCapture Party (see Section 3.6), the corresponding code is indicated in the value ‘pk_party’.
  • The ‘noisecapture_user’ table gives for each primary key ‘pk_user’, the user identifier ‘user_uuid’ (this unique identifier is randomly created each time the application is installed on a smartphone), the user creation date ‘date_creation’ (created by the remote server when uploading the data, not at the application installation), as well as the user ‘profile’ defined by the choice of a value in a list, as ‘EXPERT’, ‘NOVICE’, and ‘NONE’. The value ‘pseudo’ in the table has been created for future functionalities and is currently not used.
  • The ‘noisecapture_track_tag’ table contains for each measure defined by the primary key ‘pk_track’, the list of tags selected by the user to describe the sound environment. The identifiers of the corresponding tags are defined in the value ‘pk_tag’. The correspondence between the identifier of the ‘pk_tag’ tag and the name of the tag (‘tag_name’) is defined in the ‘noisecapture_tag’ table.
  • The ‘noisecapture_party’ table contains information about the realization of the NoiseCapture Party events [33] (see Section 3.6 for details). In principle, such event is supervised by an expert, over a limited duration and spatial extent, allowing to generate a series of measurements. It can for example be an action carried out by a Community in order to carry out a series of measures concentrated in a particular district. A NoiseCapture Party has much the same objectives as an OpenStreetMap (OSM) Mapping Party to feed the OSM global database [34]. This table gives for each NoiseCapture Party, a specific primary key ‘pk_party’ (generated by the database) returning the code of the NoiseCapture Party (‘tag’), the title ‘title’ and a description ‘description’, the spatial extent defined by a geometry ‘the_geom’, the start and end dates of the event ‘start_time’ and ‘end_time’. The boolean values ‘filter_time’ and ‘filter_area’ are used to define whether the collected data are integrated into the NoiseCapture Party set, whether or not the measurements have been made with the right NoiseCapture Party code, but outside of the temporal and and spatial limits. The value ‘layer_name’ is only used to give a name to the corresponding map layer in the web page displaying the data on the corresponding website [26]. It is important to specify that the NoiseCapture Party is technically created by the people in charge of the development of NoiseCapture. If an invalid value is used for the NoiseCapture Party code field in the ‘Description’ screen of the application (Figure 1), the code is removed, but the corresponding data are still included in the database.

3.2. User Information

3.2.1. User Profile

The use of the application according to the respect of technical procedures in acoustics is an important issue for the quality of the produced data. In order to have information on the user experience, at the installation step of the application, the user is asked to define his expertise using a 3 levels scale: ‘EXPERT’, ‘NOVICE’, or ‘NONE’. Analyzing the 76,229 different contributors in the database, over the period from 29 August 2017 to 28 August 2020 (using the field ‘date_creation’, which corresponds to the date of creation of the user in the NoiseCapture database on the remote server), 10.19% defined themselves as ‘EXPERT’, 24.78% as ‘NOVICE’, and 64.17% as ‘NONE’. A very large majority of contributors therefore have no experience in the field, which can necessarily lead to a bias in the quality of the data collected. This is an expected behavior for a citizen science project.
Note that for 653 contributors (0.86% of the total number of contributors), the profile field is empty, meaning that the information is not available during this period. This is due to an update of the application from a version prior to version 28 (the field ‘user_profile’ has been integrated from version 28), as the user is not asked to modify this field during an update. In detail, the analysis of these cases shows that most of the concerned contributors (641) were declared in the database in the first 2 months after the launch of the application, while the other 12 contributors were declared during the rest of the period.

3.2.2. User Devices

For this type of measurement application, the metrological quality of the device, whether for acoustic measurement or for other data (GPS and other sensors), is an essential aspect. On this point, the identification of the smartphone can provide useful information for a later analysis of the collected data, in postprocessing. Among the possible treatments, an a posteriori calibration of the acoustic data, for example, based on a smartphone knowledge base, offers interesting prospects for improving the quality of the acoustic indicators produced by the application [35,36]. Some works have also shown that the knowledge of the manufacturer can provide a useful information on the accuracy of the measurement [37]. This justifies the need to collect hardware-related information, namely, the ‘device_product’, the ‘device_model’, and the ‘device_manufacturer’, defined by Android documentation as the name of the overall product, the end-user-visible name for the end product and the manufacturer of the product/hardware respectively [38].
As an example, considering the Samsung Galaxy A10, which is one of the best selling Android phone, the device field will give the data of Table 3. This table shows that the commercial name of the corresponding smartphone can be declined in several device models that most of time refer to distinct version (’A10E’ for ‘SM-A102’, ‘A10’ for ‘A105’, ‘A10S’, for ‘A107’) or to the international region where they were deployed.
Over the period in question, the database references 646 distinct manufacturer names. However, the same manufacturer can appear under a different spelling; this is the case, for example, for Samsung, appearing with the following names: ‘samsung’, ‘Samsung’, and ‘SAMSUNG’. By grouping the manufacturers without taking into account the sensitivity to upper and lower case, one can identified finally 520 manufacturers (Table 4) with 5300 different smartphone models. Nevertheless, three manufacturers alone (Samsung, LGE and HUAWEI) account for about 35.2% of the models, and cumulate nearly two thirds of the tracks (65.1% or 66.3% in number of points). The top 15 manufacturers account for 90.3% of the tracks (91.1% of the points).
The distribution of measurements is more important in number of models (Table 5), as the top 15 models only have 15.9% of tracks (16.7% of points), each model accounts to only between 1.8% and 0.8% of the whole measurements. To reach half of the tracks, we have to consider 130 different models, and 1077 models to exceed 90%. In addition, Figure 6 shows that most of devices appears only few times in the database; for example, 3407 different devices are used 10 times or less; conversely, only 775 device models appear more than 50 times in the database. In detail, there are 1228 smartphones that are used only once and 677 twice.
By focusing on the two most contributing countries (US and FR, Table 6), we find a consistency between the manufacturers market share and the brands most represented in the database. This table also shows that Apple with the iPhone model (Apple and iPhone are trademarks of Apple Inc.) is a very important manufacturer in the US and in France; it suggests that the current NoiseCapture database excludes a very large number of users, that in the case of iOS users, represents a specific segment of the population, considered with higher income and education levels, in-app engagement [39]. The initial choice to select Android as the only development platform, as it represents a global market share of 80%, can thus be questioned. It would seem wise to consider an additional iOS version of the application in the future, considering the user audience, but also metrological considerations.

3.2.3. User Contribution

Table 7 illustrates the use of the application in terms of number of contributions. Slightly more than half of the contributors have contributed to the database only by 1 track, and nearly 95% by less than 10 tracks. It is likely that most of the contributors concerned by only few contributions were just interested by testing the application, before either uninstalling it or putting it aside. This table also shows that there is a small proportion of contributors who have collected a very large number of tracks, up to several thousand for some. It seems obvious that these contributors have integrated themselves into an active approach to collect measurements and that this type of user is the most interesting part of the community, a priori motivated by the collaborative approach. The animation of this specific community must be a priority in the future. This last point will be discussed in Section 4.
Considering the contributors with only one contribution, Figure 7 shows that 6155 of them (16.9%) have used the “test” tag, meaning that they were just testing the application. In addition, Table 8 shows that 14,034 (38.5%) of these “one-shot” use of the application have duration less than 20 s. These two observations may partially support our hypothesis that these one-shot contributors just want to test the application, and probably do not plan to use the application again.
Table 9 shows that for users that realize more than one contribution, the second contribution comes in the next 4.4 days, on average. However, for the major part of the contributors (158,631, 83.3%), the second contribution is realized in the same day, and in the same week for 9.9% (18,971).

3.3. Measurement Geolocalization

3.3.1. Geolocalization

In this paragraph, we present statistics and information related to the geolocalization of the NoiseCapture data. The variable ‘the_geom’, which gives the coordinates of the measurement point in the WGS 84 (EPSG:4326) map projection, has been used to perform this study.
As the country of the measurement is not in the data set, the following 2-step process has been carried out in order to define the country of origin of each measurement. First, a table called ‘noisecapture_track_frame’ was created using the PostGIS/PostgreSQL function ‘ST_EXTENT()’ [41] that returns a box that bound each track. Second, a table called ‘gadm’, mapping the administrative areas of all countries [42], has been used to create a table called ‘noisecapture_country_track’ by associating each track to the first country that contains the track bounding box.
Table 10 shows that the United States contributes more than third of NoiseCapture database, while France contributes approximately 10% of track data (8.3% point data). A strong French contribution was obviously expected, the application having been developed by French research institutes, and also because of a strong relay by national media. Conversely, it is difficult to explain the large amount of data produced by the US, except to consider that this country has a high population (3rd in the world in 2020, [43]), compared to France (22nd in the world). In addition, because some countries do not have access to Google Play or use alternative app stores, and since the NoiseCapture application is only available on Google Play store, it is not surprising that they are not found as a data producer. This is the case of China (Google Play not available in China) and Russia (an alternative app store is mainly used), for example, while they represent an important part of the world’s population (1st and 9th in the world, respectively).
Although the ranking of Peru and the Netherlands in terms of population is low, these two countries are in the top 5 in terms of number of tracks (Table 10). For these countries, the number of tracks compared to the number of contributors is very high (especially for Peru), which highlights an intensive measurement activity, which is perhaps part of a voluntary and organized action (like a NoiseCapture Party for example). A spatio-temporal analysis of the data produced in these two countries, as well as a detailed analysis of the behavior of the corresponding contributors, could eventually provide some answers. More globally, the implementation of cluster detection techniques could be an interesting way to identify organized events.
While conducting the study, it was observed that 32.5% of the tracks contain points without geolocalization (i.e., the field the_geom is empty), which represents a total of 10,783,609 points (18% of the total number of points), distributed over 141 countries. For 75% of these tracks (63,538 tracks), all the corresponding points are concerned by a lack of geolocalization (i.e., the whole track can not be geolocalized). The main reason is that the geolocalization has not enabled on the smartphone. By further analyzing, it was also observed that a large part of these tracks correspond to indoor measurements (20,045 (7.7%) of the corresponding tracks are defined with the ‘indoor’ tag). Tracks with a partial lack of geolocalized points may be due to a local loss of GPS localization, for example, when passing through a tunnel. A spatial analysis crossed with other geographical data can possibly bring elements of answer in this case.
Even when the measurement points are localized (i.e., the GPS actually transmits a position), this measurement can have a poor accuracy. Putting aside the technical quality of the hardware used in the smartphone for GPS location, this poor accuracy may be obtained when the measurement is made in an environment that is not clear enough (in or near a building, overcast sky) making it difficult to connect to a sufficient number of GPS satellites.

3.3.2. Accuracy

The ‘accuracy’ data collected by the NoiseCapture application allow one to associate a location accuracy (in meters) to each measurement. This value is obtained using the getAccuracy() function in Android [44], meaning that there is a 68% probability that the true location is inside the circle (with a radius equal to the value of the ‘accuracy’) centered at the corresponding location. The analysis of this parameter shows that the median value of ‘accuracy’ is around 8 m. It should also be mentioned that it is possible to find some measurement points with non-realistic accuracy values (such as 1.1 × 10 5  m) that may due to a wrong technical implementation of the GPS algorithm in the smartphone. For points with geolocalization, Table 11 shows that most of accuracy are under 25 m (42,313,601 points, 86.5%), which can be considered as a relevant accuracy for noise studies [45], and 35,184,828 (71.9%), 18,069,523 (36.9%), and 420,150 (0.8%) under 15 m, 5 m, and 1 m, respectively. Finally, one can observe from Figure 8 that the accuracy tends to increase (i.e., the accuracy value decrease) when the measurement duration increases. This is due to the fact that a sufficient duration may be required for the GPS receiver within the smartphone to detect GPS satellites, and then to obtain the best accuracy of location. It suggests that a NoiseCapture user should wait few seconds after starting the application (for example, ref. [14] mentions a duration of 4 seconds), before performing a measurement, in order to obtain the best geolocalization.
Last, it must be mentioned that the function getAccuracy() returns the value 0.0 when the smartphone is not able to obtain a value for the accuracy. This should not be consider as an accuracy value of 0 m.

3.3.3. Speed

The NoiseCapture data set contains information about the ‘speed’ value, which represents the speed (in meter per second) measured by the smartphone GPS at the time of the measurement point. Table 12 shows that 38.5% of tracks (65.4% of the measurement points) have a speed equal to 0. According to the Android documentation for the getSpeed() function [46], a null value is returned when the location does not have a speed; it does not mean that the speed is equal to zero, but that it is not possible to evaluate its value, even when the measurement is geolocalized. Note also that it could be interesting to also collect the estimated speed accuracy using the Android function getSpeedAccuracyMetersPerSecond() in a future release of the application.
When the speed value is different greater than 0, it means that the measurement point is moving, but it may be difficult to determine in a simple way the transportation mode that is used (walking/running, bicycle, light vehicle, public transportation, etc.), with the only knowledge of the speed value, as the speed ranges corresponding to each transportation mode may overlap [47]. Assuming people walking at a speed ranging from 0.5 km/h to 5 km/h (0.14 to 1.4 m/s, respectively), one can consider that around 14.4% of the measurements are realized during walking. One can also find speed values that correspond very clearly to measurements carried out in fast mode of transportation, including air transportation of the order of 280 m/s.
Additional analysis of the ‘speed’ information also shows several anomalies, such as negative values (26,304 measurement points, 0.04%), with the ‘−2’ (26,257 meas.) or ‘−1’ (14 meas.) values; such values may probably have a signification, but this information is missing in the Android documentation. Other negative values (33 meas) are in the range 1 , 0 and may be due to numerical accuracy.

3.4. Temporal Characteristics of Measurements

3.4.1. Measurement Timestamp

As already mentioned, the analysis developed here concerns an extraction of the database, as the official launch of the application over a period of 3 years, from 29 August 2017 to 28 August 2020 (considering all versions of the application since number 28). At the time of a measurement, the beginning of a ‘track’ is defined by the field record_UTC (given by the smartphone) and each ‘point’ of a ‘track’ is defined by the field time_location (given by the GPS).
The analysis of the entire database (i.e., between the date of the track record_UTC and the date of the first point time_location in the corresponding track), shows some measurements that are visibly incorrectly time-stamped; this corresponds to points without geolocalization (defined with the time_location=’1970-01-01’ by default). One can also observe measurements (21,897 tracks, 849,128 points) with a time shift of several hours (Table 13), but it represents less than 2% of the total number of tracks. Last, it is also possible that some users use date and location metadata scrambling tools on their smartphone to avoid tracking. The number of tracks/points concerned being however very low, one can imagine that the database analyzed here is little or not at all concerned by this type of error.
Figure 9 illustrates the distribution of the tracks in function of the hour of a day. For the entire database (Figure 9a), one can observe a moderate variation from one hour to another, which can be explained by the fact that measurements are collected in all the time zones simultaneously (assuming that measurements are done all over the word simultaneously). When focusing on the data collected in France only, Figure 9b shows, as expected, a small number of measurements during night and early morning and more measurements during day and afternoon.
In addition, Figure 10 illustrates a small variation from one day/month to another, except for ‘October’ with more tracks for the year 2018, due to an unusual and large amount of data (8470 tracks) collected on 8–9 October 2018, by few users only, localized in Peru (6326 tracks, 78,789 points). This can be due to a specific event.

3.4.2. Measurement Duration

The ‘time_length’ data (in second) are defined from the start of the measurement during a track, until the user ends the measurement. Figure 11 shows that most measurements are done with a track duration around 1–20 s (44.6%) and 1–3 min (17.3%). Only 6.6% of tracks have duration greater than 10 min. The 10 s duration corresponds to a large part of the measurement (51,098 tracks 19.6%); this is due to the fact that user can used a predefined duration, which is fixed to 10 s by default. One can note that measurement duration between 1 and 3 min has also as strong presence. Table 14 shows that only small percentage of user collecting tracks with long period move along the track (≤6.45%, when considering a minimum speed value of 0.5 m/s).
An issue can be observed regarding the total number of measuring points (59,685,328), which is greater than the sum of all the measurement track duration (∑time_length = 59,684,657), with a difference of 671 s (i.e., points). Analyzing this issue, one can observe that 82 measurement tracks (made by different users, with different devices, using different application release) have a number of points (i.e., seconds) that are not equal to time_length. In addition, one can mentioned that few points have been removed from their corresponding tracks. For now, the exact reason of such anomalies is still not defined, but may be due to an unusual behavior of the application or to numerical inaccuracies.

3.5. Smartphone Acoustic Calibration

The relevance of the collected acoustic measurements is largely based on the measurement protocol applied by the contributor as well as on the metrological quality of the smartphone. At this stage, concerning the first point, it is expected that the contributor follows the recommendations available in the application. No other information is included in the collected data in order to analyze if this measurement protocol is well followed (excepted for the calibration); this point will be discussed in Section 4. The second point has given rise to numerous studies in the literature, as it is a critical element.
It is indeed hoped that the user calibrates his/her smartphone before collecting measurements. Numerous studies, among them recent ones [48,49,50], have shown the need to make a correction on the values measured by smartphones, in order to get closer to those that would have been measured with a reference device, such as a sound level meter. However, from a statistical point of view, this condition is not as critical, since it can be expected that due to a large number of measurements collected by different smartphones, the results may statistically converge to the expected values. This hypothesis seems to be confirmed by Murphy and King, showing, that in the absence of calibration, the sound levels measured on average by a wide variety of smartphones are very close to the expected value, but at the expense of a large standard deviation. Acoustic calibration can however reduce the standard deviation.
In the NoiseCapture application, the calibration procedure consists in evaluating a correction (or a calibration gain, i.e., the ‘calibration_gain’ value) that will be applied on the input temporal signal before the postprocessing of all noise indicators, assuming a linear relationship both in frequency and in amplitude, which is of course questionable for some smartphones. Within NoiseCapture, several calibration methods are proposed and defined by the field ‘calibration_method’ in the database:
  • The most relevant solution (’Calibrator’ method) consists in using an acoustic calibrator, according to the classical rules for acoustic measurement. This solution requires an external microphone, connected to the smartphone, with a diameter compatible with the use of an acoustic calibrator. Note that using an external microphone can also improve the measurement accuracy in comparison with the internal microphone [51]; thus, this solution must be promoted to contributors. The ‘calibration_gain’ is then determined for a reference frequency and for a reference level (for example 94 dB@1 kHz) and applied, during measurements, to the entire temporal signal before processing;
  • Another method (’Reference’ method) is used to correct the sound level measured (overall or for a given frequency) by the smartphone using another measuring device (i.e., using a visual comparison), considered as a reference. The value of the gain ‘calibration_gain’ is obtained from an ambient noise measurement.
  • A third method (’CalibratedSmartPhone’ method) is used to calibrate one or more smartphones simultaneously, using an already calibrated smartphone as a reference. The procedure is fully automatic, controlled by the reference smartphone, and is based either on the measurement of the ambient noise or a pink noise generated by the reference smartphone.
  • Finally, a more recent method (’Traffic’ method) is based on the measurement of several pass-by of light road vehicles, which, by comparison with a statistical model of noise emission, makes it possible to estimate the correction to be made to make the measurement coincide with the expected statistical value [52].
The user can also directly change the value of the calibration gain in the application settings at any time. The default value of ‘calibration_gain’ is set to 0 dB as long as no calibration method has been applied, or as long as the user has not directly changed this setting. A change of this parameter will be considered as a ‘ManualSetting’ for the ‘calibration_method’ field.
The choice of the method is defined by the field ‘calibration_method’, but only since version N°49 (17 February 2020). If, since the launch of the application, several calibration methods were already available, the information on the choice of the applied method was not known and only the value of ‘calibration_gain’ was actually uploaded to the remote server. For database consistency reasons, all data collected using versions prior to version 49 of the application, the choice ‘None’ is affected to ‘calibration_method’, although a calibration method may have been used. However, since version 49, the choice ‘None’ is only affected when no calibration is performed.
Table 15 shows the distribution of the collected tracks according to the calibration method. As indicated above, the field that defines the choice of the calibration method, is only available since release 49. By analyzing the data collected before release 49, one can observe that 62,731 of the 241,532 collected tracks, i.e., ~26%, have a calibration value different from 0, meaning that the corresponding users have probably applied either a calibration method or a manual change of the calibration gain in the settings of the application.
From release 49, we can see that 7.9% of the tracks have been made by smartphones that have been calibrated (92.1% are defined by ‘None’ for the ‘calibration_method’), but for about half of them, the manual method has been applied. For these tracks, it is therefore difficult to determine how the value of the calibration gain has been evaluated. Among the other calibration methods, the one using the automatic procedure between smartphones is the most used, followed by methods using a reference and an acoustic calibrator. The traffic calibration method is the most recent and has generated little data so far. In the future, it will be important to highlight this last method, which is the only one that is able to calibrate a smartphone without the need for an external device, while offering sufficient accuracy.
The application of a calibration method is not enough to justify the quality of the measurements. The obtained value of the calibration gain (field ‘gain_calibration’) is also a very important information. Figure 12 illustrates the distribution of tracks according to this value and brings some comments. The presence of abnormally high (in absolute value), even extreme and aberrant values shows either a bad use of the calibration methods or a technical problem. The number of tracks collected with a calibration gain of zero (default value) globally reflects a lack of calibration, as it is unlikely that a smartphone is calibrated by default. Finally, we can see that 86.8% of the collected tracks have a calibration value between −10 and +10 dB, which seems rather realistic, but does not bring any certainty on the quality of the measurement.
Table 16 completes this first analysis by showing the distribution of calibration gain values according to the method (data collected since release 49). As expected, it can be observed that when a calibration method considered as ‘robust’ is applied (’CalibratedSmartPhone’, ‘Calibrator’ and ‘Traffic’), the gain is different from 0 dB, except for the ‘Reference’ method. There is also slightly less disparity in the gain values when a calibration method is applied. One can also observe that for the ‘Calibrator’ method, 90.7% of the calibration gain values are greater than 10 dB, which shows a different behavior than other methods. This may be a misunderstanding of the method, with some users attempting to calibrate their smartphone without a reference device.
The analysis of these calibration data, for example, in relation to the type of device and user profile, is in itself a separate study. The creation of a ‘validated’ database of calibration values for each smartphone model, for example, is an interesting prospect. However, this perspective study is beyond the scope of the present article, which, at this stage, only aims to present the data collected and their limits of use.

3.6. NoiseCapture Parties

As mentioned above, a NoiseCapture Party is a special event organized by a given organization, aiming to carry out measurements, generally over a limited time and a spatial area, for educational, scientific dissemination, or research purposes. The advantage of these events lies in the fact that the measurements are generally well ‘controlled’, and most of the smartphones have been previously calibrated. It can thus be considered that the collected measurements have a better quality compared to the other measurements in the database.
The list of all NoiseCapture Parties are given in Table 17, with the number of considered smartphones and the total of collected tracks and points. As expected, the ratio of calibrated smartphones is greater for NoiseCapture Parties. The total number of collected tracks and points represents 1.2% and 0.6% of all data in the database.
It must be specified that other similar events could have been organized, without having given rise to the creation of a specific tag, and without having informed the developers of the application. For example, this is the case for several recently published research works [53,54,55,56,57].

3.7. Soundscape Description

NoiseCapture allows users to complete the acoustic measurement with information about his/her own perception of the sound environment, using ‘tags’ (field   ‘noisecapture_track_tag’) for describing the noise environment and the noise source along the track. In addition, they can give an information of ‘pleasantness’ (field   ‘pleasantness’) by selecting a value (0 for ‘unpleasant’, 25, 50, 75, 100 for ‘pleasant’); this field may be empty if no value is selecting (default value).
The analysis of the tags can be particularly interesting to distinguish the measurement conditions; indeed in certain conditions, such as rainy or windy weather, the acoustic measurement may be distorted, and it is therefore interesting to have such information before analyzing the acoustic indicators. The information about the indoor/outdoor measurement is also interesting for people who would like to use the data to characterize indoor or outdoor sound environment, specifically. Last, the knowledge of the sound sources that are perceived and the evaluation of the pleasantness are also interesting data for researchers that study the notion of soundscape.
Both ‘Pleasantness’ and ‘Tags’ are supplementary info. Their use add beneficial information about the sound environment. Ideally, this information should be systematically provided by the contributors. However, Table 18 shows that only 17.5% of the tracks have both information, while 48.7% do not have any and 33.8% have either one of them (mainly the pleasantness with 30.2%). An independence test showed that there is a dependency between using both ‘Pleasantness’ and ‘Tags’: a participant using tag will use pleasantness more often.

3.7.1. Pleasantness

Figure 13a shows that most of tracks (205,416 (78.9%), equivalent to 46,623,131 points (78.1%)) are not associated with a value of pleasantness, meaning that the default empty value is not modified. Excepted for the level ‘50’, all other values are used quite uniformly. The over-representation of the level value ‘50’ can be explained by three possible reasons:
  • for the most part, users cannot judge the quality of the sound environment in a clear-cut way;
  • by default, the selection cursor is positioned on the value ‘50’ that can influence the user;
  • the user may be tempted to select the cursor, without however wanting to make a decision. Once the cursor is activated, it is no longer possible to go back and a value will be automatically validated.
The two last hypothesis can introduce a bias, suggesting that a more suitable selection mode should be proposed in a future release of the application.
The behavior of a contributor can also be analyzed in terms of his propensity to use all the possible values of the pleasantness scale (Figure 13b). This analysis shows that for 52,979 users, only 1 level is used; however, in detail, for 43,764 of them (i.e., 82.6% of 52,979), the ‘NULL’ value is used; 12,489 used 2 levels, etc. Few users therefore use the pleasantness scale, and even fewer use these different levels of the scale. In a future evolution of the application, it could be interesting to ‘motivate’ users to provide information, for example by ‘forcing’ them to give an answer, including a ‘don’t say’ answer.

3.7.2. Tags

In addition to the perception of their sound environment, users also have the possibility to specify the measurement conditions (4 tags) and the nature of the perceived sound sources (14 tags in four categories: human activity, transportation, natural, and mechanical activity). Table 19 gives a description of the tag fields in the database (‘pk_tag’ and ‘tag_name’) as well as the corresponding English description within the NoiseCapture application (see Figure 1). The list of ‘pk_tag’ is not continuous, some missing numbers correspond to tags that are no longer used since the first official release of the application.
Figure 14a shows the number of tags that are simultaneously used to describe a track. In about half of the collected tracks (124,363, 47.7%), the contributors do not use any tag to describe the measure. This is better than for the pleasantness evaluation.
Nearly 30% of the tracks contain 1 or 2 tags: when considering 1 tag only, 17,094 tracks (40.9%) are defined by an environment tag (‘test’ or ‘Indoor’); when considering 2 tags, 2061 tracks are defined by 2 environment tags, 9260 tracks by 2 source tags and 24,315 tracks by a combination of two types of tag. One can also note that a number of tracks simultaneously contains a large number of tags, even the 18 possible tags, which is not realistic. We can assume that the corresponding tracks are test measurements, but they do not necessary mention the ‘test’ tag. Figure 14b shows that this ‘test’ tag is used in 30,077 of the collected tracks, which is important. An analysis of the database, for the purpose of studying sound environments, will necessarily exclude the collected data with this tag.
The other interesting aspect is that the ‘Indoor’ tag is present in 58,967 of the tracks collected, which represents an interesting quantity for the study of sound environments in closed spaces (building, transportation), even though the initial objective of the application was to study outdoor environments. Note also, that the tags ‘Indoor’ and ‘Test’ are not independent, meaning that both tags can be used together.
Among the sound sources mentioned by the contributors, ‘voice’, ‘footsteps’, and ‘road’ are present, which is consistent with a contributor who collects measurements closed to road infrastructures, while walking and talking. Again, it is obvious that the analysis of these tags and their occurrence can provide interesting information on the perception of sound environments. However, this is beyond the scope of this article.

3.8. Noise Indicators

The purpose of the NoiseCapture application is based on the measurement of acoustic indicators for the analysis of sound environments. The data that are present in the NoiseCapture database concern the equivalent sound level L A , eq on a track, as well as the spectrum and sound level at each point of the track, measured every second. The postprocessing of these data can, in a second step, give access to percentile indicators (such as L A 10 or L A 50 ) or to sound level distributions, for example. In the following, the analysis is restricted to the data as such, and not to the sound environments.
First of all, it should be remembered that in terms of acoustic measurement, smartphone manufacturers under the Android OS must respect a number of recommendations defined in the Android Compatibility Definition [38]. In particular, they should offer (1) an audio capture with approximately flat amplitude and frequency characteristics of ±3 dB from 100 Hz to 4000 Hz, (2) an input sensitivity such that a 90 dB Sound Power Level (SPL) at 1000 Hz gives an RMS value of 2500 for 16-bit samples, and (3) a linear change of the amplitude over a range of at least 30 dB from −18 dB to +12 dB relatively to 90 dB SPL at the microphone. Some smartphones may offer superior features, but it is expected that all smartphones meet the minimum requirements.
Figure 15 shows the range of L A , eq values measured along the tracks and at each point of a track. While not visible in this figure, one can observe data with very low sound levels (a few decibels, even negative ones), which seems physically both unrealistic in a real environment, but also a priori outside the measurement capabilities of a smartphone. On the other hand, the highest levels are of the order of 125 dB, which is not unrealistic but, nevertheless, unlikely in a normal environment.
In details, Figure 15a shows that the noise levels measured on tracks can be represented as a mixture of 2 normal distributions (noted N ( mean , standard deviation ) ), a first group X 1 defined as N ( 107.4 , 3.1 ) and a second group X 2 defined as N ( 62.0 , 15.5 ) . These normal distribution are also respectively defined by a ‘gain_calibration’ of mean 73 and 0 dB (median 80 and 0 dB). Figure 15b shows similar results for the noise levels at the measurement points, as a mixture of 2 normal distributions, N ( 106.3 , 3.2 ) and N ( 50.9 , 17.2 ) with similar statistical values for the ‘gain_calibration’. For highest sound levels, it is quite evident that the calibration was not performed correctly.
This simple study shows that the range of variation of the measured sound levels is abnormally wide, the absence of calibration or a bad calibration being a probable cause.

4. Discussion and Future Developments

4.1. Synthesis

The analysis of the data collected during the first 3 years of operation since the launch of the application clearly shows that the information may be made of anomalies and uncertainties. Quantifying and reducing some of these biases is possible, either by a better knowledge and control of the user’s behavior and of the context of measurement or by improving the smartphone application.
Table 20 already proposes at this stage some simple modifications to implement within the application, mainly by checking some settings (verification of the smartphone date/time, activation of the geolocalization, user profile update, change of the ‘Pleasantness’ selection mode). Most critical aspects concern the lack of a good geolocalization and bias in the noise level measurements mainly due to a wrong or a lack of smartphone calibration. These two subjects are specifically discussed in Section 4.2 and Section 4.3. In addition, a better knowledge of the context of the measurement could also judiciously complete the collected data, or even replace certain user actions, such as the use of tags. Some suggestions will be proposed in Section 4.4. Last, increasing trust in the data also means increasing trust in the users. The animation of the community of contributors is another essential challenge. This is discussed in Section 4.5.

4.2. Increasing Localization Accuracy

In the current release of the application, localization is performed in an elementary way, and it was realized afterwards that it may not be sufficient depending of the objective of the use of data. With a view to improve the performance of the application (and therefore the quality of the data produced), the quality of GPS localization is a point on which the user must be made aware. In the application, this can for example take the form of recommendations to improve the quality of GPS location, such as activating the ‘High accuracy’ mode in the Android settings, re-calibrating the GPS via the use of a third party application, or activating additional localization functions via WiFi, Bluetooth and mobile networks.

4.3. Building a Smartphone Calibration Database

Some authors have rightly proposed to provide contributors with a database to calibrate smartphones, in order to limit bias during an acoustic measurement. Building such a database can be tedious because of the large number of smartphone models present on the market simultaneously, as well as their very rapid evolution. However, the NoiseCapture experimentation has opened new perspectives. Indeed, the analysis carried out in the present article shows that a large part of the contributions come from a limited number of manufacturers and models (three manufacturers account for about 35.2% of the models and nearly two thirds of the tracks; 15 models only have 15.9% of tracks). This information would limit the number of calibrations to be performed in the laboratory to build a calibration database. The other perspective would be to use the calibration data proposed by the contributors for their smartphone. The quality of this calibration can however be discussed, except for contributors performing their calibration during a NoiseCapture Party type event.

4.4. Collecting Information about the Context Awareness

From the very beginning of the application creation, the kind of the information sent back to the remote server was deliberately restricted to what was strictly necessary, so that it would not be considered as invasive. The study of the data collected over 3 years nevertheless shows that their use in a better controlled scientific approach would require additional information.
In particular, information on the context of the measurement, such as wind detection, activity recognition, transportation mode detection, how the smartphone is used during measurements, and place recognition [47,58,59,60,61,62] could be useful. The use of information provided by other smartphone sensors (accelerometer, orientation, brightness sensor, and proximity sensor) could also provide information on the process of the measurement [14,17]. Note also that, as mentioned in [47], specific functions are already available in Android API to identify some user activities [63], which could be a first attempt to obtain new information. To a lesser extent, it may also be interesting to collect the speed accuracy (adding a new data ‘speed_accuracy’), since this value is also available in Android API.
Providing that smartphones have sufficient resources, the integration of sound source identification algorithms can also give interesting additional information, and can advantageously replace the use of tags [64]. Otherwise, it should also be possible to include such identification as a postprocessing on the remote server, for example, by using the collected 1 s spectra. All these development perspectives must nevertheless be integrated in the total respect of the privacy of the contributors [65], in particular, in the respect of laws in specific Regions/Countries, such as the General Data Protection Regulation (GDPR) in Europe [66].

4.5. Increasing and Animating the Community of Contributors

The participatory approach is of course the main originality of the application, allowing one to considerably multiply the number of measurement points, with a large variability in time and space. Like any participatory approach, the main challenge is to maintain the initial interest of the contributor to support a research project or make their individual contribution a major social issue (i.e., noise pollution) [67], beyond a time of discovery and a few measurements. The analysis of installs/uninstalls detailed in Section 2.4 shows indeed a tendency to a negative imbalance between installs and uninstalls of the application, which suggests that it is important to propose a solution to better retain users. Moreover, the analysis of the contributors behavior in Section 3.2.3, shows that there are finally few active contributors (half of the contributors made only one measurement, mostly to test the application), and that almost half of the contributions do not exceed 20 s (38.5%, Table 8). It is therefore require to develop strategies that allow for the development of a community of very active contributors.
This must be achieved by enhancing the application in order to motivate users to regularly produce measures, for example by adding reminder notifications to contributors or by developing a more playful aspect (creation of pseudonyms, setting up a challenge or a serious game based application such as noise battle or noise quest [68], creating badges…). If the target is more oriented towards a community of professionals (i.e., the initial target), the animation of the community can be more distributed, by calling upon ‘ambassadors’ (teachers, student researchers, technical agents of communities, government services…) who will see a particular interest in organizing, for example, NoiseCapture Parties. As also mentioned by others authors [48,67,69,70], the advantage of organizing ‘controlled’ events lies in the possibility of training users to carry out measurements using a validated protocol, particularly from the point of view of the calibration of smartphones, which would increase ‘confidence’ in the measurements. Once trained, users could in turn train other users, increasing the ‘trusted’ community.
Implementing a serious game type application, or increasing interactions with contributors, should also encourage the contributor to take measurements in specific spaces and at specific times. As considered by the authors of [14], this would make it possible to compensate for a lack of measurements in certain places or at certain times.
The analyzing of the geolocalization of the measurement points also showed a inhomogeneity in the diffusion of the application throughout the world, mainly due to the initial choice of the development platform (Android) and the associated application stores (on Google Play only). If one can observed the very wide use of the application throughout the world offers (currently, the application has been used in 204 different countries), this analysis shows the need to disseminate the application even more widely in order to acquire data in some countries with large populations. This could offer a wealth of data that is particularly interesting from the point of view of evaluating sound environments in countries with very different cultures and environments.

5. Conclusions

The use of a crowd-sourcing type approach offers interesting perspectives in the analysis of sound environments, in particular because of the spatial extent and the temporal dynamics that the data collected can provide. The involvement of citizens in a collaborative approach also brings another dimension to scientific research on the subject. The initial and legitimate fears about the relevance of using such data in an environmental approach (evaluation of public policies to reduce noise nuisance, effects of noise on health, perception of noise environments) are being allayed. Studies have indeed shown the relevance of this type of approach [67,69], while underlining some important points, for example, users proactivity, critical mass of contributors, increasing of measurement accuracy or the need of organizing collective sessions of noise sensing, etc.
The development of the NoiseCapture application is fully in line with this alternative approach. Compared to similar approaches, however, the NoiseCapture approach offers a completely open source platform, ensuring total transparency on the methods of collecting and processing data, and giving the possibility to everyone to freely use the data. The sustainability of the approach was also considered, by making effort to ensure the functioning of the project over time. These specificities are certainly the reasons for the success of the approach, whether it be with many communities.
Since the launch of the application on 29 August 2017, the amount of data collected is considerable. After 3 years of operation, thanks to the participation of 74,082 contributors, the database has accumulated 260,422 tracks and 59,685,328 one-second measurement points, spread over 204 different countries (Figure 16). To our knowledge, there is no other similar experimentation.
Although the amount of data collected is considerable, any exploitation of the database for applications related to the study of sound environment requires a perfect understanding of the data, in order to limit bias in the analysis. The objective of this article was therefore to review all the data collected (nature, content, limits, etc.) and to identify specific behavior linked to the use of the application (Section 3). This analysis now provides a precise framework for the further exploitation of the data. In view of the very large amount of data collected, it is however clear that depending on the nature of the expected analysis, a large part of the data cannot be used, either because it does not present any interest for the corresponding analysis, or due to a lack of completeness and accuracy. As discussed in Section 4, in our opinion, enhancing/controlling the quality of the data and of the measurement conditions constitute two major developments for improving the database. The other major perspective consists in the animation of the contributors community to increase confidence in the data.
Thus, as soon as attention is paid to the inherent limits of the collected data, the exploitation of this database offers very interesting perspectives on the characterization of sound environments. Any relevant analysis could be useful for communities to assess the noise environment of their territory, and usefully complement regulatory requirements, such as the 2002/49/CE Directive, in Europe, relating to the assessment and management of environmental noise. As an example, a very simple analysis of the sound levels collected in France shows, without any particular treatment, an overall decrease in sound levels during the periods of lockdown related to COVID (Figure 17).
Beyond the exploitation of the database, one can also mention that the use of the NoiseCapture application with a dedicated use of the collected data (i.e., without using the NoiseCapture database, but using only the data export capabilities of the application) can be an interesting tool for scientific purposes [53,54,55,56,57].

Author Contributions

Conceptualization, J.P., E.B., G.P. and N.F.; methodology, J.P. and A.B.; software, N.F. and G.P.; formal analysis, J.P. and A.B.; investigation, J.P. and A.B.; writing—original draft preparation, J.P. and A.B.; writing—review and editing, J.P., A.B., E.B., G.P. and P.A.; visualization, A.B. and N.F.; supervision, J.P. and E.B.; project administration, J.P. and E.B.; funding acquisition, E.B. All authors have read and agreed to the published version of the manuscript.

Funding

This research was initially funded in the framework of the ENERGIC-OD Project (European Network for Redistributing Geospatial Information to user Communities—Open Data), under the ICT Policy Support Programme (ICT PSP) (CIP-ICT-PSP-2013-7) as part of the Competitiveness and Innovation Framework Programme by the European Community.

Data Availability Statement

The data presented in this study are openly available from Université Gustave Eiffel Dataverse Repository at https://0-doi-org.brum.beds.ac.uk/10.25578/J5DG3W.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
APIApplication Programming Interface
CNRSCentre national de la recherche scientifique (French National Centre for
Scientific Research)
ERDEntity relation diagrams
GPSGlobal positioning system
ICTInformation and communication Technologies
IfsttarInstitut Français des sciences et technologies des transports, de l’aménagement et
des réseaux
OSMOpenStreetMap
RNERepartition of the noise exposure
SDISpatial data infrastructure
SPLSound Power Level

References

  1. Directive 2002/49/EC of the European Parliament and of the Council of 25 June 2002 Relating to the Assessment and Management of Environmental Noise—Declaration by the Commission in the Conciliation Committee on the Directive. 2002. Available online: http://data.europa.eu/eli/dir/2002/49/oj/eng (accessed on 26 April 2021).
  2. Picaut, J.; Can, A.; Fortin, N.; Ardouin, J.; Lagrange, M. Low-Cost Sensors for Urban Noise Monitoring Networks—A Literature Review. Sensors 2020, 20, 2256. [Google Scholar] [CrossRef]
  3. Camprodon, G.; González, O.; Barberán, V.; Pérez, M.; Smári, V.; de Heras, M.A.; Bizzotto, A. Smart Citizen Kit and Station: An open environmental monitoring system for citizen participation and scientific experimentation. HardwareX 2019, 6, e00070. [Google Scholar] [CrossRef]
  4. Aiello, L.M.; Schifanella, R.; Quercia, D.; Aletta, F. Chatty maps: Constructing sound maps of urban areas from social media data. R. Soc. Open Sci. 2016, 3, 150690. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Gasco, L.; Clavel, C.; Asensio, C.; de Arcas, G. Beyond sound level monitoring: Exploitation of social media to gather citizens subjective response to noise. Sci. Total. Environ. 2019, 658, 69–79. [Google Scholar] [CrossRef] [PubMed]
  6. Mydlarz, C.; Drumm, I.; Cox, T. Application of novel techniques for the investigation of human relationships with soundscapes. In Proceedings of the INTER-NOISE and NOISE-CON Congress and Conference Proceedings, Osaka, Japan, 4–7 September 2011; Volume 2011, pp. 738–744. [Google Scholar]
  7. Santini, S.; Ostermaier, B.; Adelmann, R. On the use of sensor nodes and mobile phones for the assessment of noise pollution levels in urban environments. In Proceedings of the 2009 Sixth International Conference on Networked Sensing Systems, Pittsburgh, PA, USA, 17–19 June 2009; pp. 1–8. [Google Scholar] [CrossRef]
  8. Rana, R.K.; Chou, C.T.; Kanhere, S.S.; Bulusu, N.; Hu, W. Ear-phone: An End-to-end Participatory Urban Noise Mapping System. In Proceedings of the 9th ACM/IEEE International Conference on Information Processing in Sensor Networks, Stockholm, Sweden, 12–16 April 2010; pp. 105–116. [Google Scholar] [CrossRef]
  9. Kanjo, E. NoiseSPY: A Real-Time Mobile Phone Platform for Urban Noise Monitoring and Mapping. Mob. Netw. Appl. 2010, 15, 562–574. [Google Scholar] [CrossRef] [Green Version]
  10. Maisonneuve, N.; Stevens, M.; Ochab, B. Participatory Noise Pollution Monitoring Using Mobile Phones. Info. Pol. 2010, 15, 51–71. [Google Scholar] [CrossRef] [Green Version]
  11. Duda, P. Processing and Unification of Environmental Noise Data from Road Traffic with Spatial Dimension Collected through Mobile Phones. J. Geosci. Environ. Prot. 2016, 4, 1. [Google Scholar] [CrossRef] [Green Version]
  12. Guillaume, G.; Can, A.; Petit, G.; Fortin, N.; Palominos, S.; Gauvreau, B.; Bocher, E.; Picaut, J. Noise mapping based on participative measurements. Noise Mapp. 2016, 3, 140–156. [Google Scholar] [CrossRef] [Green Version]
  13. Radicchi, A.; Henckel, D.; Memmel, M. Citizens as smart, active sensors for a quiet and just city. The case of the “open source soundscapes” approach to identify, assess and plan “everyday quiet areas” in cities. Noise Mapp. 2016, 5, 1–20. [Google Scholar] [CrossRef] [Green Version]
  14. Zamora, W.; Vera, E.; Calafate, C.T.; Cano, J.C.; Manzoni, P. GRC-Sensing: An Architecture to Measure Acoustic Pollution Based on Crowdsensing. Sensors 2018, 18, 2596. [Google Scholar] [CrossRef] [Green Version]
  15. Brambilla, G.; Pedrielli, F. Smartphone-Based Participatory Soundscape Mapping for a More Sustainable Acoustic Environment. Sustainability 2020, 12, 7899. [Google Scholar] [CrossRef]
  16. Hachem, S.; Mallet, V.; Ventura, R.; Pathak, A.; Issarny, V.; Raverdy, P.; Bhatia, R. Monitoring Noise Pollution Using the Urban Civics Middleware. In Proceedings of the 2015 IEEE First International Conference on Big Data Computing Service and Applications, Redwood City, CA, USA, 30 March–2 April 2015; pp. 52–61. [Google Scholar] [CrossRef]
  17. Zappatore, M.; Longo, A.; Bochicchio, M.A. Crowd-sensing our Smart Cities: A Platform for Noise Monitoring and Acoustic Urban Planning. J. Commun. Softw. Syst. 2017, 13, 53–67. [Google Scholar] [CrossRef] [Green Version]
  18. Picaut, J.; Fortin, N.; Bocher, E.; Petit, G.; Aumond, P.; Guillaume, G. An open-science crowdsourcing approach for producing community noise maps using smartphones. Build. Environ. 2019, 148, 20–33. [Google Scholar] [CrossRef]
  19. ENERGIC OD—Official Website. Available online: https://www.energic-od.eu/ (accessed on 18 March 2021).
  20. Noise-Planet Website. Noise-Planet—Data. 2021. Available online: https://data.noise-planet.org/index.html (accessed on 13 January 2021).
  21. Wikipedia. Mobile Operating System. 2021. Available online: https://en.wikipedia.org/w/index.php?title=Mobile_operating_system&oldid=998546997 (accessed on 13 January 2021).
  22. Kardous, C.A.; Shaw, P.B. Evaluation of smartphone sound measurement applications. J. Acoust. Soc. Am. 2014, 135, EL186–EL192. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. NoiseCapture—Applications sur Google Play. Available online: https://play.google.com/store/apps/details?id=org.noise_planet.noisecapture (accessed on 18 January 2021).
  24. Can, A.; Gauvreau, B. Describing and classifying urban sound environments with a relevant set of physical indicators. J. Acoust. Soc. Am. 2015, 137, 208–218. [Google Scholar] [CrossRef] [PubMed]
  25. Bocher, E.; Petit, G.; Picaut, J.; Fortin, N.; Guillaume, G. Collaborative noise data collected from smartphones. Data Brief 2017, 14, 498–503. [Google Scholar] [CrossRef]
  26. Noise-Planet Website. NoiseCapture Interactive Community Map. 2021. Available online: https://noise-planet.org/map_noisecapture/index.html (accessed on 13 January 2021).
  27. Releases Ifsttar/NoiseCapture. Available online: https://github.com/Ifsttar/NoiseCapture/releases (accessed on 18 January 2021).
  28. Ifsttar/NoiseCapture: Measurement Export File, from Smartphone to Database. Available online: https://github.com/Ifsttar/NoiseCapture/blame/master/app/src/main/java/org/noise_planet/noisecapture/MeasurementExport.java#L136 (accessed on 18 January 2021).
  29. PostGIS—Spatial and Geographic Objects for PostgreSQL. Available online: https://postgis.net/ (accessed on 14 January 2021).
  30. Group, P.G.D. PostgreSQL. 2021. Available online: https://www.postgresql.org/ (accessed on 14 January 2021).
  31. Picaut, J.; Fortin, N.; Bocher, E.; Petit, G. Université Gustave Eiffel, NoiseCapture Dataverse Repository, NoiseCapture Data Extraction from 29 August 2017 until 28 August 2020 (3 Years), V1. 2021. Available online: https://0-doi-org.brum.beds.ac.uk/10.25578/J5DG3W (accessed on 18 July 2021).
  32. Open Data Commons Open Database License (ODbL) v1.0—Open Data Commons: Legal Tools for Open Data. Available online: https://opendatacommons.org/licenses/odbl/1-0/ (accessed on 18 March 2021).
  33. Noise-Planet—Scientific tools for environmental noise assessment. Available online: https://noise-planet.org/ (accessed on 18 March 2021).
  34. Mapping Parties—OpenStreetMap Wiki. Available online: https://wiki.openstreetmap.org/wiki/Mapping_parties (accessed on 15 January 2021).
  35. Zhu, Y.; Li, J.; Liu, L.; Tham, C.K. iCal: Intervention-free Calibration for Measuring Noise with Smartphones. In Proceedings of the 2015 IEEE 21st International Conference on Parallel and Distributed Systems, Melbourne, Australia, 14–17 December 2015; pp. 85–91. [Google Scholar] [CrossRef]
  36. Ventura, R.; Mallet, V.; Issarny, V.; Raverdy, P.G.; Rebhi, F. Evaluation and calibration of mobile phones for noise monitoring application. J. Acoust. Soc. Am. 2017, 142, 3084–3093. [Google Scholar] [CrossRef]
  37. Murphy, E.; King, E.A. Testing the accuracy of smartphones and sound level meter applications for measuring environmental noise. Appl. Acoust. 2016, 106, 16–22. [Google Scholar] [CrossRef] [Green Version]
  38. Android 11 Compatibility Definition. Available online: https://source.android.com/compatibility/android-cdd (accessed on 10 February 2021).
  39. Android vs. iPhone Users: The Difference between Behavior. 2017. Available online: https://buildfire.com/ios-android-users/ (accessed on 24 February 2021).
  40. Global Stats. Mobil Vendor Market Share Dec2019-Dec2020. 2021. Available online: https://gs.statcounter.com/vendor-market-share/mobile/ (accessed on 27 January 2021).
  41. The PostGIS Development Group. ‘ST Extent’ Function. Available online: https://postgis.net/docs/ST_Extent.html (accessed on 9 February 2021).
  42. GADM Data. Available online: https://gadm.org/data.html (accessed on 9 February 2021).
  43. List of Countries and Dependencies by Population. 2021. Available online: https://en.wikipedia.org/wiki/List_of_countries_and_dependencies_by_population (accessed on 18 January 2021).
  44. Android Developers. ‘getAccuracy’ Function. Available online: https://developer.android.com/reference/android/location/Location#getAccuracy() (accessed on 17 February 2021).
  45. Aumond, P.; Can, A.; Mallet, V.; De Coensel, B.; Ribeiro, C.; Botteldooren, D.; Lavandier, C. Kriging-based spatial interpolation from measurements for sound level mapping in urban areas. J. Acoust. Soc. Am. 2018, 143, 2847–2857. [Google Scholar] [CrossRef] [Green Version]
  46. Android Developers. ’getSpeed’ Function. Available online: https://developer.android.com/reference/android/location/Location#getSpeed() (accessed on 17 February 2021).
  47. Bedogni, L.; DiźFelice, M.; Bononi, L. Context-aware Android applications through transportation mode detection techniques. Wirel. Commun. Mob. Comput. 2016, 16, 2523–2541. [Google Scholar] [CrossRef] [Green Version]
  48. Aumond, P.; Lavandier, C.; Ribeiro, C.; Boix, E.G.; Kambona, K.; D’Hondt, E.; Delaitre, P. A study of the accuracy of mobile technology for measuring urban noise pollution in large scale participatory sensing campaigns. Appl. Acoust. 2017, 117, 219–226. [Google Scholar] [CrossRef]
  49. Zamora, W.; Calafate, C.T.; Cano, J.C.; Manzoni, P. Accurate Ambient Noise Assessment Using Smartphones. Sensors 2017, 17, 917. [Google Scholar] [CrossRef] [Green Version]
  50. Garg, S.; Lim, K.M.; Lee, H.P. An averaging method for accurately calibrating smartphone microphones for environmental noise measurement. Appl. Acoust. 2019, 143, 222–228. [Google Scholar] [CrossRef]
  51. Kardous, C.A.; Shaw, P.B. Evaluation of smartphone sound measurement applications (apps) using external microphones—A follow-up study. J. Acoust. Soc. Am. 2016, 140, EL327–EL333. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  52. Aumond, P.; Can, A.; Rey Gozalo, G.; Fortin, N.; Suárez, E. Method for in situ acoustic calibration of smartphone-based sound measurement applications. Appl. Acoust. 2020, 166, 107337. [Google Scholar] [CrossRef]
  53. Dubey, R.; Bharadwaj, S.; Zafar, M.; Sharma, V.; Biswas, S. Collaborative noise mapping using smartphone. ISPRS Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2020, 43, 253–260. [Google Scholar] [CrossRef]
  54. Graziuso, G.; Grimaldi, M.; Mancini, S.; Quartieri, J.; Guarnaccia, C. Crowdsourcing Data for the Elaboration of Noise Maps: A Methodological Proposal. J. Phys. Conf. Ser. 2020, 1603, 012030. [Google Scholar] [CrossRef]
  55. Mohammed, H.M.E.H.S.; Badawy, S.S.I.; Hussien, A.I.H.; Gorgy, A.A.F. Assessment of noise pollution and its effect on patients undergoing surgeries under regional anesthesia, is it time to incorporate noise monitoring to anesthesia monitors: An observational cohort study. Ain Shams J. Anesthesiol. 2020, 12, 20. [Google Scholar] [CrossRef]
  56. Sakagami, K. How did “state of emergency” declaration in Japan due to the COVID-19 pandemic affect the acoustic environment in a rather quiet residential area? UCL Open Environ. Prepr. 2020. [Google Scholar] [CrossRef]
  57. Nourmohammadi, Z.; Lilasathapornkit, T.; Ashfaq, M.; Gu, Z.; Saberi, M. Mapping Urban Environmental Performance with Emerging Data Sources: A Case of Urban Greenery and Traffic Noise in Sydney, Australia. Sustainability 2021, 13, 605. [Google Scholar] [CrossRef]
  58. Chon, Y.; Lane, N.D.; Li, F.; Cha, H.; Zhao, F. Automatically characterizing places with opportunistic crowdsensing using smartphones. In Proceedings of the 2012 ACM Conference on Ubiquitous Computing, Pittsburgh, PA, USA, 5–8 September 2012; pp. 481–490. [Google Scholar] [CrossRef]
  59. Hwang, K.; Lee, S. Environmental audio scene and activity recognition through mobile-based crowdsourcing. IEEE Trans. Consum. Electron. 2012, 58, 700–705. [Google Scholar] [CrossRef]
  60. Kendrick, P.; Cox, T.; Li, F.; Fazenda, B.; Jackson, I. Wind-induced microphone noise detection—Automatically monitoring the audio quality of field recordings. In Proceedings of the 2013 IEEE International Conference on Multimedia and Expo (ICME), San Jose, CA, USA, 15–19 July 2013; pp. 1–6. [Google Scholar] [CrossRef]
  61. Shoaib, M.; Bosch, S.; Incel, O.D.; Scholten, H.; Havinga, P.J.M. A Survey of Online Activity Recognition Using Mobile Phones. Sensors 2015, 15, 2059–2085. [Google Scholar] [CrossRef] [PubMed]
  62. Wang, W.; Chang, Q.; Li, Q.; Shi, Z.; Chen, W. Indoor-Outdoor Detection Using a Smart Phone Sensor. Sensors 2016, 16, 1563. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  63. DetectedActivity|Google APIs for Android. Available online: https://developers.google.com/android/reference/com/google/android/gms/location/DetectedActivity (accessed on 20 January 2021).
  64. Gontier, F.; Lavandier, C.; Aumond, P.; Lagrange, M.; Petiot, J.F. Estimation of the Perceived Time of Presence of Sources in Urban Acoustic Environments Using Deep Learning Techniques. Acta Acust. United Acust. 2019, 105. [Google Scholar] [CrossRef] [Green Version]
  65. Silva, M.D.; Viterbo, J.; Bernardini, F.; Maciel, C. Identifying Privacy Functional Requirements for Crowdsourcing Applications in Smart Cities. In Proceedings of the 2018 IEEE International Conference on Intelligence and Security Informatics (ISI), Miami, FL, USA, 9–11 November 2018; pp. 106–111. [Google Scholar] [CrossRef]
  66. EUR-Lex—32016R0679-EN-EUR-Lex: Regulation (EU) 2016/679 of the European Parliament and of the Council of 27 April 2016 on the Protection of Natural Persons with Regard to the Processing of Personal Data and on the Free Movement of Such Data, and Repealing Directive 95/46/EC (General Data Protection Regulation). Available online: https://eur-lex.europa.eu/eli/reg/2016/679/oj (accessed on 18 February 2021).
  67. D’Hondt, E.; Stevens, M.; Jacobs, A. Participatory noise mapping works! An evaluation of participatory sensing as an alternative to standard techniques for environmental monitoring. Pervasive Mob. Comput. 2013, 9, 681–694. [Google Scholar] [CrossRef]
  68. Martí, I.G.; Rodríguez, L.E.; Benedito, M.; Trilles, S.; Beltrán, A.; Díaz, L.; Huerta, J. Mobile Application for Noise Pollution Monitoring through Gamification Techniques. Entertainment Computing—ICEC 2012; Herrlich, M., Malaka, R., Masuch, M., Eds.; Springer: Berlin/Heidelberg, Germany, 2012; pp. 562–571. [Google Scholar] [CrossRef] [Green Version]
  69. Lefevre, B.; Agarwal, R.; Issarny, V.; Mallet, V. Mobile crowd-sensing as a resource for contextualized urban public policies: A study using three use cases on noise and soundscape monitoring. Cities Health 2019, 1–19. [Google Scholar] [CrossRef] [Green Version]
  70. Lee, H.P.; Garg, S.; Lim, K.M. Crowdsourcing of environmental noise map using calibrated smartphones. Appl. Acoust. 2020, 160, 107130. [Google Scholar] [CrossRef]
Figure 1. NoiseCapture Android application main screens. From top/down and left/right: Measurement, Description, Results, Map.
Figure 1. NoiseCapture Android application main screens. From top/down and left/right: Measurement, Description, Results, Map.
Ijerph 18 07777 g001
Figure 2. Screenshots of the NoiseCapture map website. Example of collected data representation, on the Lyon French city (from https://noise-planet.org/map_noisecapture/index.html#18/45.75387/4.84052/ (accessed on 15 July 2021)), at a higher (up) and lower (down) zoom levels.
Figure 2. Screenshots of the NoiseCapture map website. Example of collected data representation, on the Lyon French city (from https://noise-planet.org/map_noisecapture/index.html#18/45.75387/4.84052/ (accessed on 15 July 2021)), at a higher (up) and lower (down) zoom levels.
Ijerph 18 07777 g002
Figure 3. Weekly evolution of user installs and uninstalls of the NoiseCapture application (data from Google Play dashboard): (a) global data and (b) data for France and United states of America. ‘Installs’: number of users who have installed the application at least on one device; ‘Uninstalls’: users who have uninstalled the application from all their devices; ‘Active devices’: number of active devices that contains application, and which was turned on at least once in the previous 30 days; ‘Contributors’: Users who have upload data to the NoiseCapture remote server; ‘Events’: Events that may have a particular impact on the users behavior. (a) Application installs for France (FR) and United-States of America (US); (b) Application installs/uninstalls/contributors.
Figure 3. Weekly evolution of user installs and uninstalls of the NoiseCapture application (data from Google Play dashboard): (a) global data and (b) data for France and United states of America. ‘Installs’: number of users who have installed the application at least on one device; ‘Uninstalls’: users who have uninstalled the application from all their devices; ‘Active devices’: number of active devices that contains application, and which was turned on at least once in the previous 30 days; ‘Contributors’: Users who have upload data to the NoiseCapture remote server; ‘Events’: Events that may have a particular impact on the users behavior. (a) Application installs for France (FR) and United-States of America (US); (b) Application installs/uninstalls/contributors.
Ijerph 18 07777 g003
Figure 4. Weekly ratio between the number of contributors to the NoiseCapture database and NoiseCapture installs.
Figure 4. Weekly ratio between the number of contributors to the NoiseCapture database and NoiseCapture installs.
Ijerph 18 07777 g004
Figure 5. Entity relation diagrams (ERD) of the NoiseCapture PostGIS database. The type of each field in the tables is mentioned: ‘123’ for ‘float’ values, ‘ABC’ for ‘text’ chains, ‘timestamptz’ for time stamp date, ‘🗹’ for ‘boolean’ value. The ‘key’ yellow symbol is used to display primary keys of the table, whose corresponding names are displayed in bold.
Figure 5. Entity relation diagrams (ERD) of the NoiseCapture PostGIS database. The type of each field in the tables is mentioned: ‘123’ for ‘float’ values, ‘ABC’ for ‘text’ chains, ‘timestamptz’ for time stamp date, ‘🗹’ for ‘boolean’ value. The ‘key’ yellow symbol is used to display primary keys of the table, whose corresponding names are displayed in bold.
Ijerph 18 07777 g005
Figure 6. Distribution of the number of occurrences of a smartphone model in the database. As an example, 775 device models appear more than 50 times in the database.
Figure 6. Distribution of the number of occurrences of a smartphone model in the database. As an example, 775 device models appear more than 50 times in the database.
Ijerph 18 07777 g006
Figure 7. Distribution of the tags used for tracks collected by user who have 1 only contribution to the database.
Figure 7. Distribution of the tags used for tracks collected by user who have 1 only contribution to the database.
Ijerph 18 07777 g007
Figure 8. Evolution of the median value of the accuracy of geolocalization (for all measurement points with geolocalization) in function of time, since the first second of measurement.
Figure 8. Evolution of the median value of the accuracy of geolocalization (for all measurement points with geolocalization) in function of time, since the first second of measurement.
Ijerph 18 07777 g008
Figure 9. Distribution of tracks per hour of the day, for the entire database and for the data collected in France only. (a) Distribution of tracks per hour of the day, for the whole database. (b) Distribution of tracks per hour of the day for data collected in France only.
Figure 9. Distribution of tracks per hour of the day, for the entire database and for the data collected in France only. (a) Distribution of tracks per hour of the day, for the whole database. (b) Distribution of tracks per hour of the day for data collected in France only.
Ijerph 18 07777 g009
Figure 10. Distribution of track measurements in function of the day and the month, over the 3 years of the collected data. (a) Distribution of tracks per day of the week. (b) Distribution of tracks per month.
Figure 10. Distribution of track measurements in function of the day and the month, over the 3 years of the collected data. (a) Distribution of tracks per day of the week. (b) Distribution of tracks per month.
Ijerph 18 07777 g010aIjerph 18 07777 g010b
Figure 11. Distribution of tracks in function of the duration of the measurement (‘time_length’).
Figure 11. Distribution of tracks in function of the duration of the measurement (‘time_length’).
Ijerph 18 07777 g011
Figure 12. Number of tracks in function of ‘gain_calibration’ value.
Figure 12. Number of tracks in function of ‘gain_calibration’ value.
Ijerph 18 07777 g012
Figure 13. User evaluation of pleasantness on a track: (a) Distribution of pleasantness values on tracks. (b) Distribution of number of pleasantness levels used by contributors.
Figure 13. User evaluation of pleasantness on a track: (a) Distribution of pleasantness values on tracks. (b) Distribution of number of pleasantness levels used by contributors.
Ijerph 18 07777 g013aIjerph 18 07777 g013b
Figure 14. Use of soundscape tags by contributors. (a) Number of tags simultaneously used in a track. (b) Tags name.
Figure 14. Use of soundscape tags by contributors. (a) Number of tags simultaneously used in a track. (b) Tags name.
Ijerph 18 07777 g014aIjerph 18 07777 g014b
Figure 15. Distribution of noise levels (a) on the 260,422 tracks and (b) on the 59,685,328 points.
Figure 15. Distribution of noise levels (a) on the 260,422 tracks and (b) on the 59,685,328 points.
Ijerph 18 07777 g015
Figure 16. ‘Heatmap’ representation of the NoiseCapture data collected around the world.
Figure 16. ‘Heatmap’ representation of the NoiseCapture data collected around the world.
Ijerph 18 07777 g016
Figure 17. Distribution of mean ‘noise_level’ in France by week accompanied with 2 verticals lines that represent the start and end of the first lockdown.
Figure 17. Distribution of mean ‘noise_level’ in France by week accompanied with 2 verticals lines that represent the start and end of the first lockdown.
Ijerph 18 07777 g017
Table 1. NoiseCapture application releases. Each new version, defined by a version release (for example ‘51’) and a version number (for example ‘1.2.15’) proposes changes (bug corrections, user interface enhancement, etc.). The reader can refer to the detailed list of fixes in each (pre-)release from the GitHub source code management platform [27]. The changes made on the data export, from the smartphone to the data server (adding new data, patches, etc.), are detailed in the history of the source code file MeasurementExport.java [28]. The date of publication of the application on Google Play is also provided for information (corresponding official public release are indicated in bold with symbol *).
Table 1. NoiseCapture application releases. Each new version, defined by a version release (for example ‘51’) and a version number (for example ‘1.2.15’) proposes changes (bug corrections, user interface enhancement, etc.). The reader can refer to the detailed list of fixes in each (pre-)release from the GitHub source code management platform [27]. The changes made on the data export, from the smartphone to the data server (adding new data, patches, etc.), are detailed in the history of the source code file MeasurementExport.java [28]. The date of publication of the application on Google Play is also provided for information (corresponding official public release are indicated in bold with symbol *).
Release (Number)Source Code PublicationSTATUSApplication PublicationComments
51 * (1.2.15)3 July 2020Release7 July 2020Fix automated measurement upload
49 (1.2.13)17 February 2020Pre-release Add calibration method using road traffic
Add ‘calibration_method’
45 * (1.2.9)27 March 2019Release26 March 2019Calibration in L A , eq instead of L eq
43 * (1.2.7)16 November 2018Release16 November 2018Minor changes
35 * (1.1.3) Release20 April 2018Minor changes
34 * (1.1.2) Release29 January 2018Minor changes
33 * (1.1.0)23 November 2017Pre-release24 November 2017Ability to use a calibrated sMarchtphone
to automatically calibrate other sMarchtphone(s)
32 * (1.0.4) Release6 November 2017Minor changes
31 * (1.0.3) Release6 October 2017Minor changes
30 (1.0.2)18 September 2017Pre-release Add NoiseCapture Party functionalities
29 * (1.0.1) Release31 August 2017Minor changes
28 * (1.0.0)23 August 2017Release29 August 2017Official first release
Add ‘user_profile’
Table 2. Distribution of collected data per release, from 29 August 2017 to 28 August 2020. The number of collected data during NoiseCapture Parties (see Section 3.6) is also indicated, as well as in terms of percentage of the total number of data. Releases in bold with symbol * correspond to public releases on Google Play. Note that the total number of contributors in this table (74,082) does not correspond to the total of unique contributors, as a contributor may have use several release of the application.
Table 2. Distribution of collected data per release, from 29 August 2017 to 28 August 2020. The number of collected data during NoiseCapture Parties (see Section 3.6) is also indicated, as well as in terms of percentage of the total number of data. Releases in bold with symbol * correspond to public releases on Google Play. Note that the total number of contributors in this table (74,082) does not correspond to the total of unique contributors, as a contributor may have use several release of the application.
ReleaseContributorsTracksPoints
TotalPartyTotalPartyTotalParty
28 *2635446,268
29 *270589911,588,156
30357 (20.0%)416133 (32.0%)140,62711,523 (8.2%)
31 *14322 (0.1%)44266 (0.1%)957,920770 (0.1%)
32 *15533 (0.1%)47469 (0.2%)847,0931556 (0.2%)
33 *54424 (0.1%)19,22552 (0.3%)3,530,3498819 (0.2%)
34 *905313 (0.1%)28,60767 (0.2%)6,121,15418,793 (0.3%)
35 *15,73467 (0.4%)68,911921 (1.3%)12,465,115117,797 (0.9%)
36461861
37115520,732
38371774
3922108
40112
411197
421110
43 *11,16982 (0.7%)37,960643 (1.7%)9,309,93489,794 (1.0%)
446334276
45 *23,331183 (0.8%)67,7651306 (1.9%)18,629,005142,168 (0.8%)
46375030
4741 (25.0%)61 (16.6%)21,101134 (0.6%)
483128597
4925233235,882
50440 0.0%)146
51 *35342 (0.1%)18,6534 (0.02%)5,750,091221 (0.003%)
Total74,082364 (0.5%)260,4223142 (1.2%)59,685,328391,575 (0.6%)
Table 3. Collected device information for the Samsung Galaxy A10 Android phone, as well as, the count of corresponding phones in the NoiseCapture database.
Table 3. Collected device information for the Samsung Galaxy A10 Android phone, as well as, the count of corresponding phones in the NoiseCapture database.
‘device_model’‘device_product’‘device_manufacturer’Count
SM-A102Na10ekxsamsung1
SM-A102Ua10esqsamsung880
SM-A102U1a10euesamsung1
SM-A102Wa10ecssamsung4
SM-A105Fa10ddsamsung53
SM-A105FNa10eeasamsung176
SM-A105Ga10dxsamsung31
SM-A105Ma10ubsamsung97
SM-A107Fa10sxxsamsung508
SM-A107Ma10subsamsung23
Table 4. Top 15 of smartphone manufacturers (‘device_manufacturer’, case insensitive) in the NoiseCapture database. The number of corresponding distinct device models (‘device_model’), the number of tracks, as well as the cumulative number of tracks are also given. Note that this table do not regroup data from the same manufacturer but with a different writing (upper/lower case, as for ‘Samsung’ and ‘samsung’).
Table 4. Top 15 of smartphone manufacturers (‘device_manufacturer’, case insensitive) in the NoiseCapture database. The number of corresponding distinct device models (‘device_model’), the number of tracks, as well as the cumulative number of tracks are also given. Note that this table do not regroup data from the same manufacturer but with a different writing (upper/lower case, as for ‘Samsung’ and ‘samsung’).
RankDevice_MANUFACTURERNb of ModelsNb of Tracks%Cumul. Nb of Tracks%
1samsung1032101,42038.9%101,42038.9%
2LGE38336,28813.9%137,70852.9%
3HUAWEI45431,93712.2%169,64565.1%
4motorola12617,8226.8%187,46771.9%
5ZTE17112,8404.9%200,30776.9%
6Xiaomi10663342.4%206,64179.3%
7TCL19151802.0%211,82181.3%
8Sony16751162.0%216,93783.3%
9OPPO9137421.4%220,67984.7%
10WIKO7332231.2%223,90286.0%
11asus11528701.1%226,77287.1%
12HTC13024060.9%229,17888.0%
13HMD Global4122080.8%231,38688.8%
14LENOVO10818810.7%235,06589.6%
15OnePlus3217980.7%235,06590.3%
Table 5. Top 5 of smartphone model (‘device_model’) in the NoiseCapture database.
Table 5. Top 5 of smartphone model (‘device_model’) in the NoiseCapture database.
RankDevice_MODEL‘Device_MANUFACTURER’Nb of Tracks%Cumulative Nb of Tracks%
1ANE-LX3samsung47291.8%47291.8%
2SM-G930Fsamsung37221.4%84513.2%
3LM-X210(G)LGE34791.3%11,9304.6%
4SM-A520Fsamsung32051.2%15,1355.8%
5Z982ZTE28901.1%18,0256.9%
6SM-G935Fsamsung28541.1%20,8798.0%
7Moto E (4)motorola27481.0%23,6279.1%
8SM-N950Usamsung23840.9%26,0119.9%
9moto e5 playmotorola23610.9%28,37210.9%
10SM-G950Fsamsung23010.9%30,67311.8%
11SM-J327T1samsung22970.9%32,97012.6%
12LGMP260LGE22130.8%35,18313.5%
13SM-S327VLsamsung22000.8%37,38314.3%
14VTR-L09samsung20950.8%39,47815.1%
15SM-J727T1HUAWEI20480.8%41,52615.9%
Table 6. Top 6 smartphone manufacturers (‘device_manufacturer’, case-insensitive) for USA and France between August 2017 and August 2020. Top 6 manufacturers data are from Statcounter Global Stats website (licensed under a Creative Commons Attribution-Share Alike 3.0 Unported License) [40].
Table 6. Top 6 smartphone manufacturers (‘device_manufacturer’, case-insensitive) for USA and France between August 2017 and August 2020. Top 6 manufacturers data are from Statcounter Global Stats website (licensed under a Creative Commons Attribution-Share Alike 3.0 Unported License) [40].
CountryDevice_MANUFACTURERNumber of Tracks%Top 6 Manufacturer in Country
United Statessamsung32,34136.6%IPhone
LGE23,66826.8%Samsung
motorola925710.5%LGE
ZTE81549.2%Motorola
TCL21172.4%Google
Alcatel12891.4%ZTE
Francesamsung12,89946.2%Samsung
HUAWEI486417.4%IPhone
WIKO17316.2%HUAWEI
Xiaomi12944.6%Sony
Sony12544.5%Xiaomi
motorola10933.9%WIKO
Table 7. Distribution of the number of contributors in function of the number of track measurements. The number of corresponding points is also given.
Table 7. Distribution of the number of contributors in function of the number of track measurements. The number of corresponding points is also given.
Number of TracksNumber of Contributors%Number of Points%
136,40552.0%8,709,87214.6%
2–1030,04343.0%24,106,57840.1%
11–5030634.4%14,779,03324.7%
51–1002360.3%3,915,3106.5%
101–10001430.2%8,016,51713.4%
>100080.1%158,0180.7%
Total69,898 59,685,328
Table 8. Distribution of the time length (in second) in function of the number of track measurements for tracks collected by user who had 1 contribution to the database.
Table 8. Distribution of the time length (in second) in function of the number of track measurements for tracks collected by user who had 1 contribution to the database.
Time LengthNumber of Tracks%
1–2014,03438.5%
21–60945226.0%
61–300924225.4%
301–60015374.2%
601–9005341.5%
901–12003150.8%
1201–18003681.0%
1801–24002130.6%
2401–30001110.3%
3001–36001030.3%
>36004961.4%
Total36,405
Table 9. Duration between two successive measurements for users who have more than one contribution.
Table 9. Duration between two successive measurements for users who have more than one contribution.
Duration between 2 Successive Measurements (Day)Number of Tracks%
0158,63183.3%
171403.7%
2–711,8316.2%
8–1439112.0%
14–2119231.0%
21–3014500.8%
31–6022561.2%
61–9010630.6%
91–18012690.7%
181–3657680.4%
>3652820.1%
Total188,794100%
Table 10. Distribution of the collected data per country and ranking (first ranks are displayed in bold). Population per country (percentage of world population) data are from in [43].
Table 10. Distribution of the collected data per country and ranking (first ranks are displayed in bold). Population per country (percentage of world population) data are from in [43].
CountryPopulation (Rank)Contributors (Rank)Tracks (Rank)Points (Rank)Points/Track
China17.9% (1)58 (54)354 (43)158,797 (26)448.6
India17.5% (2)894 (4)2241 (16)243,778 (22)108.9
United States4.2% (3)29,108 (1)88,341 (1)22,676,833 (1)256.7
Indonesia3.4% (4)91 (45)199 (55)20,244 (65)101.7
Pakistan2.8% (5)124 (32)321 (45)27,035 (57)84.2
Brazil2.7% (6)448 (14)1503 (20)244,482 (21)162.6
Nigeria2.7% (7)33 (67)66 (78)9319 (75)141.2
Bangladesh2.1% (8)160 (26)572 (28)332,316 (19)581
Russia1.86% (9)174 (24)850 (24)94,277 (33)110.2
Germany1.1% (19)790 (7)3093 (7)1,216,164 (5)393.2
France0.9% (20)5516 (2)27,911 (2)4,972,054 (2)178.1
United Kingdom0.8% (21)1164 (3)4693 (4)2,067,182 (3)440.5
Canada0.5% (37)792 (6)2512 (15)1,551,808 (4)617.7
Peru0.4% (42)77 (48)11,231 (3)138,716 (27)12.3
Netherlands0.2% (67)435 (15)3409 (5)413,897 (16)121.4
Table 11. Distribution of accuracy for the point with geolocalization.
Table 11. Distribution of accuracy for the point with geolocalization.
‘accuracy’Number of Points%
000
0 , 1 420,1500.9
1 , 2 1,400,4662.8
2 , 3 3,574,0047.3
3 , 4 7,406,06015.1
4 , 5 5,268,84310.8
5 , 10 11,142,30522.8
10 , 15 5,973,00012.2
15 , 25 7,128,77314.6
25 , 35 1,725,1693.5
35 , 50 1,192,9082.4
50 , 100 1,426,4752.9
>1002,243,5664.6
Total48,901,719
Table 12. Distribution of ‘speed’ for points with geolocalization.
Table 12. Distribution of ‘speed’ for points with geolocalization.
‘speed’Number of Points%
031,986,10265.4%
0 , 1.4 837,66218.2%
1.4 , 4.2 2,058,5854.2%
>4.25,919,37012.2%
Total48,901,719
Table 13. Time shift (in hour) between record_utc and time_location, for the tracks are are 100% geolocalized.
Table 13. Time shift (in hour) between record_utc and time_location, for the tracks are are 100% geolocalized.
Time Shift (h)Number of Tracks%
<−247970.4
−241510.07
−23350.02%
−2240.002%
−2030.002%
−1980.004%
−1880.004%
−1510.0005%
−1410.0005%
−1330.002%
−12250.012%
−11170.008%
−1080.004%
−960.003%
−830.002%
−7100.005%
−670.004%
−560.003%
−4100.005%
−3670.034%
−2470.023%
−12580.13%
0194,69098.54%
14790.242
2540.027%
3280.013%
4150.008
5100.005%
660.003%
760.003%
8100.005%
930.002%
10100.005%
1140.002%
1210.0005%
1330.002%
1420.001%
1580.004%
16110.006%
1880.004%
1940.002%
2050.002%
2120.001%
2230.002%
23180.009%
24110.006%
>247020.355%
Total197,568
Table 14. Total number of collected data and distribution per time length and part of the measurement tracks and points that have been collected in motion (only data with geolocalization are considered, with a speed greater than 0.5 m/s).
Table 14. Total number of collected data and distribution per time length and part of the measurement tracks and points that have been collected in motion (only data with geolocalization are considered, with a speed greater than 0.5 m/s).
‘time_length’TracksPoints
TotalMovingTotalMoving
1 , 3 min21,9634477 (20.4%)679,675437,640 (64.4%)
3 , 5 min71901575 (21.9%)516,389343,035 (66.4%)
5 , 10 min72042029 (28.16%)1,080,596769,976 (71.2%)
10 , 20 min52311678 (32.1%)1,661,7531,189,641 (71.6%)
20 , 30 min2083752 (36.1%)1,237,893925,499 (74.7%%)
30 , 60 min2362741 (31.4)2,094,6251,446,091 (69.0%)
1 , 5 h2026466 (22.9%)3,694,6831,912,894 (51.7%)
>5 h14219 (13.4%)752,485189,075 (25.1%)
Table 15. Distribution of tracks in function of the calibration method, before and after release 49 (’n.a.’ for ‘not available’).
Table 15. Distribution of tracks in function of the calibration method, before and after release 49 (’n.a.’ for ‘not available’).
‘calibration_method’Since ReleaseNb of Tracks before R49 (%)Nb of Tracks Since R49 (%)
CalibratedSmartPhone33n.a.277 (1.4%)
Calibrator28n.a.139 (0.7%)
ManualSetting28n.a.838 (4.4%)
None28241,53217,395 (92.1%)
Reference28n.a.167 (0.9%)
Traffic49n.a.74 (0.5%)
Total 241,53218,890
Table 16. Cross table of the number of collected tracks in function of the calibration method and the calibration gain (data from release 49).
Table 16. Cross table of the number of collected tracks in function of the calibration method and the calibration gain (data from release 49).
Calibration Method/Gain<−10 dB[−10, −5] dB[−5, 0] dB0 dB[0, 5] dB[5, 10] dB>10 dBTotal
CalibratedSmartPhone-61 (22.0%)136 (49.1%)-75 (27.1%)3 (1.1%)2 (0.7%)277
Calibrator1 (0.7%)2 (1.4%)3 (2.2%)-7 (5.0%)-126 (90.7%)139
ManualSetting56 (6.7%)13 (1.6%)150 (17.9%)126 (15.%)156 (18.6%)54 (6.4%)283 (33.8%)838
None559 (2.4%)911 (5.3%)865 (5.1%)13,857 (79.9%)485 (2.9%)227 (1.5%)491 (2.9%)17,395
Reference9 (5.4%)21 (12.6%)38 (22.7%)23 (13.8%)57 (34.1%)7 (4.2%)12 (7.2%)167
Traffic20 (27.0%)14 (18.9%)28 (37.8%)-6 (8.1%)5 (6.8%)1 (1.4%)74
Total6451022122014,00678629691518,890
Table 17. List of NoiseCapture Parties. More information are located in the ‘noisecapture_party’ table of the database. The number of contributors, as well as the total of collected tracks and points are given. The organization in charge of the NoiseCapture Party is also mentioned (Noise-Planet is the organization in charge of the development of NoiseCapture application). While the CICAM NoiseCapture Party has been planned, it was canceled due to the pandemic situation (i.e., there are no corresponding tracks).
Table 17. List of NoiseCapture Parties. More information are located in the ‘noisecapture_party’ table of the database. The number of contributors, as well as the total of collected tracks and points are given. The organization in charge of the NoiseCapture Party is also mentioned (Noise-Planet is the organization in charge of the development of NoiseCapture application). While the CICAM NoiseCapture Party has been planned, it was canceled due to the pandemic situation (i.e., there are no corresponding tracks).
‘pk_party’‘tag’Organization‘filter_area’‘filter_time’ContributorsTracksPoints
1SNDIGITALWEEKNoise-PlanetTRUEFALSE713311,523
2ANQESNoise-PlanetTRUETRUE4294479
3FDS2017Noise-PlanetTRUETRUE261239
5IMS2018Noise-PlanetTRUEFALSE136718,793
6UDCUniversidade da CoruñaTRUETRUE8566879
9TEST44Noise-PlanetTRUETRUE1391
10UNISAUniversity of Salerno, ItalyTRUETRUE1314915,912
11PNRGMNoise-PlanetTRUETRUE2136089
12AMSOUNDSWaag Technology & SocietyTRUETRUE218693
13PNRGMParc Naturel du MorbihanTRUETRUE1410021,470
14FDSSTRASNoise-PlanetTRUETRUE5312967
15AGGLOBASTIANoise-PlanetFALSETRUE1950759,838
17FDSNTSNoise-PlanetTRUETRUE7665916
18H2020Noise-PlanetTRUETRUE118922,060
19UDCUniversidade da Coruña, SpainFALSETRUE201385866
20MSANoise-PlanetTRUETRUE991885
21GEO2019Noise-PlanetTRUETRUE4342063,521
22IMS2019Noise-PlanetTRUETRUE2319217,309
23FPSLYONoise-PlanetTRUETRUE113410,285
24SSSOROLL2019Generalitat de CatalunyaFALSETRUE6837236,272
26UNISAUniversity of Salerno, ItalyTRUETRUE2033223,220
27FDSSTRASNoise-PlanetTRUETRUE371771
28H2020Noise-PlanetTRUETRUE93932,948
29UDCUniversidade da Coruña, SpainFALSETRUE9732099
30MSANoise-PlanetTRUETRUE10103665
31CICAMEPN, Quito, EcuadorFALSETRUE---
32UDC_COVIDUniversidade da Coruña, SpainTRUETRUE3324914,785
Table 18. Cross table between pleasantness and tags.
Table 18. Cross table between pleasantness and tags.
Tag/PleasantnessUsedNot Used
Used45,549 (17.5%)78,814 (30.2%)
Not used9457 (3.6%)126,602 (48.7%)
Table 19. Tags description: ‘pk_tag’ and ‘tag_name’ are the primary key and the name of the tags. The ‘Description’ correspond to the name of the tag in the corresponding NoiseCapture screen.
Table 19. Tags description: ‘pk_tag’ and ‘tag_name’ are the primary key and the name of the tags. The ‘Description’ correspond to the name of the tag in the corresponding NoiseCapture screen.
CategoryMeasurement Conditions
‘pk_tag’161323
‘tag_name’testindoorrainwind
DescriptionTestIndoorRainWind
Human Activity Sources
‘pk_tag’18302028
‘tag_name’chattingchildrenfootstepsmusic
DescriptionVoiceChildrenFootstepsMusic
CategoryTransportation Sources
‘pk_tag’27322635
‘tag_name’roadrailair_trafficmarine_traffic
DescriptionRoadRailAir T.Marine
Natural Sources
‘pk_tag’343329
‘tag_name’wateranimalsvegetation
DescriptionWaterAnimalsVegetation
Mechanical Activity Sources
‘pk_tag’243631
‘tag_name’worksalarmsindustrial
DescriptionWorksAlarmsIndustrial
Table 20. Possible enhancements of the NoiseCapture application and database.
Table 20. Possible enhancements of the NoiseCapture application and database.
DataUncertainties/BiasPossible SourcesPossible Solutions
User profileProfile information is emptyCannot evaluate the expertise of the contributorIn the app: update the field during an app update if the field is empty.
GeolocalizationNo geolocalization of a trackThe geolocalization is turned offIn the app: add a message for turning on the geolocalization
Indoor measurementsIn the app: wait for future methodologies (Indoor positioning System) and high sensitivity GPS for indoor localization
No geolocalization of a point in a trackLocal loss of geolocalizationUse GIS methodologies to re-locate the point within the track
Inhomogeneous worldwide coverageNo access to Google PlayUse alternative app stores
AccuracyValue equal to ‘0.0’No geolocalizationIn the app: add a message for turning on the geolocalization
Extreme (not realistic) valuesUnknownNo known solutions
Large (but realistic) values In the app: ask contributors to wait for a better localization before starting the measurement
SpeedValue equal to ‘0.0’No geolocalizationIn the app: add a message for turning on the geolocalization
Negative valuesUnknownNo known solutions
No evaluation of the accuracy of the speed valueIn the app: use the Android function getSpeedAccuracyMetersPerSecond() to store this missing information.
TimestampWrong dateThe geolocalization is turned offIn the app: add a message for turning on the geolocalization
Wrong phone settingIn the app: check that the date is correct and add a message if not
CalibrationThe calibration method is not knownThe information about the selected calibration method is collected since the version 49 onlyNo solution
Extreme (not realistic) valuesNo calibration method usedIn the app: send a notification to calibrate the smartphone
In the app: check the calibration value and send a notication if the value seems incorrect
In the app/remote server: create a smartphone model calibration database
PleasantnessPossible bias at level 50%The default value is fixed at a pleasantness of 50%In the app: change the selection mode for the pleasantness without default value
Noise levelsExtreme (not realistic) valuesCalibration is not correctImprove the calibration of the smartphone
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Picaut, J.; Boumchich, A.; Bocher, E.; Fortin, N.; Petit, G.; Aumond, P. A Smartphone-Based Crowd-Sourced Database for Environmental Noise Assessment. Int. J. Environ. Res. Public Health 2021, 18, 7777. https://0-doi-org.brum.beds.ac.uk/10.3390/ijerph18157777

AMA Style

Picaut J, Boumchich A, Bocher E, Fortin N, Petit G, Aumond P. A Smartphone-Based Crowd-Sourced Database for Environmental Noise Assessment. International Journal of Environmental Research and Public Health. 2021; 18(15):7777. https://0-doi-org.brum.beds.ac.uk/10.3390/ijerph18157777

Chicago/Turabian Style

Picaut, Judicaël, Ayoub Boumchich, Erwan Bocher, Nicolas Fortin, Gwendall Petit, and Pierre Aumond. 2021. "A Smartphone-Based Crowd-Sourced Database for Environmental Noise Assessment" International Journal of Environmental Research and Public Health 18, no. 15: 7777. https://0-doi-org.brum.beds.ac.uk/10.3390/ijerph18157777

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop